Lewdiculous
commited on
Commit
•
4ec3888
1
Parent(s):
8f98c87
Update README.md
Browse files
README.md
CHANGED
@@ -31,7 +31,7 @@ GGUF-IQ-Imatrix quants for [NeverSleep/Llama-3-Lumimaid-8B-v0.1-OAS](https://hug
|
|
31 |
> If there are any issues or questions let me know.
|
32 |
|
33 |
> [!NOTE]
|
34 |
-
> For **8GB VRAM** GPUs, I recommend the **Q4_K_M-imat** quant for up to 12288 context sizes.
|
35 |
|
36 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/65d4cf2693a0a3744a27536c/JUxfdTot7v7LTdIGYyzYM.png)
|
37 |
|
|
|
31 |
> If there are any issues or questions let me know.
|
32 |
|
33 |
> [!NOTE]
|
34 |
+
> For **8GB VRAM** GPUs, I recommend the **Q4_K_M-imat** (4.89 BPW) quant for up to 12288 context sizes.
|
35 |
|
36 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/65d4cf2693a0a3744a27536c/JUxfdTot7v7LTdIGYyzYM.png)
|
37 |
|