Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
Custom GGUF Quants with iMatrix for :
|
2 |
+
https://huggingface.co/MarsupialAI/LaDameBlanche-v2-95b
|
3 |
+
|
4 |
+
- Q8_0 used as quant base : https://huggingface.co/mradermacher/LaDameBlanche-v2-95b-GGUF
|
5 |
+
- iMatrix here : https://huggingface.co/mradermacher/LaDameBlanche-v2-95b-i1-GGUF
|
6 |
+
|
7 |
+
(Yes, I'm lazy, but I can live with a 0.01ppl bump ^^)
|
8 |
+
|
9 |
+
The model is a great merge, sensical and creative, imho working better for lesser requirements than the 100b+ Miqu which are worthy only for those with 48GB VRAM or more.
|
10 |
+
|
11 |
+
In IQ2_LR(2.7BPW, for 8k context with 36GB VRAM and an IGP running the OS display), ARC Challenge at 57, ARC Easy at 77, PPL 512 at 4.5860.
|
12 |
+
|
13 |
+
Mesdames et messieurs, vous êtes servis!
|