MarsupialAI
commited on
Commit
•
434a9a4
1
Parent(s):
cd15117
Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,5 @@
|
|
|
|
|
|
1 |
Some folks are claiming there's something funky going on with GGUF quanting for Llama 3 models. I don't disagree.
|
2 |
|
3 |
Some of those people are speculating that it has something to do with converting the raw weights from bf16 to fp16 instead
|
@@ -22,4 +24,32 @@ None specified: size 29.9GB PPL @ ???? 9.5158 +/- 0.15418 PPL @ Q4km 9.62
|
|
22 |
As you can see, converting to fp32 has no meaningful effect on PPL compared to converting to fp16. PPL is identical at full weight,
|
23 |
and the miniscule loss shown at Q4km is will within the margin of error. There will no doubt be some people who will claim
|
24 |
"PpL iSn'T gOoD eNoUgH!!1!". For those people, I have uploaded all GGUFs used in this test. Feel free to use those files to do
|
25 |
-
more extensive testing on your own time. I consider the matter resolved until somebody can conclusively demonstrate otherwise.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Initial Testing
|
2 |
+
|
3 |
Some folks are claiming there's something funky going on with GGUF quanting for Llama 3 models. I don't disagree.
|
4 |
|
5 |
Some of those people are speculating that it has something to do with converting the raw weights from bf16 to fp16 instead
|
|
|
24 |
As you can see, converting to fp32 has no meaningful effect on PPL compared to converting to fp16. PPL is identical at full weight,
|
25 |
and the miniscule loss shown at Q4km is will within the margin of error. There will no doubt be some people who will claim
|
26 |
"PpL iSn'T gOoD eNoUgH!!1!". For those people, I have uploaded all GGUFs used in this test. Feel free to use those files to do
|
27 |
+
more extensive testing on your own time. I consider the matter resolved until somebody can conclusively demonstrate otherwise.
|
28 |
+
|
29 |
+
|
30 |
+
# Continued Experiments 2024-05-11
|
31 |
+
|
32 |
+
As an imatrix enjoyer, it has been bugging me whether the precision of the quant used to generate the imatrix actually
|
33 |
+
matters. Scuttlebut says "yes, but only a little". Logically, I don't think it should matter to a meaningful extent. PPL
|
34 |
+
scales, so a value that is relatively important at fp16 should also register as relatively important at Q8 or even Q4.
|
35 |
+
|
36 |
+
To test this theory properly, I took failspy/Llama-3-8B-Instruct-abliterated and converted it to GGUF in both fp16 and fp32
|
37 |
+
formats. I then quantized each of those GGUFs to both Q8_0 and Q4_0. I then generated imatrices for each of those six
|
38 |
+
GGUFs. Then I created eight GGUFs quantized at Q4_k_m:
|
39 |
+
|
40 |
+
- fp32 GGUF, fp32 imatrix
|
41 |
+
- fp16 GGUF, fp16 imatrix
|
42 |
+
- fp32 GGUF, fp32->Q8 imatrix
|
43 |
+
- fp16 GGUF, fp16->Q8 imatrix
|
44 |
+
- fp32 GGUF, fp32->Q4 imatrix
|
45 |
+
- fp16 GGUF, fp16->Q4 imatrix
|
46 |
+
- fp32 GGUF, no imatrix
|
47 |
+
- fp16 GGUF, no imatrix
|
48 |
+
|
49 |
+
I ran PPL against all 8 quants, as well as the full fp16 and fp32 GGUFs. Results:
|
50 |
+
|
51 |
+
<in progress>
|
52 |
+
|
53 |
+
Conclusion:
|
54 |
+
|
55 |
+
<in progress>
|