MarsupialAI commited on
Commit
434a9a4
1 Parent(s): cd15117

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +31 -1
README.md CHANGED
@@ -1,3 +1,5 @@
 
 
1
  Some folks are claiming there's something funky going on with GGUF quanting for Llama 3 models. I don't disagree.
2
 
3
  Some of those people are speculating that it has something to do with converting the raw weights from bf16 to fp16 instead
@@ -22,4 +24,32 @@ None specified: size 29.9GB PPL @ ???? 9.5158 +/- 0.15418 PPL @ Q4km 9.62
22
  As you can see, converting to fp32 has no meaningful effect on PPL compared to converting to fp16. PPL is identical at full weight,
23
  and the miniscule loss shown at Q4km is will within the margin of error. There will no doubt be some people who will claim
24
  "PpL iSn'T gOoD eNoUgH!!1!". For those people, I have uploaded all GGUFs used in this test. Feel free to use those files to do
25
- more extensive testing on your own time. I consider the matter resolved until somebody can conclusively demonstrate otherwise.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Initial Testing
2
+
3
  Some folks are claiming there's something funky going on with GGUF quanting for Llama 3 models. I don't disagree.
4
 
5
  Some of those people are speculating that it has something to do with converting the raw weights from bf16 to fp16 instead
 
24
  As you can see, converting to fp32 has no meaningful effect on PPL compared to converting to fp16. PPL is identical at full weight,
25
  and the miniscule loss shown at Q4km is will within the margin of error. There will no doubt be some people who will claim
26
  "PpL iSn'T gOoD eNoUgH!!1!". For those people, I have uploaded all GGUFs used in this test. Feel free to use those files to do
27
+ more extensive testing on your own time. I consider the matter resolved until somebody can conclusively demonstrate otherwise.
28
+
29
+
30
+ # Continued Experiments 2024-05-11
31
+
32
+ As an imatrix enjoyer, it has been bugging me whether the precision of the quant used to generate the imatrix actually
33
+ matters. Scuttlebut says "yes, but only a little". Logically, I don't think it should matter to a meaningful extent. PPL
34
+ scales, so a value that is relatively important at fp16 should also register as relatively important at Q8 or even Q4.
35
+
36
+ To test this theory properly, I took failspy/Llama-3-8B-Instruct-abliterated and converted it to GGUF in both fp16 and fp32
37
+ formats. I then quantized each of those GGUFs to both Q8_0 and Q4_0. I then generated imatrices for each of those six
38
+ GGUFs. Then I created eight GGUFs quantized at Q4_k_m:
39
+
40
+ - fp32 GGUF, fp32 imatrix
41
+ - fp16 GGUF, fp16 imatrix
42
+ - fp32 GGUF, fp32->Q8 imatrix
43
+ - fp16 GGUF, fp16->Q8 imatrix
44
+ - fp32 GGUF, fp32->Q4 imatrix
45
+ - fp16 GGUF, fp16->Q4 imatrix
46
+ - fp32 GGUF, no imatrix
47
+ - fp16 GGUF, no imatrix
48
+
49
+ I ran PPL against all 8 quants, as well as the full fp16 and fp32 GGUFs. Results:
50
+
51
+ <in progress>
52
+
53
+ Conclusion:
54
+
55
+ <in progress>