bartowski commited on
Commit
d1f22aa
1 Parent(s): 7abe463

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -6
README.md CHANGED
@@ -20,11 +20,10 @@ Original model: https://huggingface.co/google/gemma-2-27b-it
20
 
21
  All quants made using imatrix option with dataset from [here](https://gist.github.com/bartowski1182/eb213dccb3571f863da82e99418f81e8)
22
 
23
- Experimental quants are made with `--output-tensor-type f16 --token-embedding-type f16` per [ZeroWw](https://huggingface.co/ZeroWw)'s suggestion, please provide any feedback on quality differences you spot.
24
-
25
  ## What's new
26
 
27
  - June 21 2024: Contains latest tokenizer fixes, which addressed a few oddities from the original fix, should be closest to correct performance yet. Also has metadata for SWA and logit softcapping.
 
28
 
29
  ## Prompt format
30
 
@@ -43,27 +42,34 @@ Note that this model does not support a System prompt.
43
  | -------- | ---------- | --------- | ----------- |
44
  | [gemma-2-27b-it-Q8_0_L.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q8_1.gguf) | Q8_0_L | 30.04GB | *Experimental*, uses f16 for embed and output weights. Please provide any feedback of differences. Extremely high quality, generally unneeded but max available quant. |
45
  | [gemma-2-27b-it-Q8_0.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q8_0.gguf) | Q8_0 | 28.93GB | Extremely high quality, generally unneeded but max available quant. |
46
- | [gemma-2-27b-it-Q6_K_L.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q6_K_L.gguf) | Q6_K_L | 23.73GB | *Experimental*, uses f16 for embed and output weights. Please provide any feedback of differences. Very high quality, near perfect, *recommended*. |
47
  | [gemma-2-27b-it-Q6_K.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q6_K.gguf) | Q6_K | 22.34GB | Very high quality, near perfect, *recommended*. |
48
- | [gemma-2-27b-it-Q5_K_L.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q5_K_L.gguf) | Q5_K_L | 20.79GB | *Experimental*, uses f16 for embed and output weights. Please provide any feedback of differences. High quality, *recommended*. |
49
  | [gemma-2-27b-it-Q5_K_M.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q5_K_M.gguf) | Q5_K_M | 19.40GB | High quality, *recommended*. |
50
  | [gemma-2-27b-it-Q5_K_S.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q5_K_S.gguf) | Q5_K_S | 18.88GB | High quality, *recommended*. |
51
- | [gemma-2-27b-it-Q4_K_L.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q4_K_L.gguf) | Q4_K_L | 18.03GB | *Experimental*, uses f16 for embed and output weights. Please provide any feedback of differences. Good quality, uses about 4.83 bits per weight, *recommended*. |
52
  | [gemma-2-27b-it-Q4_K_M.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q4_K_M.gguf) | Q4_K_M | 16.64GB | Good quality, uses about 4.83 bits per weight, *recommended*. |
53
  | [gemma-2-27b-it-Q4_K_S.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q4_K_S.gguf) | Q4_K_S | 15.73GB | Slightly lower quality with more space savings, *recommended*. |
54
  | [gemma-2-27b-it-IQ4_XS.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-IQ4_XS.gguf) | IQ4_XS | 14.81GB | Decent quality, smaller than Q4_K_S with similar performance, *recommended*. |
55
- | [gemma-2-27b-it-Q3_K_XL.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q3_K_XL.gguf) | Q3_K_XL | 15.91GB | *Experimental*, uses f16 for embed and output weights. Please provide any feedback of differences. Lower quality but usable, good for low RAM availability. |
56
  | [gemma-2-27b-it-Q3_K_L.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q3_K_L.gguf) | Q3_K_L | 14.51GB | Lower quality but usable, good for low RAM availability. |
57
  | [gemma-2-27b-it-Q3_K_M.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q3_K_M.gguf) | Q3_K_M | 13.42GB | Even lower quality. |
58
  | [gemma-2-27b-it-IQ3_M.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-IQ3_M.gguf) | IQ3_M | 12.45GB | Medium-low quality, new method with decent performance comparable to Q3_K_M. |
59
  | [gemma-2-27b-it-Q3_K_S.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q3_K_S.gguf) | Q3_K_S | 12.16GB | Low quality, not recommended. |
60
  | [gemma-2-27b-it-IQ3_XS.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-IQ3_XS.gguf) | IQ3_XS | 11.55GB | Lower quality, new method with decent performance, slightly better than Q3_K_S. |
61
  | [gemma-2-27b-it-IQ3_XXS.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-IQ3_XXS.gguf) | IQ3_XXS | 10.75GB | Lower quality, new method with decent performance, comparable to Q3 quants. |
 
62
  | [gemma-2-27b-it-Q2_K.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q2_K.gguf) | Q2_K | 10.44GB | Very low quality but surprisingly usable. |
63
  | [gemma-2-27b-it-IQ2_M.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-IQ2_M.gguf) | IQ2_M | 9.39GB | Very low quality, uses SOTA techniques to also be surprisingly usable. |
64
  | [gemma-2-27b-it-IQ2_S.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-IQ2_S.gguf) | IQ2_S | 8.65GB | Very low quality, uses SOTA techniques to be usable. |
65
  | [gemma-2-27b-it-IQ2_XS.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-IQ2_XS.gguf) | IQ2_XS | 8.39GB | Very low quality, uses SOTA techniques to be usable. |
66
 
 
 
 
 
 
 
67
  ## Downloading using huggingface-cli
68
 
69
  First, make sure you have hugginface-cli installed:
 
20
 
21
  All quants made using imatrix option with dataset from [here](https://gist.github.com/bartowski1182/eb213dccb3571f863da82e99418f81e8)
22
 
 
 
23
  ## What's new
24
 
25
  - June 21 2024: Contains latest tokenizer fixes, which addressed a few oddities from the original fix, should be closest to correct performance yet. Also has metadata for SWA and logit softcapping.
26
+ - July 3 2024: Updated the experimental quants to newer method, Q8 for embed/output, yields higher quality at much lower size than f16 (left Q8_0_L since Q8_0 is already Q8 embed/output)
27
 
28
  ## Prompt format
29
 
 
42
  | -------- | ---------- | --------- | ----------- |
43
  | [gemma-2-27b-it-Q8_0_L.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q8_1.gguf) | Q8_0_L | 30.04GB | *Experimental*, uses f16 for embed and output weights. Please provide any feedback of differences. Extremely high quality, generally unneeded but max available quant. |
44
  | [gemma-2-27b-it-Q8_0.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q8_0.gguf) | Q8_0 | 28.93GB | Extremely high quality, generally unneeded but max available quant. |
45
+ | [gemma-2-27b-it-Q6_K_L.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q6_K_L.gguf) | Q6_K_L | 22.62GB | Uses Q8_0 for embed and output weights. Very high quality, near perfect, *recommended*. |
46
  | [gemma-2-27b-it-Q6_K.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q6_K.gguf) | Q6_K | 22.34GB | Very high quality, near perfect, *recommended*. |
47
+ | [gemma-2-27b-it-Q5_K_L.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q5_K_L.gguf) | Q5_K_L | 19.69GB | Uses Q8_0 for embed and output weights. High quality, *recommended*. |
48
  | [gemma-2-27b-it-Q5_K_M.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q5_K_M.gguf) | Q5_K_M | 19.40GB | High quality, *recommended*. |
49
  | [gemma-2-27b-it-Q5_K_S.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q5_K_S.gguf) | Q5_K_S | 18.88GB | High quality, *recommended*. |
50
+ | [gemma-2-27b-it-Q4_K_L.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q4_K_L.gguf) | Q4_K_L | 16.93GB | Uses Q8_0 for embed and output weights. Good quality, uses about 4.83 bits per weight, *recommended*. |
51
  | [gemma-2-27b-it-Q4_K_M.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q4_K_M.gguf) | Q4_K_M | 16.64GB | Good quality, uses about 4.83 bits per weight, *recommended*. |
52
  | [gemma-2-27b-it-Q4_K_S.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q4_K_S.gguf) | Q4_K_S | 15.73GB | Slightly lower quality with more space savings, *recommended*. |
53
  | [gemma-2-27b-it-IQ4_XS.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-IQ4_XS.gguf) | IQ4_XS | 14.81GB | Decent quality, smaller than Q4_K_S with similar performance, *recommended*. |
54
+ | [gemma-2-27b-it-Q3_K_XL.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q3_K_XL.gguf) | Q3_K_XL | 14.80GB | Uses Q8_0 for embed and output weights. Lower quality but usable, good for low RAM availability. |
55
  | [gemma-2-27b-it-Q3_K_L.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q3_K_L.gguf) | Q3_K_L | 14.51GB | Lower quality but usable, good for low RAM availability. |
56
  | [gemma-2-27b-it-Q3_K_M.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q3_K_M.gguf) | Q3_K_M | 13.42GB | Even lower quality. |
57
  | [gemma-2-27b-it-IQ3_M.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-IQ3_M.gguf) | IQ3_M | 12.45GB | Medium-low quality, new method with decent performance comparable to Q3_K_M. |
58
  | [gemma-2-27b-it-Q3_K_S.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q3_K_S.gguf) | Q3_K_S | 12.16GB | Low quality, not recommended. |
59
  | [gemma-2-27b-it-IQ3_XS.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-IQ3_XS.gguf) | IQ3_XS | 11.55GB | Lower quality, new method with decent performance, slightly better than Q3_K_S. |
60
  | [gemma-2-27b-it-IQ3_XXS.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-IQ3_XXS.gguf) | IQ3_XXS | 10.75GB | Lower quality, new method with decent performance, comparable to Q3 quants. |
61
+ | [gemma-2-27b-it-Q2_K_L.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q2_K_L.gguf) | Q2_K_L | 10.73GB | Uses Q8_0 for embed and output weights. Very low quality but surprisingly usable. |
62
  | [gemma-2-27b-it-Q2_K.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-Q2_K.gguf) | Q2_K | 10.44GB | Very low quality but surprisingly usable. |
63
  | [gemma-2-27b-it-IQ2_M.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-IQ2_M.gguf) | IQ2_M | 9.39GB | Very low quality, uses SOTA techniques to also be surprisingly usable. |
64
  | [gemma-2-27b-it-IQ2_S.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-IQ2_S.gguf) | IQ2_S | 8.65GB | Very low quality, uses SOTA techniques to be usable. |
65
  | [gemma-2-27b-it-IQ2_XS.gguf](https://huggingface.co/bartowski/gemma-2-27b-it-GGUF/blob/main/gemma-2-27b-it-IQ2_XS.gguf) | IQ2_XS | 8.39GB | Very low quality, uses SOTA techniques to be usable. |
66
 
67
+ ## Credits
68
+
69
+ Thank you kalomaze and Dampf for assistance in creating the imatrix calibration dataset
70
+
71
+ Thank you ZeroWw for the inspiration to experiment with embed/output
72
+
73
  ## Downloading using huggingface-cli
74
 
75
  First, make sure you have hugginface-cli installed: