Update README.md
Browse files
README.md
CHANGED
@@ -5,7 +5,7 @@ GPT4-X-Alpasta-30b working with Oobabooga's Text Generation Webui and KoboldAI.
|
|
5 |
<p><strong>What's included</strong></p>
|
6 |
|
7 |
<P>GPTQ: 2 quantized versions. One quantized --true-sequential and act-order optimizations, and the other was quantized using --true-sequential --groupsize 128 optimizations (coming soon)</P>
|
8 |
-
<P>GGML:
|
9 |
|
10 |
<p><strong>GPU/GPTQ Usage</strong></p>
|
11 |
<p>To use with your GPU using GPTQ pick one of the .safetensors along with all of the .jsons and .model files.</p>
|
@@ -31,10 +31,10 @@ GPT4-X-Alpasta-30b working with Oobabooga's Text Generation Webui and KoboldAI.
|
|
31 |
|
32 |
<p><strong><font size="4">--true-sequential --groupsize 128</font></strong></p>
|
33 |
|
34 |
-
<strong>Wikitext2</strong>:
|
35 |
|
36 |
-
<strong>Ptb-New</strong>:
|
37 |
|
38 |
-
<strong>C4-New</strong>:
|
39 |
|
40 |
<strong>Note</strong>: This version uses <i>--groupsize 128</i>, resulting in better evaluations. However, it consumes more VRAM.
|
|
|
5 |
<p><strong>What's included</strong></p>
|
6 |
|
7 |
<P>GPTQ: 2 quantized versions. One quantized --true-sequential and act-order optimizations, and the other was quantized using --true-sequential --groupsize 128 optimizations (coming soon)</P>
|
8 |
+
<P>GGML: 2 quantized versions. One quantized using q4_1, and the other was quantized using q5_0.</P>
|
9 |
|
10 |
<p><strong>GPU/GPTQ Usage</strong></p>
|
11 |
<p>To use with your GPU using GPTQ pick one of the .safetensors along with all of the .jsons and .model files.</p>
|
|
|
31 |
|
32 |
<p><strong><font size="4">--true-sequential --groupsize 128</font></strong></p>
|
33 |
|
34 |
+
<strong>Wikitext2</strong>: 4.70257568359375
|
35 |
|
36 |
+
<strong>Ptb-New</strong>: 9.323467254638672
|
37 |
|
38 |
+
<strong>C4-New</strong>: 7.041860580444336
|
39 |
|
40 |
<strong>Note</strong>: This version uses <i>--groupsize 128</i>, resulting in better evaluations. However, it consumes more VRAM.
|