Update README.md
Browse files
README.md
CHANGED
@@ -10,7 +10,11 @@ library_name: transformers
|
|
10 |
# GGML of:
|
11 |
Manticore-13b-Chat-Pyg by [openaccess-ai-collective](https://huggingface.co/openaccess-ai-collective/manticore-13b-chat-pyg) with the Guanaco 13b qLoRa by [TimDettmers](https://huggingface.co/timdettmers/guanaco-13b) applied through [Monero](https://huggingface.co/Monero/Manticore-13b-Chat-Pyg-Guanaco), quantized by [mindrage](https://huggingface.co/mindrage), uncensored
|
12 |
|
13 |
-
(
|
|
|
|
|
|
|
|
|
14 |
|
15 |
[link to GPTQ Version](https://huggingface.co/mindrage/Manticore-13B-Chat-Pyg-Guanaco-GPTQ-4bit-128g.no-act-order.safetensors)
|
16 |
|
|
|
10 |
# GGML of:
|
11 |
Manticore-13b-Chat-Pyg by [openaccess-ai-collective](https://huggingface.co/openaccess-ai-collective/manticore-13b-chat-pyg) with the Guanaco 13b qLoRa by [TimDettmers](https://huggingface.co/timdettmers/guanaco-13b) applied through [Monero](https://huggingface.co/Monero/Manticore-13b-Chat-Pyg-Guanaco), quantized by [mindrage](https://huggingface.co/mindrage), uncensored
|
12 |
|
13 |
+
12.06.2023: Added versions quantized with the new method (less precision loss relative to compression ratio, but slower (for now)):
|
14 |
+
q2_K, q3_KM, q4_KS, q4_KM, q5_KS
|
15 |
+
|
16 |
+
Old Quant method:
|
17 |
+
q4_0, q5_0 and q8_0 versions available
|
18 |
|
19 |
[link to GPTQ Version](https://huggingface.co/mindrage/Manticore-13B-Chat-Pyg-Guanaco-GPTQ-4bit-128g.no-act-order.safetensors)
|
20 |
|