Update README.md
Browse files
README.md
CHANGED
@@ -25,7 +25,7 @@ Made with exllamav2 0.2.3 with the default dataset.
|
|
25 |
Exl2 quants can be used with Nvidia RTX2xxx or newer GPUs on Windows/Linux or AMD on Linux.
|
26 |
This model format works the best when a model fits your GPU, otherwise it's better to use GGUF versions.
|
27 |
For example with RTX3060/12GB I could fit 4.5bpw/5bpw with Q6 cache and 16k context.
|
28 |
-
|
29 |
|
30 |
# Original model card
|
31 |
# huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2
|
|
|
25 |
Exl2 quants can be used with Nvidia RTX2xxx or newer GPUs on Windows/Linux or AMD on Linux.
|
26 |
This model format works the best when a model fits your GPU, otherwise it's better to use GGUF versions.
|
27 |
For example with RTX3060/12GB I could fit 4.5bpw/5bpw with Q6 cache and 16k context.
|
28 |
+
Use with with Text-Generation-WebUI, TabbyAPI or other apps that have exllamav2 loader.
|
29 |
|
30 |
# Original model card
|
31 |
# huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2
|