MarsupialAI
commited on
Commit
•
13688f7
1
Parent(s):
0c5f294
Update README.md
Browse files
README.md
CHANGED
@@ -34,6 +34,7 @@ Assume this has it too. This model is for entertainment purposes only.
|
|
34 |
|
35 |
FP16 and Q4_K_S GGUFs: https://huggingface.co/MarsupialAI/KitchenSink_103b_GGUF_Q4KS_FP16
|
36 |
Q2 and Q3 GGUFs: https://huggingface.co/bcse/KitchenSink-103b-GGUF
|
|
|
37 |
EXL2 quants: https://huggingface.co/AzureBlack/KitchenSink_103b-2.5bpw-6h-exl2 https://huggingface.co/AzureBlack/KitchenSink_103b-3.5bpw-6h-exl2 https://huggingface.co/AzureBlack/KitchenSink_103b-4.5bpw-6h-exl2
|
38 |
|
39 |
If you create additional quants, please let me know and I will link them here as well.
|
|
|
34 |
|
35 |
FP16 and Q4_K_S GGUFs: https://huggingface.co/MarsupialAI/KitchenSink_103b_GGUF_Q4KS_FP16
|
36 |
Q2 and Q3 GGUFs: https://huggingface.co/bcse/KitchenSink-103b-GGUF
|
37 |
+
iMatrix GGUFs: https://huggingface.co/MarsupialAI/KitchenSink_103b_iMatrix_GGUF
|
38 |
EXL2 quants: https://huggingface.co/AzureBlack/KitchenSink_103b-2.5bpw-6h-exl2 https://huggingface.co/AzureBlack/KitchenSink_103b-3.5bpw-6h-exl2 https://huggingface.co/AzureBlack/KitchenSink_103b-4.5bpw-6h-exl2
|
39 |
|
40 |
If you create additional quants, please let me know and I will link them here as well.
|