legraphista
commited on
Commit
•
b6350f4
1
Parent(s):
6ba3f19
Update README.md
Browse files
README.md
CHANGED
@@ -17,6 +17,11 @@ tags:
|
|
17 |
# DeepSeek-V2-Lite-Chat-IMat-GGUF
|
18 |
_Llama.cpp imatrix quantization of deepseek-ai/DeepSeek-V2-Lite-Chat_
|
19 |
|
|
|
|
|
|
|
|
|
|
|
20 |
- [DeepSeek-V2-Lite-Chat-IMat-GGUF](#deepseek-v2-lite-chat-imat-gguf)
|
21 |
- [Files](#files)
|
22 |
- [IMatrix](#imatrix)
|
@@ -31,10 +36,7 @@ _Llama.cpp imatrix quantization of deepseek-ai/DeepSeek-V2-Lite-Chat_
|
|
31 |
- [Why is the IMatrix not applied everywhere?](#why-is-the-imatrix-not-applied-everywhere)
|
32 |
- [How do I merge a split GGUF?](#how-do-i-merge-a-split-gguf)
|
33 |
|
34 |
-
|
35 |
-
Original dtype: `BF16` (`bfloat16`)
|
36 |
-
Quantized by: llama.cpp fork [PR 7519](https://github.com/ggerganov/llama.cpp/pull/7519)
|
37 |
-
IMatrix dataset: [here](https://gist.githubusercontent.com/legraphista/d6d93f1a254bcfc58e0af3777eaec41e/raw/d380e7002cea4a51c33fffd47db851942754e7cc/imatrix.calibration.medium.raw)
|
38 |
|
39 |
## Files
|
40 |
|
|
|
17 |
# DeepSeek-V2-Lite-Chat-IMat-GGUF
|
18 |
_Llama.cpp imatrix quantization of deepseek-ai/DeepSeek-V2-Lite-Chat_
|
19 |
|
20 |
+
Original Model: [deepseek-ai/DeepSeek-V2-Lite-Chat](https://huggingface.co/deepseek-ai/DeepSeek-V2-Lite-Chat)
|
21 |
+
Original dtype: `BF16` (`bfloat16`)
|
22 |
+
Quantized by: llama.cpp fork [PR 7519](https://github.com/ggerganov/llama.cpp/pull/7519)
|
23 |
+
IMatrix dataset: [here](https://gist.githubusercontent.com/legraphista/d6d93f1a254bcfc58e0af3777eaec41e/raw/d380e7002cea4a51c33fffd47db851942754e7cc/imatrix.calibration.medium.raw)
|
24 |
+
|
25 |
- [DeepSeek-V2-Lite-Chat-IMat-GGUF](#deepseek-v2-lite-chat-imat-gguf)
|
26 |
- [Files](#files)
|
27 |
- [IMatrix](#imatrix)
|
|
|
36 |
- [Why is the IMatrix not applied everywhere?](#why-is-the-imatrix-not-applied-everywhere)
|
37 |
- [How do I merge a split GGUF?](#how-do-i-merge-a-split-gguf)
|
38 |
|
39 |
+
---
|
|
|
|
|
|
|
40 |
|
41 |
## Files
|
42 |
|