fix: readme llama.cpp link
Browse files
README.md
CHANGED
@@ -19,7 +19,7 @@ _Llama.cpp imatrix quantization of deepseek-ai/DeepSeek-V2-Lite-Chat_
|
|
19 |
|
20 |
Original Model: [deepseek-ai/DeepSeek-V2-Lite-Chat](https://huggingface.co/deepseek-ai/DeepSeek-V2-Lite-Chat)
|
21 |
Original dtype: `BF16` (`bfloat16`)
|
22 |
-
Quantized by: llama.cpp [
|
23 |
IMatrix dataset: [here](https://gist.githubusercontent.com/legraphista/d6d93f1a254bcfc58e0af3777eaec41e/raw/d380e7002cea4a51c33fffd47db851942754e7cc/imatrix.calibration.medium.raw)
|
24 |
|
25 |
## Files
|
|
|
19 |
|
20 |
Original Model: [deepseek-ai/DeepSeek-V2-Lite-Chat](https://huggingface.co/deepseek-ai/DeepSeek-V2-Lite-Chat)
|
21 |
Original dtype: `BF16` (`bfloat16`)
|
22 |
+
Quantized by: llama.cpp fork [PR 7519](https://github.com/ggerganov/llama.cpp/pull/7519)
|
23 |
IMatrix dataset: [here](https://gist.githubusercontent.com/legraphista/d6d93f1a254bcfc58e0af3777eaec41e/raw/d380e7002cea4a51c33fffd47db851942754e7cc/imatrix.calibration.medium.raw)
|
24 |
|
25 |
## Files
|