Update README.md
Browse files
README.md
CHANGED
@@ -4,6 +4,25 @@ license_name: deepseek
|
|
4 |
license_link: https://huggingface.co/deepseek-ai/deepseek-coder-33b-base/blob/main/LICENSE
|
5 |
---
|
6 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
7 |
|
8 |
# This model is now live (We'll always be serving the newest model on our web app)!
|
9 |
Access at: https://www.whiterabbitneo.com/
|
|
|
4 |
license_link: https://huggingface.co/deepseek-ai/deepseek-coder-33b-base/blob/main/LICENSE
|
5 |
---
|
6 |
|
7 |
+
Q8_0 GGUF quantization of [WhiteRabbitNeo/WhiteRabbitNeo-33B-v1.5](https://huggingface.co/WhiteRabbitNeo/WhiteRabbitNeo-33B-v1.5)
|
8 |
+
converted and quantized with llama.cpp
|
9 |
+
|
10 |
+
Please note there is a **bug** in the `convert.py` script from recent versions of llama.cpp, that **affects models
|
11 |
+
with additional vocabulary** (like this one). This results in model that can be converted, but have some garbage
|
12 |
+
and cannot be used for inference. Until this is fixed, you must use an older version of `convert.py`
|
13 |
+
(only this script, the rest of llama.cpp can stay with the latest version), and use the following parameters:
|
14 |
+
`--pad-vocab --vocab-type bpe`. For example. with a locally downloaded version of the original model repo:
|
15 |
+
|
16 |
+
```
|
17 |
+
python llama.cpp/convert-aa23412.py WhiteRabbitNeo-33B-v1.5 --pad-vocab --vocab-type bpe
|
18 |
+
```
|
19 |
+
|
20 |
+
The **last working version** of `convert.py` is [aa23412](https://github.com/ggerganov/llama.cpp/tree/aa2341298924ac89778252015efcb792f2df1e20)
|
21 |
+
|
22 |
+
---
|
23 |
+
|
24 |
+
# ORGINAL MODEL CARD
|
25 |
+
|
26 |
|
27 |
# This model is now live (We'll always be serving the newest model on our web app)!
|
28 |
Access at: https://www.whiterabbitneo.com/
|