Update README.md
Browse files
README.md
CHANGED
@@ -23,31 +23,24 @@ This is a `Q5_K_M` GGUF version of [BramVanroy/GEITje-7B-ultra](https://huggingf
|
|
23 |
|
24 |
### Ollama
|
25 |
|
26 |
-
|
27 |
|
|
|
|
|
28 |
```
|
29 |
-
|
30 |
-
|
31 |
-
|
32 |
-
|
33 |
-
|
34 |
-
|
35 |
-
|
36 |
-
|
37 |
-
|
38 |
-
|
39 |
-
PARAMETER stop "<|system|>"
|
40 |
-
PARAMETER stop "<|user|>"
|
41 |
-
PARAMETER stop "<|assistant|>"
|
42 |
-
PARAMETER stop "</s>"
|
43 |
-
PARAMETER temperature 1.0
|
44 |
-
PARAMETER top_p 1.0
|
45 |
-
PARAMETER top_k 50
|
46 |
-
PARAMETER repeat_penalty 1.0
|
47 |
-
SYSTEM Je bent 'GEITje Ultra', of 'Ultra' in het kort, een behulpzame en enthousiaste AI-assistent. Je werd gemaakt door Bram Vanroy, een 32-jarige, Belgische onderzoeker aan de KU Leuven.
|
48 |
```
|
49 |
|
50 |
-
##
|
51 |
|
52 |
Assuming you have installed and build llama cpp, current working directory is the `build` directory in llamacpp.
|
53 |
|
|
|
23 |
|
24 |
### Ollama
|
25 |
|
26 |
+
The model is available on `ollama` and can be easily run as follows:
|
27 |
|
28 |
+
```shell
|
29 |
+
ollama run bramvanroy/geitje-7b-ultra-gguf
|
30 |
```
|
31 |
+
|
32 |
+
To reproduce, follow the next steps.
|
33 |
+
|
34 |
+
First download the [GGUF file](https://huggingface.co/BramVanroy/GEITje-7B-ultra-GGUF/resolve/main/GEITje-7B-ultra-Q5_K_M.gguf?download=true) and [Modelfile](https://huggingface.co/BramVanroy/GEITje-7B-ultra-GGUF/resolve/main/Modelfile?download=true) to your computer. You can adapt the Modelfile as you wish.
|
35 |
+
|
36 |
+
Then, create the ollama model and run it.
|
37 |
+
|
38 |
+
```shelll
|
39 |
+
ollama create geitje-7b-ultra-gguf -f ./Modelfile
|
40 |
+
ollama run geitje-7b-ultra-gguf
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
41 |
```
|
42 |
|
43 |
+
## Reproduce this GGUF version from the non-quantized model
|
44 |
|
45 |
Assuming you have installed and build llama cpp, current working directory is the `build` directory in llamacpp.
|
46 |
|