BramVanroy commited on
Commit
63c6e32
·
verified ·
1 Parent(s): 8b96289

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -21
README.md CHANGED
@@ -23,31 +23,24 @@ This is a `Q5_K_M` GGUF version of [BramVanroy/GEITje-7B-ultra](https://huggingf
23
 
24
  ### Ollama
25
 
26
- First download the [GGUF file](https://huggingface.co/BramVanroy/GEITje-7B-ultra-GGUF/resolve/main/GEITje-7B-ultra-Q5_K_M.gguf?download=true) to your computer.
27
 
 
 
28
  ```
29
- FROM GEITje-7B-ultra-Q5_K_M.gguf
30
- TEMPLATE """{{- if .System }}
31
- <|system|>
32
- {{ .System }}</s>
33
- {{ end }}
34
- <|user|>
35
- {{ .Prompt }}</s>
36
- <|assistant|>
37
- """
38
- PARAMETER num_ctx 8192
39
- PARAMETER stop "<|system|>"
40
- PARAMETER stop "<|user|>"
41
- PARAMETER stop "<|assistant|>"
42
- PARAMETER stop "</s>"
43
- PARAMETER temperature 1.0
44
- PARAMETER top_p 1.0
45
- PARAMETER top_k 50
46
- PARAMETER repeat_penalty 1.0
47
- SYSTEM Je bent 'GEITje Ultra', of 'Ultra' in het kort, een behulpzame en enthousiaste AI-assistent. Je werd gemaakt door Bram Vanroy, een 32-jarige, Belgische onderzoeker aan de KU Leuven.
48
  ```
49
 
50
- ## Repro
51
 
52
  Assuming you have installed and build llama cpp, current working directory is the `build` directory in llamacpp.
53
 
 
23
 
24
  ### Ollama
25
 
26
+ The model is available on `ollama` and can be easily run as follows:
27
 
28
+ ```shell
29
+ ollama run bramvanroy/geitje-7b-ultra-gguf
30
  ```
31
+
32
+ To reproduce, follow the next steps.
33
+
34
+ First download the [GGUF file](https://huggingface.co/BramVanroy/GEITje-7B-ultra-GGUF/resolve/main/GEITje-7B-ultra-Q5_K_M.gguf?download=true) and [Modelfile](https://huggingface.co/BramVanroy/GEITje-7B-ultra-GGUF/resolve/main/Modelfile?download=true) to your computer. You can adapt the Modelfile as you wish.
35
+
36
+ Then, create the ollama model and run it.
37
+
38
+ ```shelll
39
+ ollama create geitje-7b-ultra-gguf -f ./Modelfile
40
+ ollama run geitje-7b-ultra-gguf
 
 
 
 
 
 
 
 
 
41
  ```
42
 
43
+ ## Reproduce this GGUF version from the non-quantized model
44
 
45
  Assuming you have installed and build llama cpp, current working directory is the `build` directory in llamacpp.
46