More documentation
Browse files
README.md
CHANGED
@@ -183,8 +183,8 @@ Personal recommendations are to use an [i1-Q4_K_M](https://www.reddit.com/r/Loca
|
|
183 |
num_ctx = 4096
|
184 |
repeat_penalty = 1.2
|
185 |
temperature = 0.85
|
186 |
-
top_k = 0 # Change to 40 if you're roleplaying
|
187 |
top_p = 1
|
188 |
```
|
189 |
|
190 |
-
Other recommendations can be found on [this paper on mobile LLMs](https://openreview.net/pdf?id=ahVsd1hy2W)
|
|
|
183 |
num_ctx = 4096
|
184 |
repeat_penalty = 1.2
|
185 |
temperature = 0.85
|
186 |
+
top_k = 0 # Change to 40+ if you're roleplaying
|
187 |
top_p = 1
|
188 |
```
|
189 |
|
190 |
+
Other recommendations can be found on [this paper on mobile LLMs](https://openreview.net/pdf?id=ahVsd1hy2W), [this paper on balancing model parameters](https://arxiv.org/html/2408.13586v1), and [this Reddit post about tweaking Llama 3.1 parameters](https://www.reddit.com/r/LocalLLaMA/comments/1ej1zrl/try_these_settings_for_llama_31_for_longer_or/).
|