Use a low temperature mistral for tasks (#652)
Browse files- .env.template +12 -1
.env.template
CHANGED
@@ -213,7 +213,18 @@ MODELS=`[
|
|
213 |
|
214 |
OLD_MODELS=`[{"name":"bigcode/starcoder"}, {"name":"OpenAssistant/oasst-sft-6-llama-30b-xor"}, {"name":"HuggingFaceH4/zephyr-7b-alpha"}]`
|
215 |
|
216 |
-
TASK_MODEL
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
217 |
|
218 |
|
219 |
APP_BASE="/chat"
|
|
|
213 |
|
214 |
OLD_MODELS=`[{"name":"bigcode/starcoder"}, {"name":"OpenAssistant/oasst-sft-6-llama-30b-xor"}, {"name":"HuggingFaceH4/zephyr-7b-alpha"}]`
|
215 |
|
216 |
+
TASK_MODEL=`{
|
217 |
+
"name": "mistralai/Mistral-7B-Instruct-v0.2",
|
218 |
+
"chatPromptTemplate" : "<s>{{#each messages}}{{#ifUser}}[INST] {{#if @first}}{{#if @root.preprompt}}{{@root.preprompt}}\n{{/if}}{{/if}}{{content}} [/INST]{{/ifUser}}{{#ifAssistant}}{{content}}</s>{{/ifAssistant}}{{/each}}",
|
219 |
+
"parameters": {
|
220 |
+
"temperature": 0.1,
|
221 |
+
"top_p": 0.95,
|
222 |
+
"repetition_penalty": 1.2,
|
223 |
+
"top_k": 50,
|
224 |
+
"truncate": 3072,
|
225 |
+
"max_new_tokens": 1024,
|
226 |
+
"stop": ["</s>"]
|
227 |
+
}}`
|
228 |
|
229 |
|
230 |
APP_BASE="/chat"
|