Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
@@ -26,7 +26,7 @@ max_new_tokens=1000
|
|
26 |
top_p=0.92
|
27 |
repetition_penalty=1.7
|
28 |
|
29 |
-
model_name = "Inagua/code-model"
|
30 |
|
31 |
llm = LLM(model_name, max_model_len=4096)
|
32 |
|
@@ -134,7 +134,8 @@ class MistralChatBot:
|
|
134 |
self.system_prompt = system_prompt
|
135 |
|
136 |
def predict(self, user_message, context):
|
137 |
-
detailed_prompt = """### Question ###\n""" + user_message + "\n\n### Contexte ###\n" + context + "\n\n### Formule ###\n"
|
|
|
138 |
prompts = [detailed_prompt]
|
139 |
outputs = llm.generate(prompts, sampling_params, use_tqdm = False)
|
140 |
generated_text = outputs[0].outputs[0].text
|
|
|
26 |
top_p=0.92
|
27 |
repetition_penalty=1.7
|
28 |
|
29 |
+
model_name = "Inagua/code-model-2"
|
30 |
|
31 |
llm = LLM(model_name, max_model_len=4096)
|
32 |
|
|
|
134 |
self.system_prompt = system_prompt
|
135 |
|
136 |
def predict(self, user_message, context):
|
137 |
+
#detailed_prompt = """### Question ###\n""" + user_message + "\n\n### Contexte ###\n" + context + "\n\n### Formule ###\n"
|
138 |
+
detailed_prompt = """### Question ###\n""" + user_message + "\n\n### Field ###\n"
|
139 |
prompts = [detailed_prompt]
|
140 |
outputs = llm.generate(prompts, sampling_params, use_tqdm = False)
|
141 |
generated_text = outputs[0].outputs[0].text
|