Spaces:
Sleeping
Sleeping
Add Minueza-32M-Chat model
Browse files
app.py
CHANGED
@@ -33,12 +33,13 @@ def generate(
|
|
33 |
elif model_name == "Felladrin/TinyMistral-248M-SFT-v4":
|
34 |
outputs = pipe(prompt, max_length=1024, use_cache=True, penalty_alpha=0.5, top_k=5, repetition_penalty=1.001)
|
35 |
else:
|
36 |
-
outputs = pipe(prompt, max_length=1024, do_sample=True, temperature=0.
|
37 |
|
38 |
return outputs[0]["generated_text"]
|
39 |
|
40 |
model_choices = [
|
41 |
"Felladrin/Llama-160M-Chat-v1",
|
|
|
42 |
"Felladrin/Smol-Llama-101M-Chat-v1",
|
43 |
"Felladrin/TinyMistral-248M-SFT-v4",
|
44 |
"Felladrin/Pythia-31M-Chat-v1",
|
|
|
33 |
elif model_name == "Felladrin/TinyMistral-248M-SFT-v4":
|
34 |
outputs = pipe(prompt, max_length=1024, use_cache=True, penalty_alpha=0.5, top_k=5, repetition_penalty=1.001)
|
35 |
else:
|
36 |
+
outputs = pipe(prompt, max_length=1024, do_sample=True, temperature=0.7, top_k=35, top_p=0.5, repetition_penalty=1.176)
|
37 |
|
38 |
return outputs[0]["generated_text"]
|
39 |
|
40 |
model_choices = [
|
41 |
"Felladrin/Llama-160M-Chat-v1",
|
42 |
+
"Felladrin/Minueza-32M-Chat",
|
43 |
"Felladrin/Smol-Llama-101M-Chat-v1",
|
44 |
"Felladrin/TinyMistral-248M-SFT-v4",
|
45 |
"Felladrin/Pythia-31M-Chat-v1",
|