Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -10,10 +10,7 @@ tokenizer = AutoTokenizer.from_pretrained(model_path, trust_remote_code=False)
|
|
10 |
|
11 |
|
12 |
def respond(
|
13 |
-
message
|
14 |
-
max_tokens,
|
15 |
-
temperature,
|
16 |
-
top_p,
|
17 |
):
|
18 |
formatted_prompt = f"<s>[INST]{message}[/INST]"
|
19 |
|
@@ -23,9 +20,7 @@ def respond(
|
|
23 |
result = model.generate(
|
24 |
tokens,
|
25 |
do_sample=False,
|
26 |
-
max_new_tokens=
|
27 |
-
temperature=temperature,
|
28 |
-
top_p=top_p,
|
29 |
)
|
30 |
|
31 |
response = tokenizer.decode(result[0], skip_special_tokens=True)
|
|
|
10 |
|
11 |
|
12 |
def respond(
|
13 |
+
message
|
|
|
|
|
|
|
14 |
):
|
15 |
formatted_prompt = f"<s>[INST]{message}[/INST]"
|
16 |
|
|
|
20 |
result = model.generate(
|
21 |
tokens,
|
22 |
do_sample=False,
|
23 |
+
max_new_tokens=200
|
|
|
|
|
24 |
)
|
25 |
|
26 |
response = tokenizer.decode(result[0], skip_special_tokens=True)
|