Spaces:
Sleeping
Sleeping
hf llm
Browse files- backend.py +1 -1
backend.py
CHANGED
@@ -63,7 +63,7 @@ os.environ['TOKENIZERS_PARALLELISM'] = 'false'
|
|
63 |
llm = HuggingFaceLLM(
|
64 |
context_window=4096,
|
65 |
max_new_tokens=256,
|
66 |
-
generate_kwargs={"temperature": 0.
|
67 |
system_prompt=system_prompt,
|
68 |
tokenizer_name="meta-llama/Llama-2-7b-chat-hf",
|
69 |
model_name="meta-llama/Llama-2-7b-chat-hf",
|
|
|
63 |
llm = HuggingFaceLLM(
|
64 |
context_window=4096,
|
65 |
max_new_tokens=256,
|
66 |
+
generate_kwargs={"temperature": 0.1, "do_sample": True},
|
67 |
system_prompt=system_prompt,
|
68 |
tokenizer_name="meta-llama/Llama-2-7b-chat-hf",
|
69 |
model_name="meta-llama/Llama-2-7b-chat-hf",
|