Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
update HF models
Browse files
app_hf.py
CHANGED
@@ -2,26 +2,16 @@ from utils import get_app
|
|
2 |
|
3 |
demo = get_app(
|
4 |
models=[
|
5 |
-
"Qwen/Qwen2.5-Coder-32B-Instruct",
|
6 |
-
"Qwen/Qwen2.5-72B-Instruct",
|
7 |
-
"meta-llama/Llama-3.1-70B-Instruct",
|
8 |
-
"mistralai/Mixtral-8x7B-Instruct-v0.1",
|
9 |
-
"meta-llama/Llama-3.1-8B-Instruct",
|
10 |
-
"google/gemma-2-9b-it",
|
11 |
-
"mistralai/Mistral-7B-v0.1",
|
12 |
-
"meta-llama/Llama-2-7b-chat-hf",
|
13 |
-
"meta-llama/Llama-3.2-3B-Instruct",
|
14 |
-
"meta-llama/Llama-3.2-1B-Instruct",
|
15 |
-
"Qwen/Qwen2.5-1.5B-Instruct",
|
16 |
"microsoft/Phi-3.5-mini-instruct",
|
17 |
"HuggingFaceTB/SmolLM2-1.7B-Instruct",
|
18 |
"google/gemma-2-2b-it",
|
19 |
-
"meta-llama/Llama-3.2-3B",
|
20 |
-
"meta-llama/Llama-3.2-1B",
|
21 |
"openai-community/gpt2",
|
|
|
|
|
22 |
],
|
23 |
default_model="HuggingFaceTB/SmolLM2-1.7B-Instruct",
|
24 |
src="models",
|
|
|
25 |
)
|
26 |
|
27 |
if __name__ == "__main__":
|
|
|
2 |
|
3 |
demo = get_app(
|
4 |
models=[
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
5 |
"microsoft/Phi-3.5-mini-instruct",
|
6 |
"HuggingFaceTB/SmolLM2-1.7B-Instruct",
|
7 |
"google/gemma-2-2b-it",
|
|
|
|
|
8 |
"openai-community/gpt2",
|
9 |
+
"microsoft/phi-2",
|
10 |
+
"TinyLlama/TinyLlama-1.1B-Chat-v1.0",
|
11 |
],
|
12 |
default_model="HuggingFaceTB/SmolLM2-1.7B-Instruct",
|
13 |
src="models",
|
14 |
+
cache_examples=False
|
15 |
)
|
16 |
|
17 |
if __name__ == "__main__":
|