use deepthought-8b-llama-v0.01-alpha-Q4_0_8_8.gguf
Browse files
app.py
CHANGED
@@ -5,7 +5,7 @@ from typing import Tuple, List
|
|
5 |
import time
|
6 |
|
7 |
DESCRIPTION = f"""
|
8 |
-
# Chat with
|
9 |
"""
|
10 |
|
11 |
MAX_MAX_NEW_TOKENS = 1024
|
@@ -13,8 +13,8 @@ DEFAULT_MAX_NEW_TOKENS = 200
|
|
13 |
|
14 |
# Download the GGUF file
|
15 |
model_path = hf_hub_download(
|
16 |
-
repo_id="
|
17 |
-
filename="
|
18 |
repo_type="model"
|
19 |
)
|
20 |
# Load the GGUF model
|
|
|
5 |
import time
|
6 |
|
7 |
DESCRIPTION = f"""
|
8 |
+
# Chat with Deepthought 8B as GGUF on CPU
|
9 |
"""
|
10 |
|
11 |
MAX_MAX_NEW_TOKENS = 1024
|
|
|
13 |
|
14 |
# Download the GGUF file
|
15 |
model_path = hf_hub_download(
|
16 |
+
repo_id="bartowski/deepthought-8b-llama-v0.01-alpha-GGUF",
|
17 |
+
filename="deepthought-8b-llama-v0.01-alpha-Q4_0_8_8.gguf",
|
18 |
repo_type="model"
|
19 |
)
|
20 |
# Load the GGUF model
|