TobDeBer commited on
Commit
cd90565
·
verified ·
1 Parent(s): c3727be

use deepthought-8b-llama-v0.01-alpha-Q4_0_8_8.gguf

Browse files
Files changed (1) hide show
  1. app.py +3 -3
app.py CHANGED
@@ -5,7 +5,7 @@ from typing import Tuple, List
5
  import time
6
 
7
  DESCRIPTION = f"""
8
- # Chat with Arco 500M as GGUF on CPU
9
  """
10
 
11
  MAX_MAX_NEW_TOKENS = 1024
@@ -13,8 +13,8 @@ DEFAULT_MAX_NEW_TOKENS = 200
13
 
14
  # Download the GGUF file
15
  model_path = hf_hub_download(
16
- repo_id="TobDeBer/arco-Q4_K_M-GGUF",
17
- filename="arco-q4_k_m.gguf",
18
  repo_type="model"
19
  )
20
  # Load the GGUF model
 
5
  import time
6
 
7
  DESCRIPTION = f"""
8
+ # Chat with Deepthought 8B as GGUF on CPU
9
  """
10
 
11
  MAX_MAX_NEW_TOKENS = 1024
 
13
 
14
  # Download the GGUF file
15
  model_path = hf_hub_download(
16
+ repo_id="bartowski/deepthought-8b-llama-v0.01-alpha-GGUF",
17
+ filename="deepthought-8b-llama-v0.01-alpha-Q4_0_8_8.gguf",
18
  repo_type="model"
19
  )
20
  # Load the GGUF model