Spaces:
Paused
Paused
Commit
·
6ddfc7c
1
Parent(s):
a2ba236
Update app.py
Browse files
app.py
CHANGED
@@ -3,14 +3,22 @@ from transformers import AutoTokenizer, pipeline, logging
|
|
3 |
from auto_gptq import AutoGPTQForCausalLM, BaseQuantizeConfig
|
4 |
from huggingface_hub import snapshot_download
|
5 |
|
6 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
7 |
|
8 |
|
9 |
quantized_model_dir = "FPHam/Jackson_The_Formalizer_V2_13b_GPTQ"
|
10 |
|
11 |
snapshot_download(repo_id=quantized_model_dir, local_dir=local_folder, local_dir_use_symlinks=False)
|
12 |
|
13 |
-
model_basename = "/model/Jackson2-4bit-128g-GPTQ.safetensors"
|
14 |
|
15 |
use_strict = False
|
16 |
|
|
|
3 |
from auto_gptq import AutoGPTQForCausalLM, BaseQuantizeConfig
|
4 |
from huggingface_hub import snapshot_download
|
5 |
|
6 |
+
import shutil
|
7 |
+
import os
|
8 |
+
|
9 |
+
cwd = os.getcwd()
|
10 |
+
cachedir = cwd+'/cache'
|
11 |
+
os.mkdir(cachedir)
|
12 |
+
os.environ['HF_HOME'] = cachedir
|
13 |
+
|
14 |
+
local_folder = cachedir + "/model"
|
15 |
|
16 |
|
17 |
quantized_model_dir = "FPHam/Jackson_The_Formalizer_V2_13b_GPTQ"
|
18 |
|
19 |
snapshot_download(repo_id=quantized_model_dir, local_dir=local_folder, local_dir_use_symlinks=False)
|
20 |
|
21 |
+
model_basename = cachedir + "/model/Jackson2-4bit-128g-GPTQ.safetensors"
|
22 |
|
23 |
use_strict = False
|
24 |
|