Spaces:
Paused
Paused
Commit
·
a110fec
1
Parent(s):
060489c
Update app.py
Browse files
app.py
CHANGED
@@ -16,7 +16,7 @@ local_folder = cachedir + "/model"
|
|
16 |
|
17 |
quantized_model_dir = "FPHam/Jackson_The_Formalizer_V2_13b_GPTQ"
|
18 |
|
19 |
-
|
20 |
|
21 |
model_basename = cachedir + "/model/Jackson2-4bit-128g-GPTQ"
|
22 |
|
@@ -24,7 +24,7 @@ use_strict = False
|
|
24 |
|
25 |
use_triton = False
|
26 |
|
27 |
-
tokenizer = AutoTokenizer.from_pretrained(
|
28 |
|
29 |
quantize_config = BaseQuantizeConfig(
|
30 |
bits=4,
|
@@ -32,7 +32,7 @@ quantize_config = BaseQuantizeConfig(
|
|
32 |
desc_act=False
|
33 |
)
|
34 |
|
35 |
-
model = AutoGPTQForCausalLM.from_quantized(
|
36 |
use_safetensors=True,
|
37 |
strict=use_strict,
|
38 |
model_basename=model_basename,
|
|
|
16 |
|
17 |
quantized_model_dir = "FPHam/Jackson_The_Formalizer_V2_13b_GPTQ"
|
18 |
|
19 |
+
snapshot_download(repo_id=quantized_model_dir, local_dir=local_folder, local_dir_use_symlinks=True)
|
20 |
|
21 |
model_basename = cachedir + "/model/Jackson2-4bit-128g-GPTQ"
|
22 |
|
|
|
24 |
|
25 |
use_triton = False
|
26 |
|
27 |
+
tokenizer = AutoTokenizer.from_pretrained(local_foler, use_fast=False)
|
28 |
|
29 |
quantize_config = BaseQuantizeConfig(
|
30 |
bits=4,
|
|
|
32 |
desc_act=False
|
33 |
)
|
34 |
|
35 |
+
model = AutoGPTQForCausalLM.from_quantized(local_folder,
|
36 |
use_safetensors=True,
|
37 |
strict=use_strict,
|
38 |
model_basename=model_basename,
|