correct model name, and add use_safetensors=False in from_quanted
Browse files- web_quant.py +2 -2
web_quant.py
CHANGED
@@ -9,7 +9,7 @@ import gradio as gr
|
|
9 |
from threading import Thread
|
10 |
|
11 |
def load_model(model_name):
|
12 |
-
model = AutoGPTQForCausalLM.from_quantized(model_name, device_map="auto")
|
13 |
tokenizer = AutoTokenizer.from_pretrained(model_name, padding_side="right", use_fast=False)
|
14 |
return model, tokenizer
|
15 |
|
@@ -108,6 +108,6 @@ def main(args):
|
|
108 |
|
109 |
if __name__ == '__main__':
|
110 |
parser = argparse.ArgumentParser()
|
111 |
-
parser.add_argument("--model-name", type=str, default="FreedomIntelligence/AceGPT-
|
112 |
args = parser.parse_args()
|
113 |
main(args)
|
|
|
9 |
from threading import Thread
|
10 |
|
11 |
def load_model(model_name):
|
12 |
+
model = AutoGPTQForCausalLM.from_quantized(model_name, device_map="auto", use_safetensors=False)
|
13 |
tokenizer = AutoTokenizer.from_pretrained(model_name, padding_side="right", use_fast=False)
|
14 |
return model, tokenizer
|
15 |
|
|
|
108 |
|
109 |
if __name__ == '__main__':
|
110 |
parser = argparse.ArgumentParser()
|
111 |
+
parser.add_argument("--model-name", type=str, default="FreedomIntelligence/AceGPT-7b-chat-GPTQ")
|
112 |
args = parser.parse_args()
|
113 |
main(args)
|