Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -2,15 +2,8 @@ import gradio as gr
|
|
2 |
import torch
|
3 |
from transformers import AutoTokenizer, AutoModelForCausalLM, AutoConfig
|
4 |
|
5 |
-
tokenizer = AutoTokenizer.from_pretrained("
|
6 |
-
|
7 |
-
# # config.quantization_config["use_exllama"] = True
|
8 |
-
# config.quantization_config["disable_exllama"] = True
|
9 |
-
# config.quantization_config["exllama_config"] = {"version":2}
|
10 |
-
# # model = AutoModelForCausalLM.from_config(config, trust_remote_code=True)
|
11 |
-
config = AutoConfig.from_pretrained("TheBloke/SOLAR-10.7B-Instruct-v1.0-uncensored-GPTQ")
|
12 |
-
config.quantization_config["disable_exllama"] = True
|
13 |
-
model = AutoModelForCausalLM.from_config(config)
|
14 |
|
15 |
def generate_response(prompt):
|
16 |
conversation = [{'role': 'user', 'content': prompt}]
|
|
|
2 |
import torch
|
3 |
from transformers import AutoTokenizer, AutoModelForCausalLM, AutoConfig
|
4 |
|
5 |
+
tokenizer = AutoTokenizer.from_pretrained("kimnt93/chat-llama2-1b-1.0")
|
6 |
+
model = AutoModelForCausalLM.from_pretrained("kimnt93/chat-llama2-1b-1.0")
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
7 |
|
8 |
def generate_response(prompt):
|
9 |
conversation = [{'role': 'user', 'content': prompt}]
|