alfredplpl commited on
Commit
7572cf2
·
verified ·
1 Parent(s): 9ab7453

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -1
app.py CHANGED
@@ -45,7 +45,7 @@ h1 {
45
 
46
  # Load the tokenizer and model
47
  tokenizer = AutoTokenizer.from_pretrained("llm-jp/llm-jp-13b-instruct-full-ac_001_16x-dolly-ichikara_004_001_single-oasst-oasst2-v2.0")
48
- model = AutoModelForCausalLM.from_pretrained("llm-jp/llm-jp-13b-instruct-full-ac_001_16x-dolly-ichikara_004_001_single-oasst-oasst2-v2.0", device_map="cuda", torch_dtype=torch.bfloat16)
49
  model=model.eval()
50
 
51
  @spaces.GPU()
@@ -70,6 +70,8 @@ def chat_llm_jp_v2(message: str,
70
  conversation.extend([{"role": "user", "content": user}, {"role": "assistant", "content": assistant}])
71
  conversation.append({"role": "user", "content": message})
72
 
 
 
73
  streamer = TextIteratorStreamer(tokenizer, timeout=10.0, skip_prompt=True, skip_special_tokens=True)
74
 
75
  generate_kwargs = dict(
 
45
 
46
  # Load the tokenizer and model
47
  tokenizer = AutoTokenizer.from_pretrained("llm-jp/llm-jp-13b-instruct-full-ac_001_16x-dolly-ichikara_004_001_single-oasst-oasst2-v2.0")
48
+ model = AutoModelForCausalLM.from_pretrained("llm-jp/llm-jp-13b-instruct-full-ac_001_16x-dolly-ichikara_004_001_single-oasst-oasst2-v2.0", device_map="auto", torch_dtype=torch.bfloat16)
49
  model=model.eval()
50
 
51
  @spaces.GPU()
 
70
  conversation.extend([{"role": "user", "content": user}, {"role": "assistant", "content": assistant}])
71
  conversation.append({"role": "user", "content": message})
72
 
73
+ input_ids = tokenizer.apply_chat_template(conversation, return_tensors="pt").to(model.device)
74
+
75
  streamer = TextIteratorStreamer(tokenizer, timeout=10.0, skip_prompt=True, skip_special_tokens=True)
76
 
77
  generate_kwargs = dict(