phongtran commited on
Commit
f83af8e
·
1 Parent(s): 15c194c
Files changed (1) hide show
  1. app.py +2 -3
app.py CHANGED
@@ -6,7 +6,6 @@ import os
6
  model_path = "vinai/PhoGPT-7B5-Instruct"
7
 
8
  config = AutoConfig.from_pretrained(model_path, trust_remote_code=True, token=os.environ['HK_TOKEN'])
9
- config.init_device = "cuda"
10
  # config.attn_config['attn_impl'] = 'triton' # Enable if "triton" installed!
11
 
12
  model = AutoModelForCausalLM.from_pretrained(
@@ -23,8 +22,8 @@ def answer(input_prompt):
23
  input_ids = tokenizer(input_prompt, return_tensors="pt")
24
 
25
  outputs = model.generate(
26
- inputs=input_ids["input_ids"].to("cuda"),
27
- attention_mask=input_ids["attention_mask"].to("cuda"),
28
  do_sample=True,
29
  temperature=1.0,
30
  top_k=50,
 
6
  model_path = "vinai/PhoGPT-7B5-Instruct"
7
 
8
  config = AutoConfig.from_pretrained(model_path, trust_remote_code=True, token=os.environ['HK_TOKEN'])
 
9
  # config.attn_config['attn_impl'] = 'triton' # Enable if "triton" installed!
10
 
11
  model = AutoModelForCausalLM.from_pretrained(
 
22
  input_ids = tokenizer(input_prompt, return_tensors="pt")
23
 
24
  outputs = model.generate(
25
+ inputs=input_ids["input_ids"].to("cpu"),
26
+ attention_mask=input_ids["attention_mask"].to("cpu"),
27
  do_sample=True,
28
  temperature=1.0,
29
  top_k=50,