LLMproj1 commited on
Commit
014c274
·
verified ·
1 Parent(s): 0f8ec6d

Update model.py

Browse files
Files changed (1) hide show
  1. model.py +3 -3
model.py CHANGED
@@ -13,7 +13,7 @@ os.system('pip install "unsloth[colab-new] @ git+https://github.com/unslothai/un
13
 
14
  import gradio as gr
15
 
16
- from transformers import TextStreamer
17
  from unsloth import FastLanguageModel
18
  from google.colab import drive
19
 
@@ -52,10 +52,10 @@ def chat_alpaca(message: str, history: list, temperature: float, max_new_tokens:
52
  inputs = tokenizer([prompt], return_tensors="pt").to("cuda")
53
 
54
  # Define the streamer
55
- text_streamer = TextStreamer(tokenizer)
56
 
57
  # Generate the response
58
- outputs = model.generate(**inputs, streamer=text_streamer, max_new_tokens=max_new_tokens, temperature=temperature)
59
  response = tokenizer.batch_decode(outputs, skip_special_tokens=True)[0]
60
 
61
  # Return the response
 
13
 
14
  import gradio as gr
15
 
16
+ # from transformers import TextStreamer
17
  from unsloth import FastLanguageModel
18
  from google.colab import drive
19
 
 
52
  inputs = tokenizer([prompt], return_tensors="pt").to("cuda")
53
 
54
  # Define the streamer
55
+ # text_streamer = TextStreamer(tokenizer)
56
 
57
  # Generate the response
58
+ outputs = model.generate(**inputs, max_new_tokens=max_new_tokens, temperature=temperature)
59
  response = tokenizer.batch_decode(outputs, skip_special_tokens=True)[0]
60
 
61
  # Return the response