matteosz commited on
Commit
206fd23
·
1 Parent(s): edd396d

revert unquantized

Browse files
Files changed (1) hide show
  1. app.py +11 -4
app.py CHANGED
@@ -1,15 +1,22 @@
1
  import gradio as gr
2
- from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
 
 
3
 
4
- checkpoint = 'matteosz/phi3-mini-4k-sft-dpo-quant'
5
 
6
  tokenizer = AutoTokenizer.from_pretrained(checkpoint, trust_remote_code=True)
7
  tokenizer.model_max_length = 256
8
- model = AutoModelForCausalLM.from_pretrained(
 
 
 
 
9
  checkpoint,
10
  trust_remote_code=True,
11
  device_map='auto',
12
- ).eval()
 
13
 
14
  pipe = pipeline(
15
  "text-generation",
 
1
  import gradio as gr
2
+ from transformers import AutoTokenizer, pipeline
3
+ from peft import AutoPeftModelForCausalLM
4
+ import torch
5
 
6
+ checkpoint = 'ernestoBocini/Phi3-mini-DPO-Tuned'
7
 
8
  tokenizer = AutoTokenizer.from_pretrained(checkpoint, trust_remote_code=True)
9
  tokenizer.model_max_length = 256
10
+ tokenizer.pad_token = tokenizer.unk_token
11
+ tokenizer.pad_token_id = tokenizer.convert_tokens_to_ids(tokenizer.pad_token)
12
+ tokenizer.padding_side = 'left'
13
+
14
+ model = AutoPeftModelForCausalLM.from_pretrained(
15
  checkpoint,
16
  trust_remote_code=True,
17
  device_map='auto',
18
+ torch_dtype=torch.bfloat16
19
+ ).merge_and_unload().to(torch.float16).eval()
20
 
21
  pipe = pipeline(
22
  "text-generation",