Ravi21 commited on
Commit
9f34020
1 Parent(s): a00d624

Update model.py

Browse files
Files changed (1) hide show
  1. model.py +63 -0
model.py CHANGED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from threading import Thread
2
+ from typing import Iterator
3
+
4
+ import torch
5
+ from transformers import AutoConfig, AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
6
+ model="/kaggle/input/deberta-v3-large-hf-weights"
7
+ if torch.cuda.is_available():
8
+ config = AutoConfig.from_pretrained(model_id)
9
+ config.pretraining_tp = 1
10
+ model = AutoModelForCausalLM.from_pretrained(
11
+ model_id,
12
+ config=config,
13
+ torch_dtype=torch.float16,
14
+ load_in_4bit=True,
15
+ device_map='auto'
16
+ )
17
+ else:
18
+ model = None
19
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
20
+ def preprocess(sample):
21
+ first_sentences = [sample["prompt"]] * 5
22
+ second_sentences = [sample[option] for option in "ABCDE"]
23
+ tokenized_sentences = tokenizer(first_sentences, second_sentences, truncation=True, padding=True, return_tensors="pt")
24
+ sample["input_ids"] = tokenized_sentences["input_ids"]
25
+ sample["attention_mask"] = tokenized_sentences["attention_mask"]
26
+ return sample
27
+
28
+ def get_input_token_length(message: str, chat_history: list[tuple[str, str]], system_prompt: str) -> int:
29
+ prompt = get_prompt(message, chat_history, system_prompt)
30
+ input_ids = tokenizer([prompt], return_tensors='np', add_special_tokens=False)['input_ids']
31
+ return input_ids.shape[-1]
32
+
33
+ def run(message: str,
34
+ chat_history: list[tuple[str, str]],
35
+ system_prompt: str,
36
+ max_new_tokens: int = 1024,
37
+ temperature: float = 0.8,
38
+ top_p: float = 0.95,
39
+ top_k: int = 50) -> Iterator[str]:
40
+ prompt = get_prompt(message, chat_history, system_prompt)
41
+ inputs = tokenizer([prompt], return_tensors='pt', add_special_tokens=False).to('cuda')
42
+
43
+ streamer = TextIteratorStreamer(tokenizer,
44
+ timeout=10.,
45
+ skip_prompt=True,
46
+ skip_special_tokens=True)
47
+ generate_kwargs = dict(
48
+ inputs,
49
+ streamer=streamer,
50
+ max_new_tokens=max_new_tokens,
51
+ do_sample=True,
52
+ top_p=top_p,
53
+ top_k=top_k,
54
+ temperature=temperature,
55
+ num_beams=1,
56
+ )
57
+ t = Thread(target=model.generate, kwargs=generate_kwargs)
58
+ t.start()
59
+
60
+ outputs = []
61
+ for text in streamer:
62
+ outputs.append(text)
63
+ yield ''.join(outputs)