shibing624 commited on
Commit
f2a5f61
1 Parent(s): 007f62d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -4
README.md CHANGED
@@ -104,21 +104,20 @@ First, you pass your input through the transformer model, then you get the gener
104
 
105
  Install package:
106
  ```
107
- pip install transformers
108
  ```
109
 
110
  ```python
111
  import sys
 
112
  from peft import PeftModel
113
  from transformers import AutoModelForCausalLM, AutoTokenizer, GenerationConfig
114
 
115
 
116
  model = AutoModelForCausalLM.from_pretrained("baichuan-inc/Baichuan-13B-Chat", device_map='auto', trust_remote_code=True)
117
  model.generation_config = GenerationConfig.from_pretrained("baichuan-inc/Baichuan-13B-Chat", trust_remote_code=True)
118
-
119
  tokenizer = AutoTokenizer.from_pretrained("baichuan-inc/Baichuan-13B-Chat", trust_remote_code=True)
120
  model = PeftModel.from_pretrained(model, "shibing624/vicuna-baichuan-13b-chat-lora")
121
-
122
  device = "cuda" if torch.cuda.is_available() else "cpu"
123
 
124
  def generate_prompt(instruction):
@@ -131,7 +130,7 @@ for s in sents:
131
  inputs = tokenizer(q, return_tensors="pt")
132
  inputs = inputs.to(device=device)
133
 
134
- generate_ids = ref_model.generate(
135
  **inputs,
136
  max_new_tokens=120,
137
  )
 
104
 
105
  Install package:
106
  ```
107
+ pip install transformers -U
108
  ```
109
 
110
  ```python
111
  import sys
112
+ import torch
113
  from peft import PeftModel
114
  from transformers import AutoModelForCausalLM, AutoTokenizer, GenerationConfig
115
 
116
 
117
  model = AutoModelForCausalLM.from_pretrained("baichuan-inc/Baichuan-13B-Chat", device_map='auto', trust_remote_code=True)
118
  model.generation_config = GenerationConfig.from_pretrained("baichuan-inc/Baichuan-13B-Chat", trust_remote_code=True)
 
119
  tokenizer = AutoTokenizer.from_pretrained("baichuan-inc/Baichuan-13B-Chat", trust_remote_code=True)
120
  model = PeftModel.from_pretrained(model, "shibing624/vicuna-baichuan-13b-chat-lora")
 
121
  device = "cuda" if torch.cuda.is_available() else "cpu"
122
 
123
  def generate_prompt(instruction):
 
130
  inputs = tokenizer(q, return_tensors="pt")
131
  inputs = inputs.to(device=device)
132
 
133
+ generate_ids = model.generate(
134
  **inputs,
135
  max_new_tokens=120,
136
  )