shibing624
commited on
Commit
•
f2a5f61
1
Parent(s):
007f62d
Update README.md
Browse files
README.md
CHANGED
@@ -104,21 +104,20 @@ First, you pass your input through the transformer model, then you get the gener
|
|
104 |
|
105 |
Install package:
|
106 |
```
|
107 |
-
pip install transformers
|
108 |
```
|
109 |
|
110 |
```python
|
111 |
import sys
|
|
|
112 |
from peft import PeftModel
|
113 |
from transformers import AutoModelForCausalLM, AutoTokenizer, GenerationConfig
|
114 |
|
115 |
|
116 |
model = AutoModelForCausalLM.from_pretrained("baichuan-inc/Baichuan-13B-Chat", device_map='auto', trust_remote_code=True)
|
117 |
model.generation_config = GenerationConfig.from_pretrained("baichuan-inc/Baichuan-13B-Chat", trust_remote_code=True)
|
118 |
-
|
119 |
tokenizer = AutoTokenizer.from_pretrained("baichuan-inc/Baichuan-13B-Chat", trust_remote_code=True)
|
120 |
model = PeftModel.from_pretrained(model, "shibing624/vicuna-baichuan-13b-chat-lora")
|
121 |
-
|
122 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
123 |
|
124 |
def generate_prompt(instruction):
|
@@ -131,7 +130,7 @@ for s in sents:
|
|
131 |
inputs = tokenizer(q, return_tensors="pt")
|
132 |
inputs = inputs.to(device=device)
|
133 |
|
134 |
-
generate_ids =
|
135 |
**inputs,
|
136 |
max_new_tokens=120,
|
137 |
)
|
|
|
104 |
|
105 |
Install package:
|
106 |
```
|
107 |
+
pip install transformers -U
|
108 |
```
|
109 |
|
110 |
```python
|
111 |
import sys
|
112 |
+
import torch
|
113 |
from peft import PeftModel
|
114 |
from transformers import AutoModelForCausalLM, AutoTokenizer, GenerationConfig
|
115 |
|
116 |
|
117 |
model = AutoModelForCausalLM.from_pretrained("baichuan-inc/Baichuan-13B-Chat", device_map='auto', trust_remote_code=True)
|
118 |
model.generation_config = GenerationConfig.from_pretrained("baichuan-inc/Baichuan-13B-Chat", trust_remote_code=True)
|
|
|
119 |
tokenizer = AutoTokenizer.from_pretrained("baichuan-inc/Baichuan-13B-Chat", trust_remote_code=True)
|
120 |
model = PeftModel.from_pretrained(model, "shibing624/vicuna-baichuan-13b-chat-lora")
|
|
|
121 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
122 |
|
123 |
def generate_prompt(instruction):
|
|
|
130 |
inputs = tokenizer(q, return_tensors="pt")
|
131 |
inputs = inputs.to(device=device)
|
132 |
|
133 |
+
generate_ids = model.generate(
|
134 |
**inputs,
|
135 |
max_new_tokens=120,
|
136 |
)
|