Update README.md
Browse files
README.md
CHANGED
@@ -59,14 +59,11 @@ To run inference with the model in full `bfloat16` precision you need approximat
|
|
59 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
60 |
import transformers
|
61 |
import torch
|
62 |
-
from peft import PeftModel
|
63 |
|
64 |
-
model = "
|
65 |
-
adapters_name = 'vilm/vulture-40B'
|
66 |
|
67 |
tokenizer = AutoTokenizer.from_pretrained(model)
|
68 |
m = AutoModelForCausalLM.from_pretrained(model, torch_dtype=torch.bfloat16, device_map="auto" )
|
69 |
-
m = PeftModel.from_pretrained(m, adapters_name)
|
70 |
|
71 |
prompt = "A chat between a curious user and an artificial intelligence assistant.\n\nUSER:Thành phố Hồ Chí Minh nằm ở đâu?<|endoftext|>ASSISTANT:"
|
72 |
|
|
|
59 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
60 |
import transformers
|
61 |
import torch
|
|
|
62 |
|
63 |
+
model = "vilm/vulture-40B"
|
|
|
64 |
|
65 |
tokenizer = AutoTokenizer.from_pretrained(model)
|
66 |
m = AutoModelForCausalLM.from_pretrained(model, torch_dtype=torch.bfloat16, device_map="auto" )
|
|
|
67 |
|
68 |
prompt = "A chat between a curious user and an artificial intelligence assistant.\n\nUSER:Thành phố Hồ Chí Minh nằm ở đâu?<|endoftext|>ASSISTANT:"
|
69 |
|