qnguyen3 commited on
Commit
f4a0293
·
1 Parent(s): 30dbf29

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -4
README.md CHANGED
@@ -59,14 +59,11 @@ To run inference with the model in full `bfloat16` precision you need approximat
59
  from transformers import AutoTokenizer, AutoModelForCausalLM
60
  import transformers
61
  import torch
62
- from peft import PeftModel
63
 
64
- model = "tiiuae/falcon-40B"
65
- adapters_name = 'vilm/vulture-40B'
66
 
67
  tokenizer = AutoTokenizer.from_pretrained(model)
68
  m = AutoModelForCausalLM.from_pretrained(model, torch_dtype=torch.bfloat16, device_map="auto" )
69
- m = PeftModel.from_pretrained(m, adapters_name)
70
 
71
  prompt = "A chat between a curious user and an artificial intelligence assistant.\n\nUSER:Thành phố Hồ Chí Minh nằm ở đâu?<|endoftext|>ASSISTANT:"
72
 
 
59
  from transformers import AutoTokenizer, AutoModelForCausalLM
60
  import transformers
61
  import torch
 
62
 
63
+ model = "vilm/vulture-40B"
 
64
 
65
  tokenizer = AutoTokenizer.from_pretrained(model)
66
  m = AutoModelForCausalLM.from_pretrained(model, torch_dtype=torch.bfloat16, device_map="auto" )
 
67
 
68
  prompt = "A chat between a curious user and an artificial intelligence assistant.\n\nUSER:Thành phố Hồ Chí Minh nằm ở đâu?<|endoftext|>ASSISTANT:"
69