Text Generation
Transformers
PyTorch
Safetensors
English
llama
biology
medical
text-generation-inference
Inference Endpoints
AdaptLLM commited on
Commit
9871a1a
1 Parent(s): 41b951f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -45,8 +45,8 @@ For example, to chat with the biomedicine model:
45
  ```python
46
  from transformers import AutoModelForCausalLM, AutoTokenizer
47
 
48
- model = AutoModelForCausalLM.from_pretrained("AdaptLLM/medicine-chat")
49
- tokenizer = AutoTokenizer.from_pretrained("AdaptLLM/medicine-chat", use_fast=False)
50
 
51
  # Put your input here:
52
  user_input = '''Question: Which of the following is an example of monosomy?
@@ -58,11 +58,11 @@ Options:
58
 
59
  Please provide your choice first and then provide explanations if possible.'''
60
 
61
- # We use the prompt template of LLaMA-2-Chat demo
62
- prompt = f"<s>[INST] <<SYS>>\nYou are a helpful, respectful and honest assistant. Always answer as helpfully as possible, while being safe. Your answers should not include any harmful, unethical, racist, sexist, toxic, dangerous, or illegal content. Please ensure that your responses are socially unbiased and positive in nature.\n\nIf a question does not make any sense, or is not factually coherent, explain why instead of answering something not correct. If you don't know the answer to a question, please don't share false information.\n<</SYS>>\n\n{user_input} [/INST]"
63
 
64
  inputs = tokenizer(prompt, return_tensors="pt", add_special_tokens=False).input_ids.to(model.device)
65
- outputs = model.generate(input_ids=inputs, max_length=4096)[0]
66
 
67
  answer_start = int(inputs.shape[-1])
68
  pred = tokenizer.decode(outputs[answer_start:], skip_special_tokens=True)
 
45
  ```python
46
  from transformers import AutoModelForCausalLM, AutoTokenizer
47
 
48
+ model = AutoModelForCausalLM.from_pretrained("AdaptLLM/medicine-LLM-13B")
49
+ tokenizer = AutoTokenizer.from_pretrained("AdaptLLM/medicine-LLM-13B", use_fast=False)
50
 
51
  # Put your input here:
52
  user_input = '''Question: Which of the following is an example of monosomy?
 
58
 
59
  Please provide your choice first and then provide explanations if possible.'''
60
 
61
+ # Simply use your input as the prompt
62
+ prompt = user_input
63
 
64
  inputs = tokenizer(prompt, return_tensors="pt", add_special_tokens=False).input_ids.to(model.device)
65
+ outputs = model.generate(input_ids=inputs, max_length=2048)[0]
66
 
67
  answer_start = int(inputs.shape[-1])
68
  pred = tokenizer.decode(outputs[answer_start:], skip_special_tokens=True)