ArvindSharma18 commited on
Commit
ed966df
1 Parent(s): aa8ab9f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -7
README.md CHANGED
@@ -21,18 +21,17 @@ tags:
21
  ```python
22
  from unsloth import FastLanguageModel
23
  import torch
24
- max_seq_length = 4096 # Choose any! We auto support RoPE Scaling internally!
25
- dtype = torch.float16 # None for auto detection. Float16 for Tesla T4, V100, Bfloat16 for Ampere+
26
- load_in_4bit = True # Use 4bit quantization to reduce memory usage. Can be False.
27
 
28
  model, tokenizer = FastLanguageModel.from_pretrained(
29
- model_name = "ArvindSharma18/Phi-3-mini-4k-instruct-bnb-4bit-Clinical-Trail-Merged", # "unsloth/mistral-7b" for 16bit loading
30
  max_seq_length = max_seq_length,
31
  dtype = dtype,
32
- load_in_4bit = load_in_4bit,
33
- # token = "hf_...", # use one if using gated models like meta-llama/Llama-2-7b-hf
34
  )
35
- FastLanguageModel.for_inference(model) # Enable native 2x faster inference
36
  inputs = tokenizer(
37
  [
38
  "Official Title: Randomized Trial of Usual Care vs. Specialized, Phase-specific Care for Youth at Risk for Psychosis"
 
21
  ```python
22
  from unsloth import FastLanguageModel
23
  import torch
24
+ max_seq_length = 4096
25
+ dtype = torch.float16
26
+ load_in_4bit = True
27
 
28
  model, tokenizer = FastLanguageModel.from_pretrained(
29
+ model_name = "ArvindSharma18/Phi-3-mini-4k-instruct-bnb-4bit-Clinical-Trail-Merged",
30
  max_seq_length = max_seq_length,
31
  dtype = dtype,
32
+ load_in_4bit = load_in_4bit
 
33
  )
34
+ FastLanguageModel.for_inference(model)
35
  inputs = tokenizer(
36
  [
37
  "Official Title: Randomized Trial of Usual Care vs. Specialized, Phase-specific Care for Youth at Risk for Psychosis"