Update README.md
Browse files
README.md
CHANGED
@@ -63,8 +63,8 @@ dtype: bfloat16
|
|
63 |
from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer
|
64 |
|
65 |
# Model & Tokenizer
|
66 |
-
MODEL_NAME =
|
67 |
-
model = AutoModelForCausalLM.from_pretrained(MODEL_NAME, device_map='cuda:
|
68 |
tok = AutoTokenizer.from_pretrained(MODEL_NAME)
|
69 |
|
70 |
# Inference
|
@@ -84,7 +84,7 @@ from transformers import AutoTokenizer
|
|
84 |
import transformers
|
85 |
import torch
|
86 |
|
87 |
-
model =
|
88 |
|
89 |
tokenizer = AutoTokenizer.from_pretrained(model)
|
90 |
pipeline = transformers.pipeline(
|
|
|
63 |
from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer
|
64 |
|
65 |
# Model & Tokenizer
|
66 |
+
MODEL_NAME = 'Kukedlc/NeuralTopBench-7B-ties'
|
67 |
+
model = AutoModelForCausalLM.from_pretrained(MODEL_NAME, device_map='cuda:0', load_in_4bit=True)
|
68 |
tok = AutoTokenizer.from_pretrained(MODEL_NAME)
|
69 |
|
70 |
# Inference
|
|
|
84 |
import transformers
|
85 |
import torch
|
86 |
|
87 |
+
model = 'Kukedlc/NeuralTopBench-7B-ties'
|
88 |
|
89 |
tokenizer = AutoTokenizer.from_pretrained(model)
|
90 |
pipeline = transformers.pipeline(
|