Update README.md
Browse files
README.md
CHANGED
@@ -68,8 +68,8 @@ The model can be used with HuggingFace's `transformers` library:
|
|
68 |
import torch
|
69 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
70 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
71 |
-
model = AutoModelForCausalLM.from_pretrained("
|
72 |
-
tokenizer = AutoTokenizer.from_pretrained("
|
73 |
def generate_response(message: str, temperature: float = 0.4, repetition_penalty: float = 1.12) -> str:
|
74 |
# Convert message to PyTorch tensors
|
75 |
input_ids = tokenizer.encode(
|
|
|
68 |
import torch
|
69 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
70 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
71 |
+
model = AutoModelForCausalLM.from_pretrained("OuteAI/Lite-Oute-1-65M").to(device)
|
72 |
+
tokenizer = AutoTokenizer.from_pretrained("OuteAI/Lite-Oute-1-65M")
|
73 |
def generate_response(message: str, temperature: float = 0.4, repetition_penalty: float = 1.12) -> str:
|
74 |
# Convert message to PyTorch tensors
|
75 |
input_ids = tokenizer.encode(
|