Update chatbot_bedrock.py
Browse files- chatbot_bedrock.py +2 -2
chatbot_bedrock.py
CHANGED
@@ -7,14 +7,14 @@ from transformers import AutoModelForCausalLM, AutoTokenizer
|
|
7 |
import streamlit as st
|
8 |
from langchain_community.llms import HuggingFaceHub
|
9 |
|
10 |
-
|
11 |
|
12 |
@st.cache_resource
|
13 |
def load_model():
|
14 |
#tokenizer = AutoTokenizer.from_pretrained("KvrParaskevi/Hotel-Assistant-Attempt4-Llama-2-7b")
|
15 |
#model = AutoModelForCausalLM.from_pretrained("KvrParaskevi/Hotel-Assistant-Attempt4-Llama-2-7b")
|
16 |
model = HuggingFaceHub(
|
17 |
-
repo_id=
|
18 |
task="text-generation",
|
19 |
model_kwargs={
|
20 |
"max_new_tokens": 512,
|
|
|
7 |
import streamlit as st
|
8 |
from langchain_community.llms import HuggingFaceHub
|
9 |
|
10 |
+
my_model_id = os.getenv('MODEL_REPO_ID', 'Default Value')
|
11 |
|
12 |
@st.cache_resource
|
13 |
def load_model():
|
14 |
#tokenizer = AutoTokenizer.from_pretrained("KvrParaskevi/Hotel-Assistant-Attempt4-Llama-2-7b")
|
15 |
#model = AutoModelForCausalLM.from_pretrained("KvrParaskevi/Hotel-Assistant-Attempt4-Llama-2-7b")
|
16 |
model = HuggingFaceHub(
|
17 |
+
repo_id=my_model_id,
|
18 |
task="text-generation",
|
19 |
model_kwargs={
|
20 |
"max_new_tokens": 512,
|