from langchain_openai import OpenAI | |
from transformers import AutoModelForCausalLM | |
import os | |
os.environ["TOKENIZERS_PARALLELISM"] = "true" | |
if not "OPENAI_API_KEY" in os.environ: | |
from config_key import OPENAI_API_KEY | |
os.environ['OPENAI_API_KEY'] = OPENAI_API_KEY | |
openai_llm = OpenAI(temperature=0, model="gpt-3.5-turbo-instruct") | |
# opensource_llm = AutoModelForCausalLM.from_pretrained("meta-llama/Llama-2-7b-hf") #LAMA MODEL | |