Spaces:
Sleeping
Sleeping
Upload app.py with huggingface_hub
Browse files
app.py
CHANGED
@@ -1,12 +1,12 @@
|
|
1 |
import os
|
2 |
import gradio as gr
|
3 |
-
from
|
4 |
from langchain import LLMChain, PromptTemplate
|
5 |
from langchain.memory import ConversationBufferMemory
|
6 |
|
7 |
OPENAI_API_KEY=os.getenv('OPENAI_API_KEY')
|
8 |
|
9 |
-
template = """
|
10 |
{chat_history}
|
11 |
User: {user_message}
|
12 |
Chatbot:"""
|
@@ -18,7 +18,7 @@ prompt = PromptTemplate(
|
|
18 |
memory = ConversationBufferMemory(memory_key="chat_history")
|
19 |
|
20 |
llm_chain = LLMChain(
|
21 |
-
llm=ChatOpenAI(temperature='
|
22 |
prompt=prompt,
|
23 |
verbose=True,
|
24 |
memory=memory,
|
@@ -31,4 +31,4 @@ def get_text_response(user_message,history):
|
|
31 |
demo = gr.ChatInterface(get_text_response)
|
32 |
|
33 |
if __name__ == "__main__":
|
34 |
-
demo.launch() #To
|
|
|
1 |
import os
|
2 |
import gradio as gr
|
3 |
+
from langchain.chat_models import ChatOpenAI
|
4 |
from langchain import LLMChain, PromptTemplate
|
5 |
from langchain.memory import ConversationBufferMemory
|
6 |
|
7 |
OPENAI_API_KEY=os.getenv('OPENAI_API_KEY')
|
8 |
|
9 |
+
template = """You are a helpful assistant to answer all user queries.
|
10 |
{chat_history}
|
11 |
User: {user_message}
|
12 |
Chatbot:"""
|
|
|
18 |
memory = ConversationBufferMemory(memory_key="chat_history")
|
19 |
|
20 |
llm_chain = LLMChain(
|
21 |
+
llm=ChatOpenAI(temperature='0.5', model_name="gpt-3.5-turbo"),
|
22 |
prompt=prompt,
|
23 |
verbose=True,
|
24 |
memory=memory,
|
|
|
31 |
demo = gr.ChatInterface(get_text_response)
|
32 |
|
33 |
if __name__ == "__main__":
|
34 |
+
demo.launch() #To create a public link, set `share=True` in `launch()`. To enable errors and logs, set `debug=True` in `launch()`.
|