Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -804,7 +804,10 @@ Standalone question: {question}""")
|
|
804 |
chain = ConversationalRetrievalChain.from_llm(
|
805 |
condense_question_prompt=standalone_question_prompt,
|
806 |
combine_docs_chain_kwargs={'prompt': answer_prompt},
|
807 |
-
condense_question_llm=condense_question_llm,
|
|
|
|
|
|
|
808 |
|
809 |
memory=memory,
|
810 |
retriever = base_retriever_HF, #changed this
|
@@ -813,7 +816,8 @@ Standalone question: {question}""")
|
|
813 |
llm=instantiate_LLM(
|
814 |
LLM_provider="Google",api_key=google_api_key,temperature=0.5,
|
815 |
model_name="gemini-pro"),
|
816 |
-
chain_type=
|
|
|
817 |
verbose= True,
|
818 |
return_source_documents=True
|
819 |
)
|
@@ -826,8 +830,6 @@ Standalone question: {question}""")
|
|
826 |
|
827 |
|
828 |
|
829 |
-
|
830 |
-
|
831 |
def submit_message(prompt, prompt_template, temperature, max_tokens, context_length, state):
|
832 |
|
833 |
|
|
|
804 |
chain = ConversationalRetrievalChain.from_llm(
|
805 |
condense_question_prompt=standalone_question_prompt,
|
806 |
combine_docs_chain_kwargs={'prompt': answer_prompt},
|
807 |
+
#condense_question_llm=condense_question_llm,
|
808 |
+
condense_question_llm=instantiate_LLM(
|
809 |
+
LLM_provider="Google",api_key=google_api_key,temperature=0.1,
|
810 |
+
model_name="gemini-pro"),
|
811 |
|
812 |
memory=memory,
|
813 |
retriever = base_retriever_HF, #changed this
|
|
|
816 |
llm=instantiate_LLM(
|
817 |
LLM_provider="Google",api_key=google_api_key,temperature=0.5,
|
818 |
model_name="gemini-pro"),
|
819 |
+
chain_type= "stuff",
|
820 |
+
#chain_type= chain_type,
|
821 |
verbose= True,
|
822 |
return_source_documents=True
|
823 |
)
|
|
|
830 |
|
831 |
|
832 |
|
|
|
|
|
833 |
def submit_message(prompt, prompt_template, temperature, max_tokens, context_length, state):
|
834 |
|
835 |
|