Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -10,7 +10,7 @@ from langchain.llms import HuggingFacePipeline
|
|
10 |
from langchain.memory import ConversationBufferMemory
|
11 |
|
12 |
from transformers import AutoTokenizer, pipeline
|
13 |
-
import
|
14 |
|
15 |
# Lista de modelos 100% abertos e gratuitos
|
16 |
list_llm = [
|
@@ -125,15 +125,15 @@ def demo():
|
|
125 |
outputs=[vector_db, process_status]
|
126 |
)
|
127 |
|
128 |
-
def load_model(model, temp):
|
129 |
-
if
|
130 |
return None, "Por favor, processe os documentos primeiro."
|
131 |
-
qa = initialize_llmchain(list_llm[list_llm_simple.index(model)], temp, 512, 3,
|
132 |
return qa, "Modelo carregado!"
|
133 |
|
134 |
load_model_btn.click(
|
135 |
load_model,
|
136 |
-
inputs=[model_selector, temperature],
|
137 |
outputs=[qa_chain, model_status]
|
138 |
)
|
139 |
|
|
|
10 |
from langchain.memory import ConversationBufferMemory
|
11 |
|
12 |
from transformers import AutoTokenizer, pipeline
|
13 |
+
import torch
|
14 |
|
15 |
# Lista de modelos 100% abertos e gratuitos
|
16 |
list_llm = [
|
|
|
125 |
outputs=[vector_db, process_status]
|
126 |
)
|
127 |
|
128 |
+
def load_model(model, temp, vector_db_state):
|
129 |
+
if vector_db_state is None:
|
130 |
return None, "Por favor, processe os documentos primeiro."
|
131 |
+
qa = initialize_llmchain(list_llm[list_llm_simple.index(model)], temp, 512, 3, vector_db_state)
|
132 |
return qa, "Modelo carregado!"
|
133 |
|
134 |
load_model_btn.click(
|
135 |
load_model,
|
136 |
+
inputs=[model_selector, temperature, vector_db],
|
137 |
outputs=[qa_chain, model_status]
|
138 |
)
|
139 |
|