Spaces:
Sleeping
Sleeping
Commit
·
bae5b02
1
Parent(s):
500107e
Update app.py
Browse files
app.py
CHANGED
@@ -4,7 +4,8 @@ import requests
|
|
4 |
import csv
|
5 |
import os
|
6 |
import langchain
|
7 |
-
import
|
|
|
8 |
|
9 |
|
10 |
|
@@ -63,7 +64,7 @@ def submit_message(prompt, prompt_template, temperature, max_tokens, context_len
|
|
63 |
|
64 |
# completion = openai.ChatCompletion.create(model="gpt-3.5-turbo", messages=system_prompt + history[-context_length*2:] + [prompt_msg], temperature=temperature, max_tokens=max_tokens)
|
65 |
|
66 |
-
# vectordb = "
|
67 |
|
68 |
|
69 |
|
@@ -115,10 +116,7 @@ with gr.Blocks(css=css) as demo:
|
|
115 |
Ask questions of all of them, or pick your guru.""",
|
116 |
elem_id="header")
|
117 |
|
118 |
-
|
119 |
-
dirs = os.listdir( path )
|
120 |
-
for file in dirs:
|
121 |
-
gr.Markdown(file)
|
122 |
|
123 |
|
124 |
with gr.Row():
|
|
|
4 |
import csv
|
5 |
import os
|
6 |
import langchain
|
7 |
+
import chromadb
|
8 |
+
#import faiss
|
9 |
|
10 |
|
11 |
|
|
|
64 |
|
65 |
# completion = openai.ChatCompletion.create(model="gpt-3.5-turbo", messages=system_prompt + history[-context_length*2:] + [prompt_msg], temperature=temperature, max_tokens=max_tokens)
|
66 |
|
67 |
+
# vectordb = "./embeddings"
|
68 |
|
69 |
|
70 |
|
|
|
116 |
Ask questions of all of them, or pick your guru.""",
|
117 |
elem_id="header")
|
118 |
|
119 |
+
|
|
|
|
|
|
|
120 |
|
121 |
|
122 |
with gr.Row():
|