Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
@@ -3,14 +3,10 @@ import gradio as gr
|
|
3 |
import copy
|
4 |
import time
|
5 |
import llama_cpp
|
|
|
6 |
from llama_cpp import Llama
|
7 |
from huggingface_hub import hf_hub_download
|
8 |
|
9 |
-
with open("ingest.py", "r") as file:
|
10 |
-
code = file.read()
|
11 |
-
exec(code)
|
12 |
-
|
13 |
-
|
14 |
llm = Llama(
|
15 |
model_path=hf_hub_download(
|
16 |
repo_id=os.environ.get("REPO_ID", "TheBloke/Llama-2-7b-Chat-GGUF"),
|
@@ -72,4 +68,7 @@ demo = gr.ChatInterface(
|
|
72 |
clear_btn="Clear",
|
73 |
)
|
74 |
demo.queue(concurrency_count=1, max_size=5)
|
75 |
-
|
|
|
|
|
|
|
|
3 |
import copy
|
4 |
import time
|
5 |
import llama_cpp
|
6 |
+
import ingest
|
7 |
from llama_cpp import Llama
|
8 |
from huggingface_hub import hf_hub_download
|
9 |
|
|
|
|
|
|
|
|
|
|
|
10 |
llm = Llama(
|
11 |
model_path=hf_hub_download(
|
12 |
repo_id=os.environ.get("REPO_ID", "TheBloke/Llama-2-7b-Chat-GGUF"),
|
|
|
68 |
clear_btn="Clear",
|
69 |
)
|
70 |
demo.queue(concurrency_count=1, max_size=5)
|
71 |
+
|
72 |
+
demo.launch()
|
73 |
+
|
74 |
+
ingest.main()
|