Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -40,10 +40,10 @@ llm = Llama(model_path = 'Llama-2-ko-7B-chat-gguf-q4_0.bin',
|
|
40 |
)
|
41 |
# App code
|
42 |
def chat(x):
|
43 |
-
prom = f"๋ค์์ A์ B์ ์ญํ ๊ทน์ด์ผ. ๋๋ B์ผ. A์ ๋ํํ๊ณ ์์ด. ์น๊ตฌ์๊ฒ ์น๊ทผํ๊ณ ๊ฐ๊ฒฐํ๊ฒ ์ ๋๋ตํด์ค.\n\n### A:\n{x}\n\n### B:\n"
|
44 |
-
output = llm(prom, max_tokens=20, stop=["###"], echo=True)
|
45 |
-
return output['choices'][0]['text'][len(prom):-1]
|
46 |
-
|
47 |
|
48 |
with gr.Blocks() as demo:
|
49 |
count = 0
|
|
|
40 |
)
|
41 |
# App code
|
42 |
def chat(x):
|
43 |
+
#prom = f"๋ค์์ A์ B์ ์ญํ ๊ทน์ด์ผ. ๋๋ B์ผ. A์ ๋ํํ๊ณ ์์ด. ์น๊ตฌ์๊ฒ ์น๊ทผํ๊ณ ๊ฐ๊ฒฐํ๊ฒ ์ ๋๋ตํด์ค.\n\n### A:\n{x}\n\n### B:\n"
|
44 |
+
#output = llm(prom, max_tokens=20, stop=["###"], echo=True)
|
45 |
+
#return output['choices'][0]['text'][len(prom):-1]
|
46 |
+
return "AI ์๋ต์
๋๋ค."
|
47 |
|
48 |
with gr.Blocks() as demo:
|
49 |
count = 0
|