Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -335,6 +335,7 @@ def search_glossary(query):
|
|
335 |
api_name="/ask_llm"
|
336 |
)
|
337 |
st.markdown(result)
|
|
|
338 |
st.code(result, language="python", line_numbers=True)
|
339 |
# ๐ ArXiv RAG researcher expert ~-<>-~ Paper Summary & Ask LLM - api_name: /ask_llm
|
340 |
result2 = client.predict(
|
@@ -344,6 +345,7 @@ def search_glossary(query):
|
|
344 |
api_name="/ask_llm"
|
345 |
)
|
346 |
st.markdown(result2)
|
|
|
347 |
st.code(result2, language="python", line_numbers=True)
|
348 |
# ๐ ArXiv RAG researcher expert ~-<>-~ Paper Summary & Ask LLM - api_name: /ask_llm
|
349 |
result3 = client.predict(
|
@@ -353,6 +355,7 @@ def search_glossary(query):
|
|
353 |
api_name="/ask_llm"
|
354 |
)
|
355 |
st.markdown(result3)
|
|
|
356 |
st.code(result3, language="python", line_numbers=True)
|
357 |
# ๐ ArXiv RAG researcher expert ~-<>-~ Paper Summary & Ask LLM - api_name: /update_with_rag_md
|
358 |
response2 = client.predict(
|
@@ -363,9 +366,31 @@ def search_glossary(query):
|
|
363 |
api_name="/update_with_rag_md"
|
364 |
)
|
365 |
st.markdown(response2[0])
|
366 |
-
st.code(response2[0], language="python", line_numbers=True, wrap_lines=True)
|
|
|
367 |
st.markdown(response2[1])
|
368 |
st.code(response2[1], language="python", line_numbers=True, wrap_lines=True)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
369 |
return result, result2, result3, response2
|
370 |
|
371 |
|
|
|
335 |
api_name="/ask_llm"
|
336 |
)
|
337 |
st.markdown(result)
|
338 |
+
|
339 |
st.code(result, language="python", line_numbers=True)
|
340 |
# ๐ ArXiv RAG researcher expert ~-<>-~ Paper Summary & Ask LLM - api_name: /ask_llm
|
341 |
result2 = client.predict(
|
|
|
345 |
api_name="/ask_llm"
|
346 |
)
|
347 |
st.markdown(result2)
|
348 |
+
|
349 |
st.code(result2, language="python", line_numbers=True)
|
350 |
# ๐ ArXiv RAG researcher expert ~-<>-~ Paper Summary & Ask LLM - api_name: /ask_llm
|
351 |
result3 = client.predict(
|
|
|
355 |
api_name="/ask_llm"
|
356 |
)
|
357 |
st.markdown(result3)
|
358 |
+
|
359 |
st.code(result3, language="python", line_numbers=True)
|
360 |
# ๐ ArXiv RAG researcher expert ~-<>-~ Paper Summary & Ask LLM - api_name: /update_with_rag_md
|
361 |
response2 = client.predict(
|
|
|
366 |
api_name="/update_with_rag_md"
|
367 |
)
|
368 |
st.markdown(response2[0])
|
369 |
+
st.code(response2[0], language="python", line_numbers=True, wrap_lines=True)
|
370 |
+
|
371 |
st.markdown(response2[1])
|
372 |
st.code(response2[1], language="python", line_numbers=True, wrap_lines=True)
|
373 |
+
|
374 |
+
|
375 |
+
# Persist AI Results to Markdown Files
|
376 |
+
filename = generate_filename(result, "md")
|
377 |
+
create_file(filename, query, result)
|
378 |
+
st.session_state.chat_history.append({"assistant": query, "ArXiV": result})
|
379 |
+
|
380 |
+
filename = generate_filename(result2, "md")
|
381 |
+
create_file(filename, query, result2)
|
382 |
+
st.session_state.chat_history.append({"assistant": query, "ArXiV": result2})
|
383 |
+
|
384 |
+
filename = generate_filename(result3, "md")
|
385 |
+
create_file(filename, query, result3)
|
386 |
+
st.session_state.chat_history.append({"assistant": query, "ArXiV": result3})
|
387 |
+
|
388 |
+
filename = generate_filename(response2, "md")
|
389 |
+
create_file(filename, query, response2)
|
390 |
+
st.session_state.chat_history.append({"assistant": query, "ArXiV": response2})
|
391 |
+
|
392 |
+
|
393 |
+
|
394 |
return result, result2, result3, response2
|
395 |
|
396 |
|