test
Browse files- app.py +25 -1
- requirements.txt +3 -1
app.py
CHANGED
@@ -5,8 +5,16 @@ from llama_index.llms.huggingface import HuggingFaceLLM
|
|
5 |
from llama_index.llms.mistralai import MistralAI
|
6 |
from llama_index.llms.openai import OpenAI
|
7 |
|
|
|
|
|
|
|
|
|
|
|
8 |
from streamlit_pdf_viewer import pdf_viewer
|
9 |
|
|
|
|
|
|
|
10 |
st.set_page_config(layout="wide")
|
11 |
|
12 |
with st.sidebar:
|
@@ -41,7 +49,7 @@ with st.sidebar:
|
|
41 |
else:
|
42 |
llm_list = []
|
43 |
|
44 |
-
|
45 |
label="Select LLM Model",
|
46 |
options=llm_list,
|
47 |
index=0
|
@@ -56,6 +64,22 @@ with st.sidebar:
|
|
56 |
step=0.05,
|
57 |
)
|
58 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
59 |
# Enter Token
|
60 |
token = st.text_input(
|
61 |
"Enter your token",
|
|
|
5 |
from llama_index.llms.mistralai import MistralAI
|
6 |
from llama_index.llms.openai import OpenAI
|
7 |
|
8 |
+
from llama_index.core import (
|
9 |
+
VectorStoreIndex,
|
10 |
+
Settings,
|
11 |
+
)
|
12 |
+
|
13 |
from streamlit_pdf_viewer import pdf_viewer
|
14 |
|
15 |
+
# Global configurations
|
16 |
+
from llama_index.core import set_global_handler
|
17 |
+
set_global_handler("langfuse")
|
18 |
st.set_page_config(layout="wide")
|
19 |
|
20 |
with st.sidebar:
|
|
|
49 |
else:
|
50 |
llm_list = []
|
51 |
|
52 |
+
llm_name = st.selectbox(
|
53 |
label="Select LLM Model",
|
54 |
options=llm_list,
|
55 |
index=0
|
|
|
64 |
step=0.05,
|
65 |
)
|
66 |
|
67 |
+
max_output_tokens = 4096
|
68 |
+
|
69 |
+
# Create LLM
|
70 |
+
if provider == 'openai':
|
71 |
+
llm = OpenAI(
|
72 |
+
model=llm_name,
|
73 |
+
temperature=temperature,
|
74 |
+
max_tokens=max_tokens
|
75 |
+
)
|
76 |
+
# Global tokenization needs to be consistent with LLM
|
77 |
+
# https://docs.llamaindex.ai/en/stable/module_guides/models/llms/
|
78 |
+
Settings.tokenizer = tiktoken.encoding_for_model(llm_name).encode
|
79 |
+
Settings.num_output = max_tokens
|
80 |
+
Settings.context_window = 4096 # max possible
|
81 |
+
|
82 |
+
|
83 |
# Enter Token
|
84 |
token = st.text_input(
|
85 |
"Enter your token",
|
requirements.txt
CHANGED
@@ -5,4 +5,6 @@ llama-index-llms-gemini
|
|
5 |
llama-index-llms-huggingface
|
6 |
llama-index-llms-huggingface-api
|
7 |
llama-index-llms-mistralai
|
8 |
-
llama-index-llms-openai
|
|
|
|
|
|
5 |
llama-index-llms-huggingface
|
6 |
llama-index-llms-huggingface-api
|
7 |
llama-index-llms-mistralai
|
8 |
+
llama-index-llms-openai
|
9 |
+
tiktoken
|
10 |
+
llama-parse
|