mahynski commited on
Commit
227b789
·
1 Parent(s): 6a201c5

moved imports

Browse files
Files changed (1) hide show
  1. app.py +11 -13
app.py CHANGED
@@ -3,25 +3,13 @@ import os
3
  import tiktoken
4
  import streamlit as st
5
 
6
- from llama_index.llms.gemini import Gemini
7
- from llama_index.llms.huggingface import HuggingFaceLLM
8
- from llama_index.llms.huggingface_api import HuggingFaceInferenceAPI
9
- from llama_index.llms.mistralai import MistralAI
10
- from llama_index.llms.openai import OpenAI
11
-
12
- from llama_index.embeddings.openai import OpenAIEmbedding
13
- from llama_index.embeddings.huggingface import HuggingFaceEmbedding, HuggingFaceInferenceAPIEmbedding
14
- from llama_index.embeddings.mistralai import MistralAIEmbedding
15
-
16
  from llama_index.core import (
17
  VectorStoreIndex,
18
  Settings,
19
  )
20
 
21
  from llama_parse import LlamaParse
22
-
23
- from transformers import AutoTokenizer
24
-
25
  from streamlit_pdf_viewer import pdf_viewer
26
 
27
  MAX_OUTPUT_TOKENS = 2048
@@ -97,6 +85,10 @@ def main():
97
  raise NotImplementedError(f"{provider} is not supported yet")
98
  elif provider == 'huggingface':
99
  if llm_name is not None and embed_name is not None:
 
 
 
 
100
  os.environ['HFTOKEN'] = str(llm_key)
101
  Settings.llm = HuggingFaceInferenceAPI(
102
  model_name=llm_name,
@@ -114,6 +106,9 @@ def main():
114
  )
115
  Settings.context_window = 4096
116
  elif provider == 'mistralai':
 
 
 
117
  os.environ['MISTRAL_API_KEY'] = str(llm_key)
118
  Settings.llm = MistralAI(
119
  model=llm_name,
@@ -130,6 +125,9 @@ def main():
130
  )
131
  Settings.context_window = 4096 # max possible
132
  elif provider == 'openai':
 
 
 
133
  os.environ["OPENAI_API_KEY"] = str(llm_key)
134
  Settings.llm = OpenAI(
135
  model=llm_name,
 
3
  import tiktoken
4
  import streamlit as st
5
 
6
+ # from llama_index.llms.gemini import Gemini
 
 
 
 
 
 
 
 
 
7
  from llama_index.core import (
8
  VectorStoreIndex,
9
  Settings,
10
  )
11
 
12
  from llama_parse import LlamaParse
 
 
 
13
  from streamlit_pdf_viewer import pdf_viewer
14
 
15
  MAX_OUTPUT_TOKENS = 2048
 
85
  raise NotImplementedError(f"{provider} is not supported yet")
86
  elif provider == 'huggingface':
87
  if llm_name is not None and embed_name is not None:
88
+ from llama_index.llms.huggingface_api import HuggingFaceInferenceAPI
89
+ from llama_index.embeddings.huggingface import HuggingFaceInferenceAPIEmbedding
90
+ from transformers import AutoTokenizer
91
+
92
  os.environ['HFTOKEN'] = str(llm_key)
93
  Settings.llm = HuggingFaceInferenceAPI(
94
  model_name=llm_name,
 
106
  )
107
  Settings.context_window = 4096
108
  elif provider == 'mistralai':
109
+ from llama_index.llms.mistralai import MistralAI
110
+ from llama_index.embeddings.mistralai import MistralAIEmbedding
111
+
112
  os.environ['MISTRAL_API_KEY'] = str(llm_key)
113
  Settings.llm = MistralAI(
114
  model=llm_name,
 
125
  )
126
  Settings.context_window = 4096 # max possible
127
  elif provider == 'openai':
128
+ from llama_index.llms.openai import OpenAI
129
+ from llama_index.embeddings.openai import OpenAIEmbedding
130
+
131
  os.environ["OPENAI_API_KEY"] = str(llm_key)
132
  Settings.llm = OpenAI(
133
  model=llm_name,