barunsaha commited on
Commit
856ad6d
·
1 Parent(s): 482e077

Add gemini-2.0-flash-exp LLM

Browse files
Files changed (2) hide show
  1. app.py +1 -1
  2. global_config.py +7 -2
app.py CHANGED
@@ -159,7 +159,7 @@ with st.sidebar:
159
  else:
160
  # The LLMs
161
  llm_provider_to_use = st.sidebar.selectbox(
162
- label='2: Select an LLM to use:',
163
  options=[f'{k} ({v["description"]})' for k, v in GlobalConfig.VALID_MODELS.items()],
164
  index=GlobalConfig.DEFAULT_MODEL_INDEX,
165
  help=GlobalConfig.LLM_PROVIDER_HELP,
 
159
  else:
160
  # The LLMs
161
  llm_provider_to_use = st.sidebar.selectbox(
162
+ label='2: Select a suitable LLM to use:\n\n(Gemini and Mistral-Nemo are recommended)',
163
  options=[f'{k} ({v["description"]})' for k, v in GlobalConfig.VALID_MODELS.items()],
164
  index=GlobalConfig.DEFAULT_MODEL_INDEX,
165
  help=GlobalConfig.LLM_PROVIDER_HELP,
global_config.py CHANGED
@@ -34,7 +34,12 @@ class GlobalConfig:
34
  'paid': True,
35
  },
36
  '[gg]gemini-1.5-flash-002': {
37
- 'description': 'faster response',
 
 
 
 
 
38
  'max_new_tokens': 8192,
39
  'paid': True,
40
  },
@@ -55,7 +60,7 @@ class GlobalConfig:
55
  '- **[gg]**: Google Gemini API\n'
56
  '- **[hf]**: Hugging Face Inference API\n'
57
  )
58
- DEFAULT_MODEL_INDEX = 2
59
  LLM_MODEL_TEMPERATURE = 0.2
60
  LLM_MODEL_MIN_OUTPUT_LENGTH = 100
61
  LLM_MODEL_MAX_INPUT_LENGTH = 400 # characters
 
34
  'paid': True,
35
  },
36
  '[gg]gemini-1.5-flash-002': {
37
+ 'description': 'faster, detailed',
38
+ 'max_new_tokens': 8192,
39
+ 'paid': True,
40
+ },
41
+ '[gg]gemini-2.0-flash-exp': {
42
+ 'description': 'fast, detailed',
43
  'max_new_tokens': 8192,
44
  'paid': True,
45
  },
 
60
  '- **[gg]**: Google Gemini API\n'
61
  '- **[hf]**: Hugging Face Inference API\n'
62
  )
63
+ DEFAULT_MODEL_INDEX = 3
64
  LLM_MODEL_TEMPERATURE = 0.2
65
  LLM_MODEL_MIN_OUTPUT_LENGTH = 100
66
  LLM_MODEL_MAX_INPUT_LENGTH = 400 # characters