clefourrier HF staff commited on
Commit
613696b
1 Parent(s): a66fcca

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +7 -7
app.py CHANGED
@@ -297,7 +297,7 @@ with demo:
297
  )
298
 
299
  with gr.Tabs(elem_classes="tab-buttons") as tabs:
300
- with gr.TabItem("🏅 LLM Benchmark (lite)", elem_id="llm-benchmark-tab-table", id=0):
301
  leaderboard_table_lite = gr.components.Dataframe(
302
  value=leaderboard_df[COLS_LITE],
303
  headers=COLS_LITE,
@@ -319,7 +319,7 @@ with demo:
319
  leaderboard_table_lite,
320
  )
321
 
322
- with gr.TabItem("📊 Extended view", elem_id="llm-benchmark-tab-table", id=1):
323
  leaderboard_table = gr.components.Dataframe(
324
  value=leaderboard_df,
325
  headers=COLS,
@@ -341,16 +341,16 @@ with demo:
341
  [hidden_leaderboard_table_for_search, search_bar],
342
  leaderboard_table,
343
  )
344
- with gr.TabItem("About", elem_id="llm-benchmark-tab-table", id=2):
345
  gr.Markdown(LLM_BENCHMARKS_TEXT, elem_classes="markdown-text")
346
 
347
- with gr.TabItem("✉️✨ Submit here! ", elem_id="llm-benchmark-tab-table", id=3):
348
  with gr.Column():
349
  with gr.Row():
350
  gr.Markdown(EVALUATION_QUEUE_TEXT, elem_classes="markdown-text")
351
 
352
  with gr.Column():
353
- with gr.Accordion(f"✅ Finished Evaluations: {len(finished_eval_queue_df)}", open=False):
354
  with gr.Row():
355
  finished_eval_table = gr.components.Dataframe(
356
  value=finished_eval_queue_df,
@@ -358,7 +358,7 @@ with demo:
358
  datatype=EVAL_TYPES,
359
  max_rows=5,
360
  )
361
- with gr.Accordion(f"🔄 Running Evaluation Queue: {len(running_eval_queue_df)}", open=False):
362
  with gr.Row():
363
  running_eval_table = gr.components.Dataframe(
364
  value=running_eval_queue_df,
@@ -367,7 +367,7 @@ with demo:
367
  max_rows=5,
368
  )
369
 
370
- with gr.Accordion(f"⏳ Pending Evaluation Queue: {len(pending_eval_queue_df)}", open=False):
371
  with gr.Row():
372
  pending_eval_table = gr.components.Dataframe(
373
  value=pending_eval_queue_df,
 
297
  )
298
 
299
  with gr.Tabs(elem_classes="tab-buttons") as tabs:
300
+ with gr.TabItem("🏅 LLM Benchmark", elem_id="llm-benchmark-tab-table", id=0):
301
  leaderboard_table_lite = gr.components.Dataframe(
302
  value=leaderboard_df[COLS_LITE],
303
  headers=COLS_LITE,
 
319
  leaderboard_table_lite,
320
  )
321
 
322
+ with gr.TabItem("🔍 Extended model view", elem_id="llm-benchmark-tab-table", id=1):
323
  leaderboard_table = gr.components.Dataframe(
324
  value=leaderboard_df,
325
  headers=COLS,
 
341
  [hidden_leaderboard_table_for_search, search_bar],
342
  leaderboard_table,
343
  )
344
+ with gr.TabItem("📝 About", elem_id="llm-benchmark-tab-table", id=2):
345
  gr.Markdown(LLM_BENCHMARKS_TEXT, elem_classes="markdown-text")
346
 
347
+ with gr.TabItem("🚀 Submit here! ", elem_id="llm-benchmark-tab-table", id=3):
348
  with gr.Column():
349
  with gr.Row():
350
  gr.Markdown(EVALUATION_QUEUE_TEXT, elem_classes="markdown-text")
351
 
352
  with gr.Column():
353
+ with gr.Accordion(f"✅ Finished Evaluations ({len(finished_eval_queue_df)})", open=False):
354
  with gr.Row():
355
  finished_eval_table = gr.components.Dataframe(
356
  value=finished_eval_queue_df,
 
358
  datatype=EVAL_TYPES,
359
  max_rows=5,
360
  )
361
+ with gr.Accordion(f"🔄 Running Evaluation Queue ({len(running_eval_queue_df)})", open=False):
362
  with gr.Row():
363
  running_eval_table = gr.components.Dataframe(
364
  value=running_eval_queue_df,
 
367
  max_rows=5,
368
  )
369
 
370
+ with gr.Accordion(f"⏳ Pending Evaluation Queue ({len(pending_eval_queue_df)})", open=False):
371
  with gr.Row():
372
  pending_eval_table = gr.components.Dataframe(
373
  value=pending_eval_queue_df,