Spaces:
AIR-Bench
/
Restarting on CPU Upgrade

nan commited on
Commit
ba13e25
1 Parent(s): a7c0332

refactor: refactor the envs

Browse files
Files changed (4) hide show
  1. app.py +2 -4
  2. src/benchmarks.py +0 -3
  3. src/envs.py +2 -0
  4. src/loaders.py +1 -1
app.py CHANGED
@@ -8,15 +8,13 @@ from src.about import (
8
  from src.benchmarks import (
9
  qa_benchmark_dict,
10
  long_doc_benchmark_dict,
11
- METRIC_LIST,
12
- DEFAULT_METRIC_QA,
13
- DEFAULT_METRIC_LONG_DOC
14
  )
15
  from src.display.css_html_js import custom_css
16
  from src.envs import (
17
  API,
18
  EVAL_RESULTS_PATH,
19
- REPO_ID
20
  )
21
  from src.loaders import (
22
  load_eval_results
 
8
  from src.benchmarks import (
9
  qa_benchmark_dict,
10
  long_doc_benchmark_dict,
11
+ METRIC_LIST
 
 
12
  )
13
  from src.display.css_html_js import custom_css
14
  from src.envs import (
15
  API,
16
  EVAL_RESULTS_PATH,
17
+ REPO_ID, DEFAULT_METRIC_QA, DEFAULT_METRIC_LONG_DOC
18
  )
19
  from src.loaders import (
20
  load_eval_results
src/benchmarks.py CHANGED
@@ -3,9 +3,6 @@ from enum import Enum
3
 
4
  from air_benchmark.tasks.tasks import BenchmarkTable
5
 
6
- DEFAULT_METRIC_QA = "ndcg_at_10"
7
- DEFAULT_METRIC_LONG_DOC = "recall_at_10"
8
-
9
  METRIC_LIST = [
10
  "ndcg_at_1",
11
  "ndcg_at_3",
 
3
 
4
  from air_benchmark.tasks.tasks import BenchmarkTable
5
 
 
 
 
6
  METRIC_LIST = [
7
  "ndcg_at_1",
8
  "ndcg_at_3",
src/envs.py CHANGED
@@ -31,3 +31,5 @@ BENCHMARK_VERSION_LIST = [
31
  ]
32
 
33
  LATEST_BENCHMARK_VERSION = BENCHMARK_VERSION_LIST[-1]
 
 
 
31
  ]
32
 
33
  LATEST_BENCHMARK_VERSION = BENCHMARK_VERSION_LIST[-1]
34
+ DEFAULT_METRIC_QA = "ndcg_at_10"
35
+ DEFAULT_METRIC_LONG_DOC = "recall_at_10"
src/loaders.py CHANGED
@@ -3,7 +3,7 @@ from typing import List
3
 
4
  import pandas as pd
5
 
6
- from src.benchmarks import DEFAULT_METRIC_QA, DEFAULT_METRIC_LONG_DOC
7
  from src.display.columns import COL_NAME_REVISION, COL_NAME_TIMESTAMP, COL_NAME_IS_ANONYMOUS
8
 
9
  from src.models import FullEvalResult, LeaderboardDataStore
 
3
 
4
  import pandas as pd
5
 
6
+ from src.envs import DEFAULT_METRIC_QA, DEFAULT_METRIC_LONG_DOC
7
  from src.display.columns import COL_NAME_REVISION, COL_NAME_TIMESTAMP, COL_NAME_IS_ANONYMOUS
8
 
9
  from src.models import FullEvalResult, LeaderboardDataStore