Spaces:
Restarting
on
CPU Upgrade
Restarting
on
CPU Upgrade
refactor: refactor the envs
Browse files- app.py +2 -4
- src/benchmarks.py +0 -3
- src/envs.py +2 -0
- src/loaders.py +1 -1
app.py
CHANGED
@@ -8,15 +8,13 @@ from src.about import (
|
|
8 |
from src.benchmarks import (
|
9 |
qa_benchmark_dict,
|
10 |
long_doc_benchmark_dict,
|
11 |
-
METRIC_LIST
|
12 |
-
DEFAULT_METRIC_QA,
|
13 |
-
DEFAULT_METRIC_LONG_DOC
|
14 |
)
|
15 |
from src.display.css_html_js import custom_css
|
16 |
from src.envs import (
|
17 |
API,
|
18 |
EVAL_RESULTS_PATH,
|
19 |
-
REPO_ID
|
20 |
)
|
21 |
from src.loaders import (
|
22 |
load_eval_results
|
|
|
8 |
from src.benchmarks import (
|
9 |
qa_benchmark_dict,
|
10 |
long_doc_benchmark_dict,
|
11 |
+
METRIC_LIST
|
|
|
|
|
12 |
)
|
13 |
from src.display.css_html_js import custom_css
|
14 |
from src.envs import (
|
15 |
API,
|
16 |
EVAL_RESULTS_PATH,
|
17 |
+
REPO_ID, DEFAULT_METRIC_QA, DEFAULT_METRIC_LONG_DOC
|
18 |
)
|
19 |
from src.loaders import (
|
20 |
load_eval_results
|
src/benchmarks.py
CHANGED
@@ -3,9 +3,6 @@ from enum import Enum
|
|
3 |
|
4 |
from air_benchmark.tasks.tasks import BenchmarkTable
|
5 |
|
6 |
-
DEFAULT_METRIC_QA = "ndcg_at_10"
|
7 |
-
DEFAULT_METRIC_LONG_DOC = "recall_at_10"
|
8 |
-
|
9 |
METRIC_LIST = [
|
10 |
"ndcg_at_1",
|
11 |
"ndcg_at_3",
|
|
|
3 |
|
4 |
from air_benchmark.tasks.tasks import BenchmarkTable
|
5 |
|
|
|
|
|
|
|
6 |
METRIC_LIST = [
|
7 |
"ndcg_at_1",
|
8 |
"ndcg_at_3",
|
src/envs.py
CHANGED
@@ -31,3 +31,5 @@ BENCHMARK_VERSION_LIST = [
|
|
31 |
]
|
32 |
|
33 |
LATEST_BENCHMARK_VERSION = BENCHMARK_VERSION_LIST[-1]
|
|
|
|
|
|
31 |
]
|
32 |
|
33 |
LATEST_BENCHMARK_VERSION = BENCHMARK_VERSION_LIST[-1]
|
34 |
+
DEFAULT_METRIC_QA = "ndcg_at_10"
|
35 |
+
DEFAULT_METRIC_LONG_DOC = "recall_at_10"
|
src/loaders.py
CHANGED
@@ -3,7 +3,7 @@ from typing import List
|
|
3 |
|
4 |
import pandas as pd
|
5 |
|
6 |
-
from src.
|
7 |
from src.display.columns import COL_NAME_REVISION, COL_NAME_TIMESTAMP, COL_NAME_IS_ANONYMOUS
|
8 |
|
9 |
from src.models import FullEvalResult, LeaderboardDataStore
|
|
|
3 |
|
4 |
import pandas as pd
|
5 |
|
6 |
+
from src.envs import DEFAULT_METRIC_QA, DEFAULT_METRIC_LONG_DOC
|
7 |
from src.display.columns import COL_NAME_REVISION, COL_NAME_TIMESTAMP, COL_NAME_IS_ANONYMOUS
|
8 |
|
9 |
from src.models import FullEvalResult, LeaderboardDataStore
|