hi-melnikov commited on
Commit
1bd2e0f
1 Parent(s): cd5319b

remove unused envs

Browse files
src/envs.py CHANGED
@@ -35,10 +35,6 @@ API = HfApi(token=H4_TOKEN)
35
 
36
  # useless env
37
  EVAL_REQUESTS_PATH = os.path.join(HF_HOME, "data/eval-queue")
38
- DYNAMIC_INFO_PATH = os.path.join(HF_HOME, "dynamic-info")
39
- DYNAMIC_INFO_FILE_PATH = os.path.join(DYNAMIC_INFO_PATH, "model_infos.json")
40
- EVAL_REQUESTS_PATH_PRIVATE = "eval-queue-private"
41
- EVAL_RESULTS_PATH_PRIVATE = "eval-results-private"
42
  PATH_TO_COLLECTION = "open-llm-leaderboard/llm-leaderboard-best-models-652d6c7965a4619fb5c27a03"
43
 
44
  # Rate limit variables
 
35
 
36
  # useless env
37
  EVAL_REQUESTS_PATH = os.path.join(HF_HOME, "data/eval-queue")
 
 
 
 
38
  PATH_TO_COLLECTION = "open-llm-leaderboard/llm-leaderboard-best-models-652d6c7965a4619fb5c27a03"
39
 
40
  # Rate limit variables
src/leaderboard/build_leaderboard.py CHANGED
@@ -7,7 +7,7 @@ import time
7
  import pandas as pd
8
  from huggingface_hub import snapshot_download
9
 
10
- from src.envs import DATA_ARENA_PATH, EVAL_RESULTS_PATH, HF_HOME
11
 
12
  # Configure logging
13
  logging.basicConfig(level=logging.INFO, format="%(asctime)s - %(levelname)s - %(message)s")
@@ -67,17 +67,6 @@ def download_openbench():
67
  print("\nJudgement in openbench-eval")
68
  subprocess.run(["ls", f"{DATA_ARENA_PATH}/model_judgment/gpt-4-1106-preview"], check=False)
69
 
70
- # copy the model answers / judgements to data folder
71
- subprocess.run(
72
- [
73
- "rsync",
74
- "-azPvh",
75
- f"{EVAL_RESULTS_PATH}/*",
76
- f"{DATA_ARENA_PATH}/",
77
- ],
78
- check=True,
79
- )
80
-
81
 
82
  def build_leadearboard_df():
83
  # Retrieve the leaderboard DataFrame
 
7
  import pandas as pd
8
  from huggingface_hub import snapshot_download
9
 
10
+ from src.envs import DATA_ARENA_PATH, HF_HOME
11
 
12
  # Configure logging
13
  logging.basicConfig(level=logging.INFO, format="%(asctime)s - %(levelname)s - %(message)s")
 
67
  print("\nJudgement in openbench-eval")
68
  subprocess.run(["ls", f"{DATA_ARENA_PATH}/model_judgment/gpt-4-1106-preview"], check=False)
69
 
 
 
 
 
 
 
 
 
 
 
 
70
 
71
  def build_leadearboard_df():
72
  # Retrieve the leaderboard DataFrame