hi-melnikov commited on
Commit
ca451af
1 Parent(s): 1f38597

added mkdirs and changed some env

Browse files
src/envs.py CHANGED
@@ -27,21 +27,21 @@ if not os.access(HF_HOME, os.W_OK):
27
  else:
28
  print("Write access confirmed for HF_HOME")
29
 
30
- EVAL_REQUESTS_PATH = os.path.join(HF_HOME, "eval-queue")
31
- EVAL_RESULTS_PATH = os.path.join(HF_HOME, "eval-results")
 
 
 
 
 
 
32
  DYNAMIC_INFO_PATH = os.path.join(HF_HOME, "dynamic-info")
33
  DYNAMIC_INFO_FILE_PATH = os.path.join(DYNAMIC_INFO_PATH, "model_infos.json")
34
-
35
  EVAL_REQUESTS_PATH_PRIVATE = "eval-queue-private"
36
  EVAL_RESULTS_PATH_PRIVATE = "eval-results-private"
37
-
38
  PATH_TO_COLLECTION = "open-llm-leaderboard/llm-leaderboard-best-models-652d6c7965a4619fb5c27a03"
39
 
40
  # Rate limit variables
41
  RATE_LIMIT_PERIOD = 7
42
  RATE_LIMIT_QUOTA = 5
43
  HAS_HIGHER_RATE_LIMIT = ["TheBloke"]
44
-
45
- RESET_JUDGEMENT_ENV = "RESET_JUDGEMENT"
46
-
47
- API = HfApi(token=H4_TOKEN)
 
27
  else:
28
  print("Write access confirmed for HF_HOME")
29
 
30
+ EVAL_RESULTS_PATH = os.path.join(HF_HOME, "openbench")
31
+
32
+ RESET_JUDGEMENT_ENV = "RESET_JUDGEMENT"
33
+
34
+ API = HfApi(token=H4_TOKEN)
35
+
36
+ # useless env
37
+ EVAL_REQUESTS_PATH = os.path.join(HF_HOME, "data/eval-queue")
38
  DYNAMIC_INFO_PATH = os.path.join(HF_HOME, "dynamic-info")
39
  DYNAMIC_INFO_FILE_PATH = os.path.join(DYNAMIC_INFO_PATH, "model_infos.json")
 
40
  EVAL_REQUESTS_PATH_PRIVATE = "eval-queue-private"
41
  EVAL_RESULTS_PATH_PRIVATE = "eval-results-private"
 
42
  PATH_TO_COLLECTION = "open-llm-leaderboard/llm-leaderboard-best-models-652d6c7965a4619fb5c27a03"
43
 
44
  # Rate limit variables
45
  RATE_LIMIT_PERIOD = 7
46
  RATE_LIMIT_QUOTA = 5
47
  HAS_HIGHER_RATE_LIMIT = ["TheBloke"]
 
 
 
 
src/leaderboard/build_leaderboard.py CHANGED
@@ -53,6 +53,11 @@ def download_dataset(repo_id, local_dir, repo_type="dataset", max_attempts=3, ba
53
 
54
  def download_openbench():
55
  """Downloads pre generated data"""
 
 
 
 
 
56
  # download answers of different models that we trust
57
  download_dataset("Vikhrmodels/openbench-eval", EVAL_RESULTS_PATH)
58
 
 
53
 
54
  def download_openbench():
55
  """Downloads pre generated data"""
56
+ os.makedirs("data/arena-hard-v0.1/model_answer/internal/", exist_ok=True)
57
+ os.makedirs("data/arena-hard-v0.1/model_answer/external/", exist_ok=True)
58
+ os.makedirs("data/arena-hard-v0.1/model_answer/model_judgement/", exist_ok=True)
59
+ os.makedirs(EVAL_RESULTS_PATH, exist_ok=True)
60
+
61
  # download answers of different models that we trust
62
  download_dataset("Vikhrmodels/openbench-eval", EVAL_RESULTS_PATH)
63