t0-0 commited on
Commit
b246f21
·
1 Parent(s): bbe855d

Changed the timing of model architecture retrieval to when submitting.

Browse files
src/leaderboard/read_evals.py CHANGED
@@ -8,7 +8,6 @@ import dateutil
8
 
9
  from src.display.formatting import make_clickable_model
10
  from src.display.utils import AutoEvalColumn, Backend, ModelType, Tasks, Version, WeightType
11
- from src.submission.check_validity import is_model_on_hub
12
 
13
 
14
  @dataclass
@@ -90,15 +89,6 @@ class EvalResult:
90
  result_key = f"{model}_{precision}_({num_few_shots}shots)_{add_special_tokens}"
91
  full_model = "/".join(org_and_model)
92
 
93
- still_on_hub, _, model_config = is_model_on_hub(
94
- full_model, config.get("model_sha", "main"), trust_remote_code=True, test_tokenizer=False
95
- )
96
- architecture = "?"
97
- if model_config is not None:
98
- architectures = getattr(model_config, "architectures", None)
99
- if architectures:
100
- architecture = ";".join(architectures)
101
-
102
  if "scores" not in data:
103
  raise KeyError(f"'scores' key not found in JSON file: {json_filepath}")
104
 
@@ -117,7 +107,6 @@ class EvalResult:
117
  results=results,
118
  precision=precision,
119
  revision=revision,
120
- architecture=architecture,
121
  num_few_shots=num_few_shots,
122
  add_special_tokens=add_special_tokens,
123
  llm_jp_eval_version=version,
@@ -136,6 +125,7 @@ class EvalResult:
136
  self.likes = request.get("likes", 0)
137
  self.num_params = request.get("params", 0)
138
  self.date = request.get("submitted_time", "")
 
139
  except Exception:
140
  print(f"Could not find request file for {self.org}/{self.model} with precision {self.precision}")
141
 
 
8
 
9
  from src.display.formatting import make_clickable_model
10
  from src.display.utils import AutoEvalColumn, Backend, ModelType, Tasks, Version, WeightType
 
11
 
12
 
13
  @dataclass
 
89
  result_key = f"{model}_{precision}_({num_few_shots}shots)_{add_special_tokens}"
90
  full_model = "/".join(org_and_model)
91
 
 
 
 
 
 
 
 
 
 
92
  if "scores" not in data:
93
  raise KeyError(f"'scores' key not found in JSON file: {json_filepath}")
94
 
 
107
  results=results,
108
  precision=precision,
109
  revision=revision,
 
110
  num_few_shots=num_few_shots,
111
  add_special_tokens=add_special_tokens,
112
  llm_jp_eval_version=version,
 
125
  self.likes = request.get("likes", 0)
126
  self.num_params = request.get("params", 0)
127
  self.date = request.get("submitted_time", "")
128
+ self.architecture = request.get("architecture", "?")
129
  except Exception:
130
  print(f"Could not find request file for {self.org}/{self.model} with precision {self.precision}")
131
 
src/submission/submit.py CHANGED
@@ -47,7 +47,14 @@ def add_new_eval(
47
  revision = "main"
48
 
49
  # Is the model on the hub?
50
- model_on_hub, error, _ = is_model_on_hub(model_name=model, revision=revision, token=TOKEN, test_tokenizer=True)
 
 
 
 
 
 
 
51
  if not model_on_hub:
52
  return styled_error(f'Model "{model}" {error}')
53
 
@@ -85,6 +92,7 @@ def add_new_eval(
85
  "private": False,
86
  "add_special_tokens": add_special_tokens,
87
  "llm_jp_eval_version": current_version,
 
88
  }
89
 
90
  # Check for duplicate submission
 
47
  revision = "main"
48
 
49
  # Is the model on the hub?
50
+ model_on_hub, error, model_config = is_model_on_hub(
51
+ model_name=model, revision=revision, token=TOKEN, test_tokenizer=True
52
+ )
53
+ architecture = "?"
54
+ if model_config is not None:
55
+ architectures = getattr(model_config, "architectures", None)
56
+ if architectures:
57
+ architecture = ";".join(architectures)
58
  if not model_on_hub:
59
  return styled_error(f'Model "{model}" {error}')
60
 
 
92
  "private": False,
93
  "add_special_tokens": add_special_tokens,
94
  "llm_jp_eval_version": current_version,
95
+ "architecture": architecture,
96
  }
97
 
98
  # Check for duplicate submission