Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
Clémentine
commited on
Commit
•
5491f2d
1
Parent(s):
ed118a6
updated GPTQ display!
Browse files- app.py +9 -9
- model_info_cache.pkl +2 -2
- model_size_cache.pkl +2 -2
- src/display_models/get_model_metadata.py +4 -2
- src/display_models/read_results.py +2 -0
app.py
CHANGED
@@ -253,14 +253,14 @@ def select_columns(df: pd.DataFrame, columns: list) -> pd.DataFrame:
|
|
253 |
|
254 |
|
255 |
NUMERIC_INTERVALS = {
|
256 |
-
"
|
257 |
-
"0~1.
|
258 |
-
"1.5~
|
259 |
-
"3~
|
260 |
-
"7~
|
261 |
-
"13~
|
262 |
-
"35~
|
263 |
-
"
|
264 |
}
|
265 |
|
266 |
|
@@ -378,7 +378,7 @@ with demo:
|
|
378 |
elem_id="filter-columns-precision",
|
379 |
)
|
380 |
filter_columns_size = gr.CheckboxGroup(
|
381 |
-
label="Model sizes",
|
382 |
choices=list(NUMERIC_INTERVALS.keys()),
|
383 |
value=list(NUMERIC_INTERVALS.keys()),
|
384 |
interactive=True,
|
|
|
253 |
|
254 |
|
255 |
NUMERIC_INTERVALS = {
|
256 |
+
"?": pd.Interval(-1, 0, closed="right"),
|
257 |
+
"0~1.5": pd.Interval(0, 1.5, closed="right"),
|
258 |
+
"1.5~3": pd.Interval(1.5, 3, closed="right"),
|
259 |
+
"3~7": pd.Interval(3, 7, closed="right"),
|
260 |
+
"7~13": pd.Interval(7, 13, closed="right"),
|
261 |
+
"13~35": pd.Interval(13, 35, closed="right"),
|
262 |
+
"35~60": pd.Interval(35, 60, closed="right"),
|
263 |
+
"60+": pd.Interval(60, 10000, closed="right"),
|
264 |
}
|
265 |
|
266 |
|
|
|
378 |
elem_id="filter-columns-precision",
|
379 |
)
|
380 |
filter_columns_size = gr.CheckboxGroup(
|
381 |
+
label="Model sizes (in billions of parameters)",
|
382 |
choices=list(NUMERIC_INTERVALS.keys()),
|
383 |
value=list(NUMERIC_INTERVALS.keys()),
|
384 |
interactive=True,
|
model_info_cache.pkl
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f98e18ea0fcf11737e108f966f6d8b09120c6a2c231b70e9e242e9bba3145a47
|
3 |
+
size 3780284
|
model_size_cache.pkl
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ace7167a258f711fa7ffeaadddc6ebef8ccb92651dce8b805228c2f18c988958
|
3 |
+
size 75324
|
src/display_models/get_model_metadata.py
CHANGED
@@ -45,13 +45,15 @@ def get_model_infos_from_hub(leaderboard_data: List[dict]):
|
|
45 |
model_data[AutoEvalColumn.license.name] = None
|
46 |
model_data[AutoEvalColumn.likes.name] = None
|
47 |
if model_name not in model_size_cache:
|
48 |
-
|
|
|
49 |
model_data[AutoEvalColumn.params.name] = model_size_cache[model_name]
|
50 |
|
51 |
model_data[AutoEvalColumn.license.name] = get_model_license(model_info)
|
52 |
model_data[AutoEvalColumn.likes.name] = get_model_likes(model_info)
|
53 |
if model_name not in model_size_cache:
|
54 |
-
|
|
|
55 |
model_data[AutoEvalColumn.params.name] = model_size_cache[model_name]
|
56 |
|
57 |
# save cache to disk in pickle format
|
|
|
45 |
model_data[AutoEvalColumn.license.name] = None
|
46 |
model_data[AutoEvalColumn.likes.name] = None
|
47 |
if model_name not in model_size_cache:
|
48 |
+
size_factor = 8 if model_data["Precision"] == "GPTQ" else 1
|
49 |
+
model_size_cache[model_name] = size_factor * get_model_size(model_name, None)
|
50 |
model_data[AutoEvalColumn.params.name] = model_size_cache[model_name]
|
51 |
|
52 |
model_data[AutoEvalColumn.license.name] = get_model_license(model_info)
|
53 |
model_data[AutoEvalColumn.likes.name] = get_model_likes(model_info)
|
54 |
if model_name not in model_size_cache:
|
55 |
+
size_factor = 8 if model_data["Precision"] == "GPTQ" else 1
|
56 |
+
model_size_cache[model_name] = size_factor * get_model_size(model_name, model_info)
|
57 |
model_data[AutoEvalColumn.params.name] = model_size_cache[model_name]
|
58 |
|
59 |
# save cache to disk in pickle format
|
src/display_models/read_results.py
CHANGED
@@ -81,6 +81,8 @@ def parse_eval_result(json_filepath: str) -> Tuple[str, list[dict]]:
|
|
81 |
model_split = model.split("/", 1)
|
82 |
|
83 |
precision = config.get("model_dtype")
|
|
|
|
|
84 |
|
85 |
model = model_split[-1]
|
86 |
|
|
|
81 |
model_split = model.split("/", 1)
|
82 |
|
83 |
precision = config.get("model_dtype")
|
84 |
+
if precision == "None":
|
85 |
+
precision = "GPTQ"
|
86 |
|
87 |
model = model_split[-1]
|
88 |
|