NicoNico6
commited on
Commit
•
09f6986
1
Parent(s):
9a4cbfe
update
Browse files- quant_strategy.json +57 -57
quant_strategy.json
CHANGED
@@ -1,7 +1,7 @@
|
|
1 |
{
|
2 |
"measurement": {
|
3 |
"model.layers.0": {
|
4 |
-
"accuracy": 0.
|
5 |
"total_bits": 319680672,
|
6 |
"q_proj": {
|
7 |
"group_size": {
|
@@ -89,14 +89,14 @@
|
|
89 |
}
|
90 |
},
|
91 |
"model.layers.1": {
|
92 |
-
"accuracy": 0.
|
93 |
-
"total_bits":
|
94 |
"q_proj": {
|
95 |
"group_size": {
|
96 |
-
"
|
97 |
},
|
98 |
"bits": [
|
99 |
-
|
100 |
],
|
101 |
"bits_prop": [
|
102 |
1
|
@@ -105,10 +105,10 @@
|
|
105 |
},
|
106 |
"k_proj": {
|
107 |
"group_size": {
|
108 |
-
"
|
109 |
},
|
110 |
"bits": [
|
111 |
-
|
112 |
],
|
113 |
"bits_prop": [
|
114 |
1
|
@@ -177,7 +177,7 @@
|
|
177 |
}
|
178 |
},
|
179 |
"model.layers.2": {
|
180 |
-
"accuracy": 0.
|
181 |
"total_bits": 319680672,
|
182 |
"q_proj": {
|
183 |
"group_size": {
|
@@ -265,7 +265,7 @@
|
|
265 |
}
|
266 |
},
|
267 |
"model.layers.3": {
|
268 |
-
"accuracy": 0.
|
269 |
"total_bits": 319680672,
|
270 |
"q_proj": {
|
271 |
"group_size": {
|
@@ -353,7 +353,7 @@
|
|
353 |
}
|
354 |
},
|
355 |
"model.layers.4": {
|
356 |
-
"accuracy": 0.
|
357 |
"total_bits": 319680672,
|
358 |
"q_proj": {
|
359 |
"group_size": {
|
@@ -441,7 +441,7 @@
|
|
441 |
}
|
442 |
},
|
443 |
"model.layers.5": {
|
444 |
-
"accuracy": 0.
|
445 |
"total_bits": 319680672,
|
446 |
"q_proj": {
|
447 |
"group_size": {
|
@@ -529,7 +529,7 @@
|
|
529 |
}
|
530 |
},
|
531 |
"model.layers.6": {
|
532 |
-
"accuracy": 0.
|
533 |
"total_bits": 319680672,
|
534 |
"q_proj": {
|
535 |
"group_size": {
|
@@ -617,7 +617,7 @@
|
|
617 |
}
|
618 |
},
|
619 |
"model.layers.7": {
|
620 |
-
"accuracy": 0.
|
621 |
"total_bits": 319680672,
|
622 |
"q_proj": {
|
623 |
"group_size": {
|
@@ -705,7 +705,7 @@
|
|
705 |
}
|
706 |
},
|
707 |
"model.layers.8": {
|
708 |
-
"accuracy": 0.
|
709 |
"total_bits": 319680672,
|
710 |
"q_proj": {
|
711 |
"group_size": {
|
@@ -793,7 +793,7 @@
|
|
793 |
}
|
794 |
},
|
795 |
"model.layers.9": {
|
796 |
-
"accuracy": 0.
|
797 |
"total_bits": 319680672,
|
798 |
"q_proj": {
|
799 |
"group_size": {
|
@@ -881,7 +881,7 @@
|
|
881 |
}
|
882 |
},
|
883 |
"model.layers.10": {
|
884 |
-
"accuracy": 0.
|
885 |
"total_bits": 319680672,
|
886 |
"q_proj": {
|
887 |
"group_size": {
|
@@ -969,7 +969,7 @@
|
|
969 |
}
|
970 |
},
|
971 |
"model.layers.11": {
|
972 |
-
"accuracy": 0.
|
973 |
"total_bits": 319680672,
|
974 |
"q_proj": {
|
975 |
"group_size": {
|
@@ -1057,7 +1057,7 @@
|
|
1057 |
}
|
1058 |
},
|
1059 |
"model.layers.12": {
|
1060 |
-
"accuracy": 0.
|
1061 |
"total_bits": 319680672,
|
1062 |
"q_proj": {
|
1063 |
"group_size": {
|
@@ -1145,7 +1145,7 @@
|
|
1145 |
}
|
1146 |
},
|
1147 |
"model.layers.13": {
|
1148 |
-
"accuracy": 0.
|
1149 |
"total_bits": 319680672,
|
1150 |
"q_proj": {
|
1151 |
"group_size": {
|
@@ -1233,7 +1233,7 @@
|
|
1233 |
}
|
1234 |
},
|
1235 |
"model.layers.14": {
|
1236 |
-
"accuracy": 0.
|
1237 |
"total_bits": 319680672,
|
1238 |
"q_proj": {
|
1239 |
"group_size": {
|
@@ -1321,7 +1321,7 @@
|
|
1321 |
}
|
1322 |
},
|
1323 |
"model.layers.15": {
|
1324 |
-
"accuracy": 0.
|
1325 |
"total_bits": 319680672,
|
1326 |
"q_proj": {
|
1327 |
"group_size": {
|
@@ -1409,7 +1409,7 @@
|
|
1409 |
}
|
1410 |
},
|
1411 |
"model.layers.16": {
|
1412 |
-
"accuracy": 0.
|
1413 |
"total_bits": 319680672,
|
1414 |
"q_proj": {
|
1415 |
"group_size": {
|
@@ -1497,7 +1497,7 @@
|
|
1497 |
}
|
1498 |
},
|
1499 |
"model.layers.17": {
|
1500 |
-
"accuracy": 0.
|
1501 |
"total_bits": 319680672,
|
1502 |
"q_proj": {
|
1503 |
"group_size": {
|
@@ -1585,7 +1585,7 @@
|
|
1585 |
}
|
1586 |
},
|
1587 |
"model.layers.18": {
|
1588 |
-
"accuracy": 0.
|
1589 |
"total_bits": 319680672,
|
1590 |
"q_proj": {
|
1591 |
"group_size": {
|
@@ -1673,7 +1673,7 @@
|
|
1673 |
}
|
1674 |
},
|
1675 |
"model.layers.19": {
|
1676 |
-
"accuracy": 0.
|
1677 |
"total_bits": 319680672,
|
1678 |
"q_proj": {
|
1679 |
"group_size": {
|
@@ -1761,7 +1761,7 @@
|
|
1761 |
}
|
1762 |
},
|
1763 |
"model.layers.20": {
|
1764 |
-
"accuracy": 0.
|
1765 |
"total_bits": 319680672,
|
1766 |
"q_proj": {
|
1767 |
"group_size": {
|
@@ -1849,7 +1849,7 @@
|
|
1849 |
}
|
1850 |
},
|
1851 |
"model.layers.21": {
|
1852 |
-
"accuracy": 0.
|
1853 |
"total_bits": 319680672,
|
1854 |
"q_proj": {
|
1855 |
"group_size": {
|
@@ -1937,7 +1937,7 @@
|
|
1937 |
}
|
1938 |
},
|
1939 |
"model.layers.22": {
|
1940 |
-
"accuracy": 0.
|
1941 |
"total_bits": 319680672,
|
1942 |
"q_proj": {
|
1943 |
"group_size": {
|
@@ -2025,7 +2025,7 @@
|
|
2025 |
}
|
2026 |
},
|
2027 |
"model.layers.23": {
|
2028 |
-
"accuracy": 0.
|
2029 |
"total_bits": 319680672,
|
2030 |
"q_proj": {
|
2031 |
"group_size": {
|
@@ -2113,7 +2113,7 @@
|
|
2113 |
}
|
2114 |
},
|
2115 |
"model.layers.24": {
|
2116 |
-
"accuracy": 0.
|
2117 |
"total_bits": 319680672,
|
2118 |
"q_proj": {
|
2119 |
"group_size": {
|
@@ -2201,7 +2201,7 @@
|
|
2201 |
}
|
2202 |
},
|
2203 |
"model.layers.25": {
|
2204 |
-
"accuracy": 0.
|
2205 |
"total_bits": 319680672,
|
2206 |
"q_proj": {
|
2207 |
"group_size": {
|
@@ -2289,7 +2289,7 @@
|
|
2289 |
}
|
2290 |
},
|
2291 |
"model.layers.26": {
|
2292 |
-
"accuracy": 0.
|
2293 |
"total_bits": 319680672,
|
2294 |
"q_proj": {
|
2295 |
"group_size": {
|
@@ -2377,14 +2377,14 @@
|
|
2377 |
}
|
2378 |
},
|
2379 |
"model.layers.27": {
|
2380 |
-
"accuracy": 0.
|
2381 |
-
"total_bits":
|
2382 |
"q_proj": {
|
2383 |
"group_size": {
|
2384 |
-
"
|
2385 |
},
|
2386 |
"bits": [
|
2387 |
-
|
2388 |
],
|
2389 |
"bits_prop": [
|
2390 |
1
|
@@ -2465,7 +2465,7 @@
|
|
2465 |
}
|
2466 |
},
|
2467 |
"model.layers.28": {
|
2468 |
-
"accuracy": 0.
|
2469 |
"total_bits": 319680672,
|
2470 |
"q_proj": {
|
2471 |
"group_size": {
|
@@ -2553,7 +2553,7 @@
|
|
2553 |
}
|
2554 |
},
|
2555 |
"model.layers.29": {
|
2556 |
-
"accuracy": 0.
|
2557 |
"total_bits": 319680672,
|
2558 |
"q_proj": {
|
2559 |
"group_size": {
|
@@ -2641,7 +2641,7 @@
|
|
2641 |
}
|
2642 |
},
|
2643 |
"model.layers.30": {
|
2644 |
-
"accuracy": 0.
|
2645 |
"total_bits": 319680672,
|
2646 |
"q_proj": {
|
2647 |
"group_size": {
|
@@ -2729,7 +2729,7 @@
|
|
2729 |
}
|
2730 |
},
|
2731 |
"model.layers.31": {
|
2732 |
-
"accuracy": 0.
|
2733 |
"total_bits": 319680672,
|
2734 |
"q_proj": {
|
2735 |
"group_size": {
|
@@ -2817,7 +2817,7 @@
|
|
2817 |
}
|
2818 |
},
|
2819 |
"model.layers.32": {
|
2820 |
-
"accuracy": 0.
|
2821 |
"total_bits": 319680672,
|
2822 |
"q_proj": {
|
2823 |
"group_size": {
|
@@ -2905,7 +2905,7 @@
|
|
2905 |
}
|
2906 |
},
|
2907 |
"model.layers.33": {
|
2908 |
-
"accuracy": 0.
|
2909 |
"total_bits": 319680672,
|
2910 |
"q_proj": {
|
2911 |
"group_size": {
|
@@ -2993,7 +2993,7 @@
|
|
2993 |
}
|
2994 |
},
|
2995 |
"model.layers.34": {
|
2996 |
-
"accuracy": 0.
|
2997 |
"total_bits": 306779232,
|
2998 |
"q_proj": {
|
2999 |
"group_size": {
|
@@ -3081,7 +3081,7 @@
|
|
3081 |
}
|
3082 |
},
|
3083 |
"model.layers.35": {
|
3084 |
-
"accuracy": 0.
|
3085 |
"total_bits": 319680672,
|
3086 |
"q_proj": {
|
3087 |
"group_size": {
|
@@ -3169,14 +3169,14 @@
|
|
3169 |
}
|
3170 |
},
|
3171 |
"model.layers.36": {
|
3172 |
-
"accuracy": 0.
|
3173 |
-
"total_bits":
|
3174 |
"q_proj": {
|
3175 |
"group_size": {
|
3176 |
-
"
|
3177 |
},
|
3178 |
"bits": [
|
3179 |
-
|
3180 |
],
|
3181 |
"bits_prop": [
|
3182 |
1
|
@@ -3257,8 +3257,8 @@
|
|
3257 |
}
|
3258 |
},
|
3259 |
"model.layers.37": {
|
3260 |
-
"accuracy": 0.
|
3261 |
-
"total_bits":
|
3262 |
"q_proj": {
|
3263 |
"group_size": {
|
3264 |
"2": 64
|
@@ -3273,10 +3273,10 @@
|
|
3273 |
},
|
3274 |
"k_proj": {
|
3275 |
"group_size": {
|
3276 |
-
"
|
3277 |
},
|
3278 |
"bits": [
|
3279 |
-
|
3280 |
],
|
3281 |
"bits_prop": [
|
3282 |
1
|
@@ -3345,8 +3345,8 @@
|
|
3345 |
}
|
3346 |
},
|
3347 |
"model.layers.38": {
|
3348 |
-
"accuracy": 0.
|
3349 |
-
"total_bits":
|
3350 |
"q_proj": {
|
3351 |
"group_size": {
|
3352 |
"2": 64
|
@@ -3361,10 +3361,10 @@
|
|
3361 |
},
|
3362 |
"k_proj": {
|
3363 |
"group_size": {
|
3364 |
-
"
|
3365 |
},
|
3366 |
"bits": [
|
3367 |
-
|
3368 |
],
|
3369 |
"bits_prop": [
|
3370 |
1
|
@@ -3433,7 +3433,7 @@
|
|
3433 |
}
|
3434 |
},
|
3435 |
"model.layers.39": {
|
3436 |
-
"accuracy": 0.
|
3437 |
"total_bits": 293877792,
|
3438 |
"q_proj": {
|
3439 |
"group_size": {
|
|
|
1 |
{
|
2 |
"measurement": {
|
3 |
"model.layers.0": {
|
4 |
+
"accuracy": 0.9552268981933594,
|
5 |
"total_bits": 319680672,
|
6 |
"q_proj": {
|
7 |
"group_size": {
|
|
|
89 |
}
|
90 |
},
|
91 |
"model.layers.1": {
|
92 |
+
"accuracy": 0.9753798246383667,
|
93 |
+
"total_bits": 319680672,
|
94 |
"q_proj": {
|
95 |
"group_size": {
|
96 |
+
"4": 128
|
97 |
},
|
98 |
"bits": [
|
99 |
+
4
|
100 |
],
|
101 |
"bits_prop": [
|
102 |
1
|
|
|
105 |
},
|
106 |
"k_proj": {
|
107 |
"group_size": {
|
108 |
+
"4": 128
|
109 |
},
|
110 |
"bits": [
|
111 |
+
4
|
112 |
],
|
113 |
"bits_prop": [
|
114 |
1
|
|
|
177 |
}
|
178 |
},
|
179 |
"model.layers.2": {
|
180 |
+
"accuracy": 0.9768151044845581,
|
181 |
"total_bits": 319680672,
|
182 |
"q_proj": {
|
183 |
"group_size": {
|
|
|
265 |
}
|
266 |
},
|
267 |
"model.layers.3": {
|
268 |
+
"accuracy": 0.9676742553710938,
|
269 |
"total_bits": 319680672,
|
270 |
"q_proj": {
|
271 |
"group_size": {
|
|
|
353 |
}
|
354 |
},
|
355 |
"model.layers.4": {
|
356 |
+
"accuracy": 0.980401337146759,
|
357 |
"total_bits": 319680672,
|
358 |
"q_proj": {
|
359 |
"group_size": {
|
|
|
441 |
}
|
442 |
},
|
443 |
"model.layers.5": {
|
444 |
+
"accuracy": 0.9808422327041626,
|
445 |
"total_bits": 319680672,
|
446 |
"q_proj": {
|
447 |
"group_size": {
|
|
|
529 |
}
|
530 |
},
|
531 |
"model.layers.6": {
|
532 |
+
"accuracy": 0.9792896509170532,
|
533 |
"total_bits": 319680672,
|
534 |
"q_proj": {
|
535 |
"group_size": {
|
|
|
617 |
}
|
618 |
},
|
619 |
"model.layers.7": {
|
620 |
+
"accuracy": 0.97636878490448,
|
621 |
"total_bits": 319680672,
|
622 |
"q_proj": {
|
623 |
"group_size": {
|
|
|
705 |
}
|
706 |
},
|
707 |
"model.layers.8": {
|
708 |
+
"accuracy": 0.9747415781021118,
|
709 |
"total_bits": 319680672,
|
710 |
"q_proj": {
|
711 |
"group_size": {
|
|
|
793 |
}
|
794 |
},
|
795 |
"model.layers.9": {
|
796 |
+
"accuracy": 0.9734252691268921,
|
797 |
"total_bits": 319680672,
|
798 |
"q_proj": {
|
799 |
"group_size": {
|
|
|
881 |
}
|
882 |
},
|
883 |
"model.layers.10": {
|
884 |
+
"accuracy": 0.9705219268798828,
|
885 |
"total_bits": 319680672,
|
886 |
"q_proj": {
|
887 |
"group_size": {
|
|
|
969 |
}
|
970 |
},
|
971 |
"model.layers.11": {
|
972 |
+
"accuracy": 0.9484126567840576,
|
973 |
"total_bits": 319680672,
|
974 |
"q_proj": {
|
975 |
"group_size": {
|
|
|
1057 |
}
|
1058 |
},
|
1059 |
"model.layers.12": {
|
1060 |
+
"accuracy": 0.984208345413208,
|
1061 |
"total_bits": 319680672,
|
1062 |
"q_proj": {
|
1063 |
"group_size": {
|
|
|
1145 |
}
|
1146 |
},
|
1147 |
"model.layers.13": {
|
1148 |
+
"accuracy": 0.9829723238945007,
|
1149 |
"total_bits": 319680672,
|
1150 |
"q_proj": {
|
1151 |
"group_size": {
|
|
|
1233 |
}
|
1234 |
},
|
1235 |
"model.layers.14": {
|
1236 |
+
"accuracy": 0.9830234050750732,
|
1237 |
"total_bits": 319680672,
|
1238 |
"q_proj": {
|
1239 |
"group_size": {
|
|
|
1321 |
}
|
1322 |
},
|
1323 |
"model.layers.15": {
|
1324 |
+
"accuracy": 0.983147382736206,
|
1325 |
"total_bits": 319680672,
|
1326 |
"q_proj": {
|
1327 |
"group_size": {
|
|
|
1409 |
}
|
1410 |
},
|
1411 |
"model.layers.16": {
|
1412 |
+
"accuracy": 0.9823517799377441,
|
1413 |
"total_bits": 319680672,
|
1414 |
"q_proj": {
|
1415 |
"group_size": {
|
|
|
1497 |
}
|
1498 |
},
|
1499 |
"model.layers.17": {
|
1500 |
+
"accuracy": 0.9815922975540161,
|
1501 |
"total_bits": 319680672,
|
1502 |
"q_proj": {
|
1503 |
"group_size": {
|
|
|
1585 |
}
|
1586 |
},
|
1587 |
"model.layers.18": {
|
1588 |
+
"accuracy": 0.9808419346809387,
|
1589 |
"total_bits": 319680672,
|
1590 |
"q_proj": {
|
1591 |
"group_size": {
|
|
|
1673 |
}
|
1674 |
},
|
1675 |
"model.layers.19": {
|
1676 |
+
"accuracy": 0.9806683659553528,
|
1677 |
"total_bits": 319680672,
|
1678 |
"q_proj": {
|
1679 |
"group_size": {
|
|
|
1761 |
}
|
1762 |
},
|
1763 |
"model.layers.20": {
|
1764 |
+
"accuracy": 0.9794905185699463,
|
1765 |
"total_bits": 319680672,
|
1766 |
"q_proj": {
|
1767 |
"group_size": {
|
|
|
1849 |
}
|
1850 |
},
|
1851 |
"model.layers.21": {
|
1852 |
+
"accuracy": 0.9791617393493652,
|
1853 |
"total_bits": 319680672,
|
1854 |
"q_proj": {
|
1855 |
"group_size": {
|
|
|
1937 |
}
|
1938 |
},
|
1939 |
"model.layers.22": {
|
1940 |
+
"accuracy": 0.979049801826477,
|
1941 |
"total_bits": 319680672,
|
1942 |
"q_proj": {
|
1943 |
"group_size": {
|
|
|
2025 |
}
|
2026 |
},
|
2027 |
"model.layers.23": {
|
2028 |
+
"accuracy": 0.9779616594314575,
|
2029 |
"total_bits": 319680672,
|
2030 |
"q_proj": {
|
2031 |
"group_size": {
|
|
|
2113 |
}
|
2114 |
},
|
2115 |
"model.layers.24": {
|
2116 |
+
"accuracy": 0.9788224697113037,
|
2117 |
"total_bits": 319680672,
|
2118 |
"q_proj": {
|
2119 |
"group_size": {
|
|
|
2201 |
}
|
2202 |
},
|
2203 |
"model.layers.25": {
|
2204 |
+
"accuracy": 0.9496293067932129,
|
2205 |
"total_bits": 319680672,
|
2206 |
"q_proj": {
|
2207 |
"group_size": {
|
|
|
2289 |
}
|
2290 |
},
|
2291 |
"model.layers.26": {
|
2292 |
+
"accuracy": 0.9890510439872742,
|
2293 |
"total_bits": 319680672,
|
2294 |
"q_proj": {
|
2295 |
"group_size": {
|
|
|
2377 |
}
|
2378 |
},
|
2379 |
"model.layers.27": {
|
2380 |
+
"accuracy": 0.9852292537689209,
|
2381 |
+
"total_bits": 306779232,
|
2382 |
"q_proj": {
|
2383 |
"group_size": {
|
2384 |
+
"2": 64
|
2385 |
},
|
2386 |
"bits": [
|
2387 |
+
2
|
2388 |
],
|
2389 |
"bits_prop": [
|
2390 |
1
|
|
|
2465 |
}
|
2466 |
},
|
2467 |
"model.layers.28": {
|
2468 |
+
"accuracy": 0.9885027408599854,
|
2469 |
"total_bits": 319680672,
|
2470 |
"q_proj": {
|
2471 |
"group_size": {
|
|
|
2553 |
}
|
2554 |
},
|
2555 |
"model.layers.29": {
|
2556 |
+
"accuracy": 0.9883345365524292,
|
2557 |
"total_bits": 319680672,
|
2558 |
"q_proj": {
|
2559 |
"group_size": {
|
|
|
2641 |
}
|
2642 |
},
|
2643 |
"model.layers.30": {
|
2644 |
+
"accuracy": 0.9879627823829651,
|
2645 |
"total_bits": 319680672,
|
2646 |
"q_proj": {
|
2647 |
"group_size": {
|
|
|
2729 |
}
|
2730 |
},
|
2731 |
"model.layers.31": {
|
2732 |
+
"accuracy": 0.987734854221344,
|
2733 |
"total_bits": 319680672,
|
2734 |
"q_proj": {
|
2735 |
"group_size": {
|
|
|
2817 |
}
|
2818 |
},
|
2819 |
"model.layers.32": {
|
2820 |
+
"accuracy": 0.9877235889434814,
|
2821 |
"total_bits": 319680672,
|
2822 |
"q_proj": {
|
2823 |
"group_size": {
|
|
|
2905 |
}
|
2906 |
},
|
2907 |
"model.layers.33": {
|
2908 |
+
"accuracy": 0.9874624013900757,
|
2909 |
"total_bits": 319680672,
|
2910 |
"q_proj": {
|
2911 |
"group_size": {
|
|
|
2993 |
}
|
2994 |
},
|
2995 |
"model.layers.34": {
|
2996 |
+
"accuracy": 0.9844291806221008,
|
2997 |
"total_bits": 306779232,
|
2998 |
"q_proj": {
|
2999 |
"group_size": {
|
|
|
3081 |
}
|
3082 |
},
|
3083 |
"model.layers.35": {
|
3084 |
+
"accuracy": 0.9864280223846436,
|
3085 |
"total_bits": 319680672,
|
3086 |
"q_proj": {
|
3087 |
"group_size": {
|
|
|
3169 |
}
|
3170 |
},
|
3171 |
"model.layers.36": {
|
3172 |
+
"accuracy": 0.9864512085914612,
|
3173 |
+
"total_bits": 319680672,
|
3174 |
"q_proj": {
|
3175 |
"group_size": {
|
3176 |
+
"4": 128
|
3177 |
},
|
3178 |
"bits": [
|
3179 |
+
4
|
3180 |
],
|
3181 |
"bits_prop": [
|
3182 |
1
|
|
|
3257 |
}
|
3258 |
},
|
3259 |
"model.layers.37": {
|
3260 |
+
"accuracy": 0.9816058874130249,
|
3261 |
+
"total_bits": 293877792,
|
3262 |
"q_proj": {
|
3263 |
"group_size": {
|
3264 |
"2": 64
|
|
|
3273 |
},
|
3274 |
"k_proj": {
|
3275 |
"group_size": {
|
3276 |
+
"2": 64
|
3277 |
},
|
3278 |
"bits": [
|
3279 |
+
2
|
3280 |
],
|
3281 |
"bits_prop": [
|
3282 |
1
|
|
|
3345 |
}
|
3346 |
},
|
3347 |
"model.layers.38": {
|
3348 |
+
"accuracy": 0.9656969308853149,
|
3349 |
+
"total_bits": 293877792,
|
3350 |
"q_proj": {
|
3351 |
"group_size": {
|
3352 |
"2": 64
|
|
|
3361 |
},
|
3362 |
"k_proj": {
|
3363 |
"group_size": {
|
3364 |
+
"2": 64
|
3365 |
},
|
3366 |
"bits": [
|
3367 |
+
2
|
3368 |
],
|
3369 |
"bits_prop": [
|
3370 |
1
|
|
|
3433 |
}
|
3434 |
},
|
3435 |
"model.layers.39": {
|
3436 |
+
"accuracy": 0.9720205068588257,
|
3437 |
"total_bits": 293877792,
|
3438 |
"q_proj": {
|
3439 |
"group_size": {
|