diff --git "a/config.json" "b/config.json" --- "a/config.json" +++ "b/config.json" @@ -1,5 +1,5 @@ { - "_name_or_path": "Qwen/Qwen2.5-32B-Instruct", + "_name_or_path": "/home/aiscuser/yangwang/convert_model/models/VPTQ-community--Qwen2.5-32B-Instruct-v8-k65536-0-woft", "architectures": [ "Qwen2ForCausalLM" ], @@ -16,22 +16,12 @@ "num_attention_heads": 40, "num_hidden_layers": 64, "num_key_value_heads": 8, - "rms_norm_eps": 1e-06, - "rope_theta": 1000000.0, - "sliding_window": null, - "tie_word_embeddings": false, - "torch_dtype": "bfloat16", - "transformers_version": "4.37.2", - "use_cache": true, - "use_sliding_window": false, - "vocab_size": 152064, "quantization_config": { - "quant_method": "vptq", "config_for_layers": { "model.layers.0.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -55,7 +45,7 @@ "model.layers.0.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -79,7 +69,7 @@ "model.layers.0.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -103,7 +93,7 @@ "model.layers.0.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -127,7 +117,7 @@ "model.layers.0.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -151,7 +141,7 @@ "model.layers.0.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -175,7 +165,7 @@ "model.layers.0.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -199,7 +189,7 @@ "model.layers.1.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -223,7 +213,7 @@ "model.layers.1.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -247,7 +237,7 @@ "model.layers.1.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -271,7 +261,7 @@ "model.layers.1.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -295,7 +285,7 @@ "model.layers.1.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -319,7 +309,7 @@ "model.layers.1.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -343,7 +333,7 @@ "model.layers.1.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -367,7 +357,7 @@ "model.layers.10.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -391,7 +381,7 @@ "model.layers.10.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -415,7 +405,7 @@ "model.layers.10.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -439,7 +429,7 @@ "model.layers.10.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -463,7 +453,7 @@ "model.layers.10.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -487,7 +477,7 @@ "model.layers.10.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -511,7 +501,7 @@ "model.layers.10.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -535,7 +525,7 @@ "model.layers.11.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -559,7 +549,7 @@ "model.layers.11.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -583,7 +573,7 @@ "model.layers.11.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -607,7 +597,7 @@ "model.layers.11.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -631,7 +621,7 @@ "model.layers.11.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -655,7 +645,7 @@ "model.layers.11.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -679,7 +669,7 @@ "model.layers.11.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -703,7 +693,7 @@ "model.layers.12.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -727,7 +717,7 @@ "model.layers.12.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -751,7 +741,7 @@ "model.layers.12.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -775,7 +765,7 @@ "model.layers.12.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -799,7 +789,7 @@ "model.layers.12.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -823,7 +813,7 @@ "model.layers.12.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -847,7 +837,7 @@ "model.layers.12.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -871,7 +861,7 @@ "model.layers.13.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -895,7 +885,7 @@ "model.layers.13.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -919,7 +909,7 @@ "model.layers.13.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -943,7 +933,7 @@ "model.layers.13.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -967,7 +957,7 @@ "model.layers.13.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -991,7 +981,7 @@ "model.layers.13.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1015,7 +1005,7 @@ "model.layers.13.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1039,7 +1029,7 @@ "model.layers.14.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -1063,7 +1053,7 @@ "model.layers.14.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1087,7 +1077,7 @@ "model.layers.14.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1111,7 +1101,7 @@ "model.layers.14.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1135,7 +1125,7 @@ "model.layers.14.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1159,7 +1149,7 @@ "model.layers.14.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1183,7 +1173,7 @@ "model.layers.14.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1207,7 +1197,7 @@ "model.layers.15.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -1231,7 +1221,7 @@ "model.layers.15.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1255,7 +1245,7 @@ "model.layers.15.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1279,7 +1269,7 @@ "model.layers.15.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1303,7 +1293,7 @@ "model.layers.15.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1327,7 +1317,7 @@ "model.layers.15.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1351,7 +1341,7 @@ "model.layers.15.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1375,7 +1365,7 @@ "model.layers.16.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -1399,7 +1389,7 @@ "model.layers.16.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1423,7 +1413,7 @@ "model.layers.16.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1447,7 +1437,7 @@ "model.layers.16.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1471,7 +1461,7 @@ "model.layers.16.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1495,7 +1485,7 @@ "model.layers.16.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1519,7 +1509,7 @@ "model.layers.16.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1543,7 +1533,7 @@ "model.layers.17.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -1567,7 +1557,7 @@ "model.layers.17.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1591,7 +1581,7 @@ "model.layers.17.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1615,7 +1605,7 @@ "model.layers.17.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1639,7 +1629,7 @@ "model.layers.17.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1663,7 +1653,7 @@ "model.layers.17.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1687,7 +1677,7 @@ "model.layers.17.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1711,7 +1701,7 @@ "model.layers.18.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -1735,7 +1725,7 @@ "model.layers.18.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1759,7 +1749,7 @@ "model.layers.18.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1783,7 +1773,7 @@ "model.layers.18.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1807,7 +1797,7 @@ "model.layers.18.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1831,7 +1821,7 @@ "model.layers.18.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1855,7 +1845,7 @@ "model.layers.18.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1879,7 +1869,7 @@ "model.layers.19.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -1903,7 +1893,7 @@ "model.layers.19.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1927,7 +1917,7 @@ "model.layers.19.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1951,7 +1941,7 @@ "model.layers.19.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1975,7 +1965,7 @@ "model.layers.19.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -1999,7 +1989,7 @@ "model.layers.19.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2023,7 +2013,7 @@ "model.layers.19.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2047,7 +2037,7 @@ "model.layers.2.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -2071,7 +2061,7 @@ "model.layers.2.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2095,7 +2085,7 @@ "model.layers.2.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2119,7 +2109,7 @@ "model.layers.2.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2143,7 +2133,7 @@ "model.layers.2.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2167,7 +2157,7 @@ "model.layers.2.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2191,7 +2181,7 @@ "model.layers.2.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2215,7 +2205,7 @@ "model.layers.20.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -2239,7 +2229,7 @@ "model.layers.20.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2263,7 +2253,7 @@ "model.layers.20.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2287,7 +2277,7 @@ "model.layers.20.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2311,7 +2301,7 @@ "model.layers.20.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2335,7 +2325,7 @@ "model.layers.20.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2359,7 +2349,7 @@ "model.layers.20.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2383,7 +2373,7 @@ "model.layers.21.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -2407,7 +2397,7 @@ "model.layers.21.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2431,7 +2421,7 @@ "model.layers.21.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2455,7 +2445,7 @@ "model.layers.21.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2479,7 +2469,7 @@ "model.layers.21.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2503,7 +2493,7 @@ "model.layers.21.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2527,7 +2517,7 @@ "model.layers.21.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2551,7 +2541,7 @@ "model.layers.22.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -2575,7 +2565,7 @@ "model.layers.22.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2599,7 +2589,7 @@ "model.layers.22.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2623,7 +2613,7 @@ "model.layers.22.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2647,7 +2637,7 @@ "model.layers.22.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2671,7 +2661,7 @@ "model.layers.22.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2695,7 +2685,7 @@ "model.layers.22.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2719,7 +2709,7 @@ "model.layers.23.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -2743,7 +2733,7 @@ "model.layers.23.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2767,7 +2757,7 @@ "model.layers.23.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2791,7 +2781,7 @@ "model.layers.23.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2815,7 +2805,7 @@ "model.layers.23.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2839,7 +2829,7 @@ "model.layers.23.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2863,7 +2853,7 @@ "model.layers.23.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2887,7 +2877,7 @@ "model.layers.24.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -2911,7 +2901,7 @@ "model.layers.24.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2935,7 +2925,7 @@ "model.layers.24.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2959,7 +2949,7 @@ "model.layers.24.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -2983,7 +2973,7 @@ "model.layers.24.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3007,7 +2997,7 @@ "model.layers.24.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3031,7 +3021,7 @@ "model.layers.24.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3055,7 +3045,7 @@ "model.layers.25.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -3079,7 +3069,7 @@ "model.layers.25.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3103,7 +3093,7 @@ "model.layers.25.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3127,7 +3117,7 @@ "model.layers.25.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3151,7 +3141,7 @@ "model.layers.25.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3175,7 +3165,7 @@ "model.layers.25.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3199,7 +3189,7 @@ "model.layers.25.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3223,7 +3213,7 @@ "model.layers.26.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -3247,7 +3237,7 @@ "model.layers.26.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3271,7 +3261,7 @@ "model.layers.26.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3295,7 +3285,7 @@ "model.layers.26.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3319,7 +3309,7 @@ "model.layers.26.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3343,7 +3333,7 @@ "model.layers.26.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3367,7 +3357,7 @@ "model.layers.26.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3391,7 +3381,7 @@ "model.layers.27.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -3415,7 +3405,7 @@ "model.layers.27.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3439,7 +3429,7 @@ "model.layers.27.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3463,7 +3453,7 @@ "model.layers.27.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3487,7 +3477,7 @@ "model.layers.27.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3511,7 +3501,7 @@ "model.layers.27.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3535,7 +3525,7 @@ "model.layers.27.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3559,7 +3549,7 @@ "model.layers.28.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -3583,7 +3573,7 @@ "model.layers.28.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3607,7 +3597,7 @@ "model.layers.28.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3631,7 +3621,7 @@ "model.layers.28.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3655,7 +3645,7 @@ "model.layers.28.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3679,7 +3669,7 @@ "model.layers.28.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3703,7 +3693,7 @@ "model.layers.28.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3727,7 +3717,7 @@ "model.layers.29.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -3751,7 +3741,7 @@ "model.layers.29.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3775,7 +3765,7 @@ "model.layers.29.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3799,7 +3789,7 @@ "model.layers.29.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3823,7 +3813,7 @@ "model.layers.29.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3847,7 +3837,7 @@ "model.layers.29.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3871,7 +3861,7 @@ "model.layers.29.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3895,7 +3885,7 @@ "model.layers.3.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -3919,7 +3909,7 @@ "model.layers.3.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3943,7 +3933,7 @@ "model.layers.3.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3967,7 +3957,7 @@ "model.layers.3.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -3991,7 +3981,7 @@ "model.layers.3.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4015,7 +4005,7 @@ "model.layers.3.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4039,7 +4029,7 @@ "model.layers.3.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4063,7 +4053,7 @@ "model.layers.30.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -4087,7 +4077,7 @@ "model.layers.30.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4111,7 +4101,7 @@ "model.layers.30.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4135,7 +4125,7 @@ "model.layers.30.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4159,7 +4149,7 @@ "model.layers.30.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4183,7 +4173,7 @@ "model.layers.30.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4207,7 +4197,7 @@ "model.layers.30.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4231,7 +4221,7 @@ "model.layers.31.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -4255,7 +4245,7 @@ "model.layers.31.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4279,7 +4269,7 @@ "model.layers.31.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4303,7 +4293,7 @@ "model.layers.31.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4327,7 +4317,7 @@ "model.layers.31.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4351,7 +4341,7 @@ "model.layers.31.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4375,7 +4365,7 @@ "model.layers.31.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4399,7 +4389,7 @@ "model.layers.32.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -4423,7 +4413,7 @@ "model.layers.32.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4447,7 +4437,7 @@ "model.layers.32.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4471,7 +4461,7 @@ "model.layers.32.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4495,7 +4485,7 @@ "model.layers.32.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4519,7 +4509,7 @@ "model.layers.32.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4543,7 +4533,7 @@ "model.layers.32.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4567,7 +4557,7 @@ "model.layers.33.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -4591,7 +4581,7 @@ "model.layers.33.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4615,7 +4605,7 @@ "model.layers.33.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4639,7 +4629,7 @@ "model.layers.33.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4663,7 +4653,7 @@ "model.layers.33.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4687,7 +4677,7 @@ "model.layers.33.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4711,7 +4701,7 @@ "model.layers.33.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4735,7 +4725,7 @@ "model.layers.34.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -4759,7 +4749,7 @@ "model.layers.34.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4783,7 +4773,7 @@ "model.layers.34.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4807,7 +4797,7 @@ "model.layers.34.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4831,7 +4821,7 @@ "model.layers.34.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4855,7 +4845,7 @@ "model.layers.34.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4879,7 +4869,7 @@ "model.layers.34.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4903,7 +4893,7 @@ "model.layers.35.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -4927,7 +4917,7 @@ "model.layers.35.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4951,7 +4941,7 @@ "model.layers.35.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4975,7 +4965,7 @@ "model.layers.35.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -4999,7 +4989,7 @@ "model.layers.35.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5023,7 +5013,7 @@ "model.layers.35.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5047,7 +5037,7 @@ "model.layers.35.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5071,7 +5061,7 @@ "model.layers.36.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -5095,7 +5085,7 @@ "model.layers.36.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5119,7 +5109,7 @@ "model.layers.36.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5143,7 +5133,7 @@ "model.layers.36.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5167,7 +5157,7 @@ "model.layers.36.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5191,7 +5181,7 @@ "model.layers.36.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5215,7 +5205,7 @@ "model.layers.36.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5239,7 +5229,7 @@ "model.layers.37.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -5263,7 +5253,7 @@ "model.layers.37.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5287,7 +5277,7 @@ "model.layers.37.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5311,7 +5301,7 @@ "model.layers.37.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5335,7 +5325,7 @@ "model.layers.37.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5359,7 +5349,7 @@ "model.layers.37.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5383,7 +5373,7 @@ "model.layers.37.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5407,7 +5397,7 @@ "model.layers.38.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -5431,7 +5421,7 @@ "model.layers.38.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5455,7 +5445,7 @@ "model.layers.38.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5479,7 +5469,7 @@ "model.layers.38.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5503,7 +5493,7 @@ "model.layers.38.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5527,7 +5517,7 @@ "model.layers.38.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5551,7 +5541,7 @@ "model.layers.38.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5575,7 +5565,7 @@ "model.layers.39.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -5599,7 +5589,7 @@ "model.layers.39.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5623,7 +5613,7 @@ "model.layers.39.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5647,7 +5637,7 @@ "model.layers.39.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5671,7 +5661,7 @@ "model.layers.39.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5695,7 +5685,7 @@ "model.layers.39.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5719,7 +5709,7 @@ "model.layers.39.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5743,7 +5733,7 @@ "model.layers.4.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -5767,7 +5757,7 @@ "model.layers.4.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5791,7 +5781,7 @@ "model.layers.4.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5815,7 +5805,7 @@ "model.layers.4.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5839,7 +5829,7 @@ "model.layers.4.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5863,7 +5853,7 @@ "model.layers.4.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5887,7 +5877,7 @@ "model.layers.4.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5911,7 +5901,7 @@ "model.layers.40.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -5935,7 +5925,7 @@ "model.layers.40.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5959,7 +5949,7 @@ "model.layers.40.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -5983,7 +5973,7 @@ "model.layers.40.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6007,7 +5997,7 @@ "model.layers.40.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6031,7 +6021,7 @@ "model.layers.40.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6055,7 +6045,7 @@ "model.layers.40.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6079,7 +6069,7 @@ "model.layers.41.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -6103,7 +6093,7 @@ "model.layers.41.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6127,7 +6117,7 @@ "model.layers.41.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6151,7 +6141,7 @@ "model.layers.41.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6175,7 +6165,7 @@ "model.layers.41.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6199,7 +6189,7 @@ "model.layers.41.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6223,7 +6213,7 @@ "model.layers.41.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6247,7 +6237,7 @@ "model.layers.42.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -6271,7 +6261,7 @@ "model.layers.42.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6295,7 +6285,7 @@ "model.layers.42.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6319,7 +6309,7 @@ "model.layers.42.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6343,7 +6333,7 @@ "model.layers.42.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6367,7 +6357,7 @@ "model.layers.42.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6391,7 +6381,7 @@ "model.layers.42.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6415,7 +6405,7 @@ "model.layers.43.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -6439,7 +6429,7 @@ "model.layers.43.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6463,7 +6453,7 @@ "model.layers.43.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6487,7 +6477,7 @@ "model.layers.43.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6511,7 +6501,7 @@ "model.layers.43.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6535,7 +6525,7 @@ "model.layers.43.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6559,7 +6549,7 @@ "model.layers.43.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6583,7 +6573,7 @@ "model.layers.44.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -6607,7 +6597,7 @@ "model.layers.44.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6631,7 +6621,7 @@ "model.layers.44.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6655,7 +6645,7 @@ "model.layers.44.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6679,7 +6669,7 @@ "model.layers.44.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6703,7 +6693,7 @@ "model.layers.44.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6727,7 +6717,7 @@ "model.layers.44.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6751,7 +6741,7 @@ "model.layers.45.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -6775,7 +6765,7 @@ "model.layers.45.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6799,7 +6789,7 @@ "model.layers.45.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6823,7 +6813,7 @@ "model.layers.45.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6847,7 +6837,7 @@ "model.layers.45.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6871,7 +6861,7 @@ "model.layers.45.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6895,7 +6885,7 @@ "model.layers.45.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6919,7 +6909,7 @@ "model.layers.46.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -6943,7 +6933,7 @@ "model.layers.46.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6967,7 +6957,7 @@ "model.layers.46.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -6991,7 +6981,7 @@ "model.layers.46.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7015,7 +7005,7 @@ "model.layers.46.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7039,7 +7029,7 @@ "model.layers.46.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7063,7 +7053,7 @@ "model.layers.46.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7087,7 +7077,7 @@ "model.layers.47.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -7111,7 +7101,7 @@ "model.layers.47.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7135,7 +7125,7 @@ "model.layers.47.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7159,7 +7149,7 @@ "model.layers.47.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7183,7 +7173,7 @@ "model.layers.47.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7207,7 +7197,7 @@ "model.layers.47.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7231,7 +7221,7 @@ "model.layers.47.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7255,7 +7245,7 @@ "model.layers.48.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -7279,7 +7269,7 @@ "model.layers.48.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7303,7 +7293,7 @@ "model.layers.48.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7327,7 +7317,7 @@ "model.layers.48.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7351,7 +7341,7 @@ "model.layers.48.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7375,7 +7365,7 @@ "model.layers.48.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7399,7 +7389,7 @@ "model.layers.48.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7423,7 +7413,7 @@ "model.layers.49.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -7447,7 +7437,7 @@ "model.layers.49.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7471,7 +7461,7 @@ "model.layers.49.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7495,7 +7485,7 @@ "model.layers.49.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7519,7 +7509,7 @@ "model.layers.49.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7543,7 +7533,7 @@ "model.layers.49.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7567,7 +7557,7 @@ "model.layers.49.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7591,7 +7581,7 @@ "model.layers.5.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -7615,7 +7605,7 @@ "model.layers.5.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7639,7 +7629,7 @@ "model.layers.5.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7663,7 +7653,7 @@ "model.layers.5.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7687,7 +7677,7 @@ "model.layers.5.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7711,7 +7701,7 @@ "model.layers.5.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7735,7 +7725,7 @@ "model.layers.5.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7759,7 +7749,7 @@ "model.layers.50.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -7783,7 +7773,7 @@ "model.layers.50.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7807,7 +7797,7 @@ "model.layers.50.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7831,7 +7821,7 @@ "model.layers.50.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7855,7 +7845,7 @@ "model.layers.50.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7879,7 +7869,7 @@ "model.layers.50.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7903,7 +7893,7 @@ "model.layers.50.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7927,7 +7917,7 @@ "model.layers.51.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -7951,7 +7941,7 @@ "model.layers.51.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7975,7 +7965,7 @@ "model.layers.51.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -7999,7 +7989,7 @@ "model.layers.51.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8023,7 +8013,7 @@ "model.layers.51.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8047,7 +8037,7 @@ "model.layers.51.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8071,7 +8061,7 @@ "model.layers.51.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8095,7 +8085,7 @@ "model.layers.52.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -8119,7 +8109,7 @@ "model.layers.52.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8143,7 +8133,7 @@ "model.layers.52.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8167,7 +8157,7 @@ "model.layers.52.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8191,7 +8181,7 @@ "model.layers.52.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8215,7 +8205,7 @@ "model.layers.52.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8239,7 +8229,7 @@ "model.layers.52.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8263,7 +8253,7 @@ "model.layers.53.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -8287,7 +8277,7 @@ "model.layers.53.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8311,7 +8301,7 @@ "model.layers.53.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8335,7 +8325,7 @@ "model.layers.53.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8359,7 +8349,7 @@ "model.layers.53.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8383,7 +8373,7 @@ "model.layers.53.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8407,7 +8397,7 @@ "model.layers.53.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8431,7 +8421,7 @@ "model.layers.54.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -8455,7 +8445,7 @@ "model.layers.54.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8479,7 +8469,7 @@ "model.layers.54.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8503,7 +8493,7 @@ "model.layers.54.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8527,7 +8517,7 @@ "model.layers.54.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8551,7 +8541,7 @@ "model.layers.54.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8575,7 +8565,7 @@ "model.layers.54.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8599,7 +8589,7 @@ "model.layers.55.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -8623,7 +8613,7 @@ "model.layers.55.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8647,7 +8637,7 @@ "model.layers.55.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8671,7 +8661,7 @@ "model.layers.55.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8695,7 +8685,7 @@ "model.layers.55.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8719,7 +8709,7 @@ "model.layers.55.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8743,7 +8733,7 @@ "model.layers.55.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8767,7 +8757,7 @@ "model.layers.56.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -8791,7 +8781,7 @@ "model.layers.56.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8815,7 +8805,7 @@ "model.layers.56.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8839,7 +8829,7 @@ "model.layers.56.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8863,7 +8853,7 @@ "model.layers.56.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8887,7 +8877,7 @@ "model.layers.56.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8911,7 +8901,7 @@ "model.layers.56.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8935,7 +8925,7 @@ "model.layers.57.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -8959,7 +8949,7 @@ "model.layers.57.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -8983,7 +8973,7 @@ "model.layers.57.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9007,7 +8997,7 @@ "model.layers.57.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9031,7 +9021,7 @@ "model.layers.57.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9055,7 +9045,7 @@ "model.layers.57.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9079,7 +9069,7 @@ "model.layers.57.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9103,7 +9093,7 @@ "model.layers.58.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -9127,7 +9117,7 @@ "model.layers.58.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9151,7 +9141,7 @@ "model.layers.58.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9175,7 +9165,7 @@ "model.layers.58.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9199,7 +9189,7 @@ "model.layers.58.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9223,7 +9213,7 @@ "model.layers.58.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9247,7 +9237,7 @@ "model.layers.58.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9271,7 +9261,7 @@ "model.layers.59.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -9295,7 +9285,7 @@ "model.layers.59.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9319,7 +9309,7 @@ "model.layers.59.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9343,7 +9333,7 @@ "model.layers.59.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9367,7 +9357,7 @@ "model.layers.59.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9391,7 +9381,7 @@ "model.layers.59.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9415,7 +9405,7 @@ "model.layers.59.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9439,7 +9429,7 @@ "model.layers.6.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -9463,7 +9453,7 @@ "model.layers.6.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9487,7 +9477,7 @@ "model.layers.6.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9511,7 +9501,7 @@ "model.layers.6.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9535,7 +9525,7 @@ "model.layers.6.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9559,7 +9549,7 @@ "model.layers.6.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9583,7 +9573,7 @@ "model.layers.6.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9607,7 +9597,7 @@ "model.layers.60.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -9631,7 +9621,7 @@ "model.layers.60.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9655,7 +9645,7 @@ "model.layers.60.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9679,7 +9669,7 @@ "model.layers.60.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9703,7 +9693,7 @@ "model.layers.60.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9727,7 +9717,7 @@ "model.layers.60.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9751,7 +9741,7 @@ "model.layers.60.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9775,7 +9765,7 @@ "model.layers.61.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -9799,7 +9789,7 @@ "model.layers.61.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9823,7 +9813,7 @@ "model.layers.61.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9847,7 +9837,7 @@ "model.layers.61.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9871,7 +9861,7 @@ "model.layers.61.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9895,7 +9885,7 @@ "model.layers.61.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9919,7 +9909,7 @@ "model.layers.61.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9943,7 +9933,7 @@ "model.layers.62.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -9967,7 +9957,7 @@ "model.layers.62.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -9991,7 +9981,7 @@ "model.layers.62.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10015,7 +10005,7 @@ "model.layers.62.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10039,7 +10029,7 @@ "model.layers.62.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10063,7 +10053,7 @@ "model.layers.62.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10087,7 +10077,7 @@ "model.layers.62.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10111,7 +10101,7 @@ "model.layers.63.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -10135,7 +10125,7 @@ "model.layers.63.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10159,7 +10149,7 @@ "model.layers.63.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10183,7 +10173,7 @@ "model.layers.63.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10207,7 +10197,7 @@ "model.layers.63.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10231,7 +10221,7 @@ "model.layers.63.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10255,7 +10245,7 @@ "model.layers.63.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10279,7 +10269,7 @@ "model.layers.7.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -10303,7 +10293,7 @@ "model.layers.7.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10327,7 +10317,7 @@ "model.layers.7.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10351,7 +10341,7 @@ "model.layers.7.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10375,7 +10365,7 @@ "model.layers.7.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10399,7 +10389,7 @@ "model.layers.7.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10423,7 +10413,7 @@ "model.layers.7.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10447,7 +10437,7 @@ "model.layers.8.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -10471,7 +10461,7 @@ "model.layers.8.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10495,7 +10485,7 @@ "model.layers.8.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10519,7 +10509,7 @@ "model.layers.8.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10543,7 +10533,7 @@ "model.layers.8.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10567,7 +10557,7 @@ "model.layers.8.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10591,7 +10581,7 @@ "model.layers.8.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10615,7 +10605,7 @@ "model.layers.9.mlp.down_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 27648, "in_features": 27648, @@ -10639,7 +10629,7 @@ "model.layers.9.mlp.gate_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10663,7 +10653,7 @@ "model.layers.9.mlp.up_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10687,7 +10677,7 @@ "model.layers.9.self_attn.k_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10711,7 +10701,7 @@ "model.layers.9.self_attn.o_proj": { "bias": null, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10735,7 +10725,7 @@ "model.layers.9.self_attn.q_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10759,7 +10749,7 @@ "model.layers.9.self_attn.v_proj": { "bias": true, "enable_norm": true, - "enable_perm": true, + "enable_perm": false, "group_num": 1, "group_size": 5120, "in_features": 5120, @@ -10780,6 +10770,17 @@ 8 ] } - } - } -} \ No newline at end of file + }, + "quant_method": "vptq" + }, + "rms_norm_eps": 1e-06, + "rope_scaling": null, + "rope_theta": 1000000.0, + "sliding_window": null, + "tie_word_embeddings": false, + "torch_dtype": "bfloat16", + "transformers_version": "4.49.0", + "use_cache": true, + "use_sliding_window": false, + "vocab_size": 152064 +}