diff --git a/.gitattributes b/.gitattributes index a6344aac8c09253b3b630fb776ae94478aa0275b..e34b0e6dece25f8e1a71c7a81623f31e4f9f622a 100644 --- a/.gitattributes +++ b/.gitattributes @@ -33,3 +33,30 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text *.zip filter=lfs diff=lfs merge=lfs -text *.zst filter=lfs diff=lfs merge=lfs -text *tfevents* filter=lfs diff=lfs merge=lfs -text +models--LLM360--K2/blobs/0e5448b8c9a7a705501c5103af02f076036414b4cf54083c9ad17048d97221b5 filter=lfs diff=lfs merge=lfs -text +models--LLM360--K2/blobs/0ef2f96b2f9b4943f45dc1ba39f0594eb14cf128c5ff5f41db34ad07cb556e7f filter=lfs diff=lfs merge=lfs -text +models--LLM360--K2/blobs/197922ce7f7807ba59bfdd42fbbe9c67e2ff8370ffda566614cdf74463363f2b filter=lfs diff=lfs merge=lfs -text +models--LLM360--K2/blobs/1991b47bdfe3cb7e05d4a6f12d1bfb38a22fedf2da68d22269e675978940b57f filter=lfs diff=lfs merge=lfs -text +models--LLM360--K2/blobs/3bfc25078191c7795def871046f6cb059474fcef43440f4f6b9db0fd757a7f64 filter=lfs diff=lfs merge=lfs -text +models--LLM360--K2/blobs/3df9e12db23090ffcb07219d5c388eabdd3cc780587d368242b398fdf8e90b4d filter=lfs diff=lfs merge=lfs -text +models--LLM360--K2/blobs/434fd7e790c2995a1f8c0ca46f6a3a5dea2c1d7d0366ed1b47e5bdec80314442 filter=lfs diff=lfs merge=lfs -text +models--LLM360--K2/blobs/4a1ff5c1228e7d11248f2fcfbd70dc59bb26a4ed737becb63e2e130b028e1698 filter=lfs diff=lfs merge=lfs -text +models--LLM360--K2/blobs/4f3b40b1a1fef50d05c3f48ce8a22bd543fd3abe376153290bacc3d37134b463 filter=lfs diff=lfs merge=lfs -text +models--LLM360--K2/blobs/583722a14a478af056eb7190401b2fa95c90d75ab1de83acb4d5f20941a224d1 filter=lfs diff=lfs merge=lfs -text +models--LLM360--K2/blobs/5afe6368480862ccd7e5fab96e17159a6a56a6ef2f75083ac04b69df977784da filter=lfs diff=lfs merge=lfs -text +models--LLM360--K2/blobs/5b4779f96a5161dccf4a59cc5479918f931c47342fa44ab01a9272f476963083 filter=lfs diff=lfs merge=lfs -text +models--LLM360--K2/blobs/617edceaa3b26e387a4d9aec6e8104db2c81ac35e5bf47f9e518af22e956094a filter=lfs diff=lfs merge=lfs -text +models--LLM360--K2/blobs/66a1a4140058b1e5040be2ebbb351ba3c73051295dbaf353487860631be06bce filter=lfs diff=lfs merge=lfs -text +models--LLM360--K2/blobs/72b7ab829379c5ce42fddf9b8ebae0ff50f967d0cd049af34c796be1dbcc3837 filter=lfs diff=lfs merge=lfs -text +models--LLM360--K2/blobs/7c546f28758f14aea461f4eedba1dc6380ffb525bcb06328c2e7ccf4bc0914e3 filter=lfs diff=lfs merge=lfs -text +models--LLM360--K2/blobs/80860954ce526545b5eaa951ed98a9dc87576e0ce5243b53001e79a90f107ac2 filter=lfs diff=lfs merge=lfs -text +models--LLM360--K2/blobs/8fecd1dfe9745e2b6546298051b39dd02ace8b88eb0232a9d6a9beccb935f9b6 filter=lfs diff=lfs merge=lfs -text +models--LLM360--K2/blobs/9844405d62d9f809278ed1577b4364f68327282f461b4afaf75a9cebf074c325 filter=lfs diff=lfs merge=lfs -text +models--LLM360--K2/blobs/a258e4bed92e87cfeacce9077d9e2989e45436d6259e6f16f1cace7d7fcdd6d8 filter=lfs diff=lfs merge=lfs -text +models--LLM360--K2/blobs/b6c7e1e011596a62423becf7831de6885bd4cb66835c9b12a1890d82341caa5d filter=lfs diff=lfs merge=lfs -text +models--LLM360--K2/blobs/c0c42dfdce41209444c998d0f30f5bdabf5e747f9ba18f5a917a3e72c8a4956b filter=lfs diff=lfs merge=lfs -text +models--LLM360--K2/blobs/c1f7a86a6d975dad016a1692a17b3c8116c1ff4ef04362060addfc1b65aa2c08 filter=lfs diff=lfs merge=lfs -text +models--LLM360--K2/blobs/c55743fc3d25e389dc4e29f7bb2db5e21f033fc8292ffe411eed32fdff0b2498 filter=lfs diff=lfs merge=lfs -text +models--LLM360--K2/blobs/c9a048c6dfcef5561324dbded792521edfb1d9c7a03404189d2a09157c628760 filter=lfs diff=lfs merge=lfs -text +models--LLM360--K2/blobs/ca618179df960063f7086853a0de59fd2fcfa2b7a220fe089d66743aa6648b8e filter=lfs diff=lfs merge=lfs -text +models--LLM360--K2/blobs/d3ebaa9cb6f5e87aa968768383824375290dc33e56a1852e48821fe285fd8448 filter=lfs diff=lfs merge=lfs -text diff --git a/model-00001-of-00027.safetensors b/model-00001-of-00027.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1ec02f8b8d2e341931a3e85ac829e74b7571088e --- /dev/null +++ b/model-00001-of-00027.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1497ed863a5cfc16bcf61d3f523c386826180489b6949947a65aef7a336fc6f3 +size 4660464344 diff --git a/model-00002-of-00027.safetensors b/model-00002-of-00027.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..20b59736232ec37e8337b4c5c5f94cded1388357 --- /dev/null +++ b/model-00002-of-00027.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aef65505f84615f679454ea67efa72b8d64a2e7eea5234a35e2aa9cba0616ba8 +size 4857105448 diff --git a/model-00003-of-00027.safetensors b/model-00003-of-00027.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4cc6ece1000889ef6c76a7daccf486d0b7ae8f0f --- /dev/null +++ b/model-00003-of-00027.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e23d53ee9dac243cd891dd9fe5692fa87f443ec4ee68c90cbc64359a1311da14 +size 4857105448 diff --git a/model-00004-of-00027.safetensors b/model-00004-of-00027.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ba1eab46e4e7c1c06b3089f8941809775956541a --- /dev/null +++ b/model-00004-of-00027.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6155fae8d54d544fee94834dcce7ecb97c4aff8075713fad9d6503d0082d6b1b +size 4857105472 diff --git a/model-00005-of-00027.safetensors b/model-00005-of-00027.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..791423215b90ca4e3e8a5b69cc9139902ef74f21 --- /dev/null +++ b/model-00005-of-00027.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2df9c8a1119028f2199ac9c1ea82935115c36d96e0df0369d71a77bdffbf2ffe +size 4857105480 diff --git a/model-00006-of-00027.safetensors b/model-00006-of-00027.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..23cec9bedb996745db7d8ba2b7ccc1657990b857 --- /dev/null +++ b/model-00006-of-00027.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3300815504a2fb38a71d4bf6d553e22476272c6ed134288bdce3fbe8231c2e9d +size 4857105480 diff --git a/model-00007-of-00027.safetensors b/model-00007-of-00027.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..edae9af2e0d68ea5b2e0a5b27a75c3fc113f2b37 --- /dev/null +++ b/model-00007-of-00027.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:08ef97a2f57a9013dd7330eeab4f40d04df2516ad88d6a328cbf5a4541d440bf +size 4857105480 diff --git a/model-00008-of-00027.safetensors b/model-00008-of-00027.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..66dd1f3faa40339824c4b1fdfb80146f2ff2bf93 --- /dev/null +++ b/model-00008-of-00027.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3672e5445b64fd4305e533ee6ff27c99153eff29fca19d055629d6b7848935e3 +size 4857105480 diff --git a/model-00009-of-00027.safetensors b/model-00009-of-00027.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b9095b6ce66eff3435391e1e5446aab9bb45c992 --- /dev/null +++ b/model-00009-of-00027.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:87a08236b63f3c90b73aa9f61cd7fb2efb450e3f077a3c17150ae54e3e1456cb +size 4857105480 diff --git a/model-00010-of-00027.safetensors b/model-00010-of-00027.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d0933cee978c63a77853bdb6e829ce2f303a9251 --- /dev/null +++ b/model-00010-of-00027.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:02f1e631e741c0883317932f2a7a3cdb2e2da95a39d294f94140cc3f80055fd7 +size 4857105480 diff --git a/model-00011-of-00027.safetensors b/model-00011-of-00027.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c6f71e1047d3f14792d169d132bca5359177268c --- /dev/null +++ b/model-00011-of-00027.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3e799684c32ed639a43d15edee02b76b42bf6bed525f6ac925b25bd068a93c0a +size 4857105480 diff --git a/model-00012-of-00027.safetensors b/model-00012-of-00027.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..81c44ee88eb26e767e4fc1c7fcda68739bbeaaed --- /dev/null +++ b/model-00012-of-00027.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eb63571cad4d66390bc5a89d4d22be45ed35294b0866bd63e7bcc14615003e40 +size 4857105480 diff --git a/model-00013-of-00027.safetensors b/model-00013-of-00027.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..209eae4d6bebe36557c6477f353485faddc70579 --- /dev/null +++ b/model-00013-of-00027.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d1d32c378e04a84b9f264300e30d1329c10ee4916965a4f024ad57fda06cdad4 +size 4857105480 diff --git a/model-00014-of-00027.safetensors b/model-00014-of-00027.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2376a46c362681db266382a206914b5494f0015c --- /dev/null +++ b/model-00014-of-00027.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7dc362fd24b29339570efd55aeb758acd5e24709a830aa2b34a58fe8368ff49d +size 4857105480 diff --git a/model-00015-of-00027.safetensors b/model-00015-of-00027.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f2afb0e9568895c57e96ba92a25c1058583b8d49 --- /dev/null +++ b/model-00015-of-00027.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f86bc594da9d6ec7d8b1b1b875950ee5f292dcfde97b004087d23aae29abef07 +size 4857105480 diff --git a/model-00016-of-00027.safetensors b/model-00016-of-00027.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0316efec883b914c2c4e09732e3e403b274b8bce --- /dev/null +++ b/model-00016-of-00027.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4f49de4f60227a1c602131a2457c4b2e0cd19170645169a8bf76f3fa945a0044 +size 4857105480 diff --git a/model-00017-of-00027.safetensors b/model-00017-of-00027.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..53934a57cce57f4b5e55143261a31ee687a7bc2b --- /dev/null +++ b/model-00017-of-00027.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b0f463e8d7c9ad68509f3cf97ad66f5e5f734ec9b3c81041cc705379cfd4e302 +size 4857105480 diff --git a/model-00018-of-00027.safetensors b/model-00018-of-00027.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..43fa36f966b07f3699325ac5ac42bffc1ec83961 --- /dev/null +++ b/model-00018-of-00027.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:acd980f3c577d58129d794b426fce2a317540675c0132947b858a0199d6f662d +size 4857105480 diff --git a/model-00019-of-00027.safetensors b/model-00019-of-00027.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1da97ab6bb426f6a2bca4d822a3cc9d9a59a7dab --- /dev/null +++ b/model-00019-of-00027.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f40d2384c974fa419f48063a4fde485119b127f9ce9d9f43f5d4fa64508ce25d +size 4857105480 diff --git a/model-00020-of-00027.safetensors b/model-00020-of-00027.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6b20f2524d1ede9ce1eb55977107993f8f1bfd68 --- /dev/null +++ b/model-00020-of-00027.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:467e76b23dfaa0baff299440c4f8c2ea34ee2bb84baccc8af018a94f301cb246 +size 4857105480 diff --git a/model-00021-of-00027.safetensors b/model-00021-of-00027.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..01dbfad398b981ed5c00a34b0a2e21472a919190 --- /dev/null +++ b/model-00021-of-00027.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6cd617551ff3a0bc34457dcf376818232b3b3be03d3e5c4de5171856fe35d6c9 +size 4857105480 diff --git a/model-00022-of-00027.safetensors b/model-00022-of-00027.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5e529330ec56b840410a3f70801fbeed315e1ea9 --- /dev/null +++ b/model-00022-of-00027.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e71a5feaf7b028a7dc7c003b969478062e278e6b2e5d0164be144904ec103e03 +size 4857105480 diff --git a/model-00023-of-00027.safetensors b/model-00023-of-00027.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..718677a11e55187f90d14403949d7fe738193f9b --- /dev/null +++ b/model-00023-of-00027.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a555ba809ed634ec3cc465ae6dae8e934c0f6ce4d3fdd2b292ec16b2d20ecba1 +size 4857105480 diff --git a/model-00024-of-00027.safetensors b/model-00024-of-00027.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ec6dd4b2e7cd6fd37b8c55dbf8ee0e95076bd795 --- /dev/null +++ b/model-00024-of-00027.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fa0f30d00cd8fa8c758f6c85efa484cd7116813deb7a81c23591350ba6b84bdf +size 4857105480 diff --git a/model-00025-of-00027.safetensors b/model-00025-of-00027.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a1ff38507e546c60346f9d6717ee35263fba73f4 --- /dev/null +++ b/model-00025-of-00027.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b99045b5c0bb406714ac60560d54ea7b1bb0d5551f71d6daf48e53007dcd5d06 +size 4857105480 diff --git a/model-00026-of-00027.safetensors b/model-00026-of-00027.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..88bd6b2629c935a3b8308e6cad53348085d43f3b --- /dev/null +++ b/model-00026-of-00027.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd27faa36f9992a3131f1c9059e690d9c0a7736db58c5d3c26a9278e9d20a63c +size 4857105480 diff --git a/model-00027-of-00027.safetensors b/model-00027-of-00027.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c8bb50a2776b52aef8acbf41babe5df9ab357e90 --- /dev/null +++ b/model-00027-of-00027.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35fbef0cb2631804fc15da4a47e05422b0dcf63cda015e2103cf9a945319c071 +size 4484352704 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..7af830c08840a437e11d71e2f52b84d559057dd3 --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,730 @@ +{ + "metadata": { + "total_size": 130572369920 + }, + "weight_map": { + "lm_head.weight": "model-00027-of-00027.safetensors", + "model.embed_tokens.weight": "model-00001-of-00027.safetensors", + "model.layers.0.input_layernorm.weight": "model-00001-of-00027.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00001-of-00027.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00027.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00001-of-00027.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00027.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00027.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00027.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00027.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00027.safetensors", + "model.layers.1.input_layernorm.weight": "model-00001-of-00027.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00001-of-00027.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00027.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00001-of-00027.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00027.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00027.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00027.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00027.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00027.safetensors", + "model.layers.10.input_layernorm.weight": "model-00004-of-00027.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00004-of-00027.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00004-of-00027.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00004-of-00027.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00004-of-00027.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00004-of-00027.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00004-of-00027.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00004-of-00027.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00004-of-00027.safetensors", + "model.layers.11.input_layernorm.weight": "model-00005-of-00027.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00005-of-00027.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00004-of-00027.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00005-of-00027.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00005-of-00027.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00004-of-00027.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00004-of-00027.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00004-of-00027.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00004-of-00027.safetensors", + "model.layers.12.input_layernorm.weight": "model-00005-of-00027.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00005-of-00027.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00005-of-00027.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00005-of-00027.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00005-of-00027.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00005-of-00027.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00005-of-00027.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00005-of-00027.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00005-of-00027.safetensors", + "model.layers.13.input_layernorm.weight": "model-00005-of-00027.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00005-of-00027.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00005-of-00027.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00005-of-00027.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00005-of-00027.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00005-of-00027.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00005-of-00027.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00005-of-00027.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00005-of-00027.safetensors", + "model.layers.14.input_layernorm.weight": "model-00006-of-00027.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00006-of-00027.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00005-of-00027.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00006-of-00027.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00006-of-00027.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00005-of-00027.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00005-of-00027.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00005-of-00027.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00005-of-00027.safetensors", + "model.layers.15.input_layernorm.weight": "model-00006-of-00027.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00006-of-00027.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00006-of-00027.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00006-of-00027.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00006-of-00027.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00006-of-00027.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00006-of-00027.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00006-of-00027.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00006-of-00027.safetensors", + "model.layers.16.input_layernorm.weight": "model-00006-of-00027.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00006-of-00027.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00006-of-00027.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00006-of-00027.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00006-of-00027.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00006-of-00027.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00006-of-00027.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00006-of-00027.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00006-of-00027.safetensors", + "model.layers.17.input_layernorm.weight": "model-00007-of-00027.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00007-of-00027.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00006-of-00027.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00007-of-00027.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00007-of-00027.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00006-of-00027.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00006-of-00027.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00006-of-00027.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00006-of-00027.safetensors", + "model.layers.18.input_layernorm.weight": "model-00007-of-00027.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00007-of-00027.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00007-of-00027.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00007-of-00027.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00007-of-00027.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00007-of-00027.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00007-of-00027.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00007-of-00027.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00007-of-00027.safetensors", + "model.layers.19.input_layernorm.weight": "model-00007-of-00027.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00007-of-00027.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00007-of-00027.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00007-of-00027.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00007-of-00027.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00007-of-00027.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00007-of-00027.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00007-of-00027.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00007-of-00027.safetensors", + "model.layers.2.input_layernorm.weight": "model-00002-of-00027.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00002-of-00027.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00027.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00002-of-00027.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00002-of-00027.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00027.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00027.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00027.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00027.safetensors", + "model.layers.20.input_layernorm.weight": "model-00008-of-00027.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00008-of-00027.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00007-of-00027.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00008-of-00027.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00008-of-00027.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00007-of-00027.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00007-of-00027.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00007-of-00027.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00007-of-00027.safetensors", + "model.layers.21.input_layernorm.weight": "model-00008-of-00027.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00008-of-00027.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00008-of-00027.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00008-of-00027.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00008-of-00027.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00008-of-00027.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00008-of-00027.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00008-of-00027.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00008-of-00027.safetensors", + "model.layers.22.input_layernorm.weight": "model-00008-of-00027.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00008-of-00027.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00008-of-00027.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00008-of-00027.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00008-of-00027.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00008-of-00027.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00008-of-00027.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00008-of-00027.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00008-of-00027.safetensors", + "model.layers.23.input_layernorm.weight": "model-00009-of-00027.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00009-of-00027.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00008-of-00027.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00009-of-00027.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00009-of-00027.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00008-of-00027.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00008-of-00027.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00008-of-00027.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00008-of-00027.safetensors", + "model.layers.24.input_layernorm.weight": "model-00009-of-00027.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00009-of-00027.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00009-of-00027.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00009-of-00027.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00009-of-00027.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00009-of-00027.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00009-of-00027.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00009-of-00027.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00009-of-00027.safetensors", + "model.layers.25.input_layernorm.weight": "model-00009-of-00027.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00009-of-00027.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00009-of-00027.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00009-of-00027.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00009-of-00027.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00009-of-00027.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00009-of-00027.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00009-of-00027.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00009-of-00027.safetensors", + "model.layers.26.input_layernorm.weight": "model-00010-of-00027.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00010-of-00027.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00009-of-00027.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00010-of-00027.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00010-of-00027.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00009-of-00027.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00009-of-00027.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00009-of-00027.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00009-of-00027.safetensors", + "model.layers.27.input_layernorm.weight": "model-00010-of-00027.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00010-of-00027.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00010-of-00027.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00010-of-00027.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00010-of-00027.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00010-of-00027.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00010-of-00027.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00010-of-00027.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00010-of-00027.safetensors", + "model.layers.28.input_layernorm.weight": "model-00010-of-00027.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00010-of-00027.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00010-of-00027.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00010-of-00027.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00010-of-00027.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00010-of-00027.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00010-of-00027.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00010-of-00027.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00010-of-00027.safetensors", + "model.layers.29.input_layernorm.weight": "model-00011-of-00027.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00011-of-00027.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00010-of-00027.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00011-of-00027.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00011-of-00027.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00010-of-00027.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00010-of-00027.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00010-of-00027.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00010-of-00027.safetensors", + "model.layers.3.input_layernorm.weight": "model-00002-of-00027.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00002-of-00027.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00002-of-00027.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00002-of-00027.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00002-of-00027.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00002-of-00027.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00002-of-00027.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00002-of-00027.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00002-of-00027.safetensors", + "model.layers.30.input_layernorm.weight": "model-00011-of-00027.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00011-of-00027.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00011-of-00027.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00011-of-00027.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00011-of-00027.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00011-of-00027.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00011-of-00027.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00011-of-00027.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00011-of-00027.safetensors", + "model.layers.31.input_layernorm.weight": "model-00011-of-00027.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00011-of-00027.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00011-of-00027.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00011-of-00027.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00011-of-00027.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00011-of-00027.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00011-of-00027.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00011-of-00027.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00011-of-00027.safetensors", + "model.layers.32.input_layernorm.weight": "model-00012-of-00027.safetensors", + "model.layers.32.mlp.down_proj.weight": "model-00012-of-00027.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model-00011-of-00027.safetensors", + "model.layers.32.mlp.up_proj.weight": "model-00012-of-00027.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00012-of-00027.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00011-of-00027.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00011-of-00027.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00011-of-00027.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00011-of-00027.safetensors", + "model.layers.33.input_layernorm.weight": "model-00012-of-00027.safetensors", + "model.layers.33.mlp.down_proj.weight": "model-00012-of-00027.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model-00012-of-00027.safetensors", + "model.layers.33.mlp.up_proj.weight": "model-00012-of-00027.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00012-of-00027.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00012-of-00027.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00012-of-00027.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00012-of-00027.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00012-of-00027.safetensors", + "model.layers.34.input_layernorm.weight": "model-00012-of-00027.safetensors", + "model.layers.34.mlp.down_proj.weight": "model-00012-of-00027.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model-00012-of-00027.safetensors", + "model.layers.34.mlp.up_proj.weight": "model-00012-of-00027.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00012-of-00027.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00012-of-00027.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00012-of-00027.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00012-of-00027.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00012-of-00027.safetensors", + "model.layers.35.input_layernorm.weight": "model-00013-of-00027.safetensors", + "model.layers.35.mlp.down_proj.weight": "model-00013-of-00027.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model-00012-of-00027.safetensors", + "model.layers.35.mlp.up_proj.weight": "model-00013-of-00027.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00013-of-00027.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00012-of-00027.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00012-of-00027.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00012-of-00027.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00012-of-00027.safetensors", + "model.layers.36.input_layernorm.weight": "model-00013-of-00027.safetensors", + "model.layers.36.mlp.down_proj.weight": "model-00013-of-00027.safetensors", + "model.layers.36.mlp.gate_proj.weight": "model-00013-of-00027.safetensors", + "model.layers.36.mlp.up_proj.weight": "model-00013-of-00027.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00013-of-00027.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model-00013-of-00027.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model-00013-of-00027.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model-00013-of-00027.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model-00013-of-00027.safetensors", + "model.layers.37.input_layernorm.weight": "model-00013-of-00027.safetensors", + "model.layers.37.mlp.down_proj.weight": "model-00013-of-00027.safetensors", + "model.layers.37.mlp.gate_proj.weight": "model-00013-of-00027.safetensors", + "model.layers.37.mlp.up_proj.weight": "model-00013-of-00027.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00013-of-00027.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model-00013-of-00027.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model-00013-of-00027.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model-00013-of-00027.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model-00013-of-00027.safetensors", + "model.layers.38.input_layernorm.weight": "model-00014-of-00027.safetensors", + "model.layers.38.mlp.down_proj.weight": "model-00014-of-00027.safetensors", + "model.layers.38.mlp.gate_proj.weight": "model-00013-of-00027.safetensors", + "model.layers.38.mlp.up_proj.weight": "model-00014-of-00027.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00014-of-00027.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model-00013-of-00027.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model-00013-of-00027.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model-00013-of-00027.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model-00013-of-00027.safetensors", + "model.layers.39.input_layernorm.weight": "model-00014-of-00027.safetensors", + "model.layers.39.mlp.down_proj.weight": "model-00014-of-00027.safetensors", + "model.layers.39.mlp.gate_proj.weight": "model-00014-of-00027.safetensors", + "model.layers.39.mlp.up_proj.weight": "model-00014-of-00027.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00014-of-00027.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model-00014-of-00027.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model-00014-of-00027.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model-00014-of-00027.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model-00014-of-00027.safetensors", + "model.layers.4.input_layernorm.weight": "model-00002-of-00027.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00002-of-00027.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00002-of-00027.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00002-of-00027.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00002-of-00027.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00002-of-00027.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00002-of-00027.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00002-of-00027.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00002-of-00027.safetensors", + "model.layers.40.input_layernorm.weight": "model-00014-of-00027.safetensors", + "model.layers.40.mlp.down_proj.weight": "model-00014-of-00027.safetensors", + "model.layers.40.mlp.gate_proj.weight": "model-00014-of-00027.safetensors", + "model.layers.40.mlp.up_proj.weight": "model-00014-of-00027.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00014-of-00027.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model-00014-of-00027.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model-00014-of-00027.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model-00014-of-00027.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model-00014-of-00027.safetensors", + "model.layers.41.input_layernorm.weight": "model-00015-of-00027.safetensors", + "model.layers.41.mlp.down_proj.weight": "model-00015-of-00027.safetensors", + "model.layers.41.mlp.gate_proj.weight": "model-00014-of-00027.safetensors", + "model.layers.41.mlp.up_proj.weight": "model-00015-of-00027.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00015-of-00027.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model-00014-of-00027.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model-00014-of-00027.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model-00014-of-00027.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model-00014-of-00027.safetensors", + "model.layers.42.input_layernorm.weight": "model-00015-of-00027.safetensors", + "model.layers.42.mlp.down_proj.weight": "model-00015-of-00027.safetensors", + "model.layers.42.mlp.gate_proj.weight": "model-00015-of-00027.safetensors", + "model.layers.42.mlp.up_proj.weight": "model-00015-of-00027.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00015-of-00027.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model-00015-of-00027.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model-00015-of-00027.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model-00015-of-00027.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model-00015-of-00027.safetensors", + "model.layers.43.input_layernorm.weight": "model-00015-of-00027.safetensors", + "model.layers.43.mlp.down_proj.weight": "model-00015-of-00027.safetensors", + "model.layers.43.mlp.gate_proj.weight": "model-00015-of-00027.safetensors", + "model.layers.43.mlp.up_proj.weight": "model-00015-of-00027.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00015-of-00027.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model-00015-of-00027.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model-00015-of-00027.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model-00015-of-00027.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model-00015-of-00027.safetensors", + "model.layers.44.input_layernorm.weight": "model-00016-of-00027.safetensors", + "model.layers.44.mlp.down_proj.weight": "model-00016-of-00027.safetensors", + "model.layers.44.mlp.gate_proj.weight": "model-00015-of-00027.safetensors", + "model.layers.44.mlp.up_proj.weight": "model-00016-of-00027.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00016-of-00027.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model-00015-of-00027.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model-00015-of-00027.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model-00015-of-00027.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model-00015-of-00027.safetensors", + "model.layers.45.input_layernorm.weight": "model-00016-of-00027.safetensors", + "model.layers.45.mlp.down_proj.weight": "model-00016-of-00027.safetensors", + "model.layers.45.mlp.gate_proj.weight": "model-00016-of-00027.safetensors", + "model.layers.45.mlp.up_proj.weight": "model-00016-of-00027.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00016-of-00027.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model-00016-of-00027.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model-00016-of-00027.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model-00016-of-00027.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model-00016-of-00027.safetensors", + "model.layers.46.input_layernorm.weight": "model-00016-of-00027.safetensors", + "model.layers.46.mlp.down_proj.weight": "model-00016-of-00027.safetensors", + "model.layers.46.mlp.gate_proj.weight": "model-00016-of-00027.safetensors", + "model.layers.46.mlp.up_proj.weight": "model-00016-of-00027.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00016-of-00027.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model-00016-of-00027.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model-00016-of-00027.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model-00016-of-00027.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model-00016-of-00027.safetensors", + "model.layers.47.input_layernorm.weight": "model-00017-of-00027.safetensors", + "model.layers.47.mlp.down_proj.weight": "model-00017-of-00027.safetensors", + "model.layers.47.mlp.gate_proj.weight": "model-00016-of-00027.safetensors", + "model.layers.47.mlp.up_proj.weight": "model-00017-of-00027.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00017-of-00027.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model-00016-of-00027.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model-00016-of-00027.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model-00016-of-00027.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model-00016-of-00027.safetensors", + "model.layers.48.input_layernorm.weight": "model-00017-of-00027.safetensors", + "model.layers.48.mlp.down_proj.weight": "model-00017-of-00027.safetensors", + "model.layers.48.mlp.gate_proj.weight": "model-00017-of-00027.safetensors", + "model.layers.48.mlp.up_proj.weight": "model-00017-of-00027.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00017-of-00027.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model-00017-of-00027.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model-00017-of-00027.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model-00017-of-00027.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model-00017-of-00027.safetensors", + "model.layers.49.input_layernorm.weight": "model-00017-of-00027.safetensors", + "model.layers.49.mlp.down_proj.weight": "model-00017-of-00027.safetensors", + "model.layers.49.mlp.gate_proj.weight": "model-00017-of-00027.safetensors", + "model.layers.49.mlp.up_proj.weight": "model-00017-of-00027.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00017-of-00027.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model-00017-of-00027.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model-00017-of-00027.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model-00017-of-00027.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model-00017-of-00027.safetensors", + "model.layers.5.input_layernorm.weight": "model-00003-of-00027.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00003-of-00027.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00002-of-00027.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00003-of-00027.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00003-of-00027.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00002-of-00027.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00002-of-00027.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00002-of-00027.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00002-of-00027.safetensors", + "model.layers.50.input_layernorm.weight": "model-00018-of-00027.safetensors", + "model.layers.50.mlp.down_proj.weight": "model-00018-of-00027.safetensors", + "model.layers.50.mlp.gate_proj.weight": "model-00017-of-00027.safetensors", + "model.layers.50.mlp.up_proj.weight": "model-00018-of-00027.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00018-of-00027.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model-00017-of-00027.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model-00017-of-00027.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model-00017-of-00027.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model-00017-of-00027.safetensors", + "model.layers.51.input_layernorm.weight": "model-00018-of-00027.safetensors", + "model.layers.51.mlp.down_proj.weight": "model-00018-of-00027.safetensors", + "model.layers.51.mlp.gate_proj.weight": "model-00018-of-00027.safetensors", + "model.layers.51.mlp.up_proj.weight": "model-00018-of-00027.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00018-of-00027.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model-00018-of-00027.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model-00018-of-00027.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model-00018-of-00027.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model-00018-of-00027.safetensors", + "model.layers.52.input_layernorm.weight": "model-00018-of-00027.safetensors", + "model.layers.52.mlp.down_proj.weight": "model-00018-of-00027.safetensors", + "model.layers.52.mlp.gate_proj.weight": "model-00018-of-00027.safetensors", + "model.layers.52.mlp.up_proj.weight": "model-00018-of-00027.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model-00018-of-00027.safetensors", + "model.layers.52.self_attn.k_proj.weight": "model-00018-of-00027.safetensors", + "model.layers.52.self_attn.o_proj.weight": "model-00018-of-00027.safetensors", + "model.layers.52.self_attn.q_proj.weight": "model-00018-of-00027.safetensors", + "model.layers.52.self_attn.v_proj.weight": "model-00018-of-00027.safetensors", + "model.layers.53.input_layernorm.weight": "model-00019-of-00027.safetensors", + "model.layers.53.mlp.down_proj.weight": "model-00019-of-00027.safetensors", + "model.layers.53.mlp.gate_proj.weight": "model-00018-of-00027.safetensors", + "model.layers.53.mlp.up_proj.weight": "model-00019-of-00027.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model-00019-of-00027.safetensors", + "model.layers.53.self_attn.k_proj.weight": "model-00018-of-00027.safetensors", + "model.layers.53.self_attn.o_proj.weight": "model-00018-of-00027.safetensors", + "model.layers.53.self_attn.q_proj.weight": "model-00018-of-00027.safetensors", + "model.layers.53.self_attn.v_proj.weight": "model-00018-of-00027.safetensors", + "model.layers.54.input_layernorm.weight": "model-00019-of-00027.safetensors", + "model.layers.54.mlp.down_proj.weight": "model-00019-of-00027.safetensors", + "model.layers.54.mlp.gate_proj.weight": "model-00019-of-00027.safetensors", + "model.layers.54.mlp.up_proj.weight": "model-00019-of-00027.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model-00019-of-00027.safetensors", + "model.layers.54.self_attn.k_proj.weight": "model-00019-of-00027.safetensors", + "model.layers.54.self_attn.o_proj.weight": "model-00019-of-00027.safetensors", + "model.layers.54.self_attn.q_proj.weight": "model-00019-of-00027.safetensors", + "model.layers.54.self_attn.v_proj.weight": "model-00019-of-00027.safetensors", + "model.layers.55.input_layernorm.weight": "model-00019-of-00027.safetensors", + "model.layers.55.mlp.down_proj.weight": "model-00019-of-00027.safetensors", + "model.layers.55.mlp.gate_proj.weight": "model-00019-of-00027.safetensors", + "model.layers.55.mlp.up_proj.weight": "model-00019-of-00027.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model-00019-of-00027.safetensors", + "model.layers.55.self_attn.k_proj.weight": "model-00019-of-00027.safetensors", + "model.layers.55.self_attn.o_proj.weight": "model-00019-of-00027.safetensors", + "model.layers.55.self_attn.q_proj.weight": "model-00019-of-00027.safetensors", + "model.layers.55.self_attn.v_proj.weight": "model-00019-of-00027.safetensors", + "model.layers.56.input_layernorm.weight": "model-00020-of-00027.safetensors", + "model.layers.56.mlp.down_proj.weight": "model-00020-of-00027.safetensors", + "model.layers.56.mlp.gate_proj.weight": "model-00019-of-00027.safetensors", + "model.layers.56.mlp.up_proj.weight": "model-00020-of-00027.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model-00020-of-00027.safetensors", + "model.layers.56.self_attn.k_proj.weight": "model-00019-of-00027.safetensors", + "model.layers.56.self_attn.o_proj.weight": "model-00019-of-00027.safetensors", + "model.layers.56.self_attn.q_proj.weight": "model-00019-of-00027.safetensors", + "model.layers.56.self_attn.v_proj.weight": "model-00019-of-00027.safetensors", + "model.layers.57.input_layernorm.weight": "model-00020-of-00027.safetensors", + "model.layers.57.mlp.down_proj.weight": "model-00020-of-00027.safetensors", + "model.layers.57.mlp.gate_proj.weight": "model-00020-of-00027.safetensors", + "model.layers.57.mlp.up_proj.weight": "model-00020-of-00027.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model-00020-of-00027.safetensors", + "model.layers.57.self_attn.k_proj.weight": "model-00020-of-00027.safetensors", + "model.layers.57.self_attn.o_proj.weight": "model-00020-of-00027.safetensors", + "model.layers.57.self_attn.q_proj.weight": "model-00020-of-00027.safetensors", + "model.layers.57.self_attn.v_proj.weight": "model-00020-of-00027.safetensors", + "model.layers.58.input_layernorm.weight": "model-00020-of-00027.safetensors", + "model.layers.58.mlp.down_proj.weight": "model-00020-of-00027.safetensors", + "model.layers.58.mlp.gate_proj.weight": "model-00020-of-00027.safetensors", + "model.layers.58.mlp.up_proj.weight": "model-00020-of-00027.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model-00020-of-00027.safetensors", + "model.layers.58.self_attn.k_proj.weight": "model-00020-of-00027.safetensors", + "model.layers.58.self_attn.o_proj.weight": "model-00020-of-00027.safetensors", + "model.layers.58.self_attn.q_proj.weight": "model-00020-of-00027.safetensors", + "model.layers.58.self_attn.v_proj.weight": "model-00020-of-00027.safetensors", + "model.layers.59.input_layernorm.weight": "model-00021-of-00027.safetensors", + "model.layers.59.mlp.down_proj.weight": "model-00021-of-00027.safetensors", + "model.layers.59.mlp.gate_proj.weight": "model-00020-of-00027.safetensors", + "model.layers.59.mlp.up_proj.weight": "model-00021-of-00027.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model-00021-of-00027.safetensors", + "model.layers.59.self_attn.k_proj.weight": "model-00020-of-00027.safetensors", + "model.layers.59.self_attn.o_proj.weight": "model-00020-of-00027.safetensors", + "model.layers.59.self_attn.q_proj.weight": "model-00020-of-00027.safetensors", + "model.layers.59.self_attn.v_proj.weight": "model-00020-of-00027.safetensors", + "model.layers.6.input_layernorm.weight": "model-00003-of-00027.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00003-of-00027.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00003-of-00027.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00003-of-00027.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00003-of-00027.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00003-of-00027.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00003-of-00027.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00003-of-00027.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00003-of-00027.safetensors", + "model.layers.60.input_layernorm.weight": "model-00021-of-00027.safetensors", + "model.layers.60.mlp.down_proj.weight": "model-00021-of-00027.safetensors", + "model.layers.60.mlp.gate_proj.weight": "model-00021-of-00027.safetensors", + "model.layers.60.mlp.up_proj.weight": "model-00021-of-00027.safetensors", + "model.layers.60.post_attention_layernorm.weight": "model-00021-of-00027.safetensors", + "model.layers.60.self_attn.k_proj.weight": "model-00021-of-00027.safetensors", + "model.layers.60.self_attn.o_proj.weight": "model-00021-of-00027.safetensors", + "model.layers.60.self_attn.q_proj.weight": "model-00021-of-00027.safetensors", + "model.layers.60.self_attn.v_proj.weight": "model-00021-of-00027.safetensors", + "model.layers.61.input_layernorm.weight": "model-00021-of-00027.safetensors", + "model.layers.61.mlp.down_proj.weight": "model-00021-of-00027.safetensors", + "model.layers.61.mlp.gate_proj.weight": "model-00021-of-00027.safetensors", + "model.layers.61.mlp.up_proj.weight": "model-00021-of-00027.safetensors", + "model.layers.61.post_attention_layernorm.weight": "model-00021-of-00027.safetensors", + "model.layers.61.self_attn.k_proj.weight": "model-00021-of-00027.safetensors", + "model.layers.61.self_attn.o_proj.weight": "model-00021-of-00027.safetensors", + "model.layers.61.self_attn.q_proj.weight": "model-00021-of-00027.safetensors", + "model.layers.61.self_attn.v_proj.weight": "model-00021-of-00027.safetensors", + "model.layers.62.input_layernorm.weight": "model-00022-of-00027.safetensors", + "model.layers.62.mlp.down_proj.weight": "model-00022-of-00027.safetensors", + "model.layers.62.mlp.gate_proj.weight": "model-00021-of-00027.safetensors", + "model.layers.62.mlp.up_proj.weight": "model-00022-of-00027.safetensors", + "model.layers.62.post_attention_layernorm.weight": "model-00022-of-00027.safetensors", + "model.layers.62.self_attn.k_proj.weight": "model-00021-of-00027.safetensors", + "model.layers.62.self_attn.o_proj.weight": "model-00021-of-00027.safetensors", + "model.layers.62.self_attn.q_proj.weight": "model-00021-of-00027.safetensors", + "model.layers.62.self_attn.v_proj.weight": "model-00021-of-00027.safetensors", + "model.layers.63.input_layernorm.weight": "model-00022-of-00027.safetensors", + "model.layers.63.mlp.down_proj.weight": "model-00022-of-00027.safetensors", + "model.layers.63.mlp.gate_proj.weight": "model-00022-of-00027.safetensors", + "model.layers.63.mlp.up_proj.weight": "model-00022-of-00027.safetensors", + "model.layers.63.post_attention_layernorm.weight": "model-00022-of-00027.safetensors", + "model.layers.63.self_attn.k_proj.weight": "model-00022-of-00027.safetensors", + "model.layers.63.self_attn.o_proj.weight": "model-00022-of-00027.safetensors", + "model.layers.63.self_attn.q_proj.weight": "model-00022-of-00027.safetensors", + "model.layers.63.self_attn.v_proj.weight": "model-00022-of-00027.safetensors", + "model.layers.64.input_layernorm.weight": "model-00022-of-00027.safetensors", + "model.layers.64.mlp.down_proj.weight": "model-00022-of-00027.safetensors", + "model.layers.64.mlp.gate_proj.weight": "model-00022-of-00027.safetensors", + "model.layers.64.mlp.up_proj.weight": "model-00022-of-00027.safetensors", + "model.layers.64.post_attention_layernorm.weight": "model-00022-of-00027.safetensors", + "model.layers.64.self_attn.k_proj.weight": "model-00022-of-00027.safetensors", + "model.layers.64.self_attn.o_proj.weight": "model-00022-of-00027.safetensors", + "model.layers.64.self_attn.q_proj.weight": "model-00022-of-00027.safetensors", + "model.layers.64.self_attn.v_proj.weight": "model-00022-of-00027.safetensors", + "model.layers.65.input_layernorm.weight": "model-00023-of-00027.safetensors", + "model.layers.65.mlp.down_proj.weight": "model-00023-of-00027.safetensors", + "model.layers.65.mlp.gate_proj.weight": "model-00022-of-00027.safetensors", + "model.layers.65.mlp.up_proj.weight": "model-00023-of-00027.safetensors", + "model.layers.65.post_attention_layernorm.weight": "model-00023-of-00027.safetensors", + "model.layers.65.self_attn.k_proj.weight": "model-00022-of-00027.safetensors", + "model.layers.65.self_attn.o_proj.weight": "model-00022-of-00027.safetensors", + "model.layers.65.self_attn.q_proj.weight": "model-00022-of-00027.safetensors", + "model.layers.65.self_attn.v_proj.weight": "model-00022-of-00027.safetensors", + "model.layers.66.input_layernorm.weight": "model-00023-of-00027.safetensors", + "model.layers.66.mlp.down_proj.weight": "model-00023-of-00027.safetensors", + "model.layers.66.mlp.gate_proj.weight": "model-00023-of-00027.safetensors", + "model.layers.66.mlp.up_proj.weight": "model-00023-of-00027.safetensors", + "model.layers.66.post_attention_layernorm.weight": "model-00023-of-00027.safetensors", + "model.layers.66.self_attn.k_proj.weight": "model-00023-of-00027.safetensors", + "model.layers.66.self_attn.o_proj.weight": "model-00023-of-00027.safetensors", + "model.layers.66.self_attn.q_proj.weight": "model-00023-of-00027.safetensors", + "model.layers.66.self_attn.v_proj.weight": "model-00023-of-00027.safetensors", + "model.layers.67.input_layernorm.weight": "model-00023-of-00027.safetensors", + "model.layers.67.mlp.down_proj.weight": "model-00023-of-00027.safetensors", + "model.layers.67.mlp.gate_proj.weight": "model-00023-of-00027.safetensors", + "model.layers.67.mlp.up_proj.weight": "model-00023-of-00027.safetensors", + "model.layers.67.post_attention_layernorm.weight": "model-00023-of-00027.safetensors", + "model.layers.67.self_attn.k_proj.weight": "model-00023-of-00027.safetensors", + "model.layers.67.self_attn.o_proj.weight": "model-00023-of-00027.safetensors", + "model.layers.67.self_attn.q_proj.weight": "model-00023-of-00027.safetensors", + "model.layers.67.self_attn.v_proj.weight": "model-00023-of-00027.safetensors", + "model.layers.68.input_layernorm.weight": "model-00024-of-00027.safetensors", + "model.layers.68.mlp.down_proj.weight": "model-00024-of-00027.safetensors", + "model.layers.68.mlp.gate_proj.weight": "model-00023-of-00027.safetensors", + "model.layers.68.mlp.up_proj.weight": "model-00024-of-00027.safetensors", + "model.layers.68.post_attention_layernorm.weight": "model-00024-of-00027.safetensors", + "model.layers.68.self_attn.k_proj.weight": "model-00023-of-00027.safetensors", + "model.layers.68.self_attn.o_proj.weight": "model-00023-of-00027.safetensors", + "model.layers.68.self_attn.q_proj.weight": "model-00023-of-00027.safetensors", + "model.layers.68.self_attn.v_proj.weight": "model-00023-of-00027.safetensors", + "model.layers.69.input_layernorm.weight": "model-00024-of-00027.safetensors", + "model.layers.69.mlp.down_proj.weight": "model-00024-of-00027.safetensors", + "model.layers.69.mlp.gate_proj.weight": "model-00024-of-00027.safetensors", + "model.layers.69.mlp.up_proj.weight": "model-00024-of-00027.safetensors", + "model.layers.69.post_attention_layernorm.weight": "model-00024-of-00027.safetensors", + "model.layers.69.self_attn.k_proj.weight": "model-00024-of-00027.safetensors", + "model.layers.69.self_attn.o_proj.weight": "model-00024-of-00027.safetensors", + "model.layers.69.self_attn.q_proj.weight": "model-00024-of-00027.safetensors", + "model.layers.69.self_attn.v_proj.weight": "model-00024-of-00027.safetensors", + "model.layers.7.input_layernorm.weight": "model-00003-of-00027.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00003-of-00027.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00003-of-00027.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00003-of-00027.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00003-of-00027.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00003-of-00027.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00003-of-00027.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00003-of-00027.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00003-of-00027.safetensors", + "model.layers.70.input_layernorm.weight": "model-00024-of-00027.safetensors", + "model.layers.70.mlp.down_proj.weight": "model-00024-of-00027.safetensors", + "model.layers.70.mlp.gate_proj.weight": "model-00024-of-00027.safetensors", + "model.layers.70.mlp.up_proj.weight": "model-00024-of-00027.safetensors", + "model.layers.70.post_attention_layernorm.weight": "model-00024-of-00027.safetensors", + "model.layers.70.self_attn.k_proj.weight": "model-00024-of-00027.safetensors", + "model.layers.70.self_attn.o_proj.weight": "model-00024-of-00027.safetensors", + "model.layers.70.self_attn.q_proj.weight": "model-00024-of-00027.safetensors", + "model.layers.70.self_attn.v_proj.weight": "model-00024-of-00027.safetensors", + "model.layers.71.input_layernorm.weight": "model-00025-of-00027.safetensors", + "model.layers.71.mlp.down_proj.weight": "model-00025-of-00027.safetensors", + "model.layers.71.mlp.gate_proj.weight": "model-00024-of-00027.safetensors", + "model.layers.71.mlp.up_proj.weight": "model-00025-of-00027.safetensors", + "model.layers.71.post_attention_layernorm.weight": "model-00025-of-00027.safetensors", + "model.layers.71.self_attn.k_proj.weight": "model-00024-of-00027.safetensors", + "model.layers.71.self_attn.o_proj.weight": "model-00024-of-00027.safetensors", + "model.layers.71.self_attn.q_proj.weight": "model-00024-of-00027.safetensors", + "model.layers.71.self_attn.v_proj.weight": "model-00024-of-00027.safetensors", + "model.layers.72.input_layernorm.weight": "model-00025-of-00027.safetensors", + "model.layers.72.mlp.down_proj.weight": "model-00025-of-00027.safetensors", + "model.layers.72.mlp.gate_proj.weight": "model-00025-of-00027.safetensors", + "model.layers.72.mlp.up_proj.weight": "model-00025-of-00027.safetensors", + "model.layers.72.post_attention_layernorm.weight": "model-00025-of-00027.safetensors", + "model.layers.72.self_attn.k_proj.weight": "model-00025-of-00027.safetensors", + "model.layers.72.self_attn.o_proj.weight": "model-00025-of-00027.safetensors", + "model.layers.72.self_attn.q_proj.weight": "model-00025-of-00027.safetensors", + "model.layers.72.self_attn.v_proj.weight": "model-00025-of-00027.safetensors", + "model.layers.73.input_layernorm.weight": "model-00025-of-00027.safetensors", + "model.layers.73.mlp.down_proj.weight": "model-00025-of-00027.safetensors", + "model.layers.73.mlp.gate_proj.weight": "model-00025-of-00027.safetensors", + "model.layers.73.mlp.up_proj.weight": "model-00025-of-00027.safetensors", + "model.layers.73.post_attention_layernorm.weight": "model-00025-of-00027.safetensors", + "model.layers.73.self_attn.k_proj.weight": "model-00025-of-00027.safetensors", + "model.layers.73.self_attn.o_proj.weight": "model-00025-of-00027.safetensors", + "model.layers.73.self_attn.q_proj.weight": "model-00025-of-00027.safetensors", + "model.layers.73.self_attn.v_proj.weight": "model-00025-of-00027.safetensors", + "model.layers.74.input_layernorm.weight": "model-00026-of-00027.safetensors", + "model.layers.74.mlp.down_proj.weight": "model-00026-of-00027.safetensors", + "model.layers.74.mlp.gate_proj.weight": "model-00025-of-00027.safetensors", + "model.layers.74.mlp.up_proj.weight": "model-00026-of-00027.safetensors", + "model.layers.74.post_attention_layernorm.weight": "model-00026-of-00027.safetensors", + "model.layers.74.self_attn.k_proj.weight": "model-00025-of-00027.safetensors", + "model.layers.74.self_attn.o_proj.weight": "model-00025-of-00027.safetensors", + "model.layers.74.self_attn.q_proj.weight": "model-00025-of-00027.safetensors", + "model.layers.74.self_attn.v_proj.weight": "model-00025-of-00027.safetensors", + "model.layers.75.input_layernorm.weight": "model-00026-of-00027.safetensors", + "model.layers.75.mlp.down_proj.weight": "model-00026-of-00027.safetensors", + "model.layers.75.mlp.gate_proj.weight": "model-00026-of-00027.safetensors", + "model.layers.75.mlp.up_proj.weight": "model-00026-of-00027.safetensors", + "model.layers.75.post_attention_layernorm.weight": "model-00026-of-00027.safetensors", + "model.layers.75.self_attn.k_proj.weight": "model-00026-of-00027.safetensors", + "model.layers.75.self_attn.o_proj.weight": "model-00026-of-00027.safetensors", + "model.layers.75.self_attn.q_proj.weight": "model-00026-of-00027.safetensors", + "model.layers.75.self_attn.v_proj.weight": "model-00026-of-00027.safetensors", + "model.layers.76.input_layernorm.weight": "model-00026-of-00027.safetensors", + "model.layers.76.mlp.down_proj.weight": "model-00026-of-00027.safetensors", + "model.layers.76.mlp.gate_proj.weight": "model-00026-of-00027.safetensors", + "model.layers.76.mlp.up_proj.weight": "model-00026-of-00027.safetensors", + "model.layers.76.post_attention_layernorm.weight": "model-00026-of-00027.safetensors", + "model.layers.76.self_attn.k_proj.weight": "model-00026-of-00027.safetensors", + "model.layers.76.self_attn.o_proj.weight": "model-00026-of-00027.safetensors", + "model.layers.76.self_attn.q_proj.weight": "model-00026-of-00027.safetensors", + "model.layers.76.self_attn.v_proj.weight": "model-00026-of-00027.safetensors", + "model.layers.77.input_layernorm.weight": "model-00027-of-00027.safetensors", + "model.layers.77.mlp.down_proj.weight": "model-00027-of-00027.safetensors", + "model.layers.77.mlp.gate_proj.weight": "model-00026-of-00027.safetensors", + "model.layers.77.mlp.up_proj.weight": "model-00027-of-00027.safetensors", + "model.layers.77.post_attention_layernorm.weight": "model-00027-of-00027.safetensors", + "model.layers.77.self_attn.k_proj.weight": "model-00026-of-00027.safetensors", + "model.layers.77.self_attn.o_proj.weight": "model-00026-of-00027.safetensors", + "model.layers.77.self_attn.q_proj.weight": "model-00026-of-00027.safetensors", + "model.layers.77.self_attn.v_proj.weight": "model-00026-of-00027.safetensors", + "model.layers.78.input_layernorm.weight": "model-00027-of-00027.safetensors", + "model.layers.78.mlp.down_proj.weight": "model-00027-of-00027.safetensors", + "model.layers.78.mlp.gate_proj.weight": "model-00027-of-00027.safetensors", + "model.layers.78.mlp.up_proj.weight": "model-00027-of-00027.safetensors", + "model.layers.78.post_attention_layernorm.weight": "model-00027-of-00027.safetensors", + "model.layers.78.self_attn.k_proj.weight": "model-00027-of-00027.safetensors", + "model.layers.78.self_attn.o_proj.weight": "model-00027-of-00027.safetensors", + "model.layers.78.self_attn.q_proj.weight": "model-00027-of-00027.safetensors", + "model.layers.78.self_attn.v_proj.weight": "model-00027-of-00027.safetensors", + "model.layers.79.input_layernorm.weight": "model-00027-of-00027.safetensors", + "model.layers.79.mlp.down_proj.weight": "model-00027-of-00027.safetensors", + "model.layers.79.mlp.gate_proj.weight": "model-00027-of-00027.safetensors", + "model.layers.79.mlp.up_proj.weight": "model-00027-of-00027.safetensors", + "model.layers.79.post_attention_layernorm.weight": "model-00027-of-00027.safetensors", + "model.layers.79.self_attn.k_proj.weight": "model-00027-of-00027.safetensors", + "model.layers.79.self_attn.o_proj.weight": "model-00027-of-00027.safetensors", + "model.layers.79.self_attn.q_proj.weight": "model-00027-of-00027.safetensors", + "model.layers.79.self_attn.v_proj.weight": "model-00027-of-00027.safetensors", + "model.layers.8.input_layernorm.weight": "model-00004-of-00027.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00004-of-00027.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00003-of-00027.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00004-of-00027.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00004-of-00027.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00003-of-00027.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00003-of-00027.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00003-of-00027.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00003-of-00027.safetensors", + "model.layers.9.input_layernorm.weight": "model-00004-of-00027.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00004-of-00027.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00004-of-00027.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00004-of-00027.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00004-of-00027.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00004-of-00027.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00004-of-00027.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00004-of-00027.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00004-of-00027.safetensors", + "model.norm.weight": "model-00027-of-00027.safetensors" + } +} \ No newline at end of file diff --git a/models--LLM360--K2/blobs/0e5448b8c9a7a705501c5103af02f076036414b4cf54083c9ad17048d97221b5 b/models--LLM360--K2/blobs/0e5448b8c9a7a705501c5103af02f076036414b4cf54083c9ad17048d97221b5 new file mode 100644 index 0000000000000000000000000000000000000000..4334414923310eb5f2c2a36e9aded6c0a746f067 --- /dev/null +++ b/models--LLM360--K2/blobs/0e5448b8c9a7a705501c5103af02f076036414b4cf54083c9ad17048d97221b5 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e5448b8c9a7a705501c5103af02f076036414b4cf54083c9ad17048d97221b5 +size 4857111932 diff --git a/models--LLM360--K2/blobs/0ef2f96b2f9b4943f45dc1ba39f0594eb14cf128c5ff5f41db34ad07cb556e7f b/models--LLM360--K2/blobs/0ef2f96b2f9b4943f45dc1ba39f0594eb14cf128c5ff5f41db34ad07cb556e7f new file mode 100644 index 0000000000000000000000000000000000000000..88d73ac21e498fb2dd3444269a572e0b06803f71 --- /dev/null +++ b/models--LLM360--K2/blobs/0ef2f96b2f9b4943f45dc1ba39f0594eb14cf128c5ff5f41db34ad07cb556e7f @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0ef2f96b2f9b4943f45dc1ba39f0594eb14cf128c5ff5f41db34ad07cb556e7f +size 4857111932 diff --git a/models--LLM360--K2/blobs/197922ce7f7807ba59bfdd42fbbe9c67e2ff8370ffda566614cdf74463363f2b b/models--LLM360--K2/blobs/197922ce7f7807ba59bfdd42fbbe9c67e2ff8370ffda566614cdf74463363f2b new file mode 100644 index 0000000000000000000000000000000000000000..6c8c03014934a9b3c5489dcc704790d9c8fc1dff --- /dev/null +++ b/models--LLM360--K2/blobs/197922ce7f7807ba59bfdd42fbbe9c67e2ff8370ffda566614cdf74463363f2b @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:197922ce7f7807ba59bfdd42fbbe9c67e2ff8370ffda566614cdf74463363f2b +size 4857111932 diff --git a/models--LLM360--K2/blobs/1991b47bdfe3cb7e05d4a6f12d1bfb38a22fedf2da68d22269e675978940b57f b/models--LLM360--K2/blobs/1991b47bdfe3cb7e05d4a6f12d1bfb38a22fedf2da68d22269e675978940b57f new file mode 100644 index 0000000000000000000000000000000000000000..b6d4f7cb89e6d3607d16078e5054a828e471a95a --- /dev/null +++ b/models--LLM360--K2/blobs/1991b47bdfe3cb7e05d4a6f12d1bfb38a22fedf2da68d22269e675978940b57f @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1991b47bdfe3cb7e05d4a6f12d1bfb38a22fedf2da68d22269e675978940b57f +size 4857111932 diff --git a/models--LLM360--K2/blobs/2c9ca528e203129d5e7e95b0f0a915ceb8e0faea b/models--LLM360--K2/blobs/2c9ca528e203129d5e7e95b0f0a915ceb8e0faea new file mode 100644 index 0000000000000000000000000000000000000000..2c9ca528e203129d5e7e95b0f0a915ceb8e0faea --- /dev/null +++ b/models--LLM360--K2/blobs/2c9ca528e203129d5e7e95b0f0a915ceb8e0faea @@ -0,0 +1,730 @@ +{ + "metadata": { + "total_size": 130572369920 + }, + "weight_map": { + "lm_head.weight": "pytorch_model-00027-of-00027.bin", + "model.embed_tokens.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.0.input_layernorm.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.0.mlp.down_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.0.mlp.gate_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.0.mlp.up_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.0.post_attention_layernorm.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.0.self_attn.k_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.0.self_attn.o_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.0.self_attn.q_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.0.self_attn.v_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.1.input_layernorm.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.1.mlp.down_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.1.mlp.gate_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.1.mlp.up_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.1.post_attention_layernorm.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.1.self_attn.k_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.1.self_attn.o_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.1.self_attn.q_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.1.self_attn.v_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.10.input_layernorm.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.10.mlp.down_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.10.mlp.gate_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.10.mlp.up_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.10.post_attention_layernorm.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.10.self_attn.k_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.10.self_attn.o_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.10.self_attn.q_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.10.self_attn.v_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.11.input_layernorm.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.11.mlp.down_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.11.mlp.gate_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.11.mlp.up_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.11.post_attention_layernorm.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.11.self_attn.k_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.11.self_attn.o_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.11.self_attn.q_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.11.self_attn.v_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.12.input_layernorm.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.12.mlp.down_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.12.mlp.gate_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.12.mlp.up_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.12.post_attention_layernorm.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.12.self_attn.k_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.12.self_attn.o_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.12.self_attn.q_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.12.self_attn.v_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.13.input_layernorm.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.13.mlp.down_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.13.mlp.gate_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.13.mlp.up_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.13.post_attention_layernorm.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.13.self_attn.k_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.13.self_attn.o_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.13.self_attn.q_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.13.self_attn.v_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.14.input_layernorm.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.14.mlp.down_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.14.mlp.gate_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.14.mlp.up_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.14.post_attention_layernorm.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.14.self_attn.k_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.14.self_attn.o_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.14.self_attn.q_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.14.self_attn.v_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.15.input_layernorm.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.15.mlp.down_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.15.mlp.gate_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.15.mlp.up_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.15.post_attention_layernorm.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.15.self_attn.k_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.15.self_attn.o_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.15.self_attn.q_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.15.self_attn.v_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.16.input_layernorm.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.16.mlp.down_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.16.mlp.gate_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.16.mlp.up_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.16.post_attention_layernorm.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.16.self_attn.k_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.16.self_attn.o_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.16.self_attn.q_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.16.self_attn.v_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.17.input_layernorm.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.17.mlp.down_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.17.mlp.gate_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.17.mlp.up_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.17.post_attention_layernorm.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.17.self_attn.k_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.17.self_attn.o_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.17.self_attn.q_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.17.self_attn.v_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.18.input_layernorm.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.18.mlp.down_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.18.mlp.gate_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.18.mlp.up_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.18.post_attention_layernorm.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.18.self_attn.k_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.18.self_attn.o_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.18.self_attn.q_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.18.self_attn.v_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.19.input_layernorm.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.19.mlp.down_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.19.mlp.gate_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.19.mlp.up_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.19.post_attention_layernorm.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.19.self_attn.k_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.19.self_attn.o_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.19.self_attn.q_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.19.self_attn.v_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.2.input_layernorm.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.2.mlp.down_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.2.mlp.gate_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.2.mlp.up_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.2.post_attention_layernorm.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.2.self_attn.k_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.2.self_attn.o_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.2.self_attn.q_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.2.self_attn.v_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.20.input_layernorm.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.20.mlp.down_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.20.mlp.gate_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.20.mlp.up_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.20.post_attention_layernorm.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.20.self_attn.k_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.20.self_attn.o_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.20.self_attn.q_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.20.self_attn.v_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.21.input_layernorm.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.21.mlp.down_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.21.mlp.gate_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.21.mlp.up_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.21.post_attention_layernorm.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.21.self_attn.k_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.21.self_attn.o_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.21.self_attn.q_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.21.self_attn.v_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.22.input_layernorm.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.22.mlp.down_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.22.mlp.gate_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.22.mlp.up_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.22.post_attention_layernorm.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.22.self_attn.k_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.22.self_attn.o_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.22.self_attn.q_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.22.self_attn.v_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.23.input_layernorm.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.23.mlp.down_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.23.mlp.gate_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.23.mlp.up_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.23.post_attention_layernorm.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.23.self_attn.k_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.23.self_attn.o_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.23.self_attn.q_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.23.self_attn.v_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.24.input_layernorm.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.24.mlp.down_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.24.mlp.gate_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.24.mlp.up_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.24.post_attention_layernorm.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.24.self_attn.k_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.24.self_attn.o_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.24.self_attn.q_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.24.self_attn.v_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.25.input_layernorm.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.25.mlp.down_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.25.mlp.gate_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.25.mlp.up_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.25.post_attention_layernorm.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.25.self_attn.k_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.25.self_attn.o_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.25.self_attn.q_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.25.self_attn.v_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.26.input_layernorm.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.26.mlp.down_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.26.mlp.gate_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.26.mlp.up_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.26.post_attention_layernorm.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.26.self_attn.k_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.26.self_attn.o_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.26.self_attn.q_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.26.self_attn.v_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.27.input_layernorm.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.27.mlp.down_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.27.mlp.gate_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.27.mlp.up_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.27.post_attention_layernorm.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.27.self_attn.k_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.27.self_attn.o_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.27.self_attn.q_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.27.self_attn.v_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.28.input_layernorm.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.28.mlp.down_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.28.mlp.gate_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.28.mlp.up_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.28.post_attention_layernorm.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.28.self_attn.k_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.28.self_attn.o_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.28.self_attn.q_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.28.self_attn.v_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.29.input_layernorm.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.29.mlp.down_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.29.mlp.gate_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.29.mlp.up_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.29.post_attention_layernorm.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.29.self_attn.k_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.29.self_attn.o_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.29.self_attn.q_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.29.self_attn.v_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.3.input_layernorm.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.3.mlp.down_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.3.mlp.gate_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.3.mlp.up_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.3.post_attention_layernorm.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.3.self_attn.k_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.3.self_attn.o_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.3.self_attn.q_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.3.self_attn.v_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.30.input_layernorm.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.30.mlp.down_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.30.mlp.gate_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.30.mlp.up_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.30.post_attention_layernorm.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.30.self_attn.k_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.30.self_attn.o_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.30.self_attn.q_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.30.self_attn.v_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.31.input_layernorm.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.31.mlp.down_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.31.mlp.gate_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.31.mlp.up_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.31.post_attention_layernorm.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.31.self_attn.k_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.31.self_attn.o_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.31.self_attn.q_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.31.self_attn.v_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.32.input_layernorm.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.32.mlp.down_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.32.mlp.gate_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.32.mlp.up_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.32.post_attention_layernorm.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.32.self_attn.k_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.32.self_attn.o_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.32.self_attn.q_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.32.self_attn.v_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.33.input_layernorm.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.33.mlp.down_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.33.mlp.gate_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.33.mlp.up_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.33.post_attention_layernorm.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.33.self_attn.k_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.33.self_attn.o_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.33.self_attn.q_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.33.self_attn.v_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.34.input_layernorm.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.34.mlp.down_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.34.mlp.gate_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.34.mlp.up_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.34.post_attention_layernorm.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.34.self_attn.k_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.34.self_attn.o_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.34.self_attn.q_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.34.self_attn.v_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.35.input_layernorm.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.35.mlp.down_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.35.mlp.gate_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.35.mlp.up_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.35.post_attention_layernorm.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.35.self_attn.k_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.35.self_attn.o_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.35.self_attn.q_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.35.self_attn.v_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.36.input_layernorm.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.36.mlp.down_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.36.mlp.gate_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.36.mlp.up_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.36.post_attention_layernorm.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.36.self_attn.k_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.36.self_attn.o_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.36.self_attn.q_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.36.self_attn.v_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.37.input_layernorm.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.37.mlp.down_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.37.mlp.gate_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.37.mlp.up_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.37.post_attention_layernorm.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.37.self_attn.k_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.37.self_attn.o_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.37.self_attn.q_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.37.self_attn.v_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.38.input_layernorm.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.38.mlp.down_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.38.mlp.gate_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.38.mlp.up_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.38.post_attention_layernorm.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.38.self_attn.k_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.38.self_attn.o_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.38.self_attn.q_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.38.self_attn.v_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.39.input_layernorm.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.39.mlp.down_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.39.mlp.gate_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.39.mlp.up_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.39.post_attention_layernorm.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.39.self_attn.k_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.39.self_attn.o_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.39.self_attn.q_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.39.self_attn.v_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.4.input_layernorm.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.4.mlp.down_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.4.mlp.gate_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.4.mlp.up_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.4.post_attention_layernorm.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.4.self_attn.k_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.4.self_attn.o_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.4.self_attn.q_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.4.self_attn.v_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.40.input_layernorm.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.40.mlp.down_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.40.mlp.gate_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.40.mlp.up_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.40.post_attention_layernorm.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.40.self_attn.k_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.40.self_attn.o_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.40.self_attn.q_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.40.self_attn.v_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.41.input_layernorm.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.41.mlp.down_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.41.mlp.gate_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.41.mlp.up_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.41.post_attention_layernorm.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.41.self_attn.k_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.41.self_attn.o_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.41.self_attn.q_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.41.self_attn.v_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.42.input_layernorm.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.42.mlp.down_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.42.mlp.gate_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.42.mlp.up_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.42.post_attention_layernorm.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.42.self_attn.k_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.42.self_attn.o_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.42.self_attn.q_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.42.self_attn.v_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.43.input_layernorm.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.43.mlp.down_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.43.mlp.gate_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.43.mlp.up_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.43.post_attention_layernorm.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.43.self_attn.k_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.43.self_attn.o_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.43.self_attn.q_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.43.self_attn.v_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.44.input_layernorm.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.44.mlp.down_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.44.mlp.gate_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.44.mlp.up_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.44.post_attention_layernorm.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.44.self_attn.k_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.44.self_attn.o_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.44.self_attn.q_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.44.self_attn.v_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.45.input_layernorm.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.45.mlp.down_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.45.mlp.gate_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.45.mlp.up_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.45.post_attention_layernorm.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.45.self_attn.k_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.45.self_attn.o_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.45.self_attn.q_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.45.self_attn.v_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.46.input_layernorm.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.46.mlp.down_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.46.mlp.gate_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.46.mlp.up_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.46.post_attention_layernorm.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.46.self_attn.k_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.46.self_attn.o_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.46.self_attn.q_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.46.self_attn.v_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.47.input_layernorm.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.47.mlp.down_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.47.mlp.gate_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.47.mlp.up_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.47.post_attention_layernorm.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.47.self_attn.k_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.47.self_attn.o_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.47.self_attn.q_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.47.self_attn.v_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.48.input_layernorm.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.48.mlp.down_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.48.mlp.gate_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.48.mlp.up_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.48.post_attention_layernorm.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.48.self_attn.k_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.48.self_attn.o_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.48.self_attn.q_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.48.self_attn.v_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.49.input_layernorm.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.49.mlp.down_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.49.mlp.gate_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.49.mlp.up_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.49.post_attention_layernorm.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.49.self_attn.k_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.49.self_attn.o_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.49.self_attn.q_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.49.self_attn.v_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.5.input_layernorm.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.5.mlp.down_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.5.mlp.gate_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.5.mlp.up_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.5.post_attention_layernorm.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.5.self_attn.k_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.5.self_attn.o_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.5.self_attn.q_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.5.self_attn.v_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.50.input_layernorm.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.50.mlp.down_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.50.mlp.gate_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.50.mlp.up_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.50.post_attention_layernorm.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.50.self_attn.k_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.50.self_attn.o_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.50.self_attn.q_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.50.self_attn.v_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.51.input_layernorm.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.51.mlp.down_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.51.mlp.gate_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.51.mlp.up_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.51.post_attention_layernorm.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.51.self_attn.k_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.51.self_attn.o_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.51.self_attn.q_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.51.self_attn.v_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.52.input_layernorm.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.52.mlp.down_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.52.mlp.gate_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.52.mlp.up_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.52.post_attention_layernorm.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.52.self_attn.k_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.52.self_attn.o_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.52.self_attn.q_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.52.self_attn.v_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.53.input_layernorm.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.53.mlp.down_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.53.mlp.gate_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.53.mlp.up_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.53.post_attention_layernorm.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.53.self_attn.k_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.53.self_attn.o_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.53.self_attn.q_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.53.self_attn.v_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.54.input_layernorm.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.54.mlp.down_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.54.mlp.gate_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.54.mlp.up_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.54.post_attention_layernorm.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.54.self_attn.k_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.54.self_attn.o_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.54.self_attn.q_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.54.self_attn.v_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.55.input_layernorm.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.55.mlp.down_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.55.mlp.gate_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.55.mlp.up_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.55.post_attention_layernorm.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.55.self_attn.k_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.55.self_attn.o_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.55.self_attn.q_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.55.self_attn.v_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.56.input_layernorm.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.56.mlp.down_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.56.mlp.gate_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.56.mlp.up_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.56.post_attention_layernorm.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.56.self_attn.k_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.56.self_attn.o_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.56.self_attn.q_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.56.self_attn.v_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.57.input_layernorm.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.57.mlp.down_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.57.mlp.gate_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.57.mlp.up_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.57.post_attention_layernorm.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.57.self_attn.k_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.57.self_attn.o_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.57.self_attn.q_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.57.self_attn.v_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.58.input_layernorm.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.58.mlp.down_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.58.mlp.gate_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.58.mlp.up_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.58.post_attention_layernorm.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.58.self_attn.k_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.58.self_attn.o_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.58.self_attn.q_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.58.self_attn.v_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.59.input_layernorm.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.59.mlp.down_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.59.mlp.gate_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.59.mlp.up_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.59.post_attention_layernorm.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.59.self_attn.k_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.59.self_attn.o_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.59.self_attn.q_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.59.self_attn.v_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.6.input_layernorm.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.6.mlp.down_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.6.mlp.gate_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.6.mlp.up_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.6.post_attention_layernorm.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.6.self_attn.k_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.6.self_attn.o_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.6.self_attn.q_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.6.self_attn.v_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.60.input_layernorm.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.60.mlp.down_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.60.mlp.gate_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.60.mlp.up_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.60.post_attention_layernorm.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.60.self_attn.k_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.60.self_attn.o_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.60.self_attn.q_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.60.self_attn.v_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.61.input_layernorm.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.61.mlp.down_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.61.mlp.gate_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.61.mlp.up_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.61.post_attention_layernorm.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.61.self_attn.k_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.61.self_attn.o_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.61.self_attn.q_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.61.self_attn.v_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.62.input_layernorm.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.62.mlp.down_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.62.mlp.gate_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.62.mlp.up_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.62.post_attention_layernorm.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.62.self_attn.k_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.62.self_attn.o_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.62.self_attn.q_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.62.self_attn.v_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.63.input_layernorm.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.63.mlp.down_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.63.mlp.gate_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.63.mlp.up_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.63.post_attention_layernorm.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.63.self_attn.k_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.63.self_attn.o_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.63.self_attn.q_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.63.self_attn.v_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.64.input_layernorm.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.64.mlp.down_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.64.mlp.gate_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.64.mlp.up_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.64.post_attention_layernorm.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.64.self_attn.k_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.64.self_attn.o_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.64.self_attn.q_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.64.self_attn.v_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.65.input_layernorm.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.65.mlp.down_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.65.mlp.gate_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.65.mlp.up_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.65.post_attention_layernorm.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.65.self_attn.k_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.65.self_attn.o_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.65.self_attn.q_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.65.self_attn.v_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.66.input_layernorm.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.66.mlp.down_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.66.mlp.gate_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.66.mlp.up_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.66.post_attention_layernorm.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.66.self_attn.k_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.66.self_attn.o_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.66.self_attn.q_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.66.self_attn.v_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.67.input_layernorm.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.67.mlp.down_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.67.mlp.gate_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.67.mlp.up_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.67.post_attention_layernorm.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.67.self_attn.k_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.67.self_attn.o_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.67.self_attn.q_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.67.self_attn.v_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.68.input_layernorm.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.68.mlp.down_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.68.mlp.gate_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.68.mlp.up_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.68.post_attention_layernorm.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.68.self_attn.k_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.68.self_attn.o_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.68.self_attn.q_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.68.self_attn.v_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.69.input_layernorm.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.69.mlp.down_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.69.mlp.gate_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.69.mlp.up_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.69.post_attention_layernorm.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.69.self_attn.k_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.69.self_attn.o_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.69.self_attn.q_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.69.self_attn.v_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.7.input_layernorm.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.7.mlp.down_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.7.mlp.gate_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.7.mlp.up_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.7.post_attention_layernorm.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.7.self_attn.k_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.7.self_attn.o_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.7.self_attn.q_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.7.self_attn.v_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.70.input_layernorm.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.70.mlp.down_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.70.mlp.gate_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.70.mlp.up_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.70.post_attention_layernorm.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.70.self_attn.k_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.70.self_attn.o_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.70.self_attn.q_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.70.self_attn.v_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.71.input_layernorm.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.71.mlp.down_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.71.mlp.gate_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.71.mlp.up_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.71.post_attention_layernorm.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.71.self_attn.k_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.71.self_attn.o_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.71.self_attn.q_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.71.self_attn.v_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.72.input_layernorm.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.72.mlp.down_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.72.mlp.gate_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.72.mlp.up_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.72.post_attention_layernorm.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.72.self_attn.k_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.72.self_attn.o_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.72.self_attn.q_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.72.self_attn.v_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.73.input_layernorm.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.73.mlp.down_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.73.mlp.gate_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.73.mlp.up_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.73.post_attention_layernorm.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.73.self_attn.k_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.73.self_attn.o_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.73.self_attn.q_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.73.self_attn.v_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.74.input_layernorm.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.74.mlp.down_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.74.mlp.gate_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.74.mlp.up_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.74.post_attention_layernorm.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.74.self_attn.k_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.74.self_attn.o_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.74.self_attn.q_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.74.self_attn.v_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.75.input_layernorm.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.75.mlp.down_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.75.mlp.gate_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.75.mlp.up_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.75.post_attention_layernorm.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.75.self_attn.k_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.75.self_attn.o_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.75.self_attn.q_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.75.self_attn.v_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.76.input_layernorm.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.76.mlp.down_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.76.mlp.gate_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.76.mlp.up_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.76.post_attention_layernorm.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.76.self_attn.k_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.76.self_attn.o_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.76.self_attn.q_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.76.self_attn.v_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.77.input_layernorm.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.77.mlp.down_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.77.mlp.gate_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.77.mlp.up_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.77.post_attention_layernorm.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.77.self_attn.k_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.77.self_attn.o_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.77.self_attn.q_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.77.self_attn.v_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.78.input_layernorm.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.78.mlp.down_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.78.mlp.gate_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.78.mlp.up_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.78.post_attention_layernorm.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.78.self_attn.k_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.78.self_attn.o_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.78.self_attn.q_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.78.self_attn.v_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.79.input_layernorm.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.79.mlp.down_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.79.mlp.gate_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.79.mlp.up_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.79.post_attention_layernorm.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.79.self_attn.k_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.79.self_attn.o_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.79.self_attn.q_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.79.self_attn.v_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.8.input_layernorm.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.8.mlp.down_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.8.mlp.gate_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.8.mlp.up_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.8.post_attention_layernorm.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.8.self_attn.k_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.8.self_attn.o_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.8.self_attn.q_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.8.self_attn.v_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.9.input_layernorm.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.9.mlp.down_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.9.mlp.gate_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.9.mlp.up_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.9.post_attention_layernorm.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.9.self_attn.k_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.9.self_attn.o_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.9.self_attn.q_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.9.self_attn.v_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.norm.weight": "pytorch_model-00027-of-00027.bin" + } +} diff --git a/models--LLM360--K2/blobs/3bfc25078191c7795def871046f6cb059474fcef43440f4f6b9db0fd757a7f64 b/models--LLM360--K2/blobs/3bfc25078191c7795def871046f6cb059474fcef43440f4f6b9db0fd757a7f64 new file mode 100644 index 0000000000000000000000000000000000000000..de4ee1d5deabc19cac865bbc30e629233af680c7 --- /dev/null +++ b/models--LLM360--K2/blobs/3bfc25078191c7795def871046f6cb059474fcef43440f4f6b9db0fd757a7f64 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3bfc25078191c7795def871046f6cb059474fcef43440f4f6b9db0fd757a7f64 +size 4857111932 diff --git a/models--LLM360--K2/blobs/3df9e12db23090ffcb07219d5c388eabdd3cc780587d368242b398fdf8e90b4d b/models--LLM360--K2/blobs/3df9e12db23090ffcb07219d5c388eabdd3cc780587d368242b398fdf8e90b4d new file mode 100644 index 0000000000000000000000000000000000000000..8bcff7df49feabcf225a10adaf05bbd9994bab79 --- /dev/null +++ b/models--LLM360--K2/blobs/3df9e12db23090ffcb07219d5c388eabdd3cc780587d368242b398fdf8e90b4d @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3df9e12db23090ffcb07219d5c388eabdd3cc780587d368242b398fdf8e90b4d +size 4660470238 diff --git a/models--LLM360--K2/blobs/434fd7e790c2995a1f8c0ca46f6a3a5dea2c1d7d0366ed1b47e5bdec80314442 b/models--LLM360--K2/blobs/434fd7e790c2995a1f8c0ca46f6a3a5dea2c1d7d0366ed1b47e5bdec80314442 new file mode 100644 index 0000000000000000000000000000000000000000..8a71578f194be800af457ba0eb59b0bbc134f0a4 --- /dev/null +++ b/models--LLM360--K2/blobs/434fd7e790c2995a1f8c0ca46f6a3a5dea2c1d7d0366ed1b47e5bdec80314442 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:434fd7e790c2995a1f8c0ca46f6a3a5dea2c1d7d0366ed1b47e5bdec80314442 +size 4857111932 diff --git a/models--LLM360--K2/blobs/4a1ff5c1228e7d11248f2fcfbd70dc59bb26a4ed737becb63e2e130b028e1698 b/models--LLM360--K2/blobs/4a1ff5c1228e7d11248f2fcfbd70dc59bb26a4ed737becb63e2e130b028e1698 new file mode 100644 index 0000000000000000000000000000000000000000..8c7e0b459545099a7d1db2de56fe132440a78dee --- /dev/null +++ b/models--LLM360--K2/blobs/4a1ff5c1228e7d11248f2fcfbd70dc59bb26a4ed737becb63e2e130b028e1698 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a1ff5c1228e7d11248f2fcfbd70dc59bb26a4ed737becb63e2e130b028e1698 +size 4857111932 diff --git a/models--LLM360--K2/blobs/4f3b40b1a1fef50d05c3f48ce8a22bd543fd3abe376153290bacc3d37134b463 b/models--LLM360--K2/blobs/4f3b40b1a1fef50d05c3f48ce8a22bd543fd3abe376153290bacc3d37134b463 new file mode 100644 index 0000000000000000000000000000000000000000..d4027dcc4904f4c76fdd6e6f54a6cf14a68a9e2c --- /dev/null +++ b/models--LLM360--K2/blobs/4f3b40b1a1fef50d05c3f48ce8a22bd543fd3abe376153290bacc3d37134b463 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4f3b40b1a1fef50d05c3f48ce8a22bd543fd3abe376153290bacc3d37134b463 +size 4857111932 diff --git a/models--LLM360--K2/blobs/583722a14a478af056eb7190401b2fa95c90d75ab1de83acb4d5f20941a224d1 b/models--LLM360--K2/blobs/583722a14a478af056eb7190401b2fa95c90d75ab1de83acb4d5f20941a224d1 new file mode 100644 index 0000000000000000000000000000000000000000..e2479dce1209d428346e80f6471e127a130ae0db --- /dev/null +++ b/models--LLM360--K2/blobs/583722a14a478af056eb7190401b2fa95c90d75ab1de83acb4d5f20941a224d1 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:583722a14a478af056eb7190401b2fa95c90d75ab1de83acb4d5f20941a224d1 +size 4857111932 diff --git a/models--LLM360--K2/blobs/5afe6368480862ccd7e5fab96e17159a6a56a6ef2f75083ac04b69df977784da b/models--LLM360--K2/blobs/5afe6368480862ccd7e5fab96e17159a6a56a6ef2f75083ac04b69df977784da new file mode 100644 index 0000000000000000000000000000000000000000..f8e0ce37cf7d233a4ea4e6f4dd13ab9aa622d4bc --- /dev/null +++ b/models--LLM360--K2/blobs/5afe6368480862ccd7e5fab96e17159a6a56a6ef2f75083ac04b69df977784da @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5afe6368480862ccd7e5fab96e17159a6a56a6ef2f75083ac04b69df977784da +size 4857111932 diff --git a/models--LLM360--K2/blobs/5b4779f96a5161dccf4a59cc5479918f931c47342fa44ab01a9272f476963083 b/models--LLM360--K2/blobs/5b4779f96a5161dccf4a59cc5479918f931c47342fa44ab01a9272f476963083 new file mode 100644 index 0000000000000000000000000000000000000000..55734c036070462728de909e93be3f7ddda3f368 --- /dev/null +++ b/models--LLM360--K2/blobs/5b4779f96a5161dccf4a59cc5479918f931c47342fa44ab01a9272f476963083 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5b4779f96a5161dccf4a59cc5479918f931c47342fa44ab01a9272f476963083 +size 4857111932 diff --git a/models--LLM360--K2/blobs/617edceaa3b26e387a4d9aec6e8104db2c81ac35e5bf47f9e518af22e956094a b/models--LLM360--K2/blobs/617edceaa3b26e387a4d9aec6e8104db2c81ac35e5bf47f9e518af22e956094a new file mode 100644 index 0000000000000000000000000000000000000000..f31a3868f90bd8ea859270c3e40813a39146c316 --- /dev/null +++ b/models--LLM360--K2/blobs/617edceaa3b26e387a4d9aec6e8104db2c81ac35e5bf47f9e518af22e956094a @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:617edceaa3b26e387a4d9aec6e8104db2c81ac35e5bf47f9e518af22e956094a +size 4857111932 diff --git a/models--LLM360--K2/blobs/66a1a4140058b1e5040be2ebbb351ba3c73051295dbaf353487860631be06bce b/models--LLM360--K2/blobs/66a1a4140058b1e5040be2ebbb351ba3c73051295dbaf353487860631be06bce new file mode 100644 index 0000000000000000000000000000000000000000..4f4f87cdff64a76266b38ee423da29f413503d80 --- /dev/null +++ b/models--LLM360--K2/blobs/66a1a4140058b1e5040be2ebbb351ba3c73051295dbaf353487860631be06bce @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:66a1a4140058b1e5040be2ebbb351ba3c73051295dbaf353487860631be06bce +size 4857111932 diff --git a/models--LLM360--K2/blobs/72b7ab829379c5ce42fddf9b8ebae0ff50f967d0cd049af34c796be1dbcc3837 b/models--LLM360--K2/blobs/72b7ab829379c5ce42fddf9b8ebae0ff50f967d0cd049af34c796be1dbcc3837 new file mode 100644 index 0000000000000000000000000000000000000000..7c342878228fdd8814ef91a4e76152d2f92c62c5 --- /dev/null +++ b/models--LLM360--K2/blobs/72b7ab829379c5ce42fddf9b8ebae0ff50f967d0cd049af34c796be1dbcc3837 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:72b7ab829379c5ce42fddf9b8ebae0ff50f967d0cd049af34c796be1dbcc3837 +size 4857111932 diff --git a/models--LLM360--K2/blobs/7c546f28758f14aea461f4eedba1dc6380ffb525bcb06328c2e7ccf4bc0914e3 b/models--LLM360--K2/blobs/7c546f28758f14aea461f4eedba1dc6380ffb525bcb06328c2e7ccf4bc0914e3 new file mode 100644 index 0000000000000000000000000000000000000000..adccb90f2cead3388be14f91c681ed1568989e41 --- /dev/null +++ b/models--LLM360--K2/blobs/7c546f28758f14aea461f4eedba1dc6380ffb525bcb06328c2e7ccf4bc0914e3 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7c546f28758f14aea461f4eedba1dc6380ffb525bcb06328c2e7ccf4bc0914e3 +size 4857111932 diff --git a/models--LLM360--K2/blobs/7ef2e098b4cf22b9facdb8f1df3f5ef4d1416f16 b/models--LLM360--K2/blobs/7ef2e098b4cf22b9facdb8f1df3f5ef4d1416f16 new file mode 100644 index 0000000000000000000000000000000000000000..7ef2e098b4cf22b9facdb8f1df3f5ef4d1416f16 --- /dev/null +++ b/models--LLM360--K2/blobs/7ef2e098b4cf22b9facdb8f1df3f5ef4d1416f16 @@ -0,0 +1,30 @@ +{ + "_name_or_path": "huggyllama/llama-65b", + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "eos_token_id": 2, + "hidden_act": "silu", + "hidden_size": 8192, + "initializer_range": 0.02, + "intermediate_size": 22016, + "max_position_embeddings": 2048, + "max_sequence_length": 2048, + "model_type": "llama", + "num_attention_heads": 64, + "num_hidden_layers": 80, + "num_key_value_heads": 64, + "pad_token_id": 0, + "pretraining_tp": 1, + "rms_norm_eps": 1e-05, + "rope_scaling": null, + "rope_theta": 10000.0, + "tie_word_embeddings": false, + "torch_dtype": "float16", + "transformers_version": "4.39.3", + "use_cache": true, + "vocab_size": 32032 +} diff --git a/models--LLM360--K2/blobs/80860954ce526545b5eaa951ed98a9dc87576e0ce5243b53001e79a90f107ac2 b/models--LLM360--K2/blobs/80860954ce526545b5eaa951ed98a9dc87576e0ce5243b53001e79a90f107ac2 new file mode 100644 index 0000000000000000000000000000000000000000..473828bc795a1cfe82dd06384d9e310e54bcb93b --- /dev/null +++ b/models--LLM360--K2/blobs/80860954ce526545b5eaa951ed98a9dc87576e0ce5243b53001e79a90f107ac2 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:80860954ce526545b5eaa951ed98a9dc87576e0ce5243b53001e79a90f107ac2 +size 4857111932 diff --git a/models--LLM360--K2/blobs/8fecd1dfe9745e2b6546298051b39dd02ace8b88eb0232a9d6a9beccb935f9b6 b/models--LLM360--K2/blobs/8fecd1dfe9745e2b6546298051b39dd02ace8b88eb0232a9d6a9beccb935f9b6 new file mode 100644 index 0000000000000000000000000000000000000000..927da3dfc925b6b1e4a2bfa0c2ed5c33b20ee8a8 --- /dev/null +++ b/models--LLM360--K2/blobs/8fecd1dfe9745e2b6546298051b39dd02ace8b88eb0232a9d6a9beccb935f9b6 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8fecd1dfe9745e2b6546298051b39dd02ace8b88eb0232a9d6a9beccb935f9b6 +size 4857111932 diff --git a/models--LLM360--K2/blobs/9844405d62d9f809278ed1577b4364f68327282f461b4afaf75a9cebf074c325 b/models--LLM360--K2/blobs/9844405d62d9f809278ed1577b4364f68327282f461b4afaf75a9cebf074c325 new file mode 100644 index 0000000000000000000000000000000000000000..96e6b2fe882b10c748aff2749319da02d5e10070 --- /dev/null +++ b/models--LLM360--K2/blobs/9844405d62d9f809278ed1577b4364f68327282f461b4afaf75a9cebf074c325 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9844405d62d9f809278ed1577b4364f68327282f461b4afaf75a9cebf074c325 +size 4857111932 diff --git a/models--LLM360--K2/blobs/a258e4bed92e87cfeacce9077d9e2989e45436d6259e6f16f1cace7d7fcdd6d8 b/models--LLM360--K2/blobs/a258e4bed92e87cfeacce9077d9e2989e45436d6259e6f16f1cace7d7fcdd6d8 new file mode 100644 index 0000000000000000000000000000000000000000..32a98e1e90f4e5e1e55ccdb2f7cc90af1db05a28 --- /dev/null +++ b/models--LLM360--K2/blobs/a258e4bed92e87cfeacce9077d9e2989e45436d6259e6f16f1cace7d7fcdd6d8 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a258e4bed92e87cfeacce9077d9e2989e45436d6259e6f16f1cace7d7fcdd6d8 +size 4857111932 diff --git a/models--LLM360--K2/blobs/b6c7e1e011596a62423becf7831de6885bd4cb66835c9b12a1890d82341caa5d b/models--LLM360--K2/blobs/b6c7e1e011596a62423becf7831de6885bd4cb66835c9b12a1890d82341caa5d new file mode 100644 index 0000000000000000000000000000000000000000..6d69d5a1e997993d3e8db4ff0b4d8ecfeef3fc79 --- /dev/null +++ b/models--LLM360--K2/blobs/b6c7e1e011596a62423becf7831de6885bd4cb66835c9b12a1890d82341caa5d @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b6c7e1e011596a62423becf7831de6885bd4cb66835c9b12a1890d82341caa5d +size 4857111932 diff --git a/models--LLM360--K2/blobs/c0c42dfdce41209444c998d0f30f5bdabf5e747f9ba18f5a917a3e72c8a4956b b/models--LLM360--K2/blobs/c0c42dfdce41209444c998d0f30f5bdabf5e747f9ba18f5a917a3e72c8a4956b new file mode 100644 index 0000000000000000000000000000000000000000..7e0c204c4154f8e1a08d3420af0fbde0363bdbee --- /dev/null +++ b/models--LLM360--K2/blobs/c0c42dfdce41209444c998d0f30f5bdabf5e747f9ba18f5a917a3e72c8a4956b @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c0c42dfdce41209444c998d0f30f5bdabf5e747f9ba18f5a917a3e72c8a4956b +size 4484358522 diff --git a/models--LLM360--K2/blobs/c1f7a86a6d975dad016a1692a17b3c8116c1ff4ef04362060addfc1b65aa2c08 b/models--LLM360--K2/blobs/c1f7a86a6d975dad016a1692a17b3c8116c1ff4ef04362060addfc1b65aa2c08 new file mode 100644 index 0000000000000000000000000000000000000000..4f09592e86545f1e79e6062d42b0c0f1a9de5781 --- /dev/null +++ b/models--LLM360--K2/blobs/c1f7a86a6d975dad016a1692a17b3c8116c1ff4ef04362060addfc1b65aa2c08 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c1f7a86a6d975dad016a1692a17b3c8116c1ff4ef04362060addfc1b65aa2c08 +size 4857111932 diff --git a/models--LLM360--K2/blobs/c55743fc3d25e389dc4e29f7bb2db5e21f033fc8292ffe411eed32fdff0b2498 b/models--LLM360--K2/blobs/c55743fc3d25e389dc4e29f7bb2db5e21f033fc8292ffe411eed32fdff0b2498 new file mode 100644 index 0000000000000000000000000000000000000000..d9b593b7286dc65ba7c5bd87b9a457be971dda75 --- /dev/null +++ b/models--LLM360--K2/blobs/c55743fc3d25e389dc4e29f7bb2db5e21f033fc8292ffe411eed32fdff0b2498 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c55743fc3d25e389dc4e29f7bb2db5e21f033fc8292ffe411eed32fdff0b2498 +size 4857111932 diff --git a/models--LLM360--K2/blobs/c9a048c6dfcef5561324dbded792521edfb1d9c7a03404189d2a09157c628760 b/models--LLM360--K2/blobs/c9a048c6dfcef5561324dbded792521edfb1d9c7a03404189d2a09157c628760 new file mode 100644 index 0000000000000000000000000000000000000000..c8ae1d76b8fa5ee28b52bd966aaa954def23d19a --- /dev/null +++ b/models--LLM360--K2/blobs/c9a048c6dfcef5561324dbded792521edfb1d9c7a03404189d2a09157c628760 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c9a048c6dfcef5561324dbded792521edfb1d9c7a03404189d2a09157c628760 +size 4857111932 diff --git a/models--LLM360--K2/blobs/ca618179df960063f7086853a0de59fd2fcfa2b7a220fe089d66743aa6648b8e b/models--LLM360--K2/blobs/ca618179df960063f7086853a0de59fd2fcfa2b7a220fe089d66743aa6648b8e new file mode 100644 index 0000000000000000000000000000000000000000..3c866c22eac79894d0a18bc0a8239fe697810e71 --- /dev/null +++ b/models--LLM360--K2/blobs/ca618179df960063f7086853a0de59fd2fcfa2b7a220fe089d66743aa6648b8e @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ca618179df960063f7086853a0de59fd2fcfa2b7a220fe089d66743aa6648b8e +size 4857111932 diff --git a/models--LLM360--K2/blobs/d3ebaa9cb6f5e87aa968768383824375290dc33e56a1852e48821fe285fd8448 b/models--LLM360--K2/blobs/d3ebaa9cb6f5e87aa968768383824375290dc33e56a1852e48821fe285fd8448 new file mode 100644 index 0000000000000000000000000000000000000000..7f6e74fcd068d36f896bf19154e1cb2fdfa1304a --- /dev/null +++ b/models--LLM360--K2/blobs/d3ebaa9cb6f5e87aa968768383824375290dc33e56a1852e48821fe285fd8448 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d3ebaa9cb6f5e87aa968768383824375290dc33e56a1852e48821fe285fd8448 +size 4857111932 diff --git a/models--LLM360--K2/refs/main b/models--LLM360--K2/refs/main new file mode 100644 index 0000000000000000000000000000000000000000..7a0ce76bf946b4efce99069a01002bf06ddd8bff --- /dev/null +++ b/models--LLM360--K2/refs/main @@ -0,0 +1 @@ +ccd3ced964e211a9c31cb346444bce761d26519e \ No newline at end of file diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/config.json b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/config.json new file mode 100644 index 0000000000000000000000000000000000000000..7ef2e098b4cf22b9facdb8f1df3f5ef4d1416f16 --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/config.json @@ -0,0 +1,30 @@ +{ + "_name_or_path": "huggyllama/llama-65b", + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "eos_token_id": 2, + "hidden_act": "silu", + "hidden_size": 8192, + "initializer_range": 0.02, + "intermediate_size": 22016, + "max_position_embeddings": 2048, + "max_sequence_length": 2048, + "model_type": "llama", + "num_attention_heads": 64, + "num_hidden_layers": 80, + "num_key_value_heads": 64, + "pad_token_id": 0, + "pretraining_tp": 1, + "rms_norm_eps": 1e-05, + "rope_scaling": null, + "rope_theta": 10000.0, + "tie_word_embeddings": false, + "torch_dtype": "float16", + "transformers_version": "4.39.3", + "use_cache": true, + "vocab_size": 32032 +} diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00001-of-00027.bin b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00001-of-00027.bin new file mode 100644 index 0000000000000000000000000000000000000000..8bcff7df49feabcf225a10adaf05bbd9994bab79 --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00001-of-00027.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3df9e12db23090ffcb07219d5c388eabdd3cc780587d368242b398fdf8e90b4d +size 4660470238 diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00002-of-00027.bin b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00002-of-00027.bin new file mode 100644 index 0000000000000000000000000000000000000000..b6d4f7cb89e6d3607d16078e5054a828e471a95a --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00002-of-00027.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1991b47bdfe3cb7e05d4a6f12d1bfb38a22fedf2da68d22269e675978940b57f +size 4857111932 diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00003-of-00027.bin b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00003-of-00027.bin new file mode 100644 index 0000000000000000000000000000000000000000..96e6b2fe882b10c748aff2749319da02d5e10070 --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00003-of-00027.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9844405d62d9f809278ed1577b4364f68327282f461b4afaf75a9cebf074c325 +size 4857111932 diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00004-of-00027.bin b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00004-of-00027.bin new file mode 100644 index 0000000000000000000000000000000000000000..c8ae1d76b8fa5ee28b52bd966aaa954def23d19a --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00004-of-00027.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c9a048c6dfcef5561324dbded792521edfb1d9c7a03404189d2a09157c628760 +size 4857111932 diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00005-of-00027.bin b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00005-of-00027.bin new file mode 100644 index 0000000000000000000000000000000000000000..4334414923310eb5f2c2a36e9aded6c0a746f067 --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00005-of-00027.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e5448b8c9a7a705501c5103af02f076036414b4cf54083c9ad17048d97221b5 +size 4857111932 diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00006-of-00027.bin b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00006-of-00027.bin new file mode 100644 index 0000000000000000000000000000000000000000..e2479dce1209d428346e80f6471e127a130ae0db --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00006-of-00027.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:583722a14a478af056eb7190401b2fa95c90d75ab1de83acb4d5f20941a224d1 +size 4857111932 diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00007-of-00027.bin b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00007-of-00027.bin new file mode 100644 index 0000000000000000000000000000000000000000..4f09592e86545f1e79e6062d42b0c0f1a9de5781 --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00007-of-00027.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c1f7a86a6d975dad016a1692a17b3c8116c1ff4ef04362060addfc1b65aa2c08 +size 4857111932 diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00008-of-00027.bin b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00008-of-00027.bin new file mode 100644 index 0000000000000000000000000000000000000000..f8e0ce37cf7d233a4ea4e6f4dd13ab9aa622d4bc --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00008-of-00027.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5afe6368480862ccd7e5fab96e17159a6a56a6ef2f75083ac04b69df977784da +size 4857111932 diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00009-of-00027.bin b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00009-of-00027.bin new file mode 100644 index 0000000000000000000000000000000000000000..6d69d5a1e997993d3e8db4ff0b4d8ecfeef3fc79 --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00009-of-00027.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b6c7e1e011596a62423becf7831de6885bd4cb66835c9b12a1890d82341caa5d +size 4857111932 diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00010-of-00027.bin b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00010-of-00027.bin new file mode 100644 index 0000000000000000000000000000000000000000..6c8c03014934a9b3c5489dcc704790d9c8fc1dff --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00010-of-00027.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:197922ce7f7807ba59bfdd42fbbe9c67e2ff8370ffda566614cdf74463363f2b +size 4857111932 diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00011-of-00027.bin b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00011-of-00027.bin new file mode 100644 index 0000000000000000000000000000000000000000..7c342878228fdd8814ef91a4e76152d2f92c62c5 --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00011-of-00027.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:72b7ab829379c5ce42fddf9b8ebae0ff50f967d0cd049af34c796be1dbcc3837 +size 4857111932 diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00012-of-00027.bin b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00012-of-00027.bin new file mode 100644 index 0000000000000000000000000000000000000000..7f6e74fcd068d36f896bf19154e1cb2fdfa1304a --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00012-of-00027.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d3ebaa9cb6f5e87aa968768383824375290dc33e56a1852e48821fe285fd8448 +size 4857111932 diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00013-of-00027.bin b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00013-of-00027.bin new file mode 100644 index 0000000000000000000000000000000000000000..3c866c22eac79894d0a18bc0a8239fe697810e71 --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00013-of-00027.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ca618179df960063f7086853a0de59fd2fcfa2b7a220fe089d66743aa6648b8e +size 4857111932 diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00014-of-00027.bin b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00014-of-00027.bin new file mode 100644 index 0000000000000000000000000000000000000000..32a98e1e90f4e5e1e55ccdb2f7cc90af1db05a28 --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00014-of-00027.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a258e4bed92e87cfeacce9077d9e2989e45436d6259e6f16f1cace7d7fcdd6d8 +size 4857111932 diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00015-of-00027.bin b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00015-of-00027.bin new file mode 100644 index 0000000000000000000000000000000000000000..de4ee1d5deabc19cac865bbc30e629233af680c7 --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00015-of-00027.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3bfc25078191c7795def871046f6cb059474fcef43440f4f6b9db0fd757a7f64 +size 4857111932 diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00016-of-00027.bin b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00016-of-00027.bin new file mode 100644 index 0000000000000000000000000000000000000000..927da3dfc925b6b1e4a2bfa0c2ed5c33b20ee8a8 --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00016-of-00027.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8fecd1dfe9745e2b6546298051b39dd02ace8b88eb0232a9d6a9beccb935f9b6 +size 4857111932 diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00017-of-00027.bin b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00017-of-00027.bin new file mode 100644 index 0000000000000000000000000000000000000000..d9b593b7286dc65ba7c5bd87b9a457be971dda75 --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00017-of-00027.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c55743fc3d25e389dc4e29f7bb2db5e21f033fc8292ffe411eed32fdff0b2498 +size 4857111932 diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00018-of-00027.bin b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00018-of-00027.bin new file mode 100644 index 0000000000000000000000000000000000000000..f31a3868f90bd8ea859270c3e40813a39146c316 --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00018-of-00027.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:617edceaa3b26e387a4d9aec6e8104db2c81ac35e5bf47f9e518af22e956094a +size 4857111932 diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00019-of-00027.bin b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00019-of-00027.bin new file mode 100644 index 0000000000000000000000000000000000000000..8a71578f194be800af457ba0eb59b0bbc134f0a4 --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00019-of-00027.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:434fd7e790c2995a1f8c0ca46f6a3a5dea2c1d7d0366ed1b47e5bdec80314442 +size 4857111932 diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00020-of-00027.bin b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00020-of-00027.bin new file mode 100644 index 0000000000000000000000000000000000000000..d4027dcc4904f4c76fdd6e6f54a6cf14a68a9e2c --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00020-of-00027.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4f3b40b1a1fef50d05c3f48ce8a22bd543fd3abe376153290bacc3d37134b463 +size 4857111932 diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00021-of-00027.bin b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00021-of-00027.bin new file mode 100644 index 0000000000000000000000000000000000000000..55734c036070462728de909e93be3f7ddda3f368 --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00021-of-00027.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5b4779f96a5161dccf4a59cc5479918f931c47342fa44ab01a9272f476963083 +size 4857111932 diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00022-of-00027.bin b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00022-of-00027.bin new file mode 100644 index 0000000000000000000000000000000000000000..4f4f87cdff64a76266b38ee423da29f413503d80 --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00022-of-00027.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:66a1a4140058b1e5040be2ebbb351ba3c73051295dbaf353487860631be06bce +size 4857111932 diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00023-of-00027.bin b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00023-of-00027.bin new file mode 100644 index 0000000000000000000000000000000000000000..8c7e0b459545099a7d1db2de56fe132440a78dee --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00023-of-00027.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a1ff5c1228e7d11248f2fcfbd70dc59bb26a4ed737becb63e2e130b028e1698 +size 4857111932 diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00024-of-00027.bin b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00024-of-00027.bin new file mode 100644 index 0000000000000000000000000000000000000000..adccb90f2cead3388be14f91c681ed1568989e41 --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00024-of-00027.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7c546f28758f14aea461f4eedba1dc6380ffb525bcb06328c2e7ccf4bc0914e3 +size 4857111932 diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00025-of-00027.bin b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00025-of-00027.bin new file mode 100644 index 0000000000000000000000000000000000000000..473828bc795a1cfe82dd06384d9e310e54bcb93b --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00025-of-00027.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:80860954ce526545b5eaa951ed98a9dc87576e0ce5243b53001e79a90f107ac2 +size 4857111932 diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00026-of-00027.bin b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00026-of-00027.bin new file mode 100644 index 0000000000000000000000000000000000000000..88d73ac21e498fb2dd3444269a572e0b06803f71 --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00026-of-00027.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0ef2f96b2f9b4943f45dc1ba39f0594eb14cf128c5ff5f41db34ad07cb556e7f +size 4857111932 diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00027-of-00027.bin b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00027-of-00027.bin new file mode 100644 index 0000000000000000000000000000000000000000..7e0c204c4154f8e1a08d3420af0fbde0363bdbee --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model-00027-of-00027.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c0c42dfdce41209444c998d0f30f5bdabf5e747f9ba18f5a917a3e72c8a4956b +size 4484358522 diff --git a/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model.bin.index.json b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model.bin.index.json new file mode 100644 index 0000000000000000000000000000000000000000..2c9ca528e203129d5e7e95b0f0a915ceb8e0faea --- /dev/null +++ b/models--LLM360--K2/snapshots/ccd3ced964e211a9c31cb346444bce761d26519e/pytorch_model.bin.index.json @@ -0,0 +1,730 @@ +{ + "metadata": { + "total_size": 130572369920 + }, + "weight_map": { + "lm_head.weight": "pytorch_model-00027-of-00027.bin", + "model.embed_tokens.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.0.input_layernorm.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.0.mlp.down_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.0.mlp.gate_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.0.mlp.up_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.0.post_attention_layernorm.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.0.self_attn.k_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.0.self_attn.o_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.0.self_attn.q_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.0.self_attn.v_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.1.input_layernorm.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.1.mlp.down_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.1.mlp.gate_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.1.mlp.up_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.1.post_attention_layernorm.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.1.self_attn.k_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.1.self_attn.o_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.1.self_attn.q_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.1.self_attn.v_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.10.input_layernorm.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.10.mlp.down_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.10.mlp.gate_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.10.mlp.up_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.10.post_attention_layernorm.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.10.self_attn.k_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.10.self_attn.o_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.10.self_attn.q_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.10.self_attn.v_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.11.input_layernorm.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.11.mlp.down_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.11.mlp.gate_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.11.mlp.up_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.11.post_attention_layernorm.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.11.self_attn.k_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.11.self_attn.o_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.11.self_attn.q_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.11.self_attn.v_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.12.input_layernorm.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.12.mlp.down_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.12.mlp.gate_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.12.mlp.up_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.12.post_attention_layernorm.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.12.self_attn.k_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.12.self_attn.o_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.12.self_attn.q_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.12.self_attn.v_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.13.input_layernorm.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.13.mlp.down_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.13.mlp.gate_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.13.mlp.up_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.13.post_attention_layernorm.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.13.self_attn.k_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.13.self_attn.o_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.13.self_attn.q_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.13.self_attn.v_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.14.input_layernorm.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.14.mlp.down_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.14.mlp.gate_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.14.mlp.up_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.14.post_attention_layernorm.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.14.self_attn.k_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.14.self_attn.o_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.14.self_attn.q_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.14.self_attn.v_proj.weight": "pytorch_model-00005-of-00027.bin", + "model.layers.15.input_layernorm.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.15.mlp.down_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.15.mlp.gate_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.15.mlp.up_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.15.post_attention_layernorm.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.15.self_attn.k_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.15.self_attn.o_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.15.self_attn.q_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.15.self_attn.v_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.16.input_layernorm.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.16.mlp.down_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.16.mlp.gate_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.16.mlp.up_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.16.post_attention_layernorm.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.16.self_attn.k_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.16.self_attn.o_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.16.self_attn.q_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.16.self_attn.v_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.17.input_layernorm.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.17.mlp.down_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.17.mlp.gate_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.17.mlp.up_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.17.post_attention_layernorm.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.17.self_attn.k_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.17.self_attn.o_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.17.self_attn.q_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.17.self_attn.v_proj.weight": "pytorch_model-00006-of-00027.bin", + "model.layers.18.input_layernorm.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.18.mlp.down_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.18.mlp.gate_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.18.mlp.up_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.18.post_attention_layernorm.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.18.self_attn.k_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.18.self_attn.o_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.18.self_attn.q_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.18.self_attn.v_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.19.input_layernorm.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.19.mlp.down_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.19.mlp.gate_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.19.mlp.up_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.19.post_attention_layernorm.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.19.self_attn.k_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.19.self_attn.o_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.19.self_attn.q_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.19.self_attn.v_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.2.input_layernorm.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.2.mlp.down_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.2.mlp.gate_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.2.mlp.up_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.2.post_attention_layernorm.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.2.self_attn.k_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.2.self_attn.o_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.2.self_attn.q_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.2.self_attn.v_proj.weight": "pytorch_model-00001-of-00027.bin", + "model.layers.20.input_layernorm.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.20.mlp.down_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.20.mlp.gate_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.20.mlp.up_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.20.post_attention_layernorm.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.20.self_attn.k_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.20.self_attn.o_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.20.self_attn.q_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.20.self_attn.v_proj.weight": "pytorch_model-00007-of-00027.bin", + "model.layers.21.input_layernorm.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.21.mlp.down_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.21.mlp.gate_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.21.mlp.up_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.21.post_attention_layernorm.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.21.self_attn.k_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.21.self_attn.o_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.21.self_attn.q_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.21.self_attn.v_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.22.input_layernorm.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.22.mlp.down_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.22.mlp.gate_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.22.mlp.up_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.22.post_attention_layernorm.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.22.self_attn.k_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.22.self_attn.o_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.22.self_attn.q_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.22.self_attn.v_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.23.input_layernorm.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.23.mlp.down_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.23.mlp.gate_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.23.mlp.up_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.23.post_attention_layernorm.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.23.self_attn.k_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.23.self_attn.o_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.23.self_attn.q_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.23.self_attn.v_proj.weight": "pytorch_model-00008-of-00027.bin", + "model.layers.24.input_layernorm.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.24.mlp.down_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.24.mlp.gate_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.24.mlp.up_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.24.post_attention_layernorm.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.24.self_attn.k_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.24.self_attn.o_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.24.self_attn.q_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.24.self_attn.v_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.25.input_layernorm.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.25.mlp.down_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.25.mlp.gate_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.25.mlp.up_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.25.post_attention_layernorm.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.25.self_attn.k_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.25.self_attn.o_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.25.self_attn.q_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.25.self_attn.v_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.26.input_layernorm.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.26.mlp.down_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.26.mlp.gate_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.26.mlp.up_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.26.post_attention_layernorm.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.26.self_attn.k_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.26.self_attn.o_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.26.self_attn.q_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.26.self_attn.v_proj.weight": "pytorch_model-00009-of-00027.bin", + "model.layers.27.input_layernorm.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.27.mlp.down_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.27.mlp.gate_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.27.mlp.up_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.27.post_attention_layernorm.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.27.self_attn.k_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.27.self_attn.o_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.27.self_attn.q_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.27.self_attn.v_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.28.input_layernorm.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.28.mlp.down_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.28.mlp.gate_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.28.mlp.up_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.28.post_attention_layernorm.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.28.self_attn.k_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.28.self_attn.o_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.28.self_attn.q_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.28.self_attn.v_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.29.input_layernorm.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.29.mlp.down_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.29.mlp.gate_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.29.mlp.up_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.29.post_attention_layernorm.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.29.self_attn.k_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.29.self_attn.o_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.29.self_attn.q_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.29.self_attn.v_proj.weight": "pytorch_model-00010-of-00027.bin", + "model.layers.3.input_layernorm.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.3.mlp.down_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.3.mlp.gate_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.3.mlp.up_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.3.post_attention_layernorm.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.3.self_attn.k_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.3.self_attn.o_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.3.self_attn.q_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.3.self_attn.v_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.30.input_layernorm.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.30.mlp.down_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.30.mlp.gate_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.30.mlp.up_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.30.post_attention_layernorm.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.30.self_attn.k_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.30.self_attn.o_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.30.self_attn.q_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.30.self_attn.v_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.31.input_layernorm.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.31.mlp.down_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.31.mlp.gate_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.31.mlp.up_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.31.post_attention_layernorm.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.31.self_attn.k_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.31.self_attn.o_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.31.self_attn.q_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.31.self_attn.v_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.32.input_layernorm.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.32.mlp.down_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.32.mlp.gate_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.32.mlp.up_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.32.post_attention_layernorm.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.32.self_attn.k_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.32.self_attn.o_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.32.self_attn.q_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.32.self_attn.v_proj.weight": "pytorch_model-00011-of-00027.bin", + "model.layers.33.input_layernorm.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.33.mlp.down_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.33.mlp.gate_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.33.mlp.up_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.33.post_attention_layernorm.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.33.self_attn.k_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.33.self_attn.o_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.33.self_attn.q_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.33.self_attn.v_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.34.input_layernorm.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.34.mlp.down_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.34.mlp.gate_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.34.mlp.up_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.34.post_attention_layernorm.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.34.self_attn.k_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.34.self_attn.o_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.34.self_attn.q_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.34.self_attn.v_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.35.input_layernorm.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.35.mlp.down_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.35.mlp.gate_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.35.mlp.up_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.35.post_attention_layernorm.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.35.self_attn.k_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.35.self_attn.o_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.35.self_attn.q_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.35.self_attn.v_proj.weight": "pytorch_model-00012-of-00027.bin", + "model.layers.36.input_layernorm.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.36.mlp.down_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.36.mlp.gate_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.36.mlp.up_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.36.post_attention_layernorm.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.36.self_attn.k_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.36.self_attn.o_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.36.self_attn.q_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.36.self_attn.v_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.37.input_layernorm.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.37.mlp.down_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.37.mlp.gate_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.37.mlp.up_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.37.post_attention_layernorm.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.37.self_attn.k_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.37.self_attn.o_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.37.self_attn.q_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.37.self_attn.v_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.38.input_layernorm.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.38.mlp.down_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.38.mlp.gate_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.38.mlp.up_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.38.post_attention_layernorm.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.38.self_attn.k_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.38.self_attn.o_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.38.self_attn.q_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.38.self_attn.v_proj.weight": "pytorch_model-00013-of-00027.bin", + "model.layers.39.input_layernorm.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.39.mlp.down_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.39.mlp.gate_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.39.mlp.up_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.39.post_attention_layernorm.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.39.self_attn.k_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.39.self_attn.o_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.39.self_attn.q_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.39.self_attn.v_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.4.input_layernorm.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.4.mlp.down_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.4.mlp.gate_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.4.mlp.up_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.4.post_attention_layernorm.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.4.self_attn.k_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.4.self_attn.o_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.4.self_attn.q_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.4.self_attn.v_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.40.input_layernorm.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.40.mlp.down_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.40.mlp.gate_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.40.mlp.up_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.40.post_attention_layernorm.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.40.self_attn.k_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.40.self_attn.o_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.40.self_attn.q_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.40.self_attn.v_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.41.input_layernorm.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.41.mlp.down_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.41.mlp.gate_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.41.mlp.up_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.41.post_attention_layernorm.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.41.self_attn.k_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.41.self_attn.o_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.41.self_attn.q_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.41.self_attn.v_proj.weight": "pytorch_model-00014-of-00027.bin", + "model.layers.42.input_layernorm.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.42.mlp.down_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.42.mlp.gate_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.42.mlp.up_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.42.post_attention_layernorm.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.42.self_attn.k_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.42.self_attn.o_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.42.self_attn.q_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.42.self_attn.v_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.43.input_layernorm.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.43.mlp.down_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.43.mlp.gate_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.43.mlp.up_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.43.post_attention_layernorm.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.43.self_attn.k_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.43.self_attn.o_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.43.self_attn.q_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.43.self_attn.v_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.44.input_layernorm.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.44.mlp.down_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.44.mlp.gate_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.44.mlp.up_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.44.post_attention_layernorm.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.44.self_attn.k_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.44.self_attn.o_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.44.self_attn.q_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.44.self_attn.v_proj.weight": "pytorch_model-00015-of-00027.bin", + "model.layers.45.input_layernorm.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.45.mlp.down_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.45.mlp.gate_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.45.mlp.up_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.45.post_attention_layernorm.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.45.self_attn.k_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.45.self_attn.o_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.45.self_attn.q_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.45.self_attn.v_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.46.input_layernorm.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.46.mlp.down_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.46.mlp.gate_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.46.mlp.up_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.46.post_attention_layernorm.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.46.self_attn.k_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.46.self_attn.o_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.46.self_attn.q_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.46.self_attn.v_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.47.input_layernorm.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.47.mlp.down_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.47.mlp.gate_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.47.mlp.up_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.47.post_attention_layernorm.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.47.self_attn.k_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.47.self_attn.o_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.47.self_attn.q_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.47.self_attn.v_proj.weight": "pytorch_model-00016-of-00027.bin", + "model.layers.48.input_layernorm.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.48.mlp.down_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.48.mlp.gate_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.48.mlp.up_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.48.post_attention_layernorm.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.48.self_attn.k_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.48.self_attn.o_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.48.self_attn.q_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.48.self_attn.v_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.49.input_layernorm.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.49.mlp.down_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.49.mlp.gate_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.49.mlp.up_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.49.post_attention_layernorm.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.49.self_attn.k_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.49.self_attn.o_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.49.self_attn.q_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.49.self_attn.v_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.5.input_layernorm.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.5.mlp.down_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.5.mlp.gate_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.5.mlp.up_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.5.post_attention_layernorm.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.5.self_attn.k_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.5.self_attn.o_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.5.self_attn.q_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.5.self_attn.v_proj.weight": "pytorch_model-00002-of-00027.bin", + "model.layers.50.input_layernorm.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.50.mlp.down_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.50.mlp.gate_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.50.mlp.up_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.50.post_attention_layernorm.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.50.self_attn.k_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.50.self_attn.o_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.50.self_attn.q_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.50.self_attn.v_proj.weight": "pytorch_model-00017-of-00027.bin", + "model.layers.51.input_layernorm.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.51.mlp.down_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.51.mlp.gate_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.51.mlp.up_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.51.post_attention_layernorm.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.51.self_attn.k_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.51.self_attn.o_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.51.self_attn.q_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.51.self_attn.v_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.52.input_layernorm.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.52.mlp.down_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.52.mlp.gate_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.52.mlp.up_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.52.post_attention_layernorm.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.52.self_attn.k_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.52.self_attn.o_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.52.self_attn.q_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.52.self_attn.v_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.53.input_layernorm.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.53.mlp.down_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.53.mlp.gate_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.53.mlp.up_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.53.post_attention_layernorm.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.53.self_attn.k_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.53.self_attn.o_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.53.self_attn.q_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.53.self_attn.v_proj.weight": "pytorch_model-00018-of-00027.bin", + "model.layers.54.input_layernorm.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.54.mlp.down_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.54.mlp.gate_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.54.mlp.up_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.54.post_attention_layernorm.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.54.self_attn.k_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.54.self_attn.o_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.54.self_attn.q_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.54.self_attn.v_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.55.input_layernorm.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.55.mlp.down_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.55.mlp.gate_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.55.mlp.up_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.55.post_attention_layernorm.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.55.self_attn.k_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.55.self_attn.o_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.55.self_attn.q_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.55.self_attn.v_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.56.input_layernorm.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.56.mlp.down_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.56.mlp.gate_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.56.mlp.up_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.56.post_attention_layernorm.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.56.self_attn.k_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.56.self_attn.o_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.56.self_attn.q_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.56.self_attn.v_proj.weight": "pytorch_model-00019-of-00027.bin", + "model.layers.57.input_layernorm.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.57.mlp.down_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.57.mlp.gate_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.57.mlp.up_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.57.post_attention_layernorm.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.57.self_attn.k_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.57.self_attn.o_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.57.self_attn.q_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.57.self_attn.v_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.58.input_layernorm.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.58.mlp.down_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.58.mlp.gate_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.58.mlp.up_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.58.post_attention_layernorm.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.58.self_attn.k_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.58.self_attn.o_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.58.self_attn.q_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.58.self_attn.v_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.59.input_layernorm.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.59.mlp.down_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.59.mlp.gate_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.59.mlp.up_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.59.post_attention_layernorm.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.59.self_attn.k_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.59.self_attn.o_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.59.self_attn.q_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.59.self_attn.v_proj.weight": "pytorch_model-00020-of-00027.bin", + "model.layers.6.input_layernorm.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.6.mlp.down_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.6.mlp.gate_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.6.mlp.up_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.6.post_attention_layernorm.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.6.self_attn.k_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.6.self_attn.o_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.6.self_attn.q_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.6.self_attn.v_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.60.input_layernorm.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.60.mlp.down_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.60.mlp.gate_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.60.mlp.up_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.60.post_attention_layernorm.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.60.self_attn.k_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.60.self_attn.o_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.60.self_attn.q_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.60.self_attn.v_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.61.input_layernorm.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.61.mlp.down_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.61.mlp.gate_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.61.mlp.up_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.61.post_attention_layernorm.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.61.self_attn.k_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.61.self_attn.o_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.61.self_attn.q_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.61.self_attn.v_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.62.input_layernorm.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.62.mlp.down_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.62.mlp.gate_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.62.mlp.up_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.62.post_attention_layernorm.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.62.self_attn.k_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.62.self_attn.o_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.62.self_attn.q_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.62.self_attn.v_proj.weight": "pytorch_model-00021-of-00027.bin", + "model.layers.63.input_layernorm.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.63.mlp.down_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.63.mlp.gate_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.63.mlp.up_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.63.post_attention_layernorm.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.63.self_attn.k_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.63.self_attn.o_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.63.self_attn.q_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.63.self_attn.v_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.64.input_layernorm.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.64.mlp.down_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.64.mlp.gate_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.64.mlp.up_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.64.post_attention_layernorm.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.64.self_attn.k_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.64.self_attn.o_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.64.self_attn.q_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.64.self_attn.v_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.65.input_layernorm.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.65.mlp.down_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.65.mlp.gate_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.65.mlp.up_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.65.post_attention_layernorm.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.65.self_attn.k_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.65.self_attn.o_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.65.self_attn.q_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.65.self_attn.v_proj.weight": "pytorch_model-00022-of-00027.bin", + "model.layers.66.input_layernorm.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.66.mlp.down_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.66.mlp.gate_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.66.mlp.up_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.66.post_attention_layernorm.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.66.self_attn.k_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.66.self_attn.o_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.66.self_attn.q_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.66.self_attn.v_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.67.input_layernorm.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.67.mlp.down_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.67.mlp.gate_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.67.mlp.up_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.67.post_attention_layernorm.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.67.self_attn.k_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.67.self_attn.o_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.67.self_attn.q_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.67.self_attn.v_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.68.input_layernorm.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.68.mlp.down_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.68.mlp.gate_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.68.mlp.up_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.68.post_attention_layernorm.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.68.self_attn.k_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.68.self_attn.o_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.68.self_attn.q_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.68.self_attn.v_proj.weight": "pytorch_model-00023-of-00027.bin", + "model.layers.69.input_layernorm.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.69.mlp.down_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.69.mlp.gate_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.69.mlp.up_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.69.post_attention_layernorm.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.69.self_attn.k_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.69.self_attn.o_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.69.self_attn.q_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.69.self_attn.v_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.7.input_layernorm.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.7.mlp.down_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.7.mlp.gate_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.7.mlp.up_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.7.post_attention_layernorm.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.7.self_attn.k_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.7.self_attn.o_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.7.self_attn.q_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.7.self_attn.v_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.70.input_layernorm.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.70.mlp.down_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.70.mlp.gate_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.70.mlp.up_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.70.post_attention_layernorm.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.70.self_attn.k_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.70.self_attn.o_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.70.self_attn.q_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.70.self_attn.v_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.71.input_layernorm.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.71.mlp.down_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.71.mlp.gate_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.71.mlp.up_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.71.post_attention_layernorm.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.71.self_attn.k_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.71.self_attn.o_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.71.self_attn.q_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.71.self_attn.v_proj.weight": "pytorch_model-00024-of-00027.bin", + "model.layers.72.input_layernorm.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.72.mlp.down_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.72.mlp.gate_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.72.mlp.up_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.72.post_attention_layernorm.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.72.self_attn.k_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.72.self_attn.o_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.72.self_attn.q_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.72.self_attn.v_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.73.input_layernorm.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.73.mlp.down_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.73.mlp.gate_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.73.mlp.up_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.73.post_attention_layernorm.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.73.self_attn.k_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.73.self_attn.o_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.73.self_attn.q_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.73.self_attn.v_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.74.input_layernorm.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.74.mlp.down_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.74.mlp.gate_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.74.mlp.up_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.74.post_attention_layernorm.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.74.self_attn.k_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.74.self_attn.o_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.74.self_attn.q_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.74.self_attn.v_proj.weight": "pytorch_model-00025-of-00027.bin", + "model.layers.75.input_layernorm.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.75.mlp.down_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.75.mlp.gate_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.75.mlp.up_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.75.post_attention_layernorm.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.75.self_attn.k_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.75.self_attn.o_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.75.self_attn.q_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.75.self_attn.v_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.76.input_layernorm.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.76.mlp.down_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.76.mlp.gate_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.76.mlp.up_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.76.post_attention_layernorm.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.76.self_attn.k_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.76.self_attn.o_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.76.self_attn.q_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.76.self_attn.v_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.77.input_layernorm.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.77.mlp.down_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.77.mlp.gate_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.77.mlp.up_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.77.post_attention_layernorm.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.77.self_attn.k_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.77.self_attn.o_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.77.self_attn.q_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.77.self_attn.v_proj.weight": "pytorch_model-00026-of-00027.bin", + "model.layers.78.input_layernorm.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.78.mlp.down_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.78.mlp.gate_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.78.mlp.up_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.78.post_attention_layernorm.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.78.self_attn.k_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.78.self_attn.o_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.78.self_attn.q_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.78.self_attn.v_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.79.input_layernorm.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.79.mlp.down_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.79.mlp.gate_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.79.mlp.up_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.79.post_attention_layernorm.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.79.self_attn.k_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.79.self_attn.o_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.79.self_attn.q_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.79.self_attn.v_proj.weight": "pytorch_model-00027-of-00027.bin", + "model.layers.8.input_layernorm.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.8.mlp.down_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.8.mlp.gate_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.8.mlp.up_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.8.post_attention_layernorm.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.8.self_attn.k_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.8.self_attn.o_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.8.self_attn.q_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.8.self_attn.v_proj.weight": "pytorch_model-00003-of-00027.bin", + "model.layers.9.input_layernorm.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.9.mlp.down_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.9.mlp.gate_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.9.mlp.up_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.9.post_attention_layernorm.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.9.self_attn.k_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.9.self_attn.o_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.9.self_attn.q_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.layers.9.self_attn.v_proj.weight": "pytorch_model-00004-of-00027.bin", + "model.norm.weight": "pytorch_model-00027-of-00027.bin" + } +}