diff --git a/config.json b/config.json new file mode 100644 index 0000000000000000000000000000000000000000..f362a02b36eb5bc805dfc9e7dac11f481ab00915 --- /dev/null +++ b/config.json @@ -0,0 +1,29 @@ +{ + "_name_or_path": "./merged3", + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": 1, + "eos_token_id": 2, + "hidden_act": "silu", + "hidden_size": 8192, + "initializer_range": 0.02, + "intermediate_size": 28672, + "max_position_embeddings": 4096, + "model_type": "llama", + "num_attention_heads": 64, + "num_hidden_layers": 140, + "num_key_value_heads": 8, + "pad_token_id": 0, + "pretraining_tp": 1, + "rms_norm_eps": 1e-05, + "rope_scaling": null, + "rope_theta": 10000.0, + "tie_word_embeddings": false, + "torch_dtype": "float32", + "transformers_version": "4.36.2", + "use_cache": false, + "vocab_size": 32000 +} diff --git a/generation_config.json b/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..5786877b25c8295c74590b8a5aace96539c44dd2 --- /dev/null +++ b/generation_config.json @@ -0,0 +1,8 @@ +{ + "_from_model_config": true, + "bos_token_id": 1, + "eos_token_id": 2, + "pad_token_id": 0, + "transformers_version": "4.36.2", + "use_cache": false +} diff --git a/model-00001-of-00106.safetensors b/model-00001-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..da30a715e3c4fc37f2669ef50e3c55032b11b5f2 --- /dev/null +++ b/model-00001-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5e3d02ed1178dcb478b46bfaafcfa15f479525264384c4abd269543192973f4e +size 4806739440 diff --git a/model-00002-of-00106.safetensors b/model-00002-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e71b60efeb254a7f5b49ad821f0f384fa3871383 --- /dev/null +++ b/model-00002-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c86dfdd3d2cd005d6317a62ed780925fc0f8f0c4d891d051ea3a2faf76772ef6 +size 4630578440 diff --git a/model-00003-of-00106.safetensors b/model-00003-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..eacc2617a8544d68dfb50f9c3cee3c3979dd389e --- /dev/null +++ b/model-00003-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c68b3ac8eb5b33ad7e3f95dc4e2ed6aa957d1f744c72e99516d8f6d64a6e427e +size 4362142864 diff --git a/model-00004-of-00106.safetensors b/model-00004-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4d2850838cf9ccb07af21d68ac9eed186effb488 --- /dev/null +++ b/model-00004-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e70647db6a35e43000cf15380bf695d07d3fad5c21b37f40e0c734c43b903d5 +size 4966188864 diff --git a/model-00005-of-00106.safetensors b/model-00005-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2527dbb5ffafff9a0a3b3069673e423f48b70afe --- /dev/null +++ b/model-00005-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:97fe33ff1b3415106e59e0def9a8fa44b1835d33ae7292825e9e3022cee021d2 +size 4362142864 diff --git a/model-00006-of-00106.safetensors b/model-00006-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..102ad8c9e73da77cae099a0e98da05da06a76ba7 --- /dev/null +++ b/model-00006-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c5ac33ed4b0c30dd2b1e63b31b30f8a97dadae2ad5004bd38d28c4f6918e3e6c +size 4362142864 diff --git a/model-00007-of-00106.safetensors b/model-00007-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f748763627132734c2268324a3cad451f2b42a07 --- /dev/null +++ b/model-00007-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2291a5e7ec507dda1db8ba0a3c0283974cf23c6bca5bc651d3fd3a27182f09d4 +size 4966188864 diff --git a/model-00008-of-00106.safetensors b/model-00008-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9d89a4f15a115bd351a9f654e872897d5f2990db --- /dev/null +++ b/model-00008-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4d8bce52a82b0aada352eb0c919bdd1e515d599bf008ac4645b29651da7c8739 +size 4362142880 diff --git a/model-00009-of-00106.safetensors b/model-00009-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d0c25d47286ad7f9f38cf0f3454b135db9d37f8d --- /dev/null +++ b/model-00009-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0a791e5ea49f14d30d9640a29ff65a82b6c6880d008136d2509cff3c4abfe722 +size 4362142872 diff --git a/model-00010-of-00106.safetensors b/model-00010-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fa292178dc53c1b530fd6718e5b91371b14a0ad2 --- /dev/null +++ b/model-00010-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8094e42cc998daa13a13dea3e0e811a8398074d43d06a747f2722e98b6e0da01 +size 4966188880 diff --git a/model-00011-of-00106.safetensors b/model-00011-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0db0a33f4aa2ae1e7e6824c03dcff43c52d56c2c --- /dev/null +++ b/model-00011-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:462539a05f8ccc64d3468d6f83c5a1bd4180a0b88816b750243dcf87cc10718a +size 4362142872 diff --git a/model-00012-of-00106.safetensors b/model-00012-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dfa7b3376082e14f23a231d10e2307487b795189 --- /dev/null +++ b/model-00012-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9921a70ce4cd1c68935afad9a1dafab6bcea1caf0b8b028a99bf9b33a7dfa226 +size 4362142872 diff --git a/model-00013-of-00106.safetensors b/model-00013-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2c375eefa48e39a9113b3aab5d04ebc4cc743ce3 --- /dev/null +++ b/model-00013-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d270424f967098636d380c77df0f67476b7f6539897b1677282f17f883d79b21 +size 4966188880 diff --git a/model-00014-of-00106.safetensors b/model-00014-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..20b52783be4306cc576afd4253cf0ffc4c6527e1 --- /dev/null +++ b/model-00014-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9260cde85a374cf62f1c4507baf02d7423669e0c466f988bad74fcb44157fdca +size 4362142872 diff --git a/model-00015-of-00106.safetensors b/model-00015-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9e682675a7d19ef7511030ccd0a7a7c60f31d3df --- /dev/null +++ b/model-00015-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c579b083ca734d4fea6a399ae413928162f723cf3479311ba64045677a22987e +size 4362142872 diff --git a/model-00016-of-00106.safetensors b/model-00016-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..baa674770b19380397d878ea13a1fd354dcc58fa --- /dev/null +++ b/model-00016-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:90d0eefc9cb11eba17d2937f1a3794394b717d8b111eb35533cac914d3f5e83c +size 4966188880 diff --git a/model-00017-of-00106.safetensors b/model-00017-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e37a51ff4f73db115f074c644c3d056668d35afd --- /dev/null +++ b/model-00017-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:453462f72541378743e47ecf4e07ac5102c217c364494b1de0b4dca46ef75e9d +size 4362142872 diff --git a/model-00018-of-00106.safetensors b/model-00018-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..135f2fd7934d1a77464f7bad3b12743040b46c09 --- /dev/null +++ b/model-00018-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e53b08f4b761076b6aa413e43e5b988b765cf0d46d45c49371e110f13f5f037b +size 4362142872 diff --git a/model-00019-of-00106.safetensors b/model-00019-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..08531a0efdeedd6a20a956ec6c585dab28d066ee --- /dev/null +++ b/model-00019-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:65096c7783b957e40a03daeeb079a642f4c7ab62e3df92d07fca6a96bdf4861d +size 4966188880 diff --git a/model-00020-of-00106.safetensors b/model-00020-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1feb13d56bb25fa5ec67c8d0a9cf6c72f66b3fa6 --- /dev/null +++ b/model-00020-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:806a9eb3da6a1882f5a43b86a2eab21a4592c987dbbe99df64e09f522925fbe2 +size 4362142872 diff --git a/model-00021-of-00106.safetensors b/model-00021-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b9aa7b2c0e8b3391930d1d4c66e6bddc75aa26df --- /dev/null +++ b/model-00021-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8ee3c8546a50a9cecf187451590d5b1aca7009a5bd9364cfa87a5b8ddfdb4022 +size 4362142872 diff --git a/model-00022-of-00106.safetensors b/model-00022-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..da34b9246f2b7196602bb23401de519b8b7e057f --- /dev/null +++ b/model-00022-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b7c148533588907213221a276b4d673af8a97b94f5786a6fb4c038cbc39f701c +size 4966188880 diff --git a/model-00023-of-00106.safetensors b/model-00023-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bc59b51244996178401af7c3f2deae01e5a5146c --- /dev/null +++ b/model-00023-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2d322bb8653aa4929901b9119dfa0fdcd6633b05d0741b48213c6612e57b76b7 +size 4362142872 diff --git a/model-00024-of-00106.safetensors b/model-00024-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7d0f9a1ea11399252e0a2771db3d3297d92998bf --- /dev/null +++ b/model-00024-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3fb7fee679e4b646fce7df478858f7cb7570357ac81ad946564158b2833db3b1 +size 4362142872 diff --git a/model-00025-of-00106.safetensors b/model-00025-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..567826a022ea5a852f880ceb51d6e6f9cc6fdf8f --- /dev/null +++ b/model-00025-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e2ad5ee15cee34e13106d7addc028405e72ce167032f88efbea995877017118c +size 4966188880 diff --git a/model-00026-of-00106.safetensors b/model-00026-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c836b46a2a1ba853096f3dd4c6df9a77f10ba26d --- /dev/null +++ b/model-00026-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7f5c5ad57fb921e16d18be91deb9a01f9ed89cc515af0be9dcd13f3c0946e7c2 +size 4362142872 diff --git a/model-00027-of-00106.safetensors b/model-00027-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..34884225cf0259e079841a48a312c827742828e4 --- /dev/null +++ b/model-00027-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1861787259cdb9cf0dbb1cbc2dd907567655ab7dd22e146306ce1fde51ef09a1 +size 4362142872 diff --git a/model-00028-of-00106.safetensors b/model-00028-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..583cfb0d196ac88e4dc08fb4d07474d72b5cba0a --- /dev/null +++ b/model-00028-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:919ac271a48742cd63d8e334d4639a2b6affaa8abf1230af44d0ae98928baf7c +size 4966188880 diff --git a/model-00029-of-00106.safetensors b/model-00029-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5ddba782a8381ab9fb5f0bd55e16f6e613b2aa8a --- /dev/null +++ b/model-00029-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:68f62e4e0b70a2b6c0c8dc0821b9640e2058eadd2518061b70e31af35e2c677d +size 4362142872 diff --git a/model-00030-of-00106.safetensors b/model-00030-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c0a00ccb1627b60cb12d7bf3ffd7c904cb6b9418 --- /dev/null +++ b/model-00030-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:65fa1210cf65d43c47df0d41188e276190fb3b8f1656654fc01225c9e68a54f1 +size 4362142872 diff --git a/model-00031-of-00106.safetensors b/model-00031-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..580c3e705ae91397c7928c3ba2b63ce9d12bfdea --- /dev/null +++ b/model-00031-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1477296b0c43aa6f94b6419cf648d905695f3b02cca451f6760e88f63cd92f7b +size 4966188880 diff --git a/model-00032-of-00106.safetensors b/model-00032-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f68e5923a739bdd863ad4f4011a1fe65bc72d4d1 --- /dev/null +++ b/model-00032-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d527dce5cddc274ac42d91f883b894624d01fa682792bf6901648b8d2d973e59 +size 4362142872 diff --git a/model-00033-of-00106.safetensors b/model-00033-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..412e2af011279c7ab8106fb22b118fccf8560f85 --- /dev/null +++ b/model-00033-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4e54efdcf7bec565248080ccbda1629aeb0d2e4085985d996fd622e4fefb83e5 +size 4362142872 diff --git a/model-00034-of-00106.safetensors b/model-00034-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ca7ff8346a09e099234b8c3cb4fe8d5c2518fbbe --- /dev/null +++ b/model-00034-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:12eff68e1ab5553881a1129f5921dc53c1131bf2ae3a335b989ccddb2ab032bd +size 4966188880 diff --git a/model-00035-of-00106.safetensors b/model-00035-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f9ae69a50326650c9c8e779dcadbb1620439a73e --- /dev/null +++ b/model-00035-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:22c7ebabe895a7ca100fd97d6f339b7ab5ca67c4fed03a3c69659bd6a31b0747 +size 4362142872 diff --git a/model-00036-of-00106.safetensors b/model-00036-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..023a760849dd4637e80e4438137d6e61f9ab10bd --- /dev/null +++ b/model-00036-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3963ca58c8343fd096240ab13c9c6a044d1f8f63518ae00eaacd4b49d35cfb66 +size 4362142872 diff --git a/model-00037-of-00106.safetensors b/model-00037-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5f877a20589b59033e3e4514b105f3102d163c6d --- /dev/null +++ b/model-00037-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cf995ee0ccd594e56e321f6527c35d6022b6613fd0244d46299d8484296cd358 +size 4966188880 diff --git a/model-00038-of-00106.safetensors b/model-00038-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..21ffe666adecb24cf065112fe296b3fb15962eff --- /dev/null +++ b/model-00038-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab2f4957ba21721580beab6aad3d05f8a113198943fac833e66bf38c555265bb +size 4362142872 diff --git a/model-00039-of-00106.safetensors b/model-00039-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..13321e5ffc224816aa036aa06bb203ef4707dd2e --- /dev/null +++ b/model-00039-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:82fb192e9cafc7e970dc44f6f3b639414113068b7172e3981708a7027537db34 +size 4362142872 diff --git a/model-00040-of-00106.safetensors b/model-00040-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dea64b06fc726cfa96588a1a9d2f3b7624c75633 --- /dev/null +++ b/model-00040-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff283f4e5f9c8da8a8c43023cce34503f6fad46e7a4e38b53bf92c5be96819da +size 4966188880 diff --git a/model-00041-of-00106.safetensors b/model-00041-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..af5790006b206c89d32d35a37a630a07c35d6fda --- /dev/null +++ b/model-00041-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3234a503a9f6bf7e1383f2aab2ea7657ba6da313ec27412c8de6f209fc6314fa +size 4362142872 diff --git a/model-00042-of-00106.safetensors b/model-00042-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..baf922c251102ea217660b3758116984588d88c5 --- /dev/null +++ b/model-00042-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a4634579c66f12c0eadf748242e5fd02d5668f3ff0cb757b377d10eecafa63c5 +size 4362142872 diff --git a/model-00043-of-00106.safetensors b/model-00043-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..860f80f50f251f755c1c0899a3442101fdc2fb84 --- /dev/null +++ b/model-00043-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd898142c17a32a31ac8c1af864937596bba7f085452e3205243d208f04258fb +size 4966188880 diff --git a/model-00044-of-00106.safetensors b/model-00044-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..adab94c31675497ae5ab621bdbb6a585725d6f7c --- /dev/null +++ b/model-00044-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f4c8e07789b06e9648062dcf04f7d3202818692cb042c35e8f3f3306cf913276 +size 4362142872 diff --git a/model-00045-of-00106.safetensors b/model-00045-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c2fe8778c740cae41fdd0897a8edce907213027a --- /dev/null +++ b/model-00045-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec51be1731eaa80ed1dfb08746e5c1ee7c4f7f51a1ab155333d85a676018db1f +size 4362142872 diff --git a/model-00046-of-00106.safetensors b/model-00046-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d88980937465167b2893950fa8721c5f873fd726 --- /dev/null +++ b/model-00046-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f426cc6d0dd49bf68eeabe99bc21a89f77fd630acb1ffc0d9baffc70798ee5eb +size 4966188880 diff --git a/model-00047-of-00106.safetensors b/model-00047-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..65f3eed1960663eb1ea666662a2d38977f4549ea --- /dev/null +++ b/model-00047-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:63faef73302257485978506e6108c4084e086c198b751bc9b609e0ea1f7f046d +size 4362142872 diff --git a/model-00048-of-00106.safetensors b/model-00048-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..78128a4b129f692f16a10cd7725c56cda27e55a6 --- /dev/null +++ b/model-00048-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a7974ed8f4bd10765391d6c0592e23784bad0cc115a129b6cbd046bb71f2433b +size 4362142872 diff --git a/model-00049-of-00106.safetensors b/model-00049-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f47fc7b99444079d6c479c2cca92deff03a6e06f --- /dev/null +++ b/model-00049-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fc29fb22488c2bce8fc8d6284ade29af60425891612f3a94aeaeef9486de8d08 +size 4966188880 diff --git a/model-00050-of-00106.safetensors b/model-00050-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d2ac9dcaa2ffa2febd1a3ad890a65fc781b01bb0 --- /dev/null +++ b/model-00050-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ece4b6c99494f37262eec7817e8a7464ad0e20d51bdad5d6fb57a2444c1cef84 +size 4362142872 diff --git a/model-00051-of-00106.safetensors b/model-00051-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4c534ee13b91e74354f244f0aa093ce100802029 --- /dev/null +++ b/model-00051-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:325449209c55776c5bb550465ca793a0bb0721f845500ea509219ac638c0f81d +size 4362142872 diff --git a/model-00052-of-00106.safetensors b/model-00052-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fc6a031b57f6e5ecb28ce7eb6c9b555eca0ee595 --- /dev/null +++ b/model-00052-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:46b29f8cee97ae1e59c8e399d07078a157fcee58128d49a21e35035fc41f54d9 +size 4966188880 diff --git a/model-00053-of-00106.safetensors b/model-00053-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6533ec0f2fa90423f0880beb239073275dbe5582 --- /dev/null +++ b/model-00053-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e667965de75bbcc46a3b41e3826815c006819cbed20d3caa50858d6dc4118a97 +size 4362142872 diff --git a/model-00054-of-00106.safetensors b/model-00054-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5c5ed5537799b397026b713e7e29c22f7fcdd162 --- /dev/null +++ b/model-00054-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:961191e353b50663e752f15cac27f23b239de0c3a5c484208b5bc1247662d64c +size 4362142872 diff --git a/model-00055-of-00106.safetensors b/model-00055-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..72be2a1d9d8d4587b1e44e5ec33e65b5e09ec8e9 --- /dev/null +++ b/model-00055-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:65a6f795838c66a55bfad713d2c763035f4d446c38fba29d63fb5283e2e65149 +size 4966188880 diff --git a/model-00056-of-00106.safetensors b/model-00056-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f4c6aff1bb19a597cab21b462ea4aac881397e5c --- /dev/null +++ b/model-00056-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:489074e753b1df7a20f7c0a6b9a8ac7363102f9c34d3737be1db2b2bfba14e2f +size 4362142872 diff --git a/model-00057-of-00106.safetensors b/model-00057-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8dfa87822009b51e0e8045368bdee0742db33ec2 --- /dev/null +++ b/model-00057-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:45e5de7a9ed11699eacf15fa56da1949c960155bb6ddba6c4f7c831022f29bfc +size 4362142872 diff --git a/model-00058-of-00106.safetensors b/model-00058-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..01c88c448df6bf8913a9be1a0ba7f0a0e5bd1440 --- /dev/null +++ b/model-00058-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4e0ba017f5f93c5d9409db93cec065d981571f55492850a07f4b57f1226c3488 +size 4966188880 diff --git a/model-00059-of-00106.safetensors b/model-00059-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9e77a435f6746ef5f965ff4642a7ca767b6abac5 --- /dev/null +++ b/model-00059-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4abd3114b542a0f1d436b78096d80f14632372e692ab397f81979ab55d7f4670 +size 4362142872 diff --git a/model-00060-of-00106.safetensors b/model-00060-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9b37b333050453e29545935e979e6298a8d738bd --- /dev/null +++ b/model-00060-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6818eb12b2a534a310c701901d5f9fb0e015978f34c244ebb417717a166e7cd1 +size 4362142872 diff --git a/model-00061-of-00106.safetensors b/model-00061-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..91f84354f901ba529cc4c1722f15caeb35e224f2 --- /dev/null +++ b/model-00061-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a94411ea38ffec9a3e6200e5bcba31ff37fe18896ccc2624b17b17f4d2feae40 +size 4966188880 diff --git a/model-00062-of-00106.safetensors b/model-00062-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0a8a2caea67a34d262bb29db3f3c862374750b68 --- /dev/null +++ b/model-00062-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd454647be722f287c1e9699c028a5a4f72beeba9c4e71052a756da43874349a +size 4362142872 diff --git a/model-00063-of-00106.safetensors b/model-00063-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..784849ff4a55269728b52fa152ff4b47bfc2bf83 --- /dev/null +++ b/model-00063-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a990dab8294b458eefd6093af87f88e20b8fad4103d5e6d07d5ab4ff61dc65cb +size 4362142872 diff --git a/model-00064-of-00106.safetensors b/model-00064-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1c0f5b0cbeefb66633385e12abb756426c2b049b --- /dev/null +++ b/model-00064-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:71957d14c8d02c3a9ff7e56bcd6f4dcae6f816d90fe74bb81611bde06d6a3171 +size 4966188880 diff --git a/model-00065-of-00106.safetensors b/model-00065-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fc82cbe2032a8a9d95ad820adb6ff50d96d18fe5 --- /dev/null +++ b/model-00065-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb4b2bb78758e4a9b48fe645343c9fb154463edb49358f4d5917da2c04757b00 +size 4362142872 diff --git a/model-00066-of-00106.safetensors b/model-00066-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..eb6c7561734e6c008aac2b13a33299e886b9638e --- /dev/null +++ b/model-00066-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:211c7eae19e5edc340d139ef58003b55ab6db5dc426411c962553a7608bdcfa2 +size 4362142872 diff --git a/model-00067-of-00106.safetensors b/model-00067-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..72aabb1b37dd12fff0921fab54b0dcd0433badb9 --- /dev/null +++ b/model-00067-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:627450074e9646600bcd911c8c6c103a375a8bfb3a34260af31ea6bc2461b535 +size 4966188880 diff --git a/model-00068-of-00106.safetensors b/model-00068-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d21ef958d3a628df4cb595e55dc43f26e95b62ab --- /dev/null +++ b/model-00068-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b80127eb68600af927d27790e497162f35096a4bce28d3460ce80292d6e01e36 +size 4362142872 diff --git a/model-00069-of-00106.safetensors b/model-00069-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dd5907c97cfc75fc7661adf2f282589af08c8eb4 --- /dev/null +++ b/model-00069-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae8312c964e8984826106530e38dc62a71bff0b2f2fb5c1c4e3ce1a2b84c3925 +size 4362142872 diff --git a/model-00070-of-00106.safetensors b/model-00070-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e4e581bc9c0befe275685d8747d817dd8e7016d9 --- /dev/null +++ b/model-00070-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aa322e0f8bb9f978e7c9084c26105207332380c03a63dcebaa5ae8017816eb01 +size 4966188880 diff --git a/model-00071-of-00106.safetensors b/model-00071-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a86f54dd6fd974e1c35ee1977edb8a0895441ac1 --- /dev/null +++ b/model-00071-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:62b7a76792f5d96ef21abc2774dadc85c9d3d4d6b59527b8ebfa02c7528d4905 +size 4362142872 diff --git a/model-00072-of-00106.safetensors b/model-00072-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ec6adac409eb66d423de0891df71b10399cf37cd --- /dev/null +++ b/model-00072-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2f00a3f92639a08360941e63de3da9a3d196b060e70ccd3723c1b20f506cc280 +size 4362142872 diff --git a/model-00073-of-00106.safetensors b/model-00073-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ae7add7222f7d4f30f17af3a45c2990578110e30 --- /dev/null +++ b/model-00073-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7f770a0dfb5ba9e1e99607a4812a31383c9c2f809da4091b569e29d4aa60182c +size 4966188880 diff --git a/model-00074-of-00106.safetensors b/model-00074-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..aa3257848696973bbf505a435288988f5f8cdc71 --- /dev/null +++ b/model-00074-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:81686f6c4ccc80614be07b3f915ca8fd40c4e0439f4d5d345c3830327ac0c3c4 +size 4362142872 diff --git a/model-00075-of-00106.safetensors b/model-00075-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..803185bbf9b34e97d3fc44f5c82b26b1bc382a64 --- /dev/null +++ b/model-00075-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e0902b9574c63bf5afb7ce591b7132535fdcdcc6f20ecbfcff9201b9780982e2 +size 4362142872 diff --git a/model-00076-of-00106.safetensors b/model-00076-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..521a579498e9865d2b7e67a65c97eb7df6c9fc6c --- /dev/null +++ b/model-00076-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a409696e20022596c4839274790e74ff3d7bf83a3cd356934b0b3f35a4b8b91 +size 4966188896 diff --git a/model-00077-of-00106.safetensors b/model-00077-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..576262346bdbbc69465a2bdbbdf906e3b8316689 --- /dev/null +++ b/model-00077-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:48d31b8361eea78b507ac4cae1b8bcbe5b49589482fe16f0bd7d315d4f16493a +size 4362142888 diff --git a/model-00078-of-00106.safetensors b/model-00078-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3db98ea9b0871a6d59d79e38687ee9beded73f13 --- /dev/null +++ b/model-00078-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fee0351ca17f70807c642369e81240040bcdc4d3b2f525c43e189c2efd3f8592 +size 4362142880 diff --git a/model-00079-of-00106.safetensors b/model-00079-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..752012093fb80bb0dffcf87cf0d8d6e4e21c4e5b --- /dev/null +++ b/model-00079-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6737f924764eb1656421277e14dfbf26dc77cdfebe879ca24feb3c0f51d1707c +size 4966188896 diff --git a/model-00080-of-00106.safetensors b/model-00080-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6cafec7ec138ccd138a5ab9baf2ea0dfca5fcf86 --- /dev/null +++ b/model-00080-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:df673e472f6ee718fe2f66539f876edeeccdd9c10afc9799e2d3e186bd5fc31d +size 4362142888 diff --git a/model-00081-of-00106.safetensors b/model-00081-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..410eba3f2e55f3b44638a690a8413ff5bd3de7d8 --- /dev/null +++ b/model-00081-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:259cbf93038a8b66ab64363418edac336bc8872320e3047fcb8f7b3236c32736 +size 4362142880 diff --git a/model-00082-of-00106.safetensors b/model-00082-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3e0a823d9408d02a0685fa62928e1150aeeb2542 --- /dev/null +++ b/model-00082-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9e819b6ae1dd98f24b4ac9ace411f32addf4c39feb49736daf3a3f60078c3a48 +size 4966188896 diff --git a/model-00083-of-00106.safetensors b/model-00083-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..35f9dedf8f1eaa6b9436e79763075c41f3586fa6 --- /dev/null +++ b/model-00083-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ad2ce5f14640917b7a2c7d89b72887f853a80c06167ca686aa5b949733a4b697 +size 4362142888 diff --git a/model-00084-of-00106.safetensors b/model-00084-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2531b01ed0fabd3755274c423981a5716a5724b9 --- /dev/null +++ b/model-00084-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2317b4cc0408b4d935ba8b68ab5821b90c78be432d440831b8029f135a862cac +size 4362142880 diff --git a/model-00085-of-00106.safetensors b/model-00085-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..502818d8e715071343fb9c9f143976891732e6a8 --- /dev/null +++ b/model-00085-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1e3ac1115f1101ef259a06614ea02ef0e409a37272c82f8693b4c4f8485a2d23 +size 4966188896 diff --git a/model-00086-of-00106.safetensors b/model-00086-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e517ef0086cf8502d3f3a30da65eeead4373b303 --- /dev/null +++ b/model-00086-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:28b777a04d69a57b90f89255df5514f9fef6193db7c4e8303f4ecdf2c3e9fa48 +size 4362142888 diff --git a/model-00087-of-00106.safetensors b/model-00087-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cb13b72acf63b882355f5835b6816ab28397c8ff --- /dev/null +++ b/model-00087-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3715878463a9a3b29da4bb4ee0983c0cf911e5276cd0ef730c1af0578ced13f1 +size 4362142880 diff --git a/model-00088-of-00106.safetensors b/model-00088-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f1ffe9e2efb30653b533b1f56b751b8ab57b9158 --- /dev/null +++ b/model-00088-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f150cecc6cabd7e124eae062f7fe860da6d97a93ccaac87f1486a0d0f45eac1a +size 4966188896 diff --git a/model-00089-of-00106.safetensors b/model-00089-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..881f718e6fcca85b04f3bb9fe67264348da4921c --- /dev/null +++ b/model-00089-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:98034f9beec33eeec5f351012e3c314601931a50bd444852ac522d49c7c3743b +size 4362142888 diff --git a/model-00090-of-00106.safetensors b/model-00090-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e6ae269520353f34a66e24985d0aff3b6f42d312 --- /dev/null +++ b/model-00090-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:253dac8be88aaa84712a457432c89e47c7e85c299e5440bb9910fe4a18a7e105 +size 4362142880 diff --git a/model-00091-of-00106.safetensors b/model-00091-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4d4f0ec90188039340da8e8e8f1188ecc7f243de --- /dev/null +++ b/model-00091-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a8bdd131fb6b255da326f100a076a675c937c43aa2e47b8bffbf67c61035b32 +size 4966188896 diff --git a/model-00092-of-00106.safetensors b/model-00092-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ffc1900c2ac57d87d0f49b8a97a0853572ced1de --- /dev/null +++ b/model-00092-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d19d866f746e89f5eaaa2e9808ca8365cfede0eabe38fbaa85b6ff03c9e62068 +size 4362142888 diff --git a/model-00093-of-00106.safetensors b/model-00093-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..78be3a008f9c89364caa17bd22d2b7f0caadfb12 --- /dev/null +++ b/model-00093-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b76288dfca56c10928708a7ea5e6b9714e90af17a3398712d6de56553bfb3cf +size 4362142880 diff --git a/model-00094-of-00106.safetensors b/model-00094-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..45b59363b44c2e60fc90bdb255c7edca142cb6ff --- /dev/null +++ b/model-00094-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:259b7fd884bd2af86adf0105a7b359e0133319a6048bfabf4283c6eb84074e4f +size 4966188896 diff --git a/model-00095-of-00106.safetensors b/model-00095-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..de763fd7a50887c142a4100f5fa742f774b2b096 --- /dev/null +++ b/model-00095-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c121375461f85483feb83edcb4feea9ea3fe401441f09d62184eaab077972002 +size 4362142888 diff --git a/model-00096-of-00106.safetensors b/model-00096-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d2fa38f884ecb9007d5fd32828db9ba6d5a6d7d7 --- /dev/null +++ b/model-00096-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:071bf7013bed8300bf90e4ad7b35fc64f3ca6e511d61d47c918c8ff9626b9e40 +size 4362142880 diff --git a/model-00097-of-00106.safetensors b/model-00097-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c77635f3c65b0e9cd64afb363d688e393348dba3 --- /dev/null +++ b/model-00097-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e785f352b0a84e9c3813b75b2b9d25e0dea886b3e32144c500102f46f43399bf +size 4966188896 diff --git a/model-00098-of-00106.safetensors b/model-00098-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8fa509004c7275207bf9354b345a160f815ee23c --- /dev/null +++ b/model-00098-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6618d04851e9b57ca0e763bdd22a18ee3f4c648013d143af06493f86fb711c85 +size 4362142888 diff --git a/model-00099-of-00106.safetensors b/model-00099-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0b3f22ec561bac15a6a5cc1cc2e87bc0677d9414 --- /dev/null +++ b/model-00099-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:40fbc154ea9c76ea69ef47733680e7fd513b52308c7d1476b527141bb5afd802 +size 4362142880 diff --git a/model-00100-of-00106.safetensors b/model-00100-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..66d6534e84398e0bb130e1da1bb22457a82dd02d --- /dev/null +++ b/model-00100-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:38613948191fa636ea2d4355039b4da227fc774045a30e19de3de5d4f0adde78 +size 4966188896 diff --git a/model-00101-of-00106.safetensors b/model-00101-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3eca31190805e69ecbf743d5a6d7394859e3cfb9 --- /dev/null +++ b/model-00101-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6f19ec01901a83f67c0566fb0bf9f18b285b9e870be5978ccd51daf87c1dcbee +size 4362142888 diff --git a/model-00102-of-00106.safetensors b/model-00102-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a31570d5425cc672211d29e5ff325a1fdd4125bb --- /dev/null +++ b/model-00102-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5296b8d72f65f20b7329127094cd6a877269b5f6b6093b7e9a33ffbb359ca66b +size 4362142880 diff --git a/model-00103-of-00106.safetensors b/model-00103-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b5ad96704e19b8d1c0c88ba00ffff0afc4a16345 --- /dev/null +++ b/model-00103-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:19e77e0709c38438955963a4fa24971c5998fed01de70d9fb5688807f2be0acf +size 4966188896 diff --git a/model-00104-of-00106.safetensors b/model-00104-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cd1173b9e1f10eda5f48216e624078a9c82b850d --- /dev/null +++ b/model-00104-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:22b7768c709d0658e5b33930ba9ebf2c9eb2d91815e9375f75108582dfe879bd +size 4362142888 diff --git a/model-00105-of-00106.safetensors b/model-00105-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5577ad520136ba41349da594466997155e3ce8ea --- /dev/null +++ b/model-00105-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7167570648b646407b34907d27e7ca19654f664a3c4e95e5d5e854bbb8bb624 +size 4362142880 diff --git a/model-00106-of-00106.safetensors b/model-00106-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..53e98cdea10876523b66e972dac9cea1b402e636 --- /dev/null +++ b/model-00106-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b3bbf6a33ad269241006dffb23030b74fd8d5c05b99adb79652d266e624ceaa +size 1988198968 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..4205262bd427465e4a13b0284bd583c3db9c93cb --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,1270 @@ +{ + "metadata": { + "total_size": 481263648768 + }, + "weight_map": { + "lm_head.weight": "model-00106-of-00106.safetensors", + "model.embed_tokens.weight": "model-00001-of-00106.safetensors", + "model.layers.0.input_layernorm.weight": "model-00001-of-00106.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00106.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.1.input_layernorm.weight": "model-00002-of-00106.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00002-of-00106.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.10.input_layernorm.weight": "model-00009-of-00106.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00009-of-00106.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.100.input_layernorm.weight": "model-00076-of-00106.safetensors", + "model.layers.100.mlp.down_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.100.mlp.gate_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.100.mlp.up_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.100.post_attention_layernorm.weight": "model-00076-of-00106.safetensors", + "model.layers.100.self_attn.k_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.100.self_attn.o_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.100.self_attn.q_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.100.self_attn.v_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.101.input_layernorm.weight": "model-00077-of-00106.safetensors", + "model.layers.101.mlp.down_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.101.mlp.gate_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.101.mlp.up_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.101.post_attention_layernorm.weight": "model-00077-of-00106.safetensors", + "model.layers.101.self_attn.k_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.101.self_attn.o_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.101.self_attn.q_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.101.self_attn.v_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.102.input_layernorm.weight": "model-00078-of-00106.safetensors", + "model.layers.102.mlp.down_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.102.mlp.gate_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.102.mlp.up_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.102.post_attention_layernorm.weight": "model-00078-of-00106.safetensors", + "model.layers.102.self_attn.k_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.102.self_attn.o_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.102.self_attn.q_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.102.self_attn.v_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.103.input_layernorm.weight": "model-00079-of-00106.safetensors", + "model.layers.103.mlp.down_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.103.mlp.gate_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.103.mlp.up_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.103.post_attention_layernorm.weight": "model-00079-of-00106.safetensors", + "model.layers.103.self_attn.k_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.103.self_attn.o_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.103.self_attn.q_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.103.self_attn.v_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.104.input_layernorm.weight": "model-00079-of-00106.safetensors", + "model.layers.104.mlp.down_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.104.mlp.gate_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.104.mlp.up_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.104.post_attention_layernorm.weight": "model-00079-of-00106.safetensors", + "model.layers.104.self_attn.k_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.104.self_attn.o_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.104.self_attn.q_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.104.self_attn.v_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.105.input_layernorm.weight": "model-00080-of-00106.safetensors", + "model.layers.105.mlp.down_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.105.mlp.gate_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.105.mlp.up_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.105.post_attention_layernorm.weight": "model-00080-of-00106.safetensors", + "model.layers.105.self_attn.k_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.105.self_attn.o_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.105.self_attn.q_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.105.self_attn.v_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.106.input_layernorm.weight": "model-00081-of-00106.safetensors", + "model.layers.106.mlp.down_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.106.mlp.gate_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.106.mlp.up_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.106.post_attention_layernorm.weight": "model-00081-of-00106.safetensors", + "model.layers.106.self_attn.k_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.106.self_attn.o_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.106.self_attn.q_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.106.self_attn.v_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.107.input_layernorm.weight": "model-00082-of-00106.safetensors", + "model.layers.107.mlp.down_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.107.mlp.gate_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.107.mlp.up_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.107.post_attention_layernorm.weight": "model-00082-of-00106.safetensors", + "model.layers.107.self_attn.k_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.107.self_attn.o_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.107.self_attn.q_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.107.self_attn.v_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.108.input_layernorm.weight": "model-00082-of-00106.safetensors", + "model.layers.108.mlp.down_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.108.mlp.gate_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.108.mlp.up_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.108.post_attention_layernorm.weight": "model-00082-of-00106.safetensors", + "model.layers.108.self_attn.k_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.108.self_attn.o_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.108.self_attn.q_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.108.self_attn.v_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.109.input_layernorm.weight": "model-00083-of-00106.safetensors", + "model.layers.109.mlp.down_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.109.mlp.gate_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.109.mlp.up_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.109.post_attention_layernorm.weight": "model-00083-of-00106.safetensors", + "model.layers.109.self_attn.k_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.109.self_attn.o_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.109.self_attn.q_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.109.self_attn.v_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.11.input_layernorm.weight": "model-00010-of-00106.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00010-of-00106.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.110.input_layernorm.weight": "model-00084-of-00106.safetensors", + "model.layers.110.mlp.down_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.110.mlp.gate_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.110.mlp.up_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.110.post_attention_layernorm.weight": "model-00084-of-00106.safetensors", + "model.layers.110.self_attn.k_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.110.self_attn.o_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.110.self_attn.q_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.110.self_attn.v_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.111.input_layernorm.weight": "model-00085-of-00106.safetensors", + "model.layers.111.mlp.down_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.111.mlp.gate_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.111.mlp.up_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.111.post_attention_layernorm.weight": "model-00085-of-00106.safetensors", + "model.layers.111.self_attn.k_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.111.self_attn.o_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.111.self_attn.q_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.111.self_attn.v_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.112.input_layernorm.weight": "model-00085-of-00106.safetensors", + "model.layers.112.mlp.down_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.112.mlp.gate_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.112.mlp.up_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.112.post_attention_layernorm.weight": "model-00085-of-00106.safetensors", + "model.layers.112.self_attn.k_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.112.self_attn.o_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.112.self_attn.q_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.112.self_attn.v_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.113.input_layernorm.weight": "model-00086-of-00106.safetensors", + "model.layers.113.mlp.down_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.113.mlp.gate_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.113.mlp.up_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.113.post_attention_layernorm.weight": "model-00086-of-00106.safetensors", + "model.layers.113.self_attn.k_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.113.self_attn.o_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.113.self_attn.q_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.113.self_attn.v_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.114.input_layernorm.weight": "model-00087-of-00106.safetensors", + "model.layers.114.mlp.down_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.114.mlp.gate_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.114.mlp.up_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.114.post_attention_layernorm.weight": "model-00087-of-00106.safetensors", + "model.layers.114.self_attn.k_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.114.self_attn.o_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.114.self_attn.q_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.114.self_attn.v_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.115.input_layernorm.weight": "model-00088-of-00106.safetensors", + "model.layers.115.mlp.down_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.115.mlp.gate_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.115.mlp.up_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.115.post_attention_layernorm.weight": "model-00088-of-00106.safetensors", + "model.layers.115.self_attn.k_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.115.self_attn.o_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.115.self_attn.q_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.115.self_attn.v_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.116.input_layernorm.weight": "model-00088-of-00106.safetensors", + "model.layers.116.mlp.down_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.116.mlp.gate_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.116.mlp.up_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.116.post_attention_layernorm.weight": "model-00088-of-00106.safetensors", + "model.layers.116.self_attn.k_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.116.self_attn.o_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.116.self_attn.q_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.116.self_attn.v_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.117.input_layernorm.weight": "model-00089-of-00106.safetensors", + "model.layers.117.mlp.down_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.117.mlp.gate_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.117.mlp.up_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.117.post_attention_layernorm.weight": "model-00089-of-00106.safetensors", + "model.layers.117.self_attn.k_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.117.self_attn.o_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.117.self_attn.q_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.117.self_attn.v_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.118.input_layernorm.weight": "model-00090-of-00106.safetensors", + "model.layers.118.mlp.down_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.118.mlp.gate_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.118.mlp.up_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.118.post_attention_layernorm.weight": "model-00090-of-00106.safetensors", + "model.layers.118.self_attn.k_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.118.self_attn.o_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.118.self_attn.q_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.118.self_attn.v_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.119.input_layernorm.weight": "model-00091-of-00106.safetensors", + "model.layers.119.mlp.down_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.119.mlp.gate_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.119.mlp.up_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.119.post_attention_layernorm.weight": "model-00091-of-00106.safetensors", + "model.layers.119.self_attn.k_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.119.self_attn.o_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.119.self_attn.q_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.119.self_attn.v_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.12.input_layernorm.weight": "model-00010-of-00106.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00010-of-00106.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.120.input_layernorm.weight": "model-00091-of-00106.safetensors", + "model.layers.120.mlp.down_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.120.mlp.gate_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.120.mlp.up_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.120.post_attention_layernorm.weight": "model-00091-of-00106.safetensors", + "model.layers.120.self_attn.k_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.120.self_attn.o_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.120.self_attn.q_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.120.self_attn.v_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.121.input_layernorm.weight": "model-00092-of-00106.safetensors", + "model.layers.121.mlp.down_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.121.mlp.gate_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.121.mlp.up_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.121.post_attention_layernorm.weight": "model-00092-of-00106.safetensors", + "model.layers.121.self_attn.k_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.121.self_attn.o_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.121.self_attn.q_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.121.self_attn.v_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.122.input_layernorm.weight": "model-00093-of-00106.safetensors", + "model.layers.122.mlp.down_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.122.mlp.gate_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.122.mlp.up_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.122.post_attention_layernorm.weight": "model-00093-of-00106.safetensors", + "model.layers.122.self_attn.k_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.122.self_attn.o_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.122.self_attn.q_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.122.self_attn.v_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.123.input_layernorm.weight": "model-00094-of-00106.safetensors", + "model.layers.123.mlp.down_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.123.mlp.gate_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.123.mlp.up_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.123.post_attention_layernorm.weight": "model-00094-of-00106.safetensors", + "model.layers.123.self_attn.k_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.123.self_attn.o_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.123.self_attn.q_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.123.self_attn.v_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.124.input_layernorm.weight": "model-00094-of-00106.safetensors", + "model.layers.124.mlp.down_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.124.mlp.gate_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.124.mlp.up_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.124.post_attention_layernorm.weight": "model-00094-of-00106.safetensors", + "model.layers.124.self_attn.k_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.124.self_attn.o_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.124.self_attn.q_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.124.self_attn.v_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.125.input_layernorm.weight": "model-00095-of-00106.safetensors", + "model.layers.125.mlp.down_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.125.mlp.gate_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.125.mlp.up_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.125.post_attention_layernorm.weight": "model-00095-of-00106.safetensors", + "model.layers.125.self_attn.k_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.125.self_attn.o_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.125.self_attn.q_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.125.self_attn.v_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.126.input_layernorm.weight": "model-00096-of-00106.safetensors", + "model.layers.126.mlp.down_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.126.mlp.gate_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.126.mlp.up_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.126.post_attention_layernorm.weight": "model-00096-of-00106.safetensors", + "model.layers.126.self_attn.k_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.126.self_attn.o_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.126.self_attn.q_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.126.self_attn.v_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.127.input_layernorm.weight": "model-00097-of-00106.safetensors", + "model.layers.127.mlp.down_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.127.mlp.gate_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.127.mlp.up_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.127.post_attention_layernorm.weight": "model-00097-of-00106.safetensors", + "model.layers.127.self_attn.k_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.127.self_attn.o_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.127.self_attn.q_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.127.self_attn.v_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.128.input_layernorm.weight": "model-00097-of-00106.safetensors", + "model.layers.128.mlp.down_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.128.mlp.gate_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.128.mlp.up_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.128.post_attention_layernorm.weight": "model-00097-of-00106.safetensors", + "model.layers.128.self_attn.k_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.128.self_attn.o_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.128.self_attn.q_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.128.self_attn.v_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.129.input_layernorm.weight": "model-00098-of-00106.safetensors", + "model.layers.129.mlp.down_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.129.mlp.gate_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.129.mlp.up_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.129.post_attention_layernorm.weight": "model-00098-of-00106.safetensors", + "model.layers.129.self_attn.k_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.129.self_attn.o_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.129.self_attn.q_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.129.self_attn.v_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.13.input_layernorm.weight": "model-00011-of-00106.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00011-of-00106.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.130.input_layernorm.weight": "model-00099-of-00106.safetensors", + "model.layers.130.mlp.down_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.130.mlp.gate_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.130.mlp.up_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.130.post_attention_layernorm.weight": "model-00099-of-00106.safetensors", + "model.layers.130.self_attn.k_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.130.self_attn.o_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.130.self_attn.q_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.130.self_attn.v_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.131.input_layernorm.weight": "model-00100-of-00106.safetensors", + "model.layers.131.mlp.down_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.131.mlp.gate_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.131.mlp.up_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.131.post_attention_layernorm.weight": "model-00100-of-00106.safetensors", + "model.layers.131.self_attn.k_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.131.self_attn.o_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.131.self_attn.q_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.131.self_attn.v_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.132.input_layernorm.weight": "model-00100-of-00106.safetensors", + "model.layers.132.mlp.down_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.132.mlp.gate_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.132.mlp.up_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.132.post_attention_layernorm.weight": "model-00100-of-00106.safetensors", + "model.layers.132.self_attn.k_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.132.self_attn.o_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.132.self_attn.q_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.132.self_attn.v_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.133.input_layernorm.weight": "model-00101-of-00106.safetensors", + "model.layers.133.mlp.down_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.133.mlp.gate_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.133.mlp.up_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.133.post_attention_layernorm.weight": "model-00101-of-00106.safetensors", + "model.layers.133.self_attn.k_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.133.self_attn.o_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.133.self_attn.q_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.133.self_attn.v_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.134.input_layernorm.weight": "model-00102-of-00106.safetensors", + "model.layers.134.mlp.down_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.134.mlp.gate_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.134.mlp.up_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.134.post_attention_layernorm.weight": "model-00102-of-00106.safetensors", + "model.layers.134.self_attn.k_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.134.self_attn.o_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.134.self_attn.q_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.134.self_attn.v_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.135.input_layernorm.weight": "model-00103-of-00106.safetensors", + "model.layers.135.mlp.down_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.135.mlp.gate_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.135.mlp.up_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.135.post_attention_layernorm.weight": "model-00103-of-00106.safetensors", + "model.layers.135.self_attn.k_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.135.self_attn.o_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.135.self_attn.q_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.135.self_attn.v_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.136.input_layernorm.weight": "model-00103-of-00106.safetensors", + "model.layers.136.mlp.down_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.136.mlp.gate_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.136.mlp.up_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.136.post_attention_layernorm.weight": "model-00103-of-00106.safetensors", + "model.layers.136.self_attn.k_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.136.self_attn.o_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.136.self_attn.q_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.136.self_attn.v_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.137.input_layernorm.weight": "model-00104-of-00106.safetensors", + "model.layers.137.mlp.down_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.137.mlp.gate_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.137.mlp.up_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.137.post_attention_layernorm.weight": "model-00104-of-00106.safetensors", + "model.layers.137.self_attn.k_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.137.self_attn.o_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.137.self_attn.q_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.137.self_attn.v_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.138.input_layernorm.weight": "model-00105-of-00106.safetensors", + "model.layers.138.mlp.down_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.138.mlp.gate_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.138.mlp.up_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.138.post_attention_layernorm.weight": "model-00105-of-00106.safetensors", + "model.layers.138.self_attn.k_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.138.self_attn.o_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.138.self_attn.q_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.138.self_attn.v_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.139.input_layernorm.weight": "model-00106-of-00106.safetensors", + "model.layers.139.mlp.down_proj.weight": "model-00106-of-00106.safetensors", + "model.layers.139.mlp.gate_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.139.mlp.up_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.139.post_attention_layernorm.weight": "model-00106-of-00106.safetensors", + "model.layers.139.self_attn.k_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.139.self_attn.o_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.139.self_attn.q_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.139.self_attn.v_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.14.input_layernorm.weight": "model-00012-of-00106.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00012-of-00106.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.15.input_layernorm.weight": "model-00013-of-00106.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00013-of-00106.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.16.input_layernorm.weight": "model-00013-of-00106.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00013-of-00106.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.17.input_layernorm.weight": "model-00014-of-00106.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00014-of-00106.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.18.input_layernorm.weight": "model-00015-of-00106.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00015-of-00106.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.19.input_layernorm.weight": "model-00016-of-00106.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00016-of-00106.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.2.input_layernorm.weight": "model-00003-of-00106.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00003-of-00106.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.20.input_layernorm.weight": "model-00016-of-00106.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00016-of-00106.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.21.input_layernorm.weight": "model-00017-of-00106.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00017-of-00106.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.22.input_layernorm.weight": "model-00018-of-00106.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00018-of-00106.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.23.input_layernorm.weight": "model-00019-of-00106.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00019-of-00106.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.24.input_layernorm.weight": "model-00019-of-00106.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00019-of-00106.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.25.input_layernorm.weight": "model-00020-of-00106.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00020-of-00106.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.26.input_layernorm.weight": "model-00021-of-00106.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00021-of-00106.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.27.input_layernorm.weight": "model-00022-of-00106.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00022-of-00106.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.28.input_layernorm.weight": "model-00022-of-00106.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00022-of-00106.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.29.input_layernorm.weight": "model-00023-of-00106.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00023-of-00106.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.3.input_layernorm.weight": "model-00004-of-00106.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00004-of-00106.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.30.input_layernorm.weight": "model-00024-of-00106.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00024-of-00106.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.31.input_layernorm.weight": "model-00025-of-00106.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00025-of-00106.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.32.input_layernorm.weight": "model-00025-of-00106.safetensors", + "model.layers.32.mlp.down_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.32.mlp.up_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00025-of-00106.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.33.input_layernorm.weight": "model-00026-of-00106.safetensors", + "model.layers.33.mlp.down_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.33.mlp.up_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00026-of-00106.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.34.input_layernorm.weight": "model-00027-of-00106.safetensors", + "model.layers.34.mlp.down_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.34.mlp.up_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00027-of-00106.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.35.input_layernorm.weight": "model-00028-of-00106.safetensors", + "model.layers.35.mlp.down_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.35.mlp.up_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00028-of-00106.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.36.input_layernorm.weight": "model-00028-of-00106.safetensors", + "model.layers.36.mlp.down_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.36.mlp.gate_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.36.mlp.up_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00028-of-00106.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.37.input_layernorm.weight": "model-00029-of-00106.safetensors", + "model.layers.37.mlp.down_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.37.mlp.gate_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.37.mlp.up_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00029-of-00106.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.38.input_layernorm.weight": "model-00030-of-00106.safetensors", + "model.layers.38.mlp.down_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.38.mlp.gate_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.38.mlp.up_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00030-of-00106.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.39.input_layernorm.weight": "model-00031-of-00106.safetensors", + "model.layers.39.mlp.down_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.39.mlp.gate_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.39.mlp.up_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00031-of-00106.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.4.input_layernorm.weight": "model-00004-of-00106.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00004-of-00106.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.40.input_layernorm.weight": "model-00031-of-00106.safetensors", + "model.layers.40.mlp.down_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.40.mlp.gate_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.40.mlp.up_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00031-of-00106.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.41.input_layernorm.weight": "model-00032-of-00106.safetensors", + "model.layers.41.mlp.down_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.41.mlp.gate_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.41.mlp.up_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00032-of-00106.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.42.input_layernorm.weight": "model-00033-of-00106.safetensors", + "model.layers.42.mlp.down_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.42.mlp.gate_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.42.mlp.up_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00033-of-00106.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.43.input_layernorm.weight": "model-00034-of-00106.safetensors", + "model.layers.43.mlp.down_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.43.mlp.gate_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.43.mlp.up_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00034-of-00106.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.44.input_layernorm.weight": "model-00034-of-00106.safetensors", + "model.layers.44.mlp.down_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.44.mlp.gate_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.44.mlp.up_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00034-of-00106.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.45.input_layernorm.weight": "model-00035-of-00106.safetensors", + "model.layers.45.mlp.down_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.45.mlp.gate_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.45.mlp.up_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00035-of-00106.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.46.input_layernorm.weight": "model-00036-of-00106.safetensors", + "model.layers.46.mlp.down_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.46.mlp.gate_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.46.mlp.up_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00036-of-00106.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.47.input_layernorm.weight": "model-00037-of-00106.safetensors", + "model.layers.47.mlp.down_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.47.mlp.gate_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.47.mlp.up_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00037-of-00106.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.48.input_layernorm.weight": "model-00037-of-00106.safetensors", + "model.layers.48.mlp.down_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.48.mlp.gate_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.48.mlp.up_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00037-of-00106.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.49.input_layernorm.weight": "model-00038-of-00106.safetensors", + "model.layers.49.mlp.down_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.49.mlp.gate_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.49.mlp.up_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00038-of-00106.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.5.input_layernorm.weight": "model-00005-of-00106.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00005-of-00106.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.50.input_layernorm.weight": "model-00039-of-00106.safetensors", + "model.layers.50.mlp.down_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.50.mlp.gate_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.50.mlp.up_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00039-of-00106.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.51.input_layernorm.weight": "model-00040-of-00106.safetensors", + "model.layers.51.mlp.down_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.51.mlp.gate_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.51.mlp.up_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00040-of-00106.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.52.input_layernorm.weight": "model-00040-of-00106.safetensors", + "model.layers.52.mlp.down_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.52.mlp.gate_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.52.mlp.up_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model-00040-of-00106.safetensors", + "model.layers.52.self_attn.k_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.52.self_attn.o_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.52.self_attn.q_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.52.self_attn.v_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.53.input_layernorm.weight": "model-00041-of-00106.safetensors", + "model.layers.53.mlp.down_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.53.mlp.gate_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.53.mlp.up_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model-00041-of-00106.safetensors", + "model.layers.53.self_attn.k_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.53.self_attn.o_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.53.self_attn.q_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.53.self_attn.v_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.54.input_layernorm.weight": "model-00042-of-00106.safetensors", + "model.layers.54.mlp.down_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.54.mlp.gate_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.54.mlp.up_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model-00042-of-00106.safetensors", + "model.layers.54.self_attn.k_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.54.self_attn.o_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.54.self_attn.q_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.54.self_attn.v_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.55.input_layernorm.weight": "model-00043-of-00106.safetensors", + "model.layers.55.mlp.down_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.55.mlp.gate_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.55.mlp.up_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model-00043-of-00106.safetensors", + "model.layers.55.self_attn.k_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.55.self_attn.o_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.55.self_attn.q_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.55.self_attn.v_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.56.input_layernorm.weight": "model-00043-of-00106.safetensors", + "model.layers.56.mlp.down_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.56.mlp.gate_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.56.mlp.up_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model-00043-of-00106.safetensors", + "model.layers.56.self_attn.k_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.56.self_attn.o_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.56.self_attn.q_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.56.self_attn.v_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.57.input_layernorm.weight": "model-00044-of-00106.safetensors", + "model.layers.57.mlp.down_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.57.mlp.gate_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.57.mlp.up_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model-00044-of-00106.safetensors", + "model.layers.57.self_attn.k_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.57.self_attn.o_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.57.self_attn.q_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.57.self_attn.v_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.58.input_layernorm.weight": "model-00045-of-00106.safetensors", + "model.layers.58.mlp.down_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.58.mlp.gate_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.58.mlp.up_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model-00045-of-00106.safetensors", + "model.layers.58.self_attn.k_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.58.self_attn.o_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.58.self_attn.q_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.58.self_attn.v_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.59.input_layernorm.weight": "model-00046-of-00106.safetensors", + "model.layers.59.mlp.down_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.59.mlp.gate_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.59.mlp.up_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model-00046-of-00106.safetensors", + "model.layers.59.self_attn.k_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.59.self_attn.o_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.59.self_attn.q_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.59.self_attn.v_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.6.input_layernorm.weight": "model-00006-of-00106.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00006-of-00106.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.60.input_layernorm.weight": "model-00046-of-00106.safetensors", + "model.layers.60.mlp.down_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.60.mlp.gate_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.60.mlp.up_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.60.post_attention_layernorm.weight": "model-00046-of-00106.safetensors", + "model.layers.60.self_attn.k_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.60.self_attn.o_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.60.self_attn.q_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.60.self_attn.v_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.61.input_layernorm.weight": "model-00047-of-00106.safetensors", + "model.layers.61.mlp.down_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.61.mlp.gate_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.61.mlp.up_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.61.post_attention_layernorm.weight": "model-00047-of-00106.safetensors", + "model.layers.61.self_attn.k_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.61.self_attn.o_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.61.self_attn.q_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.61.self_attn.v_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.62.input_layernorm.weight": "model-00048-of-00106.safetensors", + "model.layers.62.mlp.down_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.62.mlp.gate_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.62.mlp.up_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.62.post_attention_layernorm.weight": "model-00048-of-00106.safetensors", + "model.layers.62.self_attn.k_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.62.self_attn.o_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.62.self_attn.q_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.62.self_attn.v_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.63.input_layernorm.weight": "model-00049-of-00106.safetensors", + "model.layers.63.mlp.down_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.63.mlp.gate_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.63.mlp.up_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.63.post_attention_layernorm.weight": "model-00049-of-00106.safetensors", + "model.layers.63.self_attn.k_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.63.self_attn.o_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.63.self_attn.q_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.63.self_attn.v_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.64.input_layernorm.weight": "model-00049-of-00106.safetensors", + "model.layers.64.mlp.down_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.64.mlp.gate_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.64.mlp.up_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.64.post_attention_layernorm.weight": "model-00049-of-00106.safetensors", + "model.layers.64.self_attn.k_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.64.self_attn.o_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.64.self_attn.q_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.64.self_attn.v_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.65.input_layernorm.weight": "model-00050-of-00106.safetensors", + "model.layers.65.mlp.down_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.65.mlp.gate_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.65.mlp.up_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.65.post_attention_layernorm.weight": "model-00050-of-00106.safetensors", + "model.layers.65.self_attn.k_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.65.self_attn.o_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.65.self_attn.q_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.65.self_attn.v_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.66.input_layernorm.weight": "model-00051-of-00106.safetensors", + "model.layers.66.mlp.down_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.66.mlp.gate_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.66.mlp.up_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.66.post_attention_layernorm.weight": "model-00051-of-00106.safetensors", + "model.layers.66.self_attn.k_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.66.self_attn.o_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.66.self_attn.q_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.66.self_attn.v_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.67.input_layernorm.weight": "model-00052-of-00106.safetensors", + "model.layers.67.mlp.down_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.67.mlp.gate_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.67.mlp.up_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.67.post_attention_layernorm.weight": "model-00052-of-00106.safetensors", + "model.layers.67.self_attn.k_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.67.self_attn.o_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.67.self_attn.q_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.67.self_attn.v_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.68.input_layernorm.weight": "model-00052-of-00106.safetensors", + "model.layers.68.mlp.down_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.68.mlp.gate_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.68.mlp.up_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.68.post_attention_layernorm.weight": "model-00052-of-00106.safetensors", + "model.layers.68.self_attn.k_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.68.self_attn.o_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.68.self_attn.q_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.68.self_attn.v_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.69.input_layernorm.weight": "model-00053-of-00106.safetensors", + "model.layers.69.mlp.down_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.69.mlp.gate_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.69.mlp.up_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.69.post_attention_layernorm.weight": "model-00053-of-00106.safetensors", + "model.layers.69.self_attn.k_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.69.self_attn.o_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.69.self_attn.q_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.69.self_attn.v_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.7.input_layernorm.weight": "model-00007-of-00106.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00007-of-00106.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.70.input_layernorm.weight": "model-00054-of-00106.safetensors", + "model.layers.70.mlp.down_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.70.mlp.gate_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.70.mlp.up_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.70.post_attention_layernorm.weight": "model-00054-of-00106.safetensors", + "model.layers.70.self_attn.k_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.70.self_attn.o_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.70.self_attn.q_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.70.self_attn.v_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.71.input_layernorm.weight": "model-00055-of-00106.safetensors", + "model.layers.71.mlp.down_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.71.mlp.gate_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.71.mlp.up_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.71.post_attention_layernorm.weight": "model-00055-of-00106.safetensors", + "model.layers.71.self_attn.k_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.71.self_attn.o_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.71.self_attn.q_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.71.self_attn.v_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.72.input_layernorm.weight": "model-00055-of-00106.safetensors", + "model.layers.72.mlp.down_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.72.mlp.gate_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.72.mlp.up_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.72.post_attention_layernorm.weight": "model-00055-of-00106.safetensors", + "model.layers.72.self_attn.k_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.72.self_attn.o_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.72.self_attn.q_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.72.self_attn.v_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.73.input_layernorm.weight": "model-00056-of-00106.safetensors", + "model.layers.73.mlp.down_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.73.mlp.gate_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.73.mlp.up_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.73.post_attention_layernorm.weight": "model-00056-of-00106.safetensors", + "model.layers.73.self_attn.k_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.73.self_attn.o_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.73.self_attn.q_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.73.self_attn.v_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.74.input_layernorm.weight": "model-00057-of-00106.safetensors", + "model.layers.74.mlp.down_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.74.mlp.gate_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.74.mlp.up_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.74.post_attention_layernorm.weight": "model-00057-of-00106.safetensors", + "model.layers.74.self_attn.k_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.74.self_attn.o_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.74.self_attn.q_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.74.self_attn.v_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.75.input_layernorm.weight": "model-00058-of-00106.safetensors", + "model.layers.75.mlp.down_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.75.mlp.gate_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.75.mlp.up_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.75.post_attention_layernorm.weight": "model-00058-of-00106.safetensors", + "model.layers.75.self_attn.k_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.75.self_attn.o_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.75.self_attn.q_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.75.self_attn.v_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.76.input_layernorm.weight": "model-00058-of-00106.safetensors", + "model.layers.76.mlp.down_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.76.mlp.gate_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.76.mlp.up_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.76.post_attention_layernorm.weight": "model-00058-of-00106.safetensors", + "model.layers.76.self_attn.k_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.76.self_attn.o_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.76.self_attn.q_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.76.self_attn.v_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.77.input_layernorm.weight": "model-00059-of-00106.safetensors", + "model.layers.77.mlp.down_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.77.mlp.gate_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.77.mlp.up_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.77.post_attention_layernorm.weight": "model-00059-of-00106.safetensors", + "model.layers.77.self_attn.k_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.77.self_attn.o_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.77.self_attn.q_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.77.self_attn.v_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.78.input_layernorm.weight": "model-00060-of-00106.safetensors", + "model.layers.78.mlp.down_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.78.mlp.gate_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.78.mlp.up_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.78.post_attention_layernorm.weight": "model-00060-of-00106.safetensors", + "model.layers.78.self_attn.k_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.78.self_attn.o_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.78.self_attn.q_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.78.self_attn.v_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.79.input_layernorm.weight": "model-00061-of-00106.safetensors", + "model.layers.79.mlp.down_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.79.mlp.gate_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.79.mlp.up_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.79.post_attention_layernorm.weight": "model-00061-of-00106.safetensors", + "model.layers.79.self_attn.k_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.79.self_attn.o_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.79.self_attn.q_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.79.self_attn.v_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.8.input_layernorm.weight": "model-00007-of-00106.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00007-of-00106.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.80.input_layernorm.weight": "model-00061-of-00106.safetensors", + "model.layers.80.mlp.down_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.80.mlp.gate_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.80.mlp.up_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.80.post_attention_layernorm.weight": "model-00061-of-00106.safetensors", + "model.layers.80.self_attn.k_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.80.self_attn.o_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.80.self_attn.q_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.80.self_attn.v_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.81.input_layernorm.weight": "model-00062-of-00106.safetensors", + "model.layers.81.mlp.down_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.81.mlp.gate_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.81.mlp.up_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.81.post_attention_layernorm.weight": "model-00062-of-00106.safetensors", + "model.layers.81.self_attn.k_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.81.self_attn.o_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.81.self_attn.q_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.81.self_attn.v_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.82.input_layernorm.weight": "model-00063-of-00106.safetensors", + "model.layers.82.mlp.down_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.82.mlp.gate_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.82.mlp.up_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.82.post_attention_layernorm.weight": "model-00063-of-00106.safetensors", + "model.layers.82.self_attn.k_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.82.self_attn.o_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.82.self_attn.q_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.82.self_attn.v_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.83.input_layernorm.weight": "model-00064-of-00106.safetensors", + "model.layers.83.mlp.down_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.83.mlp.gate_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.83.mlp.up_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.83.post_attention_layernorm.weight": "model-00064-of-00106.safetensors", + "model.layers.83.self_attn.k_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.83.self_attn.o_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.83.self_attn.q_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.83.self_attn.v_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.84.input_layernorm.weight": "model-00064-of-00106.safetensors", + "model.layers.84.mlp.down_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.84.mlp.gate_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.84.mlp.up_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.84.post_attention_layernorm.weight": "model-00064-of-00106.safetensors", + "model.layers.84.self_attn.k_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.84.self_attn.o_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.84.self_attn.q_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.84.self_attn.v_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.85.input_layernorm.weight": "model-00065-of-00106.safetensors", + "model.layers.85.mlp.down_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.85.mlp.gate_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.85.mlp.up_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.85.post_attention_layernorm.weight": "model-00065-of-00106.safetensors", + "model.layers.85.self_attn.k_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.85.self_attn.o_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.85.self_attn.q_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.85.self_attn.v_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.86.input_layernorm.weight": "model-00066-of-00106.safetensors", + "model.layers.86.mlp.down_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.86.mlp.gate_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.86.mlp.up_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.86.post_attention_layernorm.weight": "model-00066-of-00106.safetensors", + "model.layers.86.self_attn.k_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.86.self_attn.o_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.86.self_attn.q_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.86.self_attn.v_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.87.input_layernorm.weight": "model-00067-of-00106.safetensors", + "model.layers.87.mlp.down_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.87.mlp.gate_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.87.mlp.up_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.87.post_attention_layernorm.weight": "model-00067-of-00106.safetensors", + "model.layers.87.self_attn.k_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.87.self_attn.o_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.87.self_attn.q_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.87.self_attn.v_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.88.input_layernorm.weight": "model-00067-of-00106.safetensors", + "model.layers.88.mlp.down_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.88.mlp.gate_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.88.mlp.up_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.88.post_attention_layernorm.weight": "model-00067-of-00106.safetensors", + "model.layers.88.self_attn.k_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.88.self_attn.o_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.88.self_attn.q_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.88.self_attn.v_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.89.input_layernorm.weight": "model-00068-of-00106.safetensors", + "model.layers.89.mlp.down_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.89.mlp.gate_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.89.mlp.up_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.89.post_attention_layernorm.weight": "model-00068-of-00106.safetensors", + "model.layers.89.self_attn.k_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.89.self_attn.o_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.89.self_attn.q_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.89.self_attn.v_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.9.input_layernorm.weight": "model-00008-of-00106.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00008-of-00106.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.90.input_layernorm.weight": "model-00069-of-00106.safetensors", + "model.layers.90.mlp.down_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.90.mlp.gate_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.90.mlp.up_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.90.post_attention_layernorm.weight": "model-00069-of-00106.safetensors", + "model.layers.90.self_attn.k_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.90.self_attn.o_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.90.self_attn.q_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.90.self_attn.v_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.91.input_layernorm.weight": "model-00070-of-00106.safetensors", + "model.layers.91.mlp.down_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.91.mlp.gate_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.91.mlp.up_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.91.post_attention_layernorm.weight": "model-00070-of-00106.safetensors", + "model.layers.91.self_attn.k_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.91.self_attn.o_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.91.self_attn.q_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.91.self_attn.v_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.92.input_layernorm.weight": "model-00070-of-00106.safetensors", + "model.layers.92.mlp.down_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.92.mlp.gate_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.92.mlp.up_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.92.post_attention_layernorm.weight": "model-00070-of-00106.safetensors", + "model.layers.92.self_attn.k_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.92.self_attn.o_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.92.self_attn.q_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.92.self_attn.v_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.93.input_layernorm.weight": "model-00071-of-00106.safetensors", + "model.layers.93.mlp.down_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.93.mlp.gate_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.93.mlp.up_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.93.post_attention_layernorm.weight": "model-00071-of-00106.safetensors", + "model.layers.93.self_attn.k_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.93.self_attn.o_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.93.self_attn.q_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.93.self_attn.v_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.94.input_layernorm.weight": "model-00072-of-00106.safetensors", + "model.layers.94.mlp.down_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.94.mlp.gate_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.94.mlp.up_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.94.post_attention_layernorm.weight": "model-00072-of-00106.safetensors", + "model.layers.94.self_attn.k_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.94.self_attn.o_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.94.self_attn.q_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.94.self_attn.v_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.95.input_layernorm.weight": "model-00073-of-00106.safetensors", + "model.layers.95.mlp.down_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.95.mlp.gate_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.95.mlp.up_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.95.post_attention_layernorm.weight": "model-00073-of-00106.safetensors", + "model.layers.95.self_attn.k_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.95.self_attn.o_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.95.self_attn.q_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.95.self_attn.v_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.96.input_layernorm.weight": "model-00073-of-00106.safetensors", + "model.layers.96.mlp.down_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.96.mlp.gate_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.96.mlp.up_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.96.post_attention_layernorm.weight": "model-00073-of-00106.safetensors", + "model.layers.96.self_attn.k_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.96.self_attn.o_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.96.self_attn.q_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.96.self_attn.v_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.97.input_layernorm.weight": "model-00074-of-00106.safetensors", + "model.layers.97.mlp.down_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.97.mlp.gate_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.97.mlp.up_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.97.post_attention_layernorm.weight": "model-00074-of-00106.safetensors", + "model.layers.97.self_attn.k_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.97.self_attn.o_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.97.self_attn.q_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.97.self_attn.v_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.98.input_layernorm.weight": "model-00075-of-00106.safetensors", + "model.layers.98.mlp.down_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.98.mlp.gate_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.98.mlp.up_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.98.post_attention_layernorm.weight": "model-00075-of-00106.safetensors", + "model.layers.98.self_attn.k_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.98.self_attn.o_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.98.self_attn.q_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.98.self_attn.v_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.99.input_layernorm.weight": "model-00076-of-00106.safetensors", + "model.layers.99.mlp.down_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.99.mlp.gate_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.99.mlp.up_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.99.post_attention_layernorm.weight": "model-00076-of-00106.safetensors", + "model.layers.99.self_attn.k_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.99.self_attn.o_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.99.self_attn.q_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.99.self_attn.v_proj.weight": "model-00075-of-00106.safetensors", + "model.norm.weight": "model-00106-of-00106.safetensors" + } +}