diff --git a/README.md b/README.md index b77f5bd743cffed5133fbe2215fefd412c298d65..b83538614d0af89b2e298b9b6e38c1eca5378282 100644 --- a/README.md +++ b/README.md @@ -1,28 +1,28 @@ --- -base_model: -- sophosympatheia/Midnight-Miqu-70B-v1.5 -- NeverSleep/MiquMaid-v3-70B -- maywell/miqu-evil-dpo -- 152334H/miqu-1-70b-sf +base_model: [] library_name: transformers tags: - mergekit - merge -license: other + --- # MiquSuperdark-70B-v1 -**MiquSuperdark-70B-v1** is a merge of three of the most popular Miqu-derived models, along with Miqu itself. The goal of the merge is to create an strong, well-rounded chat model that picks up desirable traits from its constituent models without sacrificing intelligence. +This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). + +## Merge Details +### Merge Method -This is a DARE Linear merge with the following composition: -- [sophosympatheia/Midnight-Miqu-70B-v1.5](https://huggingface.co/sophosympatheia/Midnight-Miqu-70B-v1.5) at weight 0.4 -- [NeverSleep/MiquMaid-v3-70B](https://huggingface.co/NeverSleep/MiquMaid-v3-70B) at weight 0.2 -- [maywell/miqu-evil-dpo](https://huggingface.co/maywell/miqu-evil-dpo) at weight 0.2 -- [152334H/miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf) at weight 0.2 (used as base model) +This model was merged using the linear [DARE](https://arxiv.org/abs/2311.03099) merge method using /home/dylan/Documents/AI/merge/miqu-1-70b-sf as a base. -DARE Linear was chosen as the merge method based on [this HF discussion](https://huggingface.co/jukofyork/Dark-Miqu-70B/discussions/2), in which the creator of Midnight-Miqu says "*in my own testing I consistently got the best results from using a dare_linear merge when working with miqu models*". +### Models Merged -## Merge Configuration +The following models were included in the merge: +* /home/dylan/Documents/AI/merge/MiquMaid-v3-70B +* /media/dylan/SanDisk/LLMs/miqu-evil-dpo/ +* /media/dylan/SanDisk/LLMs/Midnight-Miqu-70B-v1.5 + +### Configuration The following YAML configuration was used to produce this model: @@ -45,5 +45,3 @@ models: dtype: float16 tokenizer_source: model:/home/dylan/Documents/AI/merge/miqu-1-70b-sf ``` - -The tokenizer is copied from the base model [152334H/miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf). diff --git a/config.json b/config.json new file mode 100644 index 0000000000000000000000000000000000000000..e76cf7b901ff5752549f34fe56611aed7c7971b0 --- /dev/null +++ b/config.json @@ -0,0 +1,28 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": 1, + "eos_token_id": 2, + "hidden_act": "silu", + "hidden_size": 8192, + "initializer_range": 0.02, + "intermediate_size": 28672, + "max_position_embeddings": 32764, + "model_type": "llama", + "num_attention_heads": 64, + "num_hidden_layers": 80, + "num_key_value_heads": 8, + "pad_token_id": 0, + "pretraining_tp": 1, + "rms_norm_eps": 1e-05, + "rope_scaling": null, + "rope_theta": 1000000, + "tie_word_embeddings": false, + "torch_dtype": "float16", + "transformers_version": "4.36.0", + "use_cache": true, + "vocab_size": 32000 +} diff --git a/mergekit_config.yml b/mergekit_config.yml new file mode 100644 index 0000000000000000000000000000000000000000..3bcc694fc4db056e4b8c4455584e0176dc46e581 --- /dev/null +++ b/mergekit_config.yml @@ -0,0 +1,17 @@ +merge_method: dare_linear +base_model: /home/dylan/Documents/AI/merge/miqu-1-70b-sf +models: + - model: /media/dylan/SanDisk/LLMs/Midnight-Miqu-70B-v1.5 + parameters: + weight: 0.4 + - model: /home/dylan/Documents/AI/merge/miqu-1-70b-sf + parameters: + weight: 0.2 + - model: /media/dylan/SanDisk/LLMs/miqu-evil-dpo/ + parameters: + weight: 0.2 + - model: /home/dylan/Documents/AI/merge/MiquMaid-v3-70B + parameters: + weight: 0.2 +dtype: float16 +tokenizer_source: model:/home/dylan/Documents/AI/merge/miqu-1-70b-sf \ No newline at end of file diff --git a/model-00001-of-00081.safetensors b/model-00001-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b410b555bad9e2c2d69b596466711a30c21433a5 --- /dev/null +++ b/model-00001-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b0bc44cfc8b0caaa73ca9a20ebc72de37bfda26a3f4dea8f305b25a998e19d10 +size 1988117048 diff --git a/model-00002-of-00081.safetensors b/model-00002-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..140f1acc814596bb19419f489898038cded08b6f --- /dev/null +++ b/model-00002-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4e7090501cae7b0b35ab2f8b37b3263a9c0394d8b0a6f1d35b76b77c03bac663 +size 1711309856 diff --git a/model-00003-of-00081.safetensors b/model-00003-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e4681fa4ef4c6e9a337139b942cae81eaeb46284 --- /dev/null +++ b/model-00003-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:383ba84fba3bef358a09e7c612e7b4e644d3236de5aa57a66f0a7aad53d34b3a +size 1711309856 diff --git a/model-00004-of-00081.safetensors b/model-00004-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9ab6754d4d2e3c61cfc39e1af168c6a7ec609cb1 --- /dev/null +++ b/model-00004-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:68e7f4cce99af7dbf72bc4298773ba0f5436588ea526296ea4f656b3ed1f0b65 +size 1711309864 diff --git a/model-00005-of-00081.safetensors b/model-00005-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..538f331d5ca4b00c532d28bc02f15fd042fc55b5 --- /dev/null +++ b/model-00005-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:55a9dcddcc1cae39fa3ac33dbe8ec9239686fd407ad2d9e010ae499f1e42ee68 +size 1711309864 diff --git a/model-00006-of-00081.safetensors b/model-00006-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..84abeb71e8255a86a82e76296600f9eb4f16e70c --- /dev/null +++ b/model-00006-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff8c88888900436c2076d9b2aacb17a6030cf297a2489a8f588759f509454bee +size 1711309864 diff --git a/model-00007-of-00081.safetensors b/model-00007-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bdc6bfa904f47fb362ed72efce35a61f6dcbdd79 --- /dev/null +++ b/model-00007-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2ce54d1596a6dcf8f9e1d2a134e01123a6cd32d9ca1c6766dda302418919c765 +size 1711309864 diff --git a/model-00008-of-00081.safetensors b/model-00008-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3a6a5bf67bdb50b3784b54302b32c06ea8fc1652 --- /dev/null +++ b/model-00008-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e999a9d64877e486b37e65388490e76ebdcc2df5b032f039a47c2fa165200710 +size 1711309864 diff --git a/model-00009-of-00081.safetensors b/model-00009-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e87100f709e8b94ba1a85c8cb516e2028d25c41c --- /dev/null +++ b/model-00009-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0330654c6715a4cec27551583267969ba5e1305b36c9cfadfbd1c82e35d11e50 +size 1711309864 diff --git a/model-00010-of-00081.safetensors b/model-00010-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a7a242ba331a1b5345ea091ae7dff1623ab6ae75 --- /dev/null +++ b/model-00010-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:84012fea8f9a1c4b3a45c3a3de932d26e6e1b96a90acd38b64513e0aaf7e9562 +size 1711309864 diff --git a/model-00011-of-00081.safetensors b/model-00011-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f3dd5d9fad00f405a9b65dfe14b60918ece464d1 --- /dev/null +++ b/model-00011-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:98664e29bfeaa08b98e0e61b1cc31ea734ab686ee0562d1875ee6ac6052fa048 +size 1711309864 diff --git a/model-00012-of-00081.safetensors b/model-00012-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..004dd5ef77b7ecf9579132c9750a573e080021c2 --- /dev/null +++ b/model-00012-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0bb5df95096ab6451a76ada113d504d74b9d8388c5cbbbeca4031c5636ad362e +size 1711309864 diff --git a/model-00013-of-00081.safetensors b/model-00013-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..66a174999c6aac0d09ec825af60f1e191615c219 --- /dev/null +++ b/model-00013-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b55b77aff5d211d4328372bed6167d0268af306f2475b4e76d32252c5281901e +size 1711309856 diff --git a/model-00014-of-00081.safetensors b/model-00014-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4b1d95ad1ed5c0c6c5758d88bfab65fe0490e7b5 --- /dev/null +++ b/model-00014-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9faaa7bbf647c01bee079c5d5f9e663834cb276b9178ac2b2d8b8de8eedf9f17 +size 1711309856 diff --git a/model-00015-of-00081.safetensors b/model-00015-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..24d179609d5668c6d6717459029095824473528d --- /dev/null +++ b/model-00015-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:146a105b230ba34852af7b33e9edb86f120634b21845bcbd686cd12cbf626d99 +size 1711309864 diff --git a/model-00016-of-00081.safetensors b/model-00016-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..33c24a8a1a49595978e89a1e1d094f866980e777 --- /dev/null +++ b/model-00016-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d2db1c7f533c43a9d20c9fc3c97cdd8a8f7ee13bd9ab26b2e3ab80f49615641e +size 1711309864 diff --git a/model-00017-of-00081.safetensors b/model-00017-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dc3e71a5ef2efa4aa71255b51acac11c8538f902 --- /dev/null +++ b/model-00017-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae60ff350d4ce02d68f2698b78dd9b3db9860d46dc4e08fdb0bf95a7472eac72 +size 1711309864 diff --git a/model-00018-of-00081.safetensors b/model-00018-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7bc0e902cdc9083092dffdc212b78e487411d55b --- /dev/null +++ b/model-00018-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35c24049577e17c8ae3441825d4bb51dd8803ddf31f062ab6b77138df1e8f542 +size 1711309864 diff --git a/model-00019-of-00081.safetensors b/model-00019-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..219d526c86e6b173c7d91fc6fd7f95fca4ec32f7 --- /dev/null +++ b/model-00019-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3ee60d166c9932cef7433a6df20792be146ec1941807a79348204251333a0af +size 1711309864 diff --git a/model-00020-of-00081.safetensors b/model-00020-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4831fe83c0e8c1457ecaf8f5b58aeb571f8090fc --- /dev/null +++ b/model-00020-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0aa07a5a8fd60b7c074e585927866b50cf5612e8d4bed73165008fa0e36359c3 +size 1711309864 diff --git a/model-00021-of-00081.safetensors b/model-00021-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..be84078676fbae53a97ed775ea4e89d264a44ad9 --- /dev/null +++ b/model-00021-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:98ded078f151dc2eac615214d1f563f31596bab89ff0bc1680adb01f9e79eba2 +size 1711309864 diff --git a/model-00022-of-00081.safetensors b/model-00022-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ac716f0eae1d0787a89583f9a94b1e3406c5ba2e --- /dev/null +++ b/model-00022-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9bc562141c958ea526ec5e7f301fa7c2968b5c48886a94cc7a6d2b77d72ef18f +size 1711309864 diff --git a/model-00023-of-00081.safetensors b/model-00023-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8b1794737261076b13727d379bb34a135bebde8d --- /dev/null +++ b/model-00023-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d10317c4bc2a62281b0ae5aad373e8cdd73261ce43d05f3eed61b3240bcf19a4 +size 1711309864 diff --git a/model-00024-of-00081.safetensors b/model-00024-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0393b9c55028b1450ad8e5d9162104f3633f59fc --- /dev/null +++ b/model-00024-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d8bb3c7ef05bec8a8d1fba6e3c09778b961a3ab4cd5b2d689e14193fcba73ead +size 1711309856 diff --git a/model-00025-of-00081.safetensors b/model-00025-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..04aabd0d4af149def429fd34396924e07ca75ab9 --- /dev/null +++ b/model-00025-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0a1f406ffa2a2e346c04364aea3f155a7f75212d86e12163bfe4dda320accc9c +size 1711309856 diff --git a/model-00026-of-00081.safetensors b/model-00026-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e8347d86e7ebc2dfa348444cfe39840f1ba725a7 --- /dev/null +++ b/model-00026-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ade9d8dfecc1433fbb19dffb88c400ccd34f42f213bdc99accc757abea979ed0 +size 1711309864 diff --git a/model-00027-of-00081.safetensors b/model-00027-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6ab03a48aee427cc189c56c21c59199cb91766e0 --- /dev/null +++ b/model-00027-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c0f5800913053850ee793e8895124187f3d540cf9deb5965d38006680f500468 +size 1711309864 diff --git a/model-00028-of-00081.safetensors b/model-00028-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bc1d4e4253ce0827eff9969c93baef4024612be9 --- /dev/null +++ b/model-00028-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:686915f31b6f4c898f20faa2e5d5ccfbf2d209a7848cd048449302ab7d72f329 +size 1711309864 diff --git a/model-00029-of-00081.safetensors b/model-00029-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..725b3bda59e7478b9066f203260931ddf1f27244 --- /dev/null +++ b/model-00029-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:478acd77b84b6ffb3bcc9002910034b578db2a44e8286d405bdc956db82ce4c1 +size 1711309864 diff --git a/model-00030-of-00081.safetensors b/model-00030-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d273078efa3e9d23a5db863d329c07efd4bdb876 --- /dev/null +++ b/model-00030-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:46fad18b4b17bba82306de14da6b26825ba9713ccc343531840058458d8a6505 +size 1711309864 diff --git a/model-00031-of-00081.safetensors b/model-00031-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..00178168f50303b20e21da1d2b5ec2f288585917 --- /dev/null +++ b/model-00031-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:219158cf9bbaaa4a4ece380f49015428223cc54198b41cc37f8cb809a66e7d11 +size 1711309864 diff --git a/model-00032-of-00081.safetensors b/model-00032-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..392183ef6ac11c095c45359df99a7a1c60da2f0a --- /dev/null +++ b/model-00032-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8b85ce2801bf7a10559b2d3617c6e01252e71c1a0fd9a5aa757d4faca91c77ec +size 1711309864 diff --git a/model-00033-of-00081.safetensors b/model-00033-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f23049467be0d8dcb2e476c53dfa7b4b7aa79bca --- /dev/null +++ b/model-00033-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b2a6df7cb60cfb7931b66ba79feb8fa3d26bfca470af68d5105827ed64227047 +size 1711309864 diff --git a/model-00034-of-00081.safetensors b/model-00034-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b3b97d06038424450762c6cecd36665375ba6e1e --- /dev/null +++ b/model-00034-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:43ffd2f008138380cf55cbcc4740aa7cf28f667ae2dc3bfcd0ef87d0c38f25e0 +size 1711309864 diff --git a/model-00035-of-00081.safetensors b/model-00035-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dd680eaf72608a3b1e7ad5c6e60289a4d84654b1 --- /dev/null +++ b/model-00035-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:44dca8c133b27ab59ebe6c2bde9b138e538ee66e3e79b6cae1511949e5415730 +size 1711309856 diff --git a/model-00036-of-00081.safetensors b/model-00036-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ecd20d377a186972987805901b420ef77f5431d0 --- /dev/null +++ b/model-00036-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:93e3ffe10a51fa17f1d0b31db8924d5554fe4b134542c484ab18f3763f0bbecd +size 1711309856 diff --git a/model-00037-of-00081.safetensors b/model-00037-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f281116e9fcd6688d29fcaeefc8daa332c5f7a9b --- /dev/null +++ b/model-00037-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4650f6dbc9badc41bc5756584b9a084da2a23565bd463e741339f61ad045e759 +size 1711309864 diff --git a/model-00038-of-00081.safetensors b/model-00038-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ff61d3742dd58d1897acfafb65d7976af201f53e --- /dev/null +++ b/model-00038-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a2dbb3250ddc45643b62ae3bf2406206417f68fbf28f435f74addcac36ba5828 +size 1711309864 diff --git a/model-00039-of-00081.safetensors b/model-00039-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3fb8f0e229ed18934a04dce773cc2c3fa5f7a560 --- /dev/null +++ b/model-00039-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:beb9d845096cb973ff27d48ce148435c5940aeb20337b8d7612e455a3cb3f145 +size 1711309864 diff --git a/model-00040-of-00081.safetensors b/model-00040-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..be692cd60fd6e5fe5c799ac16b57dcbda7950cd0 --- /dev/null +++ b/model-00040-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7fdaa14f8b6a3f8fccc017215fc57b4aa3375f8e4f29c55bd0496421ee3cbe8 +size 1711309864 diff --git a/model-00041-of-00081.safetensors b/model-00041-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..25526ec289fbf4c20fc42e3edcddfdf27218beda --- /dev/null +++ b/model-00041-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a94295f40e3a93083ab11756fc0c53a49fec1160d32ecb72fd58ea9b0c848f38 +size 1711309864 diff --git a/model-00042-of-00081.safetensors b/model-00042-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a1e60bb8f39c7a7e2a066c6d1c27724f8cc8680b --- /dev/null +++ b/model-00042-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:33c6217194ea34ef04fad00a2d760833646d8d6507f355b6bd2556123a8059f6 +size 1711309864 diff --git a/model-00043-of-00081.safetensors b/model-00043-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..de3bd8166a380f20290766276688f49c4b412720 --- /dev/null +++ b/model-00043-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a7083d206465589e0b81ac467250588e1ef6bdca905c4cfedf177d6e6a29de0 +size 1711309864 diff --git a/model-00044-of-00081.safetensors b/model-00044-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8080d17cc3ce45084a5eba059feb7a477b077e29 --- /dev/null +++ b/model-00044-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c8df90c99310dd115e18b0ca9c09f2265d4d07b31827f7180dddf02620054a34 +size 1711309864 diff --git a/model-00045-of-00081.safetensors b/model-00045-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3104db51ab4c407d90a40c0d4be571ed04ce1fe5 --- /dev/null +++ b/model-00045-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d78e1fc039a804cf08ab9fbc9f3fcf5990ed1b1a6523122f56eea372cc7a502d +size 1711309864 diff --git a/model-00046-of-00081.safetensors b/model-00046-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ca3706277d63a95a8e948e793de8ca54743700a8 --- /dev/null +++ b/model-00046-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:06dddee03932d104e2332258379749b8736b44419447ac24f39783ce9f36c55c +size 1711309856 diff --git a/model-00047-of-00081.safetensors b/model-00047-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..63e1dd12203f51c07f4a523569446b2a9cd4bb11 --- /dev/null +++ b/model-00047-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7f483db92e40ef7b2bdc3ae5275e88cc0633f699cce3f377da33da46d4f96809 +size 1711309856 diff --git a/model-00048-of-00081.safetensors b/model-00048-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3d534841d2341c26ae60e7a1cc26ea7b1b2d5c57 --- /dev/null +++ b/model-00048-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:93243acb92b63511ba9fd70aae7eae4cf719d8b7ae79d7377ea38e7f844552f4 +size 1711309864 diff --git a/model-00049-of-00081.safetensors b/model-00049-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a6afae8e8872c919e27500497c88c65be0495aeb --- /dev/null +++ b/model-00049-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a330a7865dbf0692436f6921498d3e5ac255284089af3a110951f125ac94c0cf +size 1711309864 diff --git a/model-00050-of-00081.safetensors b/model-00050-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c8e68c97dcf105b7fb94802aa1da36a4e55b3da6 --- /dev/null +++ b/model-00050-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:19d497886caab5be881c7c1384f1e513007ef168bc1c561b5a42bbdc57a4c8b7 +size 1711309864 diff --git a/model-00051-of-00081.safetensors b/model-00051-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b351d2e6b8441cbd2f63b18e6d2869bc63e7034c --- /dev/null +++ b/model-00051-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:692176b10606b04b1e9a27ffd5cb1142a3b8c2e6151c4b4ccbabdf2eff93a371 +size 1711309864 diff --git a/model-00052-of-00081.safetensors b/model-00052-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..abe52285e2b951235e585b94e3b7425d45a6180c --- /dev/null +++ b/model-00052-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7ce3800581fe796a27ebd8ab5967e48a702a7014a89655eb71926fa8a8ee718f +size 1711309864 diff --git a/model-00053-of-00081.safetensors b/model-00053-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..620a165f4b5d31203504947a2d3f4695d4fa391d --- /dev/null +++ b/model-00053-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f83450a03de6929e74900f9c743ecf991bc9bddeaecc49aa4f828071926ed09b +size 1711309864 diff --git a/model-00054-of-00081.safetensors b/model-00054-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3d4409eff264e8691ddfcbfb7ded13c36b89e0ba --- /dev/null +++ b/model-00054-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d342d6a048293d154e84e8d4159130e514a86114dff7d801647e21ec68614671 +size 1711309864 diff --git a/model-00055-of-00081.safetensors b/model-00055-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fc56477cb8f13afbbc6c031b0883227cc856c667 --- /dev/null +++ b/model-00055-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:84998855fabf924498d6008f1732529abdf881b782304096bf77d63515eb6ba9 +size 1711309864 diff --git a/model-00056-of-00081.safetensors b/model-00056-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dd359b4ac34a2beb43a4582b707c0c26897a11e8 --- /dev/null +++ b/model-00056-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2c005b00c35bf0d0ebbe9e4ae7cf393cbf67b4ec6dd8e6b1c7399e75a2a10af8 +size 1711309864 diff --git a/model-00057-of-00081.safetensors b/model-00057-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..26270fbfc9c3566cce119d5f19363293929cfb56 --- /dev/null +++ b/model-00057-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8408f2593422f4973c58e3e57014a93bc60d36779a7ad9f95b90325ac2a47bb9 +size 1711309856 diff --git a/model-00058-of-00081.safetensors b/model-00058-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..147e6a14825de1f7562c2a97ab186a3868dd5949 --- /dev/null +++ b/model-00058-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0fabceae042712f454f3153002ea0ac0a414b1ce48b209fbe03dbfc194d8cec3 +size 1711309856 diff --git a/model-00059-of-00081.safetensors b/model-00059-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..08bb22bd572b3b15d6ec6f3ad3e483b2a71d3d73 --- /dev/null +++ b/model-00059-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4ce05d05ebf6828922ebbc5aa79848addcb84f68f9928e0891d87f9be4773f72 +size 1711309864 diff --git a/model-00060-of-00081.safetensors b/model-00060-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..18e07a0d96d9ea30b740ee76b3a919e8b4a8470f --- /dev/null +++ b/model-00060-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:65d4479adc5c59b2266ec213d5d8200356f0933a5994b8fd7850c748f9cb2f9e +size 1711309864 diff --git a/model-00061-of-00081.safetensors b/model-00061-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..67d2ab032de540a1980e60dd5ba9a54b7e042a05 --- /dev/null +++ b/model-00061-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b0f510bf20a66388989da878618f4f68793020b9a5eaa2ad8d4c17d1fb1249be +size 1711309864 diff --git a/model-00062-of-00081.safetensors b/model-00062-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..37ec01e7a13a9c2cb36a885ea32ad07ae5a494c6 --- /dev/null +++ b/model-00062-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e341b40ac296a282a4d6cc5a3829fb716ec1dacfe860be090ca390b8e42c94fe +size 1711309864 diff --git a/model-00063-of-00081.safetensors b/model-00063-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..40a444837dceb0528d5b4f9014181939b7289eb9 --- /dev/null +++ b/model-00063-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:98780fa79ea1a0fc315c46177194de83354cf5255c8e8bcdf6c488d33bb89b47 +size 1711309864 diff --git a/model-00064-of-00081.safetensors b/model-00064-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dd745051ba0c0c7a64b2cdd7db6ec86eb22f877b --- /dev/null +++ b/model-00064-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b658f54ce1dab5341ba8c0510f06c0c934707f5801bf360e5b95c8b0e98d6fbe +size 1711309864 diff --git a/model-00065-of-00081.safetensors b/model-00065-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1dd14d8ff296dcde438444fa9faf8feef7de32b3 --- /dev/null +++ b/model-00065-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2eb132ab2c287081b9d64d9d36e1859145a05f0cac16c1bceb8f2b1b33c01bef +size 1711309864 diff --git a/model-00066-of-00081.safetensors b/model-00066-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..af3f7e4f4f77c939b6e6c11307ab41bb352d1519 --- /dev/null +++ b/model-00066-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:218736af232d262a8ee8441c91be5e172b1b4063b88068809ab79d8ccc9d6f3e +size 1711309864 diff --git a/model-00067-of-00081.safetensors b/model-00067-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..894f64ec7a77dab060d0d47b60edbac26b053908 --- /dev/null +++ b/model-00067-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1274298022e225b13f3b4f0753da74e87cc2f370156b15c12b51ce1e24b09e0e +size 1711309864 diff --git a/model-00068-of-00081.safetensors b/model-00068-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b5c488aacb988d854ea9ee090de7823c4561538c --- /dev/null +++ b/model-00068-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:55cef7fcf2bcdb94ec66bb2ba19d1f567c66dc47a7979dfa0a928f0927184221 +size 1711309856 diff --git a/model-00069-of-00081.safetensors b/model-00069-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a172c74221a6e39b6719d20786d6a5b631c1af06 --- /dev/null +++ b/model-00069-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:95dae982471a40e02ee93eb1f6ce8515a406514f597de5af9b9e26c2795115db +size 1711309856 diff --git a/model-00070-of-00081.safetensors b/model-00070-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a2776e052356d1790548bf5038b257fab02d785f --- /dev/null +++ b/model-00070-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ee2e13f581f9d274675306d1866e8bdfc9d5aa148f3b2d46e1c865b3c9ea819c +size 1711309864 diff --git a/model-00071-of-00081.safetensors b/model-00071-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d2fb8d06b0a3325b7304df1375dd4f8acc82cde2 --- /dev/null +++ b/model-00071-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e430b427d52162af37599e739b0f3696c2113352617ccd1f70482e38c3d8badd +size 1711309864 diff --git a/model-00072-of-00081.safetensors b/model-00072-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5f220a13f5b4692b8733cc7003d2c844d7977868 --- /dev/null +++ b/model-00072-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fcf176a80602229f388953b65f0195487d867ff2d712ab5294898ff066f5779c +size 1711309864 diff --git a/model-00073-of-00081.safetensors b/model-00073-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..47c411f6a46ba25670ba6df6bea9b2ad1d192642 --- /dev/null +++ b/model-00073-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59f6fa9bc3cae746a628619f7b721fb106f467302530769a9373481f31215acf +size 1711309864 diff --git a/model-00074-of-00081.safetensors b/model-00074-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0296cf7c204ba953583aeafc5e20f515b18a97e8 --- /dev/null +++ b/model-00074-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0a6a9d344e29ea5c03290c0969bf77199540e1a17c7e88d5fd19dffd7c52f008 +size 1711309864 diff --git a/model-00075-of-00081.safetensors b/model-00075-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ce652c9b5f49a05cc09efe97cbc36c3a73117361 --- /dev/null +++ b/model-00075-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d7bd9e812de327ff0dcc81991024e06f73e4393346ff61bd02d703ee09bcb9ea +size 1711309864 diff --git a/model-00076-of-00081.safetensors b/model-00076-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..775db10e1782521475abfd057b358baaf42c843a --- /dev/null +++ b/model-00076-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1478496b6463017ebcacb389066ae88c34a29d140327fe0a043a7f5447b488df +size 1711309864 diff --git a/model-00077-of-00081.safetensors b/model-00077-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7bd80b70cae02a33c66ca91ca0ff1fb14bb0bef6 --- /dev/null +++ b/model-00077-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e2a0890a77a423122a73757f32424bdcf1433b6b4612b11781ecf87675b9a250 +size 1711309864 diff --git a/model-00078-of-00081.safetensors b/model-00078-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2ebef7badfbf47c334ec694340e6b841eb887ec3 --- /dev/null +++ b/model-00078-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:201a2b93918fbdfcb203e5b0b715908e7fe9d2852a9dcd3902ada3fdba04d0d8 +size 1711309864 diff --git a/model-00079-of-00081.safetensors b/model-00079-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..24e7d1213396db0235bf877d950184d5db96d4a1 --- /dev/null +++ b/model-00079-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e7b395632ec12663121fb9aa20ba41913e3bd18b39b24b7904167508ddd80f64 +size 1711309856 diff --git a/model-00080-of-00081.safetensors b/model-00080-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9a617208aa4e2dde0eefeaadacce4607357e989e --- /dev/null +++ b/model-00080-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7e056106d21e4af57610e5636a4bd4bb49e094dde717b745696bd8a3dad90e0e +size 1711309856 diff --git a/model-00081-of-00081.safetensors b/model-00081-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e1aa4b6f306860393ca9cea9ff27a87fdcd1dfd2 --- /dev/null +++ b/model-00081-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4abd3a230ce3125102263ad6b0fe8272df2124647392c5171c7fe692098f67e1 +size 771785512 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..b3f7453d1e09f1f02cf9faec73a5e1112639dc4f --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1 @@ +{"metadata": {"mergekit_version": "0.0.4.2", "total_size": 137953296384}, "weight_map": {"lm_head.weight": "model-00001-of-00081.safetensors", "model.embed_tokens.weight": "model-00001-of-00081.safetensors", "model.layers.0.input_layernorm.weight": "model-00001-of-00081.safetensors", "model.layers.0.mlp.down_proj.weight": "model-00001-of-00081.safetensors", "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00081.safetensors", "model.layers.0.mlp.up_proj.weight": "model-00002-of-00081.safetensors", "model.layers.0.post_attention_layernorm.weight": "model-00002-of-00081.safetensors", "model.layers.0.self_attn.k_proj.weight": "model-00002-of-00081.safetensors", "model.layers.0.self_attn.o_proj.weight": "model-00002-of-00081.safetensors", "model.layers.0.self_attn.q_proj.weight": "model-00002-of-00081.safetensors", "model.layers.0.self_attn.v_proj.weight": "model-00002-of-00081.safetensors", "model.layers.1.input_layernorm.weight": "model-00002-of-00081.safetensors", "model.layers.1.mlp.down_proj.weight": "model-00002-of-00081.safetensors", "model.layers.1.mlp.gate_proj.weight": "model-00002-of-00081.safetensors", "model.layers.1.mlp.up_proj.weight": "model-00003-of-00081.safetensors", "model.layers.1.post_attention_layernorm.weight": "model-00003-of-00081.safetensors", "model.layers.1.self_attn.k_proj.weight": "model-00003-of-00081.safetensors", "model.layers.1.self_attn.o_proj.weight": "model-00003-of-00081.safetensors", "model.layers.1.self_attn.q_proj.weight": "model-00003-of-00081.safetensors", "model.layers.1.self_attn.v_proj.weight": "model-00003-of-00081.safetensors", "model.layers.10.input_layernorm.weight": "model-00003-of-00081.safetensors", "model.layers.10.mlp.down_proj.weight": "model-00003-of-00081.safetensors", "model.layers.10.mlp.gate_proj.weight": "model-00003-of-00081.safetensors", "model.layers.10.mlp.up_proj.weight": "model-00004-of-00081.safetensors", "model.layers.10.post_attention_layernorm.weight": "model-00004-of-00081.safetensors", "model.layers.10.self_attn.k_proj.weight": "model-00004-of-00081.safetensors", "model.layers.10.self_attn.o_proj.weight": "model-00004-of-00081.safetensors", "model.layers.10.self_attn.q_proj.weight": "model-00004-of-00081.safetensors", "model.layers.10.self_attn.v_proj.weight": "model-00004-of-00081.safetensors", "model.layers.11.input_layernorm.weight": "model-00004-of-00081.safetensors", "model.layers.11.mlp.down_proj.weight": "model-00004-of-00081.safetensors", "model.layers.11.mlp.gate_proj.weight": "model-00004-of-00081.safetensors", "model.layers.11.mlp.up_proj.weight": "model-00005-of-00081.safetensors", "model.layers.11.post_attention_layernorm.weight": "model-00005-of-00081.safetensors", "model.layers.11.self_attn.k_proj.weight": "model-00005-of-00081.safetensors", "model.layers.11.self_attn.o_proj.weight": "model-00005-of-00081.safetensors", "model.layers.11.self_attn.q_proj.weight": "model-00005-of-00081.safetensors", "model.layers.11.self_attn.v_proj.weight": "model-00005-of-00081.safetensors", "model.layers.12.input_layernorm.weight": "model-00005-of-00081.safetensors", "model.layers.12.mlp.down_proj.weight": "model-00005-of-00081.safetensors", "model.layers.12.mlp.gate_proj.weight": "model-00005-of-00081.safetensors", "model.layers.12.mlp.up_proj.weight": "model-00006-of-00081.safetensors", "model.layers.12.post_attention_layernorm.weight": "model-00006-of-00081.safetensors", "model.layers.12.self_attn.k_proj.weight": "model-00006-of-00081.safetensors", "model.layers.12.self_attn.o_proj.weight": "model-00006-of-00081.safetensors", "model.layers.12.self_attn.q_proj.weight": "model-00006-of-00081.safetensors", "model.layers.12.self_attn.v_proj.weight": "model-00006-of-00081.safetensors", "model.layers.13.input_layernorm.weight": "model-00006-of-00081.safetensors", "model.layers.13.mlp.down_proj.weight": "model-00006-of-00081.safetensors", "model.layers.13.mlp.gate_proj.weight": "model-00006-of-00081.safetensors", "model.layers.13.mlp.up_proj.weight": "model-00007-of-00081.safetensors", "model.layers.13.post_attention_layernorm.weight": "model-00007-of-00081.safetensors", "model.layers.13.self_attn.k_proj.weight": "model-00007-of-00081.safetensors", "model.layers.13.self_attn.o_proj.weight": "model-00007-of-00081.safetensors", "model.layers.13.self_attn.q_proj.weight": "model-00007-of-00081.safetensors", "model.layers.13.self_attn.v_proj.weight": "model-00007-of-00081.safetensors", "model.layers.14.input_layernorm.weight": "model-00007-of-00081.safetensors", "model.layers.14.mlp.down_proj.weight": "model-00007-of-00081.safetensors", "model.layers.14.mlp.gate_proj.weight": "model-00007-of-00081.safetensors", "model.layers.14.mlp.up_proj.weight": "model-00008-of-00081.safetensors", "model.layers.14.post_attention_layernorm.weight": "model-00008-of-00081.safetensors", "model.layers.14.self_attn.k_proj.weight": "model-00008-of-00081.safetensors", "model.layers.14.self_attn.o_proj.weight": "model-00008-of-00081.safetensors", "model.layers.14.self_attn.q_proj.weight": "model-00008-of-00081.safetensors", "model.layers.14.self_attn.v_proj.weight": "model-00008-of-00081.safetensors", "model.layers.15.input_layernorm.weight": "model-00008-of-00081.safetensors", "model.layers.15.mlp.down_proj.weight": "model-00008-of-00081.safetensors", "model.layers.15.mlp.gate_proj.weight": "model-00008-of-00081.safetensors", "model.layers.15.mlp.up_proj.weight": "model-00009-of-00081.safetensors", "model.layers.15.post_attention_layernorm.weight": "model-00009-of-00081.safetensors", "model.layers.15.self_attn.k_proj.weight": "model-00009-of-00081.safetensors", "model.layers.15.self_attn.o_proj.weight": "model-00009-of-00081.safetensors", "model.layers.15.self_attn.q_proj.weight": "model-00009-of-00081.safetensors", "model.layers.15.self_attn.v_proj.weight": "model-00009-of-00081.safetensors", "model.layers.16.input_layernorm.weight": "model-00009-of-00081.safetensors", "model.layers.16.mlp.down_proj.weight": "model-00009-of-00081.safetensors", "model.layers.16.mlp.gate_proj.weight": "model-00009-of-00081.safetensors", "model.layers.16.mlp.up_proj.weight": "model-00010-of-00081.safetensors", "model.layers.16.post_attention_layernorm.weight": "model-00010-of-00081.safetensors", "model.layers.16.self_attn.k_proj.weight": "model-00010-of-00081.safetensors", "model.layers.16.self_attn.o_proj.weight": "model-00010-of-00081.safetensors", "model.layers.16.self_attn.q_proj.weight": "model-00010-of-00081.safetensors", "model.layers.16.self_attn.v_proj.weight": "model-00010-of-00081.safetensors", "model.layers.17.input_layernorm.weight": "model-00010-of-00081.safetensors", "model.layers.17.mlp.down_proj.weight": "model-00010-of-00081.safetensors", "model.layers.17.mlp.gate_proj.weight": "model-00010-of-00081.safetensors", "model.layers.17.mlp.up_proj.weight": "model-00011-of-00081.safetensors", "model.layers.17.post_attention_layernorm.weight": "model-00011-of-00081.safetensors", "model.layers.17.self_attn.k_proj.weight": "model-00011-of-00081.safetensors", "model.layers.17.self_attn.o_proj.weight": "model-00011-of-00081.safetensors", "model.layers.17.self_attn.q_proj.weight": "model-00011-of-00081.safetensors", "model.layers.17.self_attn.v_proj.weight": "model-00011-of-00081.safetensors", "model.layers.18.input_layernorm.weight": "model-00011-of-00081.safetensors", "model.layers.18.mlp.down_proj.weight": "model-00011-of-00081.safetensors", "model.layers.18.mlp.gate_proj.weight": "model-00011-of-00081.safetensors", "model.layers.18.mlp.up_proj.weight": "model-00012-of-00081.safetensors", "model.layers.18.post_attention_layernorm.weight": "model-00012-of-00081.safetensors", "model.layers.18.self_attn.k_proj.weight": "model-00012-of-00081.safetensors", "model.layers.18.self_attn.o_proj.weight": "model-00012-of-00081.safetensors", "model.layers.18.self_attn.q_proj.weight": "model-00012-of-00081.safetensors", "model.layers.18.self_attn.v_proj.weight": "model-00012-of-00081.safetensors", "model.layers.19.input_layernorm.weight": "model-00012-of-00081.safetensors", "model.layers.19.mlp.down_proj.weight": "model-00012-of-00081.safetensors", "model.layers.19.mlp.gate_proj.weight": "model-00012-of-00081.safetensors", "model.layers.19.mlp.up_proj.weight": "model-00013-of-00081.safetensors", "model.layers.19.post_attention_layernorm.weight": "model-00013-of-00081.safetensors", "model.layers.19.self_attn.k_proj.weight": "model-00013-of-00081.safetensors", "model.layers.19.self_attn.o_proj.weight": "model-00013-of-00081.safetensors", "model.layers.19.self_attn.q_proj.weight": "model-00013-of-00081.safetensors", "model.layers.19.self_attn.v_proj.weight": "model-00013-of-00081.safetensors", "model.layers.2.input_layernorm.weight": "model-00013-of-00081.safetensors", "model.layers.2.mlp.down_proj.weight": "model-00013-of-00081.safetensors", "model.layers.2.mlp.gate_proj.weight": "model-00013-of-00081.safetensors", "model.layers.2.mlp.up_proj.weight": "model-00014-of-00081.safetensors", "model.layers.2.post_attention_layernorm.weight": "model-00014-of-00081.safetensors", "model.layers.2.self_attn.k_proj.weight": "model-00014-of-00081.safetensors", "model.layers.2.self_attn.o_proj.weight": "model-00014-of-00081.safetensors", "model.layers.2.self_attn.q_proj.weight": "model-00014-of-00081.safetensors", "model.layers.2.self_attn.v_proj.weight": "model-00014-of-00081.safetensors", "model.layers.20.input_layernorm.weight": "model-00014-of-00081.safetensors", "model.layers.20.mlp.down_proj.weight": "model-00014-of-00081.safetensors", "model.layers.20.mlp.gate_proj.weight": "model-00014-of-00081.safetensors", "model.layers.20.mlp.up_proj.weight": "model-00015-of-00081.safetensors", "model.layers.20.post_attention_layernorm.weight": "model-00015-of-00081.safetensors", "model.layers.20.self_attn.k_proj.weight": "model-00015-of-00081.safetensors", "model.layers.20.self_attn.o_proj.weight": "model-00015-of-00081.safetensors", "model.layers.20.self_attn.q_proj.weight": "model-00015-of-00081.safetensors", "model.layers.20.self_attn.v_proj.weight": "model-00015-of-00081.safetensors", "model.layers.21.input_layernorm.weight": "model-00015-of-00081.safetensors", "model.layers.21.mlp.down_proj.weight": "model-00015-of-00081.safetensors", "model.layers.21.mlp.gate_proj.weight": "model-00015-of-00081.safetensors", "model.layers.21.mlp.up_proj.weight": "model-00016-of-00081.safetensors", "model.layers.21.post_attention_layernorm.weight": "model-00016-of-00081.safetensors", "model.layers.21.self_attn.k_proj.weight": "model-00016-of-00081.safetensors", "model.layers.21.self_attn.o_proj.weight": "model-00016-of-00081.safetensors", "model.layers.21.self_attn.q_proj.weight": "model-00016-of-00081.safetensors", "model.layers.21.self_attn.v_proj.weight": "model-00016-of-00081.safetensors", "model.layers.22.input_layernorm.weight": "model-00016-of-00081.safetensors", "model.layers.22.mlp.down_proj.weight": "model-00016-of-00081.safetensors", "model.layers.22.mlp.gate_proj.weight": "model-00016-of-00081.safetensors", "model.layers.22.mlp.up_proj.weight": "model-00017-of-00081.safetensors", "model.layers.22.post_attention_layernorm.weight": "model-00017-of-00081.safetensors", "model.layers.22.self_attn.k_proj.weight": "model-00017-of-00081.safetensors", "model.layers.22.self_attn.o_proj.weight": "model-00017-of-00081.safetensors", "model.layers.22.self_attn.q_proj.weight": "model-00017-of-00081.safetensors", "model.layers.22.self_attn.v_proj.weight": "model-00017-of-00081.safetensors", "model.layers.23.input_layernorm.weight": "model-00017-of-00081.safetensors", "model.layers.23.mlp.down_proj.weight": "model-00017-of-00081.safetensors", "model.layers.23.mlp.gate_proj.weight": "model-00017-of-00081.safetensors", "model.layers.23.mlp.up_proj.weight": "model-00018-of-00081.safetensors", "model.layers.23.post_attention_layernorm.weight": "model-00018-of-00081.safetensors", "model.layers.23.self_attn.k_proj.weight": "model-00018-of-00081.safetensors", "model.layers.23.self_attn.o_proj.weight": "model-00018-of-00081.safetensors", "model.layers.23.self_attn.q_proj.weight": "model-00018-of-00081.safetensors", "model.layers.23.self_attn.v_proj.weight": "model-00018-of-00081.safetensors", "model.layers.24.input_layernorm.weight": "model-00018-of-00081.safetensors", "model.layers.24.mlp.down_proj.weight": "model-00018-of-00081.safetensors", "model.layers.24.mlp.gate_proj.weight": "model-00018-of-00081.safetensors", "model.layers.24.mlp.up_proj.weight": "model-00019-of-00081.safetensors", "model.layers.24.post_attention_layernorm.weight": "model-00019-of-00081.safetensors", "model.layers.24.self_attn.k_proj.weight": "model-00019-of-00081.safetensors", "model.layers.24.self_attn.o_proj.weight": "model-00019-of-00081.safetensors", "model.layers.24.self_attn.q_proj.weight": "model-00019-of-00081.safetensors", "model.layers.24.self_attn.v_proj.weight": "model-00019-of-00081.safetensors", "model.layers.25.input_layernorm.weight": "model-00019-of-00081.safetensors", "model.layers.25.mlp.down_proj.weight": "model-00019-of-00081.safetensors", "model.layers.25.mlp.gate_proj.weight": "model-00019-of-00081.safetensors", "model.layers.25.mlp.up_proj.weight": "model-00020-of-00081.safetensors", "model.layers.25.post_attention_layernorm.weight": "model-00020-of-00081.safetensors", "model.layers.25.self_attn.k_proj.weight": "model-00020-of-00081.safetensors", "model.layers.25.self_attn.o_proj.weight": "model-00020-of-00081.safetensors", "model.layers.25.self_attn.q_proj.weight": "model-00020-of-00081.safetensors", "model.layers.25.self_attn.v_proj.weight": "model-00020-of-00081.safetensors", "model.layers.26.input_layernorm.weight": "model-00020-of-00081.safetensors", "model.layers.26.mlp.down_proj.weight": "model-00020-of-00081.safetensors", "model.layers.26.mlp.gate_proj.weight": "model-00020-of-00081.safetensors", "model.layers.26.mlp.up_proj.weight": "model-00021-of-00081.safetensors", "model.layers.26.post_attention_layernorm.weight": "model-00021-of-00081.safetensors", "model.layers.26.self_attn.k_proj.weight": "model-00021-of-00081.safetensors", "model.layers.26.self_attn.o_proj.weight": "model-00021-of-00081.safetensors", "model.layers.26.self_attn.q_proj.weight": "model-00021-of-00081.safetensors", "model.layers.26.self_attn.v_proj.weight": "model-00021-of-00081.safetensors", "model.layers.27.input_layernorm.weight": "model-00021-of-00081.safetensors", "model.layers.27.mlp.down_proj.weight": "model-00021-of-00081.safetensors", "model.layers.27.mlp.gate_proj.weight": "model-00021-of-00081.safetensors", "model.layers.27.mlp.up_proj.weight": "model-00022-of-00081.safetensors", "model.layers.27.post_attention_layernorm.weight": "model-00022-of-00081.safetensors", "model.layers.27.self_attn.k_proj.weight": "model-00022-of-00081.safetensors", "model.layers.27.self_attn.o_proj.weight": "model-00022-of-00081.safetensors", "model.layers.27.self_attn.q_proj.weight": "model-00022-of-00081.safetensors", "model.layers.27.self_attn.v_proj.weight": "model-00022-of-00081.safetensors", "model.layers.28.input_layernorm.weight": "model-00022-of-00081.safetensors", "model.layers.28.mlp.down_proj.weight": "model-00022-of-00081.safetensors", "model.layers.28.mlp.gate_proj.weight": "model-00022-of-00081.safetensors", "model.layers.28.mlp.up_proj.weight": "model-00023-of-00081.safetensors", "model.layers.28.post_attention_layernorm.weight": "model-00023-of-00081.safetensors", "model.layers.28.self_attn.k_proj.weight": "model-00023-of-00081.safetensors", "model.layers.28.self_attn.o_proj.weight": "model-00023-of-00081.safetensors", "model.layers.28.self_attn.q_proj.weight": "model-00023-of-00081.safetensors", "model.layers.28.self_attn.v_proj.weight": "model-00023-of-00081.safetensors", "model.layers.29.input_layernorm.weight": "model-00023-of-00081.safetensors", "model.layers.29.mlp.down_proj.weight": "model-00023-of-00081.safetensors", "model.layers.29.mlp.gate_proj.weight": "model-00023-of-00081.safetensors", "model.layers.29.mlp.up_proj.weight": "model-00024-of-00081.safetensors", "model.layers.29.post_attention_layernorm.weight": "model-00024-of-00081.safetensors", "model.layers.29.self_attn.k_proj.weight": "model-00024-of-00081.safetensors", "model.layers.29.self_attn.o_proj.weight": "model-00024-of-00081.safetensors", "model.layers.29.self_attn.q_proj.weight": "model-00024-of-00081.safetensors", "model.layers.29.self_attn.v_proj.weight": "model-00024-of-00081.safetensors", "model.layers.3.input_layernorm.weight": "model-00024-of-00081.safetensors", "model.layers.3.mlp.down_proj.weight": "model-00024-of-00081.safetensors", "model.layers.3.mlp.gate_proj.weight": "model-00024-of-00081.safetensors", "model.layers.3.mlp.up_proj.weight": "model-00025-of-00081.safetensors", "model.layers.3.post_attention_layernorm.weight": "model-00025-of-00081.safetensors", "model.layers.3.self_attn.k_proj.weight": "model-00025-of-00081.safetensors", "model.layers.3.self_attn.o_proj.weight": "model-00025-of-00081.safetensors", "model.layers.3.self_attn.q_proj.weight": "model-00025-of-00081.safetensors", "model.layers.3.self_attn.v_proj.weight": "model-00025-of-00081.safetensors", "model.layers.30.input_layernorm.weight": "model-00025-of-00081.safetensors", "model.layers.30.mlp.down_proj.weight": "model-00025-of-00081.safetensors", "model.layers.30.mlp.gate_proj.weight": "model-00025-of-00081.safetensors", "model.layers.30.mlp.up_proj.weight": "model-00026-of-00081.safetensors", "model.layers.30.post_attention_layernorm.weight": "model-00026-of-00081.safetensors", "model.layers.30.self_attn.k_proj.weight": "model-00026-of-00081.safetensors", "model.layers.30.self_attn.o_proj.weight": "model-00026-of-00081.safetensors", "model.layers.30.self_attn.q_proj.weight": "model-00026-of-00081.safetensors", "model.layers.30.self_attn.v_proj.weight": "model-00026-of-00081.safetensors", "model.layers.31.input_layernorm.weight": "model-00026-of-00081.safetensors", "model.layers.31.mlp.down_proj.weight": "model-00026-of-00081.safetensors", "model.layers.31.mlp.gate_proj.weight": "model-00026-of-00081.safetensors", "model.layers.31.mlp.up_proj.weight": "model-00027-of-00081.safetensors", "model.layers.31.post_attention_layernorm.weight": "model-00027-of-00081.safetensors", "model.layers.31.self_attn.k_proj.weight": "model-00027-of-00081.safetensors", "model.layers.31.self_attn.o_proj.weight": "model-00027-of-00081.safetensors", "model.layers.31.self_attn.q_proj.weight": "model-00027-of-00081.safetensors", "model.layers.31.self_attn.v_proj.weight": "model-00027-of-00081.safetensors", "model.layers.32.input_layernorm.weight": "model-00027-of-00081.safetensors", "model.layers.32.mlp.down_proj.weight": "model-00027-of-00081.safetensors", "model.layers.32.mlp.gate_proj.weight": "model-00027-of-00081.safetensors", "model.layers.32.mlp.up_proj.weight": "model-00028-of-00081.safetensors", "model.layers.32.post_attention_layernorm.weight": "model-00028-of-00081.safetensors", "model.layers.32.self_attn.k_proj.weight": "model-00028-of-00081.safetensors", "model.layers.32.self_attn.o_proj.weight": "model-00028-of-00081.safetensors", "model.layers.32.self_attn.q_proj.weight": "model-00028-of-00081.safetensors", "model.layers.32.self_attn.v_proj.weight": "model-00028-of-00081.safetensors", "model.layers.33.input_layernorm.weight": "model-00028-of-00081.safetensors", "model.layers.33.mlp.down_proj.weight": "model-00028-of-00081.safetensors", "model.layers.33.mlp.gate_proj.weight": "model-00028-of-00081.safetensors", "model.layers.33.mlp.up_proj.weight": "model-00029-of-00081.safetensors", "model.layers.33.post_attention_layernorm.weight": "model-00029-of-00081.safetensors", "model.layers.33.self_attn.k_proj.weight": "model-00029-of-00081.safetensors", "model.layers.33.self_attn.o_proj.weight": "model-00029-of-00081.safetensors", "model.layers.33.self_attn.q_proj.weight": "model-00029-of-00081.safetensors", "model.layers.33.self_attn.v_proj.weight": "model-00029-of-00081.safetensors", "model.layers.34.input_layernorm.weight": "model-00029-of-00081.safetensors", "model.layers.34.mlp.down_proj.weight": "model-00029-of-00081.safetensors", "model.layers.34.mlp.gate_proj.weight": "model-00029-of-00081.safetensors", "model.layers.34.mlp.up_proj.weight": "model-00030-of-00081.safetensors", "model.layers.34.post_attention_layernorm.weight": "model-00030-of-00081.safetensors", "model.layers.34.self_attn.k_proj.weight": "model-00030-of-00081.safetensors", "model.layers.34.self_attn.o_proj.weight": "model-00030-of-00081.safetensors", "model.layers.34.self_attn.q_proj.weight": "model-00030-of-00081.safetensors", "model.layers.34.self_attn.v_proj.weight": "model-00030-of-00081.safetensors", "model.layers.35.input_layernorm.weight": "model-00030-of-00081.safetensors", "model.layers.35.mlp.down_proj.weight": "model-00030-of-00081.safetensors", "model.layers.35.mlp.gate_proj.weight": "model-00030-of-00081.safetensors", "model.layers.35.mlp.up_proj.weight": "model-00031-of-00081.safetensors", "model.layers.35.post_attention_layernorm.weight": "model-00031-of-00081.safetensors", "model.layers.35.self_attn.k_proj.weight": "model-00031-of-00081.safetensors", "model.layers.35.self_attn.o_proj.weight": "model-00031-of-00081.safetensors", "model.layers.35.self_attn.q_proj.weight": "model-00031-of-00081.safetensors", "model.layers.35.self_attn.v_proj.weight": "model-00031-of-00081.safetensors", "model.layers.36.input_layernorm.weight": "model-00031-of-00081.safetensors", "model.layers.36.mlp.down_proj.weight": "model-00031-of-00081.safetensors", "model.layers.36.mlp.gate_proj.weight": "model-00031-of-00081.safetensors", "model.layers.36.mlp.up_proj.weight": "model-00032-of-00081.safetensors", "model.layers.36.post_attention_layernorm.weight": "model-00032-of-00081.safetensors", "model.layers.36.self_attn.k_proj.weight": "model-00032-of-00081.safetensors", "model.layers.36.self_attn.o_proj.weight": "model-00032-of-00081.safetensors", "model.layers.36.self_attn.q_proj.weight": "model-00032-of-00081.safetensors", "model.layers.36.self_attn.v_proj.weight": "model-00032-of-00081.safetensors", "model.layers.37.input_layernorm.weight": "model-00032-of-00081.safetensors", "model.layers.37.mlp.down_proj.weight": "model-00032-of-00081.safetensors", "model.layers.37.mlp.gate_proj.weight": "model-00032-of-00081.safetensors", "model.layers.37.mlp.up_proj.weight": "model-00033-of-00081.safetensors", "model.layers.37.post_attention_layernorm.weight": "model-00033-of-00081.safetensors", "model.layers.37.self_attn.k_proj.weight": "model-00033-of-00081.safetensors", "model.layers.37.self_attn.o_proj.weight": "model-00033-of-00081.safetensors", "model.layers.37.self_attn.q_proj.weight": "model-00033-of-00081.safetensors", "model.layers.37.self_attn.v_proj.weight": "model-00033-of-00081.safetensors", "model.layers.38.input_layernorm.weight": "model-00033-of-00081.safetensors", "model.layers.38.mlp.down_proj.weight": "model-00033-of-00081.safetensors", "model.layers.38.mlp.gate_proj.weight": "model-00033-of-00081.safetensors", "model.layers.38.mlp.up_proj.weight": "model-00034-of-00081.safetensors", "model.layers.38.post_attention_layernorm.weight": "model-00034-of-00081.safetensors", "model.layers.38.self_attn.k_proj.weight": "model-00034-of-00081.safetensors", "model.layers.38.self_attn.o_proj.weight": "model-00034-of-00081.safetensors", "model.layers.38.self_attn.q_proj.weight": "model-00034-of-00081.safetensors", "model.layers.38.self_attn.v_proj.weight": "model-00034-of-00081.safetensors", "model.layers.39.input_layernorm.weight": "model-00034-of-00081.safetensors", "model.layers.39.mlp.down_proj.weight": "model-00034-of-00081.safetensors", "model.layers.39.mlp.gate_proj.weight": "model-00034-of-00081.safetensors", "model.layers.39.mlp.up_proj.weight": "model-00035-of-00081.safetensors", "model.layers.39.post_attention_layernorm.weight": "model-00035-of-00081.safetensors", "model.layers.39.self_attn.k_proj.weight": "model-00035-of-00081.safetensors", "model.layers.39.self_attn.o_proj.weight": "model-00035-of-00081.safetensors", "model.layers.39.self_attn.q_proj.weight": "model-00035-of-00081.safetensors", "model.layers.39.self_attn.v_proj.weight": "model-00035-of-00081.safetensors", "model.layers.4.input_layernorm.weight": "model-00035-of-00081.safetensors", "model.layers.4.mlp.down_proj.weight": "model-00035-of-00081.safetensors", "model.layers.4.mlp.gate_proj.weight": "model-00035-of-00081.safetensors", "model.layers.4.mlp.up_proj.weight": "model-00036-of-00081.safetensors", "model.layers.4.post_attention_layernorm.weight": "model-00036-of-00081.safetensors", "model.layers.4.self_attn.k_proj.weight": "model-00036-of-00081.safetensors", "model.layers.4.self_attn.o_proj.weight": "model-00036-of-00081.safetensors", "model.layers.4.self_attn.q_proj.weight": "model-00036-of-00081.safetensors", "model.layers.4.self_attn.v_proj.weight": "model-00036-of-00081.safetensors", "model.layers.40.input_layernorm.weight": "model-00036-of-00081.safetensors", "model.layers.40.mlp.down_proj.weight": "model-00036-of-00081.safetensors", "model.layers.40.mlp.gate_proj.weight": "model-00036-of-00081.safetensors", "model.layers.40.mlp.up_proj.weight": "model-00037-of-00081.safetensors", "model.layers.40.post_attention_layernorm.weight": "model-00037-of-00081.safetensors", "model.layers.40.self_attn.k_proj.weight": "model-00037-of-00081.safetensors", "model.layers.40.self_attn.o_proj.weight": "model-00037-of-00081.safetensors", "model.layers.40.self_attn.q_proj.weight": "model-00037-of-00081.safetensors", "model.layers.40.self_attn.v_proj.weight": "model-00037-of-00081.safetensors", "model.layers.41.input_layernorm.weight": "model-00037-of-00081.safetensors", "model.layers.41.mlp.down_proj.weight": "model-00037-of-00081.safetensors", "model.layers.41.mlp.gate_proj.weight": "model-00037-of-00081.safetensors", "model.layers.41.mlp.up_proj.weight": "model-00038-of-00081.safetensors", "model.layers.41.post_attention_layernorm.weight": "model-00038-of-00081.safetensors", "model.layers.41.self_attn.k_proj.weight": "model-00038-of-00081.safetensors", "model.layers.41.self_attn.o_proj.weight": "model-00038-of-00081.safetensors", "model.layers.41.self_attn.q_proj.weight": "model-00038-of-00081.safetensors", "model.layers.41.self_attn.v_proj.weight": "model-00038-of-00081.safetensors", "model.layers.42.input_layernorm.weight": "model-00038-of-00081.safetensors", "model.layers.42.mlp.down_proj.weight": "model-00038-of-00081.safetensors", "model.layers.42.mlp.gate_proj.weight": "model-00038-of-00081.safetensors", "model.layers.42.mlp.up_proj.weight": "model-00039-of-00081.safetensors", "model.layers.42.post_attention_layernorm.weight": "model-00039-of-00081.safetensors", "model.layers.42.self_attn.k_proj.weight": "model-00039-of-00081.safetensors", "model.layers.42.self_attn.o_proj.weight": "model-00039-of-00081.safetensors", "model.layers.42.self_attn.q_proj.weight": "model-00039-of-00081.safetensors", "model.layers.42.self_attn.v_proj.weight": "model-00039-of-00081.safetensors", "model.layers.43.input_layernorm.weight": "model-00039-of-00081.safetensors", "model.layers.43.mlp.down_proj.weight": "model-00039-of-00081.safetensors", "model.layers.43.mlp.gate_proj.weight": "model-00039-of-00081.safetensors", "model.layers.43.mlp.up_proj.weight": "model-00040-of-00081.safetensors", "model.layers.43.post_attention_layernorm.weight": "model-00040-of-00081.safetensors", "model.layers.43.self_attn.k_proj.weight": "model-00040-of-00081.safetensors", "model.layers.43.self_attn.o_proj.weight": "model-00040-of-00081.safetensors", "model.layers.43.self_attn.q_proj.weight": "model-00040-of-00081.safetensors", "model.layers.43.self_attn.v_proj.weight": "model-00040-of-00081.safetensors", "model.layers.44.input_layernorm.weight": "model-00040-of-00081.safetensors", "model.layers.44.mlp.down_proj.weight": "model-00040-of-00081.safetensors", "model.layers.44.mlp.gate_proj.weight": "model-00040-of-00081.safetensors", "model.layers.44.mlp.up_proj.weight": "model-00041-of-00081.safetensors", "model.layers.44.post_attention_layernorm.weight": "model-00041-of-00081.safetensors", "model.layers.44.self_attn.k_proj.weight": "model-00041-of-00081.safetensors", "model.layers.44.self_attn.o_proj.weight": "model-00041-of-00081.safetensors", "model.layers.44.self_attn.q_proj.weight": "model-00041-of-00081.safetensors", "model.layers.44.self_attn.v_proj.weight": "model-00041-of-00081.safetensors", "model.layers.45.input_layernorm.weight": "model-00041-of-00081.safetensors", "model.layers.45.mlp.down_proj.weight": "model-00041-of-00081.safetensors", "model.layers.45.mlp.gate_proj.weight": "model-00041-of-00081.safetensors", "model.layers.45.mlp.up_proj.weight": "model-00042-of-00081.safetensors", "model.layers.45.post_attention_layernorm.weight": "model-00042-of-00081.safetensors", "model.layers.45.self_attn.k_proj.weight": "model-00042-of-00081.safetensors", "model.layers.45.self_attn.o_proj.weight": "model-00042-of-00081.safetensors", "model.layers.45.self_attn.q_proj.weight": "model-00042-of-00081.safetensors", "model.layers.45.self_attn.v_proj.weight": "model-00042-of-00081.safetensors", "model.layers.46.input_layernorm.weight": "model-00042-of-00081.safetensors", "model.layers.46.mlp.down_proj.weight": "model-00042-of-00081.safetensors", "model.layers.46.mlp.gate_proj.weight": "model-00042-of-00081.safetensors", "model.layers.46.mlp.up_proj.weight": "model-00043-of-00081.safetensors", "model.layers.46.post_attention_layernorm.weight": "model-00043-of-00081.safetensors", "model.layers.46.self_attn.k_proj.weight": "model-00043-of-00081.safetensors", "model.layers.46.self_attn.o_proj.weight": "model-00043-of-00081.safetensors", "model.layers.46.self_attn.q_proj.weight": "model-00043-of-00081.safetensors", "model.layers.46.self_attn.v_proj.weight": "model-00043-of-00081.safetensors", "model.layers.47.input_layernorm.weight": "model-00043-of-00081.safetensors", "model.layers.47.mlp.down_proj.weight": "model-00043-of-00081.safetensors", "model.layers.47.mlp.gate_proj.weight": "model-00043-of-00081.safetensors", "model.layers.47.mlp.up_proj.weight": "model-00044-of-00081.safetensors", "model.layers.47.post_attention_layernorm.weight": "model-00044-of-00081.safetensors", "model.layers.47.self_attn.k_proj.weight": "model-00044-of-00081.safetensors", "model.layers.47.self_attn.o_proj.weight": "model-00044-of-00081.safetensors", "model.layers.47.self_attn.q_proj.weight": "model-00044-of-00081.safetensors", "model.layers.47.self_attn.v_proj.weight": "model-00044-of-00081.safetensors", "model.layers.48.input_layernorm.weight": "model-00044-of-00081.safetensors", "model.layers.48.mlp.down_proj.weight": "model-00044-of-00081.safetensors", "model.layers.48.mlp.gate_proj.weight": "model-00044-of-00081.safetensors", "model.layers.48.mlp.up_proj.weight": "model-00045-of-00081.safetensors", "model.layers.48.post_attention_layernorm.weight": "model-00045-of-00081.safetensors", "model.layers.48.self_attn.k_proj.weight": "model-00045-of-00081.safetensors", "model.layers.48.self_attn.o_proj.weight": "model-00045-of-00081.safetensors", "model.layers.48.self_attn.q_proj.weight": "model-00045-of-00081.safetensors", "model.layers.48.self_attn.v_proj.weight": "model-00045-of-00081.safetensors", "model.layers.49.input_layernorm.weight": "model-00045-of-00081.safetensors", "model.layers.49.mlp.down_proj.weight": "model-00045-of-00081.safetensors", "model.layers.49.mlp.gate_proj.weight": "model-00045-of-00081.safetensors", "model.layers.49.mlp.up_proj.weight": "model-00046-of-00081.safetensors", "model.layers.49.post_attention_layernorm.weight": "model-00046-of-00081.safetensors", "model.layers.49.self_attn.k_proj.weight": "model-00046-of-00081.safetensors", "model.layers.49.self_attn.o_proj.weight": "model-00046-of-00081.safetensors", "model.layers.49.self_attn.q_proj.weight": "model-00046-of-00081.safetensors", "model.layers.49.self_attn.v_proj.weight": "model-00046-of-00081.safetensors", "model.layers.5.input_layernorm.weight": "model-00046-of-00081.safetensors", "model.layers.5.mlp.down_proj.weight": "model-00046-of-00081.safetensors", "model.layers.5.mlp.gate_proj.weight": "model-00046-of-00081.safetensors", "model.layers.5.mlp.up_proj.weight": "model-00047-of-00081.safetensors", "model.layers.5.post_attention_layernorm.weight": "model-00047-of-00081.safetensors", "model.layers.5.self_attn.k_proj.weight": "model-00047-of-00081.safetensors", "model.layers.5.self_attn.o_proj.weight": "model-00047-of-00081.safetensors", "model.layers.5.self_attn.q_proj.weight": "model-00047-of-00081.safetensors", "model.layers.5.self_attn.v_proj.weight": "model-00047-of-00081.safetensors", "model.layers.50.input_layernorm.weight": "model-00047-of-00081.safetensors", "model.layers.50.mlp.down_proj.weight": "model-00047-of-00081.safetensors", "model.layers.50.mlp.gate_proj.weight": "model-00047-of-00081.safetensors", "model.layers.50.mlp.up_proj.weight": "model-00048-of-00081.safetensors", "model.layers.50.post_attention_layernorm.weight": "model-00048-of-00081.safetensors", "model.layers.50.self_attn.k_proj.weight": "model-00048-of-00081.safetensors", "model.layers.50.self_attn.o_proj.weight": "model-00048-of-00081.safetensors", "model.layers.50.self_attn.q_proj.weight": "model-00048-of-00081.safetensors", "model.layers.50.self_attn.v_proj.weight": "model-00048-of-00081.safetensors", "model.layers.51.input_layernorm.weight": "model-00048-of-00081.safetensors", "model.layers.51.mlp.down_proj.weight": "model-00048-of-00081.safetensors", "model.layers.51.mlp.gate_proj.weight": "model-00048-of-00081.safetensors", "model.layers.51.mlp.up_proj.weight": "model-00049-of-00081.safetensors", "model.layers.51.post_attention_layernorm.weight": "model-00049-of-00081.safetensors", "model.layers.51.self_attn.k_proj.weight": "model-00049-of-00081.safetensors", "model.layers.51.self_attn.o_proj.weight": "model-00049-of-00081.safetensors", "model.layers.51.self_attn.q_proj.weight": "model-00049-of-00081.safetensors", "model.layers.51.self_attn.v_proj.weight": "model-00049-of-00081.safetensors", "model.layers.52.input_layernorm.weight": "model-00049-of-00081.safetensors", "model.layers.52.mlp.down_proj.weight": "model-00049-of-00081.safetensors", "model.layers.52.mlp.gate_proj.weight": "model-00049-of-00081.safetensors", "model.layers.52.mlp.up_proj.weight": "model-00050-of-00081.safetensors", "model.layers.52.post_attention_layernorm.weight": "model-00050-of-00081.safetensors", "model.layers.52.self_attn.k_proj.weight": "model-00050-of-00081.safetensors", "model.layers.52.self_attn.o_proj.weight": "model-00050-of-00081.safetensors", "model.layers.52.self_attn.q_proj.weight": "model-00050-of-00081.safetensors", "model.layers.52.self_attn.v_proj.weight": "model-00050-of-00081.safetensors", "model.layers.53.input_layernorm.weight": "model-00050-of-00081.safetensors", "model.layers.53.mlp.down_proj.weight": "model-00050-of-00081.safetensors", "model.layers.53.mlp.gate_proj.weight": "model-00050-of-00081.safetensors", "model.layers.53.mlp.up_proj.weight": "model-00051-of-00081.safetensors", "model.layers.53.post_attention_layernorm.weight": "model-00051-of-00081.safetensors", "model.layers.53.self_attn.k_proj.weight": "model-00051-of-00081.safetensors", "model.layers.53.self_attn.o_proj.weight": "model-00051-of-00081.safetensors", "model.layers.53.self_attn.q_proj.weight": "model-00051-of-00081.safetensors", "model.layers.53.self_attn.v_proj.weight": "model-00051-of-00081.safetensors", "model.layers.54.input_layernorm.weight": "model-00051-of-00081.safetensors", "model.layers.54.mlp.down_proj.weight": "model-00051-of-00081.safetensors", "model.layers.54.mlp.gate_proj.weight": "model-00051-of-00081.safetensors", "model.layers.54.mlp.up_proj.weight": "model-00052-of-00081.safetensors", "model.layers.54.post_attention_layernorm.weight": "model-00052-of-00081.safetensors", "model.layers.54.self_attn.k_proj.weight": "model-00052-of-00081.safetensors", "model.layers.54.self_attn.o_proj.weight": "model-00052-of-00081.safetensors", "model.layers.54.self_attn.q_proj.weight": "model-00052-of-00081.safetensors", "model.layers.54.self_attn.v_proj.weight": "model-00052-of-00081.safetensors", "model.layers.55.input_layernorm.weight": "model-00052-of-00081.safetensors", "model.layers.55.mlp.down_proj.weight": "model-00052-of-00081.safetensors", "model.layers.55.mlp.gate_proj.weight": "model-00052-of-00081.safetensors", "model.layers.55.mlp.up_proj.weight": "model-00053-of-00081.safetensors", "model.layers.55.post_attention_layernorm.weight": "model-00053-of-00081.safetensors", "model.layers.55.self_attn.k_proj.weight": "model-00053-of-00081.safetensors", "model.layers.55.self_attn.o_proj.weight": "model-00053-of-00081.safetensors", "model.layers.55.self_attn.q_proj.weight": "model-00053-of-00081.safetensors", "model.layers.55.self_attn.v_proj.weight": "model-00053-of-00081.safetensors", "model.layers.56.input_layernorm.weight": "model-00053-of-00081.safetensors", "model.layers.56.mlp.down_proj.weight": "model-00053-of-00081.safetensors", "model.layers.56.mlp.gate_proj.weight": "model-00053-of-00081.safetensors", "model.layers.56.mlp.up_proj.weight": "model-00054-of-00081.safetensors", "model.layers.56.post_attention_layernorm.weight": "model-00054-of-00081.safetensors", "model.layers.56.self_attn.k_proj.weight": "model-00054-of-00081.safetensors", "model.layers.56.self_attn.o_proj.weight": "model-00054-of-00081.safetensors", "model.layers.56.self_attn.q_proj.weight": "model-00054-of-00081.safetensors", "model.layers.56.self_attn.v_proj.weight": "model-00054-of-00081.safetensors", "model.layers.57.input_layernorm.weight": "model-00054-of-00081.safetensors", "model.layers.57.mlp.down_proj.weight": "model-00054-of-00081.safetensors", "model.layers.57.mlp.gate_proj.weight": "model-00054-of-00081.safetensors", "model.layers.57.mlp.up_proj.weight": "model-00055-of-00081.safetensors", "model.layers.57.post_attention_layernorm.weight": "model-00055-of-00081.safetensors", "model.layers.57.self_attn.k_proj.weight": "model-00055-of-00081.safetensors", "model.layers.57.self_attn.o_proj.weight": "model-00055-of-00081.safetensors", "model.layers.57.self_attn.q_proj.weight": "model-00055-of-00081.safetensors", "model.layers.57.self_attn.v_proj.weight": "model-00055-of-00081.safetensors", "model.layers.58.input_layernorm.weight": "model-00055-of-00081.safetensors", "model.layers.58.mlp.down_proj.weight": "model-00055-of-00081.safetensors", "model.layers.58.mlp.gate_proj.weight": "model-00055-of-00081.safetensors", "model.layers.58.mlp.up_proj.weight": "model-00056-of-00081.safetensors", "model.layers.58.post_attention_layernorm.weight": "model-00056-of-00081.safetensors", "model.layers.58.self_attn.k_proj.weight": "model-00056-of-00081.safetensors", "model.layers.58.self_attn.o_proj.weight": "model-00056-of-00081.safetensors", "model.layers.58.self_attn.q_proj.weight": "model-00056-of-00081.safetensors", "model.layers.58.self_attn.v_proj.weight": "model-00056-of-00081.safetensors", "model.layers.59.input_layernorm.weight": "model-00056-of-00081.safetensors", "model.layers.59.mlp.down_proj.weight": "model-00056-of-00081.safetensors", "model.layers.59.mlp.gate_proj.weight": "model-00056-of-00081.safetensors", "model.layers.59.mlp.up_proj.weight": "model-00057-of-00081.safetensors", "model.layers.59.post_attention_layernorm.weight": "model-00057-of-00081.safetensors", "model.layers.59.self_attn.k_proj.weight": "model-00057-of-00081.safetensors", "model.layers.59.self_attn.o_proj.weight": "model-00057-of-00081.safetensors", "model.layers.59.self_attn.q_proj.weight": "model-00057-of-00081.safetensors", "model.layers.59.self_attn.v_proj.weight": "model-00057-of-00081.safetensors", "model.layers.6.input_layernorm.weight": "model-00057-of-00081.safetensors", "model.layers.6.mlp.down_proj.weight": "model-00057-of-00081.safetensors", "model.layers.6.mlp.gate_proj.weight": "model-00057-of-00081.safetensors", "model.layers.6.mlp.up_proj.weight": "model-00058-of-00081.safetensors", "model.layers.6.post_attention_layernorm.weight": "model-00058-of-00081.safetensors", "model.layers.6.self_attn.k_proj.weight": "model-00058-of-00081.safetensors", "model.layers.6.self_attn.o_proj.weight": "model-00058-of-00081.safetensors", "model.layers.6.self_attn.q_proj.weight": "model-00058-of-00081.safetensors", "model.layers.6.self_attn.v_proj.weight": "model-00058-of-00081.safetensors", "model.layers.60.input_layernorm.weight": "model-00058-of-00081.safetensors", "model.layers.60.mlp.down_proj.weight": "model-00058-of-00081.safetensors", "model.layers.60.mlp.gate_proj.weight": "model-00058-of-00081.safetensors", "model.layers.60.mlp.up_proj.weight": "model-00059-of-00081.safetensors", "model.layers.60.post_attention_layernorm.weight": "model-00059-of-00081.safetensors", "model.layers.60.self_attn.k_proj.weight": "model-00059-of-00081.safetensors", "model.layers.60.self_attn.o_proj.weight": "model-00059-of-00081.safetensors", "model.layers.60.self_attn.q_proj.weight": "model-00059-of-00081.safetensors", "model.layers.60.self_attn.v_proj.weight": "model-00059-of-00081.safetensors", "model.layers.61.input_layernorm.weight": "model-00059-of-00081.safetensors", "model.layers.61.mlp.down_proj.weight": "model-00059-of-00081.safetensors", "model.layers.61.mlp.gate_proj.weight": "model-00059-of-00081.safetensors", "model.layers.61.mlp.up_proj.weight": "model-00060-of-00081.safetensors", "model.layers.61.post_attention_layernorm.weight": "model-00060-of-00081.safetensors", "model.layers.61.self_attn.k_proj.weight": "model-00060-of-00081.safetensors", "model.layers.61.self_attn.o_proj.weight": "model-00060-of-00081.safetensors", "model.layers.61.self_attn.q_proj.weight": "model-00060-of-00081.safetensors", "model.layers.61.self_attn.v_proj.weight": "model-00060-of-00081.safetensors", "model.layers.62.input_layernorm.weight": "model-00060-of-00081.safetensors", "model.layers.62.mlp.down_proj.weight": "model-00060-of-00081.safetensors", "model.layers.62.mlp.gate_proj.weight": "model-00060-of-00081.safetensors", "model.layers.62.mlp.up_proj.weight": "model-00061-of-00081.safetensors", "model.layers.62.post_attention_layernorm.weight": "model-00061-of-00081.safetensors", "model.layers.62.self_attn.k_proj.weight": "model-00061-of-00081.safetensors", "model.layers.62.self_attn.o_proj.weight": "model-00061-of-00081.safetensors", "model.layers.62.self_attn.q_proj.weight": "model-00061-of-00081.safetensors", "model.layers.62.self_attn.v_proj.weight": "model-00061-of-00081.safetensors", "model.layers.63.input_layernorm.weight": "model-00061-of-00081.safetensors", "model.layers.63.mlp.down_proj.weight": "model-00061-of-00081.safetensors", "model.layers.63.mlp.gate_proj.weight": "model-00061-of-00081.safetensors", "model.layers.63.mlp.up_proj.weight": "model-00062-of-00081.safetensors", "model.layers.63.post_attention_layernorm.weight": "model-00062-of-00081.safetensors", "model.layers.63.self_attn.k_proj.weight": "model-00062-of-00081.safetensors", "model.layers.63.self_attn.o_proj.weight": "model-00062-of-00081.safetensors", "model.layers.63.self_attn.q_proj.weight": "model-00062-of-00081.safetensors", "model.layers.63.self_attn.v_proj.weight": "model-00062-of-00081.safetensors", "model.layers.64.input_layernorm.weight": "model-00062-of-00081.safetensors", "model.layers.64.mlp.down_proj.weight": "model-00062-of-00081.safetensors", "model.layers.64.mlp.gate_proj.weight": "model-00062-of-00081.safetensors", "model.layers.64.mlp.up_proj.weight": "model-00063-of-00081.safetensors", "model.layers.64.post_attention_layernorm.weight": "model-00063-of-00081.safetensors", "model.layers.64.self_attn.k_proj.weight": "model-00063-of-00081.safetensors", "model.layers.64.self_attn.o_proj.weight": "model-00063-of-00081.safetensors", "model.layers.64.self_attn.q_proj.weight": "model-00063-of-00081.safetensors", "model.layers.64.self_attn.v_proj.weight": "model-00063-of-00081.safetensors", "model.layers.65.input_layernorm.weight": "model-00063-of-00081.safetensors", "model.layers.65.mlp.down_proj.weight": "model-00063-of-00081.safetensors", "model.layers.65.mlp.gate_proj.weight": "model-00063-of-00081.safetensors", "model.layers.65.mlp.up_proj.weight": "model-00064-of-00081.safetensors", "model.layers.65.post_attention_layernorm.weight": "model-00064-of-00081.safetensors", "model.layers.65.self_attn.k_proj.weight": "model-00064-of-00081.safetensors", "model.layers.65.self_attn.o_proj.weight": "model-00064-of-00081.safetensors", "model.layers.65.self_attn.q_proj.weight": "model-00064-of-00081.safetensors", "model.layers.65.self_attn.v_proj.weight": "model-00064-of-00081.safetensors", "model.layers.66.input_layernorm.weight": "model-00064-of-00081.safetensors", "model.layers.66.mlp.down_proj.weight": "model-00064-of-00081.safetensors", "model.layers.66.mlp.gate_proj.weight": "model-00064-of-00081.safetensors", "model.layers.66.mlp.up_proj.weight": "model-00065-of-00081.safetensors", "model.layers.66.post_attention_layernorm.weight": "model-00065-of-00081.safetensors", "model.layers.66.self_attn.k_proj.weight": "model-00065-of-00081.safetensors", "model.layers.66.self_attn.o_proj.weight": "model-00065-of-00081.safetensors", "model.layers.66.self_attn.q_proj.weight": "model-00065-of-00081.safetensors", "model.layers.66.self_attn.v_proj.weight": "model-00065-of-00081.safetensors", "model.layers.67.input_layernorm.weight": "model-00065-of-00081.safetensors", "model.layers.67.mlp.down_proj.weight": "model-00065-of-00081.safetensors", "model.layers.67.mlp.gate_proj.weight": "model-00065-of-00081.safetensors", "model.layers.67.mlp.up_proj.weight": "model-00066-of-00081.safetensors", "model.layers.67.post_attention_layernorm.weight": "model-00066-of-00081.safetensors", "model.layers.67.self_attn.k_proj.weight": "model-00066-of-00081.safetensors", "model.layers.67.self_attn.o_proj.weight": "model-00066-of-00081.safetensors", "model.layers.67.self_attn.q_proj.weight": "model-00066-of-00081.safetensors", "model.layers.67.self_attn.v_proj.weight": "model-00066-of-00081.safetensors", "model.layers.68.input_layernorm.weight": "model-00066-of-00081.safetensors", "model.layers.68.mlp.down_proj.weight": "model-00066-of-00081.safetensors", "model.layers.68.mlp.gate_proj.weight": "model-00066-of-00081.safetensors", "model.layers.68.mlp.up_proj.weight": "model-00067-of-00081.safetensors", "model.layers.68.post_attention_layernorm.weight": "model-00067-of-00081.safetensors", "model.layers.68.self_attn.k_proj.weight": "model-00067-of-00081.safetensors", "model.layers.68.self_attn.o_proj.weight": "model-00067-of-00081.safetensors", "model.layers.68.self_attn.q_proj.weight": "model-00067-of-00081.safetensors", "model.layers.68.self_attn.v_proj.weight": "model-00067-of-00081.safetensors", "model.layers.69.input_layernorm.weight": "model-00067-of-00081.safetensors", "model.layers.69.mlp.down_proj.weight": "model-00067-of-00081.safetensors", "model.layers.69.mlp.gate_proj.weight": "model-00067-of-00081.safetensors", "model.layers.69.mlp.up_proj.weight": "model-00068-of-00081.safetensors", "model.layers.69.post_attention_layernorm.weight": "model-00068-of-00081.safetensors", "model.layers.69.self_attn.k_proj.weight": "model-00068-of-00081.safetensors", "model.layers.69.self_attn.o_proj.weight": "model-00068-of-00081.safetensors", "model.layers.69.self_attn.q_proj.weight": "model-00068-of-00081.safetensors", "model.layers.69.self_attn.v_proj.weight": "model-00068-of-00081.safetensors", "model.layers.7.input_layernorm.weight": "model-00068-of-00081.safetensors", "model.layers.7.mlp.down_proj.weight": "model-00068-of-00081.safetensors", "model.layers.7.mlp.gate_proj.weight": "model-00068-of-00081.safetensors", "model.layers.7.mlp.up_proj.weight": "model-00069-of-00081.safetensors", "model.layers.7.post_attention_layernorm.weight": "model-00069-of-00081.safetensors", "model.layers.7.self_attn.k_proj.weight": "model-00069-of-00081.safetensors", "model.layers.7.self_attn.o_proj.weight": "model-00069-of-00081.safetensors", "model.layers.7.self_attn.q_proj.weight": "model-00069-of-00081.safetensors", "model.layers.7.self_attn.v_proj.weight": "model-00069-of-00081.safetensors", "model.layers.70.input_layernorm.weight": "model-00069-of-00081.safetensors", "model.layers.70.mlp.down_proj.weight": "model-00069-of-00081.safetensors", "model.layers.70.mlp.gate_proj.weight": "model-00069-of-00081.safetensors", "model.layers.70.mlp.up_proj.weight": "model-00070-of-00081.safetensors", "model.layers.70.post_attention_layernorm.weight": "model-00070-of-00081.safetensors", "model.layers.70.self_attn.k_proj.weight": "model-00070-of-00081.safetensors", "model.layers.70.self_attn.o_proj.weight": "model-00070-of-00081.safetensors", "model.layers.70.self_attn.q_proj.weight": "model-00070-of-00081.safetensors", "model.layers.70.self_attn.v_proj.weight": "model-00070-of-00081.safetensors", "model.layers.71.input_layernorm.weight": "model-00070-of-00081.safetensors", "model.layers.71.mlp.down_proj.weight": "model-00070-of-00081.safetensors", "model.layers.71.mlp.gate_proj.weight": "model-00070-of-00081.safetensors", "model.layers.71.mlp.up_proj.weight": "model-00071-of-00081.safetensors", "model.layers.71.post_attention_layernorm.weight": "model-00071-of-00081.safetensors", "model.layers.71.self_attn.k_proj.weight": "model-00071-of-00081.safetensors", "model.layers.71.self_attn.o_proj.weight": "model-00071-of-00081.safetensors", "model.layers.71.self_attn.q_proj.weight": "model-00071-of-00081.safetensors", "model.layers.71.self_attn.v_proj.weight": "model-00071-of-00081.safetensors", "model.layers.72.input_layernorm.weight": "model-00071-of-00081.safetensors", "model.layers.72.mlp.down_proj.weight": "model-00071-of-00081.safetensors", "model.layers.72.mlp.gate_proj.weight": "model-00071-of-00081.safetensors", "model.layers.72.mlp.up_proj.weight": "model-00072-of-00081.safetensors", "model.layers.72.post_attention_layernorm.weight": "model-00072-of-00081.safetensors", "model.layers.72.self_attn.k_proj.weight": "model-00072-of-00081.safetensors", "model.layers.72.self_attn.o_proj.weight": "model-00072-of-00081.safetensors", "model.layers.72.self_attn.q_proj.weight": "model-00072-of-00081.safetensors", "model.layers.72.self_attn.v_proj.weight": "model-00072-of-00081.safetensors", "model.layers.73.input_layernorm.weight": "model-00072-of-00081.safetensors", "model.layers.73.mlp.down_proj.weight": "model-00072-of-00081.safetensors", "model.layers.73.mlp.gate_proj.weight": "model-00072-of-00081.safetensors", "model.layers.73.mlp.up_proj.weight": "model-00073-of-00081.safetensors", "model.layers.73.post_attention_layernorm.weight": "model-00073-of-00081.safetensors", "model.layers.73.self_attn.k_proj.weight": "model-00073-of-00081.safetensors", "model.layers.73.self_attn.o_proj.weight": "model-00073-of-00081.safetensors", "model.layers.73.self_attn.q_proj.weight": "model-00073-of-00081.safetensors", "model.layers.73.self_attn.v_proj.weight": "model-00073-of-00081.safetensors", "model.layers.74.input_layernorm.weight": "model-00073-of-00081.safetensors", "model.layers.74.mlp.down_proj.weight": "model-00073-of-00081.safetensors", "model.layers.74.mlp.gate_proj.weight": "model-00073-of-00081.safetensors", "model.layers.74.mlp.up_proj.weight": "model-00074-of-00081.safetensors", "model.layers.74.post_attention_layernorm.weight": "model-00074-of-00081.safetensors", "model.layers.74.self_attn.k_proj.weight": "model-00074-of-00081.safetensors", "model.layers.74.self_attn.o_proj.weight": "model-00074-of-00081.safetensors", "model.layers.74.self_attn.q_proj.weight": "model-00074-of-00081.safetensors", "model.layers.74.self_attn.v_proj.weight": "model-00074-of-00081.safetensors", "model.layers.75.input_layernorm.weight": "model-00074-of-00081.safetensors", "model.layers.75.mlp.down_proj.weight": "model-00074-of-00081.safetensors", "model.layers.75.mlp.gate_proj.weight": "model-00074-of-00081.safetensors", "model.layers.75.mlp.up_proj.weight": "model-00075-of-00081.safetensors", "model.layers.75.post_attention_layernorm.weight": "model-00075-of-00081.safetensors", "model.layers.75.self_attn.k_proj.weight": "model-00075-of-00081.safetensors", "model.layers.75.self_attn.o_proj.weight": "model-00075-of-00081.safetensors", "model.layers.75.self_attn.q_proj.weight": "model-00075-of-00081.safetensors", "model.layers.75.self_attn.v_proj.weight": "model-00075-of-00081.safetensors", "model.layers.76.input_layernorm.weight": "model-00075-of-00081.safetensors", "model.layers.76.mlp.down_proj.weight": "model-00075-of-00081.safetensors", "model.layers.76.mlp.gate_proj.weight": "model-00075-of-00081.safetensors", "model.layers.76.mlp.up_proj.weight": "model-00076-of-00081.safetensors", "model.layers.76.post_attention_layernorm.weight": "model-00076-of-00081.safetensors", "model.layers.76.self_attn.k_proj.weight": "model-00076-of-00081.safetensors", "model.layers.76.self_attn.o_proj.weight": "model-00076-of-00081.safetensors", "model.layers.76.self_attn.q_proj.weight": "model-00076-of-00081.safetensors", "model.layers.76.self_attn.v_proj.weight": "model-00076-of-00081.safetensors", "model.layers.77.input_layernorm.weight": "model-00076-of-00081.safetensors", "model.layers.77.mlp.down_proj.weight": "model-00076-of-00081.safetensors", "model.layers.77.mlp.gate_proj.weight": "model-00076-of-00081.safetensors", "model.layers.77.mlp.up_proj.weight": "model-00077-of-00081.safetensors", "model.layers.77.post_attention_layernorm.weight": "model-00077-of-00081.safetensors", "model.layers.77.self_attn.k_proj.weight": "model-00077-of-00081.safetensors", "model.layers.77.self_attn.o_proj.weight": "model-00077-of-00081.safetensors", "model.layers.77.self_attn.q_proj.weight": "model-00077-of-00081.safetensors", "model.layers.77.self_attn.v_proj.weight": "model-00077-of-00081.safetensors", "model.layers.78.input_layernorm.weight": "model-00077-of-00081.safetensors", "model.layers.78.mlp.down_proj.weight": "model-00077-of-00081.safetensors", "model.layers.78.mlp.gate_proj.weight": "model-00077-of-00081.safetensors", "model.layers.78.mlp.up_proj.weight": "model-00078-of-00081.safetensors", "model.layers.78.post_attention_layernorm.weight": "model-00078-of-00081.safetensors", "model.layers.78.self_attn.k_proj.weight": "model-00078-of-00081.safetensors", "model.layers.78.self_attn.o_proj.weight": "model-00078-of-00081.safetensors", "model.layers.78.self_attn.q_proj.weight": "model-00078-of-00081.safetensors", "model.layers.78.self_attn.v_proj.weight": "model-00078-of-00081.safetensors", "model.layers.79.input_layernorm.weight": "model-00078-of-00081.safetensors", "model.layers.79.mlp.down_proj.weight": "model-00078-of-00081.safetensors", "model.layers.79.mlp.gate_proj.weight": "model-00078-of-00081.safetensors", "model.layers.79.mlp.up_proj.weight": "model-00079-of-00081.safetensors", "model.layers.79.post_attention_layernorm.weight": "model-00079-of-00081.safetensors", "model.layers.79.self_attn.k_proj.weight": "model-00079-of-00081.safetensors", "model.layers.79.self_attn.o_proj.weight": "model-00079-of-00081.safetensors", "model.layers.79.self_attn.q_proj.weight": "model-00079-of-00081.safetensors", "model.layers.79.self_attn.v_proj.weight": "model-00079-of-00081.safetensors", "model.layers.8.input_layernorm.weight": "model-00079-of-00081.safetensors", "model.layers.8.mlp.down_proj.weight": "model-00079-of-00081.safetensors", "model.layers.8.mlp.gate_proj.weight": "model-00079-of-00081.safetensors", "model.layers.8.mlp.up_proj.weight": "model-00080-of-00081.safetensors", "model.layers.8.post_attention_layernorm.weight": "model-00080-of-00081.safetensors", "model.layers.8.self_attn.k_proj.weight": "model-00080-of-00081.safetensors", "model.layers.8.self_attn.o_proj.weight": "model-00080-of-00081.safetensors", "model.layers.8.self_attn.q_proj.weight": "model-00080-of-00081.safetensors", "model.layers.8.self_attn.v_proj.weight": "model-00080-of-00081.safetensors", "model.layers.9.input_layernorm.weight": "model-00080-of-00081.safetensors", "model.layers.9.mlp.down_proj.weight": "model-00080-of-00081.safetensors", "model.layers.9.mlp.gate_proj.weight": "model-00080-of-00081.safetensors", "model.layers.9.mlp.up_proj.weight": "model-00081-of-00081.safetensors", "model.layers.9.post_attention_layernorm.weight": "model-00081-of-00081.safetensors", "model.layers.9.self_attn.k_proj.weight": "model-00081-of-00081.safetensors", "model.layers.9.self_attn.o_proj.weight": "model-00081-of-00081.safetensors", "model.layers.9.self_attn.q_proj.weight": "model-00081-of-00081.safetensors", "model.layers.9.self_attn.v_proj.weight": "model-00081-of-00081.safetensors", "model.norm.weight": "model-00081-of-00081.safetensors"}} \ No newline at end of file diff --git a/special_tokens_map.json b/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..5819ea25d7b1b6340063a0629c2143c44b0452da --- /dev/null +++ b/special_tokens_map.json @@ -0,0 +1,30 @@ +{ + "bos_token": { + "content": "", + "lstrip": false, + "normalized": true, + "rstrip": false, + "single_word": false + }, + "eos_token": { + "content": "", + "lstrip": false, + "normalized": true, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": true, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": true, + "rstrip": false, + "single_word": false + } +} diff --git a/tokenizer.model b/tokenizer.model new file mode 100644 index 0000000000000000000000000000000000000000..6c00c742ce03c627d6cd5b795984876fa49fa899 --- /dev/null +++ b/tokenizer.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347 +size 499723 diff --git a/tokenizer_config.json b/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..581618b6ed7fe03353fb255433dcb9cfd605467a --- /dev/null +++ b/tokenizer_config.json @@ -0,0 +1,42 @@ +{ + "add_bos_token": true, + "add_eos_token": false, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": true, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": true, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": true, + "rstrip": false, + "single_word": false, + "special": true + } + }, + "bos_token": "", + "clean_up_tokenization_spaces": false, + "eos_token": "", + "legacy": false, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "spaces_between_special_tokens": false, + "tokenizer_class": "LlamaTokenizer", + "unk_token": "", + "use_default_system_prompt": false, + "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}" +}