diff --git a/README.md b/README.md new file mode 100644 index 0000000000000000000000000000000000000000..bc5f30d6632ac0efdc7be2e9095e9e9579af2e33 --- /dev/null +++ b/README.md @@ -0,0 +1,199 @@ +--- +library_name: transformers +tags: [] +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + +This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] \ No newline at end of file diff --git a/config.json b/config.json new file mode 100644 index 0000000000000000000000000000000000000000..fddeaf7bcf263252d13dcf11efcd572b1771e552 --- /dev/null +++ b/config.json @@ -0,0 +1,42 @@ +{ + "_name_or_path": "pixtral_converted", + "architectures": [ + "LlavaForConditionalGeneration" + ], + "ignore_index": -100, + "image_seq_length": 1, + "image_token_index": 10, + "model_type": "llava", + "multimodal_projector_bias": false, + "projector_hidden_act": "gelu", + "text_config": { + "hidden_size": 12288, + "intermediate_size": 28672, + "is_composition": true, + "max_position_embeddings": 131072, + "model_type": "mistral", + "norm_eps": 1e-05, + "num_attention_heads": 96, + "num_hidden_layers": 88, + "num_key_value_heads": 8, + "rope_theta": 1000000000.0, + "sliding_window": null, + "vocab_size": 32768 + }, + "torch_dtype": "bfloat16", + "transformers_version": "4.47.0.dev0", + "vision_config": { + "head_dim": 88, + "hidden_act": "silu", + "hidden_size": 1408, + "image_size": 1024, + "image_token_id": 10, + "intermediate_size": 6144, + "model_type": "pixtral", + "num_hidden_layers": 40, + "patch_size": 16, + "rope_theta": 10000.0 + }, + "vision_feature_layer": -1, + "vision_feature_select_strategy": "full" +} diff --git a/generation_config.json b/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..07d5fb2413f2590faac9594acda00e72c863c8fb --- /dev/null +++ b/generation_config.json @@ -0,0 +1,6 @@ +{ + "_from_model_config": true, + "bos_token_id": 1, + "eos_token_id": 2, + "transformers_version": "4.47.0.dev0" +} diff --git a/model-00001-of-00052.safetensors b/model-00001-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..eb3561bb7738726571a7726663e1fb598f858ffa --- /dev/null +++ b/model-00001-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9d2d7a1e3e556790ffb2c6340a48c0d67c5e4355b0d3053cfbd18013f8ce2e53 +size 4509220632 diff --git a/model-00002-of-00052.safetensors b/model-00002-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0b5f93da2301994173f7c21f6531b231403edc3f --- /dev/null +++ b/model-00002-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9fcbbe01e14828c044cf0ae930c826acff6ee3b8e81fd6ca5ef6c97a42fd7f3a +size 4882270024 diff --git a/model-00003-of-00052.safetensors b/model-00003-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1312723954f273bc2e8076f09d864746eab98351 --- /dev/null +++ b/model-00003-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e84a1a7c117ecee5b9be23ace65ff17aa7dcd55350794a02820a80d9e2c3a07 +size 4831889368 diff --git a/model-00004-of-00052.safetensors b/model-00004-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6e4311b56f4f727099e58ebaf7c85811f2017a1c --- /dev/null +++ b/model-00004-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5cd040875f38bc94db33c01c68af083f6288a0e8f883a6a3f2579ce01a95cd90 +size 4831938776 diff --git a/model-00005-of-00052.safetensors b/model-00005-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..63d77a444b6455e6e8321f404fd13b0d11e8b7d6 --- /dev/null +++ b/model-00005-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9338c97508c94cfd6fe5d92e9285622d46d14d7ac97fa8dec6199d3569e764d5 +size 4831938776 diff --git a/model-00006-of-00052.safetensors b/model-00006-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0cec28cbef2f1eb7b525c3b520d15778c1784d51 --- /dev/null +++ b/model-00006-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:540fb77c3fc0150cbe012a285afc57918957f2ed92542910e876ae14235143b8 +size 4882270024 diff --git a/model-00007-of-00052.safetensors b/model-00007-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7093e0db9df4aadd1ce320c88284fac9d8938174 --- /dev/null +++ b/model-00007-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c53b5751a6f73e070752bc8d7c9409442c0e1e8e3d6902c377026f71b9cfcd9b +size 4831889384 diff --git a/model-00008-of-00052.safetensors b/model-00008-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bb17db5e868145346c728bbf3901667f4c178450 --- /dev/null +++ b/model-00008-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:194ce36f2dbeb4b391857b32579d579bb668637a26f19b1ac48252f55accb753 +size 4831938792 diff --git a/model-00009-of-00052.safetensors b/model-00009-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ee51233d01050053ab10500103febfb178f5d294 --- /dev/null +++ b/model-00009-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:abb1b4063d634fdb4ee19894e89e5f9c9e21019a2b80adfb6ba14c13fa3c5406 +size 4831938800 diff --git a/model-00010-of-00052.safetensors b/model-00010-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ebc36168620e227a10251ed45aaf4f8af8d2c6d8 --- /dev/null +++ b/model-00010-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a4fc55163298b3882aa6aa1f8f229bd9857e1494cd1ea543e4f257376bd1c49 +size 4882270040 diff --git a/model-00011-of-00052.safetensors b/model-00011-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a30377fd299f75c834a42cc0c5128ceb06500a98 --- /dev/null +++ b/model-00011-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aa3f51e16047a320259ae21c6d9506266ce9bec2be0c49508f7d99386d68d5bd +size 4831889376 diff --git a/model-00012-of-00052.safetensors b/model-00012-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c25acbe847f0dcdd6eb30be0fcf28d7a24b5ae8e --- /dev/null +++ b/model-00012-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:19b5737821632a06fc0e61e24718094be4bf361cb3fc57fbaa82a6f7019f08cb +size 4831938792 diff --git a/model-00013-of-00052.safetensors b/model-00013-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a7168dcff238563668620394ab78b851a7eb37e2 --- /dev/null +++ b/model-00013-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cea09a3e29d77fddd25d7e55992d8f17a49bcf8ed9c38ca4cba517e0045d45f5 +size 4831938800 diff --git a/model-00014-of-00052.safetensors b/model-00014-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0b865cf85efcefabcc80192d1eb90b54c6a069d4 --- /dev/null +++ b/model-00014-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:075167a29ae960189491f0f250e9c0bc3d9491bc3d3c46ee6d0089c46e43f4ec +size 4882270040 diff --git a/model-00015-of-00052.safetensors b/model-00015-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..aef5f30d494e0646d8657574886fb4d6bd562281 --- /dev/null +++ b/model-00015-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:42588f8fe11c0846cfd9e96603828256cc4e8f101e54a7f24bed2f5c28cb9fe5 +size 4831889376 diff --git a/model-00016-of-00052.safetensors b/model-00016-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..25a0e4d3ad9971d8f7e0f7d40860ac7be8a6d88c --- /dev/null +++ b/model-00016-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7281b0ee3f2590bc6c19130ed8f233887f136f8fdfab9c64b34402e9919d7db5 +size 4831938792 diff --git a/model-00017-of-00052.safetensors b/model-00017-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a4b0b1e41cf0cfe3e6044f44f933d30b610e1e05 --- /dev/null +++ b/model-00017-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:06e442f95527ea9a5f82f193e9781bbbd376e305019b28d52f84426f10928d8c +size 4831938800 diff --git a/model-00018-of-00052.safetensors b/model-00018-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..681ab4617fb6e498449d82015cd7b93551dabe8e --- /dev/null +++ b/model-00018-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6dbfb7bbd99bc0d0d44b1c34112b71a81fdf1c5354835176a4fe187d00bfce1b +size 4882270040 diff --git a/model-00019-of-00052.safetensors b/model-00019-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d49c69d61c9b091414e2130aa0cd25f0ef04caf4 --- /dev/null +++ b/model-00019-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:791b54d8e7ff38c2e575df46115c7cd21a3a9defe544b9f943cdd0cdc78f3fb3 +size 4831889376 diff --git a/model-00020-of-00052.safetensors b/model-00020-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3c79dfbd98d1186807395b422ad2d15539b43593 --- /dev/null +++ b/model-00020-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35df3afdd407778ae26e1f9df6f6ea12681acd90de4d9525f4b4cde2ef30662d +size 4831938792 diff --git a/model-00021-of-00052.safetensors b/model-00021-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b466852c655bb50c1084d4ed6831c32f927b8c63 --- /dev/null +++ b/model-00021-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd7991d7936b08bf468f2f2da9dbac46224b732bedd9ab09c2ba4ff06a82557c +size 4831938800 diff --git a/model-00022-of-00052.safetensors b/model-00022-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a8d20d8daddfbf46a675b67d6a2723f286bb10ed --- /dev/null +++ b/model-00022-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c4a4473980e3bc05a3366fa638bb861519a4665f33bc1512ac6cd0cbdc7711a5 +size 4882270040 diff --git a/model-00023-of-00052.safetensors b/model-00023-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1724cf13503e865111dd020ae041e4d3991e5cca --- /dev/null +++ b/model-00023-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:65ca1eb9328ada88ad65e08c8b18d526d298bab6965a7372be537a994c5205eb +size 4831889376 diff --git a/model-00024-of-00052.safetensors b/model-00024-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..55a8541eaf157d0d087f21c56285ba2240797505 --- /dev/null +++ b/model-00024-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ba34028c5979cfd3781ca70aead3b1b2b55180d93394ce2857db4e5b0e9af8b6 +size 4831938792 diff --git a/model-00025-of-00052.safetensors b/model-00025-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..51b24cb97b7eace823699d5a61861d3735c8d256 --- /dev/null +++ b/model-00025-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8b1dee0420ed6c93f7bcaed4197655344303b338f3d9d1027fd7f036738bf932 +size 4831938800 diff --git a/model-00026-of-00052.safetensors b/model-00026-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fee1578113d89b8acc375758842220653e1f2de1 --- /dev/null +++ b/model-00026-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5cb83cb476282252ffddddabc69ea27fc08f3f7a022acefe4335419f28050e6c +size 4882270040 diff --git a/model-00027-of-00052.safetensors b/model-00027-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0c56ad205411210c9cbafd312cbf3dbed9e6a490 --- /dev/null +++ b/model-00027-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:05f852d81895c3ba645c106f21ab9fb63d2988baed0967b17266955e65c9e95a +size 4831889376 diff --git a/model-00028-of-00052.safetensors b/model-00028-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c65b0a61c3ba625f3a81a3f614f96d05a7144084 --- /dev/null +++ b/model-00028-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1859871627e14661d1ee731005e87b312d4af7aa25a436f468a7b7e4bf329dd4 +size 4831938792 diff --git a/model-00029-of-00052.safetensors b/model-00029-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..54dbd1de67f8057d02881f701a66e992ae155313 --- /dev/null +++ b/model-00029-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3b7d2b168b2caa8a0445fedf597478d62dbd6850a777b0c08e7eebed3b0d9a6a +size 4831938800 diff --git a/model-00030-of-00052.safetensors b/model-00030-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7dfde4c7fd3dcb2652dea41e735bb5a15c3d4a33 --- /dev/null +++ b/model-00030-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d4c20b8e733f6dbd8d5a300eea95b0afa3a1bc2a780768a6e274ea08cfeb1020 +size 4882270040 diff --git a/model-00031-of-00052.safetensors b/model-00031-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b71cbaff746836c4b0d6a0bd81c160f72617c1f8 --- /dev/null +++ b/model-00031-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:400f4c047a67586f764af4ae0abba7440df8107ceba3ac0fd4166f5bcaab34e1 +size 4831889376 diff --git a/model-00032-of-00052.safetensors b/model-00032-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..313277876bd0f896ba81111cb4712eec24d66c8a --- /dev/null +++ b/model-00032-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7ac94d3d776f9882c868cb20db22e8bcf79e5f0abb134d06e64ddcf6e652b81c +size 4831938792 diff --git a/model-00033-of-00052.safetensors b/model-00033-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..63fe8a8c18e8a698da3004003543a9bebdc1aa4c --- /dev/null +++ b/model-00033-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5ed94b8ba78c5d51635a47d724d64a3959e955c71187ecc23747f686dbc33d3a +size 4831938800 diff --git a/model-00034-of-00052.safetensors b/model-00034-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..449640c4f8bb77749f6783e7414eb8ab3441fc5e --- /dev/null +++ b/model-00034-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc84284a1541d12eb8d6a113a68f978a255151dffb021c86e8d4f3ec08c1f480 +size 4882270040 diff --git a/model-00035-of-00052.safetensors b/model-00035-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..881f1da7891b1e3008ee6df03aa4aa4ed6af276f --- /dev/null +++ b/model-00035-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:24493e298808e6481a72a8dbc8a4ee173d7a2bd7657f78987bbaafe85f3d5221 +size 4831889376 diff --git a/model-00036-of-00052.safetensors b/model-00036-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3055a21c3ba00411e64841d923ac8dc729c4f8bd --- /dev/null +++ b/model-00036-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a8fcdde3f303bae55b414633dbaf047f6f00806976a5b726e2d73fdd6fc14564 +size 4831938792 diff --git a/model-00037-of-00052.safetensors b/model-00037-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ee54594ab2719dd28e867ccc86a39582af901407 --- /dev/null +++ b/model-00037-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb44645fc87b76f6b69ef97217d7760955e1fd0402efa68b24fd9ffd1ee3c9a5 +size 4831938800 diff --git a/model-00038-of-00052.safetensors b/model-00038-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d151ef174f3263375ea6b1df8610a3cf23edf2ce --- /dev/null +++ b/model-00038-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9a62587682a17ba3d030c6479661d01932afe76c2eaade70383e0ed4bece8213 +size 4882270040 diff --git a/model-00039-of-00052.safetensors b/model-00039-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d9b57d6cee09beab7717a66b44536afc4a57535e --- /dev/null +++ b/model-00039-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3ea5f4b55c5dde56308cd14ab941c8912b2c71c971feda89cd199993bfef36d3 +size 4831889376 diff --git a/model-00040-of-00052.safetensors b/model-00040-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9c22d88660d79cc427d7f544a2fec411b12c13fb --- /dev/null +++ b/model-00040-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf6a29d3af8c20a2a70676407d5f8581df4e849af24eaf2760a9c38fafe8779f +size 4831938792 diff --git a/model-00041-of-00052.safetensors b/model-00041-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..04baf5c8719eb27c290ee2dd93ebd60988da6ea9 --- /dev/null +++ b/model-00041-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bff31ed650cf76fbad236fb6685f2a4c4f86db6da037cea3a73ebc5e28a1c93a +size 4831938800 diff --git a/model-00042-of-00052.safetensors b/model-00042-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..eb73b87d805a2de208452318ece491c3d745637a --- /dev/null +++ b/model-00042-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f090172fa822725090d97b55d31873fecd1ecc01606fbc4fb1357704f6760ac +size 4882270040 diff --git a/model-00043-of-00052.safetensors b/model-00043-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e5cd85c54109c5d04894e6fb7ac4639c81d05355 --- /dev/null +++ b/model-00043-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2390076a1035012086880648ee9cbc9a6f3407ac411bf79ebef557641964fc69 +size 4831889376 diff --git a/model-00044-of-00052.safetensors b/model-00044-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ad8e9c36e539c3640e9be9255f7e952c77e61f53 --- /dev/null +++ b/model-00044-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:785ba85d22ecc3d3b6b62dbcb1e843e57fa57bf0bba7877e15f5d3fda9f6100f +size 4831938792 diff --git a/model-00045-of-00052.safetensors b/model-00045-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..787b88258e9f6623fd09398da9612f5964680986 --- /dev/null +++ b/model-00045-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9f6470663d2e7c12a1af0f1c54fb2fb5176de2d47de0e6ee06bb652738ee3421 +size 4831938800 diff --git a/model-00046-of-00052.safetensors b/model-00046-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..be4c167398ddb418f99b7d88ef6d5cb5b1b33a4b --- /dev/null +++ b/model-00046-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0f28328f36a9ac4a470716488589812b656badad04877e4bfec2288660771fd9 +size 4882270040 diff --git a/model-00047-of-00052.safetensors b/model-00047-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2f9db15cd436d540146bc49e8503d1b27a307359 --- /dev/null +++ b/model-00047-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4450bdb8aac77ef388cb192dbd78a949aa18d7e2800ce112d00894a31466b7ec +size 4831889376 diff --git a/model-00048-of-00052.safetensors b/model-00048-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4b67da1031c07ebcedfcc00aa6898bd1200ba876 --- /dev/null +++ b/model-00048-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:99f803fe1adaaab1b8518279ebc881040e54abf44d6cd20a70b17cad026f3725 +size 4831938792 diff --git a/model-00049-of-00052.safetensors b/model-00049-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ae953d063ef8c89b3f00e88762aed9fcd26ad1b3 --- /dev/null +++ b/model-00049-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a845c84202b7efdba7ab703652d41d3345b5224a17139a133e7bd768fd357a7a +size 4831938800 diff --git a/model-00050-of-00052.safetensors b/model-00050-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9197bb979a6b92c3c92fc3be840abf6a2a1b58aa --- /dev/null +++ b/model-00050-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7f82b2ff1de31179a329a70ca8bba032f07920136307cc664ddb638557f50e57 +size 4882270040 diff --git a/model-00051-of-00052.safetensors b/model-00051-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2aa180b30464f2196fcafc5d7a3f807b3b4dedad --- /dev/null +++ b/model-00051-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3cb9c6d4c02060dcf6c0a2ba260653e339ddabca064aed14a81e16fdfc925a31 +size 4831889376 diff --git a/model-00052-of-00052.safetensors b/model-00052-of-00052.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d8704dee22d90d062c73fbf615e57ac7f11fecc0 --- /dev/null +++ b/model-00052-of-00052.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6c87ad43da4553c3d7fe583bd06d25ccbf0e832cf4f7eaa913e0e8b108cffb36 +size 1510023816 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..27bb066c8627eaf62ed6a7789009b371b34648f1 --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,1166 @@ +{ + "metadata": { + "total_size": 248269691648 + }, + "weight_map": { + "language_model.lm_head.weight": "model-00052-of-00052.safetensors", + "language_model.model.embed_tokens.weight": "model-00001-of-00052.safetensors", + "language_model.model.layers.0.input_layernorm.weight": "model-00002-of-00052.safetensors", + "language_model.model.layers.0.mlp.down_proj.weight": "model-00002-of-00052.safetensors", + "language_model.model.layers.0.mlp.gate_proj.weight": "model-00002-of-00052.safetensors", + "language_model.model.layers.0.mlp.up_proj.weight": "model-00002-of-00052.safetensors", + "language_model.model.layers.0.post_attention_layernorm.weight": "model-00002-of-00052.safetensors", + "language_model.model.layers.0.self_attn.k_proj.weight": "model-00001-of-00052.safetensors", + "language_model.model.layers.0.self_attn.o_proj.weight": "model-00001-of-00052.safetensors", + "language_model.model.layers.0.self_attn.q_proj.weight": "model-00001-of-00052.safetensors", + "language_model.model.layers.0.self_attn.v_proj.weight": "model-00001-of-00052.safetensors", + "language_model.model.layers.1.input_layernorm.weight": "model-00002-of-00052.safetensors", + "language_model.model.layers.1.mlp.down_proj.weight": "model-00002-of-00052.safetensors", + "language_model.model.layers.1.mlp.gate_proj.weight": "model-00002-of-00052.safetensors", + "language_model.model.layers.1.mlp.up_proj.weight": "model-00002-of-00052.safetensors", + "language_model.model.layers.1.post_attention_layernorm.weight": "model-00002-of-00052.safetensors", + "language_model.model.layers.1.self_attn.k_proj.weight": "model-00002-of-00052.safetensors", + "language_model.model.layers.1.self_attn.o_proj.weight": "model-00002-of-00052.safetensors", + "language_model.model.layers.1.self_attn.q_proj.weight": "model-00002-of-00052.safetensors", + "language_model.model.layers.1.self_attn.v_proj.weight": "model-00002-of-00052.safetensors", + "language_model.model.layers.10.input_layernorm.weight": "model-00008-of-00052.safetensors", + "language_model.model.layers.10.mlp.down_proj.weight": "model-00008-of-00052.safetensors", + "language_model.model.layers.10.mlp.gate_proj.weight": "model-00007-of-00052.safetensors", + "language_model.model.layers.10.mlp.up_proj.weight": "model-00007-of-00052.safetensors", + "language_model.model.layers.10.post_attention_layernorm.weight": "model-00008-of-00052.safetensors", + "language_model.model.layers.10.self_attn.k_proj.weight": "model-00007-of-00052.safetensors", + "language_model.model.layers.10.self_attn.o_proj.weight": "model-00007-of-00052.safetensors", + "language_model.model.layers.10.self_attn.q_proj.weight": "model-00007-of-00052.safetensors", + "language_model.model.layers.10.self_attn.v_proj.weight": "model-00007-of-00052.safetensors", + "language_model.model.layers.11.input_layernorm.weight": "model-00008-of-00052.safetensors", + "language_model.model.layers.11.mlp.down_proj.weight": "model-00008-of-00052.safetensors", + "language_model.model.layers.11.mlp.gate_proj.weight": "model-00008-of-00052.safetensors", + "language_model.model.layers.11.mlp.up_proj.weight": "model-00008-of-00052.safetensors", + "language_model.model.layers.11.post_attention_layernorm.weight": "model-00008-of-00052.safetensors", + "language_model.model.layers.11.self_attn.k_proj.weight": "model-00008-of-00052.safetensors", + "language_model.model.layers.11.self_attn.o_proj.weight": "model-00008-of-00052.safetensors", + "language_model.model.layers.11.self_attn.q_proj.weight": "model-00008-of-00052.safetensors", + "language_model.model.layers.11.self_attn.v_proj.weight": "model-00008-of-00052.safetensors", + "language_model.model.layers.12.input_layernorm.weight": "model-00009-of-00052.safetensors", + "language_model.model.layers.12.mlp.down_proj.weight": "model-00009-of-00052.safetensors", + "language_model.model.layers.12.mlp.gate_proj.weight": "model-00008-of-00052.safetensors", + "language_model.model.layers.12.mlp.up_proj.weight": "model-00009-of-00052.safetensors", + "language_model.model.layers.12.post_attention_layernorm.weight": "model-00009-of-00052.safetensors", + "language_model.model.layers.12.self_attn.k_proj.weight": "model-00008-of-00052.safetensors", + "language_model.model.layers.12.self_attn.o_proj.weight": "model-00008-of-00052.safetensors", + "language_model.model.layers.12.self_attn.q_proj.weight": "model-00008-of-00052.safetensors", + "language_model.model.layers.12.self_attn.v_proj.weight": "model-00008-of-00052.safetensors", + "language_model.model.layers.13.input_layernorm.weight": "model-00009-of-00052.safetensors", + "language_model.model.layers.13.mlp.down_proj.weight": "model-00009-of-00052.safetensors", + "language_model.model.layers.13.mlp.gate_proj.weight": "model-00009-of-00052.safetensors", + "language_model.model.layers.13.mlp.up_proj.weight": "model-00009-of-00052.safetensors", + "language_model.model.layers.13.post_attention_layernorm.weight": "model-00009-of-00052.safetensors", + "language_model.model.layers.13.self_attn.k_proj.weight": "model-00009-of-00052.safetensors", + "language_model.model.layers.13.self_attn.o_proj.weight": "model-00009-of-00052.safetensors", + "language_model.model.layers.13.self_attn.q_proj.weight": "model-00009-of-00052.safetensors", + "language_model.model.layers.13.self_attn.v_proj.weight": "model-00009-of-00052.safetensors", + "language_model.model.layers.14.input_layernorm.weight": "model-00010-of-00052.safetensors", + "language_model.model.layers.14.mlp.down_proj.weight": "model-00010-of-00052.safetensors", + "language_model.model.layers.14.mlp.gate_proj.weight": "model-00010-of-00052.safetensors", + "language_model.model.layers.14.mlp.up_proj.weight": "model-00010-of-00052.safetensors", + "language_model.model.layers.14.post_attention_layernorm.weight": "model-00010-of-00052.safetensors", + "language_model.model.layers.14.self_attn.k_proj.weight": "model-00009-of-00052.safetensors", + "language_model.model.layers.14.self_attn.o_proj.weight": "model-00009-of-00052.safetensors", + "language_model.model.layers.14.self_attn.q_proj.weight": "model-00009-of-00052.safetensors", + "language_model.model.layers.14.self_attn.v_proj.weight": "model-00009-of-00052.safetensors", + "language_model.model.layers.15.input_layernorm.weight": "model-00010-of-00052.safetensors", + "language_model.model.layers.15.mlp.down_proj.weight": "model-00010-of-00052.safetensors", + "language_model.model.layers.15.mlp.gate_proj.weight": "model-00010-of-00052.safetensors", + "language_model.model.layers.15.mlp.up_proj.weight": "model-00010-of-00052.safetensors", + "language_model.model.layers.15.post_attention_layernorm.weight": "model-00010-of-00052.safetensors", + "language_model.model.layers.15.self_attn.k_proj.weight": "model-00010-of-00052.safetensors", + "language_model.model.layers.15.self_attn.o_proj.weight": "model-00010-of-00052.safetensors", + "language_model.model.layers.15.self_attn.q_proj.weight": "model-00010-of-00052.safetensors", + "language_model.model.layers.15.self_attn.v_proj.weight": "model-00010-of-00052.safetensors", + "language_model.model.layers.16.input_layernorm.weight": "model-00011-of-00052.safetensors", + "language_model.model.layers.16.mlp.down_proj.weight": "model-00011-of-00052.safetensors", + "language_model.model.layers.16.mlp.gate_proj.weight": "model-00011-of-00052.safetensors", + "language_model.model.layers.16.mlp.up_proj.weight": "model-00011-of-00052.safetensors", + "language_model.model.layers.16.post_attention_layernorm.weight": "model-00011-of-00052.safetensors", + "language_model.model.layers.16.self_attn.k_proj.weight": "model-00011-of-00052.safetensors", + "language_model.model.layers.16.self_attn.o_proj.weight": "model-00011-of-00052.safetensors", + "language_model.model.layers.16.self_attn.q_proj.weight": "model-00011-of-00052.safetensors", + "language_model.model.layers.16.self_attn.v_proj.weight": "model-00011-of-00052.safetensors", + "language_model.model.layers.17.input_layernorm.weight": "model-00012-of-00052.safetensors", + "language_model.model.layers.17.mlp.down_proj.weight": "model-00012-of-00052.safetensors", + "language_model.model.layers.17.mlp.gate_proj.weight": "model-00011-of-00052.safetensors", + "language_model.model.layers.17.mlp.up_proj.weight": "model-00011-of-00052.safetensors", + "language_model.model.layers.17.post_attention_layernorm.weight": "model-00012-of-00052.safetensors", + "language_model.model.layers.17.self_attn.k_proj.weight": "model-00011-of-00052.safetensors", + "language_model.model.layers.17.self_attn.o_proj.weight": "model-00011-of-00052.safetensors", + "language_model.model.layers.17.self_attn.q_proj.weight": "model-00011-of-00052.safetensors", + "language_model.model.layers.17.self_attn.v_proj.weight": "model-00011-of-00052.safetensors", + "language_model.model.layers.18.input_layernorm.weight": "model-00012-of-00052.safetensors", + "language_model.model.layers.18.mlp.down_proj.weight": "model-00012-of-00052.safetensors", + "language_model.model.layers.18.mlp.gate_proj.weight": "model-00012-of-00052.safetensors", + "language_model.model.layers.18.mlp.up_proj.weight": "model-00012-of-00052.safetensors", + "language_model.model.layers.18.post_attention_layernorm.weight": "model-00012-of-00052.safetensors", + "language_model.model.layers.18.self_attn.k_proj.weight": "model-00012-of-00052.safetensors", + "language_model.model.layers.18.self_attn.o_proj.weight": "model-00012-of-00052.safetensors", + "language_model.model.layers.18.self_attn.q_proj.weight": "model-00012-of-00052.safetensors", + "language_model.model.layers.18.self_attn.v_proj.weight": "model-00012-of-00052.safetensors", + "language_model.model.layers.19.input_layernorm.weight": "model-00013-of-00052.safetensors", + "language_model.model.layers.19.mlp.down_proj.weight": "model-00013-of-00052.safetensors", + "language_model.model.layers.19.mlp.gate_proj.weight": "model-00012-of-00052.safetensors", + "language_model.model.layers.19.mlp.up_proj.weight": "model-00013-of-00052.safetensors", + "language_model.model.layers.19.post_attention_layernorm.weight": "model-00013-of-00052.safetensors", + "language_model.model.layers.19.self_attn.k_proj.weight": "model-00012-of-00052.safetensors", + "language_model.model.layers.19.self_attn.o_proj.weight": "model-00012-of-00052.safetensors", + "language_model.model.layers.19.self_attn.q_proj.weight": "model-00012-of-00052.safetensors", + "language_model.model.layers.19.self_attn.v_proj.weight": "model-00012-of-00052.safetensors", + "language_model.model.layers.2.input_layernorm.weight": "model-00003-of-00052.safetensors", + "language_model.model.layers.2.mlp.down_proj.weight": "model-00003-of-00052.safetensors", + "language_model.model.layers.2.mlp.gate_proj.weight": "model-00003-of-00052.safetensors", + "language_model.model.layers.2.mlp.up_proj.weight": "model-00003-of-00052.safetensors", + "language_model.model.layers.2.post_attention_layernorm.weight": "model-00003-of-00052.safetensors", + "language_model.model.layers.2.self_attn.k_proj.weight": "model-00003-of-00052.safetensors", + "language_model.model.layers.2.self_attn.o_proj.weight": "model-00003-of-00052.safetensors", + "language_model.model.layers.2.self_attn.q_proj.weight": "model-00003-of-00052.safetensors", + "language_model.model.layers.2.self_attn.v_proj.weight": "model-00003-of-00052.safetensors", + "language_model.model.layers.20.input_layernorm.weight": "model-00013-of-00052.safetensors", + "language_model.model.layers.20.mlp.down_proj.weight": "model-00013-of-00052.safetensors", + "language_model.model.layers.20.mlp.gate_proj.weight": "model-00013-of-00052.safetensors", + "language_model.model.layers.20.mlp.up_proj.weight": "model-00013-of-00052.safetensors", + "language_model.model.layers.20.post_attention_layernorm.weight": "model-00013-of-00052.safetensors", + "language_model.model.layers.20.self_attn.k_proj.weight": "model-00013-of-00052.safetensors", + "language_model.model.layers.20.self_attn.o_proj.weight": "model-00013-of-00052.safetensors", + "language_model.model.layers.20.self_attn.q_proj.weight": "model-00013-of-00052.safetensors", + "language_model.model.layers.20.self_attn.v_proj.weight": "model-00013-of-00052.safetensors", + "language_model.model.layers.21.input_layernorm.weight": "model-00014-of-00052.safetensors", + "language_model.model.layers.21.mlp.down_proj.weight": "model-00014-of-00052.safetensors", + "language_model.model.layers.21.mlp.gate_proj.weight": "model-00014-of-00052.safetensors", + "language_model.model.layers.21.mlp.up_proj.weight": "model-00014-of-00052.safetensors", + "language_model.model.layers.21.post_attention_layernorm.weight": "model-00014-of-00052.safetensors", + "language_model.model.layers.21.self_attn.k_proj.weight": "model-00013-of-00052.safetensors", + "language_model.model.layers.21.self_attn.o_proj.weight": "model-00013-of-00052.safetensors", + "language_model.model.layers.21.self_attn.q_proj.weight": "model-00013-of-00052.safetensors", + "language_model.model.layers.21.self_attn.v_proj.weight": "model-00013-of-00052.safetensors", + "language_model.model.layers.22.input_layernorm.weight": "model-00014-of-00052.safetensors", + "language_model.model.layers.22.mlp.down_proj.weight": "model-00014-of-00052.safetensors", + "language_model.model.layers.22.mlp.gate_proj.weight": "model-00014-of-00052.safetensors", + "language_model.model.layers.22.mlp.up_proj.weight": "model-00014-of-00052.safetensors", + "language_model.model.layers.22.post_attention_layernorm.weight": "model-00014-of-00052.safetensors", + "language_model.model.layers.22.self_attn.k_proj.weight": "model-00014-of-00052.safetensors", + "language_model.model.layers.22.self_attn.o_proj.weight": "model-00014-of-00052.safetensors", + "language_model.model.layers.22.self_attn.q_proj.weight": "model-00014-of-00052.safetensors", + "language_model.model.layers.22.self_attn.v_proj.weight": "model-00014-of-00052.safetensors", + "language_model.model.layers.23.input_layernorm.weight": "model-00015-of-00052.safetensors", + "language_model.model.layers.23.mlp.down_proj.weight": "model-00015-of-00052.safetensors", + "language_model.model.layers.23.mlp.gate_proj.weight": "model-00015-of-00052.safetensors", + "language_model.model.layers.23.mlp.up_proj.weight": "model-00015-of-00052.safetensors", + "language_model.model.layers.23.post_attention_layernorm.weight": "model-00015-of-00052.safetensors", + "language_model.model.layers.23.self_attn.k_proj.weight": "model-00015-of-00052.safetensors", + "language_model.model.layers.23.self_attn.o_proj.weight": "model-00015-of-00052.safetensors", + "language_model.model.layers.23.self_attn.q_proj.weight": "model-00015-of-00052.safetensors", + "language_model.model.layers.23.self_attn.v_proj.weight": "model-00015-of-00052.safetensors", + "language_model.model.layers.24.input_layernorm.weight": "model-00016-of-00052.safetensors", + "language_model.model.layers.24.mlp.down_proj.weight": "model-00016-of-00052.safetensors", + "language_model.model.layers.24.mlp.gate_proj.weight": "model-00015-of-00052.safetensors", + "language_model.model.layers.24.mlp.up_proj.weight": "model-00015-of-00052.safetensors", + "language_model.model.layers.24.post_attention_layernorm.weight": "model-00016-of-00052.safetensors", + "language_model.model.layers.24.self_attn.k_proj.weight": "model-00015-of-00052.safetensors", + "language_model.model.layers.24.self_attn.o_proj.weight": "model-00015-of-00052.safetensors", + "language_model.model.layers.24.self_attn.q_proj.weight": "model-00015-of-00052.safetensors", + "language_model.model.layers.24.self_attn.v_proj.weight": "model-00015-of-00052.safetensors", + "language_model.model.layers.25.input_layernorm.weight": "model-00016-of-00052.safetensors", + "language_model.model.layers.25.mlp.down_proj.weight": "model-00016-of-00052.safetensors", + "language_model.model.layers.25.mlp.gate_proj.weight": "model-00016-of-00052.safetensors", + "language_model.model.layers.25.mlp.up_proj.weight": "model-00016-of-00052.safetensors", + "language_model.model.layers.25.post_attention_layernorm.weight": "model-00016-of-00052.safetensors", + "language_model.model.layers.25.self_attn.k_proj.weight": "model-00016-of-00052.safetensors", + "language_model.model.layers.25.self_attn.o_proj.weight": "model-00016-of-00052.safetensors", + "language_model.model.layers.25.self_attn.q_proj.weight": "model-00016-of-00052.safetensors", + "language_model.model.layers.25.self_attn.v_proj.weight": "model-00016-of-00052.safetensors", + "language_model.model.layers.26.input_layernorm.weight": "model-00017-of-00052.safetensors", + "language_model.model.layers.26.mlp.down_proj.weight": "model-00017-of-00052.safetensors", + "language_model.model.layers.26.mlp.gate_proj.weight": "model-00016-of-00052.safetensors", + "language_model.model.layers.26.mlp.up_proj.weight": "model-00017-of-00052.safetensors", + "language_model.model.layers.26.post_attention_layernorm.weight": "model-00017-of-00052.safetensors", + "language_model.model.layers.26.self_attn.k_proj.weight": "model-00016-of-00052.safetensors", + "language_model.model.layers.26.self_attn.o_proj.weight": "model-00016-of-00052.safetensors", + "language_model.model.layers.26.self_attn.q_proj.weight": "model-00016-of-00052.safetensors", + "language_model.model.layers.26.self_attn.v_proj.weight": "model-00016-of-00052.safetensors", + "language_model.model.layers.27.input_layernorm.weight": "model-00017-of-00052.safetensors", + "language_model.model.layers.27.mlp.down_proj.weight": "model-00017-of-00052.safetensors", + "language_model.model.layers.27.mlp.gate_proj.weight": "model-00017-of-00052.safetensors", + "language_model.model.layers.27.mlp.up_proj.weight": "model-00017-of-00052.safetensors", + "language_model.model.layers.27.post_attention_layernorm.weight": "model-00017-of-00052.safetensors", + "language_model.model.layers.27.self_attn.k_proj.weight": "model-00017-of-00052.safetensors", + "language_model.model.layers.27.self_attn.o_proj.weight": "model-00017-of-00052.safetensors", + "language_model.model.layers.27.self_attn.q_proj.weight": "model-00017-of-00052.safetensors", + "language_model.model.layers.27.self_attn.v_proj.weight": "model-00017-of-00052.safetensors", + "language_model.model.layers.28.input_layernorm.weight": "model-00018-of-00052.safetensors", + "language_model.model.layers.28.mlp.down_proj.weight": "model-00018-of-00052.safetensors", + "language_model.model.layers.28.mlp.gate_proj.weight": "model-00018-of-00052.safetensors", + "language_model.model.layers.28.mlp.up_proj.weight": "model-00018-of-00052.safetensors", + "language_model.model.layers.28.post_attention_layernorm.weight": "model-00018-of-00052.safetensors", + "language_model.model.layers.28.self_attn.k_proj.weight": "model-00017-of-00052.safetensors", + "language_model.model.layers.28.self_attn.o_proj.weight": "model-00017-of-00052.safetensors", + "language_model.model.layers.28.self_attn.q_proj.weight": "model-00017-of-00052.safetensors", + "language_model.model.layers.28.self_attn.v_proj.weight": "model-00017-of-00052.safetensors", + "language_model.model.layers.29.input_layernorm.weight": "model-00018-of-00052.safetensors", + "language_model.model.layers.29.mlp.down_proj.weight": "model-00018-of-00052.safetensors", + "language_model.model.layers.29.mlp.gate_proj.weight": "model-00018-of-00052.safetensors", + "language_model.model.layers.29.mlp.up_proj.weight": "model-00018-of-00052.safetensors", + "language_model.model.layers.29.post_attention_layernorm.weight": "model-00018-of-00052.safetensors", + "language_model.model.layers.29.self_attn.k_proj.weight": "model-00018-of-00052.safetensors", + "language_model.model.layers.29.self_attn.o_proj.weight": "model-00018-of-00052.safetensors", + "language_model.model.layers.29.self_attn.q_proj.weight": "model-00018-of-00052.safetensors", + "language_model.model.layers.29.self_attn.v_proj.weight": "model-00018-of-00052.safetensors", + "language_model.model.layers.3.input_layernorm.weight": "model-00004-of-00052.safetensors", + "language_model.model.layers.3.mlp.down_proj.weight": "model-00004-of-00052.safetensors", + "language_model.model.layers.3.mlp.gate_proj.weight": "model-00003-of-00052.safetensors", + "language_model.model.layers.3.mlp.up_proj.weight": "model-00003-of-00052.safetensors", + "language_model.model.layers.3.post_attention_layernorm.weight": "model-00004-of-00052.safetensors", + "language_model.model.layers.3.self_attn.k_proj.weight": "model-00003-of-00052.safetensors", + "language_model.model.layers.3.self_attn.o_proj.weight": "model-00003-of-00052.safetensors", + "language_model.model.layers.3.self_attn.q_proj.weight": "model-00003-of-00052.safetensors", + "language_model.model.layers.3.self_attn.v_proj.weight": "model-00003-of-00052.safetensors", + "language_model.model.layers.30.input_layernorm.weight": "model-00019-of-00052.safetensors", + "language_model.model.layers.30.mlp.down_proj.weight": "model-00019-of-00052.safetensors", + "language_model.model.layers.30.mlp.gate_proj.weight": "model-00019-of-00052.safetensors", + "language_model.model.layers.30.mlp.up_proj.weight": "model-00019-of-00052.safetensors", + "language_model.model.layers.30.post_attention_layernorm.weight": "model-00019-of-00052.safetensors", + "language_model.model.layers.30.self_attn.k_proj.weight": "model-00019-of-00052.safetensors", + "language_model.model.layers.30.self_attn.o_proj.weight": "model-00019-of-00052.safetensors", + "language_model.model.layers.30.self_attn.q_proj.weight": "model-00019-of-00052.safetensors", + "language_model.model.layers.30.self_attn.v_proj.weight": "model-00019-of-00052.safetensors", + "language_model.model.layers.31.input_layernorm.weight": "model-00020-of-00052.safetensors", + "language_model.model.layers.31.mlp.down_proj.weight": "model-00020-of-00052.safetensors", + "language_model.model.layers.31.mlp.gate_proj.weight": "model-00019-of-00052.safetensors", + "language_model.model.layers.31.mlp.up_proj.weight": "model-00019-of-00052.safetensors", + "language_model.model.layers.31.post_attention_layernorm.weight": "model-00020-of-00052.safetensors", + "language_model.model.layers.31.self_attn.k_proj.weight": "model-00019-of-00052.safetensors", + "language_model.model.layers.31.self_attn.o_proj.weight": "model-00019-of-00052.safetensors", + "language_model.model.layers.31.self_attn.q_proj.weight": "model-00019-of-00052.safetensors", + "language_model.model.layers.31.self_attn.v_proj.weight": "model-00019-of-00052.safetensors", + "language_model.model.layers.32.input_layernorm.weight": "model-00020-of-00052.safetensors", + "language_model.model.layers.32.mlp.down_proj.weight": "model-00020-of-00052.safetensors", + "language_model.model.layers.32.mlp.gate_proj.weight": "model-00020-of-00052.safetensors", + "language_model.model.layers.32.mlp.up_proj.weight": "model-00020-of-00052.safetensors", + "language_model.model.layers.32.post_attention_layernorm.weight": "model-00020-of-00052.safetensors", + "language_model.model.layers.32.self_attn.k_proj.weight": "model-00020-of-00052.safetensors", + "language_model.model.layers.32.self_attn.o_proj.weight": "model-00020-of-00052.safetensors", + "language_model.model.layers.32.self_attn.q_proj.weight": "model-00020-of-00052.safetensors", + "language_model.model.layers.32.self_attn.v_proj.weight": "model-00020-of-00052.safetensors", + "language_model.model.layers.33.input_layernorm.weight": "model-00021-of-00052.safetensors", + "language_model.model.layers.33.mlp.down_proj.weight": "model-00021-of-00052.safetensors", + "language_model.model.layers.33.mlp.gate_proj.weight": "model-00020-of-00052.safetensors", + "language_model.model.layers.33.mlp.up_proj.weight": "model-00021-of-00052.safetensors", + "language_model.model.layers.33.post_attention_layernorm.weight": "model-00021-of-00052.safetensors", + "language_model.model.layers.33.self_attn.k_proj.weight": "model-00020-of-00052.safetensors", + "language_model.model.layers.33.self_attn.o_proj.weight": "model-00020-of-00052.safetensors", + "language_model.model.layers.33.self_attn.q_proj.weight": "model-00020-of-00052.safetensors", + "language_model.model.layers.33.self_attn.v_proj.weight": "model-00020-of-00052.safetensors", + "language_model.model.layers.34.input_layernorm.weight": "model-00021-of-00052.safetensors", + "language_model.model.layers.34.mlp.down_proj.weight": "model-00021-of-00052.safetensors", + "language_model.model.layers.34.mlp.gate_proj.weight": "model-00021-of-00052.safetensors", + "language_model.model.layers.34.mlp.up_proj.weight": "model-00021-of-00052.safetensors", + "language_model.model.layers.34.post_attention_layernorm.weight": "model-00021-of-00052.safetensors", + "language_model.model.layers.34.self_attn.k_proj.weight": "model-00021-of-00052.safetensors", + "language_model.model.layers.34.self_attn.o_proj.weight": "model-00021-of-00052.safetensors", + "language_model.model.layers.34.self_attn.q_proj.weight": "model-00021-of-00052.safetensors", + "language_model.model.layers.34.self_attn.v_proj.weight": "model-00021-of-00052.safetensors", + "language_model.model.layers.35.input_layernorm.weight": "model-00022-of-00052.safetensors", + "language_model.model.layers.35.mlp.down_proj.weight": "model-00022-of-00052.safetensors", + "language_model.model.layers.35.mlp.gate_proj.weight": "model-00022-of-00052.safetensors", + "language_model.model.layers.35.mlp.up_proj.weight": "model-00022-of-00052.safetensors", + "language_model.model.layers.35.post_attention_layernorm.weight": "model-00022-of-00052.safetensors", + "language_model.model.layers.35.self_attn.k_proj.weight": "model-00021-of-00052.safetensors", + "language_model.model.layers.35.self_attn.o_proj.weight": "model-00021-of-00052.safetensors", + "language_model.model.layers.35.self_attn.q_proj.weight": "model-00021-of-00052.safetensors", + "language_model.model.layers.35.self_attn.v_proj.weight": "model-00021-of-00052.safetensors", + "language_model.model.layers.36.input_layernorm.weight": "model-00022-of-00052.safetensors", + "language_model.model.layers.36.mlp.down_proj.weight": "model-00022-of-00052.safetensors", + "language_model.model.layers.36.mlp.gate_proj.weight": "model-00022-of-00052.safetensors", + "language_model.model.layers.36.mlp.up_proj.weight": "model-00022-of-00052.safetensors", + "language_model.model.layers.36.post_attention_layernorm.weight": "model-00022-of-00052.safetensors", + "language_model.model.layers.36.self_attn.k_proj.weight": "model-00022-of-00052.safetensors", + "language_model.model.layers.36.self_attn.o_proj.weight": "model-00022-of-00052.safetensors", + "language_model.model.layers.36.self_attn.q_proj.weight": "model-00022-of-00052.safetensors", + "language_model.model.layers.36.self_attn.v_proj.weight": "model-00022-of-00052.safetensors", + "language_model.model.layers.37.input_layernorm.weight": "model-00023-of-00052.safetensors", + "language_model.model.layers.37.mlp.down_proj.weight": "model-00023-of-00052.safetensors", + "language_model.model.layers.37.mlp.gate_proj.weight": "model-00023-of-00052.safetensors", + "language_model.model.layers.37.mlp.up_proj.weight": "model-00023-of-00052.safetensors", + "language_model.model.layers.37.post_attention_layernorm.weight": "model-00023-of-00052.safetensors", + "language_model.model.layers.37.self_attn.k_proj.weight": "model-00023-of-00052.safetensors", + "language_model.model.layers.37.self_attn.o_proj.weight": "model-00023-of-00052.safetensors", + "language_model.model.layers.37.self_attn.q_proj.weight": "model-00023-of-00052.safetensors", + "language_model.model.layers.37.self_attn.v_proj.weight": "model-00023-of-00052.safetensors", + "language_model.model.layers.38.input_layernorm.weight": "model-00024-of-00052.safetensors", + "language_model.model.layers.38.mlp.down_proj.weight": "model-00024-of-00052.safetensors", + "language_model.model.layers.38.mlp.gate_proj.weight": "model-00023-of-00052.safetensors", + "language_model.model.layers.38.mlp.up_proj.weight": "model-00023-of-00052.safetensors", + "language_model.model.layers.38.post_attention_layernorm.weight": "model-00024-of-00052.safetensors", + "language_model.model.layers.38.self_attn.k_proj.weight": "model-00023-of-00052.safetensors", + "language_model.model.layers.38.self_attn.o_proj.weight": "model-00023-of-00052.safetensors", + "language_model.model.layers.38.self_attn.q_proj.weight": "model-00023-of-00052.safetensors", + "language_model.model.layers.38.self_attn.v_proj.weight": "model-00023-of-00052.safetensors", + "language_model.model.layers.39.input_layernorm.weight": "model-00024-of-00052.safetensors", + "language_model.model.layers.39.mlp.down_proj.weight": "model-00024-of-00052.safetensors", + "language_model.model.layers.39.mlp.gate_proj.weight": "model-00024-of-00052.safetensors", + "language_model.model.layers.39.mlp.up_proj.weight": "model-00024-of-00052.safetensors", + "language_model.model.layers.39.post_attention_layernorm.weight": "model-00024-of-00052.safetensors", + "language_model.model.layers.39.self_attn.k_proj.weight": "model-00024-of-00052.safetensors", + "language_model.model.layers.39.self_attn.o_proj.weight": "model-00024-of-00052.safetensors", + "language_model.model.layers.39.self_attn.q_proj.weight": "model-00024-of-00052.safetensors", + "language_model.model.layers.39.self_attn.v_proj.weight": "model-00024-of-00052.safetensors", + "language_model.model.layers.4.input_layernorm.weight": "model-00004-of-00052.safetensors", + "language_model.model.layers.4.mlp.down_proj.weight": "model-00004-of-00052.safetensors", + "language_model.model.layers.4.mlp.gate_proj.weight": "model-00004-of-00052.safetensors", + "language_model.model.layers.4.mlp.up_proj.weight": "model-00004-of-00052.safetensors", + "language_model.model.layers.4.post_attention_layernorm.weight": "model-00004-of-00052.safetensors", + "language_model.model.layers.4.self_attn.k_proj.weight": "model-00004-of-00052.safetensors", + "language_model.model.layers.4.self_attn.o_proj.weight": "model-00004-of-00052.safetensors", + "language_model.model.layers.4.self_attn.q_proj.weight": "model-00004-of-00052.safetensors", + "language_model.model.layers.4.self_attn.v_proj.weight": "model-00004-of-00052.safetensors", + "language_model.model.layers.40.input_layernorm.weight": "model-00025-of-00052.safetensors", + "language_model.model.layers.40.mlp.down_proj.weight": "model-00025-of-00052.safetensors", + "language_model.model.layers.40.mlp.gate_proj.weight": "model-00024-of-00052.safetensors", + "language_model.model.layers.40.mlp.up_proj.weight": "model-00025-of-00052.safetensors", + "language_model.model.layers.40.post_attention_layernorm.weight": "model-00025-of-00052.safetensors", + "language_model.model.layers.40.self_attn.k_proj.weight": "model-00024-of-00052.safetensors", + "language_model.model.layers.40.self_attn.o_proj.weight": "model-00024-of-00052.safetensors", + "language_model.model.layers.40.self_attn.q_proj.weight": "model-00024-of-00052.safetensors", + "language_model.model.layers.40.self_attn.v_proj.weight": "model-00024-of-00052.safetensors", + "language_model.model.layers.41.input_layernorm.weight": "model-00025-of-00052.safetensors", + "language_model.model.layers.41.mlp.down_proj.weight": "model-00025-of-00052.safetensors", + "language_model.model.layers.41.mlp.gate_proj.weight": "model-00025-of-00052.safetensors", + "language_model.model.layers.41.mlp.up_proj.weight": "model-00025-of-00052.safetensors", + "language_model.model.layers.41.post_attention_layernorm.weight": "model-00025-of-00052.safetensors", + "language_model.model.layers.41.self_attn.k_proj.weight": "model-00025-of-00052.safetensors", + "language_model.model.layers.41.self_attn.o_proj.weight": "model-00025-of-00052.safetensors", + "language_model.model.layers.41.self_attn.q_proj.weight": "model-00025-of-00052.safetensors", + "language_model.model.layers.41.self_attn.v_proj.weight": "model-00025-of-00052.safetensors", + "language_model.model.layers.42.input_layernorm.weight": "model-00026-of-00052.safetensors", + "language_model.model.layers.42.mlp.down_proj.weight": "model-00026-of-00052.safetensors", + "language_model.model.layers.42.mlp.gate_proj.weight": "model-00026-of-00052.safetensors", + "language_model.model.layers.42.mlp.up_proj.weight": "model-00026-of-00052.safetensors", + "language_model.model.layers.42.post_attention_layernorm.weight": "model-00026-of-00052.safetensors", + "language_model.model.layers.42.self_attn.k_proj.weight": "model-00025-of-00052.safetensors", + "language_model.model.layers.42.self_attn.o_proj.weight": "model-00025-of-00052.safetensors", + "language_model.model.layers.42.self_attn.q_proj.weight": "model-00025-of-00052.safetensors", + "language_model.model.layers.42.self_attn.v_proj.weight": "model-00025-of-00052.safetensors", + "language_model.model.layers.43.input_layernorm.weight": "model-00026-of-00052.safetensors", + "language_model.model.layers.43.mlp.down_proj.weight": "model-00026-of-00052.safetensors", + "language_model.model.layers.43.mlp.gate_proj.weight": "model-00026-of-00052.safetensors", + "language_model.model.layers.43.mlp.up_proj.weight": "model-00026-of-00052.safetensors", + "language_model.model.layers.43.post_attention_layernorm.weight": "model-00026-of-00052.safetensors", + "language_model.model.layers.43.self_attn.k_proj.weight": "model-00026-of-00052.safetensors", + "language_model.model.layers.43.self_attn.o_proj.weight": "model-00026-of-00052.safetensors", + "language_model.model.layers.43.self_attn.q_proj.weight": "model-00026-of-00052.safetensors", + "language_model.model.layers.43.self_attn.v_proj.weight": "model-00026-of-00052.safetensors", + "language_model.model.layers.44.input_layernorm.weight": "model-00027-of-00052.safetensors", + "language_model.model.layers.44.mlp.down_proj.weight": "model-00027-of-00052.safetensors", + "language_model.model.layers.44.mlp.gate_proj.weight": "model-00027-of-00052.safetensors", + "language_model.model.layers.44.mlp.up_proj.weight": "model-00027-of-00052.safetensors", + "language_model.model.layers.44.post_attention_layernorm.weight": "model-00027-of-00052.safetensors", + "language_model.model.layers.44.self_attn.k_proj.weight": "model-00027-of-00052.safetensors", + "language_model.model.layers.44.self_attn.o_proj.weight": "model-00027-of-00052.safetensors", + "language_model.model.layers.44.self_attn.q_proj.weight": "model-00027-of-00052.safetensors", + "language_model.model.layers.44.self_attn.v_proj.weight": "model-00027-of-00052.safetensors", + "language_model.model.layers.45.input_layernorm.weight": "model-00028-of-00052.safetensors", + "language_model.model.layers.45.mlp.down_proj.weight": "model-00028-of-00052.safetensors", + "language_model.model.layers.45.mlp.gate_proj.weight": "model-00027-of-00052.safetensors", + "language_model.model.layers.45.mlp.up_proj.weight": "model-00027-of-00052.safetensors", + "language_model.model.layers.45.post_attention_layernorm.weight": "model-00028-of-00052.safetensors", + "language_model.model.layers.45.self_attn.k_proj.weight": "model-00027-of-00052.safetensors", + "language_model.model.layers.45.self_attn.o_proj.weight": "model-00027-of-00052.safetensors", + "language_model.model.layers.45.self_attn.q_proj.weight": "model-00027-of-00052.safetensors", + "language_model.model.layers.45.self_attn.v_proj.weight": "model-00027-of-00052.safetensors", + "language_model.model.layers.46.input_layernorm.weight": "model-00028-of-00052.safetensors", + "language_model.model.layers.46.mlp.down_proj.weight": "model-00028-of-00052.safetensors", + "language_model.model.layers.46.mlp.gate_proj.weight": "model-00028-of-00052.safetensors", + "language_model.model.layers.46.mlp.up_proj.weight": "model-00028-of-00052.safetensors", + "language_model.model.layers.46.post_attention_layernorm.weight": "model-00028-of-00052.safetensors", + "language_model.model.layers.46.self_attn.k_proj.weight": "model-00028-of-00052.safetensors", + "language_model.model.layers.46.self_attn.o_proj.weight": "model-00028-of-00052.safetensors", + "language_model.model.layers.46.self_attn.q_proj.weight": "model-00028-of-00052.safetensors", + "language_model.model.layers.46.self_attn.v_proj.weight": "model-00028-of-00052.safetensors", + "language_model.model.layers.47.input_layernorm.weight": "model-00029-of-00052.safetensors", + "language_model.model.layers.47.mlp.down_proj.weight": "model-00029-of-00052.safetensors", + "language_model.model.layers.47.mlp.gate_proj.weight": "model-00028-of-00052.safetensors", + "language_model.model.layers.47.mlp.up_proj.weight": "model-00029-of-00052.safetensors", + "language_model.model.layers.47.post_attention_layernorm.weight": "model-00029-of-00052.safetensors", + "language_model.model.layers.47.self_attn.k_proj.weight": "model-00028-of-00052.safetensors", + "language_model.model.layers.47.self_attn.o_proj.weight": "model-00028-of-00052.safetensors", + "language_model.model.layers.47.self_attn.q_proj.weight": "model-00028-of-00052.safetensors", + "language_model.model.layers.47.self_attn.v_proj.weight": "model-00028-of-00052.safetensors", + "language_model.model.layers.48.input_layernorm.weight": "model-00029-of-00052.safetensors", + "language_model.model.layers.48.mlp.down_proj.weight": "model-00029-of-00052.safetensors", + "language_model.model.layers.48.mlp.gate_proj.weight": "model-00029-of-00052.safetensors", + "language_model.model.layers.48.mlp.up_proj.weight": "model-00029-of-00052.safetensors", + "language_model.model.layers.48.post_attention_layernorm.weight": "model-00029-of-00052.safetensors", + "language_model.model.layers.48.self_attn.k_proj.weight": "model-00029-of-00052.safetensors", + "language_model.model.layers.48.self_attn.o_proj.weight": "model-00029-of-00052.safetensors", + "language_model.model.layers.48.self_attn.q_proj.weight": "model-00029-of-00052.safetensors", + "language_model.model.layers.48.self_attn.v_proj.weight": "model-00029-of-00052.safetensors", + "language_model.model.layers.49.input_layernorm.weight": "model-00030-of-00052.safetensors", + "language_model.model.layers.49.mlp.down_proj.weight": "model-00030-of-00052.safetensors", + "language_model.model.layers.49.mlp.gate_proj.weight": "model-00030-of-00052.safetensors", + "language_model.model.layers.49.mlp.up_proj.weight": "model-00030-of-00052.safetensors", + "language_model.model.layers.49.post_attention_layernorm.weight": "model-00030-of-00052.safetensors", + "language_model.model.layers.49.self_attn.k_proj.weight": "model-00029-of-00052.safetensors", + "language_model.model.layers.49.self_attn.o_proj.weight": "model-00029-of-00052.safetensors", + "language_model.model.layers.49.self_attn.q_proj.weight": "model-00029-of-00052.safetensors", + "language_model.model.layers.49.self_attn.v_proj.weight": "model-00029-of-00052.safetensors", + "language_model.model.layers.5.input_layernorm.weight": "model-00005-of-00052.safetensors", + "language_model.model.layers.5.mlp.down_proj.weight": "model-00005-of-00052.safetensors", + "language_model.model.layers.5.mlp.gate_proj.weight": "model-00004-of-00052.safetensors", + "language_model.model.layers.5.mlp.up_proj.weight": "model-00005-of-00052.safetensors", + "language_model.model.layers.5.post_attention_layernorm.weight": "model-00005-of-00052.safetensors", + "language_model.model.layers.5.self_attn.k_proj.weight": "model-00004-of-00052.safetensors", + "language_model.model.layers.5.self_attn.o_proj.weight": "model-00004-of-00052.safetensors", + "language_model.model.layers.5.self_attn.q_proj.weight": "model-00004-of-00052.safetensors", + "language_model.model.layers.5.self_attn.v_proj.weight": "model-00004-of-00052.safetensors", + "language_model.model.layers.50.input_layernorm.weight": "model-00030-of-00052.safetensors", + "language_model.model.layers.50.mlp.down_proj.weight": "model-00030-of-00052.safetensors", + "language_model.model.layers.50.mlp.gate_proj.weight": "model-00030-of-00052.safetensors", + "language_model.model.layers.50.mlp.up_proj.weight": "model-00030-of-00052.safetensors", + "language_model.model.layers.50.post_attention_layernorm.weight": "model-00030-of-00052.safetensors", + "language_model.model.layers.50.self_attn.k_proj.weight": "model-00030-of-00052.safetensors", + "language_model.model.layers.50.self_attn.o_proj.weight": "model-00030-of-00052.safetensors", + "language_model.model.layers.50.self_attn.q_proj.weight": "model-00030-of-00052.safetensors", + "language_model.model.layers.50.self_attn.v_proj.weight": "model-00030-of-00052.safetensors", + "language_model.model.layers.51.input_layernorm.weight": "model-00031-of-00052.safetensors", + "language_model.model.layers.51.mlp.down_proj.weight": "model-00031-of-00052.safetensors", + "language_model.model.layers.51.mlp.gate_proj.weight": "model-00031-of-00052.safetensors", + "language_model.model.layers.51.mlp.up_proj.weight": "model-00031-of-00052.safetensors", + "language_model.model.layers.51.post_attention_layernorm.weight": "model-00031-of-00052.safetensors", + "language_model.model.layers.51.self_attn.k_proj.weight": "model-00031-of-00052.safetensors", + "language_model.model.layers.51.self_attn.o_proj.weight": "model-00031-of-00052.safetensors", + "language_model.model.layers.51.self_attn.q_proj.weight": "model-00031-of-00052.safetensors", + "language_model.model.layers.51.self_attn.v_proj.weight": "model-00031-of-00052.safetensors", + "language_model.model.layers.52.input_layernorm.weight": "model-00032-of-00052.safetensors", + "language_model.model.layers.52.mlp.down_proj.weight": "model-00032-of-00052.safetensors", + "language_model.model.layers.52.mlp.gate_proj.weight": "model-00031-of-00052.safetensors", + "language_model.model.layers.52.mlp.up_proj.weight": "model-00031-of-00052.safetensors", + "language_model.model.layers.52.post_attention_layernorm.weight": "model-00032-of-00052.safetensors", + "language_model.model.layers.52.self_attn.k_proj.weight": "model-00031-of-00052.safetensors", + "language_model.model.layers.52.self_attn.o_proj.weight": "model-00031-of-00052.safetensors", + "language_model.model.layers.52.self_attn.q_proj.weight": "model-00031-of-00052.safetensors", + "language_model.model.layers.52.self_attn.v_proj.weight": "model-00031-of-00052.safetensors", + "language_model.model.layers.53.input_layernorm.weight": "model-00032-of-00052.safetensors", + "language_model.model.layers.53.mlp.down_proj.weight": "model-00032-of-00052.safetensors", + "language_model.model.layers.53.mlp.gate_proj.weight": "model-00032-of-00052.safetensors", + "language_model.model.layers.53.mlp.up_proj.weight": "model-00032-of-00052.safetensors", + "language_model.model.layers.53.post_attention_layernorm.weight": "model-00032-of-00052.safetensors", + "language_model.model.layers.53.self_attn.k_proj.weight": "model-00032-of-00052.safetensors", + "language_model.model.layers.53.self_attn.o_proj.weight": "model-00032-of-00052.safetensors", + "language_model.model.layers.53.self_attn.q_proj.weight": "model-00032-of-00052.safetensors", + "language_model.model.layers.53.self_attn.v_proj.weight": "model-00032-of-00052.safetensors", + "language_model.model.layers.54.input_layernorm.weight": "model-00033-of-00052.safetensors", + "language_model.model.layers.54.mlp.down_proj.weight": "model-00033-of-00052.safetensors", + "language_model.model.layers.54.mlp.gate_proj.weight": "model-00032-of-00052.safetensors", + "language_model.model.layers.54.mlp.up_proj.weight": "model-00033-of-00052.safetensors", + "language_model.model.layers.54.post_attention_layernorm.weight": "model-00033-of-00052.safetensors", + "language_model.model.layers.54.self_attn.k_proj.weight": "model-00032-of-00052.safetensors", + "language_model.model.layers.54.self_attn.o_proj.weight": "model-00032-of-00052.safetensors", + "language_model.model.layers.54.self_attn.q_proj.weight": "model-00032-of-00052.safetensors", + "language_model.model.layers.54.self_attn.v_proj.weight": "model-00032-of-00052.safetensors", + "language_model.model.layers.55.input_layernorm.weight": "model-00033-of-00052.safetensors", + "language_model.model.layers.55.mlp.down_proj.weight": "model-00033-of-00052.safetensors", + "language_model.model.layers.55.mlp.gate_proj.weight": "model-00033-of-00052.safetensors", + "language_model.model.layers.55.mlp.up_proj.weight": "model-00033-of-00052.safetensors", + "language_model.model.layers.55.post_attention_layernorm.weight": "model-00033-of-00052.safetensors", + "language_model.model.layers.55.self_attn.k_proj.weight": "model-00033-of-00052.safetensors", + "language_model.model.layers.55.self_attn.o_proj.weight": "model-00033-of-00052.safetensors", + "language_model.model.layers.55.self_attn.q_proj.weight": "model-00033-of-00052.safetensors", + "language_model.model.layers.55.self_attn.v_proj.weight": "model-00033-of-00052.safetensors", + "language_model.model.layers.56.input_layernorm.weight": "model-00034-of-00052.safetensors", + "language_model.model.layers.56.mlp.down_proj.weight": "model-00034-of-00052.safetensors", + "language_model.model.layers.56.mlp.gate_proj.weight": "model-00034-of-00052.safetensors", + "language_model.model.layers.56.mlp.up_proj.weight": "model-00034-of-00052.safetensors", + "language_model.model.layers.56.post_attention_layernorm.weight": "model-00034-of-00052.safetensors", + "language_model.model.layers.56.self_attn.k_proj.weight": "model-00033-of-00052.safetensors", + "language_model.model.layers.56.self_attn.o_proj.weight": "model-00033-of-00052.safetensors", + "language_model.model.layers.56.self_attn.q_proj.weight": "model-00033-of-00052.safetensors", + "language_model.model.layers.56.self_attn.v_proj.weight": "model-00033-of-00052.safetensors", + "language_model.model.layers.57.input_layernorm.weight": "model-00034-of-00052.safetensors", + "language_model.model.layers.57.mlp.down_proj.weight": "model-00034-of-00052.safetensors", + "language_model.model.layers.57.mlp.gate_proj.weight": "model-00034-of-00052.safetensors", + "language_model.model.layers.57.mlp.up_proj.weight": "model-00034-of-00052.safetensors", + "language_model.model.layers.57.post_attention_layernorm.weight": "model-00034-of-00052.safetensors", + "language_model.model.layers.57.self_attn.k_proj.weight": "model-00034-of-00052.safetensors", + "language_model.model.layers.57.self_attn.o_proj.weight": "model-00034-of-00052.safetensors", + "language_model.model.layers.57.self_attn.q_proj.weight": "model-00034-of-00052.safetensors", + "language_model.model.layers.57.self_attn.v_proj.weight": "model-00034-of-00052.safetensors", + "language_model.model.layers.58.input_layernorm.weight": "model-00035-of-00052.safetensors", + "language_model.model.layers.58.mlp.down_proj.weight": "model-00035-of-00052.safetensors", + "language_model.model.layers.58.mlp.gate_proj.weight": "model-00035-of-00052.safetensors", + "language_model.model.layers.58.mlp.up_proj.weight": "model-00035-of-00052.safetensors", + "language_model.model.layers.58.post_attention_layernorm.weight": "model-00035-of-00052.safetensors", + "language_model.model.layers.58.self_attn.k_proj.weight": "model-00035-of-00052.safetensors", + "language_model.model.layers.58.self_attn.o_proj.weight": "model-00035-of-00052.safetensors", + "language_model.model.layers.58.self_attn.q_proj.weight": "model-00035-of-00052.safetensors", + "language_model.model.layers.58.self_attn.v_proj.weight": "model-00035-of-00052.safetensors", + "language_model.model.layers.59.input_layernorm.weight": "model-00036-of-00052.safetensors", + "language_model.model.layers.59.mlp.down_proj.weight": "model-00036-of-00052.safetensors", + "language_model.model.layers.59.mlp.gate_proj.weight": "model-00035-of-00052.safetensors", + "language_model.model.layers.59.mlp.up_proj.weight": "model-00035-of-00052.safetensors", + "language_model.model.layers.59.post_attention_layernorm.weight": "model-00036-of-00052.safetensors", + "language_model.model.layers.59.self_attn.k_proj.weight": "model-00035-of-00052.safetensors", + "language_model.model.layers.59.self_attn.o_proj.weight": "model-00035-of-00052.safetensors", + "language_model.model.layers.59.self_attn.q_proj.weight": "model-00035-of-00052.safetensors", + "language_model.model.layers.59.self_attn.v_proj.weight": "model-00035-of-00052.safetensors", + "language_model.model.layers.6.input_layernorm.weight": "model-00005-of-00052.safetensors", + "language_model.model.layers.6.mlp.down_proj.weight": "model-00005-of-00052.safetensors", + "language_model.model.layers.6.mlp.gate_proj.weight": "model-00005-of-00052.safetensors", + "language_model.model.layers.6.mlp.up_proj.weight": "model-00005-of-00052.safetensors", + "language_model.model.layers.6.post_attention_layernorm.weight": "model-00005-of-00052.safetensors", + "language_model.model.layers.6.self_attn.k_proj.weight": "model-00005-of-00052.safetensors", + "language_model.model.layers.6.self_attn.o_proj.weight": "model-00005-of-00052.safetensors", + "language_model.model.layers.6.self_attn.q_proj.weight": "model-00005-of-00052.safetensors", + "language_model.model.layers.6.self_attn.v_proj.weight": "model-00005-of-00052.safetensors", + "language_model.model.layers.60.input_layernorm.weight": "model-00036-of-00052.safetensors", + "language_model.model.layers.60.mlp.down_proj.weight": "model-00036-of-00052.safetensors", + "language_model.model.layers.60.mlp.gate_proj.weight": "model-00036-of-00052.safetensors", + "language_model.model.layers.60.mlp.up_proj.weight": "model-00036-of-00052.safetensors", + "language_model.model.layers.60.post_attention_layernorm.weight": "model-00036-of-00052.safetensors", + "language_model.model.layers.60.self_attn.k_proj.weight": "model-00036-of-00052.safetensors", + "language_model.model.layers.60.self_attn.o_proj.weight": "model-00036-of-00052.safetensors", + "language_model.model.layers.60.self_attn.q_proj.weight": "model-00036-of-00052.safetensors", + "language_model.model.layers.60.self_attn.v_proj.weight": "model-00036-of-00052.safetensors", + "language_model.model.layers.61.input_layernorm.weight": "model-00037-of-00052.safetensors", + "language_model.model.layers.61.mlp.down_proj.weight": "model-00037-of-00052.safetensors", + "language_model.model.layers.61.mlp.gate_proj.weight": "model-00036-of-00052.safetensors", + "language_model.model.layers.61.mlp.up_proj.weight": "model-00037-of-00052.safetensors", + "language_model.model.layers.61.post_attention_layernorm.weight": "model-00037-of-00052.safetensors", + "language_model.model.layers.61.self_attn.k_proj.weight": "model-00036-of-00052.safetensors", + "language_model.model.layers.61.self_attn.o_proj.weight": "model-00036-of-00052.safetensors", + "language_model.model.layers.61.self_attn.q_proj.weight": "model-00036-of-00052.safetensors", + "language_model.model.layers.61.self_attn.v_proj.weight": "model-00036-of-00052.safetensors", + "language_model.model.layers.62.input_layernorm.weight": "model-00037-of-00052.safetensors", + "language_model.model.layers.62.mlp.down_proj.weight": "model-00037-of-00052.safetensors", + "language_model.model.layers.62.mlp.gate_proj.weight": "model-00037-of-00052.safetensors", + "language_model.model.layers.62.mlp.up_proj.weight": "model-00037-of-00052.safetensors", + "language_model.model.layers.62.post_attention_layernorm.weight": "model-00037-of-00052.safetensors", + "language_model.model.layers.62.self_attn.k_proj.weight": "model-00037-of-00052.safetensors", + "language_model.model.layers.62.self_attn.o_proj.weight": "model-00037-of-00052.safetensors", + "language_model.model.layers.62.self_attn.q_proj.weight": "model-00037-of-00052.safetensors", + "language_model.model.layers.62.self_attn.v_proj.weight": "model-00037-of-00052.safetensors", + "language_model.model.layers.63.input_layernorm.weight": "model-00038-of-00052.safetensors", + "language_model.model.layers.63.mlp.down_proj.weight": "model-00038-of-00052.safetensors", + "language_model.model.layers.63.mlp.gate_proj.weight": "model-00038-of-00052.safetensors", + "language_model.model.layers.63.mlp.up_proj.weight": "model-00038-of-00052.safetensors", + "language_model.model.layers.63.post_attention_layernorm.weight": "model-00038-of-00052.safetensors", + "language_model.model.layers.63.self_attn.k_proj.weight": "model-00037-of-00052.safetensors", + "language_model.model.layers.63.self_attn.o_proj.weight": "model-00037-of-00052.safetensors", + "language_model.model.layers.63.self_attn.q_proj.weight": "model-00037-of-00052.safetensors", + "language_model.model.layers.63.self_attn.v_proj.weight": "model-00037-of-00052.safetensors", + "language_model.model.layers.64.input_layernorm.weight": "model-00038-of-00052.safetensors", + "language_model.model.layers.64.mlp.down_proj.weight": "model-00038-of-00052.safetensors", + "language_model.model.layers.64.mlp.gate_proj.weight": "model-00038-of-00052.safetensors", + "language_model.model.layers.64.mlp.up_proj.weight": "model-00038-of-00052.safetensors", + "language_model.model.layers.64.post_attention_layernorm.weight": "model-00038-of-00052.safetensors", + "language_model.model.layers.64.self_attn.k_proj.weight": "model-00038-of-00052.safetensors", + "language_model.model.layers.64.self_attn.o_proj.weight": "model-00038-of-00052.safetensors", + "language_model.model.layers.64.self_attn.q_proj.weight": "model-00038-of-00052.safetensors", + "language_model.model.layers.64.self_attn.v_proj.weight": "model-00038-of-00052.safetensors", + "language_model.model.layers.65.input_layernorm.weight": "model-00039-of-00052.safetensors", + "language_model.model.layers.65.mlp.down_proj.weight": "model-00039-of-00052.safetensors", + "language_model.model.layers.65.mlp.gate_proj.weight": "model-00039-of-00052.safetensors", + "language_model.model.layers.65.mlp.up_proj.weight": "model-00039-of-00052.safetensors", + "language_model.model.layers.65.post_attention_layernorm.weight": "model-00039-of-00052.safetensors", + "language_model.model.layers.65.self_attn.k_proj.weight": "model-00039-of-00052.safetensors", + "language_model.model.layers.65.self_attn.o_proj.weight": "model-00039-of-00052.safetensors", + "language_model.model.layers.65.self_attn.q_proj.weight": "model-00039-of-00052.safetensors", + "language_model.model.layers.65.self_attn.v_proj.weight": "model-00039-of-00052.safetensors", + "language_model.model.layers.66.input_layernorm.weight": "model-00040-of-00052.safetensors", + "language_model.model.layers.66.mlp.down_proj.weight": "model-00040-of-00052.safetensors", + "language_model.model.layers.66.mlp.gate_proj.weight": "model-00039-of-00052.safetensors", + "language_model.model.layers.66.mlp.up_proj.weight": "model-00039-of-00052.safetensors", + "language_model.model.layers.66.post_attention_layernorm.weight": "model-00040-of-00052.safetensors", + "language_model.model.layers.66.self_attn.k_proj.weight": "model-00039-of-00052.safetensors", + "language_model.model.layers.66.self_attn.o_proj.weight": "model-00039-of-00052.safetensors", + "language_model.model.layers.66.self_attn.q_proj.weight": "model-00039-of-00052.safetensors", + "language_model.model.layers.66.self_attn.v_proj.weight": "model-00039-of-00052.safetensors", + "language_model.model.layers.67.input_layernorm.weight": "model-00040-of-00052.safetensors", + "language_model.model.layers.67.mlp.down_proj.weight": "model-00040-of-00052.safetensors", + "language_model.model.layers.67.mlp.gate_proj.weight": "model-00040-of-00052.safetensors", + "language_model.model.layers.67.mlp.up_proj.weight": "model-00040-of-00052.safetensors", + "language_model.model.layers.67.post_attention_layernorm.weight": "model-00040-of-00052.safetensors", + "language_model.model.layers.67.self_attn.k_proj.weight": "model-00040-of-00052.safetensors", + "language_model.model.layers.67.self_attn.o_proj.weight": "model-00040-of-00052.safetensors", + "language_model.model.layers.67.self_attn.q_proj.weight": "model-00040-of-00052.safetensors", + "language_model.model.layers.67.self_attn.v_proj.weight": "model-00040-of-00052.safetensors", + "language_model.model.layers.68.input_layernorm.weight": "model-00041-of-00052.safetensors", + "language_model.model.layers.68.mlp.down_proj.weight": "model-00041-of-00052.safetensors", + "language_model.model.layers.68.mlp.gate_proj.weight": "model-00040-of-00052.safetensors", + "language_model.model.layers.68.mlp.up_proj.weight": "model-00041-of-00052.safetensors", + "language_model.model.layers.68.post_attention_layernorm.weight": "model-00041-of-00052.safetensors", + "language_model.model.layers.68.self_attn.k_proj.weight": "model-00040-of-00052.safetensors", + "language_model.model.layers.68.self_attn.o_proj.weight": "model-00040-of-00052.safetensors", + "language_model.model.layers.68.self_attn.q_proj.weight": "model-00040-of-00052.safetensors", + "language_model.model.layers.68.self_attn.v_proj.weight": "model-00040-of-00052.safetensors", + "language_model.model.layers.69.input_layernorm.weight": "model-00041-of-00052.safetensors", + "language_model.model.layers.69.mlp.down_proj.weight": "model-00041-of-00052.safetensors", + "language_model.model.layers.69.mlp.gate_proj.weight": "model-00041-of-00052.safetensors", + "language_model.model.layers.69.mlp.up_proj.weight": "model-00041-of-00052.safetensors", + "language_model.model.layers.69.post_attention_layernorm.weight": "model-00041-of-00052.safetensors", + "language_model.model.layers.69.self_attn.k_proj.weight": "model-00041-of-00052.safetensors", + "language_model.model.layers.69.self_attn.o_proj.weight": "model-00041-of-00052.safetensors", + "language_model.model.layers.69.self_attn.q_proj.weight": "model-00041-of-00052.safetensors", + "language_model.model.layers.69.self_attn.v_proj.weight": "model-00041-of-00052.safetensors", + "language_model.model.layers.7.input_layernorm.weight": "model-00006-of-00052.safetensors", + "language_model.model.layers.7.mlp.down_proj.weight": "model-00006-of-00052.safetensors", + "language_model.model.layers.7.mlp.gate_proj.weight": "model-00006-of-00052.safetensors", + "language_model.model.layers.7.mlp.up_proj.weight": "model-00006-of-00052.safetensors", + "language_model.model.layers.7.post_attention_layernorm.weight": "model-00006-of-00052.safetensors", + "language_model.model.layers.7.self_attn.k_proj.weight": "model-00005-of-00052.safetensors", + "language_model.model.layers.7.self_attn.o_proj.weight": "model-00005-of-00052.safetensors", + "language_model.model.layers.7.self_attn.q_proj.weight": "model-00005-of-00052.safetensors", + "language_model.model.layers.7.self_attn.v_proj.weight": "model-00005-of-00052.safetensors", + "language_model.model.layers.70.input_layernorm.weight": "model-00042-of-00052.safetensors", + "language_model.model.layers.70.mlp.down_proj.weight": "model-00042-of-00052.safetensors", + "language_model.model.layers.70.mlp.gate_proj.weight": "model-00042-of-00052.safetensors", + "language_model.model.layers.70.mlp.up_proj.weight": "model-00042-of-00052.safetensors", + "language_model.model.layers.70.post_attention_layernorm.weight": "model-00042-of-00052.safetensors", + "language_model.model.layers.70.self_attn.k_proj.weight": "model-00041-of-00052.safetensors", + "language_model.model.layers.70.self_attn.o_proj.weight": "model-00041-of-00052.safetensors", + "language_model.model.layers.70.self_attn.q_proj.weight": "model-00041-of-00052.safetensors", + "language_model.model.layers.70.self_attn.v_proj.weight": "model-00041-of-00052.safetensors", + "language_model.model.layers.71.input_layernorm.weight": "model-00042-of-00052.safetensors", + "language_model.model.layers.71.mlp.down_proj.weight": "model-00042-of-00052.safetensors", + "language_model.model.layers.71.mlp.gate_proj.weight": "model-00042-of-00052.safetensors", + "language_model.model.layers.71.mlp.up_proj.weight": "model-00042-of-00052.safetensors", + "language_model.model.layers.71.post_attention_layernorm.weight": "model-00042-of-00052.safetensors", + "language_model.model.layers.71.self_attn.k_proj.weight": "model-00042-of-00052.safetensors", + "language_model.model.layers.71.self_attn.o_proj.weight": "model-00042-of-00052.safetensors", + "language_model.model.layers.71.self_attn.q_proj.weight": "model-00042-of-00052.safetensors", + "language_model.model.layers.71.self_attn.v_proj.weight": "model-00042-of-00052.safetensors", + "language_model.model.layers.72.input_layernorm.weight": "model-00043-of-00052.safetensors", + "language_model.model.layers.72.mlp.down_proj.weight": "model-00043-of-00052.safetensors", + "language_model.model.layers.72.mlp.gate_proj.weight": "model-00043-of-00052.safetensors", + "language_model.model.layers.72.mlp.up_proj.weight": "model-00043-of-00052.safetensors", + "language_model.model.layers.72.post_attention_layernorm.weight": "model-00043-of-00052.safetensors", + "language_model.model.layers.72.self_attn.k_proj.weight": "model-00043-of-00052.safetensors", + "language_model.model.layers.72.self_attn.o_proj.weight": "model-00043-of-00052.safetensors", + "language_model.model.layers.72.self_attn.q_proj.weight": "model-00043-of-00052.safetensors", + "language_model.model.layers.72.self_attn.v_proj.weight": "model-00043-of-00052.safetensors", + "language_model.model.layers.73.input_layernorm.weight": "model-00044-of-00052.safetensors", + "language_model.model.layers.73.mlp.down_proj.weight": "model-00044-of-00052.safetensors", + "language_model.model.layers.73.mlp.gate_proj.weight": "model-00043-of-00052.safetensors", + "language_model.model.layers.73.mlp.up_proj.weight": "model-00043-of-00052.safetensors", + "language_model.model.layers.73.post_attention_layernorm.weight": "model-00044-of-00052.safetensors", + "language_model.model.layers.73.self_attn.k_proj.weight": "model-00043-of-00052.safetensors", + "language_model.model.layers.73.self_attn.o_proj.weight": "model-00043-of-00052.safetensors", + "language_model.model.layers.73.self_attn.q_proj.weight": "model-00043-of-00052.safetensors", + "language_model.model.layers.73.self_attn.v_proj.weight": "model-00043-of-00052.safetensors", + "language_model.model.layers.74.input_layernorm.weight": "model-00044-of-00052.safetensors", + "language_model.model.layers.74.mlp.down_proj.weight": "model-00044-of-00052.safetensors", + "language_model.model.layers.74.mlp.gate_proj.weight": "model-00044-of-00052.safetensors", + "language_model.model.layers.74.mlp.up_proj.weight": "model-00044-of-00052.safetensors", + "language_model.model.layers.74.post_attention_layernorm.weight": "model-00044-of-00052.safetensors", + "language_model.model.layers.74.self_attn.k_proj.weight": "model-00044-of-00052.safetensors", + "language_model.model.layers.74.self_attn.o_proj.weight": "model-00044-of-00052.safetensors", + "language_model.model.layers.74.self_attn.q_proj.weight": "model-00044-of-00052.safetensors", + "language_model.model.layers.74.self_attn.v_proj.weight": "model-00044-of-00052.safetensors", + "language_model.model.layers.75.input_layernorm.weight": "model-00045-of-00052.safetensors", + "language_model.model.layers.75.mlp.down_proj.weight": "model-00045-of-00052.safetensors", + "language_model.model.layers.75.mlp.gate_proj.weight": "model-00044-of-00052.safetensors", + "language_model.model.layers.75.mlp.up_proj.weight": "model-00045-of-00052.safetensors", + "language_model.model.layers.75.post_attention_layernorm.weight": "model-00045-of-00052.safetensors", + "language_model.model.layers.75.self_attn.k_proj.weight": "model-00044-of-00052.safetensors", + "language_model.model.layers.75.self_attn.o_proj.weight": "model-00044-of-00052.safetensors", + "language_model.model.layers.75.self_attn.q_proj.weight": "model-00044-of-00052.safetensors", + "language_model.model.layers.75.self_attn.v_proj.weight": "model-00044-of-00052.safetensors", + "language_model.model.layers.76.input_layernorm.weight": "model-00045-of-00052.safetensors", + "language_model.model.layers.76.mlp.down_proj.weight": "model-00045-of-00052.safetensors", + "language_model.model.layers.76.mlp.gate_proj.weight": "model-00045-of-00052.safetensors", + "language_model.model.layers.76.mlp.up_proj.weight": "model-00045-of-00052.safetensors", + "language_model.model.layers.76.post_attention_layernorm.weight": "model-00045-of-00052.safetensors", + "language_model.model.layers.76.self_attn.k_proj.weight": "model-00045-of-00052.safetensors", + "language_model.model.layers.76.self_attn.o_proj.weight": "model-00045-of-00052.safetensors", + "language_model.model.layers.76.self_attn.q_proj.weight": "model-00045-of-00052.safetensors", + "language_model.model.layers.76.self_attn.v_proj.weight": "model-00045-of-00052.safetensors", + "language_model.model.layers.77.input_layernorm.weight": "model-00046-of-00052.safetensors", + "language_model.model.layers.77.mlp.down_proj.weight": "model-00046-of-00052.safetensors", + "language_model.model.layers.77.mlp.gate_proj.weight": "model-00046-of-00052.safetensors", + "language_model.model.layers.77.mlp.up_proj.weight": "model-00046-of-00052.safetensors", + "language_model.model.layers.77.post_attention_layernorm.weight": "model-00046-of-00052.safetensors", + "language_model.model.layers.77.self_attn.k_proj.weight": "model-00045-of-00052.safetensors", + "language_model.model.layers.77.self_attn.o_proj.weight": "model-00045-of-00052.safetensors", + "language_model.model.layers.77.self_attn.q_proj.weight": "model-00045-of-00052.safetensors", + "language_model.model.layers.77.self_attn.v_proj.weight": "model-00045-of-00052.safetensors", + "language_model.model.layers.78.input_layernorm.weight": "model-00046-of-00052.safetensors", + "language_model.model.layers.78.mlp.down_proj.weight": "model-00046-of-00052.safetensors", + "language_model.model.layers.78.mlp.gate_proj.weight": "model-00046-of-00052.safetensors", + "language_model.model.layers.78.mlp.up_proj.weight": "model-00046-of-00052.safetensors", + "language_model.model.layers.78.post_attention_layernorm.weight": "model-00046-of-00052.safetensors", + "language_model.model.layers.78.self_attn.k_proj.weight": "model-00046-of-00052.safetensors", + "language_model.model.layers.78.self_attn.o_proj.weight": "model-00046-of-00052.safetensors", + "language_model.model.layers.78.self_attn.q_proj.weight": "model-00046-of-00052.safetensors", + "language_model.model.layers.78.self_attn.v_proj.weight": "model-00046-of-00052.safetensors", + "language_model.model.layers.79.input_layernorm.weight": "model-00047-of-00052.safetensors", + "language_model.model.layers.79.mlp.down_proj.weight": "model-00047-of-00052.safetensors", + "language_model.model.layers.79.mlp.gate_proj.weight": "model-00047-of-00052.safetensors", + "language_model.model.layers.79.mlp.up_proj.weight": "model-00047-of-00052.safetensors", + "language_model.model.layers.79.post_attention_layernorm.weight": "model-00047-of-00052.safetensors", + "language_model.model.layers.79.self_attn.k_proj.weight": "model-00047-of-00052.safetensors", + "language_model.model.layers.79.self_attn.o_proj.weight": "model-00047-of-00052.safetensors", + "language_model.model.layers.79.self_attn.q_proj.weight": "model-00047-of-00052.safetensors", + "language_model.model.layers.79.self_attn.v_proj.weight": "model-00047-of-00052.safetensors", + "language_model.model.layers.8.input_layernorm.weight": "model-00006-of-00052.safetensors", + "language_model.model.layers.8.mlp.down_proj.weight": "model-00006-of-00052.safetensors", + "language_model.model.layers.8.mlp.gate_proj.weight": "model-00006-of-00052.safetensors", + "language_model.model.layers.8.mlp.up_proj.weight": "model-00006-of-00052.safetensors", + "language_model.model.layers.8.post_attention_layernorm.weight": "model-00006-of-00052.safetensors", + "language_model.model.layers.8.self_attn.k_proj.weight": "model-00006-of-00052.safetensors", + "language_model.model.layers.8.self_attn.o_proj.weight": "model-00006-of-00052.safetensors", + "language_model.model.layers.8.self_attn.q_proj.weight": "model-00006-of-00052.safetensors", + "language_model.model.layers.8.self_attn.v_proj.weight": "model-00006-of-00052.safetensors", + "language_model.model.layers.80.input_layernorm.weight": "model-00048-of-00052.safetensors", + "language_model.model.layers.80.mlp.down_proj.weight": "model-00048-of-00052.safetensors", + "language_model.model.layers.80.mlp.gate_proj.weight": "model-00047-of-00052.safetensors", + "language_model.model.layers.80.mlp.up_proj.weight": "model-00047-of-00052.safetensors", + "language_model.model.layers.80.post_attention_layernorm.weight": "model-00048-of-00052.safetensors", + "language_model.model.layers.80.self_attn.k_proj.weight": "model-00047-of-00052.safetensors", + "language_model.model.layers.80.self_attn.o_proj.weight": "model-00047-of-00052.safetensors", + "language_model.model.layers.80.self_attn.q_proj.weight": "model-00047-of-00052.safetensors", + "language_model.model.layers.80.self_attn.v_proj.weight": "model-00047-of-00052.safetensors", + "language_model.model.layers.81.input_layernorm.weight": "model-00048-of-00052.safetensors", + "language_model.model.layers.81.mlp.down_proj.weight": "model-00048-of-00052.safetensors", + "language_model.model.layers.81.mlp.gate_proj.weight": "model-00048-of-00052.safetensors", + "language_model.model.layers.81.mlp.up_proj.weight": "model-00048-of-00052.safetensors", + "language_model.model.layers.81.post_attention_layernorm.weight": "model-00048-of-00052.safetensors", + "language_model.model.layers.81.self_attn.k_proj.weight": "model-00048-of-00052.safetensors", + "language_model.model.layers.81.self_attn.o_proj.weight": "model-00048-of-00052.safetensors", + "language_model.model.layers.81.self_attn.q_proj.weight": "model-00048-of-00052.safetensors", + "language_model.model.layers.81.self_attn.v_proj.weight": "model-00048-of-00052.safetensors", + "language_model.model.layers.82.input_layernorm.weight": "model-00049-of-00052.safetensors", + "language_model.model.layers.82.mlp.down_proj.weight": "model-00049-of-00052.safetensors", + "language_model.model.layers.82.mlp.gate_proj.weight": "model-00048-of-00052.safetensors", + "language_model.model.layers.82.mlp.up_proj.weight": "model-00049-of-00052.safetensors", + "language_model.model.layers.82.post_attention_layernorm.weight": "model-00049-of-00052.safetensors", + "language_model.model.layers.82.self_attn.k_proj.weight": "model-00048-of-00052.safetensors", + "language_model.model.layers.82.self_attn.o_proj.weight": "model-00048-of-00052.safetensors", + "language_model.model.layers.82.self_attn.q_proj.weight": "model-00048-of-00052.safetensors", + "language_model.model.layers.82.self_attn.v_proj.weight": "model-00048-of-00052.safetensors", + "language_model.model.layers.83.input_layernorm.weight": "model-00049-of-00052.safetensors", + "language_model.model.layers.83.mlp.down_proj.weight": "model-00049-of-00052.safetensors", + "language_model.model.layers.83.mlp.gate_proj.weight": "model-00049-of-00052.safetensors", + "language_model.model.layers.83.mlp.up_proj.weight": "model-00049-of-00052.safetensors", + "language_model.model.layers.83.post_attention_layernorm.weight": "model-00049-of-00052.safetensors", + "language_model.model.layers.83.self_attn.k_proj.weight": "model-00049-of-00052.safetensors", + "language_model.model.layers.83.self_attn.o_proj.weight": "model-00049-of-00052.safetensors", + "language_model.model.layers.83.self_attn.q_proj.weight": "model-00049-of-00052.safetensors", + "language_model.model.layers.83.self_attn.v_proj.weight": "model-00049-of-00052.safetensors", + "language_model.model.layers.84.input_layernorm.weight": "model-00050-of-00052.safetensors", + "language_model.model.layers.84.mlp.down_proj.weight": "model-00050-of-00052.safetensors", + "language_model.model.layers.84.mlp.gate_proj.weight": "model-00050-of-00052.safetensors", + "language_model.model.layers.84.mlp.up_proj.weight": "model-00050-of-00052.safetensors", + "language_model.model.layers.84.post_attention_layernorm.weight": "model-00050-of-00052.safetensors", + "language_model.model.layers.84.self_attn.k_proj.weight": "model-00049-of-00052.safetensors", + "language_model.model.layers.84.self_attn.o_proj.weight": "model-00049-of-00052.safetensors", + "language_model.model.layers.84.self_attn.q_proj.weight": "model-00049-of-00052.safetensors", + "language_model.model.layers.84.self_attn.v_proj.weight": "model-00049-of-00052.safetensors", + "language_model.model.layers.85.input_layernorm.weight": "model-00050-of-00052.safetensors", + "language_model.model.layers.85.mlp.down_proj.weight": "model-00050-of-00052.safetensors", + "language_model.model.layers.85.mlp.gate_proj.weight": "model-00050-of-00052.safetensors", + "language_model.model.layers.85.mlp.up_proj.weight": "model-00050-of-00052.safetensors", + "language_model.model.layers.85.post_attention_layernorm.weight": "model-00050-of-00052.safetensors", + "language_model.model.layers.85.self_attn.k_proj.weight": "model-00050-of-00052.safetensors", + "language_model.model.layers.85.self_attn.o_proj.weight": "model-00050-of-00052.safetensors", + "language_model.model.layers.85.self_attn.q_proj.weight": "model-00050-of-00052.safetensors", + "language_model.model.layers.85.self_attn.v_proj.weight": "model-00050-of-00052.safetensors", + "language_model.model.layers.86.input_layernorm.weight": "model-00051-of-00052.safetensors", + "language_model.model.layers.86.mlp.down_proj.weight": "model-00051-of-00052.safetensors", + "language_model.model.layers.86.mlp.gate_proj.weight": "model-00051-of-00052.safetensors", + "language_model.model.layers.86.mlp.up_proj.weight": "model-00051-of-00052.safetensors", + "language_model.model.layers.86.post_attention_layernorm.weight": "model-00051-of-00052.safetensors", + "language_model.model.layers.86.self_attn.k_proj.weight": "model-00051-of-00052.safetensors", + "language_model.model.layers.86.self_attn.o_proj.weight": "model-00051-of-00052.safetensors", + "language_model.model.layers.86.self_attn.q_proj.weight": "model-00051-of-00052.safetensors", + "language_model.model.layers.86.self_attn.v_proj.weight": "model-00051-of-00052.safetensors", + "language_model.model.layers.87.input_layernorm.weight": "model-00052-of-00052.safetensors", + "language_model.model.layers.87.mlp.down_proj.weight": "model-00052-of-00052.safetensors", + "language_model.model.layers.87.mlp.gate_proj.weight": "model-00051-of-00052.safetensors", + "language_model.model.layers.87.mlp.up_proj.weight": "model-00051-of-00052.safetensors", + "language_model.model.layers.87.post_attention_layernorm.weight": "model-00052-of-00052.safetensors", + "language_model.model.layers.87.self_attn.k_proj.weight": "model-00051-of-00052.safetensors", + "language_model.model.layers.87.self_attn.o_proj.weight": "model-00051-of-00052.safetensors", + "language_model.model.layers.87.self_attn.q_proj.weight": "model-00051-of-00052.safetensors", + "language_model.model.layers.87.self_attn.v_proj.weight": "model-00051-of-00052.safetensors", + "language_model.model.layers.9.input_layernorm.weight": "model-00007-of-00052.safetensors", + "language_model.model.layers.9.mlp.down_proj.weight": "model-00007-of-00052.safetensors", + "language_model.model.layers.9.mlp.gate_proj.weight": "model-00007-of-00052.safetensors", + "language_model.model.layers.9.mlp.up_proj.weight": "model-00007-of-00052.safetensors", + "language_model.model.layers.9.post_attention_layernorm.weight": "model-00007-of-00052.safetensors", + "language_model.model.layers.9.self_attn.k_proj.weight": "model-00007-of-00052.safetensors", + "language_model.model.layers.9.self_attn.o_proj.weight": "model-00007-of-00052.safetensors", + "language_model.model.layers.9.self_attn.q_proj.weight": "model-00007-of-00052.safetensors", + "language_model.model.layers.9.self_attn.v_proj.weight": "model-00007-of-00052.safetensors", + "language_model.model.norm.weight": "model-00052-of-00052.safetensors", + "multi_modal_projector.linear_1.weight": "model-00001-of-00052.safetensors", + "multi_modal_projector.linear_2.weight": "model-00001-of-00052.safetensors", + "vision_tower.ln_pre.weight": "model-00001-of-00052.safetensors", + "vision_tower.patch_conv.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.0.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.0.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.0.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.0.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.0.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.0.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.0.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.0.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.0.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.1.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.1.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.1.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.1.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.1.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.1.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.1.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.1.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.1.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.10.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.10.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.10.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.10.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.10.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.10.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.10.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.10.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.10.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.11.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.11.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.11.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.11.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.11.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.11.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.11.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.11.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.11.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.12.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.12.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.12.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.12.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.12.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.12.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.12.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.12.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.12.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.13.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.13.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.13.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.13.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.13.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.13.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.13.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.13.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.13.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.14.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.14.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.14.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.14.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.14.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.14.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.14.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.14.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.14.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.15.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.15.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.15.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.15.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.15.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.15.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.15.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.15.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.15.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.16.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.16.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.16.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.16.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.16.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.16.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.16.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.16.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.16.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.17.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.17.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.17.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.17.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.17.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.17.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.17.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.17.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.17.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.18.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.18.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.18.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.18.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.18.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.18.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.18.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.18.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.18.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.19.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.19.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.19.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.19.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.19.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.19.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.19.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.19.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.19.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.2.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.2.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.2.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.2.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.2.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.2.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.2.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.2.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.2.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.20.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.20.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.20.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.20.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.20.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.20.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.20.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.20.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.20.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.21.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.21.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.21.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.21.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.21.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.21.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.21.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.21.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.21.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.22.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.22.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.22.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.22.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.22.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.22.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.22.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.22.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.22.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.23.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.23.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.23.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.23.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.23.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.23.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.23.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.23.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.23.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.24.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.24.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.24.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.24.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.24.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.24.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.24.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.24.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.24.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.25.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.25.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.25.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.25.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.25.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.25.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.25.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.25.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.25.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.26.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.26.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.26.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.26.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.26.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.26.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.26.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.26.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.26.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.27.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.27.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.27.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.27.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.27.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.27.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.27.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.27.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.27.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.28.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.28.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.28.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.28.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.28.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.28.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.28.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.28.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.28.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.29.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.29.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.29.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.29.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.29.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.29.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.29.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.29.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.29.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.3.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.3.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.3.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.3.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.3.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.3.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.3.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.3.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.3.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.30.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.30.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.30.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.30.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.30.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.30.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.30.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.30.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.30.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.31.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.31.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.31.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.31.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.31.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.31.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.31.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.31.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.31.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.32.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.32.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.32.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.32.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.32.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.32.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.32.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.32.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.32.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.33.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.33.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.33.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.33.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.33.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.33.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.33.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.33.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.33.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.34.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.34.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.34.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.34.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.34.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.34.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.34.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.34.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.34.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.35.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.35.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.35.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.35.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.35.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.35.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.35.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.35.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.35.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.36.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.36.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.36.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.36.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.36.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.36.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.36.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.36.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.36.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.37.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.37.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.37.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.37.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.37.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.37.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.37.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.37.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.37.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.38.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.38.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.38.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.38.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.38.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.38.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.38.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.38.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.38.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.39.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.39.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.39.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.39.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.39.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.39.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.39.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.39.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.39.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.4.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.4.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.4.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.4.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.4.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.4.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.4.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.4.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.4.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.5.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.5.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.5.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.5.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.5.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.5.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.5.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.5.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.5.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.6.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.6.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.6.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.6.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.6.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.6.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.6.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.6.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.6.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.7.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.7.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.7.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.7.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.7.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.7.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.7.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.7.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.7.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.8.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.8.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.8.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.8.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.8.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.8.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.8.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.8.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.8.ffn_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.9.attention.k_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.9.attention.o_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.9.attention.q_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.9.attention.v_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.9.attention_norm.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.9.feed_forward.down_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.9.feed_forward.gate_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.9.feed_forward.up_proj.weight": "model-00001-of-00052.safetensors", + "vision_tower.transformer.layers.9.ffn_norm.weight": "model-00001-of-00052.safetensors" + } +}