ausboss commited on
Commit
40a6901
1 Parent(s): 311dfde

Upload with huggingface_hub

Browse files
.ipynb_checkpoints/README-checkpoint.md ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Merge of [huggyllama/llama-30b](https://huggingface.co/huggyllama/llama-30b) + [kaiokendev/SuperCOT-LoRA](https://huggingface.co/kaiokendev/SuperCOT-LoRA/edit/main/README.md)
2
+
3
+ Supercot was trained to work with langchain prompting.
4
+
5
+
6
+ Load up locally in my custom LLM notebook that uses the Oobabooga modules to load up models: https://github.com/ausboss/Local-LLM-Langchain
7
+
8
+ Then you can add cells from of these other notebooks for testing: https://github.com/gkamradt/langchain-tutorials
9
+
10
+
11
+
12
+ # From Koikendev Lora page
13
+
14
+ ### Compatibility
15
+ This LoRA is compatible with any 7B, 13B or 30B 4-bit quantized LLaMa model, including ggml quantized converted bins
16
+
17
+ ### Prompting
18
+ You should prompt the LoRA the same way you would prompt Alpaca or Alpacino:
19
+
20
+ ```
21
+ Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
22
+
23
+ ### Instruction:
24
+ <instruction>
25
+
26
+ ### Input:
27
+ <any additional context. Remove this if it's not neccesary>
28
+
29
+ ### Response:
30
+ <make sure to leave a single new-line here for optimal results>
31
+ ```
32
+
33
+ Remember that with lower parameter sizes, the structure of the prompt becomes more important. The same prompt worded differently can give wildly different answers. Consider using the following suggestion suffixes to improve output quality:
34
+
35
+ - "Think through this step by step"
36
+ - "Let's think about this logically"
37
+ - "Explain your reasoning"
38
+ - "Provide details to support your answer"
39
+ - "Compare and contrast your answer with alternatives"
40
+
41
+ ### Coming Soon
42
+ - Tweet fix for 13B and 7B - lower model sizes seem to be extremely sensitive to hashtags at the end of training data responses, especially at longer cutoffs
.ipynb_checkpoints/config-checkpoint.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "huggyllama/llama-30b",
3
+ "architectures": [
4
+ "LlamaForCausalLM"
5
+ ],
6
+ "bos_token_id": 1,
7
+ "eos_token_id": 2,
8
+ "hidden_act": "silu",
9
+ "hidden_size": 6656,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 17920,
12
+ "max_position_embeddings": 2048,
13
+ "max_sequence_length": 2048,
14
+ "model_type": "llama",
15
+ "num_attention_heads": 52,
16
+ "num_hidden_layers": 60,
17
+ "pad_token_id": 0,
18
+ "rms_norm_eps": 1e-06,
19
+ "tie_word_embeddings": false,
20
+ "torch_dtype": "float16",
21
+ "transformers_version": "4.28.0",
22
+ "use_cache": true,
23
+ "vocab_size": 32000
24
+ }
.ipynb_checkpoints/generation_config-checkpoint.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "pad_token_id": 0,
6
+ "transformers_version": "4.28.0"
7
+ }
.ipynb_checkpoints/huggingface-metadata-checkpoint.txt ADDED
@@ -0,0 +1,248 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ url: https://huggingface.co/ausboss/llama-30b-supercot
2
+ branch: main
3
+ download date: 2023-05-24 03:23:45
4
+ sha256sum:
5
+ 09cf82059eaabd225a99bf38743c21347bd8f1afd60407fba21db30022e8275b pytorch_model-00001-of-00243.bin
6
+ 9d058a7dc32ebfcbd3b1aa1b0cf98b8cb236c09530e06c8a1a34bfa51f940fac pytorch_model-00002-of-00243.bin
7
+ 483ff136784e1446a1d47152ad369275b63d97ac454d45eb31bd0383de2b3be5 pytorch_model-00003-of-00243.bin
8
+ b5dc5c69cbd281b33441630fb502e371026415c49cc71cebdb065ebcac7b6eba pytorch_model-00004-of-00243.bin
9
+ 38f703f231fca57a9d743fdea91bb4e052c0f0644927ec00905662a80665601c pytorch_model-00005-of-00243.bin
10
+ 2aa57f5546e994b9259ebf08767063be4ed2512fb7e9fe98d719d3e027ebb8f4 pytorch_model-00006-of-00243.bin
11
+ ee40d76e5c51b5ecc279a1397afa50983b274c1bd0b145c7ae33af872a1e28d4 pytorch_model-00007-of-00243.bin
12
+ 518c4fdaa89fa9876c3cae8ada56b7b96b0e230cb7af370308a869e4ede07395 pytorch_model-00008-of-00243.bin
13
+ a381bb81a90c9671940886f1afa1f1738d740bdf74f79b9849dbd1b99be616b0 pytorch_model-00009-of-00243.bin
14
+ 0d725d67acabd07d3ca1c2d7508068d9ea763391658d35c4db3f5df4a78fd328 pytorch_model-00010-of-00243.bin
15
+ 885fa2668e40d8c75ff4f0506f1594db76a5260bc643db5c33bda2b03c3fa640 pytorch_model-00011-of-00243.bin
16
+ 602302d9a2e2f6e9aacc9c5593720b0366fb836a41d117f856ef936606f4c9a7 pytorch_model-00012-of-00243.bin
17
+ 3f1fe1988509adfb8d81b2ffa2ad7061264bcbde838f58c21f05e17c63cccc01 pytorch_model-00013-of-00243.bin
18
+ bb699ddfd979f80d32cd8d11ac342bd598bb50a1e6d4172fb741a0e056680218 pytorch_model-00014-of-00243.bin
19
+ 488e49dc2810cb0c8403b2891c6f81835aa6c87a3112ec83ec67368ea1740ccf pytorch_model-00015-of-00243.bin
20
+ 615e7dd2bb66158e3079cd0fa16c4a0519c3119ab967a509763db50df62890ad pytorch_model-00016-of-00243.bin
21
+ ffeb25c38f3da6737a18042ea6ec758aadf1c08fa3d10f11122b956c6f62c52e pytorch_model-00017-of-00243.bin
22
+ 899de81f2149db059c3ea2f44d9ab95e2b1632e966e22b3750b0dcb17c4f87bc pytorch_model-00018-of-00243.bin
23
+ e6319f1a3e67b5d2936933d58362962708d12f0cd796cb2ded83fcbaef502371 pytorch_model-00019-of-00243.bin
24
+ f1d0aa33e387ecdca5428e36c9be1109db674853b2871625cec6b9b6a5af4ab8 pytorch_model-00020-of-00243.bin
25
+ 5b3ae784451facf7bc237364bb005335f3be65a266529782af9b4a114cfb2837 pytorch_model-00021-of-00243.bin
26
+ 78726f86efadb520399f36d6bdc00388a8555ec8a9b0c7450b92e335236c4079 pytorch_model-00022-of-00243.bin
27
+ f4f1e3de05871c160c943ca38c26f3018a9fcd9afd1fe5d6b96c7196f5451539 pytorch_model-00023-of-00243.bin
28
+ 58385fda4dd1ddff817078bcc304548cf1ee60dd75e23f823720c271d2b84ab2 pytorch_model-00024-of-00243.bin
29
+ 35e7d0c3acc694ef071ffb46ce4e28a6109b0f65a3203d9ac52d3e8d4de2b8a1 pytorch_model-00025-of-00243.bin
30
+ f707916b73bf3e093baad418cf9030694e7b0692d0463d7df7d612d5dbae1824 pytorch_model-00026-of-00243.bin
31
+ 272a5bd29c3067980f30ddea2b1d2299bc805b7cf2932cd60b08c26ca48620c2 pytorch_model-00027-of-00243.bin
32
+ 3b2228ebf36f2b157154cf541ad45a95e12411d94ec8d96130816b7f77d7c4b5 pytorch_model-00028-of-00243.bin
33
+ d0f173c3eae5b18afa589d9f418f607ed250846830eb6360882e6de4ec7a6a9c pytorch_model-00029-of-00243.bin
34
+ 5ed6a1d34e7309b0019c74f84d4fa1f013cbb0136388b5446997899d053daad9 pytorch_model-00030-of-00243.bin
35
+ 83da5b0f1d756afca6beb4abd3b6ef6d54b97729d7b731e79634149eea752f1f pytorch_model-00031-of-00243.bin
36
+ 904fa86c833bae7b562e7d6cabecbbf2f7652d9b77f5dfe63ffb1102879eca79 pytorch_model-00032-of-00243.bin
37
+ 3ac46349de58557f72a99408f64d40475be9d762e1b92c89639b9216f374a812 pytorch_model-00033-of-00243.bin
38
+ a784cd0657d7f3d059da7953faf4ad031c912db84408e62e6354f3c5ac055009 pytorch_model-00034-of-00243.bin
39
+ af219ff22435d8331db4cce62e95dc2aff7f3d6a6f94e2412f5bd770d914d31d pytorch_model-00035-of-00243.bin
40
+ 64f205f529359c3c7aa894afd2f92c978c189adad15621c95fdeb589c8888610 pytorch_model-00036-of-00243.bin
41
+ ec644a8bff8bdb6cbb90c13baf23bc4cfdf4b0c6cd061fc8a301248150764251 pytorch_model-00037-of-00243.bin
42
+ a0654f007170e8667204132d34414670686b13be76ff5d7e5580c4cb48e63c54 pytorch_model-00038-of-00243.bin
43
+ 7ccbf439ef076c3ae403d668dd0660c7bdd3b0e9e2b8f94e3586a63ff2b89902 pytorch_model-00039-of-00243.bin
44
+ 75f0735e581714ce2c4e7435dd649e3765156b5394b7bc1e6efd6cf5379c3a79 pytorch_model-00040-of-00243.bin
45
+ f0bc1eb8f14cc3c336665c6afe3c6f64e39960d92453db2902f1cf0813f79a11 pytorch_model-00041-of-00243.bin
46
+ 3602e6f55f5d831429539fd0b0998d7cdcb68c337d288fa498b6cde5b8c52074 pytorch_model-00042-of-00243.bin
47
+ 74d9d42bec0d46c04ae048068d9d9a0283a7287a39f9d73178e66768d37b226d pytorch_model-00043-of-00243.bin
48
+ 48dd4dfd621872b4028e4ec70a3eaa768df2586865264349d70dfefb6df6d37e pytorch_model-00044-of-00243.bin
49
+ 8adb35cb6efc3d91f53211034abd47c6a8562e9f37b85aeaad9b28511efd78fd pytorch_model-00045-of-00243.bin
50
+ 7f4dd4b88af1d669fc20d55f1682d21728a1b68b075894436fc6f6f6007e5b4f pytorch_model-00046-of-00243.bin
51
+ d6a78ac519544c67c784acbdd670f68d0b7b45d9be0cf3accbb210b0505303de pytorch_model-00047-of-00243.bin
52
+ 69d34eaaa8a48b0bbd14887b4732deaeb9038cb56299fa30dbdd66bf7cf0a004 pytorch_model-00048-of-00243.bin
53
+ 94c49dcde09e908596928aae6839828ac087a5ad5ce9871fd24ec6889ecd53e7 pytorch_model-00049-of-00243.bin
54
+ 0a406bdc68423623f681091482727715007a692d7740ebe0e4b29468e2062b6d pytorch_model-00050-of-00243.bin
55
+ b485d818f5f7bd134e3df0382ea9c7c6e372e0cadba716a91d144d090058fe62 pytorch_model-00051-of-00243.bin
56
+ 6434295e65d4ef96baa8af71f44b44a99b47095b988de1bd48c7e8ba8a3c80be pytorch_model-00052-of-00243.bin
57
+ 5ac66042226aee62e71e1240c095bb49a6f50b935b9797c1d36d27911e38a21f pytorch_model-00053-of-00243.bin
58
+ e589a0b358fcba4f5e60f6a04212e2b8641b489d1584310caf15375d2426c60d pytorch_model-00054-of-00243.bin
59
+ 336e7c9aa81429b8bd6722857c1e442cc24449c64922e19c11ec300ab52780bb pytorch_model-00055-of-00243.bin
60
+ 87c4c538088b0ef6c542c3b837d0981156078453f2115bf3a2c39ae7cd37f998 pytorch_model-00056-of-00243.bin
61
+ 2d29fd39708d7cfd7d59995d0430c2cdb25f2f7ac8dd3d31f950b963d9806aa3 pytorch_model-00057-of-00243.bin
62
+ 10c9abdf748e2aee9a3f9c56789769225ec86cc4fd00c02604a3bfb6116f7acf pytorch_model-00058-of-00243.bin
63
+ 28f4ccf6a88a36e6e99514978f57cce70359cc32df66dde4091addc12ed6fd30 pytorch_model-00059-of-00243.bin
64
+ f76d024c8d574fde700a89c3e4b18fe83d2ee9c95716035b7310129cd47f5302 pytorch_model-00060-of-00243.bin
65
+ 2918b4ab52ef985a42e29e4435bbe2f128b17c5901ca4068fb3ed34cb5555d9e pytorch_model-00061-of-00243.bin
66
+ ab7bfcb3aef1e67477d6f1f68e397e90b4d5c3a82903f1a13a0cccf369dcccef pytorch_model-00062-of-00243.bin
67
+ e944d9ecc8cfdfb0dce84f78f8707006a73f24fb8fa76a9c230123425831a184 pytorch_model-00063-of-00243.bin
68
+ ee20c0ec398ceab9597d104513b90336e5f4c672cbde5a3b2ece936c8a3f5148 pytorch_model-00064-of-00243.bin
69
+ 6869b1ed0d6a2daf25939ac8ff1c696589fd62e47c114516e5726369c8e562b2 pytorch_model-00065-of-00243.bin
70
+ 29f5ead347af64a517cb43c02140b8cca9e56831c6997709ceb92209585c9426 pytorch_model-00066-of-00243.bin
71
+ f3521ab2911ac86f3a8d791bd700a4cb70425f53f2639ce5d166dbf380968112 pytorch_model-00067-of-00243.bin
72
+ 3e25fb6e55e1a627fa1bcb76c5371b8ce9b0bd233fd408a70fbc00dbb8d727c0 pytorch_model-00068-of-00243.bin
73
+ 3e9c59b9ea19e75fe7de7ada5e41dc8c38bc95b0925eb1f41cd493d024ae6899 pytorch_model-00069-of-00243.bin
74
+ 0aceb706fe626a8f26a52535ba9f164735aa2b9018d6f71266f07f5102bde48f pytorch_model-00070-of-00243.bin
75
+ 43a5999497f7aee7be2f3943fb259bb283a457bbab83d08833bbc4ca55ea44d9 pytorch_model-00071-of-00243.bin
76
+ 341304b4d0a6d8d616b6527e9a27940ff6fdff744e9d5d33648222fd788245fa pytorch_model-00072-of-00243.bin
77
+ ebf99e3cc61828079936a9c5b6ece50d6ca76b34f5b8e7b5544eb95771068198 pytorch_model-00073-of-00243.bin
78
+ afa4171a1f12702ae7ff0cdc59cae74d3211c2110a1fb4cedc9b67d5915e4519 pytorch_model-00074-of-00243.bin
79
+ a6980b5aa243a7c2cbd55ae561221272c32c3670f34964933f004934939eed5e pytorch_model-00075-of-00243.bin
80
+ c61bdce8f9252a08ecc3b4af42ee699b943ff7fe7086ca5f5c8e363ebda18d08 pytorch_model-00076-of-00243.bin
81
+ 82139e62b584a952ef1610e251f78c9f6175d70891b7ef3240c89ca3175df5b7 pytorch_model-00077-of-00243.bin
82
+ 0f8e26f8d0eb89d64609fe3846c3fc2b904704ce82b7b67037e4f105ea4b00e1 pytorch_model-00078-of-00243.bin
83
+ 43a4343a8010a2e773b82a054a4b37669690a4b361d6a29eee62aa184279a22c pytorch_model-00079-of-00243.bin
84
+ 85fc194be080bc29eace87a662f14afea9376a10b936daa89bdbc03162bfca89 pytorch_model-00080-of-00243.bin
85
+ 016d9d5873f09d06760dabce4acb167a7115d2487fdb09a19dd0f51e195c755c pytorch_model-00081-of-00243.bin
86
+ fa4b253a4d70d0b6e9619fe1ca4194a64c6cdb1cc6e30e79b47a6d84068bdf52 pytorch_model-00082-of-00243.bin
87
+ f401abb6ebbeeaeee501f58ef182a2f0df87303e0eb3e1ef61891a564f4fe4af pytorch_model-00083-of-00243.bin
88
+ 5565664a2f03217978813660e9ca63925e5abae9beef71b2d10772f436cdcb4e pytorch_model-00084-of-00243.bin
89
+ 70bd36a3f6a7f45aae6e9b4b56c5889701533915d10a6b8020c93cb3ed13fb36 pytorch_model-00085-of-00243.bin
90
+ f0d8592e9dbadd34d090b9c309d8d2781695ecf4e6d7005addd25b13630dd611 pytorch_model-00086-of-00243.bin
91
+ 533d20debbd4f0429f12f99ea5ce6a5b88e286af7c58604555ee7bef6c38b286 pytorch_model-00087-of-00243.bin
92
+ 511177391ac994c00b1e39c424f8843f794c0348a9b56418c7c1584bc30d29f4 pytorch_model-00088-of-00243.bin
93
+ 4f4c446298218fd821f5b5b9796b46650badf697fd81fa9d5af3cfa017ea2ede pytorch_model-00089-of-00243.bin
94
+ 69935ce1a0ae99cafd3318adb16cd46005c691214d66b9c1ba2654f04bb6ee48 pytorch_model-00090-of-00243.bin
95
+ f8f6e703981fe6c81d09909ca4730cd50855563fc9c67804923791a19f53fe13 pytorch_model-00091-of-00243.bin
96
+ f4f6c8d6d43d0758840be87cf81ca1936a6ec5e7db567d7e6e28be1482ea3990 pytorch_model-00092-of-00243.bin
97
+ e2f13f781f5bc19ff0774b0d7d37af9ae162c372ec7b57cc6f01768349206f74 pytorch_model-00093-of-00243.bin
98
+ b640b907f0977649c55b5f620cefaeae0e2ab3ffa9d34004a3eaee5e5ad163e3 pytorch_model-00094-of-00243.bin
99
+ 7028c4e1472979988f960b51b7a8872720c12e4ae77bdfc5ed2992448a497419 pytorch_model-00095-of-00243.bin
100
+ 6c2c69a678f85a987e12eca058b60d9712f0c134fe9daf9ba45de5f8ecbf905a pytorch_model-00096-of-00243.bin
101
+ 30747c23a557c14dca4766adedd1555529d1409045f8e8af5d8ba2912b500356 pytorch_model-00097-of-00243.bin
102
+ 97e97a502d9354ecbc2731f93bec4b1d766636fc7b89f2be9749db719ac07400 pytorch_model-00098-of-00243.bin
103
+ 0b7f79b28aeaddb8edcedf9ed249720c05977f13a0f7c27fe410403c0412fc0a pytorch_model-00099-of-00243.bin
104
+ 08ebae4bfebe60c18b235bc0f6e4ee6b1659df839f1c9ddbcfad5e1398d2e923 pytorch_model-00100-of-00243.bin
105
+ 8c8adb82ce7bef686b2cf6b64a207750249d9f725bed0b6b78249cbdd8911e48 pytorch_model-00101-of-00243.bin
106
+ ba24db819cf91d0f2ebc5b1ac07208cbaa71893e1cb6d8f7e0e7543173535946 pytorch_model-00102-of-00243.bin
107
+ 74af29ea6dc24e0a1936ea080f3b70582912e0450c767a30c5360e3c62f143b9 pytorch_model-00103-of-00243.bin
108
+ fe068509fbae8dcfaffd178f8b830a7b3c3ddc839afff0974db4c62a701ff332 pytorch_model-00104-of-00243.bin
109
+ ce9f24ae56b070092e99cc119a5900b1195ad20f6a9a459a3781775f66d2266d pytorch_model-00105-of-00243.bin
110
+ 77e7ae120c4c15a43394dcc82bdcba5fe75c844c39144d6917904111bc7cef6b pytorch_model-00106-of-00243.bin
111
+ dd149a6139ec2067097aebf058b4fd9b7030a49dddb8a66e323fc91000196bd6 pytorch_model-00107-of-00243.bin
112
+ 15bc0b43afa0c818e80835fbf1e8c4213f63c0eeed1f91ed2a0c66bf9e142470 pytorch_model-00108-of-00243.bin
113
+ 09e753f0642a1c904ef732b342894f3bd4d3f68f46989495c477eced1debde94 pytorch_model-00109-of-00243.bin
114
+ f3cf460b6c86e2a35fd60e3715cf9756f54c3db4bb4abc2f07c08b79002b4e19 pytorch_model-00110-of-00243.bin
115
+ 0a353b27752a908fdc4374d041cbb5a6c912a771698d87b0f3c17874b8e488c1 pytorch_model-00111-of-00243.bin
116
+ 63778cc7a8bc989390fd94d668c436b7ad69c86ff2f69d22f73a40bd965d3e1c pytorch_model-00112-of-00243.bin
117
+ 11bc10f9e50374ecd18dec38dc6d0e666fb7548249bb2ec5ed67ff01778c2170 pytorch_model-00113-of-00243.bin
118
+ ca5885767c24be36b32cea538f90d53e05f8450224b0bb788aadd5a6dd89bd34 pytorch_model-00114-of-00243.bin
119
+ 961620e8c9efb910851816c1dd259cb4030f529597ebc830ce539597a384b623 pytorch_model-00115-of-00243.bin
120
+ 51ee0ac8161ccc17facabf1958ed412db37f8f5b0a190f28fe5aa16a23ade55d pytorch_model-00116-of-00243.bin
121
+ ad0f45f599f266e34bfaa8639b1d531744a2d841305805474c45729f0d9b777e pytorch_model-00117-of-00243.bin
122
+ 58ac3d11f22bf981e59e856c6d78c505b0edff247e88323039a697d8d4ad57a4 pytorch_model-00118-of-00243.bin
123
+ 1f5ab60a4639fe55130a4e8c2909c7b2e1411d8f2cbd2b00ec8341f5c9a98f06 pytorch_model-00119-of-00243.bin
124
+ 94b4074f17ba08cfe46399d073badc2457d7ca0e3583ab690074de03ccbb62aa pytorch_model-00120-of-00243.bin
125
+ 0a48a967e3edb8b500eaa12947a32f32f26a24b67e60d3d402afd7481df81516 pytorch_model-00121-of-00243.bin
126
+ fd9eb815daa468dd8be675c7248f2ff4542bbc53984562280181a01ab5a62708 pytorch_model-00122-of-00243.bin
127
+ fffe12cc8e0346cfc8a8230359b8ba7927294408e132218731971d90d5737b41 pytorch_model-00123-of-00243.bin
128
+ 987380de8f157fb20e2e80aba57a8e27716e3729a736decc3758cd0d4dc73e21 pytorch_model-00124-of-00243.bin
129
+ 6ba2b9b8a71af19c38436d7340a58fff7978575fef2deb9f89b0adb9163b2977 pytorch_model-00125-of-00243.bin
130
+ 232f12b40cb60e6c2fb6e55bcba8a855286c16de51d1c7b0fe8c483678c04ff3 pytorch_model-00126-of-00243.bin
131
+ c57a3f4efb3da5660563061253fa71662e6d93e17934588eb3303691a2cb3bf3 pytorch_model-00127-of-00243.bin
132
+ 76b14de090b4bef2c78b6ea383f18f7b6ff5dbd9dccecc55fbdfc90d71d719f3 pytorch_model-00128-of-00243.bin
133
+ 02ff1662e0386bb56929643ff59ef8f5434277e65c5ed78fe6c78cfa68c561b0 pytorch_model-00129-of-00243.bin
134
+ a3e78ee0aa271f0f79b956344faca8e4f161383eb557bd1e8a5c48f26f13c5e6 pytorch_model-00130-of-00243.bin
135
+ 01f54fd5816fc61838ec064b5cafc4c25eaa3b87937ede268b8785c426ccdd7d pytorch_model-00131-of-00243.bin
136
+ 4ecaab016076677b83190110e275271ec4a53b3d736d945170bb1376098701ff pytorch_model-00132-of-00243.bin
137
+ 73d18ed728d97ace5cdd89d9d0c2c76e359ceed3ad05fa03f3038b35d481db27 pytorch_model-00133-of-00243.bin
138
+ 8b8de0770ce14282f96b36f09868aa1b8f1396f2ea2cc683e81d9b7dbdce3799 pytorch_model-00134-of-00243.bin
139
+ 0c2e0f85e72a5e3ae5194a91edfcd141b9ebeb8654cc28769b29e79faaa7f7fe pytorch_model-00135-of-00243.bin
140
+ 33e2206373661164a1c13008c67e2cdce7967368ebd0b2f20377ccda74723656 pytorch_model-00136-of-00243.bin
141
+ 48d5ab98d950b69cb2807c93f97029b2019011cd7bbbc08c9f08d3390ef6f205 pytorch_model-00137-of-00243.bin
142
+ 6435c742d6fdcbbd382d3905e46406d5af16d63d48d323301828ef8a26210fc3 pytorch_model-00138-of-00243.bin
143
+ c4c2593e0dcf66f315079300056c2a3f4b89c0909274bbc520e5f23cb34ad7a3 pytorch_model-00139-of-00243.bin
144
+ da00a68a0900c492dece277a26a9e12fae77bb1edfe32603f12c9be5ca85217a pytorch_model-00140-of-00243.bin
145
+ 35d379ec3521bdfdb1789ad36ea80173053fcc3728c608be8760cea5805a2be2 pytorch_model-00141-of-00243.bin
146
+ 4aba717fa9caa634aee268a2bfef4efc615a363545c23dbb997d12388b12a82d pytorch_model-00142-of-00243.bin
147
+ b51a06c89c8a475845c018beb6520db6ffb4a717b2bb5cb376f4e912dde124c1 pytorch_model-00143-of-00243.bin
148
+ ea46a5b3336bc4eb812bc1bafe1f40371585b40f5632113ca396b1064c38dd30 pytorch_model-00144-of-00243.bin
149
+ b346ac3f89c61c6083439c3757cedec730b672c18a35c679e08547ab873f4ecb pytorch_model-00145-of-00243.bin
150
+ cdf36f5445079764746a016d559d4700352e6a4e4d252d19b9d3e4b9f3552b14 pytorch_model-00146-of-00243.bin
151
+ 6ab71611a456425ec1dee8a49f31621edb18701d876531c63ec7f19032ebaab9 pytorch_model-00147-of-00243.bin
152
+ 530f68272df243bb3970dad56606c4f657706dfece505a7f97c4987d94c8691c pytorch_model-00148-of-00243.bin
153
+ 3b9671d920c9506367ff83f6aa8e20c9cc3bfc955d6235e8291f200144ce1d2e pytorch_model-00149-of-00243.bin
154
+ cb244fe146c2e37c1a69cb2b30bf11da20082b3cad8b195e000f224764425b9a pytorch_model-00150-of-00243.bin
155
+ 4f0bbcaabbc664e59d3a96572d0d746c2c5017c089d80ab1d6cb66047f86313e pytorch_model-00151-of-00243.bin
156
+ 0f07bf8e97696a087a4e03db68a497ee7f8e44b8fe285717372b80e966ef6908 pytorch_model-00152-of-00243.bin
157
+ 66ab23c030464f4b6885d54eee5189819f0399b7ce1d65cba2872f945c444c0b pytorch_model-00153-of-00243.bin
158
+ d10391a3fce4e7ef21a5fd041f3553b9b2b5e35c273dae79915bbec065fef613 pytorch_model-00154-of-00243.bin
159
+ 30484b1ffc896bfa63bcd3525931821e7b984e3ceaae3a42c64d13ff92349a4e pytorch_model-00155-of-00243.bin
160
+ b3fd3b87151dfe5ab73598e6e48ae3331c6108e1ebbad5eecade29a6bbb314fc pytorch_model-00156-of-00243.bin
161
+ 6a03cf30b4a4ad61a6385c6d9a169643edbfaa949459684c61f446ae361239f3 pytorch_model-00157-of-00243.bin
162
+ 311fa925f07530e5cebcde5059bdb845db52b7f8da00623f48dc7da705a6a2df pytorch_model-00158-of-00243.bin
163
+ 77545e68e432bacfb0bec0bc9476ad3375cfc1827da8efa9b353f9fbc758a8b8 pytorch_model-00159-of-00243.bin
164
+ 74adb9c4554adc71b6eedc3b5a8b141b3613c1177ef0adb5bfaeeaa42bbd85a9 pytorch_model-00160-of-00243.bin
165
+ 01b5812179799a61884ee85985ef8d8ecd00037eda2fad672dfdc2878b18a46d pytorch_model-00161-of-00243.bin
166
+ 9650fadc97c002bbcdec1154bdd422fe12d18c61b9441ed74a1b68f545c2788b pytorch_model-00162-of-00243.bin
167
+ 788094824dd1823f531e8cec250979556a72d22ba9d1182dbfcd3db429214b89 pytorch_model-00163-of-00243.bin
168
+ d755e0f69d15678a2b361212ab3d1780bf975da6d33360a5d3ed96389a02f039 pytorch_model-00164-of-00243.bin
169
+ 5694a4c2b11ed581cdaea73cef531c26c987eedd59c221be052e8f3e443e8cfe pytorch_model-00165-of-00243.bin
170
+ 6b0ca82e0a531a4a8f03b9ed6c5fa0c850ec1b3fd5d5f54a2c85f0e553355326 pytorch_model-00166-of-00243.bin
171
+ bb9afe8715554aa01b606a0b23f231b25c2762c99a2e59ca503d0e1d8b4f6a8a pytorch_model-00167-of-00243.bin
172
+ 1c131a096d7d7363f475e6367480f5665c87111b6c58ba925474ce2ffc123799 pytorch_model-00168-of-00243.bin
173
+ eeaf8e409e220b251a5d9e3fac2735c13668e5da016d7c317677d4f74fe12b00 pytorch_model-00169-of-00243.bin
174
+ 08c43e5cc128bfbbf6dd8f30e6d5bcc616c2054f341062e562927acb99922bf1 pytorch_model-00170-of-00243.bin
175
+ 8f5b277bca1c0e05c0bd7601f4d3301d6856ff4fde62daa17a81707f60928ec5 pytorch_model-00171-of-00243.bin
176
+ 30999caa7ced67cad5a441e3719f4ae4435a2c5591e17b08c0c2f07a18c5df5d pytorch_model-00172-of-00243.bin
177
+ dab1edecb2176f68468d683afe638b840f970120f9610a831c4bb5966b2f491a pytorch_model-00173-of-00243.bin
178
+ c82cee7ac9ecd9f6f071c244f4cf6e6e8459f8d4bb22fa38c00087195add4e6c pytorch_model-00174-of-00243.bin
179
+ 37213a8bb345cd94e10963669f683a8c1a2378b6ff8374d863f2a09b2b0e05a1 pytorch_model-00175-of-00243.bin
180
+ 7e80625d461a8818933861646a94e7a0d29c9ce532debea258caafb10dc4c09f pytorch_model-00176-of-00243.bin
181
+ 4cd85828447cce382ec197771c5a4417adb37262c49c60c4440262e3c3fb45dc pytorch_model-00177-of-00243.bin
182
+ 8c3622803909e984166c781aca5033d9fc751b76873552cdc21ec6101a371fca pytorch_model-00178-of-00243.bin
183
+ f7c185f4628d8ec71522b8ba2d2478d4d65d8a1de2b0f0854974929594a0cf23 pytorch_model-00179-of-00243.bin
184
+ a6269af0e3be281ad0b2e77fc064eadf56ff0ea50f2a1297ad93db9a76b202af pytorch_model-00180-of-00243.bin
185
+ b1a21ae3f3d8883734b3e8ec37eac5cc79c16a13c96e8ab945548356d0d3e447 pytorch_model-00181-of-00243.bin
186
+ 65dfebebecac857d45e49f881b44fe52f38283da648c67b68cf8f34aaa9d7388 pytorch_model-00182-of-00243.bin
187
+ 24dfe8116221ae9cb90ab393471d34d30f84daff0ca1632c3fc705d6c8355ba0 pytorch_model-00183-of-00243.bin
188
+ a02d1a5570ec7831e5c56a4e8fa9813d98415dd406f4c2fd48a3c885bdb3187b pytorch_model-00184-of-00243.bin
189
+ 03340f30578794690dc58b97d17879f04219034f5466c4a45e3951f857c79fa1 pytorch_model-00185-of-00243.bin
190
+ 7f5bfcc28404d85f0dffcc3ef0b014585307ade9010a004b509772b8266ec74f pytorch_model-00186-of-00243.bin
191
+ b3844cf4044bb1ffb02d8e9b7b1b7fd4b4c124b010c04c1f8ff8a78e03ff320f pytorch_model-00187-of-00243.bin
192
+ 4089b0dbd91b0f2b10e7dd5fcbca4dde9faffafd661695c2da5beee8370072a3 pytorch_model-00188-of-00243.bin
193
+ 1666768c16bfc76a11fedc3525358b9e429ae50c115b4f706960d65910715543 pytorch_model-00189-of-00243.bin
194
+ 1affffa5fccbcc2166f666eeed7aced8d38743574ea7b6201fcd149272bb6246 pytorch_model-00190-of-00243.bin
195
+ ad41ac209d56068a6078429b6b18135176a9ca4ec4c24568094e1eb9ebce0f60 pytorch_model-00191-of-00243.bin
196
+ 7ca1cab9541d33f8cbb692b5f319698c44205da210e3cdc673c9c0a2b7104b43 pytorch_model-00192-of-00243.bin
197
+ 46d2a9a508115d14c4b4f9580a99f2b83b24222b6da6836d293f35885b610875 pytorch_model-00193-of-00243.bin
198
+ 174844ba09a4a7471bba31188e01397be6c40e050ea685eaaa6bae22f2bb11c4 pytorch_model-00194-of-00243.bin
199
+ 00f948d55679a438369d81791520688569c94a1b3621cb10d4454638fbcbe4db pytorch_model-00195-of-00243.bin
200
+ 2faff3cfe1c5d530d6665a4063be2f7d26e05f0c6904cbbd7d091905d973ff79 pytorch_model-00196-of-00243.bin
201
+ 895b6d403638393612d7d0f5cb963c80631c153f03c3d4966869057630c6f94e pytorch_model-00197-of-00243.bin
202
+ 7dc6f3b6c479c1e775a810637d02051c8e448a8c4f6a9cd26788242dc6e515ac pytorch_model-00198-of-00243.bin
203
+ e0700fd25fe1be375cf7df82194120c17a154ad5c57ebfb9975f42db0aef5f48 pytorch_model-00199-of-00243.bin
204
+ 43ea69f26fa3fc99f36a789e84022ff32a61ba3be9dac4b92e007b7361d843ca pytorch_model-00200-of-00243.bin
205
+ f97ffd772f8a328990d30289216d6028e6f544790af0cb81afd60205484b1c46 pytorch_model-00201-of-00243.bin
206
+ e1591dea37d51f02161ec7296a0787cac06d73decf4709a2ec02363ed937ddd9 pytorch_model-00202-of-00243.bin
207
+ ca9e94ee2efbcc51fe478068340599fca381ec4eaa40892e095ace61509e5bde pytorch_model-00203-of-00243.bin
208
+ 37564c62aa89a31a666c3ee10dfb69bb88d6b07c2a99c0b8c16cd56913d0faf4 pytorch_model-00204-of-00243.bin
209
+ 214824198bc777c3b026b7a164678aa7ed34a597db6fd033edaa94358395f8ae pytorch_model-00205-of-00243.bin
210
+ cee45752533144a37844d8eda345b613dae3f94d9da2f98f768dc3065f157a87 pytorch_model-00206-of-00243.bin
211
+ c9721ed5865c998b36f622f5a2c5b135fb7cd5d5e989f0ad8b6d2e9a36bce3b5 pytorch_model-00207-of-00243.bin
212
+ fd1323c856b7c584a0877890cd80cbddd449f4071a1180de51ac7441df33cfaf pytorch_model-00208-of-00243.bin
213
+ b67108c6c45e1c0426a2e890f6e15f63e5fbc9dc03b033f613b42bf7d86279ff pytorch_model-00209-of-00243.bin
214
+ 9a3486126b67295c5533a653e354ff25b5324b8339f9d815f7ebfbca791f2378 pytorch_model-00210-of-00243.bin
215
+ 5b624bbf42939baf3ba4d7ba25672ae363e7932d8df18b411321ea9dbab3d6c4 pytorch_model-00211-of-00243.bin
216
+ 5005a92e9e6ccac697122bc5b8b9a0a21b4ca55e97fbe4018cfdc5fb5e869a06 pytorch_model-00212-of-00243.bin
217
+ ff9154478a5fd9a3b2cf5266a302497da01a538b28fbb62181d45c92d2c6c2ae pytorch_model-00213-of-00243.bin
218
+ 4659295ce30f7ddb2c9673cf7d86280fa74869b1a08352f38406c1fbe56f51f5 pytorch_model-00214-of-00243.bin
219
+ 794d1149608c2bb5b0007749e2474fdf7e61c150693310a1918163dbd4babfad pytorch_model-00215-of-00243.bin
220
+ 8132b955c65d85dd5d15662ec17484143f683b78ca0af405a24725fd9393baeb pytorch_model-00216-of-00243.bin
221
+ 640c7e2e8dca0350525ab5a51870f309b9b0203e72884be1034ef2ce74481e6e pytorch_model-00217-of-00243.bin
222
+ e7007217a8100f86b12894c49bac1106af5e565b03e1741b277dd02a64374fb0 pytorch_model-00218-of-00243.bin
223
+ c62bd6c10e27bf8d8caf92f3f13b26d92487f66a3e62f4af9e49c7eba63167b8 pytorch_model-00219-of-00243.bin
224
+ 1e657f439b4a2db41cdaeadcdb88f5eb48a8cd726ac460da9cdefd0502b3eb9b pytorch_model-00220-of-00243.bin
225
+ de028b5f2a19b85c5be2d0654ae2337516220ec27669ffc8ef17c43104bb3224 pytorch_model-00221-of-00243.bin
226
+ 3ad8fc90c1fb9bb43cfe5b2401bfb7b290ce8c704bbc5515f1ce4a5873f95139 pytorch_model-00222-of-00243.bin
227
+ 9b989a86fa6dcf0436cc974981f6d3c8811132e92a26166a25ea457c72b3db06 pytorch_model-00223-of-00243.bin
228
+ aca6597466a65388286be22891e5474fd1e79eda6b2b1b3f968163a7b1a4393d pytorch_model-00224-of-00243.bin
229
+ fd0f867516832221d24705969d3593e5a1811ee761b8fbf7c57bcc1ad1dfc8a5 pytorch_model-00225-of-00243.bin
230
+ 6794576954ec8970ff09499a7c0de486532326826f850bbdbf2e3fa815f69e1e pytorch_model-00226-of-00243.bin
231
+ 4f7cef542e67021df4ff812429ea0491ef967122d94afa3cdd83b41181fa3c75 pytorch_model-00227-of-00243.bin
232
+ fee6c54c08d1efabc58eaa318651a625199cea878e49dd9257ec8d493009a779 pytorch_model-00228-of-00243.bin
233
+ ae514ec50b760014c719a16b9509b5cb3b0d9f82dfd4d489cc3d3484e8d6a9a7 pytorch_model-00229-of-00243.bin
234
+ 6d4b7fb5ec49a09cafc0977126e0d22a7acd1f029dff96efe130c665373619f6 pytorch_model-00230-of-00243.bin
235
+ fcfa6ba6f24e9ae498769fd4a3e20c2f183ea290640de2f205a378d50c44c457 pytorch_model-00231-of-00243.bin
236
+ 9f1994823e8a45ccdd4030a3abbbed28026a08fa5908ccacf7cb0f9976ad8976 pytorch_model-00232-of-00243.bin
237
+ cf64ed719e75ab225838749b8885e00c23b280a1c750b9f695aac00ad89152aa pytorch_model-00233-of-00243.bin
238
+ befc340950c3750e9fb79be5f7a3ba7ecb5bcc5422bd327e9309fb440dbbb5be pytorch_model-00234-of-00243.bin
239
+ abc57ae3f55a7ee8fb3fac3e36531b1e577fb4916a6b1c3fa88d7ce9969aed5f pytorch_model-00235-of-00243.bin
240
+ 29bb9ec1e0db7a8d1dba82cfc430a6d8a6f1620dd0006a36de3a8f90cb943bac pytorch_model-00236-of-00243.bin
241
+ 14d20d28ee07dbaba393a8d82aa17683bc31d988fb4f22d261336dd5096ae71c pytorch_model-00237-of-00243.bin
242
+ 58df1115168fe8d60dffd81f7e0a2d5b0997601ca73f75cfab9ba128a3a3d496 pytorch_model-00238-of-00243.bin
243
+ 7f42e2a869414c1cba14c3bb0b3e72fabc06a8a77144dac233c366c1883073b8 pytorch_model-00239-of-00243.bin
244
+ a4faa71e384023a2e1e21609bf5b6967925df8f2b64d825183fad2fb2418985a pytorch_model-00240-of-00243.bin
245
+ 530296bfc60095c9f4840776a6ee24558efddcf19bd1a7ebe42895c02d7e912c pytorch_model-00241-of-00243.bin
246
+ 8bc75fb2c6486f123c89701022ed29145c1246fe109a3defa5d470b35074a988 pytorch_model-00242-of-00243.bin
247
+ 39f587e500e18a9db4797a87945ef84b9dc7bb989506f7d322144b270918358c pytorch_model-00243-of-00243.bin
248
+ 9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347 tokenizer.model
.ipynb_checkpoints/pytorch_model.bin.index-checkpoint.json ADDED
@@ -0,0 +1,610 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 65057902592
4
+ },
5
+ "weight_map": {
6
+ "lm_head.weight": "pytorch_model-00243-of-00243.bin",
7
+ "model.embed_tokens.weight": "pytorch_model-00002-of-00243.bin",
8
+ "model.layers.0.input_layernorm.weight": "pytorch_model-00006-of-00243.bin",
9
+ "model.layers.0.mlp.down_proj.weight": "pytorch_model-00005-of-00243.bin",
10
+ "model.layers.0.mlp.gate_proj.weight": "pytorch_model-00004-of-00243.bin",
11
+ "model.layers.0.mlp.up_proj.weight": "pytorch_model-00006-of-00243.bin",
12
+ "model.layers.0.post_attention_layernorm.weight": "pytorch_model-00006-of-00243.bin",
13
+ "model.layers.0.self_attn.k_proj.weight": "pytorch_model-00003-of-00243.bin",
14
+ "model.layers.0.self_attn.o_proj.weight": "pytorch_model-00003-of-00243.bin",
15
+ "model.layers.0.self_attn.q_proj.weight": "pytorch_model-00003-of-00243.bin",
16
+ "model.layers.0.self_attn.rotary_emb.inv_freq": "pytorch_model-00003-of-00243.bin",
17
+ "model.layers.0.self_attn.v_proj.weight": "pytorch_model-00003-of-00243.bin",
18
+ "model.layers.1.input_layernorm.weight": "pytorch_model-00010-of-00243.bin",
19
+ "model.layers.1.mlp.down_proj.weight": "pytorch_model-00009-of-00243.bin",
20
+ "model.layers.1.mlp.gate_proj.weight": "pytorch_model-00008-of-00243.bin",
21
+ "model.layers.1.mlp.up_proj.weight": "pytorch_model-00010-of-00243.bin",
22
+ "model.layers.1.post_attention_layernorm.weight": "pytorch_model-00010-of-00243.bin",
23
+ "model.layers.1.self_attn.k_proj.weight": "pytorch_model-00007-of-00243.bin",
24
+ "model.layers.1.self_attn.o_proj.weight": "pytorch_model-00007-of-00243.bin",
25
+ "model.layers.1.self_attn.q_proj.weight": "pytorch_model-00006-of-00243.bin",
26
+ "model.layers.1.self_attn.rotary_emb.inv_freq": "pytorch_model-00007-of-00243.bin",
27
+ "model.layers.1.self_attn.v_proj.weight": "pytorch_model-00007-of-00243.bin",
28
+ "model.layers.10.input_layernorm.weight": "pytorch_model-00046-of-00243.bin",
29
+ "model.layers.10.mlp.down_proj.weight": "pytorch_model-00045-of-00243.bin",
30
+ "model.layers.10.mlp.gate_proj.weight": "pytorch_model-00044-of-00243.bin",
31
+ "model.layers.10.mlp.up_proj.weight": "pytorch_model-00046-of-00243.bin",
32
+ "model.layers.10.post_attention_layernorm.weight": "pytorch_model-00046-of-00243.bin",
33
+ "model.layers.10.self_attn.k_proj.weight": "pytorch_model-00043-of-00243.bin",
34
+ "model.layers.10.self_attn.o_proj.weight": "pytorch_model-00043-of-00243.bin",
35
+ "model.layers.10.self_attn.q_proj.weight": "pytorch_model-00042-of-00243.bin",
36
+ "model.layers.10.self_attn.rotary_emb.inv_freq": "pytorch_model-00043-of-00243.bin",
37
+ "model.layers.10.self_attn.v_proj.weight": "pytorch_model-00043-of-00243.bin",
38
+ "model.layers.11.input_layernorm.weight": "pytorch_model-00050-of-00243.bin",
39
+ "model.layers.11.mlp.down_proj.weight": "pytorch_model-00049-of-00243.bin",
40
+ "model.layers.11.mlp.gate_proj.weight": "pytorch_model-00048-of-00243.bin",
41
+ "model.layers.11.mlp.up_proj.weight": "pytorch_model-00050-of-00243.bin",
42
+ "model.layers.11.post_attention_layernorm.weight": "pytorch_model-00050-of-00243.bin",
43
+ "model.layers.11.self_attn.k_proj.weight": "pytorch_model-00047-of-00243.bin",
44
+ "model.layers.11.self_attn.o_proj.weight": "pytorch_model-00047-of-00243.bin",
45
+ "model.layers.11.self_attn.q_proj.weight": "pytorch_model-00046-of-00243.bin",
46
+ "model.layers.11.self_attn.rotary_emb.inv_freq": "pytorch_model-00047-of-00243.bin",
47
+ "model.layers.11.self_attn.v_proj.weight": "pytorch_model-00047-of-00243.bin",
48
+ "model.layers.12.input_layernorm.weight": "pytorch_model-00054-of-00243.bin",
49
+ "model.layers.12.mlp.down_proj.weight": "pytorch_model-00053-of-00243.bin",
50
+ "model.layers.12.mlp.gate_proj.weight": "pytorch_model-00052-of-00243.bin",
51
+ "model.layers.12.mlp.up_proj.weight": "pytorch_model-00054-of-00243.bin",
52
+ "model.layers.12.post_attention_layernorm.weight": "pytorch_model-00054-of-00243.bin",
53
+ "model.layers.12.self_attn.k_proj.weight": "pytorch_model-00051-of-00243.bin",
54
+ "model.layers.12.self_attn.o_proj.weight": "pytorch_model-00051-of-00243.bin",
55
+ "model.layers.12.self_attn.q_proj.weight": "pytorch_model-00050-of-00243.bin",
56
+ "model.layers.12.self_attn.rotary_emb.inv_freq": "pytorch_model-00051-of-00243.bin",
57
+ "model.layers.12.self_attn.v_proj.weight": "pytorch_model-00051-of-00243.bin",
58
+ "model.layers.13.input_layernorm.weight": "pytorch_model-00058-of-00243.bin",
59
+ "model.layers.13.mlp.down_proj.weight": "pytorch_model-00057-of-00243.bin",
60
+ "model.layers.13.mlp.gate_proj.weight": "pytorch_model-00056-of-00243.bin",
61
+ "model.layers.13.mlp.up_proj.weight": "pytorch_model-00058-of-00243.bin",
62
+ "model.layers.13.post_attention_layernorm.weight": "pytorch_model-00058-of-00243.bin",
63
+ "model.layers.13.self_attn.k_proj.weight": "pytorch_model-00055-of-00243.bin",
64
+ "model.layers.13.self_attn.o_proj.weight": "pytorch_model-00055-of-00243.bin",
65
+ "model.layers.13.self_attn.q_proj.weight": "pytorch_model-00054-of-00243.bin",
66
+ "model.layers.13.self_attn.rotary_emb.inv_freq": "pytorch_model-00055-of-00243.bin",
67
+ "model.layers.13.self_attn.v_proj.weight": "pytorch_model-00055-of-00243.bin",
68
+ "model.layers.14.input_layernorm.weight": "pytorch_model-00062-of-00243.bin",
69
+ "model.layers.14.mlp.down_proj.weight": "pytorch_model-00061-of-00243.bin",
70
+ "model.layers.14.mlp.gate_proj.weight": "pytorch_model-00060-of-00243.bin",
71
+ "model.layers.14.mlp.up_proj.weight": "pytorch_model-00062-of-00243.bin",
72
+ "model.layers.14.post_attention_layernorm.weight": "pytorch_model-00062-of-00243.bin",
73
+ "model.layers.14.self_attn.k_proj.weight": "pytorch_model-00059-of-00243.bin",
74
+ "model.layers.14.self_attn.o_proj.weight": "pytorch_model-00059-of-00243.bin",
75
+ "model.layers.14.self_attn.q_proj.weight": "pytorch_model-00058-of-00243.bin",
76
+ "model.layers.14.self_attn.rotary_emb.inv_freq": "pytorch_model-00059-of-00243.bin",
77
+ "model.layers.14.self_attn.v_proj.weight": "pytorch_model-00059-of-00243.bin",
78
+ "model.layers.15.input_layernorm.weight": "pytorch_model-00066-of-00243.bin",
79
+ "model.layers.15.mlp.down_proj.weight": "pytorch_model-00065-of-00243.bin",
80
+ "model.layers.15.mlp.gate_proj.weight": "pytorch_model-00064-of-00243.bin",
81
+ "model.layers.15.mlp.up_proj.weight": "pytorch_model-00066-of-00243.bin",
82
+ "model.layers.15.post_attention_layernorm.weight": "pytorch_model-00066-of-00243.bin",
83
+ "model.layers.15.self_attn.k_proj.weight": "pytorch_model-00063-of-00243.bin",
84
+ "model.layers.15.self_attn.o_proj.weight": "pytorch_model-00063-of-00243.bin",
85
+ "model.layers.15.self_attn.q_proj.weight": "pytorch_model-00062-of-00243.bin",
86
+ "model.layers.15.self_attn.rotary_emb.inv_freq": "pytorch_model-00063-of-00243.bin",
87
+ "model.layers.15.self_attn.v_proj.weight": "pytorch_model-00063-of-00243.bin",
88
+ "model.layers.16.input_layernorm.weight": "pytorch_model-00070-of-00243.bin",
89
+ "model.layers.16.mlp.down_proj.weight": "pytorch_model-00069-of-00243.bin",
90
+ "model.layers.16.mlp.gate_proj.weight": "pytorch_model-00068-of-00243.bin",
91
+ "model.layers.16.mlp.up_proj.weight": "pytorch_model-00070-of-00243.bin",
92
+ "model.layers.16.post_attention_layernorm.weight": "pytorch_model-00070-of-00243.bin",
93
+ "model.layers.16.self_attn.k_proj.weight": "pytorch_model-00067-of-00243.bin",
94
+ "model.layers.16.self_attn.o_proj.weight": "pytorch_model-00067-of-00243.bin",
95
+ "model.layers.16.self_attn.q_proj.weight": "pytorch_model-00066-of-00243.bin",
96
+ "model.layers.16.self_attn.rotary_emb.inv_freq": "pytorch_model-00067-of-00243.bin",
97
+ "model.layers.16.self_attn.v_proj.weight": "pytorch_model-00067-of-00243.bin",
98
+ "model.layers.17.input_layernorm.weight": "pytorch_model-00074-of-00243.bin",
99
+ "model.layers.17.mlp.down_proj.weight": "pytorch_model-00073-of-00243.bin",
100
+ "model.layers.17.mlp.gate_proj.weight": "pytorch_model-00072-of-00243.bin",
101
+ "model.layers.17.mlp.up_proj.weight": "pytorch_model-00074-of-00243.bin",
102
+ "model.layers.17.post_attention_layernorm.weight": "pytorch_model-00074-of-00243.bin",
103
+ "model.layers.17.self_attn.k_proj.weight": "pytorch_model-00071-of-00243.bin",
104
+ "model.layers.17.self_attn.o_proj.weight": "pytorch_model-00071-of-00243.bin",
105
+ "model.layers.17.self_attn.q_proj.weight": "pytorch_model-00070-of-00243.bin",
106
+ "model.layers.17.self_attn.rotary_emb.inv_freq": "pytorch_model-00071-of-00243.bin",
107
+ "model.layers.17.self_attn.v_proj.weight": "pytorch_model-00071-of-00243.bin",
108
+ "model.layers.18.input_layernorm.weight": "pytorch_model-00078-of-00243.bin",
109
+ "model.layers.18.mlp.down_proj.weight": "pytorch_model-00077-of-00243.bin",
110
+ "model.layers.18.mlp.gate_proj.weight": "pytorch_model-00076-of-00243.bin",
111
+ "model.layers.18.mlp.up_proj.weight": "pytorch_model-00078-of-00243.bin",
112
+ "model.layers.18.post_attention_layernorm.weight": "pytorch_model-00078-of-00243.bin",
113
+ "model.layers.18.self_attn.k_proj.weight": "pytorch_model-00075-of-00243.bin",
114
+ "model.layers.18.self_attn.o_proj.weight": "pytorch_model-00075-of-00243.bin",
115
+ "model.layers.18.self_attn.q_proj.weight": "pytorch_model-00074-of-00243.bin",
116
+ "model.layers.18.self_attn.rotary_emb.inv_freq": "pytorch_model-00075-of-00243.bin",
117
+ "model.layers.18.self_attn.v_proj.weight": "pytorch_model-00075-of-00243.bin",
118
+ "model.layers.19.input_layernorm.weight": "pytorch_model-00082-of-00243.bin",
119
+ "model.layers.19.mlp.down_proj.weight": "pytorch_model-00081-of-00243.bin",
120
+ "model.layers.19.mlp.gate_proj.weight": "pytorch_model-00080-of-00243.bin",
121
+ "model.layers.19.mlp.up_proj.weight": "pytorch_model-00082-of-00243.bin",
122
+ "model.layers.19.post_attention_layernorm.weight": "pytorch_model-00082-of-00243.bin",
123
+ "model.layers.19.self_attn.k_proj.weight": "pytorch_model-00079-of-00243.bin",
124
+ "model.layers.19.self_attn.o_proj.weight": "pytorch_model-00079-of-00243.bin",
125
+ "model.layers.19.self_attn.q_proj.weight": "pytorch_model-00078-of-00243.bin",
126
+ "model.layers.19.self_attn.rotary_emb.inv_freq": "pytorch_model-00079-of-00243.bin",
127
+ "model.layers.19.self_attn.v_proj.weight": "pytorch_model-00079-of-00243.bin",
128
+ "model.layers.2.input_layernorm.weight": "pytorch_model-00014-of-00243.bin",
129
+ "model.layers.2.mlp.down_proj.weight": "pytorch_model-00013-of-00243.bin",
130
+ "model.layers.2.mlp.gate_proj.weight": "pytorch_model-00012-of-00243.bin",
131
+ "model.layers.2.mlp.up_proj.weight": "pytorch_model-00014-of-00243.bin",
132
+ "model.layers.2.post_attention_layernorm.weight": "pytorch_model-00014-of-00243.bin",
133
+ "model.layers.2.self_attn.k_proj.weight": "pytorch_model-00011-of-00243.bin",
134
+ "model.layers.2.self_attn.o_proj.weight": "pytorch_model-00011-of-00243.bin",
135
+ "model.layers.2.self_attn.q_proj.weight": "pytorch_model-00010-of-00243.bin",
136
+ "model.layers.2.self_attn.rotary_emb.inv_freq": "pytorch_model-00011-of-00243.bin",
137
+ "model.layers.2.self_attn.v_proj.weight": "pytorch_model-00011-of-00243.bin",
138
+ "model.layers.20.input_layernorm.weight": "pytorch_model-00086-of-00243.bin",
139
+ "model.layers.20.mlp.down_proj.weight": "pytorch_model-00085-of-00243.bin",
140
+ "model.layers.20.mlp.gate_proj.weight": "pytorch_model-00084-of-00243.bin",
141
+ "model.layers.20.mlp.up_proj.weight": "pytorch_model-00086-of-00243.bin",
142
+ "model.layers.20.post_attention_layernorm.weight": "pytorch_model-00086-of-00243.bin",
143
+ "model.layers.20.self_attn.k_proj.weight": "pytorch_model-00083-of-00243.bin",
144
+ "model.layers.20.self_attn.o_proj.weight": "pytorch_model-00083-of-00243.bin",
145
+ "model.layers.20.self_attn.q_proj.weight": "pytorch_model-00082-of-00243.bin",
146
+ "model.layers.20.self_attn.rotary_emb.inv_freq": "pytorch_model-00083-of-00243.bin",
147
+ "model.layers.20.self_attn.v_proj.weight": "pytorch_model-00083-of-00243.bin",
148
+ "model.layers.21.input_layernorm.weight": "pytorch_model-00090-of-00243.bin",
149
+ "model.layers.21.mlp.down_proj.weight": "pytorch_model-00089-of-00243.bin",
150
+ "model.layers.21.mlp.gate_proj.weight": "pytorch_model-00088-of-00243.bin",
151
+ "model.layers.21.mlp.up_proj.weight": "pytorch_model-00090-of-00243.bin",
152
+ "model.layers.21.post_attention_layernorm.weight": "pytorch_model-00090-of-00243.bin",
153
+ "model.layers.21.self_attn.k_proj.weight": "pytorch_model-00087-of-00243.bin",
154
+ "model.layers.21.self_attn.o_proj.weight": "pytorch_model-00087-of-00243.bin",
155
+ "model.layers.21.self_attn.q_proj.weight": "pytorch_model-00086-of-00243.bin",
156
+ "model.layers.21.self_attn.rotary_emb.inv_freq": "pytorch_model-00087-of-00243.bin",
157
+ "model.layers.21.self_attn.v_proj.weight": "pytorch_model-00087-of-00243.bin",
158
+ "model.layers.22.input_layernorm.weight": "pytorch_model-00094-of-00243.bin",
159
+ "model.layers.22.mlp.down_proj.weight": "pytorch_model-00093-of-00243.bin",
160
+ "model.layers.22.mlp.gate_proj.weight": "pytorch_model-00092-of-00243.bin",
161
+ "model.layers.22.mlp.up_proj.weight": "pytorch_model-00094-of-00243.bin",
162
+ "model.layers.22.post_attention_layernorm.weight": "pytorch_model-00094-of-00243.bin",
163
+ "model.layers.22.self_attn.k_proj.weight": "pytorch_model-00091-of-00243.bin",
164
+ "model.layers.22.self_attn.o_proj.weight": "pytorch_model-00091-of-00243.bin",
165
+ "model.layers.22.self_attn.q_proj.weight": "pytorch_model-00090-of-00243.bin",
166
+ "model.layers.22.self_attn.rotary_emb.inv_freq": "pytorch_model-00091-of-00243.bin",
167
+ "model.layers.22.self_attn.v_proj.weight": "pytorch_model-00091-of-00243.bin",
168
+ "model.layers.23.input_layernorm.weight": "pytorch_model-00098-of-00243.bin",
169
+ "model.layers.23.mlp.down_proj.weight": "pytorch_model-00097-of-00243.bin",
170
+ "model.layers.23.mlp.gate_proj.weight": "pytorch_model-00096-of-00243.bin",
171
+ "model.layers.23.mlp.up_proj.weight": "pytorch_model-00098-of-00243.bin",
172
+ "model.layers.23.post_attention_layernorm.weight": "pytorch_model-00098-of-00243.bin",
173
+ "model.layers.23.self_attn.k_proj.weight": "pytorch_model-00095-of-00243.bin",
174
+ "model.layers.23.self_attn.o_proj.weight": "pytorch_model-00095-of-00243.bin",
175
+ "model.layers.23.self_attn.q_proj.weight": "pytorch_model-00094-of-00243.bin",
176
+ "model.layers.23.self_attn.rotary_emb.inv_freq": "pytorch_model-00095-of-00243.bin",
177
+ "model.layers.23.self_attn.v_proj.weight": "pytorch_model-00095-of-00243.bin",
178
+ "model.layers.24.input_layernorm.weight": "pytorch_model-00102-of-00243.bin",
179
+ "model.layers.24.mlp.down_proj.weight": "pytorch_model-00101-of-00243.bin",
180
+ "model.layers.24.mlp.gate_proj.weight": "pytorch_model-00100-of-00243.bin",
181
+ "model.layers.24.mlp.up_proj.weight": "pytorch_model-00102-of-00243.bin",
182
+ "model.layers.24.post_attention_layernorm.weight": "pytorch_model-00102-of-00243.bin",
183
+ "model.layers.24.self_attn.k_proj.weight": "pytorch_model-00099-of-00243.bin",
184
+ "model.layers.24.self_attn.o_proj.weight": "pytorch_model-00099-of-00243.bin",
185
+ "model.layers.24.self_attn.q_proj.weight": "pytorch_model-00098-of-00243.bin",
186
+ "model.layers.24.self_attn.rotary_emb.inv_freq": "pytorch_model-00099-of-00243.bin",
187
+ "model.layers.24.self_attn.v_proj.weight": "pytorch_model-00099-of-00243.bin",
188
+ "model.layers.25.input_layernorm.weight": "pytorch_model-00106-of-00243.bin",
189
+ "model.layers.25.mlp.down_proj.weight": "pytorch_model-00105-of-00243.bin",
190
+ "model.layers.25.mlp.gate_proj.weight": "pytorch_model-00104-of-00243.bin",
191
+ "model.layers.25.mlp.up_proj.weight": "pytorch_model-00106-of-00243.bin",
192
+ "model.layers.25.post_attention_layernorm.weight": "pytorch_model-00106-of-00243.bin",
193
+ "model.layers.25.self_attn.k_proj.weight": "pytorch_model-00103-of-00243.bin",
194
+ "model.layers.25.self_attn.o_proj.weight": "pytorch_model-00103-of-00243.bin",
195
+ "model.layers.25.self_attn.q_proj.weight": "pytorch_model-00102-of-00243.bin",
196
+ "model.layers.25.self_attn.rotary_emb.inv_freq": "pytorch_model-00103-of-00243.bin",
197
+ "model.layers.25.self_attn.v_proj.weight": "pytorch_model-00103-of-00243.bin",
198
+ "model.layers.26.input_layernorm.weight": "pytorch_model-00110-of-00243.bin",
199
+ "model.layers.26.mlp.down_proj.weight": "pytorch_model-00109-of-00243.bin",
200
+ "model.layers.26.mlp.gate_proj.weight": "pytorch_model-00108-of-00243.bin",
201
+ "model.layers.26.mlp.up_proj.weight": "pytorch_model-00110-of-00243.bin",
202
+ "model.layers.26.post_attention_layernorm.weight": "pytorch_model-00110-of-00243.bin",
203
+ "model.layers.26.self_attn.k_proj.weight": "pytorch_model-00107-of-00243.bin",
204
+ "model.layers.26.self_attn.o_proj.weight": "pytorch_model-00107-of-00243.bin",
205
+ "model.layers.26.self_attn.q_proj.weight": "pytorch_model-00106-of-00243.bin",
206
+ "model.layers.26.self_attn.rotary_emb.inv_freq": "pytorch_model-00107-of-00243.bin",
207
+ "model.layers.26.self_attn.v_proj.weight": "pytorch_model-00107-of-00243.bin",
208
+ "model.layers.27.input_layernorm.weight": "pytorch_model-00114-of-00243.bin",
209
+ "model.layers.27.mlp.down_proj.weight": "pytorch_model-00113-of-00243.bin",
210
+ "model.layers.27.mlp.gate_proj.weight": "pytorch_model-00112-of-00243.bin",
211
+ "model.layers.27.mlp.up_proj.weight": "pytorch_model-00114-of-00243.bin",
212
+ "model.layers.27.post_attention_layernorm.weight": "pytorch_model-00114-of-00243.bin",
213
+ "model.layers.27.self_attn.k_proj.weight": "pytorch_model-00111-of-00243.bin",
214
+ "model.layers.27.self_attn.o_proj.weight": "pytorch_model-00111-of-00243.bin",
215
+ "model.layers.27.self_attn.q_proj.weight": "pytorch_model-00110-of-00243.bin",
216
+ "model.layers.27.self_attn.rotary_emb.inv_freq": "pytorch_model-00111-of-00243.bin",
217
+ "model.layers.27.self_attn.v_proj.weight": "pytorch_model-00111-of-00243.bin",
218
+ "model.layers.28.input_layernorm.weight": "pytorch_model-00118-of-00243.bin",
219
+ "model.layers.28.mlp.down_proj.weight": "pytorch_model-00117-of-00243.bin",
220
+ "model.layers.28.mlp.gate_proj.weight": "pytorch_model-00116-of-00243.bin",
221
+ "model.layers.28.mlp.up_proj.weight": "pytorch_model-00118-of-00243.bin",
222
+ "model.layers.28.post_attention_layernorm.weight": "pytorch_model-00118-of-00243.bin",
223
+ "model.layers.28.self_attn.k_proj.weight": "pytorch_model-00115-of-00243.bin",
224
+ "model.layers.28.self_attn.o_proj.weight": "pytorch_model-00115-of-00243.bin",
225
+ "model.layers.28.self_attn.q_proj.weight": "pytorch_model-00114-of-00243.bin",
226
+ "model.layers.28.self_attn.rotary_emb.inv_freq": "pytorch_model-00115-of-00243.bin",
227
+ "model.layers.28.self_attn.v_proj.weight": "pytorch_model-00115-of-00243.bin",
228
+ "model.layers.29.input_layernorm.weight": "pytorch_model-00122-of-00243.bin",
229
+ "model.layers.29.mlp.down_proj.weight": "pytorch_model-00121-of-00243.bin",
230
+ "model.layers.29.mlp.gate_proj.weight": "pytorch_model-00120-of-00243.bin",
231
+ "model.layers.29.mlp.up_proj.weight": "pytorch_model-00122-of-00243.bin",
232
+ "model.layers.29.post_attention_layernorm.weight": "pytorch_model-00122-of-00243.bin",
233
+ "model.layers.29.self_attn.k_proj.weight": "pytorch_model-00119-of-00243.bin",
234
+ "model.layers.29.self_attn.o_proj.weight": "pytorch_model-00119-of-00243.bin",
235
+ "model.layers.29.self_attn.q_proj.weight": "pytorch_model-00118-of-00243.bin",
236
+ "model.layers.29.self_attn.rotary_emb.inv_freq": "pytorch_model-00119-of-00243.bin",
237
+ "model.layers.29.self_attn.v_proj.weight": "pytorch_model-00119-of-00243.bin",
238
+ "model.layers.3.input_layernorm.weight": "pytorch_model-00018-of-00243.bin",
239
+ "model.layers.3.mlp.down_proj.weight": "pytorch_model-00017-of-00243.bin",
240
+ "model.layers.3.mlp.gate_proj.weight": "pytorch_model-00016-of-00243.bin",
241
+ "model.layers.3.mlp.up_proj.weight": "pytorch_model-00018-of-00243.bin",
242
+ "model.layers.3.post_attention_layernorm.weight": "pytorch_model-00018-of-00243.bin",
243
+ "model.layers.3.self_attn.k_proj.weight": "pytorch_model-00015-of-00243.bin",
244
+ "model.layers.3.self_attn.o_proj.weight": "pytorch_model-00015-of-00243.bin",
245
+ "model.layers.3.self_attn.q_proj.weight": "pytorch_model-00014-of-00243.bin",
246
+ "model.layers.3.self_attn.rotary_emb.inv_freq": "pytorch_model-00015-of-00243.bin",
247
+ "model.layers.3.self_attn.v_proj.weight": "pytorch_model-00015-of-00243.bin",
248
+ "model.layers.30.input_layernorm.weight": "pytorch_model-00126-of-00243.bin",
249
+ "model.layers.30.mlp.down_proj.weight": "pytorch_model-00125-of-00243.bin",
250
+ "model.layers.30.mlp.gate_proj.weight": "pytorch_model-00124-of-00243.bin",
251
+ "model.layers.30.mlp.up_proj.weight": "pytorch_model-00126-of-00243.bin",
252
+ "model.layers.30.post_attention_layernorm.weight": "pytorch_model-00126-of-00243.bin",
253
+ "model.layers.30.self_attn.k_proj.weight": "pytorch_model-00123-of-00243.bin",
254
+ "model.layers.30.self_attn.o_proj.weight": "pytorch_model-00123-of-00243.bin",
255
+ "model.layers.30.self_attn.q_proj.weight": "pytorch_model-00122-of-00243.bin",
256
+ "model.layers.30.self_attn.rotary_emb.inv_freq": "pytorch_model-00123-of-00243.bin",
257
+ "model.layers.30.self_attn.v_proj.weight": "pytorch_model-00123-of-00243.bin",
258
+ "model.layers.31.input_layernorm.weight": "pytorch_model-00130-of-00243.bin",
259
+ "model.layers.31.mlp.down_proj.weight": "pytorch_model-00129-of-00243.bin",
260
+ "model.layers.31.mlp.gate_proj.weight": "pytorch_model-00128-of-00243.bin",
261
+ "model.layers.31.mlp.up_proj.weight": "pytorch_model-00130-of-00243.bin",
262
+ "model.layers.31.post_attention_layernorm.weight": "pytorch_model-00130-of-00243.bin",
263
+ "model.layers.31.self_attn.k_proj.weight": "pytorch_model-00127-of-00243.bin",
264
+ "model.layers.31.self_attn.o_proj.weight": "pytorch_model-00127-of-00243.bin",
265
+ "model.layers.31.self_attn.q_proj.weight": "pytorch_model-00126-of-00243.bin",
266
+ "model.layers.31.self_attn.rotary_emb.inv_freq": "pytorch_model-00127-of-00243.bin",
267
+ "model.layers.31.self_attn.v_proj.weight": "pytorch_model-00127-of-00243.bin",
268
+ "model.layers.32.input_layernorm.weight": "pytorch_model-00134-of-00243.bin",
269
+ "model.layers.32.mlp.down_proj.weight": "pytorch_model-00133-of-00243.bin",
270
+ "model.layers.32.mlp.gate_proj.weight": "pytorch_model-00132-of-00243.bin",
271
+ "model.layers.32.mlp.up_proj.weight": "pytorch_model-00134-of-00243.bin",
272
+ "model.layers.32.post_attention_layernorm.weight": "pytorch_model-00134-of-00243.bin",
273
+ "model.layers.32.self_attn.k_proj.weight": "pytorch_model-00131-of-00243.bin",
274
+ "model.layers.32.self_attn.o_proj.weight": "pytorch_model-00131-of-00243.bin",
275
+ "model.layers.32.self_attn.q_proj.weight": "pytorch_model-00130-of-00243.bin",
276
+ "model.layers.32.self_attn.rotary_emb.inv_freq": "pytorch_model-00131-of-00243.bin",
277
+ "model.layers.32.self_attn.v_proj.weight": "pytorch_model-00131-of-00243.bin",
278
+ "model.layers.33.input_layernorm.weight": "pytorch_model-00138-of-00243.bin",
279
+ "model.layers.33.mlp.down_proj.weight": "pytorch_model-00137-of-00243.bin",
280
+ "model.layers.33.mlp.gate_proj.weight": "pytorch_model-00136-of-00243.bin",
281
+ "model.layers.33.mlp.up_proj.weight": "pytorch_model-00138-of-00243.bin",
282
+ "model.layers.33.post_attention_layernorm.weight": "pytorch_model-00138-of-00243.bin",
283
+ "model.layers.33.self_attn.k_proj.weight": "pytorch_model-00135-of-00243.bin",
284
+ "model.layers.33.self_attn.o_proj.weight": "pytorch_model-00135-of-00243.bin",
285
+ "model.layers.33.self_attn.q_proj.weight": "pytorch_model-00134-of-00243.bin",
286
+ "model.layers.33.self_attn.rotary_emb.inv_freq": "pytorch_model-00135-of-00243.bin",
287
+ "model.layers.33.self_attn.v_proj.weight": "pytorch_model-00135-of-00243.bin",
288
+ "model.layers.34.input_layernorm.weight": "pytorch_model-00142-of-00243.bin",
289
+ "model.layers.34.mlp.down_proj.weight": "pytorch_model-00141-of-00243.bin",
290
+ "model.layers.34.mlp.gate_proj.weight": "pytorch_model-00140-of-00243.bin",
291
+ "model.layers.34.mlp.up_proj.weight": "pytorch_model-00142-of-00243.bin",
292
+ "model.layers.34.post_attention_layernorm.weight": "pytorch_model-00142-of-00243.bin",
293
+ "model.layers.34.self_attn.k_proj.weight": "pytorch_model-00139-of-00243.bin",
294
+ "model.layers.34.self_attn.o_proj.weight": "pytorch_model-00139-of-00243.bin",
295
+ "model.layers.34.self_attn.q_proj.weight": "pytorch_model-00138-of-00243.bin",
296
+ "model.layers.34.self_attn.rotary_emb.inv_freq": "pytorch_model-00139-of-00243.bin",
297
+ "model.layers.34.self_attn.v_proj.weight": "pytorch_model-00139-of-00243.bin",
298
+ "model.layers.35.input_layernorm.weight": "pytorch_model-00146-of-00243.bin",
299
+ "model.layers.35.mlp.down_proj.weight": "pytorch_model-00145-of-00243.bin",
300
+ "model.layers.35.mlp.gate_proj.weight": "pytorch_model-00144-of-00243.bin",
301
+ "model.layers.35.mlp.up_proj.weight": "pytorch_model-00146-of-00243.bin",
302
+ "model.layers.35.post_attention_layernorm.weight": "pytorch_model-00146-of-00243.bin",
303
+ "model.layers.35.self_attn.k_proj.weight": "pytorch_model-00143-of-00243.bin",
304
+ "model.layers.35.self_attn.o_proj.weight": "pytorch_model-00143-of-00243.bin",
305
+ "model.layers.35.self_attn.q_proj.weight": "pytorch_model-00142-of-00243.bin",
306
+ "model.layers.35.self_attn.rotary_emb.inv_freq": "pytorch_model-00143-of-00243.bin",
307
+ "model.layers.35.self_attn.v_proj.weight": "pytorch_model-00143-of-00243.bin",
308
+ "model.layers.36.input_layernorm.weight": "pytorch_model-00150-of-00243.bin",
309
+ "model.layers.36.mlp.down_proj.weight": "pytorch_model-00149-of-00243.bin",
310
+ "model.layers.36.mlp.gate_proj.weight": "pytorch_model-00148-of-00243.bin",
311
+ "model.layers.36.mlp.up_proj.weight": "pytorch_model-00150-of-00243.bin",
312
+ "model.layers.36.post_attention_layernorm.weight": "pytorch_model-00150-of-00243.bin",
313
+ "model.layers.36.self_attn.k_proj.weight": "pytorch_model-00147-of-00243.bin",
314
+ "model.layers.36.self_attn.o_proj.weight": "pytorch_model-00147-of-00243.bin",
315
+ "model.layers.36.self_attn.q_proj.weight": "pytorch_model-00146-of-00243.bin",
316
+ "model.layers.36.self_attn.rotary_emb.inv_freq": "pytorch_model-00147-of-00243.bin",
317
+ "model.layers.36.self_attn.v_proj.weight": "pytorch_model-00147-of-00243.bin",
318
+ "model.layers.37.input_layernorm.weight": "pytorch_model-00154-of-00243.bin",
319
+ "model.layers.37.mlp.down_proj.weight": "pytorch_model-00153-of-00243.bin",
320
+ "model.layers.37.mlp.gate_proj.weight": "pytorch_model-00152-of-00243.bin",
321
+ "model.layers.37.mlp.up_proj.weight": "pytorch_model-00154-of-00243.bin",
322
+ "model.layers.37.post_attention_layernorm.weight": "pytorch_model-00154-of-00243.bin",
323
+ "model.layers.37.self_attn.k_proj.weight": "pytorch_model-00151-of-00243.bin",
324
+ "model.layers.37.self_attn.o_proj.weight": "pytorch_model-00151-of-00243.bin",
325
+ "model.layers.37.self_attn.q_proj.weight": "pytorch_model-00150-of-00243.bin",
326
+ "model.layers.37.self_attn.rotary_emb.inv_freq": "pytorch_model-00151-of-00243.bin",
327
+ "model.layers.37.self_attn.v_proj.weight": "pytorch_model-00151-of-00243.bin",
328
+ "model.layers.38.input_layernorm.weight": "pytorch_model-00158-of-00243.bin",
329
+ "model.layers.38.mlp.down_proj.weight": "pytorch_model-00157-of-00243.bin",
330
+ "model.layers.38.mlp.gate_proj.weight": "pytorch_model-00156-of-00243.bin",
331
+ "model.layers.38.mlp.up_proj.weight": "pytorch_model-00158-of-00243.bin",
332
+ "model.layers.38.post_attention_layernorm.weight": "pytorch_model-00158-of-00243.bin",
333
+ "model.layers.38.self_attn.k_proj.weight": "pytorch_model-00155-of-00243.bin",
334
+ "model.layers.38.self_attn.o_proj.weight": "pytorch_model-00155-of-00243.bin",
335
+ "model.layers.38.self_attn.q_proj.weight": "pytorch_model-00154-of-00243.bin",
336
+ "model.layers.38.self_attn.rotary_emb.inv_freq": "pytorch_model-00155-of-00243.bin",
337
+ "model.layers.38.self_attn.v_proj.weight": "pytorch_model-00155-of-00243.bin",
338
+ "model.layers.39.input_layernorm.weight": "pytorch_model-00162-of-00243.bin",
339
+ "model.layers.39.mlp.down_proj.weight": "pytorch_model-00161-of-00243.bin",
340
+ "model.layers.39.mlp.gate_proj.weight": "pytorch_model-00160-of-00243.bin",
341
+ "model.layers.39.mlp.up_proj.weight": "pytorch_model-00162-of-00243.bin",
342
+ "model.layers.39.post_attention_layernorm.weight": "pytorch_model-00162-of-00243.bin",
343
+ "model.layers.39.self_attn.k_proj.weight": "pytorch_model-00159-of-00243.bin",
344
+ "model.layers.39.self_attn.o_proj.weight": "pytorch_model-00159-of-00243.bin",
345
+ "model.layers.39.self_attn.q_proj.weight": "pytorch_model-00158-of-00243.bin",
346
+ "model.layers.39.self_attn.rotary_emb.inv_freq": "pytorch_model-00159-of-00243.bin",
347
+ "model.layers.39.self_attn.v_proj.weight": "pytorch_model-00159-of-00243.bin",
348
+ "model.layers.4.input_layernorm.weight": "pytorch_model-00022-of-00243.bin",
349
+ "model.layers.4.mlp.down_proj.weight": "pytorch_model-00021-of-00243.bin",
350
+ "model.layers.4.mlp.gate_proj.weight": "pytorch_model-00020-of-00243.bin",
351
+ "model.layers.4.mlp.up_proj.weight": "pytorch_model-00022-of-00243.bin",
352
+ "model.layers.4.post_attention_layernorm.weight": "pytorch_model-00022-of-00243.bin",
353
+ "model.layers.4.self_attn.k_proj.weight": "pytorch_model-00019-of-00243.bin",
354
+ "model.layers.4.self_attn.o_proj.weight": "pytorch_model-00019-of-00243.bin",
355
+ "model.layers.4.self_attn.q_proj.weight": "pytorch_model-00018-of-00243.bin",
356
+ "model.layers.4.self_attn.rotary_emb.inv_freq": "pytorch_model-00019-of-00243.bin",
357
+ "model.layers.4.self_attn.v_proj.weight": "pytorch_model-00019-of-00243.bin",
358
+ "model.layers.40.input_layernorm.weight": "pytorch_model-00166-of-00243.bin",
359
+ "model.layers.40.mlp.down_proj.weight": "pytorch_model-00165-of-00243.bin",
360
+ "model.layers.40.mlp.gate_proj.weight": "pytorch_model-00164-of-00243.bin",
361
+ "model.layers.40.mlp.up_proj.weight": "pytorch_model-00166-of-00243.bin",
362
+ "model.layers.40.post_attention_layernorm.weight": "pytorch_model-00166-of-00243.bin",
363
+ "model.layers.40.self_attn.k_proj.weight": "pytorch_model-00163-of-00243.bin",
364
+ "model.layers.40.self_attn.o_proj.weight": "pytorch_model-00163-of-00243.bin",
365
+ "model.layers.40.self_attn.q_proj.weight": "pytorch_model-00162-of-00243.bin",
366
+ "model.layers.40.self_attn.rotary_emb.inv_freq": "pytorch_model-00163-of-00243.bin",
367
+ "model.layers.40.self_attn.v_proj.weight": "pytorch_model-00163-of-00243.bin",
368
+ "model.layers.41.input_layernorm.weight": "pytorch_model-00170-of-00243.bin",
369
+ "model.layers.41.mlp.down_proj.weight": "pytorch_model-00169-of-00243.bin",
370
+ "model.layers.41.mlp.gate_proj.weight": "pytorch_model-00168-of-00243.bin",
371
+ "model.layers.41.mlp.up_proj.weight": "pytorch_model-00170-of-00243.bin",
372
+ "model.layers.41.post_attention_layernorm.weight": "pytorch_model-00170-of-00243.bin",
373
+ "model.layers.41.self_attn.k_proj.weight": "pytorch_model-00167-of-00243.bin",
374
+ "model.layers.41.self_attn.o_proj.weight": "pytorch_model-00167-of-00243.bin",
375
+ "model.layers.41.self_attn.q_proj.weight": "pytorch_model-00166-of-00243.bin",
376
+ "model.layers.41.self_attn.rotary_emb.inv_freq": "pytorch_model-00167-of-00243.bin",
377
+ "model.layers.41.self_attn.v_proj.weight": "pytorch_model-00167-of-00243.bin",
378
+ "model.layers.42.input_layernorm.weight": "pytorch_model-00174-of-00243.bin",
379
+ "model.layers.42.mlp.down_proj.weight": "pytorch_model-00173-of-00243.bin",
380
+ "model.layers.42.mlp.gate_proj.weight": "pytorch_model-00172-of-00243.bin",
381
+ "model.layers.42.mlp.up_proj.weight": "pytorch_model-00174-of-00243.bin",
382
+ "model.layers.42.post_attention_layernorm.weight": "pytorch_model-00174-of-00243.bin",
383
+ "model.layers.42.self_attn.k_proj.weight": "pytorch_model-00171-of-00243.bin",
384
+ "model.layers.42.self_attn.o_proj.weight": "pytorch_model-00171-of-00243.bin",
385
+ "model.layers.42.self_attn.q_proj.weight": "pytorch_model-00170-of-00243.bin",
386
+ "model.layers.42.self_attn.rotary_emb.inv_freq": "pytorch_model-00171-of-00243.bin",
387
+ "model.layers.42.self_attn.v_proj.weight": "pytorch_model-00171-of-00243.bin",
388
+ "model.layers.43.input_layernorm.weight": "pytorch_model-00178-of-00243.bin",
389
+ "model.layers.43.mlp.down_proj.weight": "pytorch_model-00177-of-00243.bin",
390
+ "model.layers.43.mlp.gate_proj.weight": "pytorch_model-00176-of-00243.bin",
391
+ "model.layers.43.mlp.up_proj.weight": "pytorch_model-00178-of-00243.bin",
392
+ "model.layers.43.post_attention_layernorm.weight": "pytorch_model-00178-of-00243.bin",
393
+ "model.layers.43.self_attn.k_proj.weight": "pytorch_model-00175-of-00243.bin",
394
+ "model.layers.43.self_attn.o_proj.weight": "pytorch_model-00175-of-00243.bin",
395
+ "model.layers.43.self_attn.q_proj.weight": "pytorch_model-00174-of-00243.bin",
396
+ "model.layers.43.self_attn.rotary_emb.inv_freq": "pytorch_model-00175-of-00243.bin",
397
+ "model.layers.43.self_attn.v_proj.weight": "pytorch_model-00175-of-00243.bin",
398
+ "model.layers.44.input_layernorm.weight": "pytorch_model-00182-of-00243.bin",
399
+ "model.layers.44.mlp.down_proj.weight": "pytorch_model-00181-of-00243.bin",
400
+ "model.layers.44.mlp.gate_proj.weight": "pytorch_model-00180-of-00243.bin",
401
+ "model.layers.44.mlp.up_proj.weight": "pytorch_model-00182-of-00243.bin",
402
+ "model.layers.44.post_attention_layernorm.weight": "pytorch_model-00182-of-00243.bin",
403
+ "model.layers.44.self_attn.k_proj.weight": "pytorch_model-00179-of-00243.bin",
404
+ "model.layers.44.self_attn.o_proj.weight": "pytorch_model-00179-of-00243.bin",
405
+ "model.layers.44.self_attn.q_proj.weight": "pytorch_model-00178-of-00243.bin",
406
+ "model.layers.44.self_attn.rotary_emb.inv_freq": "pytorch_model-00179-of-00243.bin",
407
+ "model.layers.44.self_attn.v_proj.weight": "pytorch_model-00179-of-00243.bin",
408
+ "model.layers.45.input_layernorm.weight": "pytorch_model-00186-of-00243.bin",
409
+ "model.layers.45.mlp.down_proj.weight": "pytorch_model-00185-of-00243.bin",
410
+ "model.layers.45.mlp.gate_proj.weight": "pytorch_model-00184-of-00243.bin",
411
+ "model.layers.45.mlp.up_proj.weight": "pytorch_model-00186-of-00243.bin",
412
+ "model.layers.45.post_attention_layernorm.weight": "pytorch_model-00186-of-00243.bin",
413
+ "model.layers.45.self_attn.k_proj.weight": "pytorch_model-00183-of-00243.bin",
414
+ "model.layers.45.self_attn.o_proj.weight": "pytorch_model-00183-of-00243.bin",
415
+ "model.layers.45.self_attn.q_proj.weight": "pytorch_model-00182-of-00243.bin",
416
+ "model.layers.45.self_attn.rotary_emb.inv_freq": "pytorch_model-00183-of-00243.bin",
417
+ "model.layers.45.self_attn.v_proj.weight": "pytorch_model-00183-of-00243.bin",
418
+ "model.layers.46.input_layernorm.weight": "pytorch_model-00190-of-00243.bin",
419
+ "model.layers.46.mlp.down_proj.weight": "pytorch_model-00189-of-00243.bin",
420
+ "model.layers.46.mlp.gate_proj.weight": "pytorch_model-00188-of-00243.bin",
421
+ "model.layers.46.mlp.up_proj.weight": "pytorch_model-00190-of-00243.bin",
422
+ "model.layers.46.post_attention_layernorm.weight": "pytorch_model-00190-of-00243.bin",
423
+ "model.layers.46.self_attn.k_proj.weight": "pytorch_model-00187-of-00243.bin",
424
+ "model.layers.46.self_attn.o_proj.weight": "pytorch_model-00187-of-00243.bin",
425
+ "model.layers.46.self_attn.q_proj.weight": "pytorch_model-00186-of-00243.bin",
426
+ "model.layers.46.self_attn.rotary_emb.inv_freq": "pytorch_model-00187-of-00243.bin",
427
+ "model.layers.46.self_attn.v_proj.weight": "pytorch_model-00187-of-00243.bin",
428
+ "model.layers.47.input_layernorm.weight": "pytorch_model-00194-of-00243.bin",
429
+ "model.layers.47.mlp.down_proj.weight": "pytorch_model-00193-of-00243.bin",
430
+ "model.layers.47.mlp.gate_proj.weight": "pytorch_model-00192-of-00243.bin",
431
+ "model.layers.47.mlp.up_proj.weight": "pytorch_model-00194-of-00243.bin",
432
+ "model.layers.47.post_attention_layernorm.weight": "pytorch_model-00194-of-00243.bin",
433
+ "model.layers.47.self_attn.k_proj.weight": "pytorch_model-00191-of-00243.bin",
434
+ "model.layers.47.self_attn.o_proj.weight": "pytorch_model-00191-of-00243.bin",
435
+ "model.layers.47.self_attn.q_proj.weight": "pytorch_model-00190-of-00243.bin",
436
+ "model.layers.47.self_attn.rotary_emb.inv_freq": "pytorch_model-00191-of-00243.bin",
437
+ "model.layers.47.self_attn.v_proj.weight": "pytorch_model-00191-of-00243.bin",
438
+ "model.layers.48.input_layernorm.weight": "pytorch_model-00198-of-00243.bin",
439
+ "model.layers.48.mlp.down_proj.weight": "pytorch_model-00197-of-00243.bin",
440
+ "model.layers.48.mlp.gate_proj.weight": "pytorch_model-00196-of-00243.bin",
441
+ "model.layers.48.mlp.up_proj.weight": "pytorch_model-00198-of-00243.bin",
442
+ "model.layers.48.post_attention_layernorm.weight": "pytorch_model-00198-of-00243.bin",
443
+ "model.layers.48.self_attn.k_proj.weight": "pytorch_model-00195-of-00243.bin",
444
+ "model.layers.48.self_attn.o_proj.weight": "pytorch_model-00195-of-00243.bin",
445
+ "model.layers.48.self_attn.q_proj.weight": "pytorch_model-00194-of-00243.bin",
446
+ "model.layers.48.self_attn.rotary_emb.inv_freq": "pytorch_model-00195-of-00243.bin",
447
+ "model.layers.48.self_attn.v_proj.weight": "pytorch_model-00195-of-00243.bin",
448
+ "model.layers.49.input_layernorm.weight": "pytorch_model-00202-of-00243.bin",
449
+ "model.layers.49.mlp.down_proj.weight": "pytorch_model-00201-of-00243.bin",
450
+ "model.layers.49.mlp.gate_proj.weight": "pytorch_model-00200-of-00243.bin",
451
+ "model.layers.49.mlp.up_proj.weight": "pytorch_model-00202-of-00243.bin",
452
+ "model.layers.49.post_attention_layernorm.weight": "pytorch_model-00202-of-00243.bin",
453
+ "model.layers.49.self_attn.k_proj.weight": "pytorch_model-00199-of-00243.bin",
454
+ "model.layers.49.self_attn.o_proj.weight": "pytorch_model-00199-of-00243.bin",
455
+ "model.layers.49.self_attn.q_proj.weight": "pytorch_model-00198-of-00243.bin",
456
+ "model.layers.49.self_attn.rotary_emb.inv_freq": "pytorch_model-00199-of-00243.bin",
457
+ "model.layers.49.self_attn.v_proj.weight": "pytorch_model-00199-of-00243.bin",
458
+ "model.layers.5.input_layernorm.weight": "pytorch_model-00026-of-00243.bin",
459
+ "model.layers.5.mlp.down_proj.weight": "pytorch_model-00025-of-00243.bin",
460
+ "model.layers.5.mlp.gate_proj.weight": "pytorch_model-00024-of-00243.bin",
461
+ "model.layers.5.mlp.up_proj.weight": "pytorch_model-00026-of-00243.bin",
462
+ "model.layers.5.post_attention_layernorm.weight": "pytorch_model-00026-of-00243.bin",
463
+ "model.layers.5.self_attn.k_proj.weight": "pytorch_model-00023-of-00243.bin",
464
+ "model.layers.5.self_attn.o_proj.weight": "pytorch_model-00023-of-00243.bin",
465
+ "model.layers.5.self_attn.q_proj.weight": "pytorch_model-00022-of-00243.bin",
466
+ "model.layers.5.self_attn.rotary_emb.inv_freq": "pytorch_model-00023-of-00243.bin",
467
+ "model.layers.5.self_attn.v_proj.weight": "pytorch_model-00023-of-00243.bin",
468
+ "model.layers.50.input_layernorm.weight": "pytorch_model-00206-of-00243.bin",
469
+ "model.layers.50.mlp.down_proj.weight": "pytorch_model-00205-of-00243.bin",
470
+ "model.layers.50.mlp.gate_proj.weight": "pytorch_model-00204-of-00243.bin",
471
+ "model.layers.50.mlp.up_proj.weight": "pytorch_model-00206-of-00243.bin",
472
+ "model.layers.50.post_attention_layernorm.weight": "pytorch_model-00206-of-00243.bin",
473
+ "model.layers.50.self_attn.k_proj.weight": "pytorch_model-00203-of-00243.bin",
474
+ "model.layers.50.self_attn.o_proj.weight": "pytorch_model-00203-of-00243.bin",
475
+ "model.layers.50.self_attn.q_proj.weight": "pytorch_model-00202-of-00243.bin",
476
+ "model.layers.50.self_attn.rotary_emb.inv_freq": "pytorch_model-00203-of-00243.bin",
477
+ "model.layers.50.self_attn.v_proj.weight": "pytorch_model-00203-of-00243.bin",
478
+ "model.layers.51.input_layernorm.weight": "pytorch_model-00210-of-00243.bin",
479
+ "model.layers.51.mlp.down_proj.weight": "pytorch_model-00209-of-00243.bin",
480
+ "model.layers.51.mlp.gate_proj.weight": "pytorch_model-00208-of-00243.bin",
481
+ "model.layers.51.mlp.up_proj.weight": "pytorch_model-00210-of-00243.bin",
482
+ "model.layers.51.post_attention_layernorm.weight": "pytorch_model-00210-of-00243.bin",
483
+ "model.layers.51.self_attn.k_proj.weight": "pytorch_model-00207-of-00243.bin",
484
+ "model.layers.51.self_attn.o_proj.weight": "pytorch_model-00207-of-00243.bin",
485
+ "model.layers.51.self_attn.q_proj.weight": "pytorch_model-00206-of-00243.bin",
486
+ "model.layers.51.self_attn.rotary_emb.inv_freq": "pytorch_model-00207-of-00243.bin",
487
+ "model.layers.51.self_attn.v_proj.weight": "pytorch_model-00207-of-00243.bin",
488
+ "model.layers.52.input_layernorm.weight": "pytorch_model-00214-of-00243.bin",
489
+ "model.layers.52.mlp.down_proj.weight": "pytorch_model-00213-of-00243.bin",
490
+ "model.layers.52.mlp.gate_proj.weight": "pytorch_model-00212-of-00243.bin",
491
+ "model.layers.52.mlp.up_proj.weight": "pytorch_model-00214-of-00243.bin",
492
+ "model.layers.52.post_attention_layernorm.weight": "pytorch_model-00214-of-00243.bin",
493
+ "model.layers.52.self_attn.k_proj.weight": "pytorch_model-00211-of-00243.bin",
494
+ "model.layers.52.self_attn.o_proj.weight": "pytorch_model-00211-of-00243.bin",
495
+ "model.layers.52.self_attn.q_proj.weight": "pytorch_model-00210-of-00243.bin",
496
+ "model.layers.52.self_attn.rotary_emb.inv_freq": "pytorch_model-00211-of-00243.bin",
497
+ "model.layers.52.self_attn.v_proj.weight": "pytorch_model-00211-of-00243.bin",
498
+ "model.layers.53.input_layernorm.weight": "pytorch_model-00218-of-00243.bin",
499
+ "model.layers.53.mlp.down_proj.weight": "pytorch_model-00217-of-00243.bin",
500
+ "model.layers.53.mlp.gate_proj.weight": "pytorch_model-00216-of-00243.bin",
501
+ "model.layers.53.mlp.up_proj.weight": "pytorch_model-00218-of-00243.bin",
502
+ "model.layers.53.post_attention_layernorm.weight": "pytorch_model-00218-of-00243.bin",
503
+ "model.layers.53.self_attn.k_proj.weight": "pytorch_model-00215-of-00243.bin",
504
+ "model.layers.53.self_attn.o_proj.weight": "pytorch_model-00215-of-00243.bin",
505
+ "model.layers.53.self_attn.q_proj.weight": "pytorch_model-00214-of-00243.bin",
506
+ "model.layers.53.self_attn.rotary_emb.inv_freq": "pytorch_model-00215-of-00243.bin",
507
+ "model.layers.53.self_attn.v_proj.weight": "pytorch_model-00215-of-00243.bin",
508
+ "model.layers.54.input_layernorm.weight": "pytorch_model-00222-of-00243.bin",
509
+ "model.layers.54.mlp.down_proj.weight": "pytorch_model-00221-of-00243.bin",
510
+ "model.layers.54.mlp.gate_proj.weight": "pytorch_model-00220-of-00243.bin",
511
+ "model.layers.54.mlp.up_proj.weight": "pytorch_model-00222-of-00243.bin",
512
+ "model.layers.54.post_attention_layernorm.weight": "pytorch_model-00222-of-00243.bin",
513
+ "model.layers.54.self_attn.k_proj.weight": "pytorch_model-00219-of-00243.bin",
514
+ "model.layers.54.self_attn.o_proj.weight": "pytorch_model-00219-of-00243.bin",
515
+ "model.layers.54.self_attn.q_proj.weight": "pytorch_model-00218-of-00243.bin",
516
+ "model.layers.54.self_attn.rotary_emb.inv_freq": "pytorch_model-00219-of-00243.bin",
517
+ "model.layers.54.self_attn.v_proj.weight": "pytorch_model-00219-of-00243.bin",
518
+ "model.layers.55.input_layernorm.weight": "pytorch_model-00226-of-00243.bin",
519
+ "model.layers.55.mlp.down_proj.weight": "pytorch_model-00225-of-00243.bin",
520
+ "model.layers.55.mlp.gate_proj.weight": "pytorch_model-00224-of-00243.bin",
521
+ "model.layers.55.mlp.up_proj.weight": "pytorch_model-00226-of-00243.bin",
522
+ "model.layers.55.post_attention_layernorm.weight": "pytorch_model-00226-of-00243.bin",
523
+ "model.layers.55.self_attn.k_proj.weight": "pytorch_model-00223-of-00243.bin",
524
+ "model.layers.55.self_attn.o_proj.weight": "pytorch_model-00223-of-00243.bin",
525
+ "model.layers.55.self_attn.q_proj.weight": "pytorch_model-00222-of-00243.bin",
526
+ "model.layers.55.self_attn.rotary_emb.inv_freq": "pytorch_model-00223-of-00243.bin",
527
+ "model.layers.55.self_attn.v_proj.weight": "pytorch_model-00223-of-00243.bin",
528
+ "model.layers.56.input_layernorm.weight": "pytorch_model-00230-of-00243.bin",
529
+ "model.layers.56.mlp.down_proj.weight": "pytorch_model-00229-of-00243.bin",
530
+ "model.layers.56.mlp.gate_proj.weight": "pytorch_model-00228-of-00243.bin",
531
+ "model.layers.56.mlp.up_proj.weight": "pytorch_model-00230-of-00243.bin",
532
+ "model.layers.56.post_attention_layernorm.weight": "pytorch_model-00230-of-00243.bin",
533
+ "model.layers.56.self_attn.k_proj.weight": "pytorch_model-00227-of-00243.bin",
534
+ "model.layers.56.self_attn.o_proj.weight": "pytorch_model-00227-of-00243.bin",
535
+ "model.layers.56.self_attn.q_proj.weight": "pytorch_model-00226-of-00243.bin",
536
+ "model.layers.56.self_attn.rotary_emb.inv_freq": "pytorch_model-00227-of-00243.bin",
537
+ "model.layers.56.self_attn.v_proj.weight": "pytorch_model-00227-of-00243.bin",
538
+ "model.layers.57.input_layernorm.weight": "pytorch_model-00234-of-00243.bin",
539
+ "model.layers.57.mlp.down_proj.weight": "pytorch_model-00233-of-00243.bin",
540
+ "model.layers.57.mlp.gate_proj.weight": "pytorch_model-00232-of-00243.bin",
541
+ "model.layers.57.mlp.up_proj.weight": "pytorch_model-00234-of-00243.bin",
542
+ "model.layers.57.post_attention_layernorm.weight": "pytorch_model-00234-of-00243.bin",
543
+ "model.layers.57.self_attn.k_proj.weight": "pytorch_model-00231-of-00243.bin",
544
+ "model.layers.57.self_attn.o_proj.weight": "pytorch_model-00231-of-00243.bin",
545
+ "model.layers.57.self_attn.q_proj.weight": "pytorch_model-00230-of-00243.bin",
546
+ "model.layers.57.self_attn.rotary_emb.inv_freq": "pytorch_model-00231-of-00243.bin",
547
+ "model.layers.57.self_attn.v_proj.weight": "pytorch_model-00231-of-00243.bin",
548
+ "model.layers.58.input_layernorm.weight": "pytorch_model-00238-of-00243.bin",
549
+ "model.layers.58.mlp.down_proj.weight": "pytorch_model-00237-of-00243.bin",
550
+ "model.layers.58.mlp.gate_proj.weight": "pytorch_model-00236-of-00243.bin",
551
+ "model.layers.58.mlp.up_proj.weight": "pytorch_model-00238-of-00243.bin",
552
+ "model.layers.58.post_attention_layernorm.weight": "pytorch_model-00238-of-00243.bin",
553
+ "model.layers.58.self_attn.k_proj.weight": "pytorch_model-00235-of-00243.bin",
554
+ "model.layers.58.self_attn.o_proj.weight": "pytorch_model-00235-of-00243.bin",
555
+ "model.layers.58.self_attn.q_proj.weight": "pytorch_model-00234-of-00243.bin",
556
+ "model.layers.58.self_attn.rotary_emb.inv_freq": "pytorch_model-00235-of-00243.bin",
557
+ "model.layers.58.self_attn.v_proj.weight": "pytorch_model-00235-of-00243.bin",
558
+ "model.layers.59.input_layernorm.weight": "pytorch_model-00242-of-00243.bin",
559
+ "model.layers.59.mlp.down_proj.weight": "pytorch_model-00241-of-00243.bin",
560
+ "model.layers.59.mlp.gate_proj.weight": "pytorch_model-00240-of-00243.bin",
561
+ "model.layers.59.mlp.up_proj.weight": "pytorch_model-00242-of-00243.bin",
562
+ "model.layers.59.post_attention_layernorm.weight": "pytorch_model-00242-of-00243.bin",
563
+ "model.layers.59.self_attn.k_proj.weight": "pytorch_model-00239-of-00243.bin",
564
+ "model.layers.59.self_attn.o_proj.weight": "pytorch_model-00239-of-00243.bin",
565
+ "model.layers.59.self_attn.q_proj.weight": "pytorch_model-00238-of-00243.bin",
566
+ "model.layers.59.self_attn.rotary_emb.inv_freq": "pytorch_model-00239-of-00243.bin",
567
+ "model.layers.59.self_attn.v_proj.weight": "pytorch_model-00239-of-00243.bin",
568
+ "model.layers.6.input_layernorm.weight": "pytorch_model-00030-of-00243.bin",
569
+ "model.layers.6.mlp.down_proj.weight": "pytorch_model-00029-of-00243.bin",
570
+ "model.layers.6.mlp.gate_proj.weight": "pytorch_model-00028-of-00243.bin",
571
+ "model.layers.6.mlp.up_proj.weight": "pytorch_model-00030-of-00243.bin",
572
+ "model.layers.6.post_attention_layernorm.weight": "pytorch_model-00030-of-00243.bin",
573
+ "model.layers.6.self_attn.k_proj.weight": "pytorch_model-00027-of-00243.bin",
574
+ "model.layers.6.self_attn.o_proj.weight": "pytorch_model-00027-of-00243.bin",
575
+ "model.layers.6.self_attn.q_proj.weight": "pytorch_model-00026-of-00243.bin",
576
+ "model.layers.6.self_attn.rotary_emb.inv_freq": "pytorch_model-00027-of-00243.bin",
577
+ "model.layers.6.self_attn.v_proj.weight": "pytorch_model-00027-of-00243.bin",
578
+ "model.layers.7.input_layernorm.weight": "pytorch_model-00034-of-00243.bin",
579
+ "model.layers.7.mlp.down_proj.weight": "pytorch_model-00033-of-00243.bin",
580
+ "model.layers.7.mlp.gate_proj.weight": "pytorch_model-00032-of-00243.bin",
581
+ "model.layers.7.mlp.up_proj.weight": "pytorch_model-00034-of-00243.bin",
582
+ "model.layers.7.post_attention_layernorm.weight": "pytorch_model-00034-of-00243.bin",
583
+ "model.layers.7.self_attn.k_proj.weight": "pytorch_model-00031-of-00243.bin",
584
+ "model.layers.7.self_attn.o_proj.weight": "pytorch_model-00031-of-00243.bin",
585
+ "model.layers.7.self_attn.q_proj.weight": "pytorch_model-00030-of-00243.bin",
586
+ "model.layers.7.self_attn.rotary_emb.inv_freq": "pytorch_model-00031-of-00243.bin",
587
+ "model.layers.7.self_attn.v_proj.weight": "pytorch_model-00031-of-00243.bin",
588
+ "model.layers.8.input_layernorm.weight": "pytorch_model-00038-of-00243.bin",
589
+ "model.layers.8.mlp.down_proj.weight": "pytorch_model-00037-of-00243.bin",
590
+ "model.layers.8.mlp.gate_proj.weight": "pytorch_model-00036-of-00243.bin",
591
+ "model.layers.8.mlp.up_proj.weight": "pytorch_model-00038-of-00243.bin",
592
+ "model.layers.8.post_attention_layernorm.weight": "pytorch_model-00038-of-00243.bin",
593
+ "model.layers.8.self_attn.k_proj.weight": "pytorch_model-00035-of-00243.bin",
594
+ "model.layers.8.self_attn.o_proj.weight": "pytorch_model-00035-of-00243.bin",
595
+ "model.layers.8.self_attn.q_proj.weight": "pytorch_model-00034-of-00243.bin",
596
+ "model.layers.8.self_attn.rotary_emb.inv_freq": "pytorch_model-00035-of-00243.bin",
597
+ "model.layers.8.self_attn.v_proj.weight": "pytorch_model-00035-of-00243.bin",
598
+ "model.layers.9.input_layernorm.weight": "pytorch_model-00042-of-00243.bin",
599
+ "model.layers.9.mlp.down_proj.weight": "pytorch_model-00041-of-00243.bin",
600
+ "model.layers.9.mlp.gate_proj.weight": "pytorch_model-00040-of-00243.bin",
601
+ "model.layers.9.mlp.up_proj.weight": "pytorch_model-00042-of-00243.bin",
602
+ "model.layers.9.post_attention_layernorm.weight": "pytorch_model-00042-of-00243.bin",
603
+ "model.layers.9.self_attn.k_proj.weight": "pytorch_model-00039-of-00243.bin",
604
+ "model.layers.9.self_attn.o_proj.weight": "pytorch_model-00039-of-00243.bin",
605
+ "model.layers.9.self_attn.q_proj.weight": "pytorch_model-00038-of-00243.bin",
606
+ "model.layers.9.self_attn.rotary_emb.inv_freq": "pytorch_model-00039-of-00243.bin",
607
+ "model.layers.9.self_attn.v_proj.weight": "pytorch_model-00039-of-00243.bin",
608
+ "model.norm.weight": "pytorch_model-00242-of-00243.bin"
609
+ }
610
+ }
.ipynb_checkpoints/special_tokens_map-checkpoint.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "unk_token": {
17
+ "content": "<unk>",
18
+ "lstrip": false,
19
+ "normalized": true,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
+ }
.ipynb_checkpoints/tokenizer_config-checkpoint.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "bos_token": {
5
+ "__type": "AddedToken",
6
+ "content": "<s>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "clean_up_tokenization_spaces": false,
13
+ "eos_token": {
14
+ "__type": "AddedToken",
15
+ "content": "</s>",
16
+ "lstrip": false,
17
+ "normalized": true,
18
+ "rstrip": false,
19
+ "single_word": false
20
+ },
21
+ "model_max_length": 2048,
22
+ "pad_token": null,
23
+ "sp_model_kwargs": {},
24
+ "tokenizer_class": "LlamaTokenizer",
25
+ "unk_token": {
26
+ "__type": "AddedToken",
27
+ "content": "<unk>",
28
+ "lstrip": false,
29
+ "normalized": true,
30
+ "rstrip": false,
31
+ "single_word": false
32
+ }
33
+ }
4bit.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a8db3fbb9af7b86d2742aad0aea95773bca7529aa03ef06f737a1ef0b8998e6
3
+ size 16940128404
README.md ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Merge of [huggyllama/llama-30b](https://huggingface.co/huggyllama/llama-30b) + [kaiokendev/SuperCOT-LoRA](https://huggingface.co/kaiokendev/SuperCOT-LoRA/edit/main/README.md)
2
+
3
+ Supercot was trained to work with langchain prompting.
4
+
5
+
6
+ Load up locally in my custom LLM notebook that uses the Oobabooga modules to load up models: https://github.com/ausboss/Local-LLM-Langchain
7
+
8
+ Then you can add cells from of these other notebooks for testing: https://github.com/gkamradt/langchain-tutorials
9
+
10
+
11
+
12
+ # From Koikendev Lora page
13
+
14
+ ### Compatibility
15
+ This LoRA is compatible with any 7B, 13B or 30B 4-bit quantized LLaMa model, including ggml quantized converted bins
16
+
17
+ ### Prompting
18
+ You should prompt the LoRA the same way you would prompt Alpaca or Alpacino:
19
+
20
+ ```
21
+ Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
22
+
23
+ ### Instruction:
24
+ <instruction>
25
+
26
+ ### Input:
27
+ <any additional context. Remove this if it's not neccesary>
28
+
29
+ ### Response:
30
+ <make sure to leave a single new-line here for optimal results>
31
+ ```
32
+
33
+ Remember that with lower parameter sizes, the structure of the prompt becomes more important. The same prompt worded differently can give wildly different answers. Consider using the following suggestion suffixes to improve output quality:
34
+
35
+ - "Think through this step by step"
36
+ - "Let's think about this logically"
37
+ - "Explain your reasoning"
38
+ - "Provide details to support your answer"
39
+ - "Compare and contrast your answer with alternatives"
40
+
41
+ ### Coming Soon
42
+ - Tweet fix for 13B and 7B - lower model sizes seem to be extremely sensitive to hashtags at the end of training data responses, especially at longer cutoffs
config.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "huggyllama/llama-30b",
3
+ "architectures": [
4
+ "LlamaForCausalLM"
5
+ ],
6
+ "bos_token_id": 1,
7
+ "eos_token_id": 2,
8
+ "hidden_act": "silu",
9
+ "hidden_size": 6656,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 17920,
12
+ "max_position_embeddings": 2048,
13
+ "max_sequence_length": 2048,
14
+ "model_type": "llama",
15
+ "num_attention_heads": 52,
16
+ "num_hidden_layers": 60,
17
+ "pad_token_id": 0,
18
+ "rms_norm_eps": 1e-06,
19
+ "tie_word_embeddings": false,
20
+ "torch_dtype": "float16",
21
+ "transformers_version": "4.28.0",
22
+ "use_cache": true,
23
+ "vocab_size": 32000
24
+ }
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "pad_token_id": 0,
6
+ "transformers_version": "4.28.0"
7
+ }
huggingface-metadata.txt ADDED
@@ -0,0 +1,248 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ url: https://huggingface.co/ausboss/llama-30b-supercot
2
+ branch: main
3
+ download date: 2023-05-24 03:23:45
4
+ sha256sum:
5
+ 09cf82059eaabd225a99bf38743c21347bd8f1afd60407fba21db30022e8275b pytorch_model-00001-of-00243.bin
6
+ 9d058a7dc32ebfcbd3b1aa1b0cf98b8cb236c09530e06c8a1a34bfa51f940fac pytorch_model-00002-of-00243.bin
7
+ 483ff136784e1446a1d47152ad369275b63d97ac454d45eb31bd0383de2b3be5 pytorch_model-00003-of-00243.bin
8
+ b5dc5c69cbd281b33441630fb502e371026415c49cc71cebdb065ebcac7b6eba pytorch_model-00004-of-00243.bin
9
+ 38f703f231fca57a9d743fdea91bb4e052c0f0644927ec00905662a80665601c pytorch_model-00005-of-00243.bin
10
+ 2aa57f5546e994b9259ebf08767063be4ed2512fb7e9fe98d719d3e027ebb8f4 pytorch_model-00006-of-00243.bin
11
+ ee40d76e5c51b5ecc279a1397afa50983b274c1bd0b145c7ae33af872a1e28d4 pytorch_model-00007-of-00243.bin
12
+ 518c4fdaa89fa9876c3cae8ada56b7b96b0e230cb7af370308a869e4ede07395 pytorch_model-00008-of-00243.bin
13
+ a381bb81a90c9671940886f1afa1f1738d740bdf74f79b9849dbd1b99be616b0 pytorch_model-00009-of-00243.bin
14
+ 0d725d67acabd07d3ca1c2d7508068d9ea763391658d35c4db3f5df4a78fd328 pytorch_model-00010-of-00243.bin
15
+ 885fa2668e40d8c75ff4f0506f1594db76a5260bc643db5c33bda2b03c3fa640 pytorch_model-00011-of-00243.bin
16
+ 602302d9a2e2f6e9aacc9c5593720b0366fb836a41d117f856ef936606f4c9a7 pytorch_model-00012-of-00243.bin
17
+ 3f1fe1988509adfb8d81b2ffa2ad7061264bcbde838f58c21f05e17c63cccc01 pytorch_model-00013-of-00243.bin
18
+ bb699ddfd979f80d32cd8d11ac342bd598bb50a1e6d4172fb741a0e056680218 pytorch_model-00014-of-00243.bin
19
+ 488e49dc2810cb0c8403b2891c6f81835aa6c87a3112ec83ec67368ea1740ccf pytorch_model-00015-of-00243.bin
20
+ 615e7dd2bb66158e3079cd0fa16c4a0519c3119ab967a509763db50df62890ad pytorch_model-00016-of-00243.bin
21
+ ffeb25c38f3da6737a18042ea6ec758aadf1c08fa3d10f11122b956c6f62c52e pytorch_model-00017-of-00243.bin
22
+ 899de81f2149db059c3ea2f44d9ab95e2b1632e966e22b3750b0dcb17c4f87bc pytorch_model-00018-of-00243.bin
23
+ e6319f1a3e67b5d2936933d58362962708d12f0cd796cb2ded83fcbaef502371 pytorch_model-00019-of-00243.bin
24
+ f1d0aa33e387ecdca5428e36c9be1109db674853b2871625cec6b9b6a5af4ab8 pytorch_model-00020-of-00243.bin
25
+ 5b3ae784451facf7bc237364bb005335f3be65a266529782af9b4a114cfb2837 pytorch_model-00021-of-00243.bin
26
+ 78726f86efadb520399f36d6bdc00388a8555ec8a9b0c7450b92e335236c4079 pytorch_model-00022-of-00243.bin
27
+ f4f1e3de05871c160c943ca38c26f3018a9fcd9afd1fe5d6b96c7196f5451539 pytorch_model-00023-of-00243.bin
28
+ 58385fda4dd1ddff817078bcc304548cf1ee60dd75e23f823720c271d2b84ab2 pytorch_model-00024-of-00243.bin
29
+ 35e7d0c3acc694ef071ffb46ce4e28a6109b0f65a3203d9ac52d3e8d4de2b8a1 pytorch_model-00025-of-00243.bin
30
+ f707916b73bf3e093baad418cf9030694e7b0692d0463d7df7d612d5dbae1824 pytorch_model-00026-of-00243.bin
31
+ 272a5bd29c3067980f30ddea2b1d2299bc805b7cf2932cd60b08c26ca48620c2 pytorch_model-00027-of-00243.bin
32
+ 3b2228ebf36f2b157154cf541ad45a95e12411d94ec8d96130816b7f77d7c4b5 pytorch_model-00028-of-00243.bin
33
+ d0f173c3eae5b18afa589d9f418f607ed250846830eb6360882e6de4ec7a6a9c pytorch_model-00029-of-00243.bin
34
+ 5ed6a1d34e7309b0019c74f84d4fa1f013cbb0136388b5446997899d053daad9 pytorch_model-00030-of-00243.bin
35
+ 83da5b0f1d756afca6beb4abd3b6ef6d54b97729d7b731e79634149eea752f1f pytorch_model-00031-of-00243.bin
36
+ 904fa86c833bae7b562e7d6cabecbbf2f7652d9b77f5dfe63ffb1102879eca79 pytorch_model-00032-of-00243.bin
37
+ 3ac46349de58557f72a99408f64d40475be9d762e1b92c89639b9216f374a812 pytorch_model-00033-of-00243.bin
38
+ a784cd0657d7f3d059da7953faf4ad031c912db84408e62e6354f3c5ac055009 pytorch_model-00034-of-00243.bin
39
+ af219ff22435d8331db4cce62e95dc2aff7f3d6a6f94e2412f5bd770d914d31d pytorch_model-00035-of-00243.bin
40
+ 64f205f529359c3c7aa894afd2f92c978c189adad15621c95fdeb589c8888610 pytorch_model-00036-of-00243.bin
41
+ ec644a8bff8bdb6cbb90c13baf23bc4cfdf4b0c6cd061fc8a301248150764251 pytorch_model-00037-of-00243.bin
42
+ a0654f007170e8667204132d34414670686b13be76ff5d7e5580c4cb48e63c54 pytorch_model-00038-of-00243.bin
43
+ 7ccbf439ef076c3ae403d668dd0660c7bdd3b0e9e2b8f94e3586a63ff2b89902 pytorch_model-00039-of-00243.bin
44
+ 75f0735e581714ce2c4e7435dd649e3765156b5394b7bc1e6efd6cf5379c3a79 pytorch_model-00040-of-00243.bin
45
+ f0bc1eb8f14cc3c336665c6afe3c6f64e39960d92453db2902f1cf0813f79a11 pytorch_model-00041-of-00243.bin
46
+ 3602e6f55f5d831429539fd0b0998d7cdcb68c337d288fa498b6cde5b8c52074 pytorch_model-00042-of-00243.bin
47
+ 74d9d42bec0d46c04ae048068d9d9a0283a7287a39f9d73178e66768d37b226d pytorch_model-00043-of-00243.bin
48
+ 48dd4dfd621872b4028e4ec70a3eaa768df2586865264349d70dfefb6df6d37e pytorch_model-00044-of-00243.bin
49
+ 8adb35cb6efc3d91f53211034abd47c6a8562e9f37b85aeaad9b28511efd78fd pytorch_model-00045-of-00243.bin
50
+ 7f4dd4b88af1d669fc20d55f1682d21728a1b68b075894436fc6f6f6007e5b4f pytorch_model-00046-of-00243.bin
51
+ d6a78ac519544c67c784acbdd670f68d0b7b45d9be0cf3accbb210b0505303de pytorch_model-00047-of-00243.bin
52
+ 69d34eaaa8a48b0bbd14887b4732deaeb9038cb56299fa30dbdd66bf7cf0a004 pytorch_model-00048-of-00243.bin
53
+ 94c49dcde09e908596928aae6839828ac087a5ad5ce9871fd24ec6889ecd53e7 pytorch_model-00049-of-00243.bin
54
+ 0a406bdc68423623f681091482727715007a692d7740ebe0e4b29468e2062b6d pytorch_model-00050-of-00243.bin
55
+ b485d818f5f7bd134e3df0382ea9c7c6e372e0cadba716a91d144d090058fe62 pytorch_model-00051-of-00243.bin
56
+ 6434295e65d4ef96baa8af71f44b44a99b47095b988de1bd48c7e8ba8a3c80be pytorch_model-00052-of-00243.bin
57
+ 5ac66042226aee62e71e1240c095bb49a6f50b935b9797c1d36d27911e38a21f pytorch_model-00053-of-00243.bin
58
+ e589a0b358fcba4f5e60f6a04212e2b8641b489d1584310caf15375d2426c60d pytorch_model-00054-of-00243.bin
59
+ 336e7c9aa81429b8bd6722857c1e442cc24449c64922e19c11ec300ab52780bb pytorch_model-00055-of-00243.bin
60
+ 87c4c538088b0ef6c542c3b837d0981156078453f2115bf3a2c39ae7cd37f998 pytorch_model-00056-of-00243.bin
61
+ 2d29fd39708d7cfd7d59995d0430c2cdb25f2f7ac8dd3d31f950b963d9806aa3 pytorch_model-00057-of-00243.bin
62
+ 10c9abdf748e2aee9a3f9c56789769225ec86cc4fd00c02604a3bfb6116f7acf pytorch_model-00058-of-00243.bin
63
+ 28f4ccf6a88a36e6e99514978f57cce70359cc32df66dde4091addc12ed6fd30 pytorch_model-00059-of-00243.bin
64
+ f76d024c8d574fde700a89c3e4b18fe83d2ee9c95716035b7310129cd47f5302 pytorch_model-00060-of-00243.bin
65
+ 2918b4ab52ef985a42e29e4435bbe2f128b17c5901ca4068fb3ed34cb5555d9e pytorch_model-00061-of-00243.bin
66
+ ab7bfcb3aef1e67477d6f1f68e397e90b4d5c3a82903f1a13a0cccf369dcccef pytorch_model-00062-of-00243.bin
67
+ e944d9ecc8cfdfb0dce84f78f8707006a73f24fb8fa76a9c230123425831a184 pytorch_model-00063-of-00243.bin
68
+ ee20c0ec398ceab9597d104513b90336e5f4c672cbde5a3b2ece936c8a3f5148 pytorch_model-00064-of-00243.bin
69
+ 6869b1ed0d6a2daf25939ac8ff1c696589fd62e47c114516e5726369c8e562b2 pytorch_model-00065-of-00243.bin
70
+ 29f5ead347af64a517cb43c02140b8cca9e56831c6997709ceb92209585c9426 pytorch_model-00066-of-00243.bin
71
+ f3521ab2911ac86f3a8d791bd700a4cb70425f53f2639ce5d166dbf380968112 pytorch_model-00067-of-00243.bin
72
+ 3e25fb6e55e1a627fa1bcb76c5371b8ce9b0bd233fd408a70fbc00dbb8d727c0 pytorch_model-00068-of-00243.bin
73
+ 3e9c59b9ea19e75fe7de7ada5e41dc8c38bc95b0925eb1f41cd493d024ae6899 pytorch_model-00069-of-00243.bin
74
+ 0aceb706fe626a8f26a52535ba9f164735aa2b9018d6f71266f07f5102bde48f pytorch_model-00070-of-00243.bin
75
+ 43a5999497f7aee7be2f3943fb259bb283a457bbab83d08833bbc4ca55ea44d9 pytorch_model-00071-of-00243.bin
76
+ 341304b4d0a6d8d616b6527e9a27940ff6fdff744e9d5d33648222fd788245fa pytorch_model-00072-of-00243.bin
77
+ ebf99e3cc61828079936a9c5b6ece50d6ca76b34f5b8e7b5544eb95771068198 pytorch_model-00073-of-00243.bin
78
+ afa4171a1f12702ae7ff0cdc59cae74d3211c2110a1fb4cedc9b67d5915e4519 pytorch_model-00074-of-00243.bin
79
+ a6980b5aa243a7c2cbd55ae561221272c32c3670f34964933f004934939eed5e pytorch_model-00075-of-00243.bin
80
+ c61bdce8f9252a08ecc3b4af42ee699b943ff7fe7086ca5f5c8e363ebda18d08 pytorch_model-00076-of-00243.bin
81
+ 82139e62b584a952ef1610e251f78c9f6175d70891b7ef3240c89ca3175df5b7 pytorch_model-00077-of-00243.bin
82
+ 0f8e26f8d0eb89d64609fe3846c3fc2b904704ce82b7b67037e4f105ea4b00e1 pytorch_model-00078-of-00243.bin
83
+ 43a4343a8010a2e773b82a054a4b37669690a4b361d6a29eee62aa184279a22c pytorch_model-00079-of-00243.bin
84
+ 85fc194be080bc29eace87a662f14afea9376a10b936daa89bdbc03162bfca89 pytorch_model-00080-of-00243.bin
85
+ 016d9d5873f09d06760dabce4acb167a7115d2487fdb09a19dd0f51e195c755c pytorch_model-00081-of-00243.bin
86
+ fa4b253a4d70d0b6e9619fe1ca4194a64c6cdb1cc6e30e79b47a6d84068bdf52 pytorch_model-00082-of-00243.bin
87
+ f401abb6ebbeeaeee501f58ef182a2f0df87303e0eb3e1ef61891a564f4fe4af pytorch_model-00083-of-00243.bin
88
+ 5565664a2f03217978813660e9ca63925e5abae9beef71b2d10772f436cdcb4e pytorch_model-00084-of-00243.bin
89
+ 70bd36a3f6a7f45aae6e9b4b56c5889701533915d10a6b8020c93cb3ed13fb36 pytorch_model-00085-of-00243.bin
90
+ f0d8592e9dbadd34d090b9c309d8d2781695ecf4e6d7005addd25b13630dd611 pytorch_model-00086-of-00243.bin
91
+ 533d20debbd4f0429f12f99ea5ce6a5b88e286af7c58604555ee7bef6c38b286 pytorch_model-00087-of-00243.bin
92
+ 511177391ac994c00b1e39c424f8843f794c0348a9b56418c7c1584bc30d29f4 pytorch_model-00088-of-00243.bin
93
+ 4f4c446298218fd821f5b5b9796b46650badf697fd81fa9d5af3cfa017ea2ede pytorch_model-00089-of-00243.bin
94
+ 69935ce1a0ae99cafd3318adb16cd46005c691214d66b9c1ba2654f04bb6ee48 pytorch_model-00090-of-00243.bin
95
+ f8f6e703981fe6c81d09909ca4730cd50855563fc9c67804923791a19f53fe13 pytorch_model-00091-of-00243.bin
96
+ f4f6c8d6d43d0758840be87cf81ca1936a6ec5e7db567d7e6e28be1482ea3990 pytorch_model-00092-of-00243.bin
97
+ e2f13f781f5bc19ff0774b0d7d37af9ae162c372ec7b57cc6f01768349206f74 pytorch_model-00093-of-00243.bin
98
+ b640b907f0977649c55b5f620cefaeae0e2ab3ffa9d34004a3eaee5e5ad163e3 pytorch_model-00094-of-00243.bin
99
+ 7028c4e1472979988f960b51b7a8872720c12e4ae77bdfc5ed2992448a497419 pytorch_model-00095-of-00243.bin
100
+ 6c2c69a678f85a987e12eca058b60d9712f0c134fe9daf9ba45de5f8ecbf905a pytorch_model-00096-of-00243.bin
101
+ 30747c23a557c14dca4766adedd1555529d1409045f8e8af5d8ba2912b500356 pytorch_model-00097-of-00243.bin
102
+ 97e97a502d9354ecbc2731f93bec4b1d766636fc7b89f2be9749db719ac07400 pytorch_model-00098-of-00243.bin
103
+ 0b7f79b28aeaddb8edcedf9ed249720c05977f13a0f7c27fe410403c0412fc0a pytorch_model-00099-of-00243.bin
104
+ 08ebae4bfebe60c18b235bc0f6e4ee6b1659df839f1c9ddbcfad5e1398d2e923 pytorch_model-00100-of-00243.bin
105
+ 8c8adb82ce7bef686b2cf6b64a207750249d9f725bed0b6b78249cbdd8911e48 pytorch_model-00101-of-00243.bin
106
+ ba24db819cf91d0f2ebc5b1ac07208cbaa71893e1cb6d8f7e0e7543173535946 pytorch_model-00102-of-00243.bin
107
+ 74af29ea6dc24e0a1936ea080f3b70582912e0450c767a30c5360e3c62f143b9 pytorch_model-00103-of-00243.bin
108
+ fe068509fbae8dcfaffd178f8b830a7b3c3ddc839afff0974db4c62a701ff332 pytorch_model-00104-of-00243.bin
109
+ ce9f24ae56b070092e99cc119a5900b1195ad20f6a9a459a3781775f66d2266d pytorch_model-00105-of-00243.bin
110
+ 77e7ae120c4c15a43394dcc82bdcba5fe75c844c39144d6917904111bc7cef6b pytorch_model-00106-of-00243.bin
111
+ dd149a6139ec2067097aebf058b4fd9b7030a49dddb8a66e323fc91000196bd6 pytorch_model-00107-of-00243.bin
112
+ 15bc0b43afa0c818e80835fbf1e8c4213f63c0eeed1f91ed2a0c66bf9e142470 pytorch_model-00108-of-00243.bin
113
+ 09e753f0642a1c904ef732b342894f3bd4d3f68f46989495c477eced1debde94 pytorch_model-00109-of-00243.bin
114
+ f3cf460b6c86e2a35fd60e3715cf9756f54c3db4bb4abc2f07c08b79002b4e19 pytorch_model-00110-of-00243.bin
115
+ 0a353b27752a908fdc4374d041cbb5a6c912a771698d87b0f3c17874b8e488c1 pytorch_model-00111-of-00243.bin
116
+ 63778cc7a8bc989390fd94d668c436b7ad69c86ff2f69d22f73a40bd965d3e1c pytorch_model-00112-of-00243.bin
117
+ 11bc10f9e50374ecd18dec38dc6d0e666fb7548249bb2ec5ed67ff01778c2170 pytorch_model-00113-of-00243.bin
118
+ ca5885767c24be36b32cea538f90d53e05f8450224b0bb788aadd5a6dd89bd34 pytorch_model-00114-of-00243.bin
119
+ 961620e8c9efb910851816c1dd259cb4030f529597ebc830ce539597a384b623 pytorch_model-00115-of-00243.bin
120
+ 51ee0ac8161ccc17facabf1958ed412db37f8f5b0a190f28fe5aa16a23ade55d pytorch_model-00116-of-00243.bin
121
+ ad0f45f599f266e34bfaa8639b1d531744a2d841305805474c45729f0d9b777e pytorch_model-00117-of-00243.bin
122
+ 58ac3d11f22bf981e59e856c6d78c505b0edff247e88323039a697d8d4ad57a4 pytorch_model-00118-of-00243.bin
123
+ 1f5ab60a4639fe55130a4e8c2909c7b2e1411d8f2cbd2b00ec8341f5c9a98f06 pytorch_model-00119-of-00243.bin
124
+ 94b4074f17ba08cfe46399d073badc2457d7ca0e3583ab690074de03ccbb62aa pytorch_model-00120-of-00243.bin
125
+ 0a48a967e3edb8b500eaa12947a32f32f26a24b67e60d3d402afd7481df81516 pytorch_model-00121-of-00243.bin
126
+ fd9eb815daa468dd8be675c7248f2ff4542bbc53984562280181a01ab5a62708 pytorch_model-00122-of-00243.bin
127
+ fffe12cc8e0346cfc8a8230359b8ba7927294408e132218731971d90d5737b41 pytorch_model-00123-of-00243.bin
128
+ 987380de8f157fb20e2e80aba57a8e27716e3729a736decc3758cd0d4dc73e21 pytorch_model-00124-of-00243.bin
129
+ 6ba2b9b8a71af19c38436d7340a58fff7978575fef2deb9f89b0adb9163b2977 pytorch_model-00125-of-00243.bin
130
+ 232f12b40cb60e6c2fb6e55bcba8a855286c16de51d1c7b0fe8c483678c04ff3 pytorch_model-00126-of-00243.bin
131
+ c57a3f4efb3da5660563061253fa71662e6d93e17934588eb3303691a2cb3bf3 pytorch_model-00127-of-00243.bin
132
+ 76b14de090b4bef2c78b6ea383f18f7b6ff5dbd9dccecc55fbdfc90d71d719f3 pytorch_model-00128-of-00243.bin
133
+ 02ff1662e0386bb56929643ff59ef8f5434277e65c5ed78fe6c78cfa68c561b0 pytorch_model-00129-of-00243.bin
134
+ a3e78ee0aa271f0f79b956344faca8e4f161383eb557bd1e8a5c48f26f13c5e6 pytorch_model-00130-of-00243.bin
135
+ 01f54fd5816fc61838ec064b5cafc4c25eaa3b87937ede268b8785c426ccdd7d pytorch_model-00131-of-00243.bin
136
+ 4ecaab016076677b83190110e275271ec4a53b3d736d945170bb1376098701ff pytorch_model-00132-of-00243.bin
137
+ 73d18ed728d97ace5cdd89d9d0c2c76e359ceed3ad05fa03f3038b35d481db27 pytorch_model-00133-of-00243.bin
138
+ 8b8de0770ce14282f96b36f09868aa1b8f1396f2ea2cc683e81d9b7dbdce3799 pytorch_model-00134-of-00243.bin
139
+ 0c2e0f85e72a5e3ae5194a91edfcd141b9ebeb8654cc28769b29e79faaa7f7fe pytorch_model-00135-of-00243.bin
140
+ 33e2206373661164a1c13008c67e2cdce7967368ebd0b2f20377ccda74723656 pytorch_model-00136-of-00243.bin
141
+ 48d5ab98d950b69cb2807c93f97029b2019011cd7bbbc08c9f08d3390ef6f205 pytorch_model-00137-of-00243.bin
142
+ 6435c742d6fdcbbd382d3905e46406d5af16d63d48d323301828ef8a26210fc3 pytorch_model-00138-of-00243.bin
143
+ c4c2593e0dcf66f315079300056c2a3f4b89c0909274bbc520e5f23cb34ad7a3 pytorch_model-00139-of-00243.bin
144
+ da00a68a0900c492dece277a26a9e12fae77bb1edfe32603f12c9be5ca85217a pytorch_model-00140-of-00243.bin
145
+ 35d379ec3521bdfdb1789ad36ea80173053fcc3728c608be8760cea5805a2be2 pytorch_model-00141-of-00243.bin
146
+ 4aba717fa9caa634aee268a2bfef4efc615a363545c23dbb997d12388b12a82d pytorch_model-00142-of-00243.bin
147
+ b51a06c89c8a475845c018beb6520db6ffb4a717b2bb5cb376f4e912dde124c1 pytorch_model-00143-of-00243.bin
148
+ ea46a5b3336bc4eb812bc1bafe1f40371585b40f5632113ca396b1064c38dd30 pytorch_model-00144-of-00243.bin
149
+ b346ac3f89c61c6083439c3757cedec730b672c18a35c679e08547ab873f4ecb pytorch_model-00145-of-00243.bin
150
+ cdf36f5445079764746a016d559d4700352e6a4e4d252d19b9d3e4b9f3552b14 pytorch_model-00146-of-00243.bin
151
+ 6ab71611a456425ec1dee8a49f31621edb18701d876531c63ec7f19032ebaab9 pytorch_model-00147-of-00243.bin
152
+ 530f68272df243bb3970dad56606c4f657706dfece505a7f97c4987d94c8691c pytorch_model-00148-of-00243.bin
153
+ 3b9671d920c9506367ff83f6aa8e20c9cc3bfc955d6235e8291f200144ce1d2e pytorch_model-00149-of-00243.bin
154
+ cb244fe146c2e37c1a69cb2b30bf11da20082b3cad8b195e000f224764425b9a pytorch_model-00150-of-00243.bin
155
+ 4f0bbcaabbc664e59d3a96572d0d746c2c5017c089d80ab1d6cb66047f86313e pytorch_model-00151-of-00243.bin
156
+ 0f07bf8e97696a087a4e03db68a497ee7f8e44b8fe285717372b80e966ef6908 pytorch_model-00152-of-00243.bin
157
+ 66ab23c030464f4b6885d54eee5189819f0399b7ce1d65cba2872f945c444c0b pytorch_model-00153-of-00243.bin
158
+ d10391a3fce4e7ef21a5fd041f3553b9b2b5e35c273dae79915bbec065fef613 pytorch_model-00154-of-00243.bin
159
+ 30484b1ffc896bfa63bcd3525931821e7b984e3ceaae3a42c64d13ff92349a4e pytorch_model-00155-of-00243.bin
160
+ b3fd3b87151dfe5ab73598e6e48ae3331c6108e1ebbad5eecade29a6bbb314fc pytorch_model-00156-of-00243.bin
161
+ 6a03cf30b4a4ad61a6385c6d9a169643edbfaa949459684c61f446ae361239f3 pytorch_model-00157-of-00243.bin
162
+ 311fa925f07530e5cebcde5059bdb845db52b7f8da00623f48dc7da705a6a2df pytorch_model-00158-of-00243.bin
163
+ 77545e68e432bacfb0bec0bc9476ad3375cfc1827da8efa9b353f9fbc758a8b8 pytorch_model-00159-of-00243.bin
164
+ 74adb9c4554adc71b6eedc3b5a8b141b3613c1177ef0adb5bfaeeaa42bbd85a9 pytorch_model-00160-of-00243.bin
165
+ 01b5812179799a61884ee85985ef8d8ecd00037eda2fad672dfdc2878b18a46d pytorch_model-00161-of-00243.bin
166
+ 9650fadc97c002bbcdec1154bdd422fe12d18c61b9441ed74a1b68f545c2788b pytorch_model-00162-of-00243.bin
167
+ 788094824dd1823f531e8cec250979556a72d22ba9d1182dbfcd3db429214b89 pytorch_model-00163-of-00243.bin
168
+ d755e0f69d15678a2b361212ab3d1780bf975da6d33360a5d3ed96389a02f039 pytorch_model-00164-of-00243.bin
169
+ 5694a4c2b11ed581cdaea73cef531c26c987eedd59c221be052e8f3e443e8cfe pytorch_model-00165-of-00243.bin
170
+ 6b0ca82e0a531a4a8f03b9ed6c5fa0c850ec1b3fd5d5f54a2c85f0e553355326 pytorch_model-00166-of-00243.bin
171
+ bb9afe8715554aa01b606a0b23f231b25c2762c99a2e59ca503d0e1d8b4f6a8a pytorch_model-00167-of-00243.bin
172
+ 1c131a096d7d7363f475e6367480f5665c87111b6c58ba925474ce2ffc123799 pytorch_model-00168-of-00243.bin
173
+ eeaf8e409e220b251a5d9e3fac2735c13668e5da016d7c317677d4f74fe12b00 pytorch_model-00169-of-00243.bin
174
+ 08c43e5cc128bfbbf6dd8f30e6d5bcc616c2054f341062e562927acb99922bf1 pytorch_model-00170-of-00243.bin
175
+ 8f5b277bca1c0e05c0bd7601f4d3301d6856ff4fde62daa17a81707f60928ec5 pytorch_model-00171-of-00243.bin
176
+ 30999caa7ced67cad5a441e3719f4ae4435a2c5591e17b08c0c2f07a18c5df5d pytorch_model-00172-of-00243.bin
177
+ dab1edecb2176f68468d683afe638b840f970120f9610a831c4bb5966b2f491a pytorch_model-00173-of-00243.bin
178
+ c82cee7ac9ecd9f6f071c244f4cf6e6e8459f8d4bb22fa38c00087195add4e6c pytorch_model-00174-of-00243.bin
179
+ 37213a8bb345cd94e10963669f683a8c1a2378b6ff8374d863f2a09b2b0e05a1 pytorch_model-00175-of-00243.bin
180
+ 7e80625d461a8818933861646a94e7a0d29c9ce532debea258caafb10dc4c09f pytorch_model-00176-of-00243.bin
181
+ 4cd85828447cce382ec197771c5a4417adb37262c49c60c4440262e3c3fb45dc pytorch_model-00177-of-00243.bin
182
+ 8c3622803909e984166c781aca5033d9fc751b76873552cdc21ec6101a371fca pytorch_model-00178-of-00243.bin
183
+ f7c185f4628d8ec71522b8ba2d2478d4d65d8a1de2b0f0854974929594a0cf23 pytorch_model-00179-of-00243.bin
184
+ a6269af0e3be281ad0b2e77fc064eadf56ff0ea50f2a1297ad93db9a76b202af pytorch_model-00180-of-00243.bin
185
+ b1a21ae3f3d8883734b3e8ec37eac5cc79c16a13c96e8ab945548356d0d3e447 pytorch_model-00181-of-00243.bin
186
+ 65dfebebecac857d45e49f881b44fe52f38283da648c67b68cf8f34aaa9d7388 pytorch_model-00182-of-00243.bin
187
+ 24dfe8116221ae9cb90ab393471d34d30f84daff0ca1632c3fc705d6c8355ba0 pytorch_model-00183-of-00243.bin
188
+ a02d1a5570ec7831e5c56a4e8fa9813d98415dd406f4c2fd48a3c885bdb3187b pytorch_model-00184-of-00243.bin
189
+ 03340f30578794690dc58b97d17879f04219034f5466c4a45e3951f857c79fa1 pytorch_model-00185-of-00243.bin
190
+ 7f5bfcc28404d85f0dffcc3ef0b014585307ade9010a004b509772b8266ec74f pytorch_model-00186-of-00243.bin
191
+ b3844cf4044bb1ffb02d8e9b7b1b7fd4b4c124b010c04c1f8ff8a78e03ff320f pytorch_model-00187-of-00243.bin
192
+ 4089b0dbd91b0f2b10e7dd5fcbca4dde9faffafd661695c2da5beee8370072a3 pytorch_model-00188-of-00243.bin
193
+ 1666768c16bfc76a11fedc3525358b9e429ae50c115b4f706960d65910715543 pytorch_model-00189-of-00243.bin
194
+ 1affffa5fccbcc2166f666eeed7aced8d38743574ea7b6201fcd149272bb6246 pytorch_model-00190-of-00243.bin
195
+ ad41ac209d56068a6078429b6b18135176a9ca4ec4c24568094e1eb9ebce0f60 pytorch_model-00191-of-00243.bin
196
+ 7ca1cab9541d33f8cbb692b5f319698c44205da210e3cdc673c9c0a2b7104b43 pytorch_model-00192-of-00243.bin
197
+ 46d2a9a508115d14c4b4f9580a99f2b83b24222b6da6836d293f35885b610875 pytorch_model-00193-of-00243.bin
198
+ 174844ba09a4a7471bba31188e01397be6c40e050ea685eaaa6bae22f2bb11c4 pytorch_model-00194-of-00243.bin
199
+ 00f948d55679a438369d81791520688569c94a1b3621cb10d4454638fbcbe4db pytorch_model-00195-of-00243.bin
200
+ 2faff3cfe1c5d530d6665a4063be2f7d26e05f0c6904cbbd7d091905d973ff79 pytorch_model-00196-of-00243.bin
201
+ 895b6d403638393612d7d0f5cb963c80631c153f03c3d4966869057630c6f94e pytorch_model-00197-of-00243.bin
202
+ 7dc6f3b6c479c1e775a810637d02051c8e448a8c4f6a9cd26788242dc6e515ac pytorch_model-00198-of-00243.bin
203
+ e0700fd25fe1be375cf7df82194120c17a154ad5c57ebfb9975f42db0aef5f48 pytorch_model-00199-of-00243.bin
204
+ 43ea69f26fa3fc99f36a789e84022ff32a61ba3be9dac4b92e007b7361d843ca pytorch_model-00200-of-00243.bin
205
+ f97ffd772f8a328990d30289216d6028e6f544790af0cb81afd60205484b1c46 pytorch_model-00201-of-00243.bin
206
+ e1591dea37d51f02161ec7296a0787cac06d73decf4709a2ec02363ed937ddd9 pytorch_model-00202-of-00243.bin
207
+ ca9e94ee2efbcc51fe478068340599fca381ec4eaa40892e095ace61509e5bde pytorch_model-00203-of-00243.bin
208
+ 37564c62aa89a31a666c3ee10dfb69bb88d6b07c2a99c0b8c16cd56913d0faf4 pytorch_model-00204-of-00243.bin
209
+ 214824198bc777c3b026b7a164678aa7ed34a597db6fd033edaa94358395f8ae pytorch_model-00205-of-00243.bin
210
+ cee45752533144a37844d8eda345b613dae3f94d9da2f98f768dc3065f157a87 pytorch_model-00206-of-00243.bin
211
+ c9721ed5865c998b36f622f5a2c5b135fb7cd5d5e989f0ad8b6d2e9a36bce3b5 pytorch_model-00207-of-00243.bin
212
+ fd1323c856b7c584a0877890cd80cbddd449f4071a1180de51ac7441df33cfaf pytorch_model-00208-of-00243.bin
213
+ b67108c6c45e1c0426a2e890f6e15f63e5fbc9dc03b033f613b42bf7d86279ff pytorch_model-00209-of-00243.bin
214
+ 9a3486126b67295c5533a653e354ff25b5324b8339f9d815f7ebfbca791f2378 pytorch_model-00210-of-00243.bin
215
+ 5b624bbf42939baf3ba4d7ba25672ae363e7932d8df18b411321ea9dbab3d6c4 pytorch_model-00211-of-00243.bin
216
+ 5005a92e9e6ccac697122bc5b8b9a0a21b4ca55e97fbe4018cfdc5fb5e869a06 pytorch_model-00212-of-00243.bin
217
+ ff9154478a5fd9a3b2cf5266a302497da01a538b28fbb62181d45c92d2c6c2ae pytorch_model-00213-of-00243.bin
218
+ 4659295ce30f7ddb2c9673cf7d86280fa74869b1a08352f38406c1fbe56f51f5 pytorch_model-00214-of-00243.bin
219
+ 794d1149608c2bb5b0007749e2474fdf7e61c150693310a1918163dbd4babfad pytorch_model-00215-of-00243.bin
220
+ 8132b955c65d85dd5d15662ec17484143f683b78ca0af405a24725fd9393baeb pytorch_model-00216-of-00243.bin
221
+ 640c7e2e8dca0350525ab5a51870f309b9b0203e72884be1034ef2ce74481e6e pytorch_model-00217-of-00243.bin
222
+ e7007217a8100f86b12894c49bac1106af5e565b03e1741b277dd02a64374fb0 pytorch_model-00218-of-00243.bin
223
+ c62bd6c10e27bf8d8caf92f3f13b26d92487f66a3e62f4af9e49c7eba63167b8 pytorch_model-00219-of-00243.bin
224
+ 1e657f439b4a2db41cdaeadcdb88f5eb48a8cd726ac460da9cdefd0502b3eb9b pytorch_model-00220-of-00243.bin
225
+ de028b5f2a19b85c5be2d0654ae2337516220ec27669ffc8ef17c43104bb3224 pytorch_model-00221-of-00243.bin
226
+ 3ad8fc90c1fb9bb43cfe5b2401bfb7b290ce8c704bbc5515f1ce4a5873f95139 pytorch_model-00222-of-00243.bin
227
+ 9b989a86fa6dcf0436cc974981f6d3c8811132e92a26166a25ea457c72b3db06 pytorch_model-00223-of-00243.bin
228
+ aca6597466a65388286be22891e5474fd1e79eda6b2b1b3f968163a7b1a4393d pytorch_model-00224-of-00243.bin
229
+ fd0f867516832221d24705969d3593e5a1811ee761b8fbf7c57bcc1ad1dfc8a5 pytorch_model-00225-of-00243.bin
230
+ 6794576954ec8970ff09499a7c0de486532326826f850bbdbf2e3fa815f69e1e pytorch_model-00226-of-00243.bin
231
+ 4f7cef542e67021df4ff812429ea0491ef967122d94afa3cdd83b41181fa3c75 pytorch_model-00227-of-00243.bin
232
+ fee6c54c08d1efabc58eaa318651a625199cea878e49dd9257ec8d493009a779 pytorch_model-00228-of-00243.bin
233
+ ae514ec50b760014c719a16b9509b5cb3b0d9f82dfd4d489cc3d3484e8d6a9a7 pytorch_model-00229-of-00243.bin
234
+ 6d4b7fb5ec49a09cafc0977126e0d22a7acd1f029dff96efe130c665373619f6 pytorch_model-00230-of-00243.bin
235
+ fcfa6ba6f24e9ae498769fd4a3e20c2f183ea290640de2f205a378d50c44c457 pytorch_model-00231-of-00243.bin
236
+ 9f1994823e8a45ccdd4030a3abbbed28026a08fa5908ccacf7cb0f9976ad8976 pytorch_model-00232-of-00243.bin
237
+ cf64ed719e75ab225838749b8885e00c23b280a1c750b9f695aac00ad89152aa pytorch_model-00233-of-00243.bin
238
+ befc340950c3750e9fb79be5f7a3ba7ecb5bcc5422bd327e9309fb440dbbb5be pytorch_model-00234-of-00243.bin
239
+ abc57ae3f55a7ee8fb3fac3e36531b1e577fb4916a6b1c3fa88d7ce9969aed5f pytorch_model-00235-of-00243.bin
240
+ 29bb9ec1e0db7a8d1dba82cfc430a6d8a6f1620dd0006a36de3a8f90cb943bac pytorch_model-00236-of-00243.bin
241
+ 14d20d28ee07dbaba393a8d82aa17683bc31d988fb4f22d261336dd5096ae71c pytorch_model-00237-of-00243.bin
242
+ 58df1115168fe8d60dffd81f7e0a2d5b0997601ca73f75cfab9ba128a3a3d496 pytorch_model-00238-of-00243.bin
243
+ 7f42e2a869414c1cba14c3bb0b3e72fabc06a8a77144dac233c366c1883073b8 pytorch_model-00239-of-00243.bin
244
+ a4faa71e384023a2e1e21609bf5b6967925df8f2b64d825183fad2fb2418985a pytorch_model-00240-of-00243.bin
245
+ 530296bfc60095c9f4840776a6ee24558efddcf19bd1a7ebe42895c02d7e912c pytorch_model-00241-of-00243.bin
246
+ 8bc75fb2c6486f123c89701022ed29145c1246fe109a3defa5d470b35074a988 pytorch_model-00242-of-00243.bin
247
+ 39f587e500e18a9db4797a87945ef84b9dc7bb989506f7d322144b270918358c pytorch_model-00243-of-00243.bin
248
+ 9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347 tokenizer.model
pytorch_model.bin.index.json ADDED
@@ -0,0 +1,610 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 65057902592
4
+ },
5
+ "weight_map": {
6
+ "lm_head.weight": "pytorch_model-00243-of-00243.bin",
7
+ "model.embed_tokens.weight": "pytorch_model-00002-of-00243.bin",
8
+ "model.layers.0.input_layernorm.weight": "pytorch_model-00006-of-00243.bin",
9
+ "model.layers.0.mlp.down_proj.weight": "pytorch_model-00005-of-00243.bin",
10
+ "model.layers.0.mlp.gate_proj.weight": "pytorch_model-00004-of-00243.bin",
11
+ "model.layers.0.mlp.up_proj.weight": "pytorch_model-00006-of-00243.bin",
12
+ "model.layers.0.post_attention_layernorm.weight": "pytorch_model-00006-of-00243.bin",
13
+ "model.layers.0.self_attn.k_proj.weight": "pytorch_model-00003-of-00243.bin",
14
+ "model.layers.0.self_attn.o_proj.weight": "pytorch_model-00003-of-00243.bin",
15
+ "model.layers.0.self_attn.q_proj.weight": "pytorch_model-00003-of-00243.bin",
16
+ "model.layers.0.self_attn.rotary_emb.inv_freq": "pytorch_model-00003-of-00243.bin",
17
+ "model.layers.0.self_attn.v_proj.weight": "pytorch_model-00003-of-00243.bin",
18
+ "model.layers.1.input_layernorm.weight": "pytorch_model-00010-of-00243.bin",
19
+ "model.layers.1.mlp.down_proj.weight": "pytorch_model-00009-of-00243.bin",
20
+ "model.layers.1.mlp.gate_proj.weight": "pytorch_model-00008-of-00243.bin",
21
+ "model.layers.1.mlp.up_proj.weight": "pytorch_model-00010-of-00243.bin",
22
+ "model.layers.1.post_attention_layernorm.weight": "pytorch_model-00010-of-00243.bin",
23
+ "model.layers.1.self_attn.k_proj.weight": "pytorch_model-00007-of-00243.bin",
24
+ "model.layers.1.self_attn.o_proj.weight": "pytorch_model-00007-of-00243.bin",
25
+ "model.layers.1.self_attn.q_proj.weight": "pytorch_model-00006-of-00243.bin",
26
+ "model.layers.1.self_attn.rotary_emb.inv_freq": "pytorch_model-00007-of-00243.bin",
27
+ "model.layers.1.self_attn.v_proj.weight": "pytorch_model-00007-of-00243.bin",
28
+ "model.layers.10.input_layernorm.weight": "pytorch_model-00046-of-00243.bin",
29
+ "model.layers.10.mlp.down_proj.weight": "pytorch_model-00045-of-00243.bin",
30
+ "model.layers.10.mlp.gate_proj.weight": "pytorch_model-00044-of-00243.bin",
31
+ "model.layers.10.mlp.up_proj.weight": "pytorch_model-00046-of-00243.bin",
32
+ "model.layers.10.post_attention_layernorm.weight": "pytorch_model-00046-of-00243.bin",
33
+ "model.layers.10.self_attn.k_proj.weight": "pytorch_model-00043-of-00243.bin",
34
+ "model.layers.10.self_attn.o_proj.weight": "pytorch_model-00043-of-00243.bin",
35
+ "model.layers.10.self_attn.q_proj.weight": "pytorch_model-00042-of-00243.bin",
36
+ "model.layers.10.self_attn.rotary_emb.inv_freq": "pytorch_model-00043-of-00243.bin",
37
+ "model.layers.10.self_attn.v_proj.weight": "pytorch_model-00043-of-00243.bin",
38
+ "model.layers.11.input_layernorm.weight": "pytorch_model-00050-of-00243.bin",
39
+ "model.layers.11.mlp.down_proj.weight": "pytorch_model-00049-of-00243.bin",
40
+ "model.layers.11.mlp.gate_proj.weight": "pytorch_model-00048-of-00243.bin",
41
+ "model.layers.11.mlp.up_proj.weight": "pytorch_model-00050-of-00243.bin",
42
+ "model.layers.11.post_attention_layernorm.weight": "pytorch_model-00050-of-00243.bin",
43
+ "model.layers.11.self_attn.k_proj.weight": "pytorch_model-00047-of-00243.bin",
44
+ "model.layers.11.self_attn.o_proj.weight": "pytorch_model-00047-of-00243.bin",
45
+ "model.layers.11.self_attn.q_proj.weight": "pytorch_model-00046-of-00243.bin",
46
+ "model.layers.11.self_attn.rotary_emb.inv_freq": "pytorch_model-00047-of-00243.bin",
47
+ "model.layers.11.self_attn.v_proj.weight": "pytorch_model-00047-of-00243.bin",
48
+ "model.layers.12.input_layernorm.weight": "pytorch_model-00054-of-00243.bin",
49
+ "model.layers.12.mlp.down_proj.weight": "pytorch_model-00053-of-00243.bin",
50
+ "model.layers.12.mlp.gate_proj.weight": "pytorch_model-00052-of-00243.bin",
51
+ "model.layers.12.mlp.up_proj.weight": "pytorch_model-00054-of-00243.bin",
52
+ "model.layers.12.post_attention_layernorm.weight": "pytorch_model-00054-of-00243.bin",
53
+ "model.layers.12.self_attn.k_proj.weight": "pytorch_model-00051-of-00243.bin",
54
+ "model.layers.12.self_attn.o_proj.weight": "pytorch_model-00051-of-00243.bin",
55
+ "model.layers.12.self_attn.q_proj.weight": "pytorch_model-00050-of-00243.bin",
56
+ "model.layers.12.self_attn.rotary_emb.inv_freq": "pytorch_model-00051-of-00243.bin",
57
+ "model.layers.12.self_attn.v_proj.weight": "pytorch_model-00051-of-00243.bin",
58
+ "model.layers.13.input_layernorm.weight": "pytorch_model-00058-of-00243.bin",
59
+ "model.layers.13.mlp.down_proj.weight": "pytorch_model-00057-of-00243.bin",
60
+ "model.layers.13.mlp.gate_proj.weight": "pytorch_model-00056-of-00243.bin",
61
+ "model.layers.13.mlp.up_proj.weight": "pytorch_model-00058-of-00243.bin",
62
+ "model.layers.13.post_attention_layernorm.weight": "pytorch_model-00058-of-00243.bin",
63
+ "model.layers.13.self_attn.k_proj.weight": "pytorch_model-00055-of-00243.bin",
64
+ "model.layers.13.self_attn.o_proj.weight": "pytorch_model-00055-of-00243.bin",
65
+ "model.layers.13.self_attn.q_proj.weight": "pytorch_model-00054-of-00243.bin",
66
+ "model.layers.13.self_attn.rotary_emb.inv_freq": "pytorch_model-00055-of-00243.bin",
67
+ "model.layers.13.self_attn.v_proj.weight": "pytorch_model-00055-of-00243.bin",
68
+ "model.layers.14.input_layernorm.weight": "pytorch_model-00062-of-00243.bin",
69
+ "model.layers.14.mlp.down_proj.weight": "pytorch_model-00061-of-00243.bin",
70
+ "model.layers.14.mlp.gate_proj.weight": "pytorch_model-00060-of-00243.bin",
71
+ "model.layers.14.mlp.up_proj.weight": "pytorch_model-00062-of-00243.bin",
72
+ "model.layers.14.post_attention_layernorm.weight": "pytorch_model-00062-of-00243.bin",
73
+ "model.layers.14.self_attn.k_proj.weight": "pytorch_model-00059-of-00243.bin",
74
+ "model.layers.14.self_attn.o_proj.weight": "pytorch_model-00059-of-00243.bin",
75
+ "model.layers.14.self_attn.q_proj.weight": "pytorch_model-00058-of-00243.bin",
76
+ "model.layers.14.self_attn.rotary_emb.inv_freq": "pytorch_model-00059-of-00243.bin",
77
+ "model.layers.14.self_attn.v_proj.weight": "pytorch_model-00059-of-00243.bin",
78
+ "model.layers.15.input_layernorm.weight": "pytorch_model-00066-of-00243.bin",
79
+ "model.layers.15.mlp.down_proj.weight": "pytorch_model-00065-of-00243.bin",
80
+ "model.layers.15.mlp.gate_proj.weight": "pytorch_model-00064-of-00243.bin",
81
+ "model.layers.15.mlp.up_proj.weight": "pytorch_model-00066-of-00243.bin",
82
+ "model.layers.15.post_attention_layernorm.weight": "pytorch_model-00066-of-00243.bin",
83
+ "model.layers.15.self_attn.k_proj.weight": "pytorch_model-00063-of-00243.bin",
84
+ "model.layers.15.self_attn.o_proj.weight": "pytorch_model-00063-of-00243.bin",
85
+ "model.layers.15.self_attn.q_proj.weight": "pytorch_model-00062-of-00243.bin",
86
+ "model.layers.15.self_attn.rotary_emb.inv_freq": "pytorch_model-00063-of-00243.bin",
87
+ "model.layers.15.self_attn.v_proj.weight": "pytorch_model-00063-of-00243.bin",
88
+ "model.layers.16.input_layernorm.weight": "pytorch_model-00070-of-00243.bin",
89
+ "model.layers.16.mlp.down_proj.weight": "pytorch_model-00069-of-00243.bin",
90
+ "model.layers.16.mlp.gate_proj.weight": "pytorch_model-00068-of-00243.bin",
91
+ "model.layers.16.mlp.up_proj.weight": "pytorch_model-00070-of-00243.bin",
92
+ "model.layers.16.post_attention_layernorm.weight": "pytorch_model-00070-of-00243.bin",
93
+ "model.layers.16.self_attn.k_proj.weight": "pytorch_model-00067-of-00243.bin",
94
+ "model.layers.16.self_attn.o_proj.weight": "pytorch_model-00067-of-00243.bin",
95
+ "model.layers.16.self_attn.q_proj.weight": "pytorch_model-00066-of-00243.bin",
96
+ "model.layers.16.self_attn.rotary_emb.inv_freq": "pytorch_model-00067-of-00243.bin",
97
+ "model.layers.16.self_attn.v_proj.weight": "pytorch_model-00067-of-00243.bin",
98
+ "model.layers.17.input_layernorm.weight": "pytorch_model-00074-of-00243.bin",
99
+ "model.layers.17.mlp.down_proj.weight": "pytorch_model-00073-of-00243.bin",
100
+ "model.layers.17.mlp.gate_proj.weight": "pytorch_model-00072-of-00243.bin",
101
+ "model.layers.17.mlp.up_proj.weight": "pytorch_model-00074-of-00243.bin",
102
+ "model.layers.17.post_attention_layernorm.weight": "pytorch_model-00074-of-00243.bin",
103
+ "model.layers.17.self_attn.k_proj.weight": "pytorch_model-00071-of-00243.bin",
104
+ "model.layers.17.self_attn.o_proj.weight": "pytorch_model-00071-of-00243.bin",
105
+ "model.layers.17.self_attn.q_proj.weight": "pytorch_model-00070-of-00243.bin",
106
+ "model.layers.17.self_attn.rotary_emb.inv_freq": "pytorch_model-00071-of-00243.bin",
107
+ "model.layers.17.self_attn.v_proj.weight": "pytorch_model-00071-of-00243.bin",
108
+ "model.layers.18.input_layernorm.weight": "pytorch_model-00078-of-00243.bin",
109
+ "model.layers.18.mlp.down_proj.weight": "pytorch_model-00077-of-00243.bin",
110
+ "model.layers.18.mlp.gate_proj.weight": "pytorch_model-00076-of-00243.bin",
111
+ "model.layers.18.mlp.up_proj.weight": "pytorch_model-00078-of-00243.bin",
112
+ "model.layers.18.post_attention_layernorm.weight": "pytorch_model-00078-of-00243.bin",
113
+ "model.layers.18.self_attn.k_proj.weight": "pytorch_model-00075-of-00243.bin",
114
+ "model.layers.18.self_attn.o_proj.weight": "pytorch_model-00075-of-00243.bin",
115
+ "model.layers.18.self_attn.q_proj.weight": "pytorch_model-00074-of-00243.bin",
116
+ "model.layers.18.self_attn.rotary_emb.inv_freq": "pytorch_model-00075-of-00243.bin",
117
+ "model.layers.18.self_attn.v_proj.weight": "pytorch_model-00075-of-00243.bin",
118
+ "model.layers.19.input_layernorm.weight": "pytorch_model-00082-of-00243.bin",
119
+ "model.layers.19.mlp.down_proj.weight": "pytorch_model-00081-of-00243.bin",
120
+ "model.layers.19.mlp.gate_proj.weight": "pytorch_model-00080-of-00243.bin",
121
+ "model.layers.19.mlp.up_proj.weight": "pytorch_model-00082-of-00243.bin",
122
+ "model.layers.19.post_attention_layernorm.weight": "pytorch_model-00082-of-00243.bin",
123
+ "model.layers.19.self_attn.k_proj.weight": "pytorch_model-00079-of-00243.bin",
124
+ "model.layers.19.self_attn.o_proj.weight": "pytorch_model-00079-of-00243.bin",
125
+ "model.layers.19.self_attn.q_proj.weight": "pytorch_model-00078-of-00243.bin",
126
+ "model.layers.19.self_attn.rotary_emb.inv_freq": "pytorch_model-00079-of-00243.bin",
127
+ "model.layers.19.self_attn.v_proj.weight": "pytorch_model-00079-of-00243.bin",
128
+ "model.layers.2.input_layernorm.weight": "pytorch_model-00014-of-00243.bin",
129
+ "model.layers.2.mlp.down_proj.weight": "pytorch_model-00013-of-00243.bin",
130
+ "model.layers.2.mlp.gate_proj.weight": "pytorch_model-00012-of-00243.bin",
131
+ "model.layers.2.mlp.up_proj.weight": "pytorch_model-00014-of-00243.bin",
132
+ "model.layers.2.post_attention_layernorm.weight": "pytorch_model-00014-of-00243.bin",
133
+ "model.layers.2.self_attn.k_proj.weight": "pytorch_model-00011-of-00243.bin",
134
+ "model.layers.2.self_attn.o_proj.weight": "pytorch_model-00011-of-00243.bin",
135
+ "model.layers.2.self_attn.q_proj.weight": "pytorch_model-00010-of-00243.bin",
136
+ "model.layers.2.self_attn.rotary_emb.inv_freq": "pytorch_model-00011-of-00243.bin",
137
+ "model.layers.2.self_attn.v_proj.weight": "pytorch_model-00011-of-00243.bin",
138
+ "model.layers.20.input_layernorm.weight": "pytorch_model-00086-of-00243.bin",
139
+ "model.layers.20.mlp.down_proj.weight": "pytorch_model-00085-of-00243.bin",
140
+ "model.layers.20.mlp.gate_proj.weight": "pytorch_model-00084-of-00243.bin",
141
+ "model.layers.20.mlp.up_proj.weight": "pytorch_model-00086-of-00243.bin",
142
+ "model.layers.20.post_attention_layernorm.weight": "pytorch_model-00086-of-00243.bin",
143
+ "model.layers.20.self_attn.k_proj.weight": "pytorch_model-00083-of-00243.bin",
144
+ "model.layers.20.self_attn.o_proj.weight": "pytorch_model-00083-of-00243.bin",
145
+ "model.layers.20.self_attn.q_proj.weight": "pytorch_model-00082-of-00243.bin",
146
+ "model.layers.20.self_attn.rotary_emb.inv_freq": "pytorch_model-00083-of-00243.bin",
147
+ "model.layers.20.self_attn.v_proj.weight": "pytorch_model-00083-of-00243.bin",
148
+ "model.layers.21.input_layernorm.weight": "pytorch_model-00090-of-00243.bin",
149
+ "model.layers.21.mlp.down_proj.weight": "pytorch_model-00089-of-00243.bin",
150
+ "model.layers.21.mlp.gate_proj.weight": "pytorch_model-00088-of-00243.bin",
151
+ "model.layers.21.mlp.up_proj.weight": "pytorch_model-00090-of-00243.bin",
152
+ "model.layers.21.post_attention_layernorm.weight": "pytorch_model-00090-of-00243.bin",
153
+ "model.layers.21.self_attn.k_proj.weight": "pytorch_model-00087-of-00243.bin",
154
+ "model.layers.21.self_attn.o_proj.weight": "pytorch_model-00087-of-00243.bin",
155
+ "model.layers.21.self_attn.q_proj.weight": "pytorch_model-00086-of-00243.bin",
156
+ "model.layers.21.self_attn.rotary_emb.inv_freq": "pytorch_model-00087-of-00243.bin",
157
+ "model.layers.21.self_attn.v_proj.weight": "pytorch_model-00087-of-00243.bin",
158
+ "model.layers.22.input_layernorm.weight": "pytorch_model-00094-of-00243.bin",
159
+ "model.layers.22.mlp.down_proj.weight": "pytorch_model-00093-of-00243.bin",
160
+ "model.layers.22.mlp.gate_proj.weight": "pytorch_model-00092-of-00243.bin",
161
+ "model.layers.22.mlp.up_proj.weight": "pytorch_model-00094-of-00243.bin",
162
+ "model.layers.22.post_attention_layernorm.weight": "pytorch_model-00094-of-00243.bin",
163
+ "model.layers.22.self_attn.k_proj.weight": "pytorch_model-00091-of-00243.bin",
164
+ "model.layers.22.self_attn.o_proj.weight": "pytorch_model-00091-of-00243.bin",
165
+ "model.layers.22.self_attn.q_proj.weight": "pytorch_model-00090-of-00243.bin",
166
+ "model.layers.22.self_attn.rotary_emb.inv_freq": "pytorch_model-00091-of-00243.bin",
167
+ "model.layers.22.self_attn.v_proj.weight": "pytorch_model-00091-of-00243.bin",
168
+ "model.layers.23.input_layernorm.weight": "pytorch_model-00098-of-00243.bin",
169
+ "model.layers.23.mlp.down_proj.weight": "pytorch_model-00097-of-00243.bin",
170
+ "model.layers.23.mlp.gate_proj.weight": "pytorch_model-00096-of-00243.bin",
171
+ "model.layers.23.mlp.up_proj.weight": "pytorch_model-00098-of-00243.bin",
172
+ "model.layers.23.post_attention_layernorm.weight": "pytorch_model-00098-of-00243.bin",
173
+ "model.layers.23.self_attn.k_proj.weight": "pytorch_model-00095-of-00243.bin",
174
+ "model.layers.23.self_attn.o_proj.weight": "pytorch_model-00095-of-00243.bin",
175
+ "model.layers.23.self_attn.q_proj.weight": "pytorch_model-00094-of-00243.bin",
176
+ "model.layers.23.self_attn.rotary_emb.inv_freq": "pytorch_model-00095-of-00243.bin",
177
+ "model.layers.23.self_attn.v_proj.weight": "pytorch_model-00095-of-00243.bin",
178
+ "model.layers.24.input_layernorm.weight": "pytorch_model-00102-of-00243.bin",
179
+ "model.layers.24.mlp.down_proj.weight": "pytorch_model-00101-of-00243.bin",
180
+ "model.layers.24.mlp.gate_proj.weight": "pytorch_model-00100-of-00243.bin",
181
+ "model.layers.24.mlp.up_proj.weight": "pytorch_model-00102-of-00243.bin",
182
+ "model.layers.24.post_attention_layernorm.weight": "pytorch_model-00102-of-00243.bin",
183
+ "model.layers.24.self_attn.k_proj.weight": "pytorch_model-00099-of-00243.bin",
184
+ "model.layers.24.self_attn.o_proj.weight": "pytorch_model-00099-of-00243.bin",
185
+ "model.layers.24.self_attn.q_proj.weight": "pytorch_model-00098-of-00243.bin",
186
+ "model.layers.24.self_attn.rotary_emb.inv_freq": "pytorch_model-00099-of-00243.bin",
187
+ "model.layers.24.self_attn.v_proj.weight": "pytorch_model-00099-of-00243.bin",
188
+ "model.layers.25.input_layernorm.weight": "pytorch_model-00106-of-00243.bin",
189
+ "model.layers.25.mlp.down_proj.weight": "pytorch_model-00105-of-00243.bin",
190
+ "model.layers.25.mlp.gate_proj.weight": "pytorch_model-00104-of-00243.bin",
191
+ "model.layers.25.mlp.up_proj.weight": "pytorch_model-00106-of-00243.bin",
192
+ "model.layers.25.post_attention_layernorm.weight": "pytorch_model-00106-of-00243.bin",
193
+ "model.layers.25.self_attn.k_proj.weight": "pytorch_model-00103-of-00243.bin",
194
+ "model.layers.25.self_attn.o_proj.weight": "pytorch_model-00103-of-00243.bin",
195
+ "model.layers.25.self_attn.q_proj.weight": "pytorch_model-00102-of-00243.bin",
196
+ "model.layers.25.self_attn.rotary_emb.inv_freq": "pytorch_model-00103-of-00243.bin",
197
+ "model.layers.25.self_attn.v_proj.weight": "pytorch_model-00103-of-00243.bin",
198
+ "model.layers.26.input_layernorm.weight": "pytorch_model-00110-of-00243.bin",
199
+ "model.layers.26.mlp.down_proj.weight": "pytorch_model-00109-of-00243.bin",
200
+ "model.layers.26.mlp.gate_proj.weight": "pytorch_model-00108-of-00243.bin",
201
+ "model.layers.26.mlp.up_proj.weight": "pytorch_model-00110-of-00243.bin",
202
+ "model.layers.26.post_attention_layernorm.weight": "pytorch_model-00110-of-00243.bin",
203
+ "model.layers.26.self_attn.k_proj.weight": "pytorch_model-00107-of-00243.bin",
204
+ "model.layers.26.self_attn.o_proj.weight": "pytorch_model-00107-of-00243.bin",
205
+ "model.layers.26.self_attn.q_proj.weight": "pytorch_model-00106-of-00243.bin",
206
+ "model.layers.26.self_attn.rotary_emb.inv_freq": "pytorch_model-00107-of-00243.bin",
207
+ "model.layers.26.self_attn.v_proj.weight": "pytorch_model-00107-of-00243.bin",
208
+ "model.layers.27.input_layernorm.weight": "pytorch_model-00114-of-00243.bin",
209
+ "model.layers.27.mlp.down_proj.weight": "pytorch_model-00113-of-00243.bin",
210
+ "model.layers.27.mlp.gate_proj.weight": "pytorch_model-00112-of-00243.bin",
211
+ "model.layers.27.mlp.up_proj.weight": "pytorch_model-00114-of-00243.bin",
212
+ "model.layers.27.post_attention_layernorm.weight": "pytorch_model-00114-of-00243.bin",
213
+ "model.layers.27.self_attn.k_proj.weight": "pytorch_model-00111-of-00243.bin",
214
+ "model.layers.27.self_attn.o_proj.weight": "pytorch_model-00111-of-00243.bin",
215
+ "model.layers.27.self_attn.q_proj.weight": "pytorch_model-00110-of-00243.bin",
216
+ "model.layers.27.self_attn.rotary_emb.inv_freq": "pytorch_model-00111-of-00243.bin",
217
+ "model.layers.27.self_attn.v_proj.weight": "pytorch_model-00111-of-00243.bin",
218
+ "model.layers.28.input_layernorm.weight": "pytorch_model-00118-of-00243.bin",
219
+ "model.layers.28.mlp.down_proj.weight": "pytorch_model-00117-of-00243.bin",
220
+ "model.layers.28.mlp.gate_proj.weight": "pytorch_model-00116-of-00243.bin",
221
+ "model.layers.28.mlp.up_proj.weight": "pytorch_model-00118-of-00243.bin",
222
+ "model.layers.28.post_attention_layernorm.weight": "pytorch_model-00118-of-00243.bin",
223
+ "model.layers.28.self_attn.k_proj.weight": "pytorch_model-00115-of-00243.bin",
224
+ "model.layers.28.self_attn.o_proj.weight": "pytorch_model-00115-of-00243.bin",
225
+ "model.layers.28.self_attn.q_proj.weight": "pytorch_model-00114-of-00243.bin",
226
+ "model.layers.28.self_attn.rotary_emb.inv_freq": "pytorch_model-00115-of-00243.bin",
227
+ "model.layers.28.self_attn.v_proj.weight": "pytorch_model-00115-of-00243.bin",
228
+ "model.layers.29.input_layernorm.weight": "pytorch_model-00122-of-00243.bin",
229
+ "model.layers.29.mlp.down_proj.weight": "pytorch_model-00121-of-00243.bin",
230
+ "model.layers.29.mlp.gate_proj.weight": "pytorch_model-00120-of-00243.bin",
231
+ "model.layers.29.mlp.up_proj.weight": "pytorch_model-00122-of-00243.bin",
232
+ "model.layers.29.post_attention_layernorm.weight": "pytorch_model-00122-of-00243.bin",
233
+ "model.layers.29.self_attn.k_proj.weight": "pytorch_model-00119-of-00243.bin",
234
+ "model.layers.29.self_attn.o_proj.weight": "pytorch_model-00119-of-00243.bin",
235
+ "model.layers.29.self_attn.q_proj.weight": "pytorch_model-00118-of-00243.bin",
236
+ "model.layers.29.self_attn.rotary_emb.inv_freq": "pytorch_model-00119-of-00243.bin",
237
+ "model.layers.29.self_attn.v_proj.weight": "pytorch_model-00119-of-00243.bin",
238
+ "model.layers.3.input_layernorm.weight": "pytorch_model-00018-of-00243.bin",
239
+ "model.layers.3.mlp.down_proj.weight": "pytorch_model-00017-of-00243.bin",
240
+ "model.layers.3.mlp.gate_proj.weight": "pytorch_model-00016-of-00243.bin",
241
+ "model.layers.3.mlp.up_proj.weight": "pytorch_model-00018-of-00243.bin",
242
+ "model.layers.3.post_attention_layernorm.weight": "pytorch_model-00018-of-00243.bin",
243
+ "model.layers.3.self_attn.k_proj.weight": "pytorch_model-00015-of-00243.bin",
244
+ "model.layers.3.self_attn.o_proj.weight": "pytorch_model-00015-of-00243.bin",
245
+ "model.layers.3.self_attn.q_proj.weight": "pytorch_model-00014-of-00243.bin",
246
+ "model.layers.3.self_attn.rotary_emb.inv_freq": "pytorch_model-00015-of-00243.bin",
247
+ "model.layers.3.self_attn.v_proj.weight": "pytorch_model-00015-of-00243.bin",
248
+ "model.layers.30.input_layernorm.weight": "pytorch_model-00126-of-00243.bin",
249
+ "model.layers.30.mlp.down_proj.weight": "pytorch_model-00125-of-00243.bin",
250
+ "model.layers.30.mlp.gate_proj.weight": "pytorch_model-00124-of-00243.bin",
251
+ "model.layers.30.mlp.up_proj.weight": "pytorch_model-00126-of-00243.bin",
252
+ "model.layers.30.post_attention_layernorm.weight": "pytorch_model-00126-of-00243.bin",
253
+ "model.layers.30.self_attn.k_proj.weight": "pytorch_model-00123-of-00243.bin",
254
+ "model.layers.30.self_attn.o_proj.weight": "pytorch_model-00123-of-00243.bin",
255
+ "model.layers.30.self_attn.q_proj.weight": "pytorch_model-00122-of-00243.bin",
256
+ "model.layers.30.self_attn.rotary_emb.inv_freq": "pytorch_model-00123-of-00243.bin",
257
+ "model.layers.30.self_attn.v_proj.weight": "pytorch_model-00123-of-00243.bin",
258
+ "model.layers.31.input_layernorm.weight": "pytorch_model-00130-of-00243.bin",
259
+ "model.layers.31.mlp.down_proj.weight": "pytorch_model-00129-of-00243.bin",
260
+ "model.layers.31.mlp.gate_proj.weight": "pytorch_model-00128-of-00243.bin",
261
+ "model.layers.31.mlp.up_proj.weight": "pytorch_model-00130-of-00243.bin",
262
+ "model.layers.31.post_attention_layernorm.weight": "pytorch_model-00130-of-00243.bin",
263
+ "model.layers.31.self_attn.k_proj.weight": "pytorch_model-00127-of-00243.bin",
264
+ "model.layers.31.self_attn.o_proj.weight": "pytorch_model-00127-of-00243.bin",
265
+ "model.layers.31.self_attn.q_proj.weight": "pytorch_model-00126-of-00243.bin",
266
+ "model.layers.31.self_attn.rotary_emb.inv_freq": "pytorch_model-00127-of-00243.bin",
267
+ "model.layers.31.self_attn.v_proj.weight": "pytorch_model-00127-of-00243.bin",
268
+ "model.layers.32.input_layernorm.weight": "pytorch_model-00134-of-00243.bin",
269
+ "model.layers.32.mlp.down_proj.weight": "pytorch_model-00133-of-00243.bin",
270
+ "model.layers.32.mlp.gate_proj.weight": "pytorch_model-00132-of-00243.bin",
271
+ "model.layers.32.mlp.up_proj.weight": "pytorch_model-00134-of-00243.bin",
272
+ "model.layers.32.post_attention_layernorm.weight": "pytorch_model-00134-of-00243.bin",
273
+ "model.layers.32.self_attn.k_proj.weight": "pytorch_model-00131-of-00243.bin",
274
+ "model.layers.32.self_attn.o_proj.weight": "pytorch_model-00131-of-00243.bin",
275
+ "model.layers.32.self_attn.q_proj.weight": "pytorch_model-00130-of-00243.bin",
276
+ "model.layers.32.self_attn.rotary_emb.inv_freq": "pytorch_model-00131-of-00243.bin",
277
+ "model.layers.32.self_attn.v_proj.weight": "pytorch_model-00131-of-00243.bin",
278
+ "model.layers.33.input_layernorm.weight": "pytorch_model-00138-of-00243.bin",
279
+ "model.layers.33.mlp.down_proj.weight": "pytorch_model-00137-of-00243.bin",
280
+ "model.layers.33.mlp.gate_proj.weight": "pytorch_model-00136-of-00243.bin",
281
+ "model.layers.33.mlp.up_proj.weight": "pytorch_model-00138-of-00243.bin",
282
+ "model.layers.33.post_attention_layernorm.weight": "pytorch_model-00138-of-00243.bin",
283
+ "model.layers.33.self_attn.k_proj.weight": "pytorch_model-00135-of-00243.bin",
284
+ "model.layers.33.self_attn.o_proj.weight": "pytorch_model-00135-of-00243.bin",
285
+ "model.layers.33.self_attn.q_proj.weight": "pytorch_model-00134-of-00243.bin",
286
+ "model.layers.33.self_attn.rotary_emb.inv_freq": "pytorch_model-00135-of-00243.bin",
287
+ "model.layers.33.self_attn.v_proj.weight": "pytorch_model-00135-of-00243.bin",
288
+ "model.layers.34.input_layernorm.weight": "pytorch_model-00142-of-00243.bin",
289
+ "model.layers.34.mlp.down_proj.weight": "pytorch_model-00141-of-00243.bin",
290
+ "model.layers.34.mlp.gate_proj.weight": "pytorch_model-00140-of-00243.bin",
291
+ "model.layers.34.mlp.up_proj.weight": "pytorch_model-00142-of-00243.bin",
292
+ "model.layers.34.post_attention_layernorm.weight": "pytorch_model-00142-of-00243.bin",
293
+ "model.layers.34.self_attn.k_proj.weight": "pytorch_model-00139-of-00243.bin",
294
+ "model.layers.34.self_attn.o_proj.weight": "pytorch_model-00139-of-00243.bin",
295
+ "model.layers.34.self_attn.q_proj.weight": "pytorch_model-00138-of-00243.bin",
296
+ "model.layers.34.self_attn.rotary_emb.inv_freq": "pytorch_model-00139-of-00243.bin",
297
+ "model.layers.34.self_attn.v_proj.weight": "pytorch_model-00139-of-00243.bin",
298
+ "model.layers.35.input_layernorm.weight": "pytorch_model-00146-of-00243.bin",
299
+ "model.layers.35.mlp.down_proj.weight": "pytorch_model-00145-of-00243.bin",
300
+ "model.layers.35.mlp.gate_proj.weight": "pytorch_model-00144-of-00243.bin",
301
+ "model.layers.35.mlp.up_proj.weight": "pytorch_model-00146-of-00243.bin",
302
+ "model.layers.35.post_attention_layernorm.weight": "pytorch_model-00146-of-00243.bin",
303
+ "model.layers.35.self_attn.k_proj.weight": "pytorch_model-00143-of-00243.bin",
304
+ "model.layers.35.self_attn.o_proj.weight": "pytorch_model-00143-of-00243.bin",
305
+ "model.layers.35.self_attn.q_proj.weight": "pytorch_model-00142-of-00243.bin",
306
+ "model.layers.35.self_attn.rotary_emb.inv_freq": "pytorch_model-00143-of-00243.bin",
307
+ "model.layers.35.self_attn.v_proj.weight": "pytorch_model-00143-of-00243.bin",
308
+ "model.layers.36.input_layernorm.weight": "pytorch_model-00150-of-00243.bin",
309
+ "model.layers.36.mlp.down_proj.weight": "pytorch_model-00149-of-00243.bin",
310
+ "model.layers.36.mlp.gate_proj.weight": "pytorch_model-00148-of-00243.bin",
311
+ "model.layers.36.mlp.up_proj.weight": "pytorch_model-00150-of-00243.bin",
312
+ "model.layers.36.post_attention_layernorm.weight": "pytorch_model-00150-of-00243.bin",
313
+ "model.layers.36.self_attn.k_proj.weight": "pytorch_model-00147-of-00243.bin",
314
+ "model.layers.36.self_attn.o_proj.weight": "pytorch_model-00147-of-00243.bin",
315
+ "model.layers.36.self_attn.q_proj.weight": "pytorch_model-00146-of-00243.bin",
316
+ "model.layers.36.self_attn.rotary_emb.inv_freq": "pytorch_model-00147-of-00243.bin",
317
+ "model.layers.36.self_attn.v_proj.weight": "pytorch_model-00147-of-00243.bin",
318
+ "model.layers.37.input_layernorm.weight": "pytorch_model-00154-of-00243.bin",
319
+ "model.layers.37.mlp.down_proj.weight": "pytorch_model-00153-of-00243.bin",
320
+ "model.layers.37.mlp.gate_proj.weight": "pytorch_model-00152-of-00243.bin",
321
+ "model.layers.37.mlp.up_proj.weight": "pytorch_model-00154-of-00243.bin",
322
+ "model.layers.37.post_attention_layernorm.weight": "pytorch_model-00154-of-00243.bin",
323
+ "model.layers.37.self_attn.k_proj.weight": "pytorch_model-00151-of-00243.bin",
324
+ "model.layers.37.self_attn.o_proj.weight": "pytorch_model-00151-of-00243.bin",
325
+ "model.layers.37.self_attn.q_proj.weight": "pytorch_model-00150-of-00243.bin",
326
+ "model.layers.37.self_attn.rotary_emb.inv_freq": "pytorch_model-00151-of-00243.bin",
327
+ "model.layers.37.self_attn.v_proj.weight": "pytorch_model-00151-of-00243.bin",
328
+ "model.layers.38.input_layernorm.weight": "pytorch_model-00158-of-00243.bin",
329
+ "model.layers.38.mlp.down_proj.weight": "pytorch_model-00157-of-00243.bin",
330
+ "model.layers.38.mlp.gate_proj.weight": "pytorch_model-00156-of-00243.bin",
331
+ "model.layers.38.mlp.up_proj.weight": "pytorch_model-00158-of-00243.bin",
332
+ "model.layers.38.post_attention_layernorm.weight": "pytorch_model-00158-of-00243.bin",
333
+ "model.layers.38.self_attn.k_proj.weight": "pytorch_model-00155-of-00243.bin",
334
+ "model.layers.38.self_attn.o_proj.weight": "pytorch_model-00155-of-00243.bin",
335
+ "model.layers.38.self_attn.q_proj.weight": "pytorch_model-00154-of-00243.bin",
336
+ "model.layers.38.self_attn.rotary_emb.inv_freq": "pytorch_model-00155-of-00243.bin",
337
+ "model.layers.38.self_attn.v_proj.weight": "pytorch_model-00155-of-00243.bin",
338
+ "model.layers.39.input_layernorm.weight": "pytorch_model-00162-of-00243.bin",
339
+ "model.layers.39.mlp.down_proj.weight": "pytorch_model-00161-of-00243.bin",
340
+ "model.layers.39.mlp.gate_proj.weight": "pytorch_model-00160-of-00243.bin",
341
+ "model.layers.39.mlp.up_proj.weight": "pytorch_model-00162-of-00243.bin",
342
+ "model.layers.39.post_attention_layernorm.weight": "pytorch_model-00162-of-00243.bin",
343
+ "model.layers.39.self_attn.k_proj.weight": "pytorch_model-00159-of-00243.bin",
344
+ "model.layers.39.self_attn.o_proj.weight": "pytorch_model-00159-of-00243.bin",
345
+ "model.layers.39.self_attn.q_proj.weight": "pytorch_model-00158-of-00243.bin",
346
+ "model.layers.39.self_attn.rotary_emb.inv_freq": "pytorch_model-00159-of-00243.bin",
347
+ "model.layers.39.self_attn.v_proj.weight": "pytorch_model-00159-of-00243.bin",
348
+ "model.layers.4.input_layernorm.weight": "pytorch_model-00022-of-00243.bin",
349
+ "model.layers.4.mlp.down_proj.weight": "pytorch_model-00021-of-00243.bin",
350
+ "model.layers.4.mlp.gate_proj.weight": "pytorch_model-00020-of-00243.bin",
351
+ "model.layers.4.mlp.up_proj.weight": "pytorch_model-00022-of-00243.bin",
352
+ "model.layers.4.post_attention_layernorm.weight": "pytorch_model-00022-of-00243.bin",
353
+ "model.layers.4.self_attn.k_proj.weight": "pytorch_model-00019-of-00243.bin",
354
+ "model.layers.4.self_attn.o_proj.weight": "pytorch_model-00019-of-00243.bin",
355
+ "model.layers.4.self_attn.q_proj.weight": "pytorch_model-00018-of-00243.bin",
356
+ "model.layers.4.self_attn.rotary_emb.inv_freq": "pytorch_model-00019-of-00243.bin",
357
+ "model.layers.4.self_attn.v_proj.weight": "pytorch_model-00019-of-00243.bin",
358
+ "model.layers.40.input_layernorm.weight": "pytorch_model-00166-of-00243.bin",
359
+ "model.layers.40.mlp.down_proj.weight": "pytorch_model-00165-of-00243.bin",
360
+ "model.layers.40.mlp.gate_proj.weight": "pytorch_model-00164-of-00243.bin",
361
+ "model.layers.40.mlp.up_proj.weight": "pytorch_model-00166-of-00243.bin",
362
+ "model.layers.40.post_attention_layernorm.weight": "pytorch_model-00166-of-00243.bin",
363
+ "model.layers.40.self_attn.k_proj.weight": "pytorch_model-00163-of-00243.bin",
364
+ "model.layers.40.self_attn.o_proj.weight": "pytorch_model-00163-of-00243.bin",
365
+ "model.layers.40.self_attn.q_proj.weight": "pytorch_model-00162-of-00243.bin",
366
+ "model.layers.40.self_attn.rotary_emb.inv_freq": "pytorch_model-00163-of-00243.bin",
367
+ "model.layers.40.self_attn.v_proj.weight": "pytorch_model-00163-of-00243.bin",
368
+ "model.layers.41.input_layernorm.weight": "pytorch_model-00170-of-00243.bin",
369
+ "model.layers.41.mlp.down_proj.weight": "pytorch_model-00169-of-00243.bin",
370
+ "model.layers.41.mlp.gate_proj.weight": "pytorch_model-00168-of-00243.bin",
371
+ "model.layers.41.mlp.up_proj.weight": "pytorch_model-00170-of-00243.bin",
372
+ "model.layers.41.post_attention_layernorm.weight": "pytorch_model-00170-of-00243.bin",
373
+ "model.layers.41.self_attn.k_proj.weight": "pytorch_model-00167-of-00243.bin",
374
+ "model.layers.41.self_attn.o_proj.weight": "pytorch_model-00167-of-00243.bin",
375
+ "model.layers.41.self_attn.q_proj.weight": "pytorch_model-00166-of-00243.bin",
376
+ "model.layers.41.self_attn.rotary_emb.inv_freq": "pytorch_model-00167-of-00243.bin",
377
+ "model.layers.41.self_attn.v_proj.weight": "pytorch_model-00167-of-00243.bin",
378
+ "model.layers.42.input_layernorm.weight": "pytorch_model-00174-of-00243.bin",
379
+ "model.layers.42.mlp.down_proj.weight": "pytorch_model-00173-of-00243.bin",
380
+ "model.layers.42.mlp.gate_proj.weight": "pytorch_model-00172-of-00243.bin",
381
+ "model.layers.42.mlp.up_proj.weight": "pytorch_model-00174-of-00243.bin",
382
+ "model.layers.42.post_attention_layernorm.weight": "pytorch_model-00174-of-00243.bin",
383
+ "model.layers.42.self_attn.k_proj.weight": "pytorch_model-00171-of-00243.bin",
384
+ "model.layers.42.self_attn.o_proj.weight": "pytorch_model-00171-of-00243.bin",
385
+ "model.layers.42.self_attn.q_proj.weight": "pytorch_model-00170-of-00243.bin",
386
+ "model.layers.42.self_attn.rotary_emb.inv_freq": "pytorch_model-00171-of-00243.bin",
387
+ "model.layers.42.self_attn.v_proj.weight": "pytorch_model-00171-of-00243.bin",
388
+ "model.layers.43.input_layernorm.weight": "pytorch_model-00178-of-00243.bin",
389
+ "model.layers.43.mlp.down_proj.weight": "pytorch_model-00177-of-00243.bin",
390
+ "model.layers.43.mlp.gate_proj.weight": "pytorch_model-00176-of-00243.bin",
391
+ "model.layers.43.mlp.up_proj.weight": "pytorch_model-00178-of-00243.bin",
392
+ "model.layers.43.post_attention_layernorm.weight": "pytorch_model-00178-of-00243.bin",
393
+ "model.layers.43.self_attn.k_proj.weight": "pytorch_model-00175-of-00243.bin",
394
+ "model.layers.43.self_attn.o_proj.weight": "pytorch_model-00175-of-00243.bin",
395
+ "model.layers.43.self_attn.q_proj.weight": "pytorch_model-00174-of-00243.bin",
396
+ "model.layers.43.self_attn.rotary_emb.inv_freq": "pytorch_model-00175-of-00243.bin",
397
+ "model.layers.43.self_attn.v_proj.weight": "pytorch_model-00175-of-00243.bin",
398
+ "model.layers.44.input_layernorm.weight": "pytorch_model-00182-of-00243.bin",
399
+ "model.layers.44.mlp.down_proj.weight": "pytorch_model-00181-of-00243.bin",
400
+ "model.layers.44.mlp.gate_proj.weight": "pytorch_model-00180-of-00243.bin",
401
+ "model.layers.44.mlp.up_proj.weight": "pytorch_model-00182-of-00243.bin",
402
+ "model.layers.44.post_attention_layernorm.weight": "pytorch_model-00182-of-00243.bin",
403
+ "model.layers.44.self_attn.k_proj.weight": "pytorch_model-00179-of-00243.bin",
404
+ "model.layers.44.self_attn.o_proj.weight": "pytorch_model-00179-of-00243.bin",
405
+ "model.layers.44.self_attn.q_proj.weight": "pytorch_model-00178-of-00243.bin",
406
+ "model.layers.44.self_attn.rotary_emb.inv_freq": "pytorch_model-00179-of-00243.bin",
407
+ "model.layers.44.self_attn.v_proj.weight": "pytorch_model-00179-of-00243.bin",
408
+ "model.layers.45.input_layernorm.weight": "pytorch_model-00186-of-00243.bin",
409
+ "model.layers.45.mlp.down_proj.weight": "pytorch_model-00185-of-00243.bin",
410
+ "model.layers.45.mlp.gate_proj.weight": "pytorch_model-00184-of-00243.bin",
411
+ "model.layers.45.mlp.up_proj.weight": "pytorch_model-00186-of-00243.bin",
412
+ "model.layers.45.post_attention_layernorm.weight": "pytorch_model-00186-of-00243.bin",
413
+ "model.layers.45.self_attn.k_proj.weight": "pytorch_model-00183-of-00243.bin",
414
+ "model.layers.45.self_attn.o_proj.weight": "pytorch_model-00183-of-00243.bin",
415
+ "model.layers.45.self_attn.q_proj.weight": "pytorch_model-00182-of-00243.bin",
416
+ "model.layers.45.self_attn.rotary_emb.inv_freq": "pytorch_model-00183-of-00243.bin",
417
+ "model.layers.45.self_attn.v_proj.weight": "pytorch_model-00183-of-00243.bin",
418
+ "model.layers.46.input_layernorm.weight": "pytorch_model-00190-of-00243.bin",
419
+ "model.layers.46.mlp.down_proj.weight": "pytorch_model-00189-of-00243.bin",
420
+ "model.layers.46.mlp.gate_proj.weight": "pytorch_model-00188-of-00243.bin",
421
+ "model.layers.46.mlp.up_proj.weight": "pytorch_model-00190-of-00243.bin",
422
+ "model.layers.46.post_attention_layernorm.weight": "pytorch_model-00190-of-00243.bin",
423
+ "model.layers.46.self_attn.k_proj.weight": "pytorch_model-00187-of-00243.bin",
424
+ "model.layers.46.self_attn.o_proj.weight": "pytorch_model-00187-of-00243.bin",
425
+ "model.layers.46.self_attn.q_proj.weight": "pytorch_model-00186-of-00243.bin",
426
+ "model.layers.46.self_attn.rotary_emb.inv_freq": "pytorch_model-00187-of-00243.bin",
427
+ "model.layers.46.self_attn.v_proj.weight": "pytorch_model-00187-of-00243.bin",
428
+ "model.layers.47.input_layernorm.weight": "pytorch_model-00194-of-00243.bin",
429
+ "model.layers.47.mlp.down_proj.weight": "pytorch_model-00193-of-00243.bin",
430
+ "model.layers.47.mlp.gate_proj.weight": "pytorch_model-00192-of-00243.bin",
431
+ "model.layers.47.mlp.up_proj.weight": "pytorch_model-00194-of-00243.bin",
432
+ "model.layers.47.post_attention_layernorm.weight": "pytorch_model-00194-of-00243.bin",
433
+ "model.layers.47.self_attn.k_proj.weight": "pytorch_model-00191-of-00243.bin",
434
+ "model.layers.47.self_attn.o_proj.weight": "pytorch_model-00191-of-00243.bin",
435
+ "model.layers.47.self_attn.q_proj.weight": "pytorch_model-00190-of-00243.bin",
436
+ "model.layers.47.self_attn.rotary_emb.inv_freq": "pytorch_model-00191-of-00243.bin",
437
+ "model.layers.47.self_attn.v_proj.weight": "pytorch_model-00191-of-00243.bin",
438
+ "model.layers.48.input_layernorm.weight": "pytorch_model-00198-of-00243.bin",
439
+ "model.layers.48.mlp.down_proj.weight": "pytorch_model-00197-of-00243.bin",
440
+ "model.layers.48.mlp.gate_proj.weight": "pytorch_model-00196-of-00243.bin",
441
+ "model.layers.48.mlp.up_proj.weight": "pytorch_model-00198-of-00243.bin",
442
+ "model.layers.48.post_attention_layernorm.weight": "pytorch_model-00198-of-00243.bin",
443
+ "model.layers.48.self_attn.k_proj.weight": "pytorch_model-00195-of-00243.bin",
444
+ "model.layers.48.self_attn.o_proj.weight": "pytorch_model-00195-of-00243.bin",
445
+ "model.layers.48.self_attn.q_proj.weight": "pytorch_model-00194-of-00243.bin",
446
+ "model.layers.48.self_attn.rotary_emb.inv_freq": "pytorch_model-00195-of-00243.bin",
447
+ "model.layers.48.self_attn.v_proj.weight": "pytorch_model-00195-of-00243.bin",
448
+ "model.layers.49.input_layernorm.weight": "pytorch_model-00202-of-00243.bin",
449
+ "model.layers.49.mlp.down_proj.weight": "pytorch_model-00201-of-00243.bin",
450
+ "model.layers.49.mlp.gate_proj.weight": "pytorch_model-00200-of-00243.bin",
451
+ "model.layers.49.mlp.up_proj.weight": "pytorch_model-00202-of-00243.bin",
452
+ "model.layers.49.post_attention_layernorm.weight": "pytorch_model-00202-of-00243.bin",
453
+ "model.layers.49.self_attn.k_proj.weight": "pytorch_model-00199-of-00243.bin",
454
+ "model.layers.49.self_attn.o_proj.weight": "pytorch_model-00199-of-00243.bin",
455
+ "model.layers.49.self_attn.q_proj.weight": "pytorch_model-00198-of-00243.bin",
456
+ "model.layers.49.self_attn.rotary_emb.inv_freq": "pytorch_model-00199-of-00243.bin",
457
+ "model.layers.49.self_attn.v_proj.weight": "pytorch_model-00199-of-00243.bin",
458
+ "model.layers.5.input_layernorm.weight": "pytorch_model-00026-of-00243.bin",
459
+ "model.layers.5.mlp.down_proj.weight": "pytorch_model-00025-of-00243.bin",
460
+ "model.layers.5.mlp.gate_proj.weight": "pytorch_model-00024-of-00243.bin",
461
+ "model.layers.5.mlp.up_proj.weight": "pytorch_model-00026-of-00243.bin",
462
+ "model.layers.5.post_attention_layernorm.weight": "pytorch_model-00026-of-00243.bin",
463
+ "model.layers.5.self_attn.k_proj.weight": "pytorch_model-00023-of-00243.bin",
464
+ "model.layers.5.self_attn.o_proj.weight": "pytorch_model-00023-of-00243.bin",
465
+ "model.layers.5.self_attn.q_proj.weight": "pytorch_model-00022-of-00243.bin",
466
+ "model.layers.5.self_attn.rotary_emb.inv_freq": "pytorch_model-00023-of-00243.bin",
467
+ "model.layers.5.self_attn.v_proj.weight": "pytorch_model-00023-of-00243.bin",
468
+ "model.layers.50.input_layernorm.weight": "pytorch_model-00206-of-00243.bin",
469
+ "model.layers.50.mlp.down_proj.weight": "pytorch_model-00205-of-00243.bin",
470
+ "model.layers.50.mlp.gate_proj.weight": "pytorch_model-00204-of-00243.bin",
471
+ "model.layers.50.mlp.up_proj.weight": "pytorch_model-00206-of-00243.bin",
472
+ "model.layers.50.post_attention_layernorm.weight": "pytorch_model-00206-of-00243.bin",
473
+ "model.layers.50.self_attn.k_proj.weight": "pytorch_model-00203-of-00243.bin",
474
+ "model.layers.50.self_attn.o_proj.weight": "pytorch_model-00203-of-00243.bin",
475
+ "model.layers.50.self_attn.q_proj.weight": "pytorch_model-00202-of-00243.bin",
476
+ "model.layers.50.self_attn.rotary_emb.inv_freq": "pytorch_model-00203-of-00243.bin",
477
+ "model.layers.50.self_attn.v_proj.weight": "pytorch_model-00203-of-00243.bin",
478
+ "model.layers.51.input_layernorm.weight": "pytorch_model-00210-of-00243.bin",
479
+ "model.layers.51.mlp.down_proj.weight": "pytorch_model-00209-of-00243.bin",
480
+ "model.layers.51.mlp.gate_proj.weight": "pytorch_model-00208-of-00243.bin",
481
+ "model.layers.51.mlp.up_proj.weight": "pytorch_model-00210-of-00243.bin",
482
+ "model.layers.51.post_attention_layernorm.weight": "pytorch_model-00210-of-00243.bin",
483
+ "model.layers.51.self_attn.k_proj.weight": "pytorch_model-00207-of-00243.bin",
484
+ "model.layers.51.self_attn.o_proj.weight": "pytorch_model-00207-of-00243.bin",
485
+ "model.layers.51.self_attn.q_proj.weight": "pytorch_model-00206-of-00243.bin",
486
+ "model.layers.51.self_attn.rotary_emb.inv_freq": "pytorch_model-00207-of-00243.bin",
487
+ "model.layers.51.self_attn.v_proj.weight": "pytorch_model-00207-of-00243.bin",
488
+ "model.layers.52.input_layernorm.weight": "pytorch_model-00214-of-00243.bin",
489
+ "model.layers.52.mlp.down_proj.weight": "pytorch_model-00213-of-00243.bin",
490
+ "model.layers.52.mlp.gate_proj.weight": "pytorch_model-00212-of-00243.bin",
491
+ "model.layers.52.mlp.up_proj.weight": "pytorch_model-00214-of-00243.bin",
492
+ "model.layers.52.post_attention_layernorm.weight": "pytorch_model-00214-of-00243.bin",
493
+ "model.layers.52.self_attn.k_proj.weight": "pytorch_model-00211-of-00243.bin",
494
+ "model.layers.52.self_attn.o_proj.weight": "pytorch_model-00211-of-00243.bin",
495
+ "model.layers.52.self_attn.q_proj.weight": "pytorch_model-00210-of-00243.bin",
496
+ "model.layers.52.self_attn.rotary_emb.inv_freq": "pytorch_model-00211-of-00243.bin",
497
+ "model.layers.52.self_attn.v_proj.weight": "pytorch_model-00211-of-00243.bin",
498
+ "model.layers.53.input_layernorm.weight": "pytorch_model-00218-of-00243.bin",
499
+ "model.layers.53.mlp.down_proj.weight": "pytorch_model-00217-of-00243.bin",
500
+ "model.layers.53.mlp.gate_proj.weight": "pytorch_model-00216-of-00243.bin",
501
+ "model.layers.53.mlp.up_proj.weight": "pytorch_model-00218-of-00243.bin",
502
+ "model.layers.53.post_attention_layernorm.weight": "pytorch_model-00218-of-00243.bin",
503
+ "model.layers.53.self_attn.k_proj.weight": "pytorch_model-00215-of-00243.bin",
504
+ "model.layers.53.self_attn.o_proj.weight": "pytorch_model-00215-of-00243.bin",
505
+ "model.layers.53.self_attn.q_proj.weight": "pytorch_model-00214-of-00243.bin",
506
+ "model.layers.53.self_attn.rotary_emb.inv_freq": "pytorch_model-00215-of-00243.bin",
507
+ "model.layers.53.self_attn.v_proj.weight": "pytorch_model-00215-of-00243.bin",
508
+ "model.layers.54.input_layernorm.weight": "pytorch_model-00222-of-00243.bin",
509
+ "model.layers.54.mlp.down_proj.weight": "pytorch_model-00221-of-00243.bin",
510
+ "model.layers.54.mlp.gate_proj.weight": "pytorch_model-00220-of-00243.bin",
511
+ "model.layers.54.mlp.up_proj.weight": "pytorch_model-00222-of-00243.bin",
512
+ "model.layers.54.post_attention_layernorm.weight": "pytorch_model-00222-of-00243.bin",
513
+ "model.layers.54.self_attn.k_proj.weight": "pytorch_model-00219-of-00243.bin",
514
+ "model.layers.54.self_attn.o_proj.weight": "pytorch_model-00219-of-00243.bin",
515
+ "model.layers.54.self_attn.q_proj.weight": "pytorch_model-00218-of-00243.bin",
516
+ "model.layers.54.self_attn.rotary_emb.inv_freq": "pytorch_model-00219-of-00243.bin",
517
+ "model.layers.54.self_attn.v_proj.weight": "pytorch_model-00219-of-00243.bin",
518
+ "model.layers.55.input_layernorm.weight": "pytorch_model-00226-of-00243.bin",
519
+ "model.layers.55.mlp.down_proj.weight": "pytorch_model-00225-of-00243.bin",
520
+ "model.layers.55.mlp.gate_proj.weight": "pytorch_model-00224-of-00243.bin",
521
+ "model.layers.55.mlp.up_proj.weight": "pytorch_model-00226-of-00243.bin",
522
+ "model.layers.55.post_attention_layernorm.weight": "pytorch_model-00226-of-00243.bin",
523
+ "model.layers.55.self_attn.k_proj.weight": "pytorch_model-00223-of-00243.bin",
524
+ "model.layers.55.self_attn.o_proj.weight": "pytorch_model-00223-of-00243.bin",
525
+ "model.layers.55.self_attn.q_proj.weight": "pytorch_model-00222-of-00243.bin",
526
+ "model.layers.55.self_attn.rotary_emb.inv_freq": "pytorch_model-00223-of-00243.bin",
527
+ "model.layers.55.self_attn.v_proj.weight": "pytorch_model-00223-of-00243.bin",
528
+ "model.layers.56.input_layernorm.weight": "pytorch_model-00230-of-00243.bin",
529
+ "model.layers.56.mlp.down_proj.weight": "pytorch_model-00229-of-00243.bin",
530
+ "model.layers.56.mlp.gate_proj.weight": "pytorch_model-00228-of-00243.bin",
531
+ "model.layers.56.mlp.up_proj.weight": "pytorch_model-00230-of-00243.bin",
532
+ "model.layers.56.post_attention_layernorm.weight": "pytorch_model-00230-of-00243.bin",
533
+ "model.layers.56.self_attn.k_proj.weight": "pytorch_model-00227-of-00243.bin",
534
+ "model.layers.56.self_attn.o_proj.weight": "pytorch_model-00227-of-00243.bin",
535
+ "model.layers.56.self_attn.q_proj.weight": "pytorch_model-00226-of-00243.bin",
536
+ "model.layers.56.self_attn.rotary_emb.inv_freq": "pytorch_model-00227-of-00243.bin",
537
+ "model.layers.56.self_attn.v_proj.weight": "pytorch_model-00227-of-00243.bin",
538
+ "model.layers.57.input_layernorm.weight": "pytorch_model-00234-of-00243.bin",
539
+ "model.layers.57.mlp.down_proj.weight": "pytorch_model-00233-of-00243.bin",
540
+ "model.layers.57.mlp.gate_proj.weight": "pytorch_model-00232-of-00243.bin",
541
+ "model.layers.57.mlp.up_proj.weight": "pytorch_model-00234-of-00243.bin",
542
+ "model.layers.57.post_attention_layernorm.weight": "pytorch_model-00234-of-00243.bin",
543
+ "model.layers.57.self_attn.k_proj.weight": "pytorch_model-00231-of-00243.bin",
544
+ "model.layers.57.self_attn.o_proj.weight": "pytorch_model-00231-of-00243.bin",
545
+ "model.layers.57.self_attn.q_proj.weight": "pytorch_model-00230-of-00243.bin",
546
+ "model.layers.57.self_attn.rotary_emb.inv_freq": "pytorch_model-00231-of-00243.bin",
547
+ "model.layers.57.self_attn.v_proj.weight": "pytorch_model-00231-of-00243.bin",
548
+ "model.layers.58.input_layernorm.weight": "pytorch_model-00238-of-00243.bin",
549
+ "model.layers.58.mlp.down_proj.weight": "pytorch_model-00237-of-00243.bin",
550
+ "model.layers.58.mlp.gate_proj.weight": "pytorch_model-00236-of-00243.bin",
551
+ "model.layers.58.mlp.up_proj.weight": "pytorch_model-00238-of-00243.bin",
552
+ "model.layers.58.post_attention_layernorm.weight": "pytorch_model-00238-of-00243.bin",
553
+ "model.layers.58.self_attn.k_proj.weight": "pytorch_model-00235-of-00243.bin",
554
+ "model.layers.58.self_attn.o_proj.weight": "pytorch_model-00235-of-00243.bin",
555
+ "model.layers.58.self_attn.q_proj.weight": "pytorch_model-00234-of-00243.bin",
556
+ "model.layers.58.self_attn.rotary_emb.inv_freq": "pytorch_model-00235-of-00243.bin",
557
+ "model.layers.58.self_attn.v_proj.weight": "pytorch_model-00235-of-00243.bin",
558
+ "model.layers.59.input_layernorm.weight": "pytorch_model-00242-of-00243.bin",
559
+ "model.layers.59.mlp.down_proj.weight": "pytorch_model-00241-of-00243.bin",
560
+ "model.layers.59.mlp.gate_proj.weight": "pytorch_model-00240-of-00243.bin",
561
+ "model.layers.59.mlp.up_proj.weight": "pytorch_model-00242-of-00243.bin",
562
+ "model.layers.59.post_attention_layernorm.weight": "pytorch_model-00242-of-00243.bin",
563
+ "model.layers.59.self_attn.k_proj.weight": "pytorch_model-00239-of-00243.bin",
564
+ "model.layers.59.self_attn.o_proj.weight": "pytorch_model-00239-of-00243.bin",
565
+ "model.layers.59.self_attn.q_proj.weight": "pytorch_model-00238-of-00243.bin",
566
+ "model.layers.59.self_attn.rotary_emb.inv_freq": "pytorch_model-00239-of-00243.bin",
567
+ "model.layers.59.self_attn.v_proj.weight": "pytorch_model-00239-of-00243.bin",
568
+ "model.layers.6.input_layernorm.weight": "pytorch_model-00030-of-00243.bin",
569
+ "model.layers.6.mlp.down_proj.weight": "pytorch_model-00029-of-00243.bin",
570
+ "model.layers.6.mlp.gate_proj.weight": "pytorch_model-00028-of-00243.bin",
571
+ "model.layers.6.mlp.up_proj.weight": "pytorch_model-00030-of-00243.bin",
572
+ "model.layers.6.post_attention_layernorm.weight": "pytorch_model-00030-of-00243.bin",
573
+ "model.layers.6.self_attn.k_proj.weight": "pytorch_model-00027-of-00243.bin",
574
+ "model.layers.6.self_attn.o_proj.weight": "pytorch_model-00027-of-00243.bin",
575
+ "model.layers.6.self_attn.q_proj.weight": "pytorch_model-00026-of-00243.bin",
576
+ "model.layers.6.self_attn.rotary_emb.inv_freq": "pytorch_model-00027-of-00243.bin",
577
+ "model.layers.6.self_attn.v_proj.weight": "pytorch_model-00027-of-00243.bin",
578
+ "model.layers.7.input_layernorm.weight": "pytorch_model-00034-of-00243.bin",
579
+ "model.layers.7.mlp.down_proj.weight": "pytorch_model-00033-of-00243.bin",
580
+ "model.layers.7.mlp.gate_proj.weight": "pytorch_model-00032-of-00243.bin",
581
+ "model.layers.7.mlp.up_proj.weight": "pytorch_model-00034-of-00243.bin",
582
+ "model.layers.7.post_attention_layernorm.weight": "pytorch_model-00034-of-00243.bin",
583
+ "model.layers.7.self_attn.k_proj.weight": "pytorch_model-00031-of-00243.bin",
584
+ "model.layers.7.self_attn.o_proj.weight": "pytorch_model-00031-of-00243.bin",
585
+ "model.layers.7.self_attn.q_proj.weight": "pytorch_model-00030-of-00243.bin",
586
+ "model.layers.7.self_attn.rotary_emb.inv_freq": "pytorch_model-00031-of-00243.bin",
587
+ "model.layers.7.self_attn.v_proj.weight": "pytorch_model-00031-of-00243.bin",
588
+ "model.layers.8.input_layernorm.weight": "pytorch_model-00038-of-00243.bin",
589
+ "model.layers.8.mlp.down_proj.weight": "pytorch_model-00037-of-00243.bin",
590
+ "model.layers.8.mlp.gate_proj.weight": "pytorch_model-00036-of-00243.bin",
591
+ "model.layers.8.mlp.up_proj.weight": "pytorch_model-00038-of-00243.bin",
592
+ "model.layers.8.post_attention_layernorm.weight": "pytorch_model-00038-of-00243.bin",
593
+ "model.layers.8.self_attn.k_proj.weight": "pytorch_model-00035-of-00243.bin",
594
+ "model.layers.8.self_attn.o_proj.weight": "pytorch_model-00035-of-00243.bin",
595
+ "model.layers.8.self_attn.q_proj.weight": "pytorch_model-00034-of-00243.bin",
596
+ "model.layers.8.self_attn.rotary_emb.inv_freq": "pytorch_model-00035-of-00243.bin",
597
+ "model.layers.8.self_attn.v_proj.weight": "pytorch_model-00035-of-00243.bin",
598
+ "model.layers.9.input_layernorm.weight": "pytorch_model-00042-of-00243.bin",
599
+ "model.layers.9.mlp.down_proj.weight": "pytorch_model-00041-of-00243.bin",
600
+ "model.layers.9.mlp.gate_proj.weight": "pytorch_model-00040-of-00243.bin",
601
+ "model.layers.9.mlp.up_proj.weight": "pytorch_model-00042-of-00243.bin",
602
+ "model.layers.9.post_attention_layernorm.weight": "pytorch_model-00042-of-00243.bin",
603
+ "model.layers.9.self_attn.k_proj.weight": "pytorch_model-00039-of-00243.bin",
604
+ "model.layers.9.self_attn.o_proj.weight": "pytorch_model-00039-of-00243.bin",
605
+ "model.layers.9.self_attn.q_proj.weight": "pytorch_model-00038-of-00243.bin",
606
+ "model.layers.9.self_attn.rotary_emb.inv_freq": "pytorch_model-00039-of-00243.bin",
607
+ "model.layers.9.self_attn.v_proj.weight": "pytorch_model-00039-of-00243.bin",
608
+ "model.norm.weight": "pytorch_model-00242-of-00243.bin"
609
+ }
610
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "unk_token": {
17
+ "content": "<unk>",
18
+ "lstrip": false,
19
+ "normalized": true,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
tokenizer_config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "bos_token": {
5
+ "__type": "AddedToken",
6
+ "content": "<s>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "clean_up_tokenization_spaces": false,
13
+ "eos_token": {
14
+ "__type": "AddedToken",
15
+ "content": "</s>",
16
+ "lstrip": false,
17
+ "normalized": true,
18
+ "rstrip": false,
19
+ "single_word": false
20
+ },
21
+ "model_max_length": 2048,
22
+ "pad_token": null,
23
+ "sp_model_kwargs": {},
24
+ "tokenizer_class": "LlamaTokenizer",
25
+ "unk_token": {
26
+ "__type": "AddedToken",
27
+ "content": "<unk>",
28
+ "lstrip": false,
29
+ "normalized": true,
30
+ "rstrip": false,
31
+ "single_word": false
32
+ }
33
+ }