Emiliyan Pavlov commited on
Commit
97f0b38
0 Parent(s):

Initial commit

Browse files
.gitattributes ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
+ tokenizer.* filter=lfs diff=lfs merge=lfs -text
LICENSE ADDED
@@ -0,0 +1,78 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Gemma Terms of Use
2
+
3
+ Last modified: April 1, 2024
4
+
5
+ By using, reproducing, modifying, distributing, performing or displaying any portion or element of Gemma, Model Derivatives including via any Hosted Service, (each as defined below) (collectively, the "Gemma Services") or otherwise accepting the terms of this Agreement, you agree to be bound by this Agreement.
6
+ Section 1: DEFINITIONS
7
+ 1.1 Definitions
8
+
9
+ (a) "Agreement" or "Gemma Terms of Use" means these terms and conditions that govern the use, reproduction, Distribution or modification of the Gemma Services and any terms and conditions incorporated by reference.
10
+
11
+ (b) "Distribution" or "Distribute" means any transmission, publication, or other sharing of Gemma or Model Derivatives to a third party, including by providing or making Gemma or its functionality available as a hosted service via API, web access, or any other electronic or remote means ("Hosted Service").
12
+
13
+ (c) "Gemma" means the set of machine learning language models, trained model weights and parameters identified at ai.google.dev/gemma, regardless of the source that you obtained it from.
14
+
15
+ (d) "Google" means Google LLC.
16
+
17
+ (e) "Model Derivatives" means all (i) modifications to Gemma, (ii) works based on Gemma, or (iii) any other machine learning model which is created by transfer of patterns of the weights, parameters, operations, or Output of Gemma, to that model in order to cause that model to perform similarly to Gemma, including distillation methods that use intermediate data representations or methods based on the generation of synthetic data Outputs by Gemma for training that model. For clarity, Outputs are not deemed Model Derivatives.
18
+
19
+ (f) "Output" means the information content output of Gemma or a Model Derivative that results from operating or otherwise using Gemma or the Model Derivative, including via a Hosted Service.
20
+ 1.2
21
+
22
+ As used in this Agreement, "including" means "including without limitation".
23
+ Section 2: ELIGIBILITY AND USAGE
24
+ 2.1 Eligibility
25
+
26
+ You represent and warrant that you have the legal capacity to enter into this Agreement (including being of sufficient age of consent). If you are accessing or using any of the Gemma Services for or on behalf of a legal entity, (a) you are entering into this Agreement on behalf of yourself and that legal entity, (b) you represent and warrant that you have the authority to act on behalf of and bind that entity to this Agreement and (c) references to "you" or "your" in the remainder of this Agreement refers to both you (as an individual) and that entity.
27
+ 2.2 Use
28
+
29
+ You may use, reproduce, modify, Distribute, perform or display any of the Gemma Services only in accordance with the terms of this Agreement, and must not violate (or encourage or permit anyone else to violate) any term of this Agreement.
30
+ Section 3: DISTRIBUTION AND RESTRICTIONS
31
+ 3.1 Distribution and Redistribution
32
+
33
+ You may reproduce or Distribute copies of Gemma or Model Derivatives if you meet all of the following conditions:
34
+
35
+ You must include the use restrictions referenced in Section 3.2 as an enforceable provision in any agreement (e.g., license agreement, terms of use, etc.) governing the use and/or distribution of Gemma or Model Derivatives and you must provide notice to subsequent users you Distribute to that Gemma or Model Derivatives are subject to the use restrictions in Section 3.2.
36
+ You must provide all third party recipients of Gemma or Model Derivatives a copy of this Agreement.
37
+ You must cause any modified files to carry prominent notices stating that you modified the files.
38
+ All Distributions (other than through a Hosted Service) must be accompanied by a "Notice" text file that contains the following notice: "Gemma is provided under and subject to the Gemma Terms of Use found at ai.google.dev/gemma/terms".
39
+
40
+ You may add your own intellectual property statement to your modifications and, except as set forth in this Section, may provide additional or different terms and conditions for use, reproduction, or Distribution of your modifications, or for any such Model Derivatives as a whole, provided your use, reproduction, modification, Distribution, performance, and display of Gemma otherwise complies with the terms and conditions of this Agreement. Any additional or different terms and conditions you impose must not conflict with the terms of this Agreement.
41
+ 3.2 Use Restrictions
42
+
43
+ You must not use any of the Gemma Services:
44
+
45
+ for the restricted uses set forth in the Gemma Prohibited Use Policy at ai.google.dev/gemma/prohibited_use_policy ("Prohibited Use Policy"), which is hereby incorporated by reference into this Agreement; or
46
+ in violation of applicable laws and regulations.
47
+
48
+ To the maximum extent permitted by law, Google reserves the right to restrict (remotely or otherwise) usage of any of the Gemma Services that Google reasonably believes are in violation of this Agreement.
49
+ 3.3 Generated Output
50
+
51
+ Google claims no rights in Outputs you generate using Gemma. You and your users are solely responsible for Outputs and their subsequent uses.
52
+ Section 4: ADDITIONAL PROVISIONS
53
+ 4.1 Updates
54
+
55
+ Google may update Gemma from time to time.
56
+ 4.2 Trademarks
57
+
58
+ Nothing in this Agreement grants you any rights to use Google's trademarks, trade names, logos or to otherwise suggest endorsement or misrepresent the relationship between you and Google. Google reserves any rights not expressly granted herein.
59
+ 4.3 DISCLAIMER OF WARRANTY
60
+
61
+ UNLESS REQUIRED BY APPLICABLE LAW, THE GEMMA SERVICES, AND OUTPUTS, ARE PROVIDED ON AN "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, EITHER EXPRESS OR IMPLIED, INCLUDING ANY WARRANTIES OR CONDITIONS OF TITLE, NON-INFRINGEMENT, MERCHANTABILITY, OR FITNESS FOR A PARTICULAR PURPOSE. YOU ARE SOLELY RESPONSIBLE FOR DETERMINING THE APPROPRIATENESS OF USING, REPRODUCING, MODIFYING, PERFORMING, DISPLAYING OR DISTRIBUTING ANY OF THE GEMMA SERVICES OR OUTPUTS AND ASSUME ANY AND ALL RISKS ASSOCIATED WITH YOUR USE OR DISTRIBUTION OF ANY OF THE GEMMA SERVICES OR OUTPUTS AND YOUR EXERCISE OF RIGHTS AND PERMISSIONS UNDER THIS AGREEMENT.
62
+ 4.4 LIMITATION OF LIABILITY
63
+
64
+ TO THE FULLEST EXTENT PERMITTED BY APPLICABLE LAW, IN NO EVENT AND UNDER NO LEGAL THEORY, WHETHER IN TORT (INCLUDING NEGLIGENCE), PRODUCT LIABILITY, CONTRACT, OR OTHERWISE, UNLESS REQUIRED BY APPLICABLE LAW, SHALL GOOGLE OR ITS AFFILIATES BE LIABLE TO YOU FOR DAMAGES, INCLUDING ANY DIRECT, INDIRECT, SPECIAL, INCIDENTAL, EXEMPLARY, CONSEQUENTIAL, OR PUNITIVE DAMAGES, OR LOST PROFITS OF ANY KIND ARISING FROM THIS AGREEMENT OR RELATED TO, ANY OF THE GEMMA SERVICES OR OUTPUTS EVEN IF GOOGLE OR ITS AFFILIATES HAVE BEEN ADVISED OF THE POSSIBILITY OF SUCH DAMAGES.
65
+ 4.5 Term, Termination, and Survival
66
+
67
+ The term of this Agreement will commence upon your acceptance of this Agreement (including acceptance by your use, modification, or Distribution, reproduction, performance or display of any portion or element of the Gemma Services) and will continue in full force and effect until terminated in accordance with the terms of this Agreement. Google may terminate this Agreement if you are in breach of any term of this Agreement. Upon termination of this Agreement, you must delete and cease use and Distribution of all copies of Gemma and Model Derivatives in your possession or control. Sections 1, 2.1, 3.3, 4.2 to 4.9 shall survive the termination of this Agreement.
68
+ 4.6 Governing Law and Jurisdiction
69
+
70
+ This Agreement will be governed by the laws of the State of California without regard to choice of law principles. The UN Convention on Contracts for the International Sale of Goods does not apply to this Agreement. The state and federal courts of Santa Clara County, California shall have exclusive jurisdiction of any dispute arising out of this Agreement.
71
+ 4.7 Severability
72
+
73
+ If any provision of this Agreement is held to be invalid, illegal or unenforceable, the remaining provisions shall be unaffected thereby and remain valid as if such provision had not been set forth herein.
74
+ 4.8 Entire Agreement
75
+
76
+ This Agreement states all the terms agreed between the parties and supersedes all other agreements between the parties as of the date of acceptance relating to its subject matter.
77
+ 4.9 No Waiver
78
+ Google will not be treated as having waived any rights by not exercising (or delaying the exercise of) any rights under this Agreement.
README.md ADDED
@@ -0,0 +1,143 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ tags:
4
+ - gemma2
5
+ - instruct
6
+ - bggpt
7
+ - insait
8
+ license: gemma
9
+ language:
10
+ - bg
11
+ - en
12
+ base_model:
13
+ - google/gemma-2-27b-it
14
+ - google/gemma-2-27b
15
+ pipeline_tag: text-generation
16
+ ---
17
+
18
+ # INSAIT-Institute/BgGPT-Gemma-2-27B-IT-v1.0
19
+
20
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/637e1f8cf7e01589cc17bf7e/p6d0YFHjWCQ3S12jWqO1m.png)
21
+
22
+ INSAIT introduces **BgGPT-Gemma-2-27B-IT-v1.0**, a state-of-the-art Bulgarian language model based on **google/gemma-2-27b** and **google/gemma-2-27b-it**.
23
+ BgGPT-Gemma-2-27B-IT-v1.0 is **free to use** and distributed under the [Gemma Terms of Use](https://ai.google.dev/gemma/terms).
24
+ This model was created by [`INSAIT`](https://insait.ai/), part of Sofia University St. Kliment Ohridski, in Sofia, Bulgaria.
25
+
26
+ # Model description
27
+
28
+ The model was built on top of Google’s Gemma 2 27B open models.
29
+ It was continuously pre-trained on around 100 billion tokens (85 billion in Bulgarian) using the Branch-and-Merge strategy INSAIT presented at [EMNLP’24](https://aclanthology.org/2024.findings-emnlp.1000/),
30
+ allowing the model to gain outstanding Bulgarian cultural and linguistic capabilities while retaining its English performance.
31
+ During the pre-training stage, we use various datasets, including Bulgarian web crawl data, freely available datasets such as Wikipedia, a range of specialized Bulgarian datasets sourced by the INSAIT Institute,
32
+ and machine translations of popular English datasets.
33
+ The model was then instruction-fine-tuned on a newly constructed Bulgarian instruction dataset created using real-world conversations.
34
+ For more information check our [blogpost](https://models.bggpt.ai/blog/).
35
+
36
+ # Benchmarks and Results
37
+
38
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/65fefdc282708115868203aa/5knpdR-QDSuM3WlpRxe-M.png)
39
+
40
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/65fefdc282708115868203aa/TY8F34DpUf7uXbsFVywn2.png)
41
+
42
+ We evaluate our models on a set of standard English benchmarks, a translated version of them in Bulgarian, as well as, Bulgarian specific benchmarks we collected:
43
+
44
+ - **Winogrande challenge**: testing world knowledge and understanding
45
+ - **Hellaswag**: testing sentence completion
46
+ - **ARC Easy/Challenge**: testing logical reasoning
47
+ - **TriviaQA**: testing trivia knowledge
48
+ - **GSM-8k**: solving multiple-choice questions in high-school mathematics
49
+ - **Exams**: solving high school problems from natural and social sciences
50
+ - **MON**: contains exams across various subjects for grades 4 to 12
51
+
52
+
53
+ These benchmarks test logical reasoning, mathematics, knowledge, language understanding and other skills of the models and are provided at https://github.com/insait-institute/lm-evaluation-harness-bg.
54
+ The graphs above show the performance of BgGPT 9B and BgGPT 27B compared to other large open models. The results show the excellent abilities of both 9B and 27B models in Bulgarian, which allow them to **outperform much larger models**,
55
+ including Alibaba’s Qwen 2.5 72B and Meta’s Llama3.1 70B. Further, both BgGPT 9B and BgGPT 27B **significantly improve upon the previous version of BgGPT** based on Mistral-7B ([BgGPT-7B-Instruct-v0.2](https://huggingface.co/INSAIT-Institute/BgGPT-7B-Instruct-v0.2), shown in grey in the figure).
56
+ Finally, our models retain the **excellent English performance** inherited from the original Google Gemma 2 models upon which they are based.
57
+
58
+
59
+ # Chat Preference
60
+
61
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/65f955d0c312ee009f8262bd/fGE9slHcgDJL_Kotf_FlY.png)
62
+
63
+ In addition to benchmark evaluation, we evaluated the BgGPT 27B model in terms of chat performance on **thousands of real-world Bulgarian conversations** from around **100 different topics**.
64
+ The results show that our model **significantly surpasses** the performance of the smaller variants of commercial models, such as Anthropic’s Claude Haiku and OpenAI’s GPT-4o-mini in Bulgarian chat performance,
65
+ and is **on par** with the best commercial models, such as Anthropic’s Claude Sonnet and OpenAI’s GPT-4o **according to GPT-4o itself**.
66
+
67
+ # Instruction format
68
+
69
+ In order to leverage instruction fine-tuning, your prompt should begin with a beginning-of-sequence token `<bos>` and be formatted in the Gemma 2 chat template. `<bos>` should only be the first token in a chat sequence.
70
+
71
+ E.g.
72
+ ```
73
+ <bos><start_of_turn>user\n
74
+ Кога е основан Софийският университет?<end_of_turn>\n
75
+ <start_of_turn>model\n
76
+ ```
77
+
78
+ This format is also available as a [chat template](https://huggingface.co/docs/transformers/main/chat_templating) via the `apply_chat_template()` method:
79
+
80
+ ```python
81
+ tokenizer = AutoTokenizer.from_pretrained(
82
+ "INSAIT-Institute/BgGPT-Gemma-2-27B-IT-v1.0",
83
+ use_default_system_prompt=False,
84
+ )
85
+
86
+ messages = [
87
+ {"role": "user", "content": "Кога е основан Софийският университет?"},
88
+ ]
89
+ input_ids = tokenizer.apply_chat_template(messages, return_tensors="pt", return_dict=True).to("cuda")
90
+
91
+ outputs = model.generate(**input_ids, max_new_tokens=256)
92
+ print(tokenizer.decode(outputs[0]))
93
+
94
+ ```
95
+
96
+ # Recommended Parameters
97
+
98
+ For optimal performance, we recommend the following parameters for text generation, as we have extensively tested our model with them:
99
+
100
+ ```python
101
+ generation_params = {
102
+ "temperature": 0.1
103
+ "top_k": 20,
104
+ "repetition_penalty": 1.1
105
+ }
106
+ ```
107
+
108
+ In principle, increasing temperature should work adequately as well.
109
+
110
+ # Use in 🤗 Transformers
111
+ First install the latest version of the transformers library:
112
+ ```
113
+ pip install -U 'transformers[torch]'
114
+ ```
115
+ Then load the model in transformers:
116
+ ```python
117
+ model = AutoModelForCausalLM.from_pretrained(
118
+ "INSAIT-Institute/BgGPT-Gemma-2-27B-IT-v1.0",
119
+ torch_dtype=torch.bfloat16,
120
+ attn_implementation="eager",
121
+ device_map="auto",
122
+ )
123
+ ```
124
+ **Important Note:** Models based on Gemma 2 such as BgGPT-Gemma-2-27B-IT-v1.0 do not support flash attention. Using it results in degraded performance.
125
+
126
+ # Use with GGML / llama.cpp
127
+
128
+ The model and instructions for usage in GGUF format are available at [INSAIT-Institute/BgGPT-Gemma-2-27B-IT-v1.0-GGUF](https://huggingface.co/INSAIT-Institute/BgGPT-Gemma-2-27B-IT-v1.0-GGUF).
129
+
130
+ # Community Feedback
131
+
132
+ We welcome feedback from the community to help improve BgGPT. If you have suggestions, encounter any issues, or have ideas for improvements, please:
133
+ - Share your experience using the model through Hugging Face's community discussion feature or
134
+ - Contact us at bggpt@insait.ai
135
+
136
+ Your real-world usage and insights are valuable in helping us optimize the model's performance and behaviour for various use cases.
137
+
138
+ # Summary
139
+ - **Finetuned from:** [google/gemma-2-27b-it](https://huggingface.co/google/gemma-2-27b-it); [google/gemma-2-27b](https://huggingface.co/google/gemma-2-27b);
140
+ - **Model type:** Causal decoder-only transformer language model
141
+ - **Language:** Bulgarian and English
142
+ - **Contact:** [bggpt@insait.ai](mailto:bggpt@insait.ai)
143
+ - **License:** BgGPT is distributed under [Gemma Terms of Use](https://huggingface.co/INSAIT-Institute/BgGPT-Gemma-2-27B-IT-v1.0/raw/main/LICENSE)
config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Gemma2ForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "attn_logit_softcapping": 50.0,
8
+ "bos_token_id": 2,
9
+ "cache_implementation": "hybrid",
10
+ "eos_token_id": 1,
11
+ "final_logit_softcapping": 30.0,
12
+ "head_dim": 128,
13
+ "hidden_act": "gelu_pytorch_tanh",
14
+ "hidden_activation": "gelu_pytorch_tanh",
15
+ "hidden_size": 4608,
16
+ "initializer_range": 0.02,
17
+ "intermediate_size": 36864,
18
+ "max_position_embeddings": 8192,
19
+ "model_type": "gemma2",
20
+ "num_attention_heads": 32,
21
+ "num_hidden_layers": 46,
22
+ "num_key_value_heads": 16,
23
+ "pad_token_id": 0,
24
+ "query_pre_attn_scalar": 144,
25
+ "rms_norm_eps": 1e-06,
26
+ "rope_theta": 10000.0,
27
+ "sliding_window": 4096,
28
+ "sliding_window_size": 4096,
29
+ "torch_dtype": "bfloat16",
30
+ "transformers_version": "4.42.3",
31
+ "use_cache": true,
32
+ "vocab_size": 256000
33
+ }
generation_config.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 2,
4
+ "cache_implementation": "hybrid",
5
+ "eos_token_id": 1,
6
+ "pad_token_id": 0,
7
+ "transformers_version": "4.42.3"
8
+ }
model-00001-of-00012.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f353ff7d46c043eb8fdf1fa5b8782cda55c5a06a7ebf5ed70b70ff589008bd3e
3
+ size 4737543272
model-00002-of-00012.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc631b8d8f70771b4320eda1e23afc74487ba95209f9bf492a408310588582f3
3
+ size 4869739672
model-00003-of-00012.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54b786c54fa7cbb353cae0eb8fcae47f24899645e80de9930479cb3c33c58e27
3
+ size 4869739680
model-00004-of-00012.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a155947aad19a45ddcb958eccb400b7bcd5457a3da65af2c6c13f3c355699f0
3
+ size 4983023736
model-00005-of-00012.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4da11e9db2bb5a4c7da7a9f869ab86b2b2f0aa7cc7780b126272782ad982539c
3
+ size 4869739720
model-00006-of-00012.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9e9b053bb8e2130c421a4ca23e2a6100bdcb8da1a1c78f6b9e2d7ba7f43b8b1
3
+ size 4869739712
model-00007-of-00012.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a7fa5f4dc69755d63f38badc24f448f28c85a8f969be4beb21e75b7cdd3c346
3
+ size 4983023736
model-00008-of-00012.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d8e04ff87cce282f5d0e6f2937f4f5094a2ede9ba3eb92b55cbd4c1a6b2634d
3
+ size 4869739720
model-00009-of-00012.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:79c66a4105fbfd31898fa8ccb564aabb2c0db7a929634980b0f3744475089aea
3
+ size 4869739712
model-00010-of-00012.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b95dd25eff158a39211d2eb340566a53fc32ed22ec86c30dd371cdff578bb1e
3
+ size 4983023736
model-00011-of-00012.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55b69e23c2b5b58430ec176ad9b10de5a5a38dbabe042a6065faa682e9fb0b5c
3
+ size 4869739720
model-00012-of-00012.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0aa2d94a18b25c3d38997770ee8fa26cb5bb82100a3646a5bbc8f2230071b0bb
3
+ size 679524136
model.safetensors.index.json ADDED
@@ -0,0 +1,515 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 54454256640
4
+ },
5
+ "weight_map": {
6
+ "model.embed_tokens.weight": "model-00001-of-00012.safetensors",
7
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00012.safetensors",
8
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00012.safetensors",
9
+ "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00012.safetensors",
10
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00012.safetensors",
11
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00012.safetensors",
12
+ "model.layers.0.post_feedforward_layernorm.weight": "model-00001-of-00012.safetensors",
13
+ "model.layers.0.pre_feedforward_layernorm.weight": "model-00001-of-00012.safetensors",
14
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00012.safetensors",
15
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00012.safetensors",
16
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00012.safetensors",
17
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00012.safetensors",
18
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00012.safetensors",
19
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00012.safetensors",
20
+ "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00012.safetensors",
21
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00012.safetensors",
22
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00012.safetensors",
23
+ "model.layers.1.post_feedforward_layernorm.weight": "model-00001-of-00012.safetensors",
24
+ "model.layers.1.pre_feedforward_layernorm.weight": "model-00001-of-00012.safetensors",
25
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00012.safetensors",
26
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00012.safetensors",
27
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00012.safetensors",
28
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00012.safetensors",
29
+ "model.layers.10.input_layernorm.weight": "model-00004-of-00012.safetensors",
30
+ "model.layers.10.mlp.down_proj.weight": "model-00004-of-00012.safetensors",
31
+ "model.layers.10.mlp.gate_proj.weight": "model-00003-of-00012.safetensors",
32
+ "model.layers.10.mlp.up_proj.weight": "model-00003-of-00012.safetensors",
33
+ "model.layers.10.post_attention_layernorm.weight": "model-00004-of-00012.safetensors",
34
+ "model.layers.10.post_feedforward_layernorm.weight": "model-00004-of-00012.safetensors",
35
+ "model.layers.10.pre_feedforward_layernorm.weight": "model-00004-of-00012.safetensors",
36
+ "model.layers.10.self_attn.k_proj.weight": "model-00003-of-00012.safetensors",
37
+ "model.layers.10.self_attn.o_proj.weight": "model-00003-of-00012.safetensors",
38
+ "model.layers.10.self_attn.q_proj.weight": "model-00003-of-00012.safetensors",
39
+ "model.layers.10.self_attn.v_proj.weight": "model-00003-of-00012.safetensors",
40
+ "model.layers.11.input_layernorm.weight": "model-00004-of-00012.safetensors",
41
+ "model.layers.11.mlp.down_proj.weight": "model-00004-of-00012.safetensors",
42
+ "model.layers.11.mlp.gate_proj.weight": "model-00004-of-00012.safetensors",
43
+ "model.layers.11.mlp.up_proj.weight": "model-00004-of-00012.safetensors",
44
+ "model.layers.11.post_attention_layernorm.weight": "model-00004-of-00012.safetensors",
45
+ "model.layers.11.post_feedforward_layernorm.weight": "model-00004-of-00012.safetensors",
46
+ "model.layers.11.pre_feedforward_layernorm.weight": "model-00004-of-00012.safetensors",
47
+ "model.layers.11.self_attn.k_proj.weight": "model-00004-of-00012.safetensors",
48
+ "model.layers.11.self_attn.o_proj.weight": "model-00004-of-00012.safetensors",
49
+ "model.layers.11.self_attn.q_proj.weight": "model-00004-of-00012.safetensors",
50
+ "model.layers.11.self_attn.v_proj.weight": "model-00004-of-00012.safetensors",
51
+ "model.layers.12.input_layernorm.weight": "model-00004-of-00012.safetensors",
52
+ "model.layers.12.mlp.down_proj.weight": "model-00004-of-00012.safetensors",
53
+ "model.layers.12.mlp.gate_proj.weight": "model-00004-of-00012.safetensors",
54
+ "model.layers.12.mlp.up_proj.weight": "model-00004-of-00012.safetensors",
55
+ "model.layers.12.post_attention_layernorm.weight": "model-00004-of-00012.safetensors",
56
+ "model.layers.12.post_feedforward_layernorm.weight": "model-00004-of-00012.safetensors",
57
+ "model.layers.12.pre_feedforward_layernorm.weight": "model-00004-of-00012.safetensors",
58
+ "model.layers.12.self_attn.k_proj.weight": "model-00004-of-00012.safetensors",
59
+ "model.layers.12.self_attn.o_proj.weight": "model-00004-of-00012.safetensors",
60
+ "model.layers.12.self_attn.q_proj.weight": "model-00004-of-00012.safetensors",
61
+ "model.layers.12.self_attn.v_proj.weight": "model-00004-of-00012.safetensors",
62
+ "model.layers.13.input_layernorm.weight": "model-00004-of-00012.safetensors",
63
+ "model.layers.13.mlp.down_proj.weight": "model-00004-of-00012.safetensors",
64
+ "model.layers.13.mlp.gate_proj.weight": "model-00004-of-00012.safetensors",
65
+ "model.layers.13.mlp.up_proj.weight": "model-00004-of-00012.safetensors",
66
+ "model.layers.13.post_attention_layernorm.weight": "model-00004-of-00012.safetensors",
67
+ "model.layers.13.post_feedforward_layernorm.weight": "model-00004-of-00012.safetensors",
68
+ "model.layers.13.pre_feedforward_layernorm.weight": "model-00004-of-00012.safetensors",
69
+ "model.layers.13.self_attn.k_proj.weight": "model-00004-of-00012.safetensors",
70
+ "model.layers.13.self_attn.o_proj.weight": "model-00004-of-00012.safetensors",
71
+ "model.layers.13.self_attn.q_proj.weight": "model-00004-of-00012.safetensors",
72
+ "model.layers.13.self_attn.v_proj.weight": "model-00004-of-00012.safetensors",
73
+ "model.layers.14.input_layernorm.weight": "model-00004-of-00012.safetensors",
74
+ "model.layers.14.mlp.down_proj.weight": "model-00004-of-00012.safetensors",
75
+ "model.layers.14.mlp.gate_proj.weight": "model-00004-of-00012.safetensors",
76
+ "model.layers.14.mlp.up_proj.weight": "model-00004-of-00012.safetensors",
77
+ "model.layers.14.post_attention_layernorm.weight": "model-00004-of-00012.safetensors",
78
+ "model.layers.14.post_feedforward_layernorm.weight": "model-00004-of-00012.safetensors",
79
+ "model.layers.14.pre_feedforward_layernorm.weight": "model-00004-of-00012.safetensors",
80
+ "model.layers.14.self_attn.k_proj.weight": "model-00004-of-00012.safetensors",
81
+ "model.layers.14.self_attn.o_proj.weight": "model-00004-of-00012.safetensors",
82
+ "model.layers.14.self_attn.q_proj.weight": "model-00004-of-00012.safetensors",
83
+ "model.layers.14.self_attn.v_proj.weight": "model-00004-of-00012.safetensors",
84
+ "model.layers.15.input_layernorm.weight": "model-00005-of-00012.safetensors",
85
+ "model.layers.15.mlp.down_proj.weight": "model-00005-of-00012.safetensors",
86
+ "model.layers.15.mlp.gate_proj.weight": "model-00005-of-00012.safetensors",
87
+ "model.layers.15.mlp.up_proj.weight": "model-00005-of-00012.safetensors",
88
+ "model.layers.15.post_attention_layernorm.weight": "model-00005-of-00012.safetensors",
89
+ "model.layers.15.post_feedforward_layernorm.weight": "model-00005-of-00012.safetensors",
90
+ "model.layers.15.pre_feedforward_layernorm.weight": "model-00005-of-00012.safetensors",
91
+ "model.layers.15.self_attn.k_proj.weight": "model-00004-of-00012.safetensors",
92
+ "model.layers.15.self_attn.o_proj.weight": "model-00004-of-00012.safetensors",
93
+ "model.layers.15.self_attn.q_proj.weight": "model-00004-of-00012.safetensors",
94
+ "model.layers.15.self_attn.v_proj.weight": "model-00004-of-00012.safetensors",
95
+ "model.layers.16.input_layernorm.weight": "model-00005-of-00012.safetensors",
96
+ "model.layers.16.mlp.down_proj.weight": "model-00005-of-00012.safetensors",
97
+ "model.layers.16.mlp.gate_proj.weight": "model-00005-of-00012.safetensors",
98
+ "model.layers.16.mlp.up_proj.weight": "model-00005-of-00012.safetensors",
99
+ "model.layers.16.post_attention_layernorm.weight": "model-00005-of-00012.safetensors",
100
+ "model.layers.16.post_feedforward_layernorm.weight": "model-00005-of-00012.safetensors",
101
+ "model.layers.16.pre_feedforward_layernorm.weight": "model-00005-of-00012.safetensors",
102
+ "model.layers.16.self_attn.k_proj.weight": "model-00005-of-00012.safetensors",
103
+ "model.layers.16.self_attn.o_proj.weight": "model-00005-of-00012.safetensors",
104
+ "model.layers.16.self_attn.q_proj.weight": "model-00005-of-00012.safetensors",
105
+ "model.layers.16.self_attn.v_proj.weight": "model-00005-of-00012.safetensors",
106
+ "model.layers.17.input_layernorm.weight": "model-00005-of-00012.safetensors",
107
+ "model.layers.17.mlp.down_proj.weight": "model-00005-of-00012.safetensors",
108
+ "model.layers.17.mlp.gate_proj.weight": "model-00005-of-00012.safetensors",
109
+ "model.layers.17.mlp.up_proj.weight": "model-00005-of-00012.safetensors",
110
+ "model.layers.17.post_attention_layernorm.weight": "model-00005-of-00012.safetensors",
111
+ "model.layers.17.post_feedforward_layernorm.weight": "model-00005-of-00012.safetensors",
112
+ "model.layers.17.pre_feedforward_layernorm.weight": "model-00005-of-00012.safetensors",
113
+ "model.layers.17.self_attn.k_proj.weight": "model-00005-of-00012.safetensors",
114
+ "model.layers.17.self_attn.o_proj.weight": "model-00005-of-00012.safetensors",
115
+ "model.layers.17.self_attn.q_proj.weight": "model-00005-of-00012.safetensors",
116
+ "model.layers.17.self_attn.v_proj.weight": "model-00005-of-00012.safetensors",
117
+ "model.layers.18.input_layernorm.weight": "model-00005-of-00012.safetensors",
118
+ "model.layers.18.mlp.down_proj.weight": "model-00005-of-00012.safetensors",
119
+ "model.layers.18.mlp.gate_proj.weight": "model-00005-of-00012.safetensors",
120
+ "model.layers.18.mlp.up_proj.weight": "model-00005-of-00012.safetensors",
121
+ "model.layers.18.post_attention_layernorm.weight": "model-00005-of-00012.safetensors",
122
+ "model.layers.18.post_feedforward_layernorm.weight": "model-00005-of-00012.safetensors",
123
+ "model.layers.18.pre_feedforward_layernorm.weight": "model-00005-of-00012.safetensors",
124
+ "model.layers.18.self_attn.k_proj.weight": "model-00005-of-00012.safetensors",
125
+ "model.layers.18.self_attn.o_proj.weight": "model-00005-of-00012.safetensors",
126
+ "model.layers.18.self_attn.q_proj.weight": "model-00005-of-00012.safetensors",
127
+ "model.layers.18.self_attn.v_proj.weight": "model-00005-of-00012.safetensors",
128
+ "model.layers.19.input_layernorm.weight": "model-00006-of-00012.safetensors",
129
+ "model.layers.19.mlp.down_proj.weight": "model-00006-of-00012.safetensors",
130
+ "model.layers.19.mlp.gate_proj.weight": "model-00005-of-00012.safetensors",
131
+ "model.layers.19.mlp.up_proj.weight": "model-00006-of-00012.safetensors",
132
+ "model.layers.19.post_attention_layernorm.weight": "model-00006-of-00012.safetensors",
133
+ "model.layers.19.post_feedforward_layernorm.weight": "model-00006-of-00012.safetensors",
134
+ "model.layers.19.pre_feedforward_layernorm.weight": "model-00006-of-00012.safetensors",
135
+ "model.layers.19.self_attn.k_proj.weight": "model-00005-of-00012.safetensors",
136
+ "model.layers.19.self_attn.o_proj.weight": "model-00005-of-00012.safetensors",
137
+ "model.layers.19.self_attn.q_proj.weight": "model-00005-of-00012.safetensors",
138
+ "model.layers.19.self_attn.v_proj.weight": "model-00005-of-00012.safetensors",
139
+ "model.layers.2.input_layernorm.weight": "model-00002-of-00012.safetensors",
140
+ "model.layers.2.mlp.down_proj.weight": "model-00002-of-00012.safetensors",
141
+ "model.layers.2.mlp.gate_proj.weight": "model-00002-of-00012.safetensors",
142
+ "model.layers.2.mlp.up_proj.weight": "model-00002-of-00012.safetensors",
143
+ "model.layers.2.post_attention_layernorm.weight": "model-00002-of-00012.safetensors",
144
+ "model.layers.2.post_feedforward_layernorm.weight": "model-00002-of-00012.safetensors",
145
+ "model.layers.2.pre_feedforward_layernorm.weight": "model-00002-of-00012.safetensors",
146
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00012.safetensors",
147
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00012.safetensors",
148
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00012.safetensors",
149
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00012.safetensors",
150
+ "model.layers.20.input_layernorm.weight": "model-00006-of-00012.safetensors",
151
+ "model.layers.20.mlp.down_proj.weight": "model-00006-of-00012.safetensors",
152
+ "model.layers.20.mlp.gate_proj.weight": "model-00006-of-00012.safetensors",
153
+ "model.layers.20.mlp.up_proj.weight": "model-00006-of-00012.safetensors",
154
+ "model.layers.20.post_attention_layernorm.weight": "model-00006-of-00012.safetensors",
155
+ "model.layers.20.post_feedforward_layernorm.weight": "model-00006-of-00012.safetensors",
156
+ "model.layers.20.pre_feedforward_layernorm.weight": "model-00006-of-00012.safetensors",
157
+ "model.layers.20.self_attn.k_proj.weight": "model-00006-of-00012.safetensors",
158
+ "model.layers.20.self_attn.o_proj.weight": "model-00006-of-00012.safetensors",
159
+ "model.layers.20.self_attn.q_proj.weight": "model-00006-of-00012.safetensors",
160
+ "model.layers.20.self_attn.v_proj.weight": "model-00006-of-00012.safetensors",
161
+ "model.layers.21.input_layernorm.weight": "model-00006-of-00012.safetensors",
162
+ "model.layers.21.mlp.down_proj.weight": "model-00006-of-00012.safetensors",
163
+ "model.layers.21.mlp.gate_proj.weight": "model-00006-of-00012.safetensors",
164
+ "model.layers.21.mlp.up_proj.weight": "model-00006-of-00012.safetensors",
165
+ "model.layers.21.post_attention_layernorm.weight": "model-00006-of-00012.safetensors",
166
+ "model.layers.21.post_feedforward_layernorm.weight": "model-00006-of-00012.safetensors",
167
+ "model.layers.21.pre_feedforward_layernorm.weight": "model-00006-of-00012.safetensors",
168
+ "model.layers.21.self_attn.k_proj.weight": "model-00006-of-00012.safetensors",
169
+ "model.layers.21.self_attn.o_proj.weight": "model-00006-of-00012.safetensors",
170
+ "model.layers.21.self_attn.q_proj.weight": "model-00006-of-00012.safetensors",
171
+ "model.layers.21.self_attn.v_proj.weight": "model-00006-of-00012.safetensors",
172
+ "model.layers.22.input_layernorm.weight": "model-00006-of-00012.safetensors",
173
+ "model.layers.22.mlp.down_proj.weight": "model-00006-of-00012.safetensors",
174
+ "model.layers.22.mlp.gate_proj.weight": "model-00006-of-00012.safetensors",
175
+ "model.layers.22.mlp.up_proj.weight": "model-00006-of-00012.safetensors",
176
+ "model.layers.22.post_attention_layernorm.weight": "model-00006-of-00012.safetensors",
177
+ "model.layers.22.post_feedforward_layernorm.weight": "model-00006-of-00012.safetensors",
178
+ "model.layers.22.pre_feedforward_layernorm.weight": "model-00006-of-00012.safetensors",
179
+ "model.layers.22.self_attn.k_proj.weight": "model-00006-of-00012.safetensors",
180
+ "model.layers.22.self_attn.o_proj.weight": "model-00006-of-00012.safetensors",
181
+ "model.layers.22.self_attn.q_proj.weight": "model-00006-of-00012.safetensors",
182
+ "model.layers.22.self_attn.v_proj.weight": "model-00006-of-00012.safetensors",
183
+ "model.layers.23.input_layernorm.weight": "model-00007-of-00012.safetensors",
184
+ "model.layers.23.mlp.down_proj.weight": "model-00007-of-00012.safetensors",
185
+ "model.layers.23.mlp.gate_proj.weight": "model-00006-of-00012.safetensors",
186
+ "model.layers.23.mlp.up_proj.weight": "model-00006-of-00012.safetensors",
187
+ "model.layers.23.post_attention_layernorm.weight": "model-00007-of-00012.safetensors",
188
+ "model.layers.23.post_feedforward_layernorm.weight": "model-00007-of-00012.safetensors",
189
+ "model.layers.23.pre_feedforward_layernorm.weight": "model-00007-of-00012.safetensors",
190
+ "model.layers.23.self_attn.k_proj.weight": "model-00006-of-00012.safetensors",
191
+ "model.layers.23.self_attn.o_proj.weight": "model-00006-of-00012.safetensors",
192
+ "model.layers.23.self_attn.q_proj.weight": "model-00006-of-00012.safetensors",
193
+ "model.layers.23.self_attn.v_proj.weight": "model-00006-of-00012.safetensors",
194
+ "model.layers.24.input_layernorm.weight": "model-00007-of-00012.safetensors",
195
+ "model.layers.24.mlp.down_proj.weight": "model-00007-of-00012.safetensors",
196
+ "model.layers.24.mlp.gate_proj.weight": "model-00007-of-00012.safetensors",
197
+ "model.layers.24.mlp.up_proj.weight": "model-00007-of-00012.safetensors",
198
+ "model.layers.24.post_attention_layernorm.weight": "model-00007-of-00012.safetensors",
199
+ "model.layers.24.post_feedforward_layernorm.weight": "model-00007-of-00012.safetensors",
200
+ "model.layers.24.pre_feedforward_layernorm.weight": "model-00007-of-00012.safetensors",
201
+ "model.layers.24.self_attn.k_proj.weight": "model-00007-of-00012.safetensors",
202
+ "model.layers.24.self_attn.o_proj.weight": "model-00007-of-00012.safetensors",
203
+ "model.layers.24.self_attn.q_proj.weight": "model-00007-of-00012.safetensors",
204
+ "model.layers.24.self_attn.v_proj.weight": "model-00007-of-00012.safetensors",
205
+ "model.layers.25.input_layernorm.weight": "model-00007-of-00012.safetensors",
206
+ "model.layers.25.mlp.down_proj.weight": "model-00007-of-00012.safetensors",
207
+ "model.layers.25.mlp.gate_proj.weight": "model-00007-of-00012.safetensors",
208
+ "model.layers.25.mlp.up_proj.weight": "model-00007-of-00012.safetensors",
209
+ "model.layers.25.post_attention_layernorm.weight": "model-00007-of-00012.safetensors",
210
+ "model.layers.25.post_feedforward_layernorm.weight": "model-00007-of-00012.safetensors",
211
+ "model.layers.25.pre_feedforward_layernorm.weight": "model-00007-of-00012.safetensors",
212
+ "model.layers.25.self_attn.k_proj.weight": "model-00007-of-00012.safetensors",
213
+ "model.layers.25.self_attn.o_proj.weight": "model-00007-of-00012.safetensors",
214
+ "model.layers.25.self_attn.q_proj.weight": "model-00007-of-00012.safetensors",
215
+ "model.layers.25.self_attn.v_proj.weight": "model-00007-of-00012.safetensors",
216
+ "model.layers.26.input_layernorm.weight": "model-00007-of-00012.safetensors",
217
+ "model.layers.26.mlp.down_proj.weight": "model-00007-of-00012.safetensors",
218
+ "model.layers.26.mlp.gate_proj.weight": "model-00007-of-00012.safetensors",
219
+ "model.layers.26.mlp.up_proj.weight": "model-00007-of-00012.safetensors",
220
+ "model.layers.26.post_attention_layernorm.weight": "model-00007-of-00012.safetensors",
221
+ "model.layers.26.post_feedforward_layernorm.weight": "model-00007-of-00012.safetensors",
222
+ "model.layers.26.pre_feedforward_layernorm.weight": "model-00007-of-00012.safetensors",
223
+ "model.layers.26.self_attn.k_proj.weight": "model-00007-of-00012.safetensors",
224
+ "model.layers.26.self_attn.o_proj.weight": "model-00007-of-00012.safetensors",
225
+ "model.layers.26.self_attn.q_proj.weight": "model-00007-of-00012.safetensors",
226
+ "model.layers.26.self_attn.v_proj.weight": "model-00007-of-00012.safetensors",
227
+ "model.layers.27.input_layernorm.weight": "model-00007-of-00012.safetensors",
228
+ "model.layers.27.mlp.down_proj.weight": "model-00007-of-00012.safetensors",
229
+ "model.layers.27.mlp.gate_proj.weight": "model-00007-of-00012.safetensors",
230
+ "model.layers.27.mlp.up_proj.weight": "model-00007-of-00012.safetensors",
231
+ "model.layers.27.post_attention_layernorm.weight": "model-00007-of-00012.safetensors",
232
+ "model.layers.27.post_feedforward_layernorm.weight": "model-00007-of-00012.safetensors",
233
+ "model.layers.27.pre_feedforward_layernorm.weight": "model-00007-of-00012.safetensors",
234
+ "model.layers.27.self_attn.k_proj.weight": "model-00007-of-00012.safetensors",
235
+ "model.layers.27.self_attn.o_proj.weight": "model-00007-of-00012.safetensors",
236
+ "model.layers.27.self_attn.q_proj.weight": "model-00007-of-00012.safetensors",
237
+ "model.layers.27.self_attn.v_proj.weight": "model-00007-of-00012.safetensors",
238
+ "model.layers.28.input_layernorm.weight": "model-00008-of-00012.safetensors",
239
+ "model.layers.28.mlp.down_proj.weight": "model-00008-of-00012.safetensors",
240
+ "model.layers.28.mlp.gate_proj.weight": "model-00008-of-00012.safetensors",
241
+ "model.layers.28.mlp.up_proj.weight": "model-00008-of-00012.safetensors",
242
+ "model.layers.28.post_attention_layernorm.weight": "model-00008-of-00012.safetensors",
243
+ "model.layers.28.post_feedforward_layernorm.weight": "model-00008-of-00012.safetensors",
244
+ "model.layers.28.pre_feedforward_layernorm.weight": "model-00008-of-00012.safetensors",
245
+ "model.layers.28.self_attn.k_proj.weight": "model-00007-of-00012.safetensors",
246
+ "model.layers.28.self_attn.o_proj.weight": "model-00007-of-00012.safetensors",
247
+ "model.layers.28.self_attn.q_proj.weight": "model-00007-of-00012.safetensors",
248
+ "model.layers.28.self_attn.v_proj.weight": "model-00007-of-00012.safetensors",
249
+ "model.layers.29.input_layernorm.weight": "model-00008-of-00012.safetensors",
250
+ "model.layers.29.mlp.down_proj.weight": "model-00008-of-00012.safetensors",
251
+ "model.layers.29.mlp.gate_proj.weight": "model-00008-of-00012.safetensors",
252
+ "model.layers.29.mlp.up_proj.weight": "model-00008-of-00012.safetensors",
253
+ "model.layers.29.post_attention_layernorm.weight": "model-00008-of-00012.safetensors",
254
+ "model.layers.29.post_feedforward_layernorm.weight": "model-00008-of-00012.safetensors",
255
+ "model.layers.29.pre_feedforward_layernorm.weight": "model-00008-of-00012.safetensors",
256
+ "model.layers.29.self_attn.k_proj.weight": "model-00008-of-00012.safetensors",
257
+ "model.layers.29.self_attn.o_proj.weight": "model-00008-of-00012.safetensors",
258
+ "model.layers.29.self_attn.q_proj.weight": "model-00008-of-00012.safetensors",
259
+ "model.layers.29.self_attn.v_proj.weight": "model-00008-of-00012.safetensors",
260
+ "model.layers.3.input_layernorm.weight": "model-00002-of-00012.safetensors",
261
+ "model.layers.3.mlp.down_proj.weight": "model-00002-of-00012.safetensors",
262
+ "model.layers.3.mlp.gate_proj.weight": "model-00002-of-00012.safetensors",
263
+ "model.layers.3.mlp.up_proj.weight": "model-00002-of-00012.safetensors",
264
+ "model.layers.3.post_attention_layernorm.weight": "model-00002-of-00012.safetensors",
265
+ "model.layers.3.post_feedforward_layernorm.weight": "model-00002-of-00012.safetensors",
266
+ "model.layers.3.pre_feedforward_layernorm.weight": "model-00002-of-00012.safetensors",
267
+ "model.layers.3.self_attn.k_proj.weight": "model-00002-of-00012.safetensors",
268
+ "model.layers.3.self_attn.o_proj.weight": "model-00002-of-00012.safetensors",
269
+ "model.layers.3.self_attn.q_proj.weight": "model-00002-of-00012.safetensors",
270
+ "model.layers.3.self_attn.v_proj.weight": "model-00002-of-00012.safetensors",
271
+ "model.layers.30.input_layernorm.weight": "model-00008-of-00012.safetensors",
272
+ "model.layers.30.mlp.down_proj.weight": "model-00008-of-00012.safetensors",
273
+ "model.layers.30.mlp.gate_proj.weight": "model-00008-of-00012.safetensors",
274
+ "model.layers.30.mlp.up_proj.weight": "model-00008-of-00012.safetensors",
275
+ "model.layers.30.post_attention_layernorm.weight": "model-00008-of-00012.safetensors",
276
+ "model.layers.30.post_feedforward_layernorm.weight": "model-00008-of-00012.safetensors",
277
+ "model.layers.30.pre_feedforward_layernorm.weight": "model-00008-of-00012.safetensors",
278
+ "model.layers.30.self_attn.k_proj.weight": "model-00008-of-00012.safetensors",
279
+ "model.layers.30.self_attn.o_proj.weight": "model-00008-of-00012.safetensors",
280
+ "model.layers.30.self_attn.q_proj.weight": "model-00008-of-00012.safetensors",
281
+ "model.layers.30.self_attn.v_proj.weight": "model-00008-of-00012.safetensors",
282
+ "model.layers.31.input_layernorm.weight": "model-00008-of-00012.safetensors",
283
+ "model.layers.31.mlp.down_proj.weight": "model-00008-of-00012.safetensors",
284
+ "model.layers.31.mlp.gate_proj.weight": "model-00008-of-00012.safetensors",
285
+ "model.layers.31.mlp.up_proj.weight": "model-00008-of-00012.safetensors",
286
+ "model.layers.31.post_attention_layernorm.weight": "model-00008-of-00012.safetensors",
287
+ "model.layers.31.post_feedforward_layernorm.weight": "model-00008-of-00012.safetensors",
288
+ "model.layers.31.pre_feedforward_layernorm.weight": "model-00008-of-00012.safetensors",
289
+ "model.layers.31.self_attn.k_proj.weight": "model-00008-of-00012.safetensors",
290
+ "model.layers.31.self_attn.o_proj.weight": "model-00008-of-00012.safetensors",
291
+ "model.layers.31.self_attn.q_proj.weight": "model-00008-of-00012.safetensors",
292
+ "model.layers.31.self_attn.v_proj.weight": "model-00008-of-00012.safetensors",
293
+ "model.layers.32.input_layernorm.weight": "model-00009-of-00012.safetensors",
294
+ "model.layers.32.mlp.down_proj.weight": "model-00009-of-00012.safetensors",
295
+ "model.layers.32.mlp.gate_proj.weight": "model-00008-of-00012.safetensors",
296
+ "model.layers.32.mlp.up_proj.weight": "model-00009-of-00012.safetensors",
297
+ "model.layers.32.post_attention_layernorm.weight": "model-00009-of-00012.safetensors",
298
+ "model.layers.32.post_feedforward_layernorm.weight": "model-00009-of-00012.safetensors",
299
+ "model.layers.32.pre_feedforward_layernorm.weight": "model-00009-of-00012.safetensors",
300
+ "model.layers.32.self_attn.k_proj.weight": "model-00008-of-00012.safetensors",
301
+ "model.layers.32.self_attn.o_proj.weight": "model-00008-of-00012.safetensors",
302
+ "model.layers.32.self_attn.q_proj.weight": "model-00008-of-00012.safetensors",
303
+ "model.layers.32.self_attn.v_proj.weight": "model-00008-of-00012.safetensors",
304
+ "model.layers.33.input_layernorm.weight": "model-00009-of-00012.safetensors",
305
+ "model.layers.33.mlp.down_proj.weight": "model-00009-of-00012.safetensors",
306
+ "model.layers.33.mlp.gate_proj.weight": "model-00009-of-00012.safetensors",
307
+ "model.layers.33.mlp.up_proj.weight": "model-00009-of-00012.safetensors",
308
+ "model.layers.33.post_attention_layernorm.weight": "model-00009-of-00012.safetensors",
309
+ "model.layers.33.post_feedforward_layernorm.weight": "model-00009-of-00012.safetensors",
310
+ "model.layers.33.pre_feedforward_layernorm.weight": "model-00009-of-00012.safetensors",
311
+ "model.layers.33.self_attn.k_proj.weight": "model-00009-of-00012.safetensors",
312
+ "model.layers.33.self_attn.o_proj.weight": "model-00009-of-00012.safetensors",
313
+ "model.layers.33.self_attn.q_proj.weight": "model-00009-of-00012.safetensors",
314
+ "model.layers.33.self_attn.v_proj.weight": "model-00009-of-00012.safetensors",
315
+ "model.layers.34.input_layernorm.weight": "model-00009-of-00012.safetensors",
316
+ "model.layers.34.mlp.down_proj.weight": "model-00009-of-00012.safetensors",
317
+ "model.layers.34.mlp.gate_proj.weight": "model-00009-of-00012.safetensors",
318
+ "model.layers.34.mlp.up_proj.weight": "model-00009-of-00012.safetensors",
319
+ "model.layers.34.post_attention_layernorm.weight": "model-00009-of-00012.safetensors",
320
+ "model.layers.34.post_feedforward_layernorm.weight": "model-00009-of-00012.safetensors",
321
+ "model.layers.34.pre_feedforward_layernorm.weight": "model-00009-of-00012.safetensors",
322
+ "model.layers.34.self_attn.k_proj.weight": "model-00009-of-00012.safetensors",
323
+ "model.layers.34.self_attn.o_proj.weight": "model-00009-of-00012.safetensors",
324
+ "model.layers.34.self_attn.q_proj.weight": "model-00009-of-00012.safetensors",
325
+ "model.layers.34.self_attn.v_proj.weight": "model-00009-of-00012.safetensors",
326
+ "model.layers.35.input_layernorm.weight": "model-00009-of-00012.safetensors",
327
+ "model.layers.35.mlp.down_proj.weight": "model-00009-of-00012.safetensors",
328
+ "model.layers.35.mlp.gate_proj.weight": "model-00009-of-00012.safetensors",
329
+ "model.layers.35.mlp.up_proj.weight": "model-00009-of-00012.safetensors",
330
+ "model.layers.35.post_attention_layernorm.weight": "model-00009-of-00012.safetensors",
331
+ "model.layers.35.post_feedforward_layernorm.weight": "model-00009-of-00012.safetensors",
332
+ "model.layers.35.pre_feedforward_layernorm.weight": "model-00009-of-00012.safetensors",
333
+ "model.layers.35.self_attn.k_proj.weight": "model-00009-of-00012.safetensors",
334
+ "model.layers.35.self_attn.o_proj.weight": "model-00009-of-00012.safetensors",
335
+ "model.layers.35.self_attn.q_proj.weight": "model-00009-of-00012.safetensors",
336
+ "model.layers.35.self_attn.v_proj.weight": "model-00009-of-00012.safetensors",
337
+ "model.layers.36.input_layernorm.weight": "model-00010-of-00012.safetensors",
338
+ "model.layers.36.mlp.down_proj.weight": "model-00010-of-00012.safetensors",
339
+ "model.layers.36.mlp.gate_proj.weight": "model-00009-of-00012.safetensors",
340
+ "model.layers.36.mlp.up_proj.weight": "model-00009-of-00012.safetensors",
341
+ "model.layers.36.post_attention_layernorm.weight": "model-00010-of-00012.safetensors",
342
+ "model.layers.36.post_feedforward_layernorm.weight": "model-00010-of-00012.safetensors",
343
+ "model.layers.36.pre_feedforward_layernorm.weight": "model-00010-of-00012.safetensors",
344
+ "model.layers.36.self_attn.k_proj.weight": "model-00009-of-00012.safetensors",
345
+ "model.layers.36.self_attn.o_proj.weight": "model-00009-of-00012.safetensors",
346
+ "model.layers.36.self_attn.q_proj.weight": "model-00009-of-00012.safetensors",
347
+ "model.layers.36.self_attn.v_proj.weight": "model-00009-of-00012.safetensors",
348
+ "model.layers.37.input_layernorm.weight": "model-00010-of-00012.safetensors",
349
+ "model.layers.37.mlp.down_proj.weight": "model-00010-of-00012.safetensors",
350
+ "model.layers.37.mlp.gate_proj.weight": "model-00010-of-00012.safetensors",
351
+ "model.layers.37.mlp.up_proj.weight": "model-00010-of-00012.safetensors",
352
+ "model.layers.37.post_attention_layernorm.weight": "model-00010-of-00012.safetensors",
353
+ "model.layers.37.post_feedforward_layernorm.weight": "model-00010-of-00012.safetensors",
354
+ "model.layers.37.pre_feedforward_layernorm.weight": "model-00010-of-00012.safetensors",
355
+ "model.layers.37.self_attn.k_proj.weight": "model-00010-of-00012.safetensors",
356
+ "model.layers.37.self_attn.o_proj.weight": "model-00010-of-00012.safetensors",
357
+ "model.layers.37.self_attn.q_proj.weight": "model-00010-of-00012.safetensors",
358
+ "model.layers.37.self_attn.v_proj.weight": "model-00010-of-00012.safetensors",
359
+ "model.layers.38.input_layernorm.weight": "model-00010-of-00012.safetensors",
360
+ "model.layers.38.mlp.down_proj.weight": "model-00010-of-00012.safetensors",
361
+ "model.layers.38.mlp.gate_proj.weight": "model-00010-of-00012.safetensors",
362
+ "model.layers.38.mlp.up_proj.weight": "model-00010-of-00012.safetensors",
363
+ "model.layers.38.post_attention_layernorm.weight": "model-00010-of-00012.safetensors",
364
+ "model.layers.38.post_feedforward_layernorm.weight": "model-00010-of-00012.safetensors",
365
+ "model.layers.38.pre_feedforward_layernorm.weight": "model-00010-of-00012.safetensors",
366
+ "model.layers.38.self_attn.k_proj.weight": "model-00010-of-00012.safetensors",
367
+ "model.layers.38.self_attn.o_proj.weight": "model-00010-of-00012.safetensors",
368
+ "model.layers.38.self_attn.q_proj.weight": "model-00010-of-00012.safetensors",
369
+ "model.layers.38.self_attn.v_proj.weight": "model-00010-of-00012.safetensors",
370
+ "model.layers.39.input_layernorm.weight": "model-00010-of-00012.safetensors",
371
+ "model.layers.39.mlp.down_proj.weight": "model-00010-of-00012.safetensors",
372
+ "model.layers.39.mlp.gate_proj.weight": "model-00010-of-00012.safetensors",
373
+ "model.layers.39.mlp.up_proj.weight": "model-00010-of-00012.safetensors",
374
+ "model.layers.39.post_attention_layernorm.weight": "model-00010-of-00012.safetensors",
375
+ "model.layers.39.post_feedforward_layernorm.weight": "model-00010-of-00012.safetensors",
376
+ "model.layers.39.pre_feedforward_layernorm.weight": "model-00010-of-00012.safetensors",
377
+ "model.layers.39.self_attn.k_proj.weight": "model-00010-of-00012.safetensors",
378
+ "model.layers.39.self_attn.o_proj.weight": "model-00010-of-00012.safetensors",
379
+ "model.layers.39.self_attn.q_proj.weight": "model-00010-of-00012.safetensors",
380
+ "model.layers.39.self_attn.v_proj.weight": "model-00010-of-00012.safetensors",
381
+ "model.layers.4.input_layernorm.weight": "model-00002-of-00012.safetensors",
382
+ "model.layers.4.mlp.down_proj.weight": "model-00002-of-00012.safetensors",
383
+ "model.layers.4.mlp.gate_proj.weight": "model-00002-of-00012.safetensors",
384
+ "model.layers.4.mlp.up_proj.weight": "model-00002-of-00012.safetensors",
385
+ "model.layers.4.post_attention_layernorm.weight": "model-00002-of-00012.safetensors",
386
+ "model.layers.4.post_feedforward_layernorm.weight": "model-00002-of-00012.safetensors",
387
+ "model.layers.4.pre_feedforward_layernorm.weight": "model-00002-of-00012.safetensors",
388
+ "model.layers.4.self_attn.k_proj.weight": "model-00002-of-00012.safetensors",
389
+ "model.layers.4.self_attn.o_proj.weight": "model-00002-of-00012.safetensors",
390
+ "model.layers.4.self_attn.q_proj.weight": "model-00002-of-00012.safetensors",
391
+ "model.layers.4.self_attn.v_proj.weight": "model-00002-of-00012.safetensors",
392
+ "model.layers.40.input_layernorm.weight": "model-00010-of-00012.safetensors",
393
+ "model.layers.40.mlp.down_proj.weight": "model-00010-of-00012.safetensors",
394
+ "model.layers.40.mlp.gate_proj.weight": "model-00010-of-00012.safetensors",
395
+ "model.layers.40.mlp.up_proj.weight": "model-00010-of-00012.safetensors",
396
+ "model.layers.40.post_attention_layernorm.weight": "model-00010-of-00012.safetensors",
397
+ "model.layers.40.post_feedforward_layernorm.weight": "model-00010-of-00012.safetensors",
398
+ "model.layers.40.pre_feedforward_layernorm.weight": "model-00010-of-00012.safetensors",
399
+ "model.layers.40.self_attn.k_proj.weight": "model-00010-of-00012.safetensors",
400
+ "model.layers.40.self_attn.o_proj.weight": "model-00010-of-00012.safetensors",
401
+ "model.layers.40.self_attn.q_proj.weight": "model-00010-of-00012.safetensors",
402
+ "model.layers.40.self_attn.v_proj.weight": "model-00010-of-00012.safetensors",
403
+ "model.layers.41.input_layernorm.weight": "model-00011-of-00012.safetensors",
404
+ "model.layers.41.mlp.down_proj.weight": "model-00011-of-00012.safetensors",
405
+ "model.layers.41.mlp.gate_proj.weight": "model-00011-of-00012.safetensors",
406
+ "model.layers.41.mlp.up_proj.weight": "model-00011-of-00012.safetensors",
407
+ "model.layers.41.post_attention_layernorm.weight": "model-00011-of-00012.safetensors",
408
+ "model.layers.41.post_feedforward_layernorm.weight": "model-00011-of-00012.safetensors",
409
+ "model.layers.41.pre_feedforward_layernorm.weight": "model-00011-of-00012.safetensors",
410
+ "model.layers.41.self_attn.k_proj.weight": "model-00010-of-00012.safetensors",
411
+ "model.layers.41.self_attn.o_proj.weight": "model-00010-of-00012.safetensors",
412
+ "model.layers.41.self_attn.q_proj.weight": "model-00010-of-00012.safetensors",
413
+ "model.layers.41.self_attn.v_proj.weight": "model-00010-of-00012.safetensors",
414
+ "model.layers.42.input_layernorm.weight": "model-00011-of-00012.safetensors",
415
+ "model.layers.42.mlp.down_proj.weight": "model-00011-of-00012.safetensors",
416
+ "model.layers.42.mlp.gate_proj.weight": "model-00011-of-00012.safetensors",
417
+ "model.layers.42.mlp.up_proj.weight": "model-00011-of-00012.safetensors",
418
+ "model.layers.42.post_attention_layernorm.weight": "model-00011-of-00012.safetensors",
419
+ "model.layers.42.post_feedforward_layernorm.weight": "model-00011-of-00012.safetensors",
420
+ "model.layers.42.pre_feedforward_layernorm.weight": "model-00011-of-00012.safetensors",
421
+ "model.layers.42.self_attn.k_proj.weight": "model-00011-of-00012.safetensors",
422
+ "model.layers.42.self_attn.o_proj.weight": "model-00011-of-00012.safetensors",
423
+ "model.layers.42.self_attn.q_proj.weight": "model-00011-of-00012.safetensors",
424
+ "model.layers.42.self_attn.v_proj.weight": "model-00011-of-00012.safetensors",
425
+ "model.layers.43.input_layernorm.weight": "model-00011-of-00012.safetensors",
426
+ "model.layers.43.mlp.down_proj.weight": "model-00011-of-00012.safetensors",
427
+ "model.layers.43.mlp.gate_proj.weight": "model-00011-of-00012.safetensors",
428
+ "model.layers.43.mlp.up_proj.weight": "model-00011-of-00012.safetensors",
429
+ "model.layers.43.post_attention_layernorm.weight": "model-00011-of-00012.safetensors",
430
+ "model.layers.43.post_feedforward_layernorm.weight": "model-00011-of-00012.safetensors",
431
+ "model.layers.43.pre_feedforward_layernorm.weight": "model-00011-of-00012.safetensors",
432
+ "model.layers.43.self_attn.k_proj.weight": "model-00011-of-00012.safetensors",
433
+ "model.layers.43.self_attn.o_proj.weight": "model-00011-of-00012.safetensors",
434
+ "model.layers.43.self_attn.q_proj.weight": "model-00011-of-00012.safetensors",
435
+ "model.layers.43.self_attn.v_proj.weight": "model-00011-of-00012.safetensors",
436
+ "model.layers.44.input_layernorm.weight": "model-00011-of-00012.safetensors",
437
+ "model.layers.44.mlp.down_proj.weight": "model-00011-of-00012.safetensors",
438
+ "model.layers.44.mlp.gate_proj.weight": "model-00011-of-00012.safetensors",
439
+ "model.layers.44.mlp.up_proj.weight": "model-00011-of-00012.safetensors",
440
+ "model.layers.44.post_attention_layernorm.weight": "model-00011-of-00012.safetensors",
441
+ "model.layers.44.post_feedforward_layernorm.weight": "model-00011-of-00012.safetensors",
442
+ "model.layers.44.pre_feedforward_layernorm.weight": "model-00011-of-00012.safetensors",
443
+ "model.layers.44.self_attn.k_proj.weight": "model-00011-of-00012.safetensors",
444
+ "model.layers.44.self_attn.o_proj.weight": "model-00011-of-00012.safetensors",
445
+ "model.layers.44.self_attn.q_proj.weight": "model-00011-of-00012.safetensors",
446
+ "model.layers.44.self_attn.v_proj.weight": "model-00011-of-00012.safetensors",
447
+ "model.layers.45.input_layernorm.weight": "model-00012-of-00012.safetensors",
448
+ "model.layers.45.mlp.down_proj.weight": "model-00012-of-00012.safetensors",
449
+ "model.layers.45.mlp.gate_proj.weight": "model-00011-of-00012.safetensors",
450
+ "model.layers.45.mlp.up_proj.weight": "model-00012-of-00012.safetensors",
451
+ "model.layers.45.post_attention_layernorm.weight": "model-00012-of-00012.safetensors",
452
+ "model.layers.45.post_feedforward_layernorm.weight": "model-00012-of-00012.safetensors",
453
+ "model.layers.45.pre_feedforward_layernorm.weight": "model-00012-of-00012.safetensors",
454
+ "model.layers.45.self_attn.k_proj.weight": "model-00011-of-00012.safetensors",
455
+ "model.layers.45.self_attn.o_proj.weight": "model-00011-of-00012.safetensors",
456
+ "model.layers.45.self_attn.q_proj.weight": "model-00011-of-00012.safetensors",
457
+ "model.layers.45.self_attn.v_proj.weight": "model-00011-of-00012.safetensors",
458
+ "model.layers.5.input_layernorm.weight": "model-00002-of-00012.safetensors",
459
+ "model.layers.5.mlp.down_proj.weight": "model-00002-of-00012.safetensors",
460
+ "model.layers.5.mlp.gate_proj.weight": "model-00002-of-00012.safetensors",
461
+ "model.layers.5.mlp.up_proj.weight": "model-00002-of-00012.safetensors",
462
+ "model.layers.5.post_attention_layernorm.weight": "model-00002-of-00012.safetensors",
463
+ "model.layers.5.post_feedforward_layernorm.weight": "model-00002-of-00012.safetensors",
464
+ "model.layers.5.pre_feedforward_layernorm.weight": "model-00002-of-00012.safetensors",
465
+ "model.layers.5.self_attn.k_proj.weight": "model-00002-of-00012.safetensors",
466
+ "model.layers.5.self_attn.o_proj.weight": "model-00002-of-00012.safetensors",
467
+ "model.layers.5.self_attn.q_proj.weight": "model-00002-of-00012.safetensors",
468
+ "model.layers.5.self_attn.v_proj.weight": "model-00002-of-00012.safetensors",
469
+ "model.layers.6.input_layernorm.weight": "model-00003-of-00012.safetensors",
470
+ "model.layers.6.mlp.down_proj.weight": "model-00003-of-00012.safetensors",
471
+ "model.layers.6.mlp.gate_proj.weight": "model-00002-of-00012.safetensors",
472
+ "model.layers.6.mlp.up_proj.weight": "model-00003-of-00012.safetensors",
473
+ "model.layers.6.post_attention_layernorm.weight": "model-00003-of-00012.safetensors",
474
+ "model.layers.6.post_feedforward_layernorm.weight": "model-00003-of-00012.safetensors",
475
+ "model.layers.6.pre_feedforward_layernorm.weight": "model-00003-of-00012.safetensors",
476
+ "model.layers.6.self_attn.k_proj.weight": "model-00002-of-00012.safetensors",
477
+ "model.layers.6.self_attn.o_proj.weight": "model-00002-of-00012.safetensors",
478
+ "model.layers.6.self_attn.q_proj.weight": "model-00002-of-00012.safetensors",
479
+ "model.layers.6.self_attn.v_proj.weight": "model-00002-of-00012.safetensors",
480
+ "model.layers.7.input_layernorm.weight": "model-00003-of-00012.safetensors",
481
+ "model.layers.7.mlp.down_proj.weight": "model-00003-of-00012.safetensors",
482
+ "model.layers.7.mlp.gate_proj.weight": "model-00003-of-00012.safetensors",
483
+ "model.layers.7.mlp.up_proj.weight": "model-00003-of-00012.safetensors",
484
+ "model.layers.7.post_attention_layernorm.weight": "model-00003-of-00012.safetensors",
485
+ "model.layers.7.post_feedforward_layernorm.weight": "model-00003-of-00012.safetensors",
486
+ "model.layers.7.pre_feedforward_layernorm.weight": "model-00003-of-00012.safetensors",
487
+ "model.layers.7.self_attn.k_proj.weight": "model-00003-of-00012.safetensors",
488
+ "model.layers.7.self_attn.o_proj.weight": "model-00003-of-00012.safetensors",
489
+ "model.layers.7.self_attn.q_proj.weight": "model-00003-of-00012.safetensors",
490
+ "model.layers.7.self_attn.v_proj.weight": "model-00003-of-00012.safetensors",
491
+ "model.layers.8.input_layernorm.weight": "model-00003-of-00012.safetensors",
492
+ "model.layers.8.mlp.down_proj.weight": "model-00003-of-00012.safetensors",
493
+ "model.layers.8.mlp.gate_proj.weight": "model-00003-of-00012.safetensors",
494
+ "model.layers.8.mlp.up_proj.weight": "model-00003-of-00012.safetensors",
495
+ "model.layers.8.post_attention_layernorm.weight": "model-00003-of-00012.safetensors",
496
+ "model.layers.8.post_feedforward_layernorm.weight": "model-00003-of-00012.safetensors",
497
+ "model.layers.8.pre_feedforward_layernorm.weight": "model-00003-of-00012.safetensors",
498
+ "model.layers.8.self_attn.k_proj.weight": "model-00003-of-00012.safetensors",
499
+ "model.layers.8.self_attn.o_proj.weight": "model-00003-of-00012.safetensors",
500
+ "model.layers.8.self_attn.q_proj.weight": "model-00003-of-00012.safetensors",
501
+ "model.layers.8.self_attn.v_proj.weight": "model-00003-of-00012.safetensors",
502
+ "model.layers.9.input_layernorm.weight": "model-00003-of-00012.safetensors",
503
+ "model.layers.9.mlp.down_proj.weight": "model-00003-of-00012.safetensors",
504
+ "model.layers.9.mlp.gate_proj.weight": "model-00003-of-00012.safetensors",
505
+ "model.layers.9.mlp.up_proj.weight": "model-00003-of-00012.safetensors",
506
+ "model.layers.9.post_attention_layernorm.weight": "model-00003-of-00012.safetensors",
507
+ "model.layers.9.post_feedforward_layernorm.weight": "model-00003-of-00012.safetensors",
508
+ "model.layers.9.pre_feedforward_layernorm.weight": "model-00003-of-00012.safetensors",
509
+ "model.layers.9.self_attn.k_proj.weight": "model-00003-of-00012.safetensors",
510
+ "model.layers.9.self_attn.o_proj.weight": "model-00003-of-00012.safetensors",
511
+ "model.layers.9.self_attn.q_proj.weight": "model-00003-of-00012.safetensors",
512
+ "model.layers.9.self_attn.v_proj.weight": "model-00003-of-00012.safetensors",
513
+ "model.norm.weight": "model-00012-of-00012.safetensors"
514
+ }
515
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<start_of_turn>",
4
+ "<end_of_turn>"
5
+ ],
6
+ "bos_token": {
7
+ "content": "<bos>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false
12
+ },
13
+ "eos_token": {
14
+ "content": "<eos>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ },
20
+ "pad_token": {
21
+ "content": "<pad>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false
26
+ },
27
+ "unk_token": {
28
+ "content": "<unk>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false
33
+ }
34
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f289bc05132635a8bc7aca7aa21255efd5e18f3710f43e3cdb96bcd41be4922
3
+ size 17525357
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61a7b147390c64585d6c3543dd6fc636906c9af3865a5548f27f31aee1d4c8e2
3
+ size 4241003
tokenizer_config.json ADDED
@@ -0,0 +1,2017 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<pad>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "1": {
14
+ "content": "<eos>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "2": {
22
+ "content": "<bos>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "3": {
30
+ "content": "<unk>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "4": {
38
+ "content": "<mask>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": false
44
+ },
45
+ "5": {
46
+ "content": "<2mass>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": false
52
+ },
53
+ "6": {
54
+ "content": "[@BOS@]",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": false
60
+ },
61
+ "7": {
62
+ "content": "<unused0>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": false
68
+ },
69
+ "8": {
70
+ "content": "<unused1>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": false
76
+ },
77
+ "9": {
78
+ "content": "<unused2>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": false
84
+ },
85
+ "10": {
86
+ "content": "<unused3>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": false
92
+ },
93
+ "11": {
94
+ "content": "<unused4>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": false
100
+ },
101
+ "12": {
102
+ "content": "<unused5>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": false
108
+ },
109
+ "13": {
110
+ "content": "<unused6>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": false
116
+ },
117
+ "14": {
118
+ "content": "<unused7>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "15": {
126
+ "content": "<unused8>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "16": {
134
+ "content": "<unused9>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "17": {
142
+ "content": "<unused10>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "18": {
150
+ "content": "<unused11>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "19": {
158
+ "content": "<unused12>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "20": {
166
+ "content": "<unused13>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "21": {
174
+ "content": "<unused14>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ },
181
+ "22": {
182
+ "content": "<unused15>",
183
+ "lstrip": false,
184
+ "normalized": false,
185
+ "rstrip": false,
186
+ "single_word": false,
187
+ "special": false
188
+ },
189
+ "23": {
190
+ "content": "<unused16>",
191
+ "lstrip": false,
192
+ "normalized": false,
193
+ "rstrip": false,
194
+ "single_word": false,
195
+ "special": false
196
+ },
197
+ "24": {
198
+ "content": "<unused17>",
199
+ "lstrip": false,
200
+ "normalized": false,
201
+ "rstrip": false,
202
+ "single_word": false,
203
+ "special": false
204
+ },
205
+ "25": {
206
+ "content": "<unused18>",
207
+ "lstrip": false,
208
+ "normalized": false,
209
+ "rstrip": false,
210
+ "single_word": false,
211
+ "special": false
212
+ },
213
+ "26": {
214
+ "content": "<unused19>",
215
+ "lstrip": false,
216
+ "normalized": false,
217
+ "rstrip": false,
218
+ "single_word": false,
219
+ "special": false
220
+ },
221
+ "27": {
222
+ "content": "<unused20>",
223
+ "lstrip": false,
224
+ "normalized": false,
225
+ "rstrip": false,
226
+ "single_word": false,
227
+ "special": false
228
+ },
229
+ "28": {
230
+ "content": "<unused21>",
231
+ "lstrip": false,
232
+ "normalized": false,
233
+ "rstrip": false,
234
+ "single_word": false,
235
+ "special": false
236
+ },
237
+ "29": {
238
+ "content": "<unused22>",
239
+ "lstrip": false,
240
+ "normalized": false,
241
+ "rstrip": false,
242
+ "single_word": false,
243
+ "special": false
244
+ },
245
+ "30": {
246
+ "content": "<unused23>",
247
+ "lstrip": false,
248
+ "normalized": false,
249
+ "rstrip": false,
250
+ "single_word": false,
251
+ "special": false
252
+ },
253
+ "31": {
254
+ "content": "<unused24>",
255
+ "lstrip": false,
256
+ "normalized": false,
257
+ "rstrip": false,
258
+ "single_word": false,
259
+ "special": false
260
+ },
261
+ "32": {
262
+ "content": "<unused25>",
263
+ "lstrip": false,
264
+ "normalized": false,
265
+ "rstrip": false,
266
+ "single_word": false,
267
+ "special": false
268
+ },
269
+ "33": {
270
+ "content": "<unused26>",
271
+ "lstrip": false,
272
+ "normalized": false,
273
+ "rstrip": false,
274
+ "single_word": false,
275
+ "special": false
276
+ },
277
+ "34": {
278
+ "content": "<unused27>",
279
+ "lstrip": false,
280
+ "normalized": false,
281
+ "rstrip": false,
282
+ "single_word": false,
283
+ "special": false
284
+ },
285
+ "35": {
286
+ "content": "<unused28>",
287
+ "lstrip": false,
288
+ "normalized": false,
289
+ "rstrip": false,
290
+ "single_word": false,
291
+ "special": false
292
+ },
293
+ "36": {
294
+ "content": "<unused29>",
295
+ "lstrip": false,
296
+ "normalized": false,
297
+ "rstrip": false,
298
+ "single_word": false,
299
+ "special": false
300
+ },
301
+ "37": {
302
+ "content": "<unused30>",
303
+ "lstrip": false,
304
+ "normalized": false,
305
+ "rstrip": false,
306
+ "single_word": false,
307
+ "special": false
308
+ },
309
+ "38": {
310
+ "content": "<unused31>",
311
+ "lstrip": false,
312
+ "normalized": false,
313
+ "rstrip": false,
314
+ "single_word": false,
315
+ "special": false
316
+ },
317
+ "39": {
318
+ "content": "<unused32>",
319
+ "lstrip": false,
320
+ "normalized": false,
321
+ "rstrip": false,
322
+ "single_word": false,
323
+ "special": false
324
+ },
325
+ "40": {
326
+ "content": "<unused33>",
327
+ "lstrip": false,
328
+ "normalized": false,
329
+ "rstrip": false,
330
+ "single_word": false,
331
+ "special": false
332
+ },
333
+ "41": {
334
+ "content": "<unused34>",
335
+ "lstrip": false,
336
+ "normalized": false,
337
+ "rstrip": false,
338
+ "single_word": false,
339
+ "special": false
340
+ },
341
+ "42": {
342
+ "content": "<unused35>",
343
+ "lstrip": false,
344
+ "normalized": false,
345
+ "rstrip": false,
346
+ "single_word": false,
347
+ "special": false
348
+ },
349
+ "43": {
350
+ "content": "<unused36>",
351
+ "lstrip": false,
352
+ "normalized": false,
353
+ "rstrip": false,
354
+ "single_word": false,
355
+ "special": false
356
+ },
357
+ "44": {
358
+ "content": "<unused37>",
359
+ "lstrip": false,
360
+ "normalized": false,
361
+ "rstrip": false,
362
+ "single_word": false,
363
+ "special": false
364
+ },
365
+ "45": {
366
+ "content": "<unused38>",
367
+ "lstrip": false,
368
+ "normalized": false,
369
+ "rstrip": false,
370
+ "single_word": false,
371
+ "special": false
372
+ },
373
+ "46": {
374
+ "content": "<unused39>",
375
+ "lstrip": false,
376
+ "normalized": false,
377
+ "rstrip": false,
378
+ "single_word": false,
379
+ "special": false
380
+ },
381
+ "47": {
382
+ "content": "<unused40>",
383
+ "lstrip": false,
384
+ "normalized": false,
385
+ "rstrip": false,
386
+ "single_word": false,
387
+ "special": false
388
+ },
389
+ "48": {
390
+ "content": "<unused41>",
391
+ "lstrip": false,
392
+ "normalized": false,
393
+ "rstrip": false,
394
+ "single_word": false,
395
+ "special": false
396
+ },
397
+ "49": {
398
+ "content": "<unused42>",
399
+ "lstrip": false,
400
+ "normalized": false,
401
+ "rstrip": false,
402
+ "single_word": false,
403
+ "special": false
404
+ },
405
+ "50": {
406
+ "content": "<unused43>",
407
+ "lstrip": false,
408
+ "normalized": false,
409
+ "rstrip": false,
410
+ "single_word": false,
411
+ "special": false
412
+ },
413
+ "51": {
414
+ "content": "<unused44>",
415
+ "lstrip": false,
416
+ "normalized": false,
417
+ "rstrip": false,
418
+ "single_word": false,
419
+ "special": false
420
+ },
421
+ "52": {
422
+ "content": "<unused45>",
423
+ "lstrip": false,
424
+ "normalized": false,
425
+ "rstrip": false,
426
+ "single_word": false,
427
+ "special": false
428
+ },
429
+ "53": {
430
+ "content": "<unused46>",
431
+ "lstrip": false,
432
+ "normalized": false,
433
+ "rstrip": false,
434
+ "single_word": false,
435
+ "special": false
436
+ },
437
+ "54": {
438
+ "content": "<unused47>",
439
+ "lstrip": false,
440
+ "normalized": false,
441
+ "rstrip": false,
442
+ "single_word": false,
443
+ "special": false
444
+ },
445
+ "55": {
446
+ "content": "<unused48>",
447
+ "lstrip": false,
448
+ "normalized": false,
449
+ "rstrip": false,
450
+ "single_word": false,
451
+ "special": false
452
+ },
453
+ "56": {
454
+ "content": "<unused49>",
455
+ "lstrip": false,
456
+ "normalized": false,
457
+ "rstrip": false,
458
+ "single_word": false,
459
+ "special": false
460
+ },
461
+ "57": {
462
+ "content": "<unused50>",
463
+ "lstrip": false,
464
+ "normalized": false,
465
+ "rstrip": false,
466
+ "single_word": false,
467
+ "special": false
468
+ },
469
+ "58": {
470
+ "content": "<unused51>",
471
+ "lstrip": false,
472
+ "normalized": false,
473
+ "rstrip": false,
474
+ "single_word": false,
475
+ "special": false
476
+ },
477
+ "59": {
478
+ "content": "<unused52>",
479
+ "lstrip": false,
480
+ "normalized": false,
481
+ "rstrip": false,
482
+ "single_word": false,
483
+ "special": false
484
+ },
485
+ "60": {
486
+ "content": "<unused53>",
487
+ "lstrip": false,
488
+ "normalized": false,
489
+ "rstrip": false,
490
+ "single_word": false,
491
+ "special": false
492
+ },
493
+ "61": {
494
+ "content": "<unused54>",
495
+ "lstrip": false,
496
+ "normalized": false,
497
+ "rstrip": false,
498
+ "single_word": false,
499
+ "special": false
500
+ },
501
+ "62": {
502
+ "content": "<unused55>",
503
+ "lstrip": false,
504
+ "normalized": false,
505
+ "rstrip": false,
506
+ "single_word": false,
507
+ "special": false
508
+ },
509
+ "63": {
510
+ "content": "<unused56>",
511
+ "lstrip": false,
512
+ "normalized": false,
513
+ "rstrip": false,
514
+ "single_word": false,
515
+ "special": false
516
+ },
517
+ "64": {
518
+ "content": "<unused57>",
519
+ "lstrip": false,
520
+ "normalized": false,
521
+ "rstrip": false,
522
+ "single_word": false,
523
+ "special": false
524
+ },
525
+ "65": {
526
+ "content": "<unused58>",
527
+ "lstrip": false,
528
+ "normalized": false,
529
+ "rstrip": false,
530
+ "single_word": false,
531
+ "special": false
532
+ },
533
+ "66": {
534
+ "content": "<unused59>",
535
+ "lstrip": false,
536
+ "normalized": false,
537
+ "rstrip": false,
538
+ "single_word": false,
539
+ "special": false
540
+ },
541
+ "67": {
542
+ "content": "<unused60>",
543
+ "lstrip": false,
544
+ "normalized": false,
545
+ "rstrip": false,
546
+ "single_word": false,
547
+ "special": false
548
+ },
549
+ "68": {
550
+ "content": "<unused61>",
551
+ "lstrip": false,
552
+ "normalized": false,
553
+ "rstrip": false,
554
+ "single_word": false,
555
+ "special": false
556
+ },
557
+ "69": {
558
+ "content": "<unused62>",
559
+ "lstrip": false,
560
+ "normalized": false,
561
+ "rstrip": false,
562
+ "single_word": false,
563
+ "special": false
564
+ },
565
+ "70": {
566
+ "content": "<unused63>",
567
+ "lstrip": false,
568
+ "normalized": false,
569
+ "rstrip": false,
570
+ "single_word": false,
571
+ "special": false
572
+ },
573
+ "71": {
574
+ "content": "<unused64>",
575
+ "lstrip": false,
576
+ "normalized": false,
577
+ "rstrip": false,
578
+ "single_word": false,
579
+ "special": false
580
+ },
581
+ "72": {
582
+ "content": "<unused65>",
583
+ "lstrip": false,
584
+ "normalized": false,
585
+ "rstrip": false,
586
+ "single_word": false,
587
+ "special": false
588
+ },
589
+ "73": {
590
+ "content": "<unused66>",
591
+ "lstrip": false,
592
+ "normalized": false,
593
+ "rstrip": false,
594
+ "single_word": false,
595
+ "special": false
596
+ },
597
+ "74": {
598
+ "content": "<unused67>",
599
+ "lstrip": false,
600
+ "normalized": false,
601
+ "rstrip": false,
602
+ "single_word": false,
603
+ "special": false
604
+ },
605
+ "75": {
606
+ "content": "<unused68>",
607
+ "lstrip": false,
608
+ "normalized": false,
609
+ "rstrip": false,
610
+ "single_word": false,
611
+ "special": false
612
+ },
613
+ "76": {
614
+ "content": "<unused69>",
615
+ "lstrip": false,
616
+ "normalized": false,
617
+ "rstrip": false,
618
+ "single_word": false,
619
+ "special": false
620
+ },
621
+ "77": {
622
+ "content": "<unused70>",
623
+ "lstrip": false,
624
+ "normalized": false,
625
+ "rstrip": false,
626
+ "single_word": false,
627
+ "special": false
628
+ },
629
+ "78": {
630
+ "content": "<unused71>",
631
+ "lstrip": false,
632
+ "normalized": false,
633
+ "rstrip": false,
634
+ "single_word": false,
635
+ "special": false
636
+ },
637
+ "79": {
638
+ "content": "<unused72>",
639
+ "lstrip": false,
640
+ "normalized": false,
641
+ "rstrip": false,
642
+ "single_word": false,
643
+ "special": false
644
+ },
645
+ "80": {
646
+ "content": "<unused73>",
647
+ "lstrip": false,
648
+ "normalized": false,
649
+ "rstrip": false,
650
+ "single_word": false,
651
+ "special": false
652
+ },
653
+ "81": {
654
+ "content": "<unused74>",
655
+ "lstrip": false,
656
+ "normalized": false,
657
+ "rstrip": false,
658
+ "single_word": false,
659
+ "special": false
660
+ },
661
+ "82": {
662
+ "content": "<unused75>",
663
+ "lstrip": false,
664
+ "normalized": false,
665
+ "rstrip": false,
666
+ "single_word": false,
667
+ "special": false
668
+ },
669
+ "83": {
670
+ "content": "<unused76>",
671
+ "lstrip": false,
672
+ "normalized": false,
673
+ "rstrip": false,
674
+ "single_word": false,
675
+ "special": false
676
+ },
677
+ "84": {
678
+ "content": "<unused77>",
679
+ "lstrip": false,
680
+ "normalized": false,
681
+ "rstrip": false,
682
+ "single_word": false,
683
+ "special": false
684
+ },
685
+ "85": {
686
+ "content": "<unused78>",
687
+ "lstrip": false,
688
+ "normalized": false,
689
+ "rstrip": false,
690
+ "single_word": false,
691
+ "special": false
692
+ },
693
+ "86": {
694
+ "content": "<unused79>",
695
+ "lstrip": false,
696
+ "normalized": false,
697
+ "rstrip": false,
698
+ "single_word": false,
699
+ "special": false
700
+ },
701
+ "87": {
702
+ "content": "<unused80>",
703
+ "lstrip": false,
704
+ "normalized": false,
705
+ "rstrip": false,
706
+ "single_word": false,
707
+ "special": false
708
+ },
709
+ "88": {
710
+ "content": "<unused81>",
711
+ "lstrip": false,
712
+ "normalized": false,
713
+ "rstrip": false,
714
+ "single_word": false,
715
+ "special": false
716
+ },
717
+ "89": {
718
+ "content": "<unused82>",
719
+ "lstrip": false,
720
+ "normalized": false,
721
+ "rstrip": false,
722
+ "single_word": false,
723
+ "special": false
724
+ },
725
+ "90": {
726
+ "content": "<unused83>",
727
+ "lstrip": false,
728
+ "normalized": false,
729
+ "rstrip": false,
730
+ "single_word": false,
731
+ "special": false
732
+ },
733
+ "91": {
734
+ "content": "<unused84>",
735
+ "lstrip": false,
736
+ "normalized": false,
737
+ "rstrip": false,
738
+ "single_word": false,
739
+ "special": false
740
+ },
741
+ "92": {
742
+ "content": "<unused85>",
743
+ "lstrip": false,
744
+ "normalized": false,
745
+ "rstrip": false,
746
+ "single_word": false,
747
+ "special": false
748
+ },
749
+ "93": {
750
+ "content": "<unused86>",
751
+ "lstrip": false,
752
+ "normalized": false,
753
+ "rstrip": false,
754
+ "single_word": false,
755
+ "special": false
756
+ },
757
+ "94": {
758
+ "content": "<unused87>",
759
+ "lstrip": false,
760
+ "normalized": false,
761
+ "rstrip": false,
762
+ "single_word": false,
763
+ "special": false
764
+ },
765
+ "95": {
766
+ "content": "<unused88>",
767
+ "lstrip": false,
768
+ "normalized": false,
769
+ "rstrip": false,
770
+ "single_word": false,
771
+ "special": false
772
+ },
773
+ "96": {
774
+ "content": "<unused89>",
775
+ "lstrip": false,
776
+ "normalized": false,
777
+ "rstrip": false,
778
+ "single_word": false,
779
+ "special": false
780
+ },
781
+ "97": {
782
+ "content": "<unused90>",
783
+ "lstrip": false,
784
+ "normalized": false,
785
+ "rstrip": false,
786
+ "single_word": false,
787
+ "special": false
788
+ },
789
+ "98": {
790
+ "content": "<unused91>",
791
+ "lstrip": false,
792
+ "normalized": false,
793
+ "rstrip": false,
794
+ "single_word": false,
795
+ "special": false
796
+ },
797
+ "99": {
798
+ "content": "<unused92>",
799
+ "lstrip": false,
800
+ "normalized": false,
801
+ "rstrip": false,
802
+ "single_word": false,
803
+ "special": false
804
+ },
805
+ "100": {
806
+ "content": "<unused93>",
807
+ "lstrip": false,
808
+ "normalized": false,
809
+ "rstrip": false,
810
+ "single_word": false,
811
+ "special": false
812
+ },
813
+ "101": {
814
+ "content": "<unused94>",
815
+ "lstrip": false,
816
+ "normalized": false,
817
+ "rstrip": false,
818
+ "single_word": false,
819
+ "special": false
820
+ },
821
+ "102": {
822
+ "content": "<unused95>",
823
+ "lstrip": false,
824
+ "normalized": false,
825
+ "rstrip": false,
826
+ "single_word": false,
827
+ "special": false
828
+ },
829
+ "103": {
830
+ "content": "<unused96>",
831
+ "lstrip": false,
832
+ "normalized": false,
833
+ "rstrip": false,
834
+ "single_word": false,
835
+ "special": false
836
+ },
837
+ "104": {
838
+ "content": "<unused97>",
839
+ "lstrip": false,
840
+ "normalized": false,
841
+ "rstrip": false,
842
+ "single_word": false,
843
+ "special": false
844
+ },
845
+ "105": {
846
+ "content": "<unused98>",
847
+ "lstrip": false,
848
+ "normalized": false,
849
+ "rstrip": false,
850
+ "single_word": false,
851
+ "special": false
852
+ },
853
+ "106": {
854
+ "content": "<start_of_turn>",
855
+ "lstrip": false,
856
+ "normalized": false,
857
+ "rstrip": false,
858
+ "single_word": false,
859
+ "special": true
860
+ },
861
+ "107": {
862
+ "content": "<end_of_turn>",
863
+ "lstrip": false,
864
+ "normalized": false,
865
+ "rstrip": false,
866
+ "single_word": false,
867
+ "special": true
868
+ },
869
+ "108": {
870
+ "content": "\n",
871
+ "lstrip": false,
872
+ "normalized": false,
873
+ "rstrip": false,
874
+ "single_word": false,
875
+ "special": false
876
+ },
877
+ "109": {
878
+ "content": "\n\n",
879
+ "lstrip": false,
880
+ "normalized": false,
881
+ "rstrip": false,
882
+ "single_word": false,
883
+ "special": false
884
+ },
885
+ "110": {
886
+ "content": "\n\n\n",
887
+ "lstrip": false,
888
+ "normalized": false,
889
+ "rstrip": false,
890
+ "single_word": false,
891
+ "special": false
892
+ },
893
+ "111": {
894
+ "content": "\n\n\n\n",
895
+ "lstrip": false,
896
+ "normalized": false,
897
+ "rstrip": false,
898
+ "single_word": false,
899
+ "special": false
900
+ },
901
+ "112": {
902
+ "content": "\n\n\n\n\n",
903
+ "lstrip": false,
904
+ "normalized": false,
905
+ "rstrip": false,
906
+ "single_word": false,
907
+ "special": false
908
+ },
909
+ "113": {
910
+ "content": "\n\n\n\n\n\n",
911
+ "lstrip": false,
912
+ "normalized": false,
913
+ "rstrip": false,
914
+ "single_word": false,
915
+ "special": false
916
+ },
917
+ "114": {
918
+ "content": "\n\n\n\n\n\n\n",
919
+ "lstrip": false,
920
+ "normalized": false,
921
+ "rstrip": false,
922
+ "single_word": false,
923
+ "special": false
924
+ },
925
+ "115": {
926
+ "content": "\n\n\n\n\n\n\n\n",
927
+ "lstrip": false,
928
+ "normalized": false,
929
+ "rstrip": false,
930
+ "single_word": false,
931
+ "special": false
932
+ },
933
+ "116": {
934
+ "content": "\n\n\n\n\n\n\n\n\n",
935
+ "lstrip": false,
936
+ "normalized": false,
937
+ "rstrip": false,
938
+ "single_word": false,
939
+ "special": false
940
+ },
941
+ "117": {
942
+ "content": "\n\n\n\n\n\n\n\n\n\n",
943
+ "lstrip": false,
944
+ "normalized": false,
945
+ "rstrip": false,
946
+ "single_word": false,
947
+ "special": false
948
+ },
949
+ "118": {
950
+ "content": "\n\n\n\n\n\n\n\n\n\n\n",
951
+ "lstrip": false,
952
+ "normalized": false,
953
+ "rstrip": false,
954
+ "single_word": false,
955
+ "special": false
956
+ },
957
+ "119": {
958
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n",
959
+ "lstrip": false,
960
+ "normalized": false,
961
+ "rstrip": false,
962
+ "single_word": false,
963
+ "special": false
964
+ },
965
+ "120": {
966
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n",
967
+ "lstrip": false,
968
+ "normalized": false,
969
+ "rstrip": false,
970
+ "single_word": false,
971
+ "special": false
972
+ },
973
+ "121": {
974
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
975
+ "lstrip": false,
976
+ "normalized": false,
977
+ "rstrip": false,
978
+ "single_word": false,
979
+ "special": false
980
+ },
981
+ "122": {
982
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
983
+ "lstrip": false,
984
+ "normalized": false,
985
+ "rstrip": false,
986
+ "single_word": false,
987
+ "special": false
988
+ },
989
+ "123": {
990
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
991
+ "lstrip": false,
992
+ "normalized": false,
993
+ "rstrip": false,
994
+ "single_word": false,
995
+ "special": false
996
+ },
997
+ "124": {
998
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
999
+ "lstrip": false,
1000
+ "normalized": false,
1001
+ "rstrip": false,
1002
+ "single_word": false,
1003
+ "special": false
1004
+ },
1005
+ "125": {
1006
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1007
+ "lstrip": false,
1008
+ "normalized": false,
1009
+ "rstrip": false,
1010
+ "single_word": false,
1011
+ "special": false
1012
+ },
1013
+ "126": {
1014
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1015
+ "lstrip": false,
1016
+ "normalized": false,
1017
+ "rstrip": false,
1018
+ "single_word": false,
1019
+ "special": false
1020
+ },
1021
+ "127": {
1022
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1023
+ "lstrip": false,
1024
+ "normalized": false,
1025
+ "rstrip": false,
1026
+ "single_word": false,
1027
+ "special": false
1028
+ },
1029
+ "128": {
1030
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1031
+ "lstrip": false,
1032
+ "normalized": false,
1033
+ "rstrip": false,
1034
+ "single_word": false,
1035
+ "special": false
1036
+ },
1037
+ "129": {
1038
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1039
+ "lstrip": false,
1040
+ "normalized": false,
1041
+ "rstrip": false,
1042
+ "single_word": false,
1043
+ "special": false
1044
+ },
1045
+ "130": {
1046
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1047
+ "lstrip": false,
1048
+ "normalized": false,
1049
+ "rstrip": false,
1050
+ "single_word": false,
1051
+ "special": false
1052
+ },
1053
+ "131": {
1054
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1055
+ "lstrip": false,
1056
+ "normalized": false,
1057
+ "rstrip": false,
1058
+ "single_word": false,
1059
+ "special": false
1060
+ },
1061
+ "132": {
1062
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1063
+ "lstrip": false,
1064
+ "normalized": false,
1065
+ "rstrip": false,
1066
+ "single_word": false,
1067
+ "special": false
1068
+ },
1069
+ "133": {
1070
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1071
+ "lstrip": false,
1072
+ "normalized": false,
1073
+ "rstrip": false,
1074
+ "single_word": false,
1075
+ "special": false
1076
+ },
1077
+ "134": {
1078
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1079
+ "lstrip": false,
1080
+ "normalized": false,
1081
+ "rstrip": false,
1082
+ "single_word": false,
1083
+ "special": false
1084
+ },
1085
+ "135": {
1086
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1087
+ "lstrip": false,
1088
+ "normalized": false,
1089
+ "rstrip": false,
1090
+ "single_word": false,
1091
+ "special": false
1092
+ },
1093
+ "136": {
1094
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1095
+ "lstrip": false,
1096
+ "normalized": false,
1097
+ "rstrip": false,
1098
+ "single_word": false,
1099
+ "special": false
1100
+ },
1101
+ "137": {
1102
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1103
+ "lstrip": false,
1104
+ "normalized": false,
1105
+ "rstrip": false,
1106
+ "single_word": false,
1107
+ "special": false
1108
+ },
1109
+ "138": {
1110
+ "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1111
+ "lstrip": false,
1112
+ "normalized": false,
1113
+ "rstrip": false,
1114
+ "single_word": false,
1115
+ "special": false
1116
+ },
1117
+ "139": {
1118
+ "content": "▁▁",
1119
+ "lstrip": false,
1120
+ "normalized": false,
1121
+ "rstrip": false,
1122
+ "single_word": false,
1123
+ "special": false
1124
+ },
1125
+ "140": {
1126
+ "content": "▁▁▁",
1127
+ "lstrip": false,
1128
+ "normalized": false,
1129
+ "rstrip": false,
1130
+ "single_word": false,
1131
+ "special": false
1132
+ },
1133
+ "141": {
1134
+ "content": "▁▁▁▁",
1135
+ "lstrip": false,
1136
+ "normalized": false,
1137
+ "rstrip": false,
1138
+ "single_word": false,
1139
+ "special": false
1140
+ },
1141
+ "142": {
1142
+ "content": "▁▁▁▁▁",
1143
+ "lstrip": false,
1144
+ "normalized": false,
1145
+ "rstrip": false,
1146
+ "single_word": false,
1147
+ "special": false
1148
+ },
1149
+ "143": {
1150
+ "content": "▁▁▁▁▁▁",
1151
+ "lstrip": false,
1152
+ "normalized": false,
1153
+ "rstrip": false,
1154
+ "single_word": false,
1155
+ "special": false
1156
+ },
1157
+ "144": {
1158
+ "content": "▁▁▁▁▁▁▁",
1159
+ "lstrip": false,
1160
+ "normalized": false,
1161
+ "rstrip": false,
1162
+ "single_word": false,
1163
+ "special": false
1164
+ },
1165
+ "145": {
1166
+ "content": "▁▁▁▁▁▁▁▁",
1167
+ "lstrip": false,
1168
+ "normalized": false,
1169
+ "rstrip": false,
1170
+ "single_word": false,
1171
+ "special": false
1172
+ },
1173
+ "146": {
1174
+ "content": "▁▁▁▁▁▁▁▁▁",
1175
+ "lstrip": false,
1176
+ "normalized": false,
1177
+ "rstrip": false,
1178
+ "single_word": false,
1179
+ "special": false
1180
+ },
1181
+ "147": {
1182
+ "content": "▁▁▁▁▁▁▁▁▁▁",
1183
+ "lstrip": false,
1184
+ "normalized": false,
1185
+ "rstrip": false,
1186
+ "single_word": false,
1187
+ "special": false
1188
+ },
1189
+ "148": {
1190
+ "content": "▁▁▁▁▁▁▁▁▁▁▁",
1191
+ "lstrip": false,
1192
+ "normalized": false,
1193
+ "rstrip": false,
1194
+ "single_word": false,
1195
+ "special": false
1196
+ },
1197
+ "149": {
1198
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁",
1199
+ "lstrip": false,
1200
+ "normalized": false,
1201
+ "rstrip": false,
1202
+ "single_word": false,
1203
+ "special": false
1204
+ },
1205
+ "150": {
1206
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁",
1207
+ "lstrip": false,
1208
+ "normalized": false,
1209
+ "rstrip": false,
1210
+ "single_word": false,
1211
+ "special": false
1212
+ },
1213
+ "151": {
1214
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1215
+ "lstrip": false,
1216
+ "normalized": false,
1217
+ "rstrip": false,
1218
+ "single_word": false,
1219
+ "special": false
1220
+ },
1221
+ "152": {
1222
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1223
+ "lstrip": false,
1224
+ "normalized": false,
1225
+ "rstrip": false,
1226
+ "single_word": false,
1227
+ "special": false
1228
+ },
1229
+ "153": {
1230
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1231
+ "lstrip": false,
1232
+ "normalized": false,
1233
+ "rstrip": false,
1234
+ "single_word": false,
1235
+ "special": false
1236
+ },
1237
+ "154": {
1238
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1239
+ "lstrip": false,
1240
+ "normalized": false,
1241
+ "rstrip": false,
1242
+ "single_word": false,
1243
+ "special": false
1244
+ },
1245
+ "155": {
1246
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1247
+ "lstrip": false,
1248
+ "normalized": false,
1249
+ "rstrip": false,
1250
+ "single_word": false,
1251
+ "special": false
1252
+ },
1253
+ "156": {
1254
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1255
+ "lstrip": false,
1256
+ "normalized": false,
1257
+ "rstrip": false,
1258
+ "single_word": false,
1259
+ "special": false
1260
+ },
1261
+ "157": {
1262
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1263
+ "lstrip": false,
1264
+ "normalized": false,
1265
+ "rstrip": false,
1266
+ "single_word": false,
1267
+ "special": false
1268
+ },
1269
+ "158": {
1270
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1271
+ "lstrip": false,
1272
+ "normalized": false,
1273
+ "rstrip": false,
1274
+ "single_word": false,
1275
+ "special": false
1276
+ },
1277
+ "159": {
1278
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1279
+ "lstrip": false,
1280
+ "normalized": false,
1281
+ "rstrip": false,
1282
+ "single_word": false,
1283
+ "special": false
1284
+ },
1285
+ "160": {
1286
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1287
+ "lstrip": false,
1288
+ "normalized": false,
1289
+ "rstrip": false,
1290
+ "single_word": false,
1291
+ "special": false
1292
+ },
1293
+ "161": {
1294
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1295
+ "lstrip": false,
1296
+ "normalized": false,
1297
+ "rstrip": false,
1298
+ "single_word": false,
1299
+ "special": false
1300
+ },
1301
+ "162": {
1302
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1303
+ "lstrip": false,
1304
+ "normalized": false,
1305
+ "rstrip": false,
1306
+ "single_word": false,
1307
+ "special": false
1308
+ },
1309
+ "163": {
1310
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1311
+ "lstrip": false,
1312
+ "normalized": false,
1313
+ "rstrip": false,
1314
+ "single_word": false,
1315
+ "special": false
1316
+ },
1317
+ "164": {
1318
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1319
+ "lstrip": false,
1320
+ "normalized": false,
1321
+ "rstrip": false,
1322
+ "single_word": false,
1323
+ "special": false
1324
+ },
1325
+ "165": {
1326
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1327
+ "lstrip": false,
1328
+ "normalized": false,
1329
+ "rstrip": false,
1330
+ "single_word": false,
1331
+ "special": false
1332
+ },
1333
+ "166": {
1334
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1335
+ "lstrip": false,
1336
+ "normalized": false,
1337
+ "rstrip": false,
1338
+ "single_word": false,
1339
+ "special": false
1340
+ },
1341
+ "167": {
1342
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1343
+ "lstrip": false,
1344
+ "normalized": false,
1345
+ "rstrip": false,
1346
+ "single_word": false,
1347
+ "special": false
1348
+ },
1349
+ "168": {
1350
+ "content": "▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁▁",
1351
+ "lstrip": false,
1352
+ "normalized": false,
1353
+ "rstrip": false,
1354
+ "single_word": false,
1355
+ "special": false
1356
+ },
1357
+ "169": {
1358
+ "content": "<table>",
1359
+ "lstrip": false,
1360
+ "normalized": false,
1361
+ "rstrip": false,
1362
+ "single_word": false,
1363
+ "special": false
1364
+ },
1365
+ "170": {
1366
+ "content": "<caption>",
1367
+ "lstrip": false,
1368
+ "normalized": false,
1369
+ "rstrip": false,
1370
+ "single_word": false,
1371
+ "special": false
1372
+ },
1373
+ "171": {
1374
+ "content": "<thead>",
1375
+ "lstrip": false,
1376
+ "normalized": false,
1377
+ "rstrip": false,
1378
+ "single_word": false,
1379
+ "special": false
1380
+ },
1381
+ "172": {
1382
+ "content": "<tbody>",
1383
+ "lstrip": false,
1384
+ "normalized": false,
1385
+ "rstrip": false,
1386
+ "single_word": false,
1387
+ "special": false
1388
+ },
1389
+ "173": {
1390
+ "content": "<tfoot>",
1391
+ "lstrip": false,
1392
+ "normalized": false,
1393
+ "rstrip": false,
1394
+ "single_word": false,
1395
+ "special": false
1396
+ },
1397
+ "174": {
1398
+ "content": "<tr>",
1399
+ "lstrip": false,
1400
+ "normalized": false,
1401
+ "rstrip": false,
1402
+ "single_word": false,
1403
+ "special": false
1404
+ },
1405
+ "175": {
1406
+ "content": "<th>",
1407
+ "lstrip": false,
1408
+ "normalized": false,
1409
+ "rstrip": false,
1410
+ "single_word": false,
1411
+ "special": false
1412
+ },
1413
+ "176": {
1414
+ "content": "<td>",
1415
+ "lstrip": false,
1416
+ "normalized": false,
1417
+ "rstrip": false,
1418
+ "single_word": false,
1419
+ "special": false
1420
+ },
1421
+ "177": {
1422
+ "content": "</table>",
1423
+ "lstrip": false,
1424
+ "normalized": false,
1425
+ "rstrip": false,
1426
+ "single_word": false,
1427
+ "special": false
1428
+ },
1429
+ "178": {
1430
+ "content": "</caption>",
1431
+ "lstrip": false,
1432
+ "normalized": false,
1433
+ "rstrip": false,
1434
+ "single_word": false,
1435
+ "special": false
1436
+ },
1437
+ "179": {
1438
+ "content": "</thead>",
1439
+ "lstrip": false,
1440
+ "normalized": false,
1441
+ "rstrip": false,
1442
+ "single_word": false,
1443
+ "special": false
1444
+ },
1445
+ "180": {
1446
+ "content": "</tbody>",
1447
+ "lstrip": false,
1448
+ "normalized": false,
1449
+ "rstrip": false,
1450
+ "single_word": false,
1451
+ "special": false
1452
+ },
1453
+ "181": {
1454
+ "content": "</tfoot>",
1455
+ "lstrip": false,
1456
+ "normalized": false,
1457
+ "rstrip": false,
1458
+ "single_word": false,
1459
+ "special": false
1460
+ },
1461
+ "182": {
1462
+ "content": "</tr>",
1463
+ "lstrip": false,
1464
+ "normalized": false,
1465
+ "rstrip": false,
1466
+ "single_word": false,
1467
+ "special": false
1468
+ },
1469
+ "183": {
1470
+ "content": "</th>",
1471
+ "lstrip": false,
1472
+ "normalized": false,
1473
+ "rstrip": false,
1474
+ "single_word": false,
1475
+ "special": false
1476
+ },
1477
+ "184": {
1478
+ "content": "</td>",
1479
+ "lstrip": false,
1480
+ "normalized": false,
1481
+ "rstrip": false,
1482
+ "single_word": false,
1483
+ "special": false
1484
+ },
1485
+ "185": {
1486
+ "content": "<h1>",
1487
+ "lstrip": false,
1488
+ "normalized": false,
1489
+ "rstrip": false,
1490
+ "single_word": false,
1491
+ "special": false
1492
+ },
1493
+ "186": {
1494
+ "content": "<h2>",
1495
+ "lstrip": false,
1496
+ "normalized": false,
1497
+ "rstrip": false,
1498
+ "single_word": false,
1499
+ "special": false
1500
+ },
1501
+ "187": {
1502
+ "content": "<h3>",
1503
+ "lstrip": false,
1504
+ "normalized": false,
1505
+ "rstrip": false,
1506
+ "single_word": false,
1507
+ "special": false
1508
+ },
1509
+ "188": {
1510
+ "content": "<h4>",
1511
+ "lstrip": false,
1512
+ "normalized": false,
1513
+ "rstrip": false,
1514
+ "single_word": false,
1515
+ "special": false
1516
+ },
1517
+ "189": {
1518
+ "content": "<h5>",
1519
+ "lstrip": false,
1520
+ "normalized": false,
1521
+ "rstrip": false,
1522
+ "single_word": false,
1523
+ "special": false
1524
+ },
1525
+ "190": {
1526
+ "content": "<h6>",
1527
+ "lstrip": false,
1528
+ "normalized": false,
1529
+ "rstrip": false,
1530
+ "single_word": false,
1531
+ "special": false
1532
+ },
1533
+ "191": {
1534
+ "content": "<blockquote>",
1535
+ "lstrip": false,
1536
+ "normalized": false,
1537
+ "rstrip": false,
1538
+ "single_word": false,
1539
+ "special": false
1540
+ },
1541
+ "192": {
1542
+ "content": "</h1>",
1543
+ "lstrip": false,
1544
+ "normalized": false,
1545
+ "rstrip": false,
1546
+ "single_word": false,
1547
+ "special": false
1548
+ },
1549
+ "193": {
1550
+ "content": "</h2>",
1551
+ "lstrip": false,
1552
+ "normalized": false,
1553
+ "rstrip": false,
1554
+ "single_word": false,
1555
+ "special": false
1556
+ },
1557
+ "194": {
1558
+ "content": "</h3>",
1559
+ "lstrip": false,
1560
+ "normalized": false,
1561
+ "rstrip": false,
1562
+ "single_word": false,
1563
+ "special": false
1564
+ },
1565
+ "195": {
1566
+ "content": "</h4>",
1567
+ "lstrip": false,
1568
+ "normalized": false,
1569
+ "rstrip": false,
1570
+ "single_word": false,
1571
+ "special": false
1572
+ },
1573
+ "196": {
1574
+ "content": "</h5>",
1575
+ "lstrip": false,
1576
+ "normalized": false,
1577
+ "rstrip": false,
1578
+ "single_word": false,
1579
+ "special": false
1580
+ },
1581
+ "197": {
1582
+ "content": "</h6>",
1583
+ "lstrip": false,
1584
+ "normalized": false,
1585
+ "rstrip": false,
1586
+ "single_word": false,
1587
+ "special": false
1588
+ },
1589
+ "198": {
1590
+ "content": "</blockquote>",
1591
+ "lstrip": false,
1592
+ "normalized": false,
1593
+ "rstrip": false,
1594
+ "single_word": false,
1595
+ "special": false
1596
+ },
1597
+ "199": {
1598
+ "content": "<strong>",
1599
+ "lstrip": false,
1600
+ "normalized": false,
1601
+ "rstrip": false,
1602
+ "single_word": false,
1603
+ "special": false
1604
+ },
1605
+ "200": {
1606
+ "content": "<em>",
1607
+ "lstrip": false,
1608
+ "normalized": false,
1609
+ "rstrip": false,
1610
+ "single_word": false,
1611
+ "special": false
1612
+ },
1613
+ "201": {
1614
+ "content": "<b>",
1615
+ "lstrip": false,
1616
+ "normalized": false,
1617
+ "rstrip": false,
1618
+ "single_word": false,
1619
+ "special": false
1620
+ },
1621
+ "202": {
1622
+ "content": "<i>",
1623
+ "lstrip": false,
1624
+ "normalized": false,
1625
+ "rstrip": false,
1626
+ "single_word": false,
1627
+ "special": false
1628
+ },
1629
+ "203": {
1630
+ "content": "<u>",
1631
+ "lstrip": false,
1632
+ "normalized": false,
1633
+ "rstrip": false,
1634
+ "single_word": false,
1635
+ "special": false
1636
+ },
1637
+ "204": {
1638
+ "content": "<s>",
1639
+ "lstrip": false,
1640
+ "normalized": false,
1641
+ "rstrip": false,
1642
+ "single_word": false,
1643
+ "special": false
1644
+ },
1645
+ "205": {
1646
+ "content": "<sub>",
1647
+ "lstrip": false,
1648
+ "normalized": false,
1649
+ "rstrip": false,
1650
+ "single_word": false,
1651
+ "special": false
1652
+ },
1653
+ "206": {
1654
+ "content": "<sup>",
1655
+ "lstrip": false,
1656
+ "normalized": false,
1657
+ "rstrip": false,
1658
+ "single_word": false,
1659
+ "special": false
1660
+ },
1661
+ "207": {
1662
+ "content": "<code>",
1663
+ "lstrip": false,
1664
+ "normalized": false,
1665
+ "rstrip": false,
1666
+ "single_word": false,
1667
+ "special": false
1668
+ },
1669
+ "208": {
1670
+ "content": "</strong>",
1671
+ "lstrip": false,
1672
+ "normalized": false,
1673
+ "rstrip": false,
1674
+ "single_word": false,
1675
+ "special": false
1676
+ },
1677
+ "209": {
1678
+ "content": "</em>",
1679
+ "lstrip": false,
1680
+ "normalized": false,
1681
+ "rstrip": false,
1682
+ "single_word": false,
1683
+ "special": false
1684
+ },
1685
+ "210": {
1686
+ "content": "</b>",
1687
+ "lstrip": false,
1688
+ "normalized": false,
1689
+ "rstrip": false,
1690
+ "single_word": false,
1691
+ "special": false
1692
+ },
1693
+ "211": {
1694
+ "content": "</i>",
1695
+ "lstrip": false,
1696
+ "normalized": false,
1697
+ "rstrip": false,
1698
+ "single_word": false,
1699
+ "special": false
1700
+ },
1701
+ "212": {
1702
+ "content": "</u>",
1703
+ "lstrip": false,
1704
+ "normalized": false,
1705
+ "rstrip": false,
1706
+ "single_word": false,
1707
+ "special": false
1708
+ },
1709
+ "213": {
1710
+ "content": "</s>",
1711
+ "lstrip": false,
1712
+ "normalized": false,
1713
+ "rstrip": false,
1714
+ "single_word": false,
1715
+ "special": false
1716
+ },
1717
+ "214": {
1718
+ "content": "</sub>",
1719
+ "lstrip": false,
1720
+ "normalized": false,
1721
+ "rstrip": false,
1722
+ "single_word": false,
1723
+ "special": false
1724
+ },
1725
+ "215": {
1726
+ "content": "</sup>",
1727
+ "lstrip": false,
1728
+ "normalized": false,
1729
+ "rstrip": false,
1730
+ "single_word": false,
1731
+ "special": false
1732
+ },
1733
+ "216": {
1734
+ "content": "</code>",
1735
+ "lstrip": false,
1736
+ "normalized": false,
1737
+ "rstrip": false,
1738
+ "single_word": false,
1739
+ "special": false
1740
+ },
1741
+ "255968": {
1742
+ "content": "[toxicity=0]",
1743
+ "lstrip": false,
1744
+ "normalized": false,
1745
+ "rstrip": false,
1746
+ "single_word": false,
1747
+ "special": false
1748
+ },
1749
+ "255969": {
1750
+ "content": "\t\t",
1751
+ "lstrip": false,
1752
+ "normalized": false,
1753
+ "rstrip": false,
1754
+ "single_word": false,
1755
+ "special": false
1756
+ },
1757
+ "255970": {
1758
+ "content": "\t\t\t",
1759
+ "lstrip": false,
1760
+ "normalized": false,
1761
+ "rstrip": false,
1762
+ "single_word": false,
1763
+ "special": false
1764
+ },
1765
+ "255971": {
1766
+ "content": "\t\t\t\t",
1767
+ "lstrip": false,
1768
+ "normalized": false,
1769
+ "rstrip": false,
1770
+ "single_word": false,
1771
+ "special": false
1772
+ },
1773
+ "255972": {
1774
+ "content": "\t\t\t\t\t",
1775
+ "lstrip": false,
1776
+ "normalized": false,
1777
+ "rstrip": false,
1778
+ "single_word": false,
1779
+ "special": false
1780
+ },
1781
+ "255973": {
1782
+ "content": "\t\t\t\t\t\t",
1783
+ "lstrip": false,
1784
+ "normalized": false,
1785
+ "rstrip": false,
1786
+ "single_word": false,
1787
+ "special": false
1788
+ },
1789
+ "255974": {
1790
+ "content": "\t\t\t\t\t\t\t",
1791
+ "lstrip": false,
1792
+ "normalized": false,
1793
+ "rstrip": false,
1794
+ "single_word": false,
1795
+ "special": false
1796
+ },
1797
+ "255975": {
1798
+ "content": "\t\t\t\t\t\t\t\t",
1799
+ "lstrip": false,
1800
+ "normalized": false,
1801
+ "rstrip": false,
1802
+ "single_word": false,
1803
+ "special": false
1804
+ },
1805
+ "255976": {
1806
+ "content": "\t\t\t\t\t\t\t\t\t",
1807
+ "lstrip": false,
1808
+ "normalized": false,
1809
+ "rstrip": false,
1810
+ "single_word": false,
1811
+ "special": false
1812
+ },
1813
+ "255977": {
1814
+ "content": "\t\t\t\t\t\t\t\t\t\t",
1815
+ "lstrip": false,
1816
+ "normalized": false,
1817
+ "rstrip": false,
1818
+ "single_word": false,
1819
+ "special": false
1820
+ },
1821
+ "255978": {
1822
+ "content": "\t\t\t\t\t\t\t\t\t\t\t",
1823
+ "lstrip": false,
1824
+ "normalized": false,
1825
+ "rstrip": false,
1826
+ "single_word": false,
1827
+ "special": false
1828
+ },
1829
+ "255979": {
1830
+ "content": "\t\t\t\t\t\t\t\t\t\t\t\t",
1831
+ "lstrip": false,
1832
+ "normalized": false,
1833
+ "rstrip": false,
1834
+ "single_word": false,
1835
+ "special": false
1836
+ },
1837
+ "255980": {
1838
+ "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t",
1839
+ "lstrip": false,
1840
+ "normalized": false,
1841
+ "rstrip": false,
1842
+ "single_word": false,
1843
+ "special": false
1844
+ },
1845
+ "255981": {
1846
+ "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1847
+ "lstrip": false,
1848
+ "normalized": false,
1849
+ "rstrip": false,
1850
+ "single_word": false,
1851
+ "special": false
1852
+ },
1853
+ "255982": {
1854
+ "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1855
+ "lstrip": false,
1856
+ "normalized": false,
1857
+ "rstrip": false,
1858
+ "single_word": false,
1859
+ "special": false
1860
+ },
1861
+ "255983": {
1862
+ "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1863
+ "lstrip": false,
1864
+ "normalized": false,
1865
+ "rstrip": false,
1866
+ "single_word": false,
1867
+ "special": false
1868
+ },
1869
+ "255984": {
1870
+ "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1871
+ "lstrip": false,
1872
+ "normalized": false,
1873
+ "rstrip": false,
1874
+ "single_word": false,
1875
+ "special": false
1876
+ },
1877
+ "255985": {
1878
+ "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1879
+ "lstrip": false,
1880
+ "normalized": false,
1881
+ "rstrip": false,
1882
+ "single_word": false,
1883
+ "special": false
1884
+ },
1885
+ "255986": {
1886
+ "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1887
+ "lstrip": false,
1888
+ "normalized": false,
1889
+ "rstrip": false,
1890
+ "single_word": false,
1891
+ "special": false
1892
+ },
1893
+ "255987": {
1894
+ "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1895
+ "lstrip": false,
1896
+ "normalized": false,
1897
+ "rstrip": false,
1898
+ "single_word": false,
1899
+ "special": false
1900
+ },
1901
+ "255988": {
1902
+ "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1903
+ "lstrip": false,
1904
+ "normalized": false,
1905
+ "rstrip": false,
1906
+ "single_word": false,
1907
+ "special": false
1908
+ },
1909
+ "255989": {
1910
+ "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1911
+ "lstrip": false,
1912
+ "normalized": false,
1913
+ "rstrip": false,
1914
+ "single_word": false,
1915
+ "special": false
1916
+ },
1917
+ "255990": {
1918
+ "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1919
+ "lstrip": false,
1920
+ "normalized": false,
1921
+ "rstrip": false,
1922
+ "single_word": false,
1923
+ "special": false
1924
+ },
1925
+ "255991": {
1926
+ "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1927
+ "lstrip": false,
1928
+ "normalized": false,
1929
+ "rstrip": false,
1930
+ "single_word": false,
1931
+ "special": false
1932
+ },
1933
+ "255992": {
1934
+ "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1935
+ "lstrip": false,
1936
+ "normalized": false,
1937
+ "rstrip": false,
1938
+ "single_word": false,
1939
+ "special": false
1940
+ },
1941
+ "255993": {
1942
+ "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1943
+ "lstrip": false,
1944
+ "normalized": false,
1945
+ "rstrip": false,
1946
+ "single_word": false,
1947
+ "special": false
1948
+ },
1949
+ "255994": {
1950
+ "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1951
+ "lstrip": false,
1952
+ "normalized": false,
1953
+ "rstrip": false,
1954
+ "single_word": false,
1955
+ "special": false
1956
+ },
1957
+ "255995": {
1958
+ "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1959
+ "lstrip": false,
1960
+ "normalized": false,
1961
+ "rstrip": false,
1962
+ "single_word": false,
1963
+ "special": false
1964
+ },
1965
+ "255996": {
1966
+ "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1967
+ "lstrip": false,
1968
+ "normalized": false,
1969
+ "rstrip": false,
1970
+ "single_word": false,
1971
+ "special": false
1972
+ },
1973
+ "255997": {
1974
+ "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1975
+ "lstrip": false,
1976
+ "normalized": false,
1977
+ "rstrip": false,
1978
+ "single_word": false,
1979
+ "special": false
1980
+ },
1981
+ "255998": {
1982
+ "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1983
+ "lstrip": false,
1984
+ "normalized": false,
1985
+ "rstrip": false,
1986
+ "single_word": false,
1987
+ "special": false
1988
+ },
1989
+ "255999": {
1990
+ "content": "<unused99>",
1991
+ "lstrip": false,
1992
+ "normalized": false,
1993
+ "rstrip": false,
1994
+ "single_word": false,
1995
+ "special": false
1996
+ }
1997
+ },
1998
+ "additional_special_tokens": [
1999
+ "<start_of_turn>",
2000
+ "<end_of_turn>"
2001
+ ],
2002
+ "bos_token": "<bos>",
2003
+ "chat_template": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if (message['role'] == 'assistant') %}{% set role = 'model' %}{% else %}{% set role = message['role'] %}{% endif %}{{ '<start_of_turn>' + role + '\n' + message['content'] | trim + '<end_of_turn>\n' }}{% endfor %}{% if add_generation_prompt %}{{'<start_of_turn>model\n'}}{% endif %}",
2004
+ "clean_up_tokenization_spaces": false,
2005
+ "eos_token": "<eos>",
2006
+ "max_length": 8192,
2007
+ "model_max_length": 1000000000000000019884624838656,
2008
+ "pad_token": "<pad>",
2009
+ "sp_model_kwargs": {},
2010
+ "spaces_between_special_tokens": false,
2011
+ "stride": 0,
2012
+ "tokenizer_class": "GemmaTokenizer",
2013
+ "truncation_side": "right",
2014
+ "truncation_strategy": "longest_first",
2015
+ "unk_token": "<unk>",
2016
+ "use_default_system_prompt": false
2017
+ }