zemaia commited on
Commit
000b60a
1 Parent(s): a920d39

initial commit

Browse files
config.json ADDED
@@ -0,0 +1,68 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/tmp/tmp164lo86z",
3
+ "architectures": [
4
+ "BertForTokenClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "LABEL_0",
14
+ "1": "LABEL_1",
15
+ "2": "LABEL_2",
16
+ "3": "LABEL_3",
17
+ "4": "LABEL_4",
18
+ "5": "LABEL_5",
19
+ "6": "LABEL_6",
20
+ "7": "LABEL_7",
21
+ "8": "LABEL_8",
22
+ "9": "LABEL_9",
23
+ "10": "LABEL_10",
24
+ "11": "LABEL_11",
25
+ "12": "LABEL_12",
26
+ "13": "LABEL_13",
27
+ "14": "LABEL_14",
28
+ "15": "LABEL_15",
29
+ "16": "LABEL_16",
30
+ "17": "LABEL_17",
31
+ "18": "LABEL_18"
32
+ },
33
+ "initializer_range": 0.02,
34
+ "intermediate_size": 3072,
35
+ "label2id": {
36
+ "LABEL_0": 0,
37
+ "LABEL_1": 1,
38
+ "LABEL_10": 10,
39
+ "LABEL_11": 11,
40
+ "LABEL_12": 12,
41
+ "LABEL_13": 13,
42
+ "LABEL_14": 14,
43
+ "LABEL_15": 15,
44
+ "LABEL_16": 16,
45
+ "LABEL_17": 17,
46
+ "LABEL_18": 18,
47
+ "LABEL_2": 2,
48
+ "LABEL_3": 3,
49
+ "LABEL_4": 4,
50
+ "LABEL_5": 5,
51
+ "LABEL_6": 6,
52
+ "LABEL_7": 7,
53
+ "LABEL_8": 8,
54
+ "LABEL_9": 9
55
+ },
56
+ "layer_norm_eps": 1e-12,
57
+ "max_position_embeddings": 512,
58
+ "model_type": "bert",
59
+ "num_attention_heads": 12,
60
+ "num_hidden_layers": 6,
61
+ "pad_token_id": 0,
62
+ "position_embedding_type": "absolute",
63
+ "torch_dtype": "float32",
64
+ "transformers_version": "4.38.2",
65
+ "type_vocab_size": 2,
66
+ "use_cache": true,
67
+ "vocab_size": 29794
68
+ }
model_quantized.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8850eceffe2666f6924be58d8138ae8c3f78ea9b1fdb96fe2f3f839529354301
3
+ size 87519705
ort_config.json ADDED
@@ -0,0 +1,596 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "one_external_file": true,
3
+ "opset": null,
4
+ "optimization": {},
5
+ "optimum_version": "1.18.0",
6
+ "quantization": {
7
+ "activations_dtype": "QUInt8",
8
+ "activations_symmetric": false,
9
+ "format": "QOperator",
10
+ "is_static": false,
11
+ "mode": "IntegerOps",
12
+ "nodes_to_exclude": [],
13
+ "nodes_to_quantize": [
14
+ "/bert/Constant",
15
+ "/bert/Constant_1",
16
+ "/bert/embeddings/Constant",
17
+ "Constant_157",
18
+ "/bert/embeddings/Constant_1",
19
+ "/bert/embeddings/Constant_2",
20
+ "/bert/embeddings/Constant_3",
21
+ "/bert/embeddings/LayerNorm/Constant",
22
+ "/bert/embeddings/LayerNorm/Constant_1",
23
+ "/bert/encoder/layer.0/attention/self/Constant",
24
+ "/bert/encoder/layer.0/attention/self/Constant_1",
25
+ "/bert/encoder/layer.0/attention/self/Constant_2",
26
+ "/bert/encoder/layer.0/attention/self/Constant_3",
27
+ "/bert/encoder/layer.0/attention/self/Constant_4",
28
+ "/bert/encoder/layer.0/attention/self/Constant_5",
29
+ "/bert/encoder/layer.0/attention/self/Constant_6",
30
+ "/bert/encoder/layer.0/attention/self/Constant_7",
31
+ "/bert/encoder/layer.0/attention/self/Constant_8",
32
+ "/bert/encoder/layer.0/attention/self/Constant_9",
33
+ "/bert/encoder/layer.0/attention/self/Constant_10",
34
+ "/bert/encoder/layer.0/attention/self/Constant_11",
35
+ "/bert/encoder/layer.0/attention/self/Constant_12",
36
+ "/bert/encoder/layer.0/attention/self/Constant_13",
37
+ "/bert/encoder/layer.0/attention/self/Constant_14",
38
+ "/bert/encoder/layer.0/attention/self/Constant_15",
39
+ "/bert/encoder/layer.0/attention/output/LayerNorm/Constant",
40
+ "/bert/encoder/layer.0/attention/output/LayerNorm/Constant_1",
41
+ "/bert/encoder/layer.0/intermediate/intermediate_act_fn/Constant",
42
+ "/bert/encoder/layer.0/intermediate/intermediate_act_fn/Constant_1",
43
+ "/bert/encoder/layer.0/intermediate/intermediate_act_fn/Constant_2",
44
+ "/bert/encoder/layer.0/output/LayerNorm/Constant",
45
+ "/bert/encoder/layer.0/output/LayerNorm/Constant_1",
46
+ "/bert/encoder/layer.1/attention/self/Constant",
47
+ "/bert/encoder/layer.1/attention/self/Constant_1",
48
+ "/bert/encoder/layer.1/attention/self/Constant_2",
49
+ "/bert/encoder/layer.1/attention/self/Constant_3",
50
+ "/bert/encoder/layer.1/attention/self/Constant_4",
51
+ "/bert/encoder/layer.1/attention/self/Constant_5",
52
+ "/bert/encoder/layer.1/attention/self/Constant_6",
53
+ "/bert/encoder/layer.1/attention/self/Constant_7",
54
+ "/bert/encoder/layer.1/attention/self/Constant_8",
55
+ "/bert/encoder/layer.1/attention/self/Constant_9",
56
+ "/bert/encoder/layer.1/attention/self/Constant_10",
57
+ "/bert/encoder/layer.1/attention/self/Constant_11",
58
+ "/bert/encoder/layer.1/attention/self/Constant_12",
59
+ "/bert/encoder/layer.1/attention/self/Constant_13",
60
+ "/bert/encoder/layer.1/attention/self/Constant_14",
61
+ "/bert/encoder/layer.1/attention/self/Constant_15",
62
+ "/bert/encoder/layer.1/attention/output/LayerNorm/Constant",
63
+ "/bert/encoder/layer.1/attention/output/LayerNorm/Constant_1",
64
+ "/bert/encoder/layer.1/intermediate/intermediate_act_fn/Constant",
65
+ "/bert/encoder/layer.1/intermediate/intermediate_act_fn/Constant_1",
66
+ "/bert/encoder/layer.1/intermediate/intermediate_act_fn/Constant_2",
67
+ "/bert/encoder/layer.1/output/LayerNorm/Constant",
68
+ "/bert/encoder/layer.1/output/LayerNorm/Constant_1",
69
+ "/bert/encoder/layer.2/attention/self/Constant",
70
+ "/bert/encoder/layer.2/attention/self/Constant_1",
71
+ "/bert/encoder/layer.2/attention/self/Constant_2",
72
+ "/bert/encoder/layer.2/attention/self/Constant_3",
73
+ "/bert/encoder/layer.2/attention/self/Constant_4",
74
+ "/bert/encoder/layer.2/attention/self/Constant_5",
75
+ "/bert/encoder/layer.2/attention/self/Constant_6",
76
+ "/bert/encoder/layer.2/attention/self/Constant_7",
77
+ "/bert/encoder/layer.2/attention/self/Constant_8",
78
+ "/bert/encoder/layer.2/attention/self/Constant_9",
79
+ "/bert/encoder/layer.2/attention/self/Constant_10",
80
+ "/bert/encoder/layer.2/attention/self/Constant_11",
81
+ "/bert/encoder/layer.2/attention/self/Constant_12",
82
+ "/bert/encoder/layer.2/attention/self/Constant_13",
83
+ "/bert/encoder/layer.2/attention/self/Constant_14",
84
+ "/bert/encoder/layer.2/attention/self/Constant_15",
85
+ "/bert/encoder/layer.2/attention/output/LayerNorm/Constant",
86
+ "/bert/encoder/layer.2/attention/output/LayerNorm/Constant_1",
87
+ "/bert/encoder/layer.2/intermediate/intermediate_act_fn/Constant",
88
+ "/bert/encoder/layer.2/intermediate/intermediate_act_fn/Constant_1",
89
+ "/bert/encoder/layer.2/intermediate/intermediate_act_fn/Constant_2",
90
+ "/bert/encoder/layer.2/output/LayerNorm/Constant",
91
+ "/bert/encoder/layer.2/output/LayerNorm/Constant_1",
92
+ "/bert/encoder/layer.3/attention/self/Constant",
93
+ "/bert/encoder/layer.3/attention/self/Constant_1",
94
+ "/bert/encoder/layer.3/attention/self/Constant_2",
95
+ "/bert/encoder/layer.3/attention/self/Constant_3",
96
+ "/bert/encoder/layer.3/attention/self/Constant_4",
97
+ "/bert/encoder/layer.3/attention/self/Constant_5",
98
+ "/bert/encoder/layer.3/attention/self/Constant_6",
99
+ "/bert/encoder/layer.3/attention/self/Constant_7",
100
+ "/bert/encoder/layer.3/attention/self/Constant_8",
101
+ "/bert/encoder/layer.3/attention/self/Constant_9",
102
+ "/bert/encoder/layer.3/attention/self/Constant_10",
103
+ "/bert/encoder/layer.3/attention/self/Constant_11",
104
+ "/bert/encoder/layer.3/attention/self/Constant_12",
105
+ "/bert/encoder/layer.3/attention/self/Constant_13",
106
+ "/bert/encoder/layer.3/attention/self/Constant_14",
107
+ "/bert/encoder/layer.3/attention/self/Constant_15",
108
+ "/bert/encoder/layer.3/attention/output/LayerNorm/Constant",
109
+ "/bert/encoder/layer.3/attention/output/LayerNorm/Constant_1",
110
+ "/bert/encoder/layer.3/intermediate/intermediate_act_fn/Constant",
111
+ "/bert/encoder/layer.3/intermediate/intermediate_act_fn/Constant_1",
112
+ "/bert/encoder/layer.3/intermediate/intermediate_act_fn/Constant_2",
113
+ "/bert/encoder/layer.3/output/LayerNorm/Constant",
114
+ "/bert/encoder/layer.3/output/LayerNorm/Constant_1",
115
+ "/bert/encoder/layer.4/attention/self/Constant",
116
+ "/bert/encoder/layer.4/attention/self/Constant_1",
117
+ "/bert/encoder/layer.4/attention/self/Constant_2",
118
+ "/bert/encoder/layer.4/attention/self/Constant_3",
119
+ "/bert/encoder/layer.4/attention/self/Constant_4",
120
+ "/bert/encoder/layer.4/attention/self/Constant_5",
121
+ "/bert/encoder/layer.4/attention/self/Constant_6",
122
+ "/bert/encoder/layer.4/attention/self/Constant_7",
123
+ "/bert/encoder/layer.4/attention/self/Constant_8",
124
+ "/bert/encoder/layer.4/attention/self/Constant_9",
125
+ "/bert/encoder/layer.4/attention/self/Constant_10",
126
+ "/bert/encoder/layer.4/attention/self/Constant_11",
127
+ "/bert/encoder/layer.4/attention/self/Constant_12",
128
+ "/bert/encoder/layer.4/attention/self/Constant_13",
129
+ "/bert/encoder/layer.4/attention/self/Constant_14",
130
+ "/bert/encoder/layer.4/attention/self/Constant_15",
131
+ "/bert/encoder/layer.4/attention/output/LayerNorm/Constant",
132
+ "/bert/encoder/layer.4/attention/output/LayerNorm/Constant_1",
133
+ "/bert/encoder/layer.4/intermediate/intermediate_act_fn/Constant",
134
+ "/bert/encoder/layer.4/intermediate/intermediate_act_fn/Constant_1",
135
+ "/bert/encoder/layer.4/intermediate/intermediate_act_fn/Constant_2",
136
+ "/bert/encoder/layer.4/output/LayerNorm/Constant",
137
+ "/bert/encoder/layer.4/output/LayerNorm/Constant_1",
138
+ "/bert/encoder/layer.5/attention/self/Constant",
139
+ "/bert/encoder/layer.5/attention/self/Constant_1",
140
+ "/bert/encoder/layer.5/attention/self/Constant_2",
141
+ "/bert/encoder/layer.5/attention/self/Constant_3",
142
+ "/bert/encoder/layer.5/attention/self/Constant_4",
143
+ "/bert/encoder/layer.5/attention/self/Constant_5",
144
+ "/bert/encoder/layer.5/attention/self/Constant_6",
145
+ "/bert/encoder/layer.5/attention/self/Constant_7",
146
+ "/bert/encoder/layer.5/attention/self/Constant_8",
147
+ "/bert/encoder/layer.5/attention/self/Constant_9",
148
+ "/bert/encoder/layer.5/attention/self/Constant_10",
149
+ "/bert/encoder/layer.5/attention/self/Constant_11",
150
+ "/bert/encoder/layer.5/attention/self/Constant_12",
151
+ "/bert/encoder/layer.5/attention/self/Constant_13",
152
+ "/bert/encoder/layer.5/attention/self/Constant_14",
153
+ "/bert/encoder/layer.5/attention/self/Constant_15",
154
+ "/bert/encoder/layer.5/attention/output/LayerNorm/Constant",
155
+ "/bert/encoder/layer.5/attention/output/LayerNorm/Constant_1",
156
+ "/bert/encoder/layer.5/intermediate/intermediate_act_fn/Constant",
157
+ "/bert/encoder/layer.5/intermediate/intermediate_act_fn/Constant_1",
158
+ "/bert/encoder/layer.5/intermediate/intermediate_act_fn/Constant_2",
159
+ "/bert/encoder/layer.5/output/LayerNorm/Constant",
160
+ "/bert/encoder/layer.5/output/LayerNorm/Constant_1",
161
+ "/bert/Unsqueeze",
162
+ "/bert/embeddings/Shape",
163
+ "/bert/embeddings/word_embeddings/Gather",
164
+ "/bert/embeddings/token_type_embeddings/Gather",
165
+ "/bert/Unsqueeze_1",
166
+ "/bert/embeddings/Gather",
167
+ "/bert/embeddings/Add",
168
+ "/bert/Cast",
169
+ "/bert/embeddings/Unsqueeze",
170
+ "/bert/Sub",
171
+ "/bert/embeddings/Slice",
172
+ "/bert/Mul",
173
+ "/bert/embeddings/position_embeddings/Gather",
174
+ "/bert/embeddings/Add_1",
175
+ "/bert/embeddings/LayerNorm/ReduceMean",
176
+ "/bert/embeddings/LayerNorm/Sub",
177
+ "/bert/embeddings/LayerNorm/Pow",
178
+ "/bert/embeddings/LayerNorm/ReduceMean_1",
179
+ "/bert/embeddings/LayerNorm/Add",
180
+ "/bert/embeddings/LayerNorm/Sqrt",
181
+ "/bert/embeddings/LayerNorm/Div",
182
+ "/bert/embeddings/LayerNorm/Mul",
183
+ "/bert/embeddings/LayerNorm/Add_1",
184
+ "/bert/encoder/layer.0/attention/self/query/MatMul",
185
+ "/bert/encoder/layer.0/attention/self/key/MatMul",
186
+ "/bert/encoder/layer.0/attention/self/value/MatMul",
187
+ "/bert/encoder/layer.0/attention/self/query/Add",
188
+ "/bert/encoder/layer.0/attention/self/key/Add",
189
+ "/bert/encoder/layer.0/attention/self/value/Add",
190
+ "/bert/encoder/layer.0/attention/self/Shape_4",
191
+ "/bert/encoder/layer.0/attention/self/Shape_5",
192
+ "/bert/encoder/layer.0/attention/self/Shape",
193
+ "/bert/encoder/layer.0/attention/self/Shape_1",
194
+ "/bert/encoder/layer.0/attention/self/Shape_2",
195
+ "/bert/encoder/layer.0/attention/self/Shape_3",
196
+ "/bert/encoder/layer.0/attention/self/Gather_4",
197
+ "/bert/encoder/layer.0/attention/self/Gather_5",
198
+ "/bert/encoder/layer.0/attention/self/Gather",
199
+ "/bert/encoder/layer.0/attention/self/Gather_1",
200
+ "/bert/encoder/layer.0/attention/self/Gather_2",
201
+ "/bert/encoder/layer.0/attention/self/Gather_3",
202
+ "/bert/encoder/layer.0/attention/self/Unsqueeze_4",
203
+ "/bert/encoder/layer.0/attention/self/Unsqueeze_5",
204
+ "/bert/encoder/layer.0/attention/self/Unsqueeze",
205
+ "/bert/encoder/layer.0/attention/self/Unsqueeze_1",
206
+ "/bert/encoder/layer.0/attention/self/Unsqueeze_2",
207
+ "/bert/encoder/layer.0/attention/self/Unsqueeze_3",
208
+ "/bert/encoder/layer.0/attention/self/Concat_2",
209
+ "/bert/encoder/layer.0/attention/self/Concat",
210
+ "/bert/encoder/layer.0/attention/self/Concat_1",
211
+ "/bert/encoder/layer.0/attention/self/Reshape_2",
212
+ "/bert/encoder/layer.0/attention/self/Reshape",
213
+ "/bert/encoder/layer.0/attention/self/Reshape_1",
214
+ "/bert/encoder/layer.0/attention/self/Transpose_1",
215
+ "/bert/encoder/layer.0/attention/self/Transpose_2",
216
+ "/bert/encoder/layer.0/attention/self/Transpose",
217
+ "/bert/encoder/layer.0/attention/self/MatMul",
218
+ "/bert/encoder/layer.0/attention/self/Div",
219
+ "/bert/encoder/layer.0/attention/self/Add",
220
+ "/bert/encoder/layer.0/attention/self/Softmax",
221
+ "/bert/encoder/layer.0/attention/self/MatMul_1",
222
+ "/bert/encoder/layer.0/attention/self/Transpose_3",
223
+ "/bert/encoder/layer.0/attention/self/Shape_6",
224
+ "/bert/encoder/layer.0/attention/self/Shape_7",
225
+ "/bert/encoder/layer.0/attention/self/Gather_6",
226
+ "/bert/encoder/layer.0/attention/self/Gather_7",
227
+ "/bert/encoder/layer.0/attention/self/Unsqueeze_6",
228
+ "/bert/encoder/layer.0/attention/self/Unsqueeze_7",
229
+ "/bert/encoder/layer.0/attention/self/Concat_3",
230
+ "/bert/encoder/layer.0/attention/self/Reshape_3",
231
+ "/bert/encoder/layer.0/attention/output/dense/MatMul",
232
+ "/bert/encoder/layer.0/attention/output/dense/Add",
233
+ "/bert/encoder/layer.0/attention/output/Add",
234
+ "/bert/encoder/layer.0/attention/output/LayerNorm/ReduceMean",
235
+ "/bert/encoder/layer.0/attention/output/LayerNorm/Sub",
236
+ "/bert/encoder/layer.0/attention/output/LayerNorm/Pow",
237
+ "/bert/encoder/layer.0/attention/output/LayerNorm/ReduceMean_1",
238
+ "/bert/encoder/layer.0/attention/output/LayerNorm/Add",
239
+ "/bert/encoder/layer.0/attention/output/LayerNorm/Sqrt",
240
+ "/bert/encoder/layer.0/attention/output/LayerNorm/Div",
241
+ "/bert/encoder/layer.0/attention/output/LayerNorm/Mul",
242
+ "/bert/encoder/layer.0/attention/output/LayerNorm/Add_1",
243
+ "/bert/encoder/layer.0/intermediate/dense/MatMul",
244
+ "/bert/encoder/layer.0/intermediate/dense/Add",
245
+ "/bert/encoder/layer.0/intermediate/intermediate_act_fn/Div",
246
+ "/bert/encoder/layer.0/intermediate/intermediate_act_fn/Erf",
247
+ "/bert/encoder/layer.0/intermediate/intermediate_act_fn/Add",
248
+ "/bert/encoder/layer.0/intermediate/intermediate_act_fn/Mul",
249
+ "/bert/encoder/layer.0/intermediate/intermediate_act_fn/Mul_1",
250
+ "/bert/encoder/layer.0/output/dense/MatMul",
251
+ "/bert/encoder/layer.0/output/dense/Add",
252
+ "/bert/encoder/layer.0/output/Add",
253
+ "/bert/encoder/layer.0/output/LayerNorm/ReduceMean",
254
+ "/bert/encoder/layer.0/output/LayerNorm/Sub",
255
+ "/bert/encoder/layer.0/output/LayerNorm/Pow",
256
+ "/bert/encoder/layer.0/output/LayerNorm/ReduceMean_1",
257
+ "/bert/encoder/layer.0/output/LayerNorm/Add",
258
+ "/bert/encoder/layer.0/output/LayerNorm/Sqrt",
259
+ "/bert/encoder/layer.0/output/LayerNorm/Div",
260
+ "/bert/encoder/layer.0/output/LayerNorm/Mul",
261
+ "/bert/encoder/layer.0/output/LayerNorm/Add_1",
262
+ "/bert/encoder/layer.1/attention/self/query/MatMul",
263
+ "/bert/encoder/layer.1/attention/self/key/MatMul",
264
+ "/bert/encoder/layer.1/attention/self/value/MatMul",
265
+ "/bert/encoder/layer.1/attention/self/query/Add",
266
+ "/bert/encoder/layer.1/attention/self/key/Add",
267
+ "/bert/encoder/layer.1/attention/self/value/Add",
268
+ "/bert/encoder/layer.1/attention/self/Shape_4",
269
+ "/bert/encoder/layer.1/attention/self/Shape_5",
270
+ "/bert/encoder/layer.1/attention/self/Shape",
271
+ "/bert/encoder/layer.1/attention/self/Shape_1",
272
+ "/bert/encoder/layer.1/attention/self/Shape_2",
273
+ "/bert/encoder/layer.1/attention/self/Shape_3",
274
+ "/bert/encoder/layer.1/attention/self/Gather_4",
275
+ "/bert/encoder/layer.1/attention/self/Gather_5",
276
+ "/bert/encoder/layer.1/attention/self/Gather",
277
+ "/bert/encoder/layer.1/attention/self/Gather_1",
278
+ "/bert/encoder/layer.1/attention/self/Gather_2",
279
+ "/bert/encoder/layer.1/attention/self/Gather_3",
280
+ "/bert/encoder/layer.1/attention/self/Unsqueeze_4",
281
+ "/bert/encoder/layer.1/attention/self/Unsqueeze_5",
282
+ "/bert/encoder/layer.1/attention/self/Unsqueeze",
283
+ "/bert/encoder/layer.1/attention/self/Unsqueeze_1",
284
+ "/bert/encoder/layer.1/attention/self/Unsqueeze_2",
285
+ "/bert/encoder/layer.1/attention/self/Unsqueeze_3",
286
+ "/bert/encoder/layer.1/attention/self/Concat_2",
287
+ "/bert/encoder/layer.1/attention/self/Concat",
288
+ "/bert/encoder/layer.1/attention/self/Concat_1",
289
+ "/bert/encoder/layer.1/attention/self/Reshape_2",
290
+ "/bert/encoder/layer.1/attention/self/Reshape",
291
+ "/bert/encoder/layer.1/attention/self/Reshape_1",
292
+ "/bert/encoder/layer.1/attention/self/Transpose_1",
293
+ "/bert/encoder/layer.1/attention/self/Transpose_2",
294
+ "/bert/encoder/layer.1/attention/self/Transpose",
295
+ "/bert/encoder/layer.1/attention/self/MatMul",
296
+ "/bert/encoder/layer.1/attention/self/Div",
297
+ "/bert/encoder/layer.1/attention/self/Add",
298
+ "/bert/encoder/layer.1/attention/self/Softmax",
299
+ "/bert/encoder/layer.1/attention/self/MatMul_1",
300
+ "/bert/encoder/layer.1/attention/self/Transpose_3",
301
+ "/bert/encoder/layer.1/attention/self/Shape_6",
302
+ "/bert/encoder/layer.1/attention/self/Shape_7",
303
+ "/bert/encoder/layer.1/attention/self/Gather_6",
304
+ "/bert/encoder/layer.1/attention/self/Gather_7",
305
+ "/bert/encoder/layer.1/attention/self/Unsqueeze_6",
306
+ "/bert/encoder/layer.1/attention/self/Unsqueeze_7",
307
+ "/bert/encoder/layer.1/attention/self/Concat_3",
308
+ "/bert/encoder/layer.1/attention/self/Reshape_3",
309
+ "/bert/encoder/layer.1/attention/output/dense/MatMul",
310
+ "/bert/encoder/layer.1/attention/output/dense/Add",
311
+ "/bert/encoder/layer.1/attention/output/Add",
312
+ "/bert/encoder/layer.1/attention/output/LayerNorm/ReduceMean",
313
+ "/bert/encoder/layer.1/attention/output/LayerNorm/Sub",
314
+ "/bert/encoder/layer.1/attention/output/LayerNorm/Pow",
315
+ "/bert/encoder/layer.1/attention/output/LayerNorm/ReduceMean_1",
316
+ "/bert/encoder/layer.1/attention/output/LayerNorm/Add",
317
+ "/bert/encoder/layer.1/attention/output/LayerNorm/Sqrt",
318
+ "/bert/encoder/layer.1/attention/output/LayerNorm/Div",
319
+ "/bert/encoder/layer.1/attention/output/LayerNorm/Mul",
320
+ "/bert/encoder/layer.1/attention/output/LayerNorm/Add_1",
321
+ "/bert/encoder/layer.1/intermediate/dense/MatMul",
322
+ "/bert/encoder/layer.1/intermediate/dense/Add",
323
+ "/bert/encoder/layer.1/intermediate/intermediate_act_fn/Div",
324
+ "/bert/encoder/layer.1/intermediate/intermediate_act_fn/Erf",
325
+ "/bert/encoder/layer.1/intermediate/intermediate_act_fn/Add",
326
+ "/bert/encoder/layer.1/intermediate/intermediate_act_fn/Mul",
327
+ "/bert/encoder/layer.1/intermediate/intermediate_act_fn/Mul_1",
328
+ "/bert/encoder/layer.1/output/dense/MatMul",
329
+ "/bert/encoder/layer.1/output/dense/Add",
330
+ "/bert/encoder/layer.1/output/Add",
331
+ "/bert/encoder/layer.1/output/LayerNorm/ReduceMean",
332
+ "/bert/encoder/layer.1/output/LayerNorm/Sub",
333
+ "/bert/encoder/layer.1/output/LayerNorm/Pow",
334
+ "/bert/encoder/layer.1/output/LayerNorm/ReduceMean_1",
335
+ "/bert/encoder/layer.1/output/LayerNorm/Add",
336
+ "/bert/encoder/layer.1/output/LayerNorm/Sqrt",
337
+ "/bert/encoder/layer.1/output/LayerNorm/Div",
338
+ "/bert/encoder/layer.1/output/LayerNorm/Mul",
339
+ "/bert/encoder/layer.1/output/LayerNorm/Add_1",
340
+ "/bert/encoder/layer.2/attention/self/query/MatMul",
341
+ "/bert/encoder/layer.2/attention/self/key/MatMul",
342
+ "/bert/encoder/layer.2/attention/self/value/MatMul",
343
+ "/bert/encoder/layer.2/attention/self/query/Add",
344
+ "/bert/encoder/layer.2/attention/self/key/Add",
345
+ "/bert/encoder/layer.2/attention/self/value/Add",
346
+ "/bert/encoder/layer.2/attention/self/Shape_4",
347
+ "/bert/encoder/layer.2/attention/self/Shape_5",
348
+ "/bert/encoder/layer.2/attention/self/Shape",
349
+ "/bert/encoder/layer.2/attention/self/Shape_1",
350
+ "/bert/encoder/layer.2/attention/self/Shape_2",
351
+ "/bert/encoder/layer.2/attention/self/Shape_3",
352
+ "/bert/encoder/layer.2/attention/self/Gather_4",
353
+ "/bert/encoder/layer.2/attention/self/Gather_5",
354
+ "/bert/encoder/layer.2/attention/self/Gather",
355
+ "/bert/encoder/layer.2/attention/self/Gather_1",
356
+ "/bert/encoder/layer.2/attention/self/Gather_2",
357
+ "/bert/encoder/layer.2/attention/self/Gather_3",
358
+ "/bert/encoder/layer.2/attention/self/Unsqueeze_4",
359
+ "/bert/encoder/layer.2/attention/self/Unsqueeze_5",
360
+ "/bert/encoder/layer.2/attention/self/Unsqueeze",
361
+ "/bert/encoder/layer.2/attention/self/Unsqueeze_1",
362
+ "/bert/encoder/layer.2/attention/self/Unsqueeze_2",
363
+ "/bert/encoder/layer.2/attention/self/Unsqueeze_3",
364
+ "/bert/encoder/layer.2/attention/self/Concat_2",
365
+ "/bert/encoder/layer.2/attention/self/Concat",
366
+ "/bert/encoder/layer.2/attention/self/Concat_1",
367
+ "/bert/encoder/layer.2/attention/self/Reshape_2",
368
+ "/bert/encoder/layer.2/attention/self/Reshape",
369
+ "/bert/encoder/layer.2/attention/self/Reshape_1",
370
+ "/bert/encoder/layer.2/attention/self/Transpose_1",
371
+ "/bert/encoder/layer.2/attention/self/Transpose_2",
372
+ "/bert/encoder/layer.2/attention/self/Transpose",
373
+ "/bert/encoder/layer.2/attention/self/MatMul",
374
+ "/bert/encoder/layer.2/attention/self/Div",
375
+ "/bert/encoder/layer.2/attention/self/Add",
376
+ "/bert/encoder/layer.2/attention/self/Softmax",
377
+ "/bert/encoder/layer.2/attention/self/MatMul_1",
378
+ "/bert/encoder/layer.2/attention/self/Transpose_3",
379
+ "/bert/encoder/layer.2/attention/self/Shape_6",
380
+ "/bert/encoder/layer.2/attention/self/Shape_7",
381
+ "/bert/encoder/layer.2/attention/self/Gather_6",
382
+ "/bert/encoder/layer.2/attention/self/Gather_7",
383
+ "/bert/encoder/layer.2/attention/self/Unsqueeze_6",
384
+ "/bert/encoder/layer.2/attention/self/Unsqueeze_7",
385
+ "/bert/encoder/layer.2/attention/self/Concat_3",
386
+ "/bert/encoder/layer.2/attention/self/Reshape_3",
387
+ "/bert/encoder/layer.2/attention/output/dense/MatMul",
388
+ "/bert/encoder/layer.2/attention/output/dense/Add",
389
+ "/bert/encoder/layer.2/attention/output/Add",
390
+ "/bert/encoder/layer.2/attention/output/LayerNorm/ReduceMean",
391
+ "/bert/encoder/layer.2/attention/output/LayerNorm/Sub",
392
+ "/bert/encoder/layer.2/attention/output/LayerNorm/Pow",
393
+ "/bert/encoder/layer.2/attention/output/LayerNorm/ReduceMean_1",
394
+ "/bert/encoder/layer.2/attention/output/LayerNorm/Add",
395
+ "/bert/encoder/layer.2/attention/output/LayerNorm/Sqrt",
396
+ "/bert/encoder/layer.2/attention/output/LayerNorm/Div",
397
+ "/bert/encoder/layer.2/attention/output/LayerNorm/Mul",
398
+ "/bert/encoder/layer.2/attention/output/LayerNorm/Add_1",
399
+ "/bert/encoder/layer.2/intermediate/dense/MatMul",
400
+ "/bert/encoder/layer.2/intermediate/dense/Add",
401
+ "/bert/encoder/layer.2/intermediate/intermediate_act_fn/Div",
402
+ "/bert/encoder/layer.2/intermediate/intermediate_act_fn/Erf",
403
+ "/bert/encoder/layer.2/intermediate/intermediate_act_fn/Add",
404
+ "/bert/encoder/layer.2/intermediate/intermediate_act_fn/Mul",
405
+ "/bert/encoder/layer.2/intermediate/intermediate_act_fn/Mul_1",
406
+ "/bert/encoder/layer.2/output/dense/MatMul",
407
+ "/bert/encoder/layer.2/output/dense/Add",
408
+ "/bert/encoder/layer.2/output/Add",
409
+ "/bert/encoder/layer.2/output/LayerNorm/ReduceMean",
410
+ "/bert/encoder/layer.2/output/LayerNorm/Sub",
411
+ "/bert/encoder/layer.2/output/LayerNorm/Pow",
412
+ "/bert/encoder/layer.2/output/LayerNorm/ReduceMean_1",
413
+ "/bert/encoder/layer.2/output/LayerNorm/Add",
414
+ "/bert/encoder/layer.2/output/LayerNorm/Sqrt",
415
+ "/bert/encoder/layer.2/output/LayerNorm/Div",
416
+ "/bert/encoder/layer.2/output/LayerNorm/Mul",
417
+ "/bert/encoder/layer.2/output/LayerNorm/Add_1",
418
+ "/bert/encoder/layer.3/attention/self/query/MatMul",
419
+ "/bert/encoder/layer.3/attention/self/key/MatMul",
420
+ "/bert/encoder/layer.3/attention/self/value/MatMul",
421
+ "/bert/encoder/layer.3/attention/self/query/Add",
422
+ "/bert/encoder/layer.3/attention/self/key/Add",
423
+ "/bert/encoder/layer.3/attention/self/value/Add",
424
+ "/bert/encoder/layer.3/attention/self/Shape_4",
425
+ "/bert/encoder/layer.3/attention/self/Shape_5",
426
+ "/bert/encoder/layer.3/attention/self/Shape",
427
+ "/bert/encoder/layer.3/attention/self/Shape_1",
428
+ "/bert/encoder/layer.3/attention/self/Shape_2",
429
+ "/bert/encoder/layer.3/attention/self/Shape_3",
430
+ "/bert/encoder/layer.3/attention/self/Gather_4",
431
+ "/bert/encoder/layer.3/attention/self/Gather_5",
432
+ "/bert/encoder/layer.3/attention/self/Gather",
433
+ "/bert/encoder/layer.3/attention/self/Gather_1",
434
+ "/bert/encoder/layer.3/attention/self/Gather_2",
435
+ "/bert/encoder/layer.3/attention/self/Gather_3",
436
+ "/bert/encoder/layer.3/attention/self/Unsqueeze_4",
437
+ "/bert/encoder/layer.3/attention/self/Unsqueeze_5",
438
+ "/bert/encoder/layer.3/attention/self/Unsqueeze",
439
+ "/bert/encoder/layer.3/attention/self/Unsqueeze_1",
440
+ "/bert/encoder/layer.3/attention/self/Unsqueeze_2",
441
+ "/bert/encoder/layer.3/attention/self/Unsqueeze_3",
442
+ "/bert/encoder/layer.3/attention/self/Concat_2",
443
+ "/bert/encoder/layer.3/attention/self/Concat",
444
+ "/bert/encoder/layer.3/attention/self/Concat_1",
445
+ "/bert/encoder/layer.3/attention/self/Reshape_2",
446
+ "/bert/encoder/layer.3/attention/self/Reshape",
447
+ "/bert/encoder/layer.3/attention/self/Reshape_1",
448
+ "/bert/encoder/layer.3/attention/self/Transpose_1",
449
+ "/bert/encoder/layer.3/attention/self/Transpose_2",
450
+ "/bert/encoder/layer.3/attention/self/Transpose",
451
+ "/bert/encoder/layer.3/attention/self/MatMul",
452
+ "/bert/encoder/layer.3/attention/self/Div",
453
+ "/bert/encoder/layer.3/attention/self/Add",
454
+ "/bert/encoder/layer.3/attention/self/Softmax",
455
+ "/bert/encoder/layer.3/attention/self/MatMul_1",
456
+ "/bert/encoder/layer.3/attention/self/Transpose_3",
457
+ "/bert/encoder/layer.3/attention/self/Shape_6",
458
+ "/bert/encoder/layer.3/attention/self/Shape_7",
459
+ "/bert/encoder/layer.3/attention/self/Gather_6",
460
+ "/bert/encoder/layer.3/attention/self/Gather_7",
461
+ "/bert/encoder/layer.3/attention/self/Unsqueeze_6",
462
+ "/bert/encoder/layer.3/attention/self/Unsqueeze_7",
463
+ "/bert/encoder/layer.3/attention/self/Concat_3",
464
+ "/bert/encoder/layer.3/attention/self/Reshape_3",
465
+ "/bert/encoder/layer.3/attention/output/dense/MatMul",
466
+ "/bert/encoder/layer.3/attention/output/dense/Add",
467
+ "/bert/encoder/layer.3/attention/output/Add",
468
+ "/bert/encoder/layer.3/attention/output/LayerNorm/ReduceMean",
469
+ "/bert/encoder/layer.3/attention/output/LayerNorm/Sub",
470
+ "/bert/encoder/layer.3/attention/output/LayerNorm/Pow",
471
+ "/bert/encoder/layer.3/attention/output/LayerNorm/ReduceMean_1",
472
+ "/bert/encoder/layer.3/attention/output/LayerNorm/Add",
473
+ "/bert/encoder/layer.3/attention/output/LayerNorm/Sqrt",
474
+ "/bert/encoder/layer.3/attention/output/LayerNorm/Div",
475
+ "/bert/encoder/layer.3/attention/output/LayerNorm/Mul",
476
+ "/bert/encoder/layer.3/attention/output/LayerNorm/Add_1",
477
+ "/bert/encoder/layer.3/intermediate/dense/MatMul",
478
+ "/bert/encoder/layer.3/intermediate/dense/Add",
479
+ "/bert/encoder/layer.3/intermediate/intermediate_act_fn/Div",
480
+ "/bert/encoder/layer.3/intermediate/intermediate_act_fn/Erf",
481
+ "/bert/encoder/layer.3/intermediate/intermediate_act_fn/Add",
482
+ "/bert/encoder/layer.3/intermediate/intermediate_act_fn/Mul",
483
+ "/bert/encoder/layer.3/intermediate/intermediate_act_fn/Mul_1",
484
+ "/bert/encoder/layer.3/output/dense/MatMul",
485
+ "/bert/encoder/layer.3/output/dense/Add",
486
+ "/bert/encoder/layer.3/output/Add",
487
+ "/bert/encoder/layer.3/output/LayerNorm/ReduceMean",
488
+ "/bert/encoder/layer.3/output/LayerNorm/Sub",
489
+ "/bert/encoder/layer.3/output/LayerNorm/Pow",
490
+ "/bert/encoder/layer.3/output/LayerNorm/ReduceMean_1",
491
+ "/bert/encoder/layer.3/output/LayerNorm/Add",
492
+ "/bert/encoder/layer.3/output/LayerNorm/Sqrt",
493
+ "/bert/encoder/layer.3/output/LayerNorm/Div",
494
+ "/bert/encoder/layer.3/output/LayerNorm/Mul",
495
+ "/bert/encoder/layer.3/output/LayerNorm/Add_1",
496
+ "/bert/encoder/layer.4/attention/self/query/MatMul",
497
+ "/bert/encoder/layer.4/attention/self/key/MatMul",
498
+ "/bert/encoder/layer.4/attention/self/value/MatMul",
499
+ "/bert/encoder/layer.4/attention/self/query/Add",
500
+ "/bert/encoder/layer.4/attention/self/key/Add",
501
+ "/bert/encoder/layer.4/attention/self/value/Add",
502
+ "/bert/encoder/layer.4/attention/self/Shape_4",
503
+ "/bert/encoder/layer.4/attention/self/Shape_5",
504
+ "/bert/encoder/layer.4/attention/self/Shape",
505
+ "/bert/encoder/layer.4/attention/self/Shape_1",
506
+ "/bert/encoder/layer.4/attention/self/Shape_2",
507
+ "/bert/encoder/layer.4/attention/self/Shape_3",
508
+ "/bert/encoder/layer.4/attention/self/Gather_4",
509
+ "/bert/encoder/layer.4/attention/self/Gather_5",
510
+ "/bert/encoder/layer.4/attention/self/Gather",
511
+ "/bert/encoder/layer.4/attention/self/Gather_1",
512
+ "/bert/encoder/layer.4/attention/self/Gather_2",
513
+ "/bert/encoder/layer.4/attention/self/Gather_3",
514
+ "/bert/encoder/layer.4/attention/self/Unsqueeze_4",
515
+ "/bert/encoder/layer.4/attention/self/Unsqueeze_5",
516
+ "/bert/encoder/layer.4/attention/self/Unsqueeze",
517
+ "/bert/encoder/layer.4/attention/self/Unsqueeze_1",
518
+ "/bert/encoder/layer.4/attention/self/Unsqueeze_2",
519
+ "/bert/encoder/layer.4/attention/self/Unsqueeze_3",
520
+ "/bert/encoder/layer.4/attention/self/Concat_2",
521
+ "/bert/encoder/layer.4/attention/self/Concat",
522
+ "/bert/encoder/layer.4/attention/self/Concat_1",
523
+ "/bert/encoder/layer.4/attention/self/Reshape_2",
524
+ "/bert/encoder/layer.4/attention/self/Reshape",
525
+ "/bert/encoder/layer.4/attention/self/Reshape_1",
526
+ "/bert/encoder/layer.4/attention/self/Transpose_1",
527
+ "/bert/encoder/layer.4/attention/self/Transpose_2",
528
+ "/bert/encoder/layer.4/attention/self/Transpose",
529
+ "/bert/encoder/layer.4/attention/self/MatMul",
530
+ "/bert/encoder/layer.4/attention/self/Div",
531
+ "/bert/encoder/layer.4/attention/self/Add",
532
+ "/bert/encoder/layer.4/attention/self/Softmax",
533
+ "/bert/encoder/layer.4/attention/self/MatMul_1",
534
+ "/bert/encoder/layer.4/attention/self/Transpose_3",
535
+ "/bert/encoder/layer.4/attention/self/Shape_6",
536
+ "/bert/encoder/layer.4/attention/self/Shape_7",
537
+ "/bert/encoder/layer.4/attention/self/Gather_6",
538
+ "/bert/encoder/layer.4/attention/self/Gather_7",
539
+ "/bert/encoder/layer.4/attention/self/Unsqueeze_6",
540
+ "/bert/encoder/layer.4/attention/self/Unsqueeze_7",
541
+ "/bert/encoder/layer.4/attention/self/Concat_3",
542
+ "/bert/encoder/layer.4/attention/self/Reshape_3",
543
+ "/bert/encoder/layer.4/attention/output/dense/MatMul",
544
+ "/bert/encoder/layer.4/attention/output/dense/Add",
545
+ "/bert/encoder/layer.4/attention/output/Add",
546
+ "/bert/encoder/layer.4/attention/output/LayerNorm/ReduceMean",
547
+ "/bert/encoder/layer.4/attention/output/LayerNorm/Sub",
548
+ "/bert/encoder/layer.4/attention/output/LayerNorm/Pow",
549
+ "/bert/encoder/layer.4/attention/output/LayerNorm/ReduceMean_1",
550
+ "/bert/encoder/layer.4/attention/output/LayerNorm/Add",
551
+ "/bert/encoder/layer.4/attention/output/LayerNorm/Sqrt",
552
+ "/bert/encoder/layer.4/attention/output/LayerNorm/Div",
553
+ "/bert/encoder/layer.4/attention/output/LayerNorm/Mul",
554
+ "/bert/encoder/layer.4/attention/output/LayerNorm/Add_1",
555
+ "/bert/encoder/layer.4/intermediate/dense/MatMul",
556
+ "/bert/encoder/layer.4/intermediate/dense/Add",
557
+ "/bert/encoder/layer.4/intermediate/intermediate_act_fn/Div",
558
+ "/bert/encoder/layer.4/intermediate/intermediate_act_fn/Erf",
559
+ "/bert/encoder/layer.4/intermediate/intermediate_act_fn/Add",
560
+ "/bert/encoder/layer.4/intermediate/intermediate_act_fn/Mul",
561
+ "/bert/encoder/layer.4/intermediate/intermediate_act_fn/Mul_1",
562
+ "/bert/encoder/layer.4/output/dense/MatMul",
563
+ "/bert/encoder/layer.4/output/dense/Add",
564
+ "/bert/encoder/layer.4/output/Add",
565
+ "/bert/encoder/layer.4/output/LayerNorm/ReduceMean",
566
+ "/bert/encoder/layer.4/output/LayerNorm/Sub",
567
+ "/bert/encoder/layer.4/output/LayerNorm/Pow",
568
+ "/bert/encoder/layer.4/output/LayerNorm/ReduceMean_1",
569
+ "/bert/encoder/layer.4/output/LayerNorm/Add",
570
+ "/bert/encoder/layer.4/output/LayerNorm/Sqrt",
571
+ "/bert/encoder/layer.4/output/LayerNorm/Div",
572
+ "/bert/encoder/layer.4/output/LayerNorm/Mul",
573
+ "/bert/encoder/layer.4/output/LayerNorm/Add_1"
574
+ ],
575
+ "operators_to_quantize": [
576
+ "Conv",
577
+ "MatMul",
578
+ "Attention",
579
+ "LSTM",
580
+ "Gather",
581
+ "Transpose",
582
+ "EmbedLayerNormalization"
583
+ ],
584
+ "per_channel": false,
585
+ "qdq_add_pair_to_weight": false,
586
+ "qdq_dedicated_pair": false,
587
+ "qdq_op_type_per_channel_support_to_axis": {
588
+ "MatMul": 1
589
+ },
590
+ "reduce_range": false,
591
+ "weights_dtype": "QInt8",
592
+ "weights_symmetric": true
593
+ },
594
+ "transformers_version": "4.38.2",
595
+ "use_external_data_format": false
596
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "mask_token": {
10
+ "content": "[MASK]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "[SEP]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "[UNK]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": false,
48
+ "mask_token": "[MASK]",
49
+ "max_length": 512,
50
+ "model_max_length": 1000000000000000019884624838656,
51
+ "never_split": null,
52
+ "pad_token": "[PAD]",
53
+ "sep_token": "[SEP]",
54
+ "stride": 0,
55
+ "strip_accents": null,
56
+ "tokenize_chinese_chars": true,
57
+ "tokenizer_class": "BertTokenizer",
58
+ "truncation_side": "right",
59
+ "truncation_strategy": "longest_first",
60
+ "unk_token": "[UNK]"
61
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff