shirleylqs commited on
Commit
1b7b7e5
1 Parent(s): b56c525

Upload model and tokenizer

Browse files
config.json ADDED
@@ -0,0 +1,1429 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "mistralai/Mistral-7B-v0.1",
3
+ "architectures": [
4
+ "MistralForSequenceClassification"
5
+ ],
6
+ "bos_token_id": 1,
7
+ "eos_token_id": 2,
8
+ "hidden_act": "silu",
9
+ "hidden_size": 4096,
10
+ "id2label": {
11
+ "0": "LABEL_0",
12
+ "1": "LABEL_1",
13
+ "2": "LABEL_2",
14
+ "3": "LABEL_3",
15
+ "4": "LABEL_4",
16
+ "5": "LABEL_5",
17
+ "6": "LABEL_6",
18
+ "7": "LABEL_7",
19
+ "8": "LABEL_8",
20
+ "9": "LABEL_9",
21
+ "10": "LABEL_10",
22
+ "11": "LABEL_11",
23
+ "12": "LABEL_12",
24
+ "13": "LABEL_13",
25
+ "14": "LABEL_14",
26
+ "15": "LABEL_15",
27
+ "16": "LABEL_16",
28
+ "17": "LABEL_17",
29
+ "18": "LABEL_18",
30
+ "19": "LABEL_19",
31
+ "20": "LABEL_20",
32
+ "21": "LABEL_21",
33
+ "22": "LABEL_22",
34
+ "23": "LABEL_23",
35
+ "24": "LABEL_24",
36
+ "25": "LABEL_25",
37
+ "26": "LABEL_26",
38
+ "27": "LABEL_27",
39
+ "28": "LABEL_28",
40
+ "29": "LABEL_29",
41
+ "30": "LABEL_30",
42
+ "31": "LABEL_31",
43
+ "32": "LABEL_32",
44
+ "33": "LABEL_33",
45
+ "34": "LABEL_34",
46
+ "35": "LABEL_35",
47
+ "36": "LABEL_36",
48
+ "37": "LABEL_37",
49
+ "38": "LABEL_38",
50
+ "39": "LABEL_39",
51
+ "40": "LABEL_40",
52
+ "41": "LABEL_41",
53
+ "42": "LABEL_42",
54
+ "43": "LABEL_43",
55
+ "44": "LABEL_44",
56
+ "45": "LABEL_45",
57
+ "46": "LABEL_46",
58
+ "47": "LABEL_47",
59
+ "48": "LABEL_48",
60
+ "49": "LABEL_49",
61
+ "50": "LABEL_50",
62
+ "51": "LABEL_51",
63
+ "52": "LABEL_52",
64
+ "53": "LABEL_53",
65
+ "54": "LABEL_54",
66
+ "55": "LABEL_55",
67
+ "56": "LABEL_56",
68
+ "57": "LABEL_57",
69
+ "58": "LABEL_58",
70
+ "59": "LABEL_59",
71
+ "60": "LABEL_60",
72
+ "61": "LABEL_61",
73
+ "62": "LABEL_62",
74
+ "63": "LABEL_63",
75
+ "64": "LABEL_64",
76
+ "65": "LABEL_65",
77
+ "66": "LABEL_66",
78
+ "67": "LABEL_67",
79
+ "68": "LABEL_68",
80
+ "69": "LABEL_69",
81
+ "70": "LABEL_70",
82
+ "71": "LABEL_71",
83
+ "72": "LABEL_72",
84
+ "73": "LABEL_73",
85
+ "74": "LABEL_74",
86
+ "75": "LABEL_75",
87
+ "76": "LABEL_76",
88
+ "77": "LABEL_77",
89
+ "78": "LABEL_78",
90
+ "79": "LABEL_79",
91
+ "80": "LABEL_80",
92
+ "81": "LABEL_81",
93
+ "82": "LABEL_82",
94
+ "83": "LABEL_83",
95
+ "84": "LABEL_84",
96
+ "85": "LABEL_85",
97
+ "86": "LABEL_86",
98
+ "87": "LABEL_87",
99
+ "88": "LABEL_88",
100
+ "89": "LABEL_89",
101
+ "90": "LABEL_90",
102
+ "91": "LABEL_91",
103
+ "92": "LABEL_92",
104
+ "93": "LABEL_93",
105
+ "94": "LABEL_94",
106
+ "95": "LABEL_95",
107
+ "96": "LABEL_96",
108
+ "97": "LABEL_97",
109
+ "98": "LABEL_98",
110
+ "99": "LABEL_99",
111
+ "100": "LABEL_100",
112
+ "101": "LABEL_101",
113
+ "102": "LABEL_102",
114
+ "103": "LABEL_103",
115
+ "104": "LABEL_104",
116
+ "105": "LABEL_105",
117
+ "106": "LABEL_106",
118
+ "107": "LABEL_107",
119
+ "108": "LABEL_108",
120
+ "109": "LABEL_109",
121
+ "110": "LABEL_110",
122
+ "111": "LABEL_111",
123
+ "112": "LABEL_112",
124
+ "113": "LABEL_113",
125
+ "114": "LABEL_114",
126
+ "115": "LABEL_115",
127
+ "116": "LABEL_116",
128
+ "117": "LABEL_117",
129
+ "118": "LABEL_118",
130
+ "119": "LABEL_119",
131
+ "120": "LABEL_120",
132
+ "121": "LABEL_121",
133
+ "122": "LABEL_122",
134
+ "123": "LABEL_123",
135
+ "124": "LABEL_124",
136
+ "125": "LABEL_125",
137
+ "126": "LABEL_126",
138
+ "127": "LABEL_127",
139
+ "128": "LABEL_128",
140
+ "129": "LABEL_129",
141
+ "130": "LABEL_130",
142
+ "131": "LABEL_131",
143
+ "132": "LABEL_132",
144
+ "133": "LABEL_133",
145
+ "134": "LABEL_134",
146
+ "135": "LABEL_135",
147
+ "136": "LABEL_136",
148
+ "137": "LABEL_137",
149
+ "138": "LABEL_138",
150
+ "139": "LABEL_139",
151
+ "140": "LABEL_140",
152
+ "141": "LABEL_141",
153
+ "142": "LABEL_142",
154
+ "143": "LABEL_143",
155
+ "144": "LABEL_144",
156
+ "145": "LABEL_145",
157
+ "146": "LABEL_146",
158
+ "147": "LABEL_147",
159
+ "148": "LABEL_148",
160
+ "149": "LABEL_149",
161
+ "150": "LABEL_150",
162
+ "151": "LABEL_151",
163
+ "152": "LABEL_152",
164
+ "153": "LABEL_153",
165
+ "154": "LABEL_154",
166
+ "155": "LABEL_155",
167
+ "156": "LABEL_156",
168
+ "157": "LABEL_157",
169
+ "158": "LABEL_158",
170
+ "159": "LABEL_159",
171
+ "160": "LABEL_160",
172
+ "161": "LABEL_161",
173
+ "162": "LABEL_162",
174
+ "163": "LABEL_163",
175
+ "164": "LABEL_164",
176
+ "165": "LABEL_165",
177
+ "166": "LABEL_166",
178
+ "167": "LABEL_167",
179
+ "168": "LABEL_168",
180
+ "169": "LABEL_169",
181
+ "170": "LABEL_170",
182
+ "171": "LABEL_171",
183
+ "172": "LABEL_172",
184
+ "173": "LABEL_173",
185
+ "174": "LABEL_174",
186
+ "175": "LABEL_175",
187
+ "176": "LABEL_176",
188
+ "177": "LABEL_177",
189
+ "178": "LABEL_178",
190
+ "179": "LABEL_179",
191
+ "180": "LABEL_180",
192
+ "181": "LABEL_181",
193
+ "182": "LABEL_182",
194
+ "183": "LABEL_183",
195
+ "184": "LABEL_184",
196
+ "185": "LABEL_185",
197
+ "186": "LABEL_186",
198
+ "187": "LABEL_187",
199
+ "188": "LABEL_188",
200
+ "189": "LABEL_189",
201
+ "190": "LABEL_190",
202
+ "191": "LABEL_191",
203
+ "192": "LABEL_192",
204
+ "193": "LABEL_193",
205
+ "194": "LABEL_194",
206
+ "195": "LABEL_195",
207
+ "196": "LABEL_196",
208
+ "197": "LABEL_197",
209
+ "198": "LABEL_198",
210
+ "199": "LABEL_199",
211
+ "200": "LABEL_200",
212
+ "201": "LABEL_201",
213
+ "202": "LABEL_202",
214
+ "203": "LABEL_203",
215
+ "204": "LABEL_204",
216
+ "205": "LABEL_205",
217
+ "206": "LABEL_206",
218
+ "207": "LABEL_207",
219
+ "208": "LABEL_208",
220
+ "209": "LABEL_209",
221
+ "210": "LABEL_210",
222
+ "211": "LABEL_211",
223
+ "212": "LABEL_212",
224
+ "213": "LABEL_213",
225
+ "214": "LABEL_214",
226
+ "215": "LABEL_215",
227
+ "216": "LABEL_216",
228
+ "217": "LABEL_217",
229
+ "218": "LABEL_218",
230
+ "219": "LABEL_219",
231
+ "220": "LABEL_220",
232
+ "221": "LABEL_221",
233
+ "222": "LABEL_222",
234
+ "223": "LABEL_223",
235
+ "224": "LABEL_224",
236
+ "225": "LABEL_225",
237
+ "226": "LABEL_226",
238
+ "227": "LABEL_227",
239
+ "228": "LABEL_228",
240
+ "229": "LABEL_229",
241
+ "230": "LABEL_230",
242
+ "231": "LABEL_231",
243
+ "232": "LABEL_232",
244
+ "233": "LABEL_233",
245
+ "234": "LABEL_234",
246
+ "235": "LABEL_235",
247
+ "236": "LABEL_236",
248
+ "237": "LABEL_237",
249
+ "238": "LABEL_238",
250
+ "239": "LABEL_239",
251
+ "240": "LABEL_240",
252
+ "241": "LABEL_241",
253
+ "242": "LABEL_242",
254
+ "243": "LABEL_243",
255
+ "244": "LABEL_244",
256
+ "245": "LABEL_245",
257
+ "246": "LABEL_246",
258
+ "247": "LABEL_247",
259
+ "248": "LABEL_248",
260
+ "249": "LABEL_249",
261
+ "250": "LABEL_250",
262
+ "251": "LABEL_251",
263
+ "252": "LABEL_252",
264
+ "253": "LABEL_253",
265
+ "254": "LABEL_254",
266
+ "255": "LABEL_255",
267
+ "256": "LABEL_256",
268
+ "257": "LABEL_257",
269
+ "258": "LABEL_258",
270
+ "259": "LABEL_259",
271
+ "260": "LABEL_260",
272
+ "261": "LABEL_261",
273
+ "262": "LABEL_262",
274
+ "263": "LABEL_263",
275
+ "264": "LABEL_264",
276
+ "265": "LABEL_265",
277
+ "266": "LABEL_266",
278
+ "267": "LABEL_267",
279
+ "268": "LABEL_268",
280
+ "269": "LABEL_269",
281
+ "270": "LABEL_270",
282
+ "271": "LABEL_271",
283
+ "272": "LABEL_272",
284
+ "273": "LABEL_273",
285
+ "274": "LABEL_274",
286
+ "275": "LABEL_275",
287
+ "276": "LABEL_276",
288
+ "277": "LABEL_277",
289
+ "278": "LABEL_278",
290
+ "279": "LABEL_279",
291
+ "280": "LABEL_280",
292
+ "281": "LABEL_281",
293
+ "282": "LABEL_282",
294
+ "283": "LABEL_283",
295
+ "284": "LABEL_284",
296
+ "285": "LABEL_285",
297
+ "286": "LABEL_286",
298
+ "287": "LABEL_287",
299
+ "288": "LABEL_288",
300
+ "289": "LABEL_289",
301
+ "290": "LABEL_290",
302
+ "291": "LABEL_291",
303
+ "292": "LABEL_292",
304
+ "293": "LABEL_293",
305
+ "294": "LABEL_294",
306
+ "295": "LABEL_295",
307
+ "296": "LABEL_296",
308
+ "297": "LABEL_297",
309
+ "298": "LABEL_298",
310
+ "299": "LABEL_299",
311
+ "300": "LABEL_300",
312
+ "301": "LABEL_301",
313
+ "302": "LABEL_302",
314
+ "303": "LABEL_303",
315
+ "304": "LABEL_304",
316
+ "305": "LABEL_305",
317
+ "306": "LABEL_306",
318
+ "307": "LABEL_307",
319
+ "308": "LABEL_308",
320
+ "309": "LABEL_309",
321
+ "310": "LABEL_310",
322
+ "311": "LABEL_311",
323
+ "312": "LABEL_312",
324
+ "313": "LABEL_313",
325
+ "314": "LABEL_314",
326
+ "315": "LABEL_315",
327
+ "316": "LABEL_316",
328
+ "317": "LABEL_317",
329
+ "318": "LABEL_318",
330
+ "319": "LABEL_319",
331
+ "320": "LABEL_320",
332
+ "321": "LABEL_321",
333
+ "322": "LABEL_322",
334
+ "323": "LABEL_323",
335
+ "324": "LABEL_324",
336
+ "325": "LABEL_325",
337
+ "326": "LABEL_326",
338
+ "327": "LABEL_327",
339
+ "328": "LABEL_328",
340
+ "329": "LABEL_329",
341
+ "330": "LABEL_330",
342
+ "331": "LABEL_331",
343
+ "332": "LABEL_332",
344
+ "333": "LABEL_333",
345
+ "334": "LABEL_334",
346
+ "335": "LABEL_335",
347
+ "336": "LABEL_336",
348
+ "337": "LABEL_337",
349
+ "338": "LABEL_338",
350
+ "339": "LABEL_339",
351
+ "340": "LABEL_340",
352
+ "341": "LABEL_341",
353
+ "342": "LABEL_342",
354
+ "343": "LABEL_343",
355
+ "344": "LABEL_344",
356
+ "345": "LABEL_345",
357
+ "346": "LABEL_346",
358
+ "347": "LABEL_347",
359
+ "348": "LABEL_348",
360
+ "349": "LABEL_349",
361
+ "350": "LABEL_350",
362
+ "351": "LABEL_351",
363
+ "352": "LABEL_352",
364
+ "353": "LABEL_353",
365
+ "354": "LABEL_354",
366
+ "355": "LABEL_355",
367
+ "356": "LABEL_356",
368
+ "357": "LABEL_357",
369
+ "358": "LABEL_358",
370
+ "359": "LABEL_359",
371
+ "360": "LABEL_360",
372
+ "361": "LABEL_361",
373
+ "362": "LABEL_362",
374
+ "363": "LABEL_363",
375
+ "364": "LABEL_364",
376
+ "365": "LABEL_365",
377
+ "366": "LABEL_366",
378
+ "367": "LABEL_367",
379
+ "368": "LABEL_368",
380
+ "369": "LABEL_369",
381
+ "370": "LABEL_370",
382
+ "371": "LABEL_371",
383
+ "372": "LABEL_372",
384
+ "373": "LABEL_373",
385
+ "374": "LABEL_374",
386
+ "375": "LABEL_375",
387
+ "376": "LABEL_376",
388
+ "377": "LABEL_377",
389
+ "378": "LABEL_378",
390
+ "379": "LABEL_379",
391
+ "380": "LABEL_380",
392
+ "381": "LABEL_381",
393
+ "382": "LABEL_382",
394
+ "383": "LABEL_383",
395
+ "384": "LABEL_384",
396
+ "385": "LABEL_385",
397
+ "386": "LABEL_386",
398
+ "387": "LABEL_387",
399
+ "388": "LABEL_388",
400
+ "389": "LABEL_389",
401
+ "390": "LABEL_390",
402
+ "391": "LABEL_391",
403
+ "392": "LABEL_392",
404
+ "393": "LABEL_393",
405
+ "394": "LABEL_394",
406
+ "395": "LABEL_395",
407
+ "396": "LABEL_396",
408
+ "397": "LABEL_397",
409
+ "398": "LABEL_398",
410
+ "399": "LABEL_399",
411
+ "400": "LABEL_400",
412
+ "401": "LABEL_401",
413
+ "402": "LABEL_402",
414
+ "403": "LABEL_403",
415
+ "404": "LABEL_404",
416
+ "405": "LABEL_405",
417
+ "406": "LABEL_406",
418
+ "407": "LABEL_407",
419
+ "408": "LABEL_408",
420
+ "409": "LABEL_409",
421
+ "410": "LABEL_410",
422
+ "411": "LABEL_411",
423
+ "412": "LABEL_412",
424
+ "413": "LABEL_413",
425
+ "414": "LABEL_414",
426
+ "415": "LABEL_415",
427
+ "416": "LABEL_416",
428
+ "417": "LABEL_417",
429
+ "418": "LABEL_418",
430
+ "419": "LABEL_419",
431
+ "420": "LABEL_420",
432
+ "421": "LABEL_421",
433
+ "422": "LABEL_422",
434
+ "423": "LABEL_423",
435
+ "424": "LABEL_424",
436
+ "425": "LABEL_425",
437
+ "426": "LABEL_426",
438
+ "427": "LABEL_427",
439
+ "428": "LABEL_428",
440
+ "429": "LABEL_429",
441
+ "430": "LABEL_430",
442
+ "431": "LABEL_431",
443
+ "432": "LABEL_432",
444
+ "433": "LABEL_433",
445
+ "434": "LABEL_434",
446
+ "435": "LABEL_435",
447
+ "436": "LABEL_436",
448
+ "437": "LABEL_437",
449
+ "438": "LABEL_438",
450
+ "439": "LABEL_439",
451
+ "440": "LABEL_440",
452
+ "441": "LABEL_441",
453
+ "442": "LABEL_442",
454
+ "443": "LABEL_443",
455
+ "444": "LABEL_444",
456
+ "445": "LABEL_445",
457
+ "446": "LABEL_446",
458
+ "447": "LABEL_447",
459
+ "448": "LABEL_448",
460
+ "449": "LABEL_449",
461
+ "450": "LABEL_450",
462
+ "451": "LABEL_451",
463
+ "452": "LABEL_452",
464
+ "453": "LABEL_453",
465
+ "454": "LABEL_454",
466
+ "455": "LABEL_455",
467
+ "456": "LABEL_456",
468
+ "457": "LABEL_457",
469
+ "458": "LABEL_458",
470
+ "459": "LABEL_459",
471
+ "460": "LABEL_460",
472
+ "461": "LABEL_461",
473
+ "462": "LABEL_462",
474
+ "463": "LABEL_463",
475
+ "464": "LABEL_464",
476
+ "465": "LABEL_465",
477
+ "466": "LABEL_466",
478
+ "467": "LABEL_467",
479
+ "468": "LABEL_468",
480
+ "469": "LABEL_469",
481
+ "470": "LABEL_470",
482
+ "471": "LABEL_471",
483
+ "472": "LABEL_472",
484
+ "473": "LABEL_473",
485
+ "474": "LABEL_474",
486
+ "475": "LABEL_475",
487
+ "476": "LABEL_476",
488
+ "477": "LABEL_477",
489
+ "478": "LABEL_478",
490
+ "479": "LABEL_479",
491
+ "480": "LABEL_480",
492
+ "481": "LABEL_481",
493
+ "482": "LABEL_482",
494
+ "483": "LABEL_483",
495
+ "484": "LABEL_484",
496
+ "485": "LABEL_485",
497
+ "486": "LABEL_486",
498
+ "487": "LABEL_487",
499
+ "488": "LABEL_488",
500
+ "489": "LABEL_489",
501
+ "490": "LABEL_490",
502
+ "491": "LABEL_491",
503
+ "492": "LABEL_492",
504
+ "493": "LABEL_493",
505
+ "494": "LABEL_494",
506
+ "495": "LABEL_495",
507
+ "496": "LABEL_496",
508
+ "497": "LABEL_497",
509
+ "498": "LABEL_498",
510
+ "499": "LABEL_499",
511
+ "500": "LABEL_500",
512
+ "501": "LABEL_501",
513
+ "502": "LABEL_502",
514
+ "503": "LABEL_503",
515
+ "504": "LABEL_504",
516
+ "505": "LABEL_505",
517
+ "506": "LABEL_506",
518
+ "507": "LABEL_507",
519
+ "508": "LABEL_508",
520
+ "509": "LABEL_509",
521
+ "510": "LABEL_510",
522
+ "511": "LABEL_511",
523
+ "512": "LABEL_512",
524
+ "513": "LABEL_513",
525
+ "514": "LABEL_514",
526
+ "515": "LABEL_515",
527
+ "516": "LABEL_516",
528
+ "517": "LABEL_517",
529
+ "518": "LABEL_518",
530
+ "519": "LABEL_519",
531
+ "520": "LABEL_520",
532
+ "521": "LABEL_521",
533
+ "522": "LABEL_522",
534
+ "523": "LABEL_523",
535
+ "524": "LABEL_524",
536
+ "525": "LABEL_525",
537
+ "526": "LABEL_526",
538
+ "527": "LABEL_527",
539
+ "528": "LABEL_528",
540
+ "529": "LABEL_529",
541
+ "530": "LABEL_530",
542
+ "531": "LABEL_531",
543
+ "532": "LABEL_532",
544
+ "533": "LABEL_533",
545
+ "534": "LABEL_534",
546
+ "535": "LABEL_535",
547
+ "536": "LABEL_536",
548
+ "537": "LABEL_537",
549
+ "538": "LABEL_538",
550
+ "539": "LABEL_539",
551
+ "540": "LABEL_540",
552
+ "541": "LABEL_541",
553
+ "542": "LABEL_542",
554
+ "543": "LABEL_543",
555
+ "544": "LABEL_544",
556
+ "545": "LABEL_545",
557
+ "546": "LABEL_546",
558
+ "547": "LABEL_547",
559
+ "548": "LABEL_548",
560
+ "549": "LABEL_549",
561
+ "550": "LABEL_550",
562
+ "551": "LABEL_551",
563
+ "552": "LABEL_552",
564
+ "553": "LABEL_553",
565
+ "554": "LABEL_554",
566
+ "555": "LABEL_555",
567
+ "556": "LABEL_556",
568
+ "557": "LABEL_557",
569
+ "558": "LABEL_558",
570
+ "559": "LABEL_559",
571
+ "560": "LABEL_560",
572
+ "561": "LABEL_561",
573
+ "562": "LABEL_562",
574
+ "563": "LABEL_563",
575
+ "564": "LABEL_564",
576
+ "565": "LABEL_565",
577
+ "566": "LABEL_566",
578
+ "567": "LABEL_567",
579
+ "568": "LABEL_568",
580
+ "569": "LABEL_569",
581
+ "570": "LABEL_570",
582
+ "571": "LABEL_571",
583
+ "572": "LABEL_572",
584
+ "573": "LABEL_573",
585
+ "574": "LABEL_574",
586
+ "575": "LABEL_575",
587
+ "576": "LABEL_576",
588
+ "577": "LABEL_577",
589
+ "578": "LABEL_578",
590
+ "579": "LABEL_579",
591
+ "580": "LABEL_580",
592
+ "581": "LABEL_581",
593
+ "582": "LABEL_582",
594
+ "583": "LABEL_583",
595
+ "584": "LABEL_584",
596
+ "585": "LABEL_585",
597
+ "586": "LABEL_586",
598
+ "587": "LABEL_587",
599
+ "588": "LABEL_588",
600
+ "589": "LABEL_589",
601
+ "590": "LABEL_590",
602
+ "591": "LABEL_591",
603
+ "592": "LABEL_592",
604
+ "593": "LABEL_593",
605
+ "594": "LABEL_594",
606
+ "595": "LABEL_595",
607
+ "596": "LABEL_596",
608
+ "597": "LABEL_597",
609
+ "598": "LABEL_598",
610
+ "599": "LABEL_599",
611
+ "600": "LABEL_600",
612
+ "601": "LABEL_601",
613
+ "602": "LABEL_602",
614
+ "603": "LABEL_603",
615
+ "604": "LABEL_604",
616
+ "605": "LABEL_605",
617
+ "606": "LABEL_606",
618
+ "607": "LABEL_607",
619
+ "608": "LABEL_608",
620
+ "609": "LABEL_609",
621
+ "610": "LABEL_610",
622
+ "611": "LABEL_611",
623
+ "612": "LABEL_612",
624
+ "613": "LABEL_613",
625
+ "614": "LABEL_614",
626
+ "615": "LABEL_615",
627
+ "616": "LABEL_616",
628
+ "617": "LABEL_617",
629
+ "618": "LABEL_618",
630
+ "619": "LABEL_619",
631
+ "620": "LABEL_620",
632
+ "621": "LABEL_621",
633
+ "622": "LABEL_622",
634
+ "623": "LABEL_623",
635
+ "624": "LABEL_624",
636
+ "625": "LABEL_625",
637
+ "626": "LABEL_626",
638
+ "627": "LABEL_627",
639
+ "628": "LABEL_628",
640
+ "629": "LABEL_629",
641
+ "630": "LABEL_630",
642
+ "631": "LABEL_631",
643
+ "632": "LABEL_632",
644
+ "633": "LABEL_633",
645
+ "634": "LABEL_634",
646
+ "635": "LABEL_635",
647
+ "636": "LABEL_636",
648
+ "637": "LABEL_637",
649
+ "638": "LABEL_638",
650
+ "639": "LABEL_639",
651
+ "640": "LABEL_640",
652
+ "641": "LABEL_641",
653
+ "642": "LABEL_642",
654
+ "643": "LABEL_643",
655
+ "644": "LABEL_644",
656
+ "645": "LABEL_645",
657
+ "646": "LABEL_646",
658
+ "647": "LABEL_647",
659
+ "648": "LABEL_648",
660
+ "649": "LABEL_649",
661
+ "650": "LABEL_650",
662
+ "651": "LABEL_651",
663
+ "652": "LABEL_652",
664
+ "653": "LABEL_653",
665
+ "654": "LABEL_654",
666
+ "655": "LABEL_655",
667
+ "656": "LABEL_656",
668
+ "657": "LABEL_657",
669
+ "658": "LABEL_658",
670
+ "659": "LABEL_659",
671
+ "660": "LABEL_660",
672
+ "661": "LABEL_661",
673
+ "662": "LABEL_662",
674
+ "663": "LABEL_663",
675
+ "664": "LABEL_664",
676
+ "665": "LABEL_665",
677
+ "666": "LABEL_666",
678
+ "667": "LABEL_667",
679
+ "668": "LABEL_668",
680
+ "669": "LABEL_669",
681
+ "670": "LABEL_670",
682
+ "671": "LABEL_671",
683
+ "672": "LABEL_672",
684
+ "673": "LABEL_673",
685
+ "674": "LABEL_674",
686
+ "675": "LABEL_675",
687
+ "676": "LABEL_676",
688
+ "677": "LABEL_677",
689
+ "678": "LABEL_678",
690
+ "679": "LABEL_679",
691
+ "680": "LABEL_680",
692
+ "681": "LABEL_681",
693
+ "682": "LABEL_682",
694
+ "683": "LABEL_683",
695
+ "684": "LABEL_684",
696
+ "685": "LABEL_685",
697
+ "686": "LABEL_686",
698
+ "687": "LABEL_687",
699
+ "688": "LABEL_688",
700
+ "689": "LABEL_689",
701
+ "690": "LABEL_690",
702
+ "691": "LABEL_691",
703
+ "692": "LABEL_692",
704
+ "693": "LABEL_693",
705
+ "694": "LABEL_694",
706
+ "695": "LABEL_695",
707
+ "696": "LABEL_696",
708
+ "697": "LABEL_697",
709
+ "698": "LABEL_698"
710
+ },
711
+ "initializer_range": 0.02,
712
+ "intermediate_size": 14336,
713
+ "label2id": {
714
+ "LABEL_0": 0,
715
+ "LABEL_1": 1,
716
+ "LABEL_10": 10,
717
+ "LABEL_100": 100,
718
+ "LABEL_101": 101,
719
+ "LABEL_102": 102,
720
+ "LABEL_103": 103,
721
+ "LABEL_104": 104,
722
+ "LABEL_105": 105,
723
+ "LABEL_106": 106,
724
+ "LABEL_107": 107,
725
+ "LABEL_108": 108,
726
+ "LABEL_109": 109,
727
+ "LABEL_11": 11,
728
+ "LABEL_110": 110,
729
+ "LABEL_111": 111,
730
+ "LABEL_112": 112,
731
+ "LABEL_113": 113,
732
+ "LABEL_114": 114,
733
+ "LABEL_115": 115,
734
+ "LABEL_116": 116,
735
+ "LABEL_117": 117,
736
+ "LABEL_118": 118,
737
+ "LABEL_119": 119,
738
+ "LABEL_12": 12,
739
+ "LABEL_120": 120,
740
+ "LABEL_121": 121,
741
+ "LABEL_122": 122,
742
+ "LABEL_123": 123,
743
+ "LABEL_124": 124,
744
+ "LABEL_125": 125,
745
+ "LABEL_126": 126,
746
+ "LABEL_127": 127,
747
+ "LABEL_128": 128,
748
+ "LABEL_129": 129,
749
+ "LABEL_13": 13,
750
+ "LABEL_130": 130,
751
+ "LABEL_131": 131,
752
+ "LABEL_132": 132,
753
+ "LABEL_133": 133,
754
+ "LABEL_134": 134,
755
+ "LABEL_135": 135,
756
+ "LABEL_136": 136,
757
+ "LABEL_137": 137,
758
+ "LABEL_138": 138,
759
+ "LABEL_139": 139,
760
+ "LABEL_14": 14,
761
+ "LABEL_140": 140,
762
+ "LABEL_141": 141,
763
+ "LABEL_142": 142,
764
+ "LABEL_143": 143,
765
+ "LABEL_144": 144,
766
+ "LABEL_145": 145,
767
+ "LABEL_146": 146,
768
+ "LABEL_147": 147,
769
+ "LABEL_148": 148,
770
+ "LABEL_149": 149,
771
+ "LABEL_15": 15,
772
+ "LABEL_150": 150,
773
+ "LABEL_151": 151,
774
+ "LABEL_152": 152,
775
+ "LABEL_153": 153,
776
+ "LABEL_154": 154,
777
+ "LABEL_155": 155,
778
+ "LABEL_156": 156,
779
+ "LABEL_157": 157,
780
+ "LABEL_158": 158,
781
+ "LABEL_159": 159,
782
+ "LABEL_16": 16,
783
+ "LABEL_160": 160,
784
+ "LABEL_161": 161,
785
+ "LABEL_162": 162,
786
+ "LABEL_163": 163,
787
+ "LABEL_164": 164,
788
+ "LABEL_165": 165,
789
+ "LABEL_166": 166,
790
+ "LABEL_167": 167,
791
+ "LABEL_168": 168,
792
+ "LABEL_169": 169,
793
+ "LABEL_17": 17,
794
+ "LABEL_170": 170,
795
+ "LABEL_171": 171,
796
+ "LABEL_172": 172,
797
+ "LABEL_173": 173,
798
+ "LABEL_174": 174,
799
+ "LABEL_175": 175,
800
+ "LABEL_176": 176,
801
+ "LABEL_177": 177,
802
+ "LABEL_178": 178,
803
+ "LABEL_179": 179,
804
+ "LABEL_18": 18,
805
+ "LABEL_180": 180,
806
+ "LABEL_181": 181,
807
+ "LABEL_182": 182,
808
+ "LABEL_183": 183,
809
+ "LABEL_184": 184,
810
+ "LABEL_185": 185,
811
+ "LABEL_186": 186,
812
+ "LABEL_187": 187,
813
+ "LABEL_188": 188,
814
+ "LABEL_189": 189,
815
+ "LABEL_19": 19,
816
+ "LABEL_190": 190,
817
+ "LABEL_191": 191,
818
+ "LABEL_192": 192,
819
+ "LABEL_193": 193,
820
+ "LABEL_194": 194,
821
+ "LABEL_195": 195,
822
+ "LABEL_196": 196,
823
+ "LABEL_197": 197,
824
+ "LABEL_198": 198,
825
+ "LABEL_199": 199,
826
+ "LABEL_2": 2,
827
+ "LABEL_20": 20,
828
+ "LABEL_200": 200,
829
+ "LABEL_201": 201,
830
+ "LABEL_202": 202,
831
+ "LABEL_203": 203,
832
+ "LABEL_204": 204,
833
+ "LABEL_205": 205,
834
+ "LABEL_206": 206,
835
+ "LABEL_207": 207,
836
+ "LABEL_208": 208,
837
+ "LABEL_209": 209,
838
+ "LABEL_21": 21,
839
+ "LABEL_210": 210,
840
+ "LABEL_211": 211,
841
+ "LABEL_212": 212,
842
+ "LABEL_213": 213,
843
+ "LABEL_214": 214,
844
+ "LABEL_215": 215,
845
+ "LABEL_216": 216,
846
+ "LABEL_217": 217,
847
+ "LABEL_218": 218,
848
+ "LABEL_219": 219,
849
+ "LABEL_22": 22,
850
+ "LABEL_220": 220,
851
+ "LABEL_221": 221,
852
+ "LABEL_222": 222,
853
+ "LABEL_223": 223,
854
+ "LABEL_224": 224,
855
+ "LABEL_225": 225,
856
+ "LABEL_226": 226,
857
+ "LABEL_227": 227,
858
+ "LABEL_228": 228,
859
+ "LABEL_229": 229,
860
+ "LABEL_23": 23,
861
+ "LABEL_230": 230,
862
+ "LABEL_231": 231,
863
+ "LABEL_232": 232,
864
+ "LABEL_233": 233,
865
+ "LABEL_234": 234,
866
+ "LABEL_235": 235,
867
+ "LABEL_236": 236,
868
+ "LABEL_237": 237,
869
+ "LABEL_238": 238,
870
+ "LABEL_239": 239,
871
+ "LABEL_24": 24,
872
+ "LABEL_240": 240,
873
+ "LABEL_241": 241,
874
+ "LABEL_242": 242,
875
+ "LABEL_243": 243,
876
+ "LABEL_244": 244,
877
+ "LABEL_245": 245,
878
+ "LABEL_246": 246,
879
+ "LABEL_247": 247,
880
+ "LABEL_248": 248,
881
+ "LABEL_249": 249,
882
+ "LABEL_25": 25,
883
+ "LABEL_250": 250,
884
+ "LABEL_251": 251,
885
+ "LABEL_252": 252,
886
+ "LABEL_253": 253,
887
+ "LABEL_254": 254,
888
+ "LABEL_255": 255,
889
+ "LABEL_256": 256,
890
+ "LABEL_257": 257,
891
+ "LABEL_258": 258,
892
+ "LABEL_259": 259,
893
+ "LABEL_26": 26,
894
+ "LABEL_260": 260,
895
+ "LABEL_261": 261,
896
+ "LABEL_262": 262,
897
+ "LABEL_263": 263,
898
+ "LABEL_264": 264,
899
+ "LABEL_265": 265,
900
+ "LABEL_266": 266,
901
+ "LABEL_267": 267,
902
+ "LABEL_268": 268,
903
+ "LABEL_269": 269,
904
+ "LABEL_27": 27,
905
+ "LABEL_270": 270,
906
+ "LABEL_271": 271,
907
+ "LABEL_272": 272,
908
+ "LABEL_273": 273,
909
+ "LABEL_274": 274,
910
+ "LABEL_275": 275,
911
+ "LABEL_276": 276,
912
+ "LABEL_277": 277,
913
+ "LABEL_278": 278,
914
+ "LABEL_279": 279,
915
+ "LABEL_28": 28,
916
+ "LABEL_280": 280,
917
+ "LABEL_281": 281,
918
+ "LABEL_282": 282,
919
+ "LABEL_283": 283,
920
+ "LABEL_284": 284,
921
+ "LABEL_285": 285,
922
+ "LABEL_286": 286,
923
+ "LABEL_287": 287,
924
+ "LABEL_288": 288,
925
+ "LABEL_289": 289,
926
+ "LABEL_29": 29,
927
+ "LABEL_290": 290,
928
+ "LABEL_291": 291,
929
+ "LABEL_292": 292,
930
+ "LABEL_293": 293,
931
+ "LABEL_294": 294,
932
+ "LABEL_295": 295,
933
+ "LABEL_296": 296,
934
+ "LABEL_297": 297,
935
+ "LABEL_298": 298,
936
+ "LABEL_299": 299,
937
+ "LABEL_3": 3,
938
+ "LABEL_30": 30,
939
+ "LABEL_300": 300,
940
+ "LABEL_301": 301,
941
+ "LABEL_302": 302,
942
+ "LABEL_303": 303,
943
+ "LABEL_304": 304,
944
+ "LABEL_305": 305,
945
+ "LABEL_306": 306,
946
+ "LABEL_307": 307,
947
+ "LABEL_308": 308,
948
+ "LABEL_309": 309,
949
+ "LABEL_31": 31,
950
+ "LABEL_310": 310,
951
+ "LABEL_311": 311,
952
+ "LABEL_312": 312,
953
+ "LABEL_313": 313,
954
+ "LABEL_314": 314,
955
+ "LABEL_315": 315,
956
+ "LABEL_316": 316,
957
+ "LABEL_317": 317,
958
+ "LABEL_318": 318,
959
+ "LABEL_319": 319,
960
+ "LABEL_32": 32,
961
+ "LABEL_320": 320,
962
+ "LABEL_321": 321,
963
+ "LABEL_322": 322,
964
+ "LABEL_323": 323,
965
+ "LABEL_324": 324,
966
+ "LABEL_325": 325,
967
+ "LABEL_326": 326,
968
+ "LABEL_327": 327,
969
+ "LABEL_328": 328,
970
+ "LABEL_329": 329,
971
+ "LABEL_33": 33,
972
+ "LABEL_330": 330,
973
+ "LABEL_331": 331,
974
+ "LABEL_332": 332,
975
+ "LABEL_333": 333,
976
+ "LABEL_334": 334,
977
+ "LABEL_335": 335,
978
+ "LABEL_336": 336,
979
+ "LABEL_337": 337,
980
+ "LABEL_338": 338,
981
+ "LABEL_339": 339,
982
+ "LABEL_34": 34,
983
+ "LABEL_340": 340,
984
+ "LABEL_341": 341,
985
+ "LABEL_342": 342,
986
+ "LABEL_343": 343,
987
+ "LABEL_344": 344,
988
+ "LABEL_345": 345,
989
+ "LABEL_346": 346,
990
+ "LABEL_347": 347,
991
+ "LABEL_348": 348,
992
+ "LABEL_349": 349,
993
+ "LABEL_35": 35,
994
+ "LABEL_350": 350,
995
+ "LABEL_351": 351,
996
+ "LABEL_352": 352,
997
+ "LABEL_353": 353,
998
+ "LABEL_354": 354,
999
+ "LABEL_355": 355,
1000
+ "LABEL_356": 356,
1001
+ "LABEL_357": 357,
1002
+ "LABEL_358": 358,
1003
+ "LABEL_359": 359,
1004
+ "LABEL_36": 36,
1005
+ "LABEL_360": 360,
1006
+ "LABEL_361": 361,
1007
+ "LABEL_362": 362,
1008
+ "LABEL_363": 363,
1009
+ "LABEL_364": 364,
1010
+ "LABEL_365": 365,
1011
+ "LABEL_366": 366,
1012
+ "LABEL_367": 367,
1013
+ "LABEL_368": 368,
1014
+ "LABEL_369": 369,
1015
+ "LABEL_37": 37,
1016
+ "LABEL_370": 370,
1017
+ "LABEL_371": 371,
1018
+ "LABEL_372": 372,
1019
+ "LABEL_373": 373,
1020
+ "LABEL_374": 374,
1021
+ "LABEL_375": 375,
1022
+ "LABEL_376": 376,
1023
+ "LABEL_377": 377,
1024
+ "LABEL_378": 378,
1025
+ "LABEL_379": 379,
1026
+ "LABEL_38": 38,
1027
+ "LABEL_380": 380,
1028
+ "LABEL_381": 381,
1029
+ "LABEL_382": 382,
1030
+ "LABEL_383": 383,
1031
+ "LABEL_384": 384,
1032
+ "LABEL_385": 385,
1033
+ "LABEL_386": 386,
1034
+ "LABEL_387": 387,
1035
+ "LABEL_388": 388,
1036
+ "LABEL_389": 389,
1037
+ "LABEL_39": 39,
1038
+ "LABEL_390": 390,
1039
+ "LABEL_391": 391,
1040
+ "LABEL_392": 392,
1041
+ "LABEL_393": 393,
1042
+ "LABEL_394": 394,
1043
+ "LABEL_395": 395,
1044
+ "LABEL_396": 396,
1045
+ "LABEL_397": 397,
1046
+ "LABEL_398": 398,
1047
+ "LABEL_399": 399,
1048
+ "LABEL_4": 4,
1049
+ "LABEL_40": 40,
1050
+ "LABEL_400": 400,
1051
+ "LABEL_401": 401,
1052
+ "LABEL_402": 402,
1053
+ "LABEL_403": 403,
1054
+ "LABEL_404": 404,
1055
+ "LABEL_405": 405,
1056
+ "LABEL_406": 406,
1057
+ "LABEL_407": 407,
1058
+ "LABEL_408": 408,
1059
+ "LABEL_409": 409,
1060
+ "LABEL_41": 41,
1061
+ "LABEL_410": 410,
1062
+ "LABEL_411": 411,
1063
+ "LABEL_412": 412,
1064
+ "LABEL_413": 413,
1065
+ "LABEL_414": 414,
1066
+ "LABEL_415": 415,
1067
+ "LABEL_416": 416,
1068
+ "LABEL_417": 417,
1069
+ "LABEL_418": 418,
1070
+ "LABEL_419": 419,
1071
+ "LABEL_42": 42,
1072
+ "LABEL_420": 420,
1073
+ "LABEL_421": 421,
1074
+ "LABEL_422": 422,
1075
+ "LABEL_423": 423,
1076
+ "LABEL_424": 424,
1077
+ "LABEL_425": 425,
1078
+ "LABEL_426": 426,
1079
+ "LABEL_427": 427,
1080
+ "LABEL_428": 428,
1081
+ "LABEL_429": 429,
1082
+ "LABEL_43": 43,
1083
+ "LABEL_430": 430,
1084
+ "LABEL_431": 431,
1085
+ "LABEL_432": 432,
1086
+ "LABEL_433": 433,
1087
+ "LABEL_434": 434,
1088
+ "LABEL_435": 435,
1089
+ "LABEL_436": 436,
1090
+ "LABEL_437": 437,
1091
+ "LABEL_438": 438,
1092
+ "LABEL_439": 439,
1093
+ "LABEL_44": 44,
1094
+ "LABEL_440": 440,
1095
+ "LABEL_441": 441,
1096
+ "LABEL_442": 442,
1097
+ "LABEL_443": 443,
1098
+ "LABEL_444": 444,
1099
+ "LABEL_445": 445,
1100
+ "LABEL_446": 446,
1101
+ "LABEL_447": 447,
1102
+ "LABEL_448": 448,
1103
+ "LABEL_449": 449,
1104
+ "LABEL_45": 45,
1105
+ "LABEL_450": 450,
1106
+ "LABEL_451": 451,
1107
+ "LABEL_452": 452,
1108
+ "LABEL_453": 453,
1109
+ "LABEL_454": 454,
1110
+ "LABEL_455": 455,
1111
+ "LABEL_456": 456,
1112
+ "LABEL_457": 457,
1113
+ "LABEL_458": 458,
1114
+ "LABEL_459": 459,
1115
+ "LABEL_46": 46,
1116
+ "LABEL_460": 460,
1117
+ "LABEL_461": 461,
1118
+ "LABEL_462": 462,
1119
+ "LABEL_463": 463,
1120
+ "LABEL_464": 464,
1121
+ "LABEL_465": 465,
1122
+ "LABEL_466": 466,
1123
+ "LABEL_467": 467,
1124
+ "LABEL_468": 468,
1125
+ "LABEL_469": 469,
1126
+ "LABEL_47": 47,
1127
+ "LABEL_470": 470,
1128
+ "LABEL_471": 471,
1129
+ "LABEL_472": 472,
1130
+ "LABEL_473": 473,
1131
+ "LABEL_474": 474,
1132
+ "LABEL_475": 475,
1133
+ "LABEL_476": 476,
1134
+ "LABEL_477": 477,
1135
+ "LABEL_478": 478,
1136
+ "LABEL_479": 479,
1137
+ "LABEL_48": 48,
1138
+ "LABEL_480": 480,
1139
+ "LABEL_481": 481,
1140
+ "LABEL_482": 482,
1141
+ "LABEL_483": 483,
1142
+ "LABEL_484": 484,
1143
+ "LABEL_485": 485,
1144
+ "LABEL_486": 486,
1145
+ "LABEL_487": 487,
1146
+ "LABEL_488": 488,
1147
+ "LABEL_489": 489,
1148
+ "LABEL_49": 49,
1149
+ "LABEL_490": 490,
1150
+ "LABEL_491": 491,
1151
+ "LABEL_492": 492,
1152
+ "LABEL_493": 493,
1153
+ "LABEL_494": 494,
1154
+ "LABEL_495": 495,
1155
+ "LABEL_496": 496,
1156
+ "LABEL_497": 497,
1157
+ "LABEL_498": 498,
1158
+ "LABEL_499": 499,
1159
+ "LABEL_5": 5,
1160
+ "LABEL_50": 50,
1161
+ "LABEL_500": 500,
1162
+ "LABEL_501": 501,
1163
+ "LABEL_502": 502,
1164
+ "LABEL_503": 503,
1165
+ "LABEL_504": 504,
1166
+ "LABEL_505": 505,
1167
+ "LABEL_506": 506,
1168
+ "LABEL_507": 507,
1169
+ "LABEL_508": 508,
1170
+ "LABEL_509": 509,
1171
+ "LABEL_51": 51,
1172
+ "LABEL_510": 510,
1173
+ "LABEL_511": 511,
1174
+ "LABEL_512": 512,
1175
+ "LABEL_513": 513,
1176
+ "LABEL_514": 514,
1177
+ "LABEL_515": 515,
1178
+ "LABEL_516": 516,
1179
+ "LABEL_517": 517,
1180
+ "LABEL_518": 518,
1181
+ "LABEL_519": 519,
1182
+ "LABEL_52": 52,
1183
+ "LABEL_520": 520,
1184
+ "LABEL_521": 521,
1185
+ "LABEL_522": 522,
1186
+ "LABEL_523": 523,
1187
+ "LABEL_524": 524,
1188
+ "LABEL_525": 525,
1189
+ "LABEL_526": 526,
1190
+ "LABEL_527": 527,
1191
+ "LABEL_528": 528,
1192
+ "LABEL_529": 529,
1193
+ "LABEL_53": 53,
1194
+ "LABEL_530": 530,
1195
+ "LABEL_531": 531,
1196
+ "LABEL_532": 532,
1197
+ "LABEL_533": 533,
1198
+ "LABEL_534": 534,
1199
+ "LABEL_535": 535,
1200
+ "LABEL_536": 536,
1201
+ "LABEL_537": 537,
1202
+ "LABEL_538": 538,
1203
+ "LABEL_539": 539,
1204
+ "LABEL_54": 54,
1205
+ "LABEL_540": 540,
1206
+ "LABEL_541": 541,
1207
+ "LABEL_542": 542,
1208
+ "LABEL_543": 543,
1209
+ "LABEL_544": 544,
1210
+ "LABEL_545": 545,
1211
+ "LABEL_546": 546,
1212
+ "LABEL_547": 547,
1213
+ "LABEL_548": 548,
1214
+ "LABEL_549": 549,
1215
+ "LABEL_55": 55,
1216
+ "LABEL_550": 550,
1217
+ "LABEL_551": 551,
1218
+ "LABEL_552": 552,
1219
+ "LABEL_553": 553,
1220
+ "LABEL_554": 554,
1221
+ "LABEL_555": 555,
1222
+ "LABEL_556": 556,
1223
+ "LABEL_557": 557,
1224
+ "LABEL_558": 558,
1225
+ "LABEL_559": 559,
1226
+ "LABEL_56": 56,
1227
+ "LABEL_560": 560,
1228
+ "LABEL_561": 561,
1229
+ "LABEL_562": 562,
1230
+ "LABEL_563": 563,
1231
+ "LABEL_564": 564,
1232
+ "LABEL_565": 565,
1233
+ "LABEL_566": 566,
1234
+ "LABEL_567": 567,
1235
+ "LABEL_568": 568,
1236
+ "LABEL_569": 569,
1237
+ "LABEL_57": 57,
1238
+ "LABEL_570": 570,
1239
+ "LABEL_571": 571,
1240
+ "LABEL_572": 572,
1241
+ "LABEL_573": 573,
1242
+ "LABEL_574": 574,
1243
+ "LABEL_575": 575,
1244
+ "LABEL_576": 576,
1245
+ "LABEL_577": 577,
1246
+ "LABEL_578": 578,
1247
+ "LABEL_579": 579,
1248
+ "LABEL_58": 58,
1249
+ "LABEL_580": 580,
1250
+ "LABEL_581": 581,
1251
+ "LABEL_582": 582,
1252
+ "LABEL_583": 583,
1253
+ "LABEL_584": 584,
1254
+ "LABEL_585": 585,
1255
+ "LABEL_586": 586,
1256
+ "LABEL_587": 587,
1257
+ "LABEL_588": 588,
1258
+ "LABEL_589": 589,
1259
+ "LABEL_59": 59,
1260
+ "LABEL_590": 590,
1261
+ "LABEL_591": 591,
1262
+ "LABEL_592": 592,
1263
+ "LABEL_593": 593,
1264
+ "LABEL_594": 594,
1265
+ "LABEL_595": 595,
1266
+ "LABEL_596": 596,
1267
+ "LABEL_597": 597,
1268
+ "LABEL_598": 598,
1269
+ "LABEL_599": 599,
1270
+ "LABEL_6": 6,
1271
+ "LABEL_60": 60,
1272
+ "LABEL_600": 600,
1273
+ "LABEL_601": 601,
1274
+ "LABEL_602": 602,
1275
+ "LABEL_603": 603,
1276
+ "LABEL_604": 604,
1277
+ "LABEL_605": 605,
1278
+ "LABEL_606": 606,
1279
+ "LABEL_607": 607,
1280
+ "LABEL_608": 608,
1281
+ "LABEL_609": 609,
1282
+ "LABEL_61": 61,
1283
+ "LABEL_610": 610,
1284
+ "LABEL_611": 611,
1285
+ "LABEL_612": 612,
1286
+ "LABEL_613": 613,
1287
+ "LABEL_614": 614,
1288
+ "LABEL_615": 615,
1289
+ "LABEL_616": 616,
1290
+ "LABEL_617": 617,
1291
+ "LABEL_618": 618,
1292
+ "LABEL_619": 619,
1293
+ "LABEL_62": 62,
1294
+ "LABEL_620": 620,
1295
+ "LABEL_621": 621,
1296
+ "LABEL_622": 622,
1297
+ "LABEL_623": 623,
1298
+ "LABEL_624": 624,
1299
+ "LABEL_625": 625,
1300
+ "LABEL_626": 626,
1301
+ "LABEL_627": 627,
1302
+ "LABEL_628": 628,
1303
+ "LABEL_629": 629,
1304
+ "LABEL_63": 63,
1305
+ "LABEL_630": 630,
1306
+ "LABEL_631": 631,
1307
+ "LABEL_632": 632,
1308
+ "LABEL_633": 633,
1309
+ "LABEL_634": 634,
1310
+ "LABEL_635": 635,
1311
+ "LABEL_636": 636,
1312
+ "LABEL_637": 637,
1313
+ "LABEL_638": 638,
1314
+ "LABEL_639": 639,
1315
+ "LABEL_64": 64,
1316
+ "LABEL_640": 640,
1317
+ "LABEL_641": 641,
1318
+ "LABEL_642": 642,
1319
+ "LABEL_643": 643,
1320
+ "LABEL_644": 644,
1321
+ "LABEL_645": 645,
1322
+ "LABEL_646": 646,
1323
+ "LABEL_647": 647,
1324
+ "LABEL_648": 648,
1325
+ "LABEL_649": 649,
1326
+ "LABEL_65": 65,
1327
+ "LABEL_650": 650,
1328
+ "LABEL_651": 651,
1329
+ "LABEL_652": 652,
1330
+ "LABEL_653": 653,
1331
+ "LABEL_654": 654,
1332
+ "LABEL_655": 655,
1333
+ "LABEL_656": 656,
1334
+ "LABEL_657": 657,
1335
+ "LABEL_658": 658,
1336
+ "LABEL_659": 659,
1337
+ "LABEL_66": 66,
1338
+ "LABEL_660": 660,
1339
+ "LABEL_661": 661,
1340
+ "LABEL_662": 662,
1341
+ "LABEL_663": 663,
1342
+ "LABEL_664": 664,
1343
+ "LABEL_665": 665,
1344
+ "LABEL_666": 666,
1345
+ "LABEL_667": 667,
1346
+ "LABEL_668": 668,
1347
+ "LABEL_669": 669,
1348
+ "LABEL_67": 67,
1349
+ "LABEL_670": 670,
1350
+ "LABEL_671": 671,
1351
+ "LABEL_672": 672,
1352
+ "LABEL_673": 673,
1353
+ "LABEL_674": 674,
1354
+ "LABEL_675": 675,
1355
+ "LABEL_676": 676,
1356
+ "LABEL_677": 677,
1357
+ "LABEL_678": 678,
1358
+ "LABEL_679": 679,
1359
+ "LABEL_68": 68,
1360
+ "LABEL_680": 680,
1361
+ "LABEL_681": 681,
1362
+ "LABEL_682": 682,
1363
+ "LABEL_683": 683,
1364
+ "LABEL_684": 684,
1365
+ "LABEL_685": 685,
1366
+ "LABEL_686": 686,
1367
+ "LABEL_687": 687,
1368
+ "LABEL_688": 688,
1369
+ "LABEL_689": 689,
1370
+ "LABEL_69": 69,
1371
+ "LABEL_690": 690,
1372
+ "LABEL_691": 691,
1373
+ "LABEL_692": 692,
1374
+ "LABEL_693": 693,
1375
+ "LABEL_694": 694,
1376
+ "LABEL_695": 695,
1377
+ "LABEL_696": 696,
1378
+ "LABEL_697": 697,
1379
+ "LABEL_698": 698,
1380
+ "LABEL_7": 7,
1381
+ "LABEL_70": 70,
1382
+ "LABEL_71": 71,
1383
+ "LABEL_72": 72,
1384
+ "LABEL_73": 73,
1385
+ "LABEL_74": 74,
1386
+ "LABEL_75": 75,
1387
+ "LABEL_76": 76,
1388
+ "LABEL_77": 77,
1389
+ "LABEL_78": 78,
1390
+ "LABEL_79": 79,
1391
+ "LABEL_8": 8,
1392
+ "LABEL_80": 80,
1393
+ "LABEL_81": 81,
1394
+ "LABEL_82": 82,
1395
+ "LABEL_83": 83,
1396
+ "LABEL_84": 84,
1397
+ "LABEL_85": 85,
1398
+ "LABEL_86": 86,
1399
+ "LABEL_87": 87,
1400
+ "LABEL_88": 88,
1401
+ "LABEL_89": 89,
1402
+ "LABEL_9": 9,
1403
+ "LABEL_90": 90,
1404
+ "LABEL_91": 91,
1405
+ "LABEL_92": 92,
1406
+ "LABEL_93": 93,
1407
+ "LABEL_94": 94,
1408
+ "LABEL_95": 95,
1409
+ "LABEL_96": 96,
1410
+ "LABEL_97": 97,
1411
+ "LABEL_98": 98,
1412
+ "LABEL_99": 99
1413
+ },
1414
+ "max_position_embeddings": 32768,
1415
+ "model_type": "mistral",
1416
+ "num_attention_heads": 32,
1417
+ "num_hidden_layers": 32,
1418
+ "num_key_value_heads": 8,
1419
+ "pad_token_id": 2,
1420
+ "problem_type": "single_label_classification",
1421
+ "rms_norm_eps": 1e-05,
1422
+ "rope_theta": 10000.0,
1423
+ "sliding_window": 4096,
1424
+ "tie_word_embeddings": false,
1425
+ "torch_dtype": "float32",
1426
+ "transformers_version": "4.35.0",
1427
+ "use_cache": true,
1428
+ "vocab_size": 32000
1429
+ }
label_encoder.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f722a50a38d241c1e408476147c393c6c56a9a53f81655f7857f3c11b5249cd
3
+ size 5837
model-00001-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d81b746ca2fb720249f91d6fb77c7ec096752361f3b74fb970864d4c6f6e21d
3
+ size 4987196936
model-00002-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a31076c6bd6a0c94d82d584f14f1b0ec3281a1f761d00be0cc1c7843b31e465
3
+ size 4899116440
model-00003-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44548338690c611e74b2ada3e446eb129e4288e21f7f49dc027c587b4a8e4ce6
3
+ size 4999813120
model-00004-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f835b2e44070adab914660ae36467df76761183e223311cddb497481a1f2649
3
+ size 4999813128
model-00005-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8120a5834bc9b7fc84616de6ba540729f9c4c0093d495a230d6cbe2ad0ff23db
3
+ size 4832007496
model-00006-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8073929958a20d9156ef25f9df1c39ba30cb0530df9bd542d39fd9f64967f3c2
3
+ size 3736179296
model.safetensors.index.json ADDED
@@ -0,0 +1,298 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 28454092800
4
+ },
5
+ "weight_map": {
6
+ "model.embed_tokens.weight": "model-00001-of-00006.safetensors",
7
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00006.safetensors",
8
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00006.safetensors",
9
+ "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00006.safetensors",
10
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00006.safetensors",
11
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00006.safetensors",
12
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00006.safetensors",
13
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00006.safetensors",
14
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00006.safetensors",
15
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00006.safetensors",
16
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00006.safetensors",
17
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00006.safetensors",
18
+ "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00006.safetensors",
19
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00006.safetensors",
20
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00006.safetensors",
21
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00006.safetensors",
22
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00006.safetensors",
23
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00006.safetensors",
24
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00006.safetensors",
25
+ "model.layers.10.input_layernorm.weight": "model-00003-of-00006.safetensors",
26
+ "model.layers.10.mlp.down_proj.weight": "model-00003-of-00006.safetensors",
27
+ "model.layers.10.mlp.gate_proj.weight": "model-00002-of-00006.safetensors",
28
+ "model.layers.10.mlp.up_proj.weight": "model-00002-of-00006.safetensors",
29
+ "model.layers.10.post_attention_layernorm.weight": "model-00003-of-00006.safetensors",
30
+ "model.layers.10.self_attn.k_proj.weight": "model-00002-of-00006.safetensors",
31
+ "model.layers.10.self_attn.o_proj.weight": "model-00002-of-00006.safetensors",
32
+ "model.layers.10.self_attn.q_proj.weight": "model-00002-of-00006.safetensors",
33
+ "model.layers.10.self_attn.v_proj.weight": "model-00002-of-00006.safetensors",
34
+ "model.layers.11.input_layernorm.weight": "model-00003-of-00006.safetensors",
35
+ "model.layers.11.mlp.down_proj.weight": "model-00003-of-00006.safetensors",
36
+ "model.layers.11.mlp.gate_proj.weight": "model-00003-of-00006.safetensors",
37
+ "model.layers.11.mlp.up_proj.weight": "model-00003-of-00006.safetensors",
38
+ "model.layers.11.post_attention_layernorm.weight": "model-00003-of-00006.safetensors",
39
+ "model.layers.11.self_attn.k_proj.weight": "model-00003-of-00006.safetensors",
40
+ "model.layers.11.self_attn.o_proj.weight": "model-00003-of-00006.safetensors",
41
+ "model.layers.11.self_attn.q_proj.weight": "model-00003-of-00006.safetensors",
42
+ "model.layers.11.self_attn.v_proj.weight": "model-00003-of-00006.safetensors",
43
+ "model.layers.12.input_layernorm.weight": "model-00003-of-00006.safetensors",
44
+ "model.layers.12.mlp.down_proj.weight": "model-00003-of-00006.safetensors",
45
+ "model.layers.12.mlp.gate_proj.weight": "model-00003-of-00006.safetensors",
46
+ "model.layers.12.mlp.up_proj.weight": "model-00003-of-00006.safetensors",
47
+ "model.layers.12.post_attention_layernorm.weight": "model-00003-of-00006.safetensors",
48
+ "model.layers.12.self_attn.k_proj.weight": "model-00003-of-00006.safetensors",
49
+ "model.layers.12.self_attn.o_proj.weight": "model-00003-of-00006.safetensors",
50
+ "model.layers.12.self_attn.q_proj.weight": "model-00003-of-00006.safetensors",
51
+ "model.layers.12.self_attn.v_proj.weight": "model-00003-of-00006.safetensors",
52
+ "model.layers.13.input_layernorm.weight": "model-00003-of-00006.safetensors",
53
+ "model.layers.13.mlp.down_proj.weight": "model-00003-of-00006.safetensors",
54
+ "model.layers.13.mlp.gate_proj.weight": "model-00003-of-00006.safetensors",
55
+ "model.layers.13.mlp.up_proj.weight": "model-00003-of-00006.safetensors",
56
+ "model.layers.13.post_attention_layernorm.weight": "model-00003-of-00006.safetensors",
57
+ "model.layers.13.self_attn.k_proj.weight": "model-00003-of-00006.safetensors",
58
+ "model.layers.13.self_attn.o_proj.weight": "model-00003-of-00006.safetensors",
59
+ "model.layers.13.self_attn.q_proj.weight": "model-00003-of-00006.safetensors",
60
+ "model.layers.13.self_attn.v_proj.weight": "model-00003-of-00006.safetensors",
61
+ "model.layers.14.input_layernorm.weight": "model-00003-of-00006.safetensors",
62
+ "model.layers.14.mlp.down_proj.weight": "model-00003-of-00006.safetensors",
63
+ "model.layers.14.mlp.gate_proj.weight": "model-00003-of-00006.safetensors",
64
+ "model.layers.14.mlp.up_proj.weight": "model-00003-of-00006.safetensors",
65
+ "model.layers.14.post_attention_layernorm.weight": "model-00003-of-00006.safetensors",
66
+ "model.layers.14.self_attn.k_proj.weight": "model-00003-of-00006.safetensors",
67
+ "model.layers.14.self_attn.o_proj.weight": "model-00003-of-00006.safetensors",
68
+ "model.layers.14.self_attn.q_proj.weight": "model-00003-of-00006.safetensors",
69
+ "model.layers.14.self_attn.v_proj.weight": "model-00003-of-00006.safetensors",
70
+ "model.layers.15.input_layernorm.weight": "model-00003-of-00006.safetensors",
71
+ "model.layers.15.mlp.down_proj.weight": "model-00003-of-00006.safetensors",
72
+ "model.layers.15.mlp.gate_proj.weight": "model-00003-of-00006.safetensors",
73
+ "model.layers.15.mlp.up_proj.weight": "model-00003-of-00006.safetensors",
74
+ "model.layers.15.post_attention_layernorm.weight": "model-00003-of-00006.safetensors",
75
+ "model.layers.15.self_attn.k_proj.weight": "model-00003-of-00006.safetensors",
76
+ "model.layers.15.self_attn.o_proj.weight": "model-00003-of-00006.safetensors",
77
+ "model.layers.15.self_attn.q_proj.weight": "model-00003-of-00006.safetensors",
78
+ "model.layers.15.self_attn.v_proj.weight": "model-00003-of-00006.safetensors",
79
+ "model.layers.16.input_layernorm.weight": "model-00004-of-00006.safetensors",
80
+ "model.layers.16.mlp.down_proj.weight": "model-00004-of-00006.safetensors",
81
+ "model.layers.16.mlp.gate_proj.weight": "model-00003-of-00006.safetensors",
82
+ "model.layers.16.mlp.up_proj.weight": "model-00004-of-00006.safetensors",
83
+ "model.layers.16.post_attention_layernorm.weight": "model-00004-of-00006.safetensors",
84
+ "model.layers.16.self_attn.k_proj.weight": "model-00003-of-00006.safetensors",
85
+ "model.layers.16.self_attn.o_proj.weight": "model-00003-of-00006.safetensors",
86
+ "model.layers.16.self_attn.q_proj.weight": "model-00003-of-00006.safetensors",
87
+ "model.layers.16.self_attn.v_proj.weight": "model-00003-of-00006.safetensors",
88
+ "model.layers.17.input_layernorm.weight": "model-00004-of-00006.safetensors",
89
+ "model.layers.17.mlp.down_proj.weight": "model-00004-of-00006.safetensors",
90
+ "model.layers.17.mlp.gate_proj.weight": "model-00004-of-00006.safetensors",
91
+ "model.layers.17.mlp.up_proj.weight": "model-00004-of-00006.safetensors",
92
+ "model.layers.17.post_attention_layernorm.weight": "model-00004-of-00006.safetensors",
93
+ "model.layers.17.self_attn.k_proj.weight": "model-00004-of-00006.safetensors",
94
+ "model.layers.17.self_attn.o_proj.weight": "model-00004-of-00006.safetensors",
95
+ "model.layers.17.self_attn.q_proj.weight": "model-00004-of-00006.safetensors",
96
+ "model.layers.17.self_attn.v_proj.weight": "model-00004-of-00006.safetensors",
97
+ "model.layers.18.input_layernorm.weight": "model-00004-of-00006.safetensors",
98
+ "model.layers.18.mlp.down_proj.weight": "model-00004-of-00006.safetensors",
99
+ "model.layers.18.mlp.gate_proj.weight": "model-00004-of-00006.safetensors",
100
+ "model.layers.18.mlp.up_proj.weight": "model-00004-of-00006.safetensors",
101
+ "model.layers.18.post_attention_layernorm.weight": "model-00004-of-00006.safetensors",
102
+ "model.layers.18.self_attn.k_proj.weight": "model-00004-of-00006.safetensors",
103
+ "model.layers.18.self_attn.o_proj.weight": "model-00004-of-00006.safetensors",
104
+ "model.layers.18.self_attn.q_proj.weight": "model-00004-of-00006.safetensors",
105
+ "model.layers.18.self_attn.v_proj.weight": "model-00004-of-00006.safetensors",
106
+ "model.layers.19.input_layernorm.weight": "model-00004-of-00006.safetensors",
107
+ "model.layers.19.mlp.down_proj.weight": "model-00004-of-00006.safetensors",
108
+ "model.layers.19.mlp.gate_proj.weight": "model-00004-of-00006.safetensors",
109
+ "model.layers.19.mlp.up_proj.weight": "model-00004-of-00006.safetensors",
110
+ "model.layers.19.post_attention_layernorm.weight": "model-00004-of-00006.safetensors",
111
+ "model.layers.19.self_attn.k_proj.weight": "model-00004-of-00006.safetensors",
112
+ "model.layers.19.self_attn.o_proj.weight": "model-00004-of-00006.safetensors",
113
+ "model.layers.19.self_attn.q_proj.weight": "model-00004-of-00006.safetensors",
114
+ "model.layers.19.self_attn.v_proj.weight": "model-00004-of-00006.safetensors",
115
+ "model.layers.2.input_layernorm.weight": "model-00001-of-00006.safetensors",
116
+ "model.layers.2.mlp.down_proj.weight": "model-00001-of-00006.safetensors",
117
+ "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00006.safetensors",
118
+ "model.layers.2.mlp.up_proj.weight": "model-00001-of-00006.safetensors",
119
+ "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00006.safetensors",
120
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00006.safetensors",
121
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00006.safetensors",
122
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00006.safetensors",
123
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00006.safetensors",
124
+ "model.layers.20.input_layernorm.weight": "model-00004-of-00006.safetensors",
125
+ "model.layers.20.mlp.down_proj.weight": "model-00004-of-00006.safetensors",
126
+ "model.layers.20.mlp.gate_proj.weight": "model-00004-of-00006.safetensors",
127
+ "model.layers.20.mlp.up_proj.weight": "model-00004-of-00006.safetensors",
128
+ "model.layers.20.post_attention_layernorm.weight": "model-00004-of-00006.safetensors",
129
+ "model.layers.20.self_attn.k_proj.weight": "model-00004-of-00006.safetensors",
130
+ "model.layers.20.self_attn.o_proj.weight": "model-00004-of-00006.safetensors",
131
+ "model.layers.20.self_attn.q_proj.weight": "model-00004-of-00006.safetensors",
132
+ "model.layers.20.self_attn.v_proj.weight": "model-00004-of-00006.safetensors",
133
+ "model.layers.21.input_layernorm.weight": "model-00004-of-00006.safetensors",
134
+ "model.layers.21.mlp.down_proj.weight": "model-00004-of-00006.safetensors",
135
+ "model.layers.21.mlp.gate_proj.weight": "model-00004-of-00006.safetensors",
136
+ "model.layers.21.mlp.up_proj.weight": "model-00004-of-00006.safetensors",
137
+ "model.layers.21.post_attention_layernorm.weight": "model-00004-of-00006.safetensors",
138
+ "model.layers.21.self_attn.k_proj.weight": "model-00004-of-00006.safetensors",
139
+ "model.layers.21.self_attn.o_proj.weight": "model-00004-of-00006.safetensors",
140
+ "model.layers.21.self_attn.q_proj.weight": "model-00004-of-00006.safetensors",
141
+ "model.layers.21.self_attn.v_proj.weight": "model-00004-of-00006.safetensors",
142
+ "model.layers.22.input_layernorm.weight": "model-00005-of-00006.safetensors",
143
+ "model.layers.22.mlp.down_proj.weight": "model-00005-of-00006.safetensors",
144
+ "model.layers.22.mlp.gate_proj.weight": "model-00005-of-00006.safetensors",
145
+ "model.layers.22.mlp.up_proj.weight": "model-00005-of-00006.safetensors",
146
+ "model.layers.22.post_attention_layernorm.weight": "model-00005-of-00006.safetensors",
147
+ "model.layers.22.self_attn.k_proj.weight": "model-00004-of-00006.safetensors",
148
+ "model.layers.22.self_attn.o_proj.weight": "model-00004-of-00006.safetensors",
149
+ "model.layers.22.self_attn.q_proj.weight": "model-00004-of-00006.safetensors",
150
+ "model.layers.22.self_attn.v_proj.weight": "model-00004-of-00006.safetensors",
151
+ "model.layers.23.input_layernorm.weight": "model-00005-of-00006.safetensors",
152
+ "model.layers.23.mlp.down_proj.weight": "model-00005-of-00006.safetensors",
153
+ "model.layers.23.mlp.gate_proj.weight": "model-00005-of-00006.safetensors",
154
+ "model.layers.23.mlp.up_proj.weight": "model-00005-of-00006.safetensors",
155
+ "model.layers.23.post_attention_layernorm.weight": "model-00005-of-00006.safetensors",
156
+ "model.layers.23.self_attn.k_proj.weight": "model-00005-of-00006.safetensors",
157
+ "model.layers.23.self_attn.o_proj.weight": "model-00005-of-00006.safetensors",
158
+ "model.layers.23.self_attn.q_proj.weight": "model-00005-of-00006.safetensors",
159
+ "model.layers.23.self_attn.v_proj.weight": "model-00005-of-00006.safetensors",
160
+ "model.layers.24.input_layernorm.weight": "model-00005-of-00006.safetensors",
161
+ "model.layers.24.mlp.down_proj.weight": "model-00005-of-00006.safetensors",
162
+ "model.layers.24.mlp.gate_proj.weight": "model-00005-of-00006.safetensors",
163
+ "model.layers.24.mlp.up_proj.weight": "model-00005-of-00006.safetensors",
164
+ "model.layers.24.post_attention_layernorm.weight": "model-00005-of-00006.safetensors",
165
+ "model.layers.24.self_attn.k_proj.weight": "model-00005-of-00006.safetensors",
166
+ "model.layers.24.self_attn.o_proj.weight": "model-00005-of-00006.safetensors",
167
+ "model.layers.24.self_attn.q_proj.weight": "model-00005-of-00006.safetensors",
168
+ "model.layers.24.self_attn.v_proj.weight": "model-00005-of-00006.safetensors",
169
+ "model.layers.25.input_layernorm.weight": "model-00005-of-00006.safetensors",
170
+ "model.layers.25.mlp.down_proj.weight": "model-00005-of-00006.safetensors",
171
+ "model.layers.25.mlp.gate_proj.weight": "model-00005-of-00006.safetensors",
172
+ "model.layers.25.mlp.up_proj.weight": "model-00005-of-00006.safetensors",
173
+ "model.layers.25.post_attention_layernorm.weight": "model-00005-of-00006.safetensors",
174
+ "model.layers.25.self_attn.k_proj.weight": "model-00005-of-00006.safetensors",
175
+ "model.layers.25.self_attn.o_proj.weight": "model-00005-of-00006.safetensors",
176
+ "model.layers.25.self_attn.q_proj.weight": "model-00005-of-00006.safetensors",
177
+ "model.layers.25.self_attn.v_proj.weight": "model-00005-of-00006.safetensors",
178
+ "model.layers.26.input_layernorm.weight": "model-00005-of-00006.safetensors",
179
+ "model.layers.26.mlp.down_proj.weight": "model-00005-of-00006.safetensors",
180
+ "model.layers.26.mlp.gate_proj.weight": "model-00005-of-00006.safetensors",
181
+ "model.layers.26.mlp.up_proj.weight": "model-00005-of-00006.safetensors",
182
+ "model.layers.26.post_attention_layernorm.weight": "model-00005-of-00006.safetensors",
183
+ "model.layers.26.self_attn.k_proj.weight": "model-00005-of-00006.safetensors",
184
+ "model.layers.26.self_attn.o_proj.weight": "model-00005-of-00006.safetensors",
185
+ "model.layers.26.self_attn.q_proj.weight": "model-00005-of-00006.safetensors",
186
+ "model.layers.26.self_attn.v_proj.weight": "model-00005-of-00006.safetensors",
187
+ "model.layers.27.input_layernorm.weight": "model-00006-of-00006.safetensors",
188
+ "model.layers.27.mlp.down_proj.weight": "model-00006-of-00006.safetensors",
189
+ "model.layers.27.mlp.gate_proj.weight": "model-00005-of-00006.safetensors",
190
+ "model.layers.27.mlp.up_proj.weight": "model-00005-of-00006.safetensors",
191
+ "model.layers.27.post_attention_layernorm.weight": "model-00006-of-00006.safetensors",
192
+ "model.layers.27.self_attn.k_proj.weight": "model-00005-of-00006.safetensors",
193
+ "model.layers.27.self_attn.o_proj.weight": "model-00005-of-00006.safetensors",
194
+ "model.layers.27.self_attn.q_proj.weight": "model-00005-of-00006.safetensors",
195
+ "model.layers.27.self_attn.v_proj.weight": "model-00005-of-00006.safetensors",
196
+ "model.layers.28.input_layernorm.weight": "model-00006-of-00006.safetensors",
197
+ "model.layers.28.mlp.down_proj.weight": "model-00006-of-00006.safetensors",
198
+ "model.layers.28.mlp.gate_proj.weight": "model-00006-of-00006.safetensors",
199
+ "model.layers.28.mlp.up_proj.weight": "model-00006-of-00006.safetensors",
200
+ "model.layers.28.post_attention_layernorm.weight": "model-00006-of-00006.safetensors",
201
+ "model.layers.28.self_attn.k_proj.weight": "model-00006-of-00006.safetensors",
202
+ "model.layers.28.self_attn.o_proj.weight": "model-00006-of-00006.safetensors",
203
+ "model.layers.28.self_attn.q_proj.weight": "model-00006-of-00006.safetensors",
204
+ "model.layers.28.self_attn.v_proj.weight": "model-00006-of-00006.safetensors",
205
+ "model.layers.29.input_layernorm.weight": "model-00006-of-00006.safetensors",
206
+ "model.layers.29.mlp.down_proj.weight": "model-00006-of-00006.safetensors",
207
+ "model.layers.29.mlp.gate_proj.weight": "model-00006-of-00006.safetensors",
208
+ "model.layers.29.mlp.up_proj.weight": "model-00006-of-00006.safetensors",
209
+ "model.layers.29.post_attention_layernorm.weight": "model-00006-of-00006.safetensors",
210
+ "model.layers.29.self_attn.k_proj.weight": "model-00006-of-00006.safetensors",
211
+ "model.layers.29.self_attn.o_proj.weight": "model-00006-of-00006.safetensors",
212
+ "model.layers.29.self_attn.q_proj.weight": "model-00006-of-00006.safetensors",
213
+ "model.layers.29.self_attn.v_proj.weight": "model-00006-of-00006.safetensors",
214
+ "model.layers.3.input_layernorm.weight": "model-00001-of-00006.safetensors",
215
+ "model.layers.3.mlp.down_proj.weight": "model-00001-of-00006.safetensors",
216
+ "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00006.safetensors",
217
+ "model.layers.3.mlp.up_proj.weight": "model-00001-of-00006.safetensors",
218
+ "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00006.safetensors",
219
+ "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00006.safetensors",
220
+ "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00006.safetensors",
221
+ "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00006.safetensors",
222
+ "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00006.safetensors",
223
+ "model.layers.30.input_layernorm.weight": "model-00006-of-00006.safetensors",
224
+ "model.layers.30.mlp.down_proj.weight": "model-00006-of-00006.safetensors",
225
+ "model.layers.30.mlp.gate_proj.weight": "model-00006-of-00006.safetensors",
226
+ "model.layers.30.mlp.up_proj.weight": "model-00006-of-00006.safetensors",
227
+ "model.layers.30.post_attention_layernorm.weight": "model-00006-of-00006.safetensors",
228
+ "model.layers.30.self_attn.k_proj.weight": "model-00006-of-00006.safetensors",
229
+ "model.layers.30.self_attn.o_proj.weight": "model-00006-of-00006.safetensors",
230
+ "model.layers.30.self_attn.q_proj.weight": "model-00006-of-00006.safetensors",
231
+ "model.layers.30.self_attn.v_proj.weight": "model-00006-of-00006.safetensors",
232
+ "model.layers.31.input_layernorm.weight": "model-00006-of-00006.safetensors",
233
+ "model.layers.31.mlp.down_proj.weight": "model-00006-of-00006.safetensors",
234
+ "model.layers.31.mlp.gate_proj.weight": "model-00006-of-00006.safetensors",
235
+ "model.layers.31.mlp.up_proj.weight": "model-00006-of-00006.safetensors",
236
+ "model.layers.31.post_attention_layernorm.weight": "model-00006-of-00006.safetensors",
237
+ "model.layers.31.self_attn.k_proj.weight": "model-00006-of-00006.safetensors",
238
+ "model.layers.31.self_attn.o_proj.weight": "model-00006-of-00006.safetensors",
239
+ "model.layers.31.self_attn.q_proj.weight": "model-00006-of-00006.safetensors",
240
+ "model.layers.31.self_attn.v_proj.weight": "model-00006-of-00006.safetensors",
241
+ "model.layers.4.input_layernorm.weight": "model-00001-of-00006.safetensors",
242
+ "model.layers.4.mlp.down_proj.weight": "model-00001-of-00006.safetensors",
243
+ "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00006.safetensors",
244
+ "model.layers.4.mlp.up_proj.weight": "model-00001-of-00006.safetensors",
245
+ "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00006.safetensors",
246
+ "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00006.safetensors",
247
+ "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00006.safetensors",
248
+ "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00006.safetensors",
249
+ "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00006.safetensors",
250
+ "model.layers.5.input_layernorm.weight": "model-00002-of-00006.safetensors",
251
+ "model.layers.5.mlp.down_proj.weight": "model-00002-of-00006.safetensors",
252
+ "model.layers.5.mlp.gate_proj.weight": "model-00002-of-00006.safetensors",
253
+ "model.layers.5.mlp.up_proj.weight": "model-00002-of-00006.safetensors",
254
+ "model.layers.5.post_attention_layernorm.weight": "model-00002-of-00006.safetensors",
255
+ "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00006.safetensors",
256
+ "model.layers.5.self_attn.o_proj.weight": "model-00002-of-00006.safetensors",
257
+ "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00006.safetensors",
258
+ "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00006.safetensors",
259
+ "model.layers.6.input_layernorm.weight": "model-00002-of-00006.safetensors",
260
+ "model.layers.6.mlp.down_proj.weight": "model-00002-of-00006.safetensors",
261
+ "model.layers.6.mlp.gate_proj.weight": "model-00002-of-00006.safetensors",
262
+ "model.layers.6.mlp.up_proj.weight": "model-00002-of-00006.safetensors",
263
+ "model.layers.6.post_attention_layernorm.weight": "model-00002-of-00006.safetensors",
264
+ "model.layers.6.self_attn.k_proj.weight": "model-00002-of-00006.safetensors",
265
+ "model.layers.6.self_attn.o_proj.weight": "model-00002-of-00006.safetensors",
266
+ "model.layers.6.self_attn.q_proj.weight": "model-00002-of-00006.safetensors",
267
+ "model.layers.6.self_attn.v_proj.weight": "model-00002-of-00006.safetensors",
268
+ "model.layers.7.input_layernorm.weight": "model-00002-of-00006.safetensors",
269
+ "model.layers.7.mlp.down_proj.weight": "model-00002-of-00006.safetensors",
270
+ "model.layers.7.mlp.gate_proj.weight": "model-00002-of-00006.safetensors",
271
+ "model.layers.7.mlp.up_proj.weight": "model-00002-of-00006.safetensors",
272
+ "model.layers.7.post_attention_layernorm.weight": "model-00002-of-00006.safetensors",
273
+ "model.layers.7.self_attn.k_proj.weight": "model-00002-of-00006.safetensors",
274
+ "model.layers.7.self_attn.o_proj.weight": "model-00002-of-00006.safetensors",
275
+ "model.layers.7.self_attn.q_proj.weight": "model-00002-of-00006.safetensors",
276
+ "model.layers.7.self_attn.v_proj.weight": "model-00002-of-00006.safetensors",
277
+ "model.layers.8.input_layernorm.weight": "model-00002-of-00006.safetensors",
278
+ "model.layers.8.mlp.down_proj.weight": "model-00002-of-00006.safetensors",
279
+ "model.layers.8.mlp.gate_proj.weight": "model-00002-of-00006.safetensors",
280
+ "model.layers.8.mlp.up_proj.weight": "model-00002-of-00006.safetensors",
281
+ "model.layers.8.post_attention_layernorm.weight": "model-00002-of-00006.safetensors",
282
+ "model.layers.8.self_attn.k_proj.weight": "model-00002-of-00006.safetensors",
283
+ "model.layers.8.self_attn.o_proj.weight": "model-00002-of-00006.safetensors",
284
+ "model.layers.8.self_attn.q_proj.weight": "model-00002-of-00006.safetensors",
285
+ "model.layers.8.self_attn.v_proj.weight": "model-00002-of-00006.safetensors",
286
+ "model.layers.9.input_layernorm.weight": "model-00002-of-00006.safetensors",
287
+ "model.layers.9.mlp.down_proj.weight": "model-00002-of-00006.safetensors",
288
+ "model.layers.9.mlp.gate_proj.weight": "model-00002-of-00006.safetensors",
289
+ "model.layers.9.mlp.up_proj.weight": "model-00002-of-00006.safetensors",
290
+ "model.layers.9.post_attention_layernorm.weight": "model-00002-of-00006.safetensors",
291
+ "model.layers.9.self_attn.k_proj.weight": "model-00002-of-00006.safetensors",
292
+ "model.layers.9.self_attn.o_proj.weight": "model-00002-of-00006.safetensors",
293
+ "model.layers.9.self_attn.q_proj.weight": "model-00002-of-00006.safetensors",
294
+ "model.layers.9.self_attn.v_proj.weight": "model-00002-of-00006.safetensors",
295
+ "model.norm.weight": "model-00006-of-00006.safetensors",
296
+ "score.weight": "model-00006-of-00006.safetensors"
297
+ }
298
+ }
predict.py ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # predict.py
2
+
3
+ import torch
4
+ from transformers import AutoTokenizer, AutoModelForSequenceClassification
5
+ import pickle
6
+
7
+ model_path = 'bert-mistral-fine-tuned-model'
8
+ tokenizer = AutoTokenizer.from_pretrained(model_path)
9
+ model = AutoModelForSequenceClassification.from_pretrained(model_path)
10
+
11
+ with open(f'{model_path}/label_encoder.pkl', 'rb') as f:
12
+ label_encoder = pickle.load(f)
13
+
14
+ def predict_class(text):
15
+ inputs = tokenizer(text, return_tensors='pt', truncation=True, max_length=128)
16
+ with torch.no_grad():
17
+ outputs = model(**inputs)
18
+ logits = outputs.logits
19
+ predict_class_id = logits.argmax(-1).item()
20
+ predict_label = label_encoder.inverse_transform([predict_class_id])[0]
21
+ return predict_label
22
+
23
+ if __name__ == "__main__":
24
+ text = "purulent discharge"
25
+ predicted_label = predict_class(text)
26
+ print(predicted_label)
special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "</s>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<unk>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<s>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ }
27
+ },
28
+ "additional_special_tokens": [],
29
+ "bos_token": "<s>",
30
+ "clean_up_tokenization_spaces": false,
31
+ "eos_token": "</s>",
32
+ "legacy": true,
33
+ "model_max_length": 1000000000000000019884624838656,
34
+ "pad_token": "</s>",
35
+ "sp_model_kwargs": {},
36
+ "spaces_between_special_tokens": false,
37
+ "tokenizer_class": "LlamaTokenizer",
38
+ "unk_token": "<unk>",
39
+ "use_default_system_prompt": true
40
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62ae15baee7ac4114036c639d1dd9b95de275dc21d14d32e5e25058449c39023
3
+ size 4536