asahi417 commited on
Commit
fec8163
·
1 Parent(s): 26ee824
Files changed (2) hide show
  1. config.json +423 -0
  2. pytorch_model.bin +3 -0
config.json ADDED
@@ -0,0 +1,423 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "tner_ckpt/bionlp2004_roberta_large/best_model",
3
+ "architectures": [
4
+ "RobertaForTokenClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "crf_state_dict": {
10
+ "_constraint_mask": [
11
+ [
12
+ 1.0,
13
+ 1.0,
14
+ 0.0,
15
+ 1.0,
16
+ 0.0,
17
+ 1.0,
18
+ 0.0,
19
+ 1.0,
20
+ 0.0,
21
+ 1.0,
22
+ 0.0,
23
+ 0.0,
24
+ 1.0
25
+ ],
26
+ [
27
+ 1.0,
28
+ 1.0,
29
+ 1.0,
30
+ 1.0,
31
+ 0.0,
32
+ 1.0,
33
+ 0.0,
34
+ 1.0,
35
+ 0.0,
36
+ 1.0,
37
+ 0.0,
38
+ 0.0,
39
+ 1.0
40
+ ],
41
+ [
42
+ 1.0,
43
+ 1.0,
44
+ 1.0,
45
+ 1.0,
46
+ 0.0,
47
+ 1.0,
48
+ 0.0,
49
+ 1.0,
50
+ 0.0,
51
+ 1.0,
52
+ 0.0,
53
+ 0.0,
54
+ 1.0
55
+ ],
56
+ [
57
+ 1.0,
58
+ 1.0,
59
+ 0.0,
60
+ 1.0,
61
+ 1.0,
62
+ 1.0,
63
+ 0.0,
64
+ 1.0,
65
+ 0.0,
66
+ 1.0,
67
+ 0.0,
68
+ 0.0,
69
+ 1.0
70
+ ],
71
+ [
72
+ 1.0,
73
+ 1.0,
74
+ 0.0,
75
+ 1.0,
76
+ 1.0,
77
+ 1.0,
78
+ 0.0,
79
+ 1.0,
80
+ 0.0,
81
+ 1.0,
82
+ 0.0,
83
+ 0.0,
84
+ 1.0
85
+ ],
86
+ [
87
+ 1.0,
88
+ 1.0,
89
+ 0.0,
90
+ 1.0,
91
+ 0.0,
92
+ 1.0,
93
+ 1.0,
94
+ 1.0,
95
+ 0.0,
96
+ 1.0,
97
+ 0.0,
98
+ 0.0,
99
+ 1.0
100
+ ],
101
+ [
102
+ 1.0,
103
+ 1.0,
104
+ 0.0,
105
+ 1.0,
106
+ 0.0,
107
+ 1.0,
108
+ 1.0,
109
+ 1.0,
110
+ 0.0,
111
+ 1.0,
112
+ 0.0,
113
+ 0.0,
114
+ 1.0
115
+ ],
116
+ [
117
+ 1.0,
118
+ 1.0,
119
+ 0.0,
120
+ 1.0,
121
+ 0.0,
122
+ 1.0,
123
+ 0.0,
124
+ 1.0,
125
+ 1.0,
126
+ 1.0,
127
+ 0.0,
128
+ 0.0,
129
+ 1.0
130
+ ],
131
+ [
132
+ 1.0,
133
+ 1.0,
134
+ 0.0,
135
+ 1.0,
136
+ 0.0,
137
+ 1.0,
138
+ 0.0,
139
+ 1.0,
140
+ 1.0,
141
+ 1.0,
142
+ 0.0,
143
+ 0.0,
144
+ 1.0
145
+ ],
146
+ [
147
+ 1.0,
148
+ 1.0,
149
+ 0.0,
150
+ 1.0,
151
+ 0.0,
152
+ 1.0,
153
+ 0.0,
154
+ 1.0,
155
+ 0.0,
156
+ 1.0,
157
+ 1.0,
158
+ 0.0,
159
+ 1.0
160
+ ],
161
+ [
162
+ 1.0,
163
+ 1.0,
164
+ 0.0,
165
+ 1.0,
166
+ 0.0,
167
+ 1.0,
168
+ 0.0,
169
+ 1.0,
170
+ 0.0,
171
+ 1.0,
172
+ 1.0,
173
+ 0.0,
174
+ 1.0
175
+ ],
176
+ [
177
+ 1.0,
178
+ 1.0,
179
+ 0.0,
180
+ 1.0,
181
+ 0.0,
182
+ 1.0,
183
+ 0.0,
184
+ 1.0,
185
+ 0.0,
186
+ 1.0,
187
+ 0.0,
188
+ 0.0,
189
+ 0.0
190
+ ],
191
+ [
192
+ 0.0,
193
+ 0.0,
194
+ 0.0,
195
+ 0.0,
196
+ 0.0,
197
+ 0.0,
198
+ 0.0,
199
+ 0.0,
200
+ 0.0,
201
+ 0.0,
202
+ 0.0,
203
+ 0.0,
204
+ 0.0
205
+ ]
206
+ ],
207
+ "end_transitions": [
208
+ 1.0725135803222656,
209
+ -0.12956254184246063,
210
+ 0.371111661195755,
211
+ 0.5107485055923462,
212
+ 0.8757023215293884,
213
+ 0.2065323442220688,
214
+ 1.1444202661514282,
215
+ -0.9789606928825378,
216
+ 0.6102772951126099,
217
+ -1.5749095678329468,
218
+ 0.6528273820877075
219
+ ],
220
+ "start_transitions": [
221
+ 0.9727583527565002,
222
+ 0.15879656374454498,
223
+ 2.214200973510742,
224
+ -0.17921259999275208,
225
+ -0.24598051607608795,
226
+ 0.006965472362935543,
227
+ 1.6770257949829102,
228
+ -2.4614217281341553,
229
+ -0.6591065526008606,
230
+ 0.8901280164718628,
231
+ -0.345694363117218
232
+ ],
233
+ "transitions": [
234
+ [
235
+ -0.047160349786281586,
236
+ -0.27636075019836426,
237
+ -0.01827988587319851,
238
+ -0.2668977379798889,
239
+ -0.3786858022212982,
240
+ -0.20064036548137665,
241
+ -0.1423189789056778,
242
+ -0.4146086573600769,
243
+ -0.5416123867034912,
244
+ 0.09183064103126526,
245
+ 0.284502238035202
246
+ ],
247
+ [
248
+ 0.011403710581362247,
249
+ -0.2831323444843292,
250
+ -0.1991034895181656,
251
+ 0.07006354629993439,
252
+ 0.2171705812215805,
253
+ 0.08347953110933304,
254
+ -0.18687021732330322,
255
+ 0.06655106693506241,
256
+ 0.35198599100112915,
257
+ -0.041680119931697845,
258
+ 0.4603417217731476
259
+ ],
260
+ [
261
+ -0.009048350155353546,
262
+ 0.37783774733543396,
263
+ -0.4404703378677368,
264
+ 0.10700219124555588,
265
+ -0.631479799747467,
266
+ -0.10892672091722488,
267
+ -0.07639240473508835,
268
+ -0.007770363707095385,
269
+ 0.5599918365478516,
270
+ 0.055807191878557205,
271
+ 0.24667567014694214
272
+ ],
273
+ [
274
+ -0.28981271386146545,
275
+ -0.5259994864463806,
276
+ -0.23048432171344757,
277
+ 0.4910091161727905,
278
+ 0.12749718129634857,
279
+ 0.16045114398002625,
280
+ 0.2603921890258789,
281
+ -0.12336166203022003,
282
+ -0.08413107693195343,
283
+ -0.5930366516113281,
284
+ -0.34702208638191223
285
+ ],
286
+ [
287
+ -0.2059697061777115,
288
+ -0.05627566576004028,
289
+ 0.14756712317466736,
290
+ 0.0013249005423858762,
291
+ 0.32564589381217957,
292
+ 0.24977774918079376,
293
+ 0.006007824093103409,
294
+ -0.039906613528728485,
295
+ 0.47553882002830505,
296
+ -0.011895392090082169,
297
+ 0.03930443152785301
298
+ ],
299
+ [
300
+ -0.061014119535684586,
301
+ 0.08056101202964783,
302
+ -0.009875665418803692,
303
+ -0.22185459733009338,
304
+ 0.07587382942438126,
305
+ -0.31542763113975525,
306
+ -0.5121344923973083,
307
+ 0.1371414065361023,
308
+ 0.19443896412849426,
309
+ 0.31572508811950684,
310
+ 0.2727954387664795
311
+ ],
312
+ [
313
+ -0.5849944353103638,
314
+ -0.04580551013350487,
315
+ -0.48698607087135315,
316
+ 0.6264642477035522,
317
+ -0.18660670518875122,
318
+ -0.4097630977630615,
319
+ 0.0755583792924881,
320
+ 0.4243757426738739,
321
+ -0.271587997674942,
322
+ -0.6559296250343323,
323
+ 0.4165884852409363
324
+ ],
325
+ [
326
+ -0.4860977828502655,
327
+ -0.08637966960668564,
328
+ 0.011937582865357399,
329
+ 0.33564722537994385,
330
+ 0.14710360765457153,
331
+ -0.013451650738716125,
332
+ 0.21268104016780853,
333
+ 0.06677278131246567,
334
+ -0.1522866040468216,
335
+ 0.15147152543067932,
336
+ -0.12879641354084015
337
+ ],
338
+ [
339
+ 0.26046469807624817,
340
+ 0.5372914671897888,
341
+ -0.018869543448090553,
342
+ -0.13315194845199585,
343
+ 0.6247418522834778,
344
+ -0.10685393214225769,
345
+ 0.21527133882045746,
346
+ 0.13126051425933838,
347
+ 0.03024105355143547,
348
+ 0.07785946130752563,
349
+ 1.02604341506958
350
+ ],
351
+ [
352
+ -0.1257716566324234,
353
+ -0.18844079971313477,
354
+ 0.12964299321174622,
355
+ 0.040601469576358795,
356
+ -0.29614439606666565,
357
+ 0.27116966247558594,
358
+ 0.04347997158765793,
359
+ -0.17520152032375336,
360
+ -0.2326679527759552,
361
+ 0.28523188829421997,
362
+ 0.06743942946195602
363
+ ],
364
+ [
365
+ 0.0764397531747818,
366
+ -0.044914186000823975,
367
+ -0.17368023097515106,
368
+ 0.15744398534297943,
369
+ -0.18209576606750488,
370
+ 0.0481279119849205,
371
+ -0.1274690479040146,
372
+ 0.6103290319442749,
373
+ 0.48212194442749023,
374
+ -0.27468565106391907,
375
+ 0.23631976544857025
376
+ ]
377
+ ]
378
+ },
379
+ "eos_token_id": 2,
380
+ "hidden_act": "gelu",
381
+ "hidden_dropout_prob": 0.1,
382
+ "hidden_size": 1024,
383
+ "id2label": {
384
+ "0": "O",
385
+ "1": "B-DNA",
386
+ "2": "I-DNA",
387
+ "3": "B-protein",
388
+ "4": "I-protein",
389
+ "5": "B-cell_type",
390
+ "6": "I-cell_type",
391
+ "7": "B-cell_line",
392
+ "8": "I-cell_line",
393
+ "9": "B-RNA",
394
+ "10": "I-RNA"
395
+ },
396
+ "initializer_range": 0.02,
397
+ "intermediate_size": 4096,
398
+ "label2id": {
399
+ "B-DNA": 1,
400
+ "B-RNA": 9,
401
+ "B-cell_line": 7,
402
+ "B-cell_type": 5,
403
+ "B-protein": 3,
404
+ "I-DNA": 2,
405
+ "I-RNA": 10,
406
+ "I-cell_line": 8,
407
+ "I-cell_type": 6,
408
+ "I-protein": 4,
409
+ "O": 0
410
+ },
411
+ "layer_norm_eps": 1e-05,
412
+ "max_position_embeddings": 514,
413
+ "model_type": "roberta",
414
+ "num_attention_heads": 16,
415
+ "num_hidden_layers": 24,
416
+ "pad_token_id": 1,
417
+ "position_embedding_type": "absolute",
418
+ "torch_dtype": "float32",
419
+ "transformers_version": "4.20.1",
420
+ "type_vocab_size": 1,
421
+ "use_cache": true,
422
+ "vocab_size": 50265
423
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44b3057a5f8066f6723b94fbe639672389048eca5382b405eeb01bfb19695bee
3
+ size 1417416753