{ "architectures": [ "DebertaV2ForTokenClassification" ], "attention_probs_dropout_prob": 0.1, "bos_token_id": 0, "eos_token_id": 2, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 256, "id2label": { "0": "ADJ", "1": "ADP", "2": "ADP+DET", "3": "ADV", "4": "AUX", "5": "B-ADJ", "6": "B-ADP", "7": "B-ADP+DET", "8": "B-ADV", "9": "B-AUX", "10": "B-CCONJ", "11": "B-DET", "12": "B-INTJ", "13": "B-NOUN", "14": "B-NUM", "15": "B-PART", "16": "B-PRON", "17": "B-PROPN", "18": "B-PUNCT", "19": "B-SCONJ", "20": "B-VERB", "21": "B-X", "22": "CCONJ", "23": "DET", "24": "I-ADJ", "25": "I-ADP", "26": "I-ADP+DET", "27": "I-ADV", "28": "I-AUX", "29": "I-CCONJ", "30": "I-DET", "31": "I-INTJ", "32": "I-NOUN", "33": "I-NUM", "34": "I-PART", "35": "I-PRON", "36": "I-PROPN", "37": "I-PUNCT", "38": "I-SCONJ", "39": "I-VERB", "40": "I-X", "41": "INTJ", "42": "NOUN", "43": "NUM", "44": "PART", "45": "PRON", "46": "PROPN", "47": "PUNCT", "48": "SCONJ", "49": "SYM", "50": "VERB", "51": "X" }, "initializer_range": 0.02, "intermediate_size": 768, "label2id": { "ADJ": 0, "ADP": 1, "ADP+DET": 2, "ADV": 3, "AUX": 4, "B-ADJ": 5, "B-ADP": 6, "B-ADP+DET": 7, "B-ADV": 8, "B-AUX": 9, "B-CCONJ": 10, "B-DET": 11, "B-INTJ": 12, "B-NOUN": 13, "B-NUM": 14, "B-PART": 15, "B-PRON": 16, "B-PROPN": 17, "B-PUNCT": 18, "B-SCONJ": 19, "B-VERB": 20, "B-X": 21, "CCONJ": 22, "DET": 23, "I-ADJ": 24, "I-ADP": 25, "I-ADP+DET": 26, "I-ADV": 27, "I-AUX": 28, "I-CCONJ": 29, "I-DET": 30, "I-INTJ": 31, "I-NOUN": 32, "I-NUM": 33, "I-PART": 34, "I-PRON": 35, "I-PROPN": 36, "I-PUNCT": 37, "I-SCONJ": 38, "I-VERB": 39, "I-X": 40, "INTJ": 41, "NOUN": 42, "NUM": 43, "PART": 44, "PRON": 45, "PROPN": 46, "PUNCT": 47, "SCONJ": 48, "SYM": 49, "VERB": 50, "X": 51 }, "layer_norm_eps": 1e-07, "max_position_embeddings": 128, "max_relative_positions": -1, "model_type": "deberta-v2", "num_attention_heads": 4, "num_hidden_layers": 12, "pad_token_id": 1, "pooler_dropout": 0, "pooler_hidden_act": "gelu", "pooler_hidden_size": 256, "pos_att_type": [ "p2c", "c2p" ], "position_biased_input": false, "relative_attention": true, "task_specific_params": { "upos_multiword": { "ADP+DET": { "Al": [ "a", "lo" ], "Au": [ "a", "lo" ], "Dau": [ "de", "lo" ], "Del": [ "de", "lo" ], "Pel": [ "per", "lo" ], "Pels": [ "per", "los" ], "Peu": [ "per", "lo" ], "Sul": [ "sus", "lo" ], "Suls": [ "sus", "los" ], "ai": [ "a", "lei" ], "ais": [ "a", "leis" ], "al": [ "a", "lo" ], "als": [ "a", "los" ], "ara": [ "a", "era" ], "ath": [ "a", "eth" ], "au": [ "a", "lo" ], "aus": [ "a", "los" ], "dau": [ "de", "lo" ], "daus": [ "de", "los" ], "dei": [ "de", "lei" ], "del": [ "de", "lo" ], "dels": [ "de", "los" ], "dera": [ "de", "era" ], "deras": [ "de", "eras" ], "des": [ "de", "es" ], "deth": [ "de", "eth" ], "deu": [ "de", "lo" ], "deus": [ "de", "los" ], "entau": [ "ent\u00e0", "lo" ], "entaus": [ "ent\u00e0", "los" ], "jol": [ "jos", "lo" ], "jols": [ "jos", "los" ], "pel": [ "per", "lo" ], "pels": [ "per", "los" ], "pera": [ "per", "era" ], "peth": [ "per", "eth" ], "peu": [ "per", "lo" ], "peus": [ "per", "los" ], "sul": [ "sus", "lo" ], "suls": [ "sus", "los" ], "suu": [ "sus", "lo" ], "tath": [ "ta", "eth" ], "tau": [ "ta", "lo" ] } } }, "tokenizer_class": "BertTokenizer", "torch_dtype": "float32", "transformers_version": "4.42.4", "type_vocab_size": 0, "vocab_size": 30000 }