{ "_name_or_path": "xlm-roberta-base", "architectures": [ "XLMRobertaForTokenClassification" ], "attention_probs_dropout_prob": 0.1, "bos_token_id": 0, "classifier_dropout": null, "eos_token_id": 2, "finetuning_task": "pos", "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 768, "id2label": { "0": "`", "1": "\tSYM", "2": "IN", "3": "$", "4": "WDT", "5": "LS", "6": "\tDT", "7": "VBZ", "8": "CD", "9": "SYM", "10": "UH", "11": "VB", "12": "``", "13": "VBN", "14": "NN", "15": "sleepy\t#", "16": "terrorist\t#", "17": "TO", "18": "POS", "19": "it", "20": "NNP", "21": "(", "22": ".", "23": "CC", "24": ":", "25": "RBR", "26": "''", "27": "#", "28": "FW", "29": "MD", "30": "DT", "31": "WP$", "32": ",", "33": "RBS", "34": ")", "35": "RB", "36": "WP", "37": "WRB", "38": "@", "39": "NNS", "40": "PRP$", "41": "JJS", "42": "Ready\t#", "43": "EX", "44": "U", "45": "NNPS", "46": "\tPRP", "47": "hero\t#", "48": "JJR", "49": "JJ", "50": "PRP", "51": "PDT", "52": "VBD", "53": "RP", "54": "VBP", "55": "VBG" }, "initializer_range": 0.02, "intermediate_size": 3072, "label2id": { "\tDT": 6, "\tPRP": 46, "\tSYM": 1, "#": 27, "$": 3, "''": 26, "(": 21, ")": 34, ",": 32, ".": 22, ":": 24, "@": 38, "CC": 23, "CD": 8, "DT": 30, "EX": 43, "FW": 28, "IN": 2, "JJ": 49, "JJR": 48, "JJS": 41, "LS": 5, "MD": 29, "NN": 14, "NNP": 20, "NNPS": 45, "NNS": 39, "PDT": 51, "POS": 18, "PRP": 50, "PRP$": 40, "RB": 35, "RBR": 25, "RBS": 33, "RP": 53, "Ready\t#": 42, "SYM": 9, "TO": 17, "U": 44, "UH": 10, "VB": 11, "VBD": 52, "VBG": 55, "VBN": 13, "VBP": 54, "VBZ": 7, "WDT": 4, "WP": 36, "WP$": 31, "WRB": 37, "`": 0, "``": 12, "hero\t#": 47, "it": 19, "sleepy\t#": 15, "terrorist\t#": 16 }, "layer_norm_eps": 1e-05, "max_position_embeddings": 514, "model_type": "xlm-roberta", "num_attention_heads": 12, "num_hidden_layers": 12, "output_past": true, "pad_token_id": 1, "position_embedding_type": "absolute", "torch_dtype": "float32", "transformers_version": "4.25.1", "type_vocab_size": 1, "use_cache": true, "vocab_size": 250002 }