|
{ |
|
"_name_or_path": "VulBERTa-MLP-MVD", |
|
"architectures": [ |
|
"RobertaForSequenceClassification" |
|
], |
|
"attention_probs_dropout_prob": 0.1, |
|
"bos_token_id": 0, |
|
"classifier_dropout": null, |
|
"eos_token_id": 2, |
|
"gradient_checkpointing": false, |
|
"hidden_act": "gelu", |
|
"hidden_dropout_prob": 0.1, |
|
"hidden_size": 768, |
|
"id2label": { |
|
"0": "non-vulnerable", |
|
"1": "CWE-404", |
|
"2": "CWE-476", |
|
"3": "CWE-119", |
|
"4": "CWE-706", |
|
"5": "CWE-670", |
|
"6": "CWE-673", |
|
"7": "CWE-119, CWE-666, CWE-573", |
|
"8": "CWE-573", |
|
"9": "CWE-668", |
|
"10": "CWE-400, CWE-665, CWE-020", |
|
"11": "CWE-662", |
|
"12": "CWE-400", |
|
"13": "CWE-665", |
|
"14": "CWE-020", |
|
"15": "CWE-074", |
|
"16": "CWE-362", |
|
"17": "CWE-191", |
|
"18": "CWE-190", |
|
"19": "CWE-610", |
|
"20": "CWE-704", |
|
"21": "CWE-170", |
|
"22": "CWE-676", |
|
"23": "CWE-187", |
|
"24": "CWE-138", |
|
"25": "CWE-369", |
|
"26": "CWE-662, CWE-573", |
|
"27": "CWE-834", |
|
"28": "CWE-400, CWE-665", |
|
"29": "CWE-400, CWE-404", |
|
"30": "CWE-221", |
|
"31": "CWE-754", |
|
"32": "CWE-311", |
|
"33": "CWE-404, CWE-668", |
|
"34": "CWE-506", |
|
"35": "CWE-758", |
|
"36": "CWE-666", |
|
"37": "CWE-467", |
|
"38": "CWE-327", |
|
"39": "CWE-666, CWE-573", |
|
"40": "CWE-469" |
|
}, |
|
"initializer_range": 0.02, |
|
"intermediate_size": 3072, |
|
"label2id": { |
|
"CWE-020": 14, |
|
"CWE-074": 15, |
|
"CWE-119": 3, |
|
"CWE-119, CWE-666, CWE-573": 7, |
|
"CWE-138": 24, |
|
"CWE-170": 21, |
|
"CWE-187": 23, |
|
"CWE-190": 18, |
|
"CWE-191": 17, |
|
"CWE-221": 30, |
|
"CWE-311": 32, |
|
"CWE-327": 38, |
|
"CWE-362": 16, |
|
"CWE-369": 25, |
|
"CWE-400": 12, |
|
"CWE-400, CWE-404": 29, |
|
"CWE-400, CWE-665": 28, |
|
"CWE-400, CWE-665, CWE-020": 10, |
|
"CWE-404": 1, |
|
"CWE-404, CWE-668": 33, |
|
"CWE-467": 37, |
|
"CWE-469": 40, |
|
"CWE-476": 2, |
|
"CWE-506": 34, |
|
"CWE-573": 8, |
|
"CWE-610": 19, |
|
"CWE-662": 11, |
|
"CWE-662, CWE-573": 26, |
|
"CWE-665": 13, |
|
"CWE-666": 36, |
|
"CWE-666, CWE-573": 39, |
|
"CWE-668": 9, |
|
"CWE-670": 5, |
|
"CWE-673": 6, |
|
"CWE-676": 22, |
|
"CWE-704": 20, |
|
"CWE-706": 4, |
|
"CWE-754": 31, |
|
"CWE-758": 35, |
|
"CWE-834": 27, |
|
"non-vulnerable": 0 |
|
}, |
|
"layer_norm_eps": 1e-12, |
|
"max_position_embeddings": 1026, |
|
"model_type": "roberta", |
|
"num_attention_heads": 12, |
|
"num_hidden_layers": 12, |
|
"pad_token_id": 1, |
|
"position_embedding_type": "absolute", |
|
"torch_dtype": "float32", |
|
"transformers_version": "4.37.0.dev0", |
|
"type_vocab_size": 1, |
|
"use_cache": true, |
|
"vocab_size": 50000 |
|
} |
|
|