antypasd commited on
Commit
669aae2
·
verified ·
1 Parent(s): aee07bc

Upload RobertaForSequenceClassification

Browse files
Files changed (3) hide show
  1. README.md +3 -3
  2. config.json +11 -19
  3. model.safetensors +2 -2
README.md CHANGED
@@ -1,13 +1,13 @@
1
  ---
2
  language:
3
  - en
4
- widget:
5
- - text: Call me today to earn some money mofos!
6
  datasets:
7
  - cardiffnlp/x_sensitive
8
- license: mit
9
  metrics:
10
  - f1
 
 
11
  pipeline_tag: text-classification
12
  ---
13
 
 
1
  ---
2
  language:
3
  - en
4
+ license: mit
 
5
  datasets:
6
  - cardiffnlp/x_sensitive
 
7
  metrics:
8
  - f1
9
+ widget:
10
+ - text: Call me today to earn some money mofos!
11
  pipeline_tag: text-classification
12
  ---
13
 
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/home/dimosthenisa/Desktop/sensitive_content/code/experiments/mlm_models/troberta-large/multilabel/best_model",
3
  "architectures": [
4
  "RobertaForSequenceClassification"
5
  ],
@@ -9,35 +9,27 @@
9
  "eos_token_id": 2,
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.1,
12
- "hidden_size": 1024,
13
  "id2label": {
14
- "0": "conflictual",
15
- "1": "profanity",
16
- "2": "sex",
17
- "3": "drugs",
18
- "4": "selfharm",
19
- "5": "spam"
20
  },
21
  "initializer_range": 0.02,
22
- "intermediate_size": 4096,
23
  "label2id": {
24
- "conflictual": 0,
25
- "drugs": 3,
26
- "profanity": 1,
27
- "selfharm": 4,
28
- "sex": 2,
29
- "spam": 5
30
  },
31
  "layer_norm_eps": 1e-05,
32
  "max_position_embeddings": 514,
33
  "model_type": "roberta",
34
- "num_attention_heads": 16,
35
- "num_hidden_layers": 24,
36
  "pad_token_id": 1,
37
  "position_embedding_type": "absolute",
38
- "problem_type": "multi_label_classification",
39
  "torch_dtype": "float32",
40
- "transformers_version": "4.41.2",
41
  "type_vocab_size": 1,
42
  "use_cache": true,
43
  "vocab_size": 50265
 
1
  {
2
+ "_name_or_path": "/home/dimosthenisa/Desktop/sensitive_content/code/experiments/mlm_models/troberta-base/binary/best_model",
3
  "architectures": [
4
  "RobertaForSequenceClassification"
5
  ],
 
9
  "eos_token_id": 2,
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 768,
13
  "id2label": {
14
+ "0": "non-sensitive",
15
+ "1": "sensitive"
 
 
 
 
16
  },
17
  "initializer_range": 0.02,
18
+ "intermediate_size": 3072,
19
  "label2id": {
20
+ "non-sensitive": 0,
21
+ "sensitive": 1
 
 
 
 
22
  },
23
  "layer_norm_eps": 1e-05,
24
  "max_position_embeddings": 514,
25
  "model_type": "roberta",
26
+ "num_attention_heads": 12,
27
+ "num_hidden_layers": 12,
28
  "pad_token_id": 1,
29
  "position_embedding_type": "absolute",
30
+ "problem_type": "single_label_classification",
31
  "torch_dtype": "float32",
32
+ "transformers_version": "4.39.3",
33
  "type_vocab_size": 1,
34
  "use_cache": true,
35
  "vocab_size": 50265
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6f5e6d0e4f8e4af5a9a9a75f7f16cd15644ea235e5c7240c97fa650489fea0ec
3
- size 1421511816
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8330c1a6855310d4c770856dd79e20752452f1dda1219d34c8b1ef32861e02f
3
+ size 498612824