stulcrad commited on
Commit
b21982a
·
verified ·
1 Parent(s): 8db68de

End of training

Browse files
README.md CHANGED
@@ -25,16 +25,16 @@ model-index:
25
  metrics:
26
  - name: Precision
27
  type: precision
28
- value: 0.8641078838174274
29
  - name: Recall
30
  type: recall
31
- value: 0.8904329235702833
32
  - name: F1
33
  type: f1
34
- value: 0.877072913924717
35
  - name: Accuracy
36
  type: accuracy
37
- value: 0.9697581443217353
38
  ---
39
 
40
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -44,11 +44,11 @@ should probably proofread and complete it, then remove this comment. -->
44
 
45
  This model is a fine-tuned version of [FacebookAI/xlm-roberta-large](https://huggingface.co/FacebookAI/xlm-roberta-large) on the cnec dataset.
46
  It achieves the following results on the evaluation set:
47
- - Loss: 0.1720
48
- - Precision: 0.8641
49
- - Recall: 0.8904
50
- - F1: 0.8771
51
- - Accuracy: 0.9698
52
 
53
  ## Model description
54
 
@@ -79,16 +79,16 @@ The following hyperparameters were used during training:
79
 
80
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
81
  |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
82
- | 0.4839 | 1.0 | 581 | 0.1812 | 0.7650 | 0.8038 | 0.7839 | 0.9495 |
83
- | 0.2176 | 2.0 | 1162 | 0.1446 | 0.8225 | 0.8370 | 0.8297 | 0.9625 |
84
- | 0.1579 | 3.0 | 1743 | 0.1526 | 0.8164 | 0.8578 | 0.8366 | 0.9645 |
85
- | 0.1391 | 4.0 | 2324 | 0.1362 | 0.8495 | 0.8626 | 0.8560 | 0.9663 |
86
- | 0.1163 | 5.0 | 2905 | 0.1534 | 0.8456 | 0.8749 | 0.8600 | 0.9672 |
87
- | 0.0976 | 6.0 | 3486 | 0.1567 | 0.8454 | 0.8765 | 0.8607 | 0.9653 |
88
- | 0.0725 | 7.0 | 4067 | 0.1629 | 0.8580 | 0.8813 | 0.8695 | 0.9676 |
89
- | 0.0626 | 8.0 | 4648 | 0.1749 | 0.8664 | 0.8904 | 0.8782 | 0.9683 |
90
- | 0.0564 | 9.0 | 5229 | 0.1680 | 0.8649 | 0.8894 | 0.8769 | 0.9706 |
91
- | 0.049 | 10.0 | 5810 | 0.1720 | 0.8641 | 0.8904 | 0.8771 | 0.9698 |
92
 
93
 
94
  ### Framework versions
 
25
  metrics:
26
  - name: Precision
27
  type: precision
28
+ value: 0.8750653423941454
29
  - name: Recall
30
  type: recall
31
+ value: 0.89470871191876
32
  - name: F1
33
  type: f1
34
+ value: 0.8847780126849896
35
  - name: Accuracy
36
  type: accuracy
37
+ value: 0.9699164786446582
38
  ---
39
 
40
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
44
 
45
  This model is a fine-tuned version of [FacebookAI/xlm-roberta-large](https://huggingface.co/FacebookAI/xlm-roberta-large) on the cnec dataset.
46
  It achieves the following results on the evaluation set:
47
+ - Loss: 0.2020
48
+ - Precision: 0.8751
49
+ - Recall: 0.8947
50
+ - F1: 0.8848
51
+ - Accuracy: 0.9699
52
 
53
  ## Model description
54
 
 
79
 
80
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
81
  |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
82
+ | 0.3776 | 1.0 | 581 | 0.1732 | 0.7868 | 0.8423 | 0.8136 | 0.9580 |
83
+ | 0.1773 | 2.0 | 1162 | 0.1476 | 0.8243 | 0.8675 | 0.8453 | 0.9625 |
84
+ | 0.127 | 3.0 | 1743 | 0.1522 | 0.8373 | 0.8691 | 0.8529 | 0.9654 |
85
+ | 0.1057 | 4.0 | 2324 | 0.1516 | 0.8604 | 0.8728 | 0.8665 | 0.9665 |
86
+ | 0.0852 | 5.0 | 2905 | 0.1555 | 0.8501 | 0.8883 | 0.8688 | 0.9700 |
87
+ | 0.069 | 6.0 | 3486 | 0.1847 | 0.8637 | 0.8910 | 0.8771 | 0.9681 |
88
+ | 0.0452 | 7.0 | 4067 | 0.1751 | 0.8666 | 0.8851 | 0.8757 | 0.9682 |
89
+ | 0.0385 | 8.0 | 4648 | 0.1968 | 0.8626 | 0.8888 | 0.8755 | 0.9690 |
90
+ | 0.0326 | 9.0 | 5229 | 0.1932 | 0.8717 | 0.8936 | 0.8826 | 0.9704 |
91
+ | 0.026 | 10.0 | 5810 | 0.2020 | 0.8751 | 0.8947 | 0.8848 | 0.9699 |
92
 
93
 
94
  ### Framework versions
config.json CHANGED
@@ -3,12 +3,12 @@
3
  "architectures": [
4
  "XLMRobertaForTokenClassification"
5
  ],
6
- "attention_probs_dropout_prob": 0.25,
7
  "bos_token_id": 0,
8
  "classifier_dropout": null,
9
  "eos_token_id": 2,
10
  "hidden_act": "gelu",
11
- "hidden_dropout_prob": 0.25,
12
  "hidden_size": 1024,
13
  "id2label": {
14
  "0": "O",
 
3
  "architectures": [
4
  "XLMRobertaForTokenClassification"
5
  ],
6
+ "attention_probs_dropout_prob": 0.2,
7
  "bos_token_id": 0,
8
  "classifier_dropout": null,
9
  "eos_token_id": 2,
10
  "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.2,
12
  "hidden_size": 1024,
13
  "id2label": {
14
  "0": "O",
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6b73162d300cfa0dd216f5343b6d3bf565aec740f2b8bf03693beffc8161c530
3
  size 2235473356
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b4f28b0626e84f9d3f45947f4f00d229f16b707199e727551cee10bd973aa1a
3
  size 2235473356
tmp-checkpoint-2910/config.json ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "FacebookAI/xlm-roberta-large",
3
+ "architectures": [
4
+ "XLMRobertaForTokenClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.15,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.3,
12
+ "hidden_size": 1024,
13
+ "id2label": {
14
+ "0": "O",
15
+ "1": "B-A",
16
+ "2": "I-A",
17
+ "3": "B-G",
18
+ "4": "I-G",
19
+ "5": "B-I",
20
+ "6": "I-I",
21
+ "7": "B-M",
22
+ "8": "I-M",
23
+ "9": "B-O",
24
+ "10": "I-O",
25
+ "11": "B-P",
26
+ "12": "I-P",
27
+ "13": "B-T",
28
+ "14": "I-T"
29
+ },
30
+ "initializer_range": 0.02,
31
+ "intermediate_size": 4096,
32
+ "label2id": {
33
+ "B-A": 1,
34
+ "B-G": 3,
35
+ "B-I": 5,
36
+ "B-M": 7,
37
+ "B-O": 9,
38
+ "B-P": 11,
39
+ "B-T": 13,
40
+ "I-A": 2,
41
+ "I-G": 4,
42
+ "I-I": 6,
43
+ "I-M": 8,
44
+ "I-O": 10,
45
+ "I-P": 12,
46
+ "I-T": 14,
47
+ "O": 0
48
+ },
49
+ "layer_norm_eps": 1e-05,
50
+ "max_position_embeddings": 514,
51
+ "model_type": "xlm-roberta",
52
+ "num_attention_heads": 16,
53
+ "num_hidden_layers": 24,
54
+ "output_past": true,
55
+ "pad_token_id": 1,
56
+ "position_embedding_type": "absolute",
57
+ "torch_dtype": "float32",
58
+ "transformers_version": "4.36.2",
59
+ "type_vocab_size": 1,
60
+ "use_cache": true,
61
+ "vocab_size": 250002
62
+ }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a6202687fe3aee68ea8ab8cea2ebaf5b94ee51f9a23262a8547aff52ab6b2dbc
3
  size 4728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c60a2cfe4ca61397b29fd56734d98f8d7ac8ab9e0eb13ffd59562de2b36795d
3
  size 4728