stulcrad commited on
Commit
3566548
1 Parent(s): 999bb48

End of training

Browse files
Files changed (4) hide show
  1. README.md +19 -24
  2. config.json +1 -1
  3. model.safetensors +1 -1
  4. training_args.bin +1 -1
README.md CHANGED
@@ -25,16 +25,16 @@ model-index:
25
  metrics:
26
  - name: Precision
27
  type: precision
28
- value: 0.8641078838174274
29
  - name: Recall
30
  type: recall
31
- value: 0.8904329235702833
32
  - name: F1
33
  type: f1
34
- value: 0.877072913924717
35
  - name: Accuracy
36
  type: accuracy
37
- value: 0.970074812967581
38
  ---
39
 
40
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -44,11 +44,11 @@ should probably proofread and complete it, then remove this comment. -->
44
 
45
  This model is a fine-tuned version of [FacebookAI/xlm-roberta-large](https://huggingface.co/FacebookAI/xlm-roberta-large) on the cnec dataset.
46
  It achieves the following results on the evaluation set:
47
- - Loss: 0.1937
48
- - Precision: 0.8641
49
- - Recall: 0.8904
50
- - F1: 0.8771
51
- - Accuracy: 0.9701
52
 
53
  ## Model description
54
 
@@ -68,27 +68,22 @@ More information needed
68
 
69
  The following hyperparameters were used during training:
70
  - learning_rate: 2e-05
71
- - train_batch_size: 4
72
- - eval_batch_size: 4
73
  - seed: 42
74
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
75
  - lr_scheduler_type: linear
76
- - num_epochs: 10
77
 
78
  ### Training results
79
 
80
- | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
81
- |:-------------:|:-----:|:-----:|:---------------:|:---------:|:------:|:------:|:--------:|
82
- | 0.3037 | 1.0 | 1161 | 0.1876 | 0.7785 | 0.8113 | 0.7946 | 0.9529 |
83
- | 0.2331 | 2.0 | 2322 | 0.2008 | 0.8065 | 0.8263 | 0.8163 | 0.9569 |
84
- | 0.1828 | 3.0 | 3483 | 0.1656 | 0.8332 | 0.8648 | 0.8487 | 0.9648 |
85
- | 0.1456 | 4.0 | 4644 | 0.1659 | 0.8414 | 0.8675 | 0.8542 | 0.9643 |
86
- | 0.1237 | 5.0 | 5805 | 0.1746 | 0.8538 | 0.8899 | 0.8715 | 0.9690 |
87
- | 0.1074 | 6.0 | 6966 | 0.1782 | 0.8584 | 0.8878 | 0.8728 | 0.9691 |
88
- | 0.097 | 7.0 | 8127 | 0.1802 | 0.8517 | 0.8840 | 0.8676 | 0.9691 |
89
- | 0.072 | 8.0 | 9288 | 0.1908 | 0.8636 | 0.8867 | 0.875 | 0.9703 |
90
- | 0.067 | 9.0 | 10449 | 0.1962 | 0.8672 | 0.8936 | 0.8802 | 0.9711 |
91
- | 0.0636 | 10.0 | 11610 | 0.1937 | 0.8641 | 0.8904 | 0.8771 | 0.9701 |
92
 
93
 
94
  ### Framework versions
 
25
  metrics:
26
  - name: Precision
27
  type: precision
28
+ value: 0.8533541341653667
29
  - name: Recall
30
  type: recall
31
+ value: 0.8770710849812934
32
  - name: F1
33
  type: f1
34
+ value: 0.8650500790722193
35
  - name: Accuracy
36
  type: accuracy
37
+ value: 0.9670664608320468
38
  ---
39
 
40
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
44
 
45
  This model is a fine-tuned version of [FacebookAI/xlm-roberta-large](https://huggingface.co/FacebookAI/xlm-roberta-large) on the cnec dataset.
46
  It achieves the following results on the evaluation set:
47
+ - Loss: 0.1498
48
+ - Precision: 0.8534
49
+ - Recall: 0.8771
50
+ - F1: 0.8651
51
+ - Accuracy: 0.9671
52
 
53
  ## Model description
54
 
 
68
 
69
  The following hyperparameters were used during training:
70
  - learning_rate: 2e-05
71
+ - train_batch_size: 8
72
+ - eval_batch_size: 8
73
  - seed: 42
74
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
75
  - lr_scheduler_type: linear
76
+ - num_epochs: 5
77
 
78
  ### Training results
79
 
80
+ | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
81
+ |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
82
+ | 0.3961 | 1.0 | 581 | 0.1800 | 0.8004 | 0.8231 | 0.8116 | 0.9560 |
83
+ | 0.1772 | 2.0 | 1162 | 0.1518 | 0.8357 | 0.8648 | 0.8500 | 0.9642 |
84
+ | 0.1266 | 3.0 | 1743 | 0.1545 | 0.8377 | 0.8717 | 0.8544 | 0.9680 |
85
+ | 0.1043 | 4.0 | 2324 | 0.1472 | 0.8473 | 0.8691 | 0.8580 | 0.9656 |
86
+ | 0.0804 | 5.0 | 2905 | 0.1498 | 0.8534 | 0.8771 | 0.8651 | 0.9671 |
 
 
 
 
 
87
 
88
 
89
  ### Framework versions
config.json CHANGED
@@ -8,7 +8,7 @@
8
  "classifier_dropout": null,
9
  "eos_token_id": 2,
10
  "hidden_act": "gelu",
11
- "hidden_dropout_prob": 0.3,
12
  "hidden_size": 1024,
13
  "id2label": {
14
  "0": "O",
 
8
  "classifier_dropout": null,
9
  "eos_token_id": 2,
10
  "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.2,
12
  "hidden_size": 1024,
13
  "id2label": {
14
  "0": "O",
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5edeaf9e0b1f00c8ba12f0bd348c0ada9eb21d1a4b87f8b5afbbc53496bd6893
3
  size 2235473356
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df46b2750ed74f176adba5d7e288d0b10360c90a4b06fd1c50ebf2e7026c1ea5
3
  size 2235473356
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:88026046f15f5dcd1a7c63f9d988e107838d32a787b3c6ba70a65b6440d455c1
3
  size 4728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df1edf8fec4061938d4b10761bf8d9ecaf1a667dd1c56f2044b0941cb214f45b
3
  size 4728