Rodrigo1771 commited on
Commit
ef6226a
1 Parent(s): a0ff567

Model save

Browse files
README.md ADDED
@@ -0,0 +1,101 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: PlanTL-GOB-ES/bsc-bio-ehr-es
4
+ tags:
5
+ - generated_from_trainer
6
+ datasets:
7
+ - combined-train-distemist-dev-ner
8
+ metrics:
9
+ - precision
10
+ - recall
11
+ - f1
12
+ - accuracy
13
+ model-index:
14
+ - name: output
15
+ results:
16
+ - task:
17
+ name: Token Classification
18
+ type: token-classification
19
+ dataset:
20
+ name: combined-train-distemist-dev-ner
21
+ type: combined-train-distemist-dev-ner
22
+ config: CombinedTrainDisTEMISTDevNER
23
+ split: validation
24
+ args: CombinedTrainDisTEMISTDevNER
25
+ metrics:
26
+ - name: Precision
27
+ type: precision
28
+ value: 0.31002310289674784
29
+ - name: Recall
30
+ type: recall
31
+ value: 0.816331305568554
32
+ - name: F1
33
+ type: f1
34
+ value: 0.44938176197836166
35
+ - name: Accuracy
36
+ type: accuracy
37
+ value: 0.8498565092616749
38
+ ---
39
+
40
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
41
+ should probably proofread and complete it, then remove this comment. -->
42
+
43
+ # output
44
+
45
+ This model is a fine-tuned version of [PlanTL-GOB-ES/bsc-bio-ehr-es](https://huggingface.co/PlanTL-GOB-ES/bsc-bio-ehr-es) on the combined-train-distemist-dev-ner dataset.
46
+ It achieves the following results on the evaluation set:
47
+ - Loss: 0.9384
48
+ - Precision: 0.3100
49
+ - Recall: 0.8163
50
+ - F1: 0.4494
51
+ - Accuracy: 0.8499
52
+
53
+ ## Model description
54
+
55
+ More information needed
56
+
57
+ ## Intended uses & limitations
58
+
59
+ More information needed
60
+
61
+ ## Training and evaluation data
62
+
63
+ More information needed
64
+
65
+ ## Training procedure
66
+
67
+ ### Training hyperparameters
68
+
69
+ The following hyperparameters were used during training:
70
+ - learning_rate: 5e-05
71
+ - train_batch_size: 32
72
+ - eval_batch_size: 8
73
+ - seed: 42
74
+ - gradient_accumulation_steps: 2
75
+ - total_train_batch_size: 64
76
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
77
+ - lr_scheduler_type: linear
78
+ - num_epochs: 10.0
79
+
80
+ ### Training results
81
+
82
+ | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
83
+ |:-------------:|:------:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
84
+ | No log | 0.9988 | 425 | 0.3834 | 0.2920 | 0.7819 | 0.4252 | 0.8517 |
85
+ | 0.3349 | 2.0 | 851 | 0.5730 | 0.2681 | 0.8070 | 0.4025 | 0.8221 |
86
+ | 0.1788 | 2.9988 | 1276 | 0.5796 | 0.2848 | 0.8009 | 0.4202 | 0.8338 |
87
+ | 0.1227 | 4.0 | 1702 | 0.6591 | 0.2996 | 0.8109 | 0.4376 | 0.8388 |
88
+ | 0.0856 | 4.9988 | 2127 | 0.6266 | 0.3220 | 0.8203 | 0.4624 | 0.8602 |
89
+ | 0.0597 | 6.0 | 2553 | 0.7859 | 0.3075 | 0.8112 | 0.4460 | 0.8476 |
90
+ | 0.0597 | 6.9988 | 2978 | 0.8297 | 0.3137 | 0.8166 | 0.4532 | 0.8508 |
91
+ | 0.0458 | 8.0 | 3404 | 0.8468 | 0.3135 | 0.8205 | 0.4536 | 0.8532 |
92
+ | 0.0343 | 8.9988 | 3829 | 0.9241 | 0.3085 | 0.8182 | 0.4481 | 0.8494 |
93
+ | 0.0292 | 9.9882 | 4250 | 0.9384 | 0.3100 | 0.8163 | 0.4494 | 0.8499 |
94
+
95
+
96
+ ### Framework versions
97
+
98
+ - Transformers 4.42.4
99
+ - Pytorch 2.4.0+cu121
100
+ - Datasets 2.21.0
101
+ - Tokenizers 0.19.1
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dee8406b24dd8ea93c5ad47e89c3c3618bb8d35ff5c4692b8e2ba58443735fd0
3
  size 496262556
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0948f84443842f6cd934e63534288908b6139ed1c0d4f65942454302e5cb4afd
3
  size 496262556
tb/events.out.tfevents.1725046129.6b97e535edda.6370.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6e95865b7fc16c5b5c206f58b587bfc4cf54ab941b4946ad9235d0c58e4547d2
3
- size 11327
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9e94268d839c5a3c21781840b052643e729f1bd38c4ed1942a53d09ddd89bfc
3
+ size 12153
train.log CHANGED
@@ -1527,3 +1527,16 @@ Training completed. Do not forget to share your model on huggingface.co/models =
1527
  [INFO|trainer.py:2621] 2024-08-30 19:48:54,756 >> Loading best model from /content/dissertation/scripts/ner/output/checkpoint-2127 (score: 0.4624414693662204).
1528
 
1529
 
1530
  [INFO|trainer.py:4239] 2024-08-30 19:48:54,956 >> Waiting for the current checkpoint push to be finished, this might take a couple of minutes.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1527
  [INFO|trainer.py:2621] 2024-08-30 19:48:54,756 >> Loading best model from /content/dissertation/scripts/ner/output/checkpoint-2127 (score: 0.4624414693662204).
1528
 
1529
 
1530
  [INFO|trainer.py:4239] 2024-08-30 19:48:54,956 >> Waiting for the current checkpoint push to be finished, this might take a couple of minutes.
1531
+ [INFO|trainer.py:3478] 2024-08-30 19:48:56,764 >> Saving model checkpoint to /content/dissertation/scripts/ner/output
1532
+ [INFO|configuration_utils.py:472] 2024-08-30 19:48:56,766 >> Configuration saved in /content/dissertation/scripts/ner/output/config.json
1533
+ [INFO|modeling_utils.py:2690] 2024-08-30 19:48:58,154 >> Model weights saved in /content/dissertation/scripts/ner/output/model.safetensors
1534
+ [INFO|tokenization_utils_base.py:2574] 2024-08-30 19:48:58,155 >> tokenizer config file saved in /content/dissertation/scripts/ner/output/tokenizer_config.json
1535
+ [INFO|tokenization_utils_base.py:2583] 2024-08-30 19:48:58,156 >> Special tokens file saved in /content/dissertation/scripts/ner/output/special_tokens_map.json
1536
+ [INFO|trainer.py:3478] 2024-08-30 19:48:58,210 >> Saving model checkpoint to /content/dissertation/scripts/ner/output
1537
+ [INFO|configuration_utils.py:472] 2024-08-30 19:48:58,212 >> Configuration saved in /content/dissertation/scripts/ner/output/config.json
1538
+ [INFO|modeling_utils.py:2690] 2024-08-30 19:48:59,637 >> Model weights saved in /content/dissertation/scripts/ner/output/model.safetensors
1539
+ [INFO|tokenization_utils_base.py:2574] 2024-08-30 19:48:59,639 >> tokenizer config file saved in /content/dissertation/scripts/ner/output/tokenizer_config.json
1540
+ [INFO|tokenization_utils_base.py:2583] 2024-08-30 19:48:59,639 >> Special tokens file saved in /content/dissertation/scripts/ner/output/special_tokens_map.json
1541
+ {'eval_loss': 0.938378632068634, 'eval_precision': 0.31002310289674784, 'eval_recall': 0.816331305568554, 'eval_f1': 0.44938176197836166, 'eval_accuracy': 0.8498565092616749, 'eval_runtime': 15.4697, 'eval_samples_per_second': 440.215, 'eval_steps_per_second': 55.075, 'epoch': 9.99}
1542
+ {'train_runtime': 1205.6346, 'train_samples_per_second': 225.848, 'train_steps_per_second': 3.525, 'train_loss': 0.10639642311544979, 'epoch': 9.99}
1543
+