gustavokpc commited on
Commit
b597936
·
1 Parent(s): f75b878

Training in progress epoch 0

Browse files
Files changed (2) hide show
  1. README.md +14 -18
  2. tf_model.h5 +1 -1
README.md CHANGED
@@ -15,17 +15,17 @@ probably proofread and complete it, then remove this comment. -->
15
 
16
  This model is a fine-tuned version of [bert-large-uncased](https://huggingface.co/bert-large-uncased) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - Train Loss: 0.0547
19
- - Train Accuracy: 0.9798
20
- - Train F1 M: 0.5617
21
- - Train Precision M: 0.4043
22
- - Train Recall M: 0.9814
23
- - Validation Loss: 0.2571
24
- - Validation Accuracy: 0.9215
25
- - Validation F1 M: 0.5710
26
- - Validation Precision M: 0.4102
27
- - Validation Recall M: 0.9854
28
- - Epoch: 4
29
 
30
  ## Model description
31
 
@@ -44,23 +44,19 @@ More information needed
44
  ### Training hyperparameters
45
 
46
  The following hyperparameters were used during training:
47
- - optimizer: {'name': 'Adam', 'weight_decay': None, 'clipnorm': None, 'global_clipnorm': None, 'clipvalue': None, 'use_ema': False, 'ema_momentum': 0.99, 'ema_overwrite_frequency': None, 'jit_compile': False, 'is_legacy_optimizer': False, 'learning_rate': {'module': 'keras.optimizers.schedules', 'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 3790, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered_name': None}, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False}
48
  - training_precision: float32
49
 
50
  ### Training results
51
 
52
  | Train Loss | Train Accuracy | Train F1 M | Train Precision M | Train Recall M | Validation Loss | Validation Accuracy | Validation F1 M | Validation Precision M | Validation Recall M | Epoch |
53
  |:----------:|:--------------:|:----------:|:-----------------:|:--------------:|:---------------:|:-------------------:|:---------------:|:----------------------:|:-------------------:|:-----:|
54
- | 0.3876 | 0.8240 | 0.4634 | 0.3602 | 0.7384 | 0.2283 | 0.9096 | 0.5642 | 0.4130 | 0.9367 | 0 |
55
- | 0.1901 | 0.9290 | 0.5345 | 0.3949 | 0.8907 | 0.2162 | 0.9274 | 0.5642 | 0.4060 | 0.9693 | 1 |
56
- | 0.1235 | 0.9545 | 0.5556 | 0.4028 | 0.9576 | 0.1969 | 0.9228 | 0.5681 | 0.4118 | 0.9651 | 2 |
57
- | 0.0845 | 0.9688 | 0.5596 | 0.4031 | 0.9720 | 0.2147 | 0.9222 | 0.5715 | 0.4109 | 0.9829 | 3 |
58
- | 0.0547 | 0.9798 | 0.5617 | 0.4043 | 0.9814 | 0.2571 | 0.9215 | 0.5710 | 0.4102 | 0.9854 | 4 |
59
 
60
 
61
  ### Framework versions
62
 
63
  - Transformers 4.34.1
64
- - TensorFlow 2.14.0
65
  - Datasets 2.14.5
66
  - Tokenizers 0.14.1
 
15
 
16
  This model is a fine-tuned version of [bert-large-uncased](https://huggingface.co/bert-large-uncased) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Train Loss: 0.6885
19
+ - Train Accuracy: 0.5696
20
+ - Train F1 M: 0.4983
21
+ - Train Precision M: 0.3705
22
+ - Train Recall M: 0.9997
23
+ - Validation Loss: 0.6757
24
+ - Validation Accuracy: 0.5937
25
+ - Validation F1 M: 0.6390
26
+ - Validation Precision M: 0.4063
27
+ - Validation Recall M: 1.7203
28
+ - Epoch: 0
29
 
30
  ## Model description
31
 
 
44
  ### Training hyperparameters
45
 
46
  The following hyperparameters were used during training:
47
+ - optimizer: {'name': 'Adam', 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 15160, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False}
48
  - training_precision: float32
49
 
50
  ### Training results
51
 
52
  | Train Loss | Train Accuracy | Train F1 M | Train Precision M | Train Recall M | Validation Loss | Validation Accuracy | Validation F1 M | Validation Precision M | Validation Recall M | Epoch |
53
  |:----------:|:--------------:|:----------:|:-----------------:|:--------------:|:---------------:|:-------------------:|:---------------:|:----------------------:|:-------------------:|:-----:|
54
+ | 0.6885 | 0.5696 | 0.4983 | 0.3705 | 0.9997 | 0.6757 | 0.5937 | 0.6390 | 0.4063 | 1.7203 | 0 |
 
 
 
 
55
 
56
 
57
  ### Framework versions
58
 
59
  - Transformers 4.34.1
60
+ - TensorFlow 2.10.0
61
  - Datasets 2.14.5
62
  - Tokenizers 0.14.1
tf_model.h5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b53c710da6e3f60fb4a8bc38d0b49fdc1c583dec92191f3c346297dc10219b23
3
  size 1341127728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a555c0cd694a1dc7b189e3dfee7d19a8124cd2f6f000ef566591b84b971169d
3
  size 1341127728