phitime commited on
Commit
c0e6080
1 Parent(s): 63ecb42

End of training

Browse files
Files changed (1) hide show
  1. README.md +12 -16
README.md CHANGED
@@ -3,8 +3,6 @@ license: apache-2.0
3
  base_model: google/flan-t5-small
4
  tags:
5
  - generated_from_trainer
6
- metrics:
7
- - rouge
8
  model-index:
9
  - name: flan-t5-small-finetuned-mlsum-tr
10
  results: []
@@ -17,12 +15,17 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [google/flan-t5-small](https://huggingface.co/google/flan-t5-small) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: nan
21
- - Rouge1: 4.1461
22
- - Rouge2: 1.3002
23
- - Rougel: 3.4895
24
- - Rougelsum: 3.663
25
- - Gen Len: 14.1553
 
 
 
 
 
26
 
27
  ## Model description
28
 
@@ -47,16 +50,9 @@ The following hyperparameters were used during training:
47
  - seed: 42
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
  - lr_scheduler_type: linear
50
- - num_epochs: 1
51
  - mixed_precision_training: Native AMP
52
 
53
- ### Training results
54
-
55
- | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
56
- |:-------------:|:-----:|:-----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|
57
- | 0.0 | 1.0 | 15580 | nan | 4.1461 | 1.3002 | 3.4895 | 3.663 | 14.1553 |
58
-
59
-
60
  ### Framework versions
61
 
62
  - Transformers 4.35.2
 
3
  base_model: google/flan-t5-small
4
  tags:
5
  - generated_from_trainer
 
 
6
  model-index:
7
  - name: flan-t5-small-finetuned-mlsum-tr
8
  results: []
 
15
 
16
  This model is a fine-tuned version of [google/flan-t5-small](https://huggingface.co/google/flan-t5-small) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
+ - eval_loss: nan
19
+ - eval_rouge1: 10.3443
20
+ - eval_rouge2: 5.3615
21
+ - eval_rougeL: 8.9871
22
+ - eval_rougeLsum: 9.3134
23
+ - eval_gen_len: 16.1858
24
+ - eval_runtime: 378.4208
25
+ - eval_samples_per_second: 33.759
26
+ - eval_steps_per_second: 2.111
27
+ - epoch: 2.0
28
+ - step: 31160
29
 
30
  ## Model description
31
 
 
50
  - seed: 42
51
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
52
  - lr_scheduler_type: linear
53
+ - num_epochs: 5
54
  - mixed_precision_training: Native AMP
55
 
 
 
 
 
 
 
 
56
  ### Framework versions
57
 
58
  - Transformers 4.35.2