Ransaka commited on
Commit
766759a
1 Parent(s): 70e0bbf

End of training

Browse files
Files changed (2) hide show
  1. README.md +15 -3
  2. generation_config.json +0 -2
README.md CHANGED
@@ -12,6 +12,11 @@ should probably proofread and complete it, then remove this comment. -->
12
  # sinhala-roman-transformer
13
 
14
  This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
 
 
 
 
 
15
 
16
  ## Model description
17
 
@@ -37,16 +42,23 @@ The following hyperparameters were used during training:
37
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
38
  - lr_scheduler_type: linear
39
  - lr_scheduler_warmup_steps: 2000
40
- - num_epochs: 3.0
41
  - mixed_precision_training: Native AMP
42
 
43
  ### Training results
44
 
 
 
 
 
 
 
 
45
 
46
 
47
  ### Framework versions
48
 
49
  - Transformers 4.41.2
50
- - Pytorch 2.3.0+cu121
51
- - Datasets 2.20.0
52
  - Tokenizers 0.19.1
 
12
  # sinhala-roman-transformer
13
 
14
  This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
15
+ It achieves the following results on the evaluation set:
16
+ - Loss: 0.0029
17
+ - Rouge2 Precision: 0.0
18
+ - Rouge2 Recall: 0.0
19
+ - Rouge2 Fmeasure: 0.0
20
 
21
  ## Model description
22
 
 
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: linear
44
  - lr_scheduler_warmup_steps: 2000
45
+ - training_steps: 40000
46
  - mixed_precision_training: Native AMP
47
 
48
  ### Training results
49
 
50
+ | Training Loss | Epoch | Step | Validation Loss | Rouge2 Precision | Rouge2 Recall | Rouge2 Fmeasure |
51
+ |:-------------:|:-------:|:-----:|:---------------:|:----------------:|:-------------:|:---------------:|
52
+ | 0.006 | 5.1440 | 7500 | 0.0052 | 0.0 | 0.0 | 0.0 |
53
+ | 0.0002 | 10.2881 | 15000 | 0.0034 | 0.0 | 0.0 | 0.0 |
54
+ | 0.0 | 15.4321 | 22500 | 0.0030 | 0.0 | 0.0 | 0.0 |
55
+ | 0.0 | 20.5761 | 30000 | 0.0030 | 0.0 | 0.0 | 0.0 |
56
+ | 0.0 | 25.7202 | 37500 | 0.0029 | 0.0 | 0.0 | 0.0 |
57
 
58
 
59
  ### Framework versions
60
 
61
  - Transformers 4.41.2
62
+ - Pytorch 2.1.2
63
+ - Datasets 2.19.2
64
  - Tokenizers 0.19.1
generation_config.json CHANGED
@@ -3,8 +3,6 @@
3
  "early_stopping": true,
4
  "eos_token_id": 3,
5
  "length_penalty": 2.0,
6
- "max_length": 142,
7
- "min_length": 56,
8
  "no_repeat_ngram_size": 3,
9
  "num_beams": 4,
10
  "pad_token_id": 0,
 
3
  "early_stopping": true,
4
  "eos_token_id": 3,
5
  "length_penalty": 2.0,
 
 
6
  "no_repeat_ngram_size": 3,
7
  "num_beams": 4,
8
  "pad_token_id": 0,