MwiseChacha commited on
Commit
2ca59d8
1 Parent(s): eece6c8

End of training

Browse files
Files changed (2) hide show
  1. README.md +5 -2
  2. generation_config.json +1 -1
README.md CHANGED
@@ -41,16 +41,19 @@ The following hyperparameters were used during training:
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: linear
43
  - lr_scheduler_warmup_steps: 10
44
- - training_steps: 5
45
  - mixed_precision_training: Native AMP
46
 
47
  ### Training results
48
 
 
 
 
49
 
50
 
51
  ### Framework versions
52
 
53
- - Transformers 4.42.3
54
  - Pytorch 2.3.1+cu121
55
  - Datasets 2.20.0
56
  - Tokenizers 0.19.1
 
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: linear
43
  - lr_scheduler_warmup_steps: 10
44
+ - training_steps: 10
45
  - mixed_precision_training: Native AMP
46
 
47
  ### Training results
48
 
49
+ | Training Loss | Epoch | Step | Validation Loss |
50
+ |:-------------:|:------:|:----:|:---------------:|
51
+ | No log | 0.7407 | 10 | 0.6153 |
52
 
53
 
54
  ### Framework versions
55
 
56
+ - Transformers 4.41.2
57
  - Pytorch 2.3.1+cu121
58
  - Datasets 2.20.0
59
  - Tokenizers 0.19.1
generation_config.json CHANGED
@@ -5,5 +5,5 @@
5
  "eos_token_id": 2,
6
  "max_length": 1876,
7
  "pad_token_id": 1,
8
- "transformers_version": "4.42.3"
9
  }
 
5
  "eos_token_id": 2,
6
  "max_length": 1876,
7
  "pad_token_id": 1,
8
+ "transformers_version": "4.41.2"
9
  }