anantoj commited on
Commit
8e211ba
1 Parent(s): 7045e1b

Initial commit

Browse files
README.md ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - generated_from_trainer
5
+ model-index:
6
+ - name: results
7
+ results: []
8
+ ---
9
+
10
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
11
+ should probably proofread and complete it, then remove this comment. -->
12
+
13
+ # results
14
+
15
+ This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on the None dataset.
16
+ It achieves the following results on the evaluation set:
17
+ - Loss: 2.0868
18
+
19
+ ## Model description
20
+
21
+ More information needed
22
+
23
+ ## Intended uses & limitations
24
+
25
+ More information needed
26
+
27
+ ## Training and evaluation data
28
+
29
+ More information needed
30
+
31
+ ## Training procedure
32
+
33
+ ### Training hyperparameters
34
+
35
+ The following hyperparameters were used during training:
36
+ - learning_rate: 2e-05
37
+ - train_batch_size: 16
38
+ - eval_batch_size: 16
39
+ - seed: 42
40
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
41
+ - lr_scheduler_type: linear
42
+ - num_epochs: 3
43
+ - mixed_precision_training: Native AMP
44
+
45
+ ### Training results
46
+
47
+ | Training Loss | Epoch | Step | Validation Loss |
48
+ |:-------------:|:-----:|:-----:|:---------------:|
49
+ | 2.2583 | 1.0 | 14719 | 2.1164 |
50
+ | 2.2649 | 2.0 | 29438 | 2.0925 |
51
+ | 2.209 | 3.0 | 44157 | 2.0868 |
52
+
53
+
54
+ ### Framework versions
55
+
56
+ - Transformers 4.19.4
57
+ - Pytorch 1.11.0+cu113
58
+ - Datasets 2.3.2
59
+ - Tokenizers 0.12.1
checkpoint-44157/config.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "t5-small",
3
+ "architectures": [
4
+ "T5ForConditionalGeneration"
5
+ ],
6
+ "d_ff": 2048,
7
+ "d_kv": 64,
8
+ "d_model": 512,
9
+ "decoder_start_token_id": 0,
10
+ "dropout_rate": 0.1,
11
+ "eos_token_id": 1,
12
+ "feed_forward_proj": "relu",
13
+ "initializer_factor": 1.0,
14
+ "is_encoder_decoder": true,
15
+ "layer_norm_epsilon": 1e-06,
16
+ "model_type": "t5",
17
+ "n_positions": 512,
18
+ "num_decoder_layers": 6,
19
+ "num_heads": 8,
20
+ "num_layers": 6,
21
+ "output_past": true,
22
+ "pad_token_id": 0,
23
+ "relative_attention_max_distance": 128,
24
+ "relative_attention_num_buckets": 32,
25
+ "task_specific_params": {
26
+ "summarization": {
27
+ "early_stopping": true,
28
+ "length_penalty": 2.0,
29
+ "max_length": 200,
30
+ "min_length": 30,
31
+ "no_repeat_ngram_size": 3,
32
+ "num_beams": 4,
33
+ "prefix": "summarize: "
34
+ },
35
+ "translation_en_to_de": {
36
+ "early_stopping": true,
37
+ "max_length": 300,
38
+ "num_beams": 4,
39
+ "prefix": "translate English to German: "
40
+ },
41
+ "translation_en_to_fr": {
42
+ "early_stopping": true,
43
+ "max_length": 300,
44
+ "num_beams": 4,
45
+ "prefix": "translate English to French: "
46
+ },
47
+ "translation_en_to_ro": {
48
+ "early_stopping": true,
49
+ "max_length": 300,
50
+ "num_beams": 4,
51
+ "prefix": "translate English to Romanian: "
52
+ }
53
+ },
54
+ "torch_dtype": "float32",
55
+ "transformers_version": "4.19.4",
56
+ "use_cache": true,
57
+ "vocab_size": 32128
58
+ }
checkpoint-44157/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15c46c37a1b327317e2d9ba045448b5c85be37f9b2581c8624114bcf64b521c0
3
+ size 484129461
checkpoint-44157/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2a080ed17900659a6f7c133c72c814002125f51c35ae6c7d423421367542a0e
3
+ size 242070267
checkpoint-44157/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc3d0e710cabea991f3688f1b06042c351dcad4c6c31e7155a30be41084e42ec
3
+ size 14503
checkpoint-44157/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19ebcd5f457b4e963a47547930d77acbcf9f874a502751b8720fb128e1b3893a
3
+ size 559
checkpoint-44157/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d67dd30727a2b39ab56e811aa8b1718be5cf6a140eb779db146e24974f09c3e4
3
+ size 623
checkpoint-44157/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>", "additional_special_tokens": ["<extra_id_0>", "<extra_id_1>", "<extra_id_2>", "<extra_id_3>", "<extra_id_4>", "<extra_id_5>", "<extra_id_6>", "<extra_id_7>", "<extra_id_8>", "<extra_id_9>", "<extra_id_10>", "<extra_id_11>", "<extra_id_12>", "<extra_id_13>", "<extra_id_14>", "<extra_id_15>", "<extra_id_16>", "<extra_id_17>", "<extra_id_18>", "<extra_id_19>", "<extra_id_20>", "<extra_id_21>", "<extra_id_22>", "<extra_id_23>", "<extra_id_24>", "<extra_id_25>", "<extra_id_26>", "<extra_id_27>", "<extra_id_28>", "<extra_id_29>", "<extra_id_30>", "<extra_id_31>", "<extra_id_32>", "<extra_id_33>", "<extra_id_34>", "<extra_id_35>", "<extra_id_36>", "<extra_id_37>", "<extra_id_38>", "<extra_id_39>", "<extra_id_40>", "<extra_id_41>", "<extra_id_42>", "<extra_id_43>", "<extra_id_44>", "<extra_id_45>", "<extra_id_46>", "<extra_id_47>", "<extra_id_48>", "<extra_id_49>", "<extra_id_50>", "<extra_id_51>", "<extra_id_52>", "<extra_id_53>", "<extra_id_54>", "<extra_id_55>", "<extra_id_56>", "<extra_id_57>", "<extra_id_58>", "<extra_id_59>", "<extra_id_60>", "<extra_id_61>", "<extra_id_62>", "<extra_id_63>", "<extra_id_64>", "<extra_id_65>", "<extra_id_66>", "<extra_id_67>", "<extra_id_68>", "<extra_id_69>", "<extra_id_70>", "<extra_id_71>", "<extra_id_72>", "<extra_id_73>", "<extra_id_74>", "<extra_id_75>", "<extra_id_76>", "<extra_id_77>", "<extra_id_78>", "<extra_id_79>", "<extra_id_80>", "<extra_id_81>", "<extra_id_82>", "<extra_id_83>", "<extra_id_84>", "<extra_id_85>", "<extra_id_86>", "<extra_id_87>", "<extra_id_88>", "<extra_id_89>", "<extra_id_90>", "<extra_id_91>", "<extra_id_92>", "<extra_id_93>", "<extra_id_94>", "<extra_id_95>", "<extra_id_96>", "<extra_id_97>", "<extra_id_98>", "<extra_id_99>"]}
checkpoint-44157/spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d60acb128cf7b7f2536e8f38a5b18a05535c9e14c7a355904270e15b0945ea86
3
+ size 791656
checkpoint-44157/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-44157/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>", "extra_ids": 100, "additional_special_tokens": ["<extra_id_0>", "<extra_id_1>", "<extra_id_2>", "<extra_id_3>", "<extra_id_4>", "<extra_id_5>", "<extra_id_6>", "<extra_id_7>", "<extra_id_8>", "<extra_id_9>", "<extra_id_10>", "<extra_id_11>", "<extra_id_12>", "<extra_id_13>", "<extra_id_14>", "<extra_id_15>", "<extra_id_16>", "<extra_id_17>", "<extra_id_18>", "<extra_id_19>", "<extra_id_20>", "<extra_id_21>", "<extra_id_22>", "<extra_id_23>", "<extra_id_24>", "<extra_id_25>", "<extra_id_26>", "<extra_id_27>", "<extra_id_28>", "<extra_id_29>", "<extra_id_30>", "<extra_id_31>", "<extra_id_32>", "<extra_id_33>", "<extra_id_34>", "<extra_id_35>", "<extra_id_36>", "<extra_id_37>", "<extra_id_38>", "<extra_id_39>", "<extra_id_40>", "<extra_id_41>", "<extra_id_42>", "<extra_id_43>", "<extra_id_44>", "<extra_id_45>", "<extra_id_46>", "<extra_id_47>", "<extra_id_48>", "<extra_id_49>", "<extra_id_50>", "<extra_id_51>", "<extra_id_52>", "<extra_id_53>", "<extra_id_54>", "<extra_id_55>", "<extra_id_56>", "<extra_id_57>", "<extra_id_58>", "<extra_id_59>", "<extra_id_60>", "<extra_id_61>", "<extra_id_62>", "<extra_id_63>", "<extra_id_64>", "<extra_id_65>", "<extra_id_66>", "<extra_id_67>", "<extra_id_68>", "<extra_id_69>", "<extra_id_70>", "<extra_id_71>", "<extra_id_72>", "<extra_id_73>", "<extra_id_74>", "<extra_id_75>", "<extra_id_76>", "<extra_id_77>", "<extra_id_78>", "<extra_id_79>", "<extra_id_80>", "<extra_id_81>", "<extra_id_82>", "<extra_id_83>", "<extra_id_84>", "<extra_id_85>", "<extra_id_86>", "<extra_id_87>", "<extra_id_88>", "<extra_id_89>", "<extra_id_90>", "<extra_id_91>", "<extra_id_92>", "<extra_id_93>", "<extra_id_94>", "<extra_id_95>", "<extra_id_96>", "<extra_id_97>", "<extra_id_98>", "<extra_id_99>"], "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "t5-small", "tokenizer_class": "T5Tokenizer"}
checkpoint-44157/trainer_state.json ADDED
@@ -0,0 +1,568 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 3.0,
5
+ "global_step": 44157,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.03,
12
+ "learning_rate": 1.9774441198450984e-05,
13
+ "loss": 2.5305,
14
+ "step": 500
15
+ },
16
+ {
17
+ "epoch": 0.07,
18
+ "learning_rate": 1.9548429467581584e-05,
19
+ "loss": 2.4174,
20
+ "step": 1000
21
+ },
22
+ {
23
+ "epoch": 0.1,
24
+ "learning_rate": 1.9322417736712188e-05,
25
+ "loss": 2.4269,
26
+ "step": 1500
27
+ },
28
+ {
29
+ "epoch": 0.14,
30
+ "learning_rate": 1.909595307652241e-05,
31
+ "loss": 2.3693,
32
+ "step": 2000
33
+ },
34
+ {
35
+ "epoch": 0.17,
36
+ "learning_rate": 1.8869488416332635e-05,
37
+ "loss": 2.3566,
38
+ "step": 2500
39
+ },
40
+ {
41
+ "epoch": 0.2,
42
+ "learning_rate": 1.8643023756142857e-05,
43
+ "loss": 2.3586,
44
+ "step": 3000
45
+ },
46
+ {
47
+ "epoch": 0.24,
48
+ "learning_rate": 1.841655909595308e-05,
49
+ "loss": 2.3499,
50
+ "step": 3500
51
+ },
52
+ {
53
+ "epoch": 0.27,
54
+ "learning_rate": 1.81900944357633e-05,
55
+ "loss": 2.3594,
56
+ "step": 4000
57
+ },
58
+ {
59
+ "epoch": 0.31,
60
+ "learning_rate": 1.7963629775573522e-05,
61
+ "loss": 2.345,
62
+ "step": 4500
63
+ },
64
+ {
65
+ "epoch": 0.34,
66
+ "learning_rate": 1.7737165115383744e-05,
67
+ "loss": 2.321,
68
+ "step": 5000
69
+ },
70
+ {
71
+ "epoch": 0.37,
72
+ "learning_rate": 1.7510700455193966e-05,
73
+ "loss": 2.3003,
74
+ "step": 5500
75
+ },
76
+ {
77
+ "epoch": 0.41,
78
+ "learning_rate": 1.728423579500419e-05,
79
+ "loss": 2.3128,
80
+ "step": 6000
81
+ },
82
+ {
83
+ "epoch": 0.44,
84
+ "learning_rate": 1.7058224064134795e-05,
85
+ "loss": 2.3011,
86
+ "step": 6500
87
+ },
88
+ {
89
+ "epoch": 0.48,
90
+ "learning_rate": 1.6831759403945017e-05,
91
+ "loss": 2.3096,
92
+ "step": 7000
93
+ },
94
+ {
95
+ "epoch": 0.51,
96
+ "learning_rate": 1.660529474375524e-05,
97
+ "loss": 2.3256,
98
+ "step": 7500
99
+ },
100
+ {
101
+ "epoch": 0.54,
102
+ "learning_rate": 1.637883008356546e-05,
103
+ "loss": 2.2922,
104
+ "step": 8000
105
+ },
106
+ {
107
+ "epoch": 0.58,
108
+ "learning_rate": 1.6152365423375686e-05,
109
+ "loss": 2.3021,
110
+ "step": 8500
111
+ },
112
+ {
113
+ "epoch": 0.61,
114
+ "learning_rate": 1.5925900763185907e-05,
115
+ "loss": 2.3011,
116
+ "step": 9000
117
+ },
118
+ {
119
+ "epoch": 0.65,
120
+ "learning_rate": 1.569943610299613e-05,
121
+ "loss": 2.2814,
122
+ "step": 9500
123
+ },
124
+ {
125
+ "epoch": 0.68,
126
+ "learning_rate": 1.547297144280635e-05,
127
+ "loss": 2.2816,
128
+ "step": 10000
129
+ },
130
+ {
131
+ "epoch": 0.71,
132
+ "learning_rate": 1.5246959711936953e-05,
133
+ "loss": 2.2801,
134
+ "step": 10500
135
+ },
136
+ {
137
+ "epoch": 0.75,
138
+ "learning_rate": 1.5020947981067555e-05,
139
+ "loss": 2.2993,
140
+ "step": 11000
141
+ },
142
+ {
143
+ "epoch": 0.78,
144
+ "learning_rate": 1.4794483320877777e-05,
145
+ "loss": 2.3057,
146
+ "step": 11500
147
+ },
148
+ {
149
+ "epoch": 0.82,
150
+ "learning_rate": 1.4568018660688002e-05,
151
+ "loss": 2.2701,
152
+ "step": 12000
153
+ },
154
+ {
155
+ "epoch": 0.85,
156
+ "learning_rate": 1.4341554000498224e-05,
157
+ "loss": 2.2742,
158
+ "step": 12500
159
+ },
160
+ {
161
+ "epoch": 0.88,
162
+ "learning_rate": 1.4115089340308446e-05,
163
+ "loss": 2.2872,
164
+ "step": 13000
165
+ },
166
+ {
167
+ "epoch": 0.92,
168
+ "learning_rate": 1.388862468011867e-05,
169
+ "loss": 2.256,
170
+ "step": 13500
171
+ },
172
+ {
173
+ "epoch": 0.95,
174
+ "learning_rate": 1.3662160019928891e-05,
175
+ "loss": 2.2959,
176
+ "step": 14000
177
+ },
178
+ {
179
+ "epoch": 0.99,
180
+ "learning_rate": 1.3435695359739113e-05,
181
+ "loss": 2.2583,
182
+ "step": 14500
183
+ },
184
+ {
185
+ "epoch": 1.0,
186
+ "eval_loss": 2.116374969482422,
187
+ "eval_runtime": 353.9363,
188
+ "eval_samples_per_second": 166.341,
189
+ "eval_steps_per_second": 10.397,
190
+ "step": 14719
191
+ },
192
+ {
193
+ "epoch": 1.02,
194
+ "learning_rate": 1.3209230699549337e-05,
195
+ "loss": 2.2496,
196
+ "step": 15000
197
+ },
198
+ {
199
+ "epoch": 1.05,
200
+ "learning_rate": 1.2983218968679939e-05,
201
+ "loss": 2.2561,
202
+ "step": 15500
203
+ },
204
+ {
205
+ "epoch": 1.09,
206
+ "learning_rate": 1.275675430849016e-05,
207
+ "loss": 2.2408,
208
+ "step": 16000
209
+ },
210
+ {
211
+ "epoch": 1.12,
212
+ "learning_rate": 1.2530289648300384e-05,
213
+ "loss": 2.2835,
214
+ "step": 16500
215
+ },
216
+ {
217
+ "epoch": 1.15,
218
+ "learning_rate": 1.2303824988110606e-05,
219
+ "loss": 2.2636,
220
+ "step": 17000
221
+ },
222
+ {
223
+ "epoch": 1.19,
224
+ "learning_rate": 1.2077360327920828e-05,
225
+ "loss": 2.2751,
226
+ "step": 17500
227
+ },
228
+ {
229
+ "epoch": 1.22,
230
+ "learning_rate": 1.1850895667731051e-05,
231
+ "loss": 2.279,
232
+ "step": 18000
233
+ },
234
+ {
235
+ "epoch": 1.26,
236
+ "learning_rate": 1.1624431007541275e-05,
237
+ "loss": 2.27,
238
+ "step": 18500
239
+ },
240
+ {
241
+ "epoch": 1.29,
242
+ "learning_rate": 1.1398419276671877e-05,
243
+ "loss": 2.2436,
244
+ "step": 19000
245
+ },
246
+ {
247
+ "epoch": 1.32,
248
+ "learning_rate": 1.1171954616482099e-05,
249
+ "loss": 2.2575,
250
+ "step": 19500
251
+ },
252
+ {
253
+ "epoch": 1.36,
254
+ "learning_rate": 1.094548995629232e-05,
255
+ "loss": 2.2669,
256
+ "step": 20000
257
+ },
258
+ {
259
+ "epoch": 1.39,
260
+ "learning_rate": 1.0719025296102544e-05,
261
+ "loss": 2.269,
262
+ "step": 20500
263
+ },
264
+ {
265
+ "epoch": 1.43,
266
+ "learning_rate": 1.0492560635912768e-05,
267
+ "loss": 2.2429,
268
+ "step": 21000
269
+ },
270
+ {
271
+ "epoch": 1.46,
272
+ "learning_rate": 1.026609597572299e-05,
273
+ "loss": 2.2424,
274
+ "step": 21500
275
+ },
276
+ {
277
+ "epoch": 1.49,
278
+ "learning_rate": 1.0039631315533211e-05,
279
+ "loss": 2.2549,
280
+ "step": 22000
281
+ },
282
+ {
283
+ "epoch": 1.53,
284
+ "learning_rate": 9.813166655343435e-06,
285
+ "loss": 2.2623,
286
+ "step": 22500
287
+ },
288
+ {
289
+ "epoch": 1.56,
290
+ "learning_rate": 9.587154924474037e-06,
291
+ "loss": 2.2634,
292
+ "step": 23000
293
+ },
294
+ {
295
+ "epoch": 1.6,
296
+ "learning_rate": 9.360690264284259e-06,
297
+ "loss": 2.2674,
298
+ "step": 23500
299
+ },
300
+ {
301
+ "epoch": 1.63,
302
+ "learning_rate": 9.134225604094482e-06,
303
+ "loss": 2.2611,
304
+ "step": 24000
305
+ },
306
+ {
307
+ "epoch": 1.66,
308
+ "learning_rate": 8.907760943904704e-06,
309
+ "loss": 2.2418,
310
+ "step": 24500
311
+ },
312
+ {
313
+ "epoch": 1.7,
314
+ "learning_rate": 8.681749213035306e-06,
315
+ "loss": 2.2493,
316
+ "step": 25000
317
+ },
318
+ {
319
+ "epoch": 1.73,
320
+ "learning_rate": 8.45528455284553e-06,
321
+ "loss": 2.2259,
322
+ "step": 25500
323
+ },
324
+ {
325
+ "epoch": 1.77,
326
+ "learning_rate": 8.229272821976131e-06,
327
+ "loss": 2.2296,
328
+ "step": 26000
329
+ },
330
+ {
331
+ "epoch": 1.8,
332
+ "learning_rate": 8.002808161786353e-06,
333
+ "loss": 2.2258,
334
+ "step": 26500
335
+ },
336
+ {
337
+ "epoch": 1.83,
338
+ "learning_rate": 7.776343501596577e-06,
339
+ "loss": 2.2407,
340
+ "step": 27000
341
+ },
342
+ {
343
+ "epoch": 1.87,
344
+ "learning_rate": 7.5498788414067996e-06,
345
+ "loss": 2.2242,
346
+ "step": 27500
347
+ },
348
+ {
349
+ "epoch": 1.9,
350
+ "learning_rate": 7.323414181217021e-06,
351
+ "loss": 2.2227,
352
+ "step": 28000
353
+ },
354
+ {
355
+ "epoch": 1.94,
356
+ "learning_rate": 7.096949521027244e-06,
357
+ "loss": 2.2264,
358
+ "step": 28500
359
+ },
360
+ {
361
+ "epoch": 1.97,
362
+ "learning_rate": 6.870484860837466e-06,
363
+ "loss": 2.2649,
364
+ "step": 29000
365
+ },
366
+ {
367
+ "epoch": 2.0,
368
+ "eval_loss": 2.092484951019287,
369
+ "eval_runtime": 353.9215,
370
+ "eval_samples_per_second": 166.348,
371
+ "eval_steps_per_second": 10.398,
372
+ "step": 29438
373
+ },
374
+ {
375
+ "epoch": 2.0,
376
+ "learning_rate": 6.6440202006476895e-06,
377
+ "loss": 2.2415,
378
+ "step": 29500
379
+ },
380
+ {
381
+ "epoch": 2.04,
382
+ "learning_rate": 6.417555540457911e-06,
383
+ "loss": 2.2354,
384
+ "step": 30000
385
+ },
386
+ {
387
+ "epoch": 2.07,
388
+ "learning_rate": 6.191090880268135e-06,
389
+ "loss": 2.2296,
390
+ "step": 30500
391
+ },
392
+ {
393
+ "epoch": 2.11,
394
+ "learning_rate": 5.965079149398737e-06,
395
+ "loss": 2.241,
396
+ "step": 31000
397
+ },
398
+ {
399
+ "epoch": 2.14,
400
+ "learning_rate": 5.738614489208959e-06,
401
+ "loss": 2.2279,
402
+ "step": 31500
403
+ },
404
+ {
405
+ "epoch": 2.17,
406
+ "learning_rate": 5.512149829019182e-06,
407
+ "loss": 2.2405,
408
+ "step": 32000
409
+ },
410
+ {
411
+ "epoch": 2.21,
412
+ "learning_rate": 5.285685168829404e-06,
413
+ "loss": 2.2379,
414
+ "step": 32500
415
+ },
416
+ {
417
+ "epoch": 2.24,
418
+ "learning_rate": 5.059673437960007e-06,
419
+ "loss": 2.2386,
420
+ "step": 33000
421
+ },
422
+ {
423
+ "epoch": 2.28,
424
+ "learning_rate": 4.833208777770229e-06,
425
+ "loss": 2.2298,
426
+ "step": 33500
427
+ },
428
+ {
429
+ "epoch": 2.31,
430
+ "learning_rate": 4.6067441175804515e-06,
431
+ "loss": 2.2259,
432
+ "step": 34000
433
+ },
434
+ {
435
+ "epoch": 2.34,
436
+ "learning_rate": 4.380279457390674e-06,
437
+ "loss": 2.2537,
438
+ "step": 34500
439
+ },
440
+ {
441
+ "epoch": 2.38,
442
+ "learning_rate": 4.153814797200897e-06,
443
+ "loss": 2.2521,
444
+ "step": 35000
445
+ },
446
+ {
447
+ "epoch": 2.41,
448
+ "learning_rate": 3.927803066331499e-06,
449
+ "loss": 2.2049,
450
+ "step": 35500
451
+ },
452
+ {
453
+ "epoch": 2.45,
454
+ "learning_rate": 3.701338406141722e-06,
455
+ "loss": 2.2317,
456
+ "step": 36000
457
+ },
458
+ {
459
+ "epoch": 2.48,
460
+ "learning_rate": 3.4748737459519442e-06,
461
+ "loss": 2.2476,
462
+ "step": 36500
463
+ },
464
+ {
465
+ "epoch": 2.51,
466
+ "learning_rate": 3.248409085762167e-06,
467
+ "loss": 2.2435,
468
+ "step": 37000
469
+ },
470
+ {
471
+ "epoch": 2.55,
472
+ "learning_rate": 3.0219444255723896e-06,
473
+ "loss": 2.2523,
474
+ "step": 37500
475
+ },
476
+ {
477
+ "epoch": 2.58,
478
+ "learning_rate": 2.795479765382612e-06,
479
+ "loss": 2.2264,
480
+ "step": 38000
481
+ },
482
+ {
483
+ "epoch": 2.62,
484
+ "learning_rate": 2.5694680345132143e-06,
485
+ "loss": 2.2287,
486
+ "step": 38500
487
+ },
488
+ {
489
+ "epoch": 2.65,
490
+ "learning_rate": 2.343003374323437e-06,
491
+ "loss": 2.2238,
492
+ "step": 39000
493
+ },
494
+ {
495
+ "epoch": 2.68,
496
+ "learning_rate": 2.116991643454039e-06,
497
+ "loss": 2.2383,
498
+ "step": 39500
499
+ },
500
+ {
501
+ "epoch": 2.72,
502
+ "learning_rate": 1.8905269832642617e-06,
503
+ "loss": 2.234,
504
+ "step": 40000
505
+ },
506
+ {
507
+ "epoch": 2.75,
508
+ "learning_rate": 1.6640623230744844e-06,
509
+ "loss": 2.2298,
510
+ "step": 40500
511
+ },
512
+ {
513
+ "epoch": 2.79,
514
+ "learning_rate": 1.4375976628847069e-06,
515
+ "loss": 2.2237,
516
+ "step": 41000
517
+ },
518
+ {
519
+ "epoch": 2.82,
520
+ "learning_rate": 1.2111330026949296e-06,
521
+ "loss": 2.2489,
522
+ "step": 41500
523
+ },
524
+ {
525
+ "epoch": 2.85,
526
+ "learning_rate": 9.846683425051522e-07,
527
+ "loss": 2.2263,
528
+ "step": 42000
529
+ },
530
+ {
531
+ "epoch": 2.89,
532
+ "learning_rate": 7.582036823153747e-07,
533
+ "loss": 2.2223,
534
+ "step": 42500
535
+ },
536
+ {
537
+ "epoch": 2.92,
538
+ "learning_rate": 5.317390221255973e-07,
539
+ "loss": 2.2229,
540
+ "step": 43000
541
+ },
542
+ {
543
+ "epoch": 2.96,
544
+ "learning_rate": 3.0527436193581997e-07,
545
+ "loss": 2.2121,
546
+ "step": 43500
547
+ },
548
+ {
549
+ "epoch": 2.99,
550
+ "learning_rate": 7.880970174604252e-08,
551
+ "loss": 2.209,
552
+ "step": 44000
553
+ },
554
+ {
555
+ "epoch": 3.0,
556
+ "eval_loss": 2.086761713027954,
557
+ "eval_runtime": 355.3374,
558
+ "eval_samples_per_second": 165.685,
559
+ "eval_steps_per_second": 10.356,
560
+ "step": 44157
561
+ }
562
+ ],
563
+ "max_steps": 44157,
564
+ "num_train_epochs": 3,
565
+ "total_flos": 6.916093810478285e+16,
566
+ "trial_name": null,
567
+ "trial_params": null
568
+ }
checkpoint-44157/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19ba9d8516ad17e2c42a4b8c4e8c8f875bc32973b67b249cf10b10aaf8b8ab6e
3
+ size 3311
config.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "t5-small",
3
+ "architectures": [
4
+ "T5ForConditionalGeneration"
5
+ ],
6
+ "d_ff": 2048,
7
+ "d_kv": 64,
8
+ "d_model": 512,
9
+ "decoder_start_token_id": 0,
10
+ "dropout_rate": 0.1,
11
+ "eos_token_id": 1,
12
+ "feed_forward_proj": "relu",
13
+ "initializer_factor": 1.0,
14
+ "is_encoder_decoder": true,
15
+ "layer_norm_epsilon": 1e-06,
16
+ "model_type": "t5",
17
+ "n_positions": 512,
18
+ "num_decoder_layers": 6,
19
+ "num_heads": 8,
20
+ "num_layers": 6,
21
+ "output_past": true,
22
+ "pad_token_id": 0,
23
+ "relative_attention_max_distance": 128,
24
+ "relative_attention_num_buckets": 32,
25
+ "task_specific_params": {
26
+ "summarization": {
27
+ "early_stopping": true,
28
+ "length_penalty": 2.0,
29
+ "max_length": 200,
30
+ "min_length": 30,
31
+ "no_repeat_ngram_size": 3,
32
+ "num_beams": 4,
33
+ "prefix": "summarize: "
34
+ },
35
+ "translation_en_to_de": {
36
+ "early_stopping": true,
37
+ "max_length": 300,
38
+ "num_beams": 4,
39
+ "prefix": "translate English to German: "
40
+ },
41
+ "translation_en_to_fr": {
42
+ "early_stopping": true,
43
+ "max_length": 300,
44
+ "num_beams": 4,
45
+ "prefix": "translate English to French: "
46
+ },
47
+ "translation_en_to_ro": {
48
+ "early_stopping": true,
49
+ "max_length": 300,
50
+ "num_beams": 4,
51
+ "prefix": "translate English to Romanian: "
52
+ }
53
+ },
54
+ "torch_dtype": "float32",
55
+ "transformers_version": "4.19.4",
56
+ "use_cache": true,
57
+ "vocab_size": 32128
58
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2a080ed17900659a6f7c133c72c814002125f51c35ae6c7d423421367542a0e
3
+ size 242070267
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>", "additional_special_tokens": ["<extra_id_0>", "<extra_id_1>", "<extra_id_2>", "<extra_id_3>", "<extra_id_4>", "<extra_id_5>", "<extra_id_6>", "<extra_id_7>", "<extra_id_8>", "<extra_id_9>", "<extra_id_10>", "<extra_id_11>", "<extra_id_12>", "<extra_id_13>", "<extra_id_14>", "<extra_id_15>", "<extra_id_16>", "<extra_id_17>", "<extra_id_18>", "<extra_id_19>", "<extra_id_20>", "<extra_id_21>", "<extra_id_22>", "<extra_id_23>", "<extra_id_24>", "<extra_id_25>", "<extra_id_26>", "<extra_id_27>", "<extra_id_28>", "<extra_id_29>", "<extra_id_30>", "<extra_id_31>", "<extra_id_32>", "<extra_id_33>", "<extra_id_34>", "<extra_id_35>", "<extra_id_36>", "<extra_id_37>", "<extra_id_38>", "<extra_id_39>", "<extra_id_40>", "<extra_id_41>", "<extra_id_42>", "<extra_id_43>", "<extra_id_44>", "<extra_id_45>", "<extra_id_46>", "<extra_id_47>", "<extra_id_48>", "<extra_id_49>", "<extra_id_50>", "<extra_id_51>", "<extra_id_52>", "<extra_id_53>", "<extra_id_54>", "<extra_id_55>", "<extra_id_56>", "<extra_id_57>", "<extra_id_58>", "<extra_id_59>", "<extra_id_60>", "<extra_id_61>", "<extra_id_62>", "<extra_id_63>", "<extra_id_64>", "<extra_id_65>", "<extra_id_66>", "<extra_id_67>", "<extra_id_68>", "<extra_id_69>", "<extra_id_70>", "<extra_id_71>", "<extra_id_72>", "<extra_id_73>", "<extra_id_74>", "<extra_id_75>", "<extra_id_76>", "<extra_id_77>", "<extra_id_78>", "<extra_id_79>", "<extra_id_80>", "<extra_id_81>", "<extra_id_82>", "<extra_id_83>", "<extra_id_84>", "<extra_id_85>", "<extra_id_86>", "<extra_id_87>", "<extra_id_88>", "<extra_id_89>", "<extra_id_90>", "<extra_id_91>", "<extra_id_92>", "<extra_id_93>", "<extra_id_94>", "<extra_id_95>", "<extra_id_96>", "<extra_id_97>", "<extra_id_98>", "<extra_id_99>"]}
spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d60acb128cf7b7f2536e8f38a5b18a05535c9e14c7a355904270e15b0945ea86
3
+ size 791656
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>", "extra_ids": 100, "additional_special_tokens": ["<extra_id_0>", "<extra_id_1>", "<extra_id_2>", "<extra_id_3>", "<extra_id_4>", "<extra_id_5>", "<extra_id_6>", "<extra_id_7>", "<extra_id_8>", "<extra_id_9>", "<extra_id_10>", "<extra_id_11>", "<extra_id_12>", "<extra_id_13>", "<extra_id_14>", "<extra_id_15>", "<extra_id_16>", "<extra_id_17>", "<extra_id_18>", "<extra_id_19>", "<extra_id_20>", "<extra_id_21>", "<extra_id_22>", "<extra_id_23>", "<extra_id_24>", "<extra_id_25>", "<extra_id_26>", "<extra_id_27>", "<extra_id_28>", "<extra_id_29>", "<extra_id_30>", "<extra_id_31>", "<extra_id_32>", "<extra_id_33>", "<extra_id_34>", "<extra_id_35>", "<extra_id_36>", "<extra_id_37>", "<extra_id_38>", "<extra_id_39>", "<extra_id_40>", "<extra_id_41>", "<extra_id_42>", "<extra_id_43>", "<extra_id_44>", "<extra_id_45>", "<extra_id_46>", "<extra_id_47>", "<extra_id_48>", "<extra_id_49>", "<extra_id_50>", "<extra_id_51>", "<extra_id_52>", "<extra_id_53>", "<extra_id_54>", "<extra_id_55>", "<extra_id_56>", "<extra_id_57>", "<extra_id_58>", "<extra_id_59>", "<extra_id_60>", "<extra_id_61>", "<extra_id_62>", "<extra_id_63>", "<extra_id_64>", "<extra_id_65>", "<extra_id_66>", "<extra_id_67>", "<extra_id_68>", "<extra_id_69>", "<extra_id_70>", "<extra_id_71>", "<extra_id_72>", "<extra_id_73>", "<extra_id_74>", "<extra_id_75>", "<extra_id_76>", "<extra_id_77>", "<extra_id_78>", "<extra_id_79>", "<extra_id_80>", "<extra_id_81>", "<extra_id_82>", "<extra_id_83>", "<extra_id_84>", "<extra_id_85>", "<extra_id_86>", "<extra_id_87>", "<extra_id_88>", "<extra_id_89>", "<extra_id_90>", "<extra_id_91>", "<extra_id_92>", "<extra_id_93>", "<extra_id_94>", "<extra_id_95>", "<extra_id_96>", "<extra_id_97>", "<extra_id_98>", "<extra_id_99>"], "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "t5-small", "tokenizer_class": "T5Tokenizer"}
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:effd2643805aa0d145b33d51d879c48376bd40edd6a00105dad77efe1240a28b
3
+ size 3375