dongim04 commited on
Commit
078764c
1 Parent(s): ddb6198

End of training

Browse files
README.md CHANGED
@@ -18,8 +18,8 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [openai/whisper-base](https://huggingface.co/openai/whisper-base) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 0.1212
22
- - Wer: 3.6561
23
 
24
  ## Model description
25
 
@@ -42,60 +42,33 @@ The following hyperparameters were used during training:
42
  - train_batch_size: 16
43
  - eval_batch_size: 8
44
  - seed: 42
45
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: linear
47
- - training_steps: 4000
48
  - mixed_precision_training: Native AMP
49
 
50
  ### Training results
51
 
52
- | Training Loss | Epoch | Step | Validation Loss | Wer |
53
- |:-------------:|:-------:|:----:|:---------------:|:------:|
54
- | 0.1587 | 0.3676 | 100 | 0.1626 | 7.1105 |
55
- | 0.1464 | 0.7353 | 200 | 0.1325 | 5.7486 |
56
- | 0.0699 | 1.1029 | 300 | 0.1217 | 4.3894 |
57
- | 0.0714 | 1.4706 | 400 | 0.1147 | 4.2034 |
58
- | 0.0529 | 1.8382 | 500 | 0.1117 | 4.0358 |
59
- | 0.0315 | 2.2059 | 600 | 0.1087 | 3.8865 |
60
- | 0.0305 | 2.5735 | 700 | 0.1077 | 3.8787 |
61
- | 0.0307 | 2.9412 | 800 | 0.1031 | 3.5958 |
62
- | 0.0137 | 3.3088 | 900 | 0.1075 | 3.5304 |
63
- | 0.0125 | 3.6765 | 1000 | 0.1065 | 3.4858 |
64
- | 0.0103 | 4.0441 | 1100 | 0.1069 | 3.5592 |
65
- | 0.0066 | 4.4118 | 1200 | 0.1093 | 3.5539 |
66
- | 0.0063 | 4.7794 | 1300 | 0.1072 | 4.0332 |
67
- | 0.0043 | 5.1471 | 1400 | 0.1095 | 3.5880 |
68
- | 0.0045 | 5.5147 | 1500 | 0.1109 | 5.1672 |
69
- | 0.0048 | 5.8824 | 1600 | 0.1114 | 3.5723 |
70
- | 0.0035 | 6.25 | 1700 | 0.1128 | 3.5775 |
71
- | 0.0033 | 6.6176 | 1800 | 0.1117 | 4.6591 |
72
- | 0.0032 | 6.9853 | 1900 | 0.1132 | 3.5435 |
73
- | 0.0032 | 7.3529 | 2000 | 0.1138 | 3.5801 |
74
- | 0.0026 | 7.7206 | 2100 | 0.1151 | 3.6246 |
75
- | 0.0024 | 8.0882 | 2200 | 0.1155 | 3.6639 |
76
- | 0.0023 | 8.4559 | 2300 | 0.1167 | 3.6613 |
77
- | 0.0022 | 8.8235 | 2400 | 0.1176 | 3.6299 |
78
- | 0.0019 | 9.1912 | 2500 | 0.1177 | 3.5592 |
79
- | 0.0018 | 9.5588 | 2600 | 0.1169 | 3.5827 |
80
- | 0.0018 | 9.9265 | 2700 | 0.1175 | 3.5985 |
81
- | 0.0016 | 10.2941 | 2800 | 0.1183 | 3.6142 |
82
- | 0.0017 | 10.6618 | 2900 | 0.1190 | 3.6246 |
83
- | 0.0016 | 11.0294 | 3000 | 0.1184 | 3.6954 |
84
- | 0.0016 | 11.3971 | 3100 | 0.1192 | 3.6194 |
85
- | 0.0015 | 11.7647 | 3200 | 0.1197 | 3.6508 |
86
- | 0.0014 | 12.1324 | 3300 | 0.1202 | 3.6142 |
87
- | 0.0013 | 12.5 | 3400 | 0.1202 | 3.6194 |
88
- | 0.0014 | 12.8676 | 3500 | 0.1204 | 3.6561 |
89
- | 0.0013 | 13.2353 | 3600 | 0.1208 | 3.6351 |
90
- | 0.0014 | 13.6029 | 3700 | 0.1209 | 3.6561 |
91
- | 0.0013 | 13.9706 | 3800 | 0.1211 | 3.6456 |
92
- | 0.0014 | 14.3382 | 3900 | 0.1212 | 3.6613 |
93
- | 0.0013 | 14.7059 | 4000 | 0.1212 | 3.6561 |
94
 
95
 
96
  ### Framework versions
97
 
98
- - Transformers 4.44.2
99
- - Pytorch 2.5.0+cu121
100
  - Datasets 3.1.0
101
- - Tokenizers 0.19.1
 
18
 
19
  This model is a fine-tuned version of [openai/whisper-base](https://huggingface.co/openai/whisper-base) on the None dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.1142
22
+ - Wer: 3.7226
23
 
24
  ## Model description
25
 
 
42
  - train_batch_size: 16
43
  - eval_batch_size: 8
44
  - seed: 42
45
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
46
  - lr_scheduler_type: linear
47
+ - training_steps: 2000
48
  - mixed_precision_training: Native AMP
49
 
50
  ### Training results
51
 
52
+ | Training Loss | Epoch | Step | Validation Loss | Wer |
53
+ |:-------------:|:------:|:----:|:---------------:|:------:|
54
+ | 0.0667 | 0.3534 | 100 | 0.1267 | 4.8230 |
55
+ | 0.0693 | 0.7067 | 200 | 0.1188 | 4.1087 |
56
+ | 0.0614 | 1.0601 | 300 | 0.1133 | 3.7907 |
57
+ | 0.0274 | 1.4134 | 400 | 0.1128 | 3.7503 |
58
+ | 0.0365 | 1.7668 | 500 | 0.1113 | 3.8513 |
59
+ | 0.014 | 2.1201 | 600 | 0.1106 | 3.7201 |
60
+ | 0.0165 | 2.4735 | 700 | 0.1116 | 3.7352 |
61
+ | 0.0137 | 2.8269 | 800 | 0.1096 | 3.6999 |
62
+ | 0.007 | 3.1802 | 900 | 0.1114 | 3.6216 |
63
+ | 0.007 | 3.5336 | 1000 | 0.1130 | 3.6317 |
64
+ | 0.0077 | 3.8869 | 1100 | 0.1128 | 3.6645 |
65
+ | 0.0056 | 4.2403 | 1200 | 0.1138 | 3.7302 |
66
+ | 0.005 | 4.5936 | 1300 | 0.1142 | 3.7226 |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
67
 
68
 
69
  ### Framework versions
70
 
71
+ - Transformers 4.46.2
72
+ - Pytorch 2.5.1+cu121
73
  - Datasets 3.1.0
74
+ - Tokenizers 0.20.3
generation_config.json CHANGED
@@ -158,99 +158,10 @@
158
  "pad_token_id": 50257,
159
  "prev_sot_token_id": 50361,
160
  "return_timestamps": false,
161
- "suppress_tokens": [
162
- 1,
163
- 2,
164
- 7,
165
- 8,
166
- 9,
167
- 10,
168
- 14,
169
- 25,
170
- 26,
171
- 27,
172
- 28,
173
- 29,
174
- 31,
175
- 58,
176
- 59,
177
- 60,
178
- 61,
179
- 62,
180
- 63,
181
- 90,
182
- 91,
183
- 92,
184
- 93,
185
- 359,
186
- 503,
187
- 522,
188
- 542,
189
- 873,
190
- 893,
191
- 902,
192
- 918,
193
- 922,
194
- 931,
195
- 1350,
196
- 1853,
197
- 1982,
198
- 2460,
199
- 2627,
200
- 3246,
201
- 3253,
202
- 3268,
203
- 3536,
204
- 3846,
205
- 3961,
206
- 4183,
207
- 4667,
208
- 6585,
209
- 6647,
210
- 7273,
211
- 9061,
212
- 9383,
213
- 10428,
214
- 10929,
215
- 11938,
216
- 12033,
217
- 12331,
218
- 12562,
219
- 13793,
220
- 14157,
221
- 14635,
222
- 15265,
223
- 15618,
224
- 16553,
225
- 16604,
226
- 18362,
227
- 18956,
228
- 20075,
229
- 21675,
230
- 22520,
231
- 26130,
232
- 26161,
233
- 26435,
234
- 28279,
235
- 29464,
236
- 31650,
237
- 32302,
238
- 32470,
239
- 36865,
240
- 42863,
241
- 47425,
242
- 49870,
243
- 50254,
244
- 50258,
245
- 50358,
246
- 50359,
247
- 50360,
248
- 50361,
249
- 50362
250
- ],
251
  "task_to_id": {
252
  "transcribe": 50359,
253
  "translate": 50358
254
  },
255
- "transformers_version": "4.44.2"
256
  }
 
158
  "pad_token_id": 50257,
159
  "prev_sot_token_id": 50361,
160
  "return_timestamps": false,
161
+ "suppress_tokens": [],
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
162
  "task_to_id": {
163
  "transcribe": 50359,
164
  "translate": 50358
165
  },
166
+ "transformers_version": "4.46.2"
167
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c1ee7230035ecb1104f14b7a9de94ac2200734a71f7c234038347ff4edbe8f90
3
  size 290403936
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ffb3af42b837d1636b54074d6590583c9e44fe5e6aa8ff39ceb078c37cac371
3
  size 290403936
runs/Nov16_04-24-38_cf84fc18aa82/events.out.tfevents.1731731099.cf84fc18aa82.2855.2 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:adb9f8c1168d97baa5c0867e39ad1de554cb8675ba86098b9ec3e1dd21e6492b
3
- size 19702
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39f704db49d6f6128f6f54cc9aaff9b32cd1581925a78b0c4422b94e371991a5
3
+ size 21218