|
{ |
|
"best_metric": 1.0, |
|
"best_model_checkpoint": "../data/wav2vec2-large-xlsr-53-thai/checkpoint-1000", |
|
"epoch": 99.77827050997783, |
|
"global_step": 135000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 9.970000000000001e-05, |
|
"loss": 7.1725, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"eval_loss": 3.039940595626831, |
|
"eval_runtime": 185.7874, |
|
"eval_samples_per_second": 16.293, |
|
"eval_steps_per_second": 0.511, |
|
"eval_wer": 1.0, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 9.925763216679077e-05, |
|
"loss": 1.0788, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"eval_loss": 0.6130112409591675, |
|
"eval_runtime": 176.9194, |
|
"eval_samples_per_second": 17.109, |
|
"eval_steps_per_second": 0.537, |
|
"eval_wer": 0.6820256330103157, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 9.851303052866717e-05, |
|
"loss": 0.6339, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"eval_loss": 0.49624550342559814, |
|
"eval_runtime": 175.6715, |
|
"eval_samples_per_second": 17.231, |
|
"eval_steps_per_second": 0.541, |
|
"eval_wer": 0.6168802750859643, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 9.776842889054357e-05, |
|
"loss": 0.5295, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"eval_loss": 0.4389719069004059, |
|
"eval_runtime": 176.6072, |
|
"eval_samples_per_second": 17.14, |
|
"eval_steps_per_second": 0.538, |
|
"eval_wer": 0.577117849327915, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 3.7, |
|
"learning_rate": 9.702382725241996e-05, |
|
"loss": 0.4616, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 3.7, |
|
"eval_loss": 0.4208119809627533, |
|
"eval_runtime": 196.7481, |
|
"eval_samples_per_second": 15.385, |
|
"eval_steps_per_second": 0.483, |
|
"eval_wer": 0.5618005626758362, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 4.43, |
|
"learning_rate": 9.627922561429635e-05, |
|
"loss": 0.4206, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 4.43, |
|
"eval_loss": 0.4009915888309479, |
|
"eval_runtime": 180.4803, |
|
"eval_samples_per_second": 16.772, |
|
"eval_steps_per_second": 0.526, |
|
"eval_wer": 0.5411691153485464, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 5.17, |
|
"learning_rate": 9.553462397617276e-05, |
|
"loss": 0.3908, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 5.17, |
|
"eval_loss": 0.399889200925827, |
|
"eval_runtime": 189.7255, |
|
"eval_samples_per_second": 15.955, |
|
"eval_steps_per_second": 0.501, |
|
"eval_wer": 0.5384807752422632, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 5.91, |
|
"learning_rate": 9.479002233804914e-05, |
|
"loss": 0.3659, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 5.91, |
|
"eval_loss": 0.3719017505645752, |
|
"eval_runtime": 184.5994, |
|
"eval_samples_per_second": 16.398, |
|
"eval_steps_per_second": 0.515, |
|
"eval_wer": 0.5141606752110034, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 6.65, |
|
"learning_rate": 9.404542069992555e-05, |
|
"loss": 0.338, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 6.65, |
|
"eval_loss": 0.35533592104911804, |
|
"eval_runtime": 188.5325, |
|
"eval_samples_per_second": 16.056, |
|
"eval_steps_per_second": 0.504, |
|
"eval_wer": 0.5101594248202563, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 7.39, |
|
"learning_rate": 9.330156366344007e-05, |
|
"loss": 0.3222, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 7.39, |
|
"eval_loss": 0.40722423791885376, |
|
"eval_runtime": 195.2723, |
|
"eval_samples_per_second": 15.501, |
|
"eval_steps_per_second": 0.487, |
|
"eval_wer": 0.49309159112222567, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 8.13, |
|
"learning_rate": 9.255696202531646e-05, |
|
"loss": 0.31, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 8.13, |
|
"eval_loss": 0.37148234248161316, |
|
"eval_runtime": 182.6314, |
|
"eval_samples_per_second": 16.574, |
|
"eval_steps_per_second": 0.52, |
|
"eval_wer": 0.48715223507346045, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 8.87, |
|
"learning_rate": 9.181310498883098e-05, |
|
"loss": 0.2901, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 8.87, |
|
"eval_loss": 0.365583598613739, |
|
"eval_runtime": 175.3968, |
|
"eval_samples_per_second": 17.258, |
|
"eval_steps_per_second": 0.542, |
|
"eval_wer": 0.4882775867458581, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 9.61, |
|
"learning_rate": 9.106850335070738e-05, |
|
"loss": 0.276, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 9.61, |
|
"eval_loss": 0.37835174798965454, |
|
"eval_runtime": 175.9962, |
|
"eval_samples_per_second": 17.199, |
|
"eval_steps_per_second": 0.54, |
|
"eval_wer": 0.4897155361050328, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 10.35, |
|
"learning_rate": 9.032539091586003e-05, |
|
"loss": 0.2628, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 10.35, |
|
"eval_loss": 0.4127887189388275, |
|
"eval_runtime": 180.0512, |
|
"eval_samples_per_second": 16.812, |
|
"eval_steps_per_second": 0.528, |
|
"eval_wer": 0.48871522350734603, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 11.09, |
|
"learning_rate": 8.958078927773641e-05, |
|
"loss": 0.256, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 11.09, |
|
"eval_loss": 0.3795180022716522, |
|
"eval_runtime": 174.1363, |
|
"eval_samples_per_second": 17.383, |
|
"eval_steps_per_second": 0.546, |
|
"eval_wer": 0.46939668646452015, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 11.83, |
|
"learning_rate": 8.88361876396128e-05, |
|
"loss": 0.2455, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 11.83, |
|
"eval_loss": 0.3748489320278168, |
|
"eval_runtime": 174.8629, |
|
"eval_samples_per_second": 17.311, |
|
"eval_steps_per_second": 0.543, |
|
"eval_wer": 0.46970928415129726, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 12.56, |
|
"learning_rate": 8.809233060312732e-05, |
|
"loss": 0.2377, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 12.56, |
|
"eval_loss": 0.4371410012245178, |
|
"eval_runtime": 173.1913, |
|
"eval_samples_per_second": 17.478, |
|
"eval_steps_per_second": 0.549, |
|
"eval_wer": 0.4650203188496405, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 13.3, |
|
"learning_rate": 8.734772896500373e-05, |
|
"loss": 0.2279, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 13.3, |
|
"eval_loss": 0.38019847869873047, |
|
"eval_runtime": 177.0444, |
|
"eval_samples_per_second": 17.097, |
|
"eval_steps_per_second": 0.537, |
|
"eval_wer": 0.45339168490153176, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 14.04, |
|
"learning_rate": 8.660387192851825e-05, |
|
"loss": 0.2247, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 14.04, |
|
"eval_loss": 0.3687661588191986, |
|
"eval_runtime": 197.8618, |
|
"eval_samples_per_second": 15.299, |
|
"eval_steps_per_second": 0.48, |
|
"eval_wer": 0.4567677399187246, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 14.78, |
|
"learning_rate": 8.585927029039465e-05, |
|
"loss": 0.2139, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 14.78, |
|
"eval_loss": 0.3616667091846466, |
|
"eval_runtime": 177.8968, |
|
"eval_samples_per_second": 17.015, |
|
"eval_steps_per_second": 0.534, |
|
"eval_wer": 0.446639574867146, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 15.52, |
|
"learning_rate": 8.511541325390916e-05, |
|
"loss": 0.2072, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 15.52, |
|
"eval_loss": 0.4171736538410187, |
|
"eval_runtime": 171.8647, |
|
"eval_samples_per_second": 17.613, |
|
"eval_steps_per_second": 0.553, |
|
"eval_wer": 0.45883088465145355, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 16.26, |
|
"learning_rate": 8.437081161578556e-05, |
|
"loss": 0.1999, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 16.26, |
|
"eval_loss": 0.7755644917488098, |
|
"eval_runtime": 171.1132, |
|
"eval_samples_per_second": 17.69, |
|
"eval_steps_per_second": 0.555, |
|
"eval_wer": 0.45126602063144733, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"learning_rate": 8.362695457930007e-05, |
|
"loss": 0.1958, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"eval_loss": 0.41745948791503906, |
|
"eval_runtime": 172.9131, |
|
"eval_samples_per_second": 17.506, |
|
"eval_steps_per_second": 0.549, |
|
"eval_wer": 0.44382619568615195, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 17.74, |
|
"learning_rate": 8.288235294117647e-05, |
|
"loss": 0.1866, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 17.74, |
|
"eval_loss": 0.4865190386772156, |
|
"eval_runtime": 173.9742, |
|
"eval_samples_per_second": 17.399, |
|
"eval_steps_per_second": 0.546, |
|
"eval_wer": 0.45401688027508597, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 18.48, |
|
"learning_rate": 8.213924050632912e-05, |
|
"loss": 0.1854, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 18.48, |
|
"eval_loss": 0.5420244336128235, |
|
"eval_runtime": 175.6968, |
|
"eval_samples_per_second": 17.229, |
|
"eval_steps_per_second": 0.541, |
|
"eval_wer": 0.4475148483901219, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 19.22, |
|
"learning_rate": 8.139463886820552e-05, |
|
"loss": 0.1777, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 19.22, |
|
"eval_loss": 0.4428083300590515, |
|
"eval_runtime": 174.8946, |
|
"eval_samples_per_second": 17.308, |
|
"eval_steps_per_second": 0.543, |
|
"eval_wer": 0.4501406689590497, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 19.96, |
|
"learning_rate": 8.065003723008191e-05, |
|
"loss": 0.1764, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 19.96, |
|
"eval_loss": 0.44021618366241455, |
|
"eval_runtime": 174.1095, |
|
"eval_samples_per_second": 17.386, |
|
"eval_steps_per_second": 0.546, |
|
"eval_wer": 0.44307596123788684, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 20.69, |
|
"learning_rate": 7.990543559195831e-05, |
|
"loss": 0.1698, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 20.69, |
|
"eval_loss": 0.48001521825790405, |
|
"eval_runtime": 173.9598, |
|
"eval_samples_per_second": 17.401, |
|
"eval_steps_per_second": 0.546, |
|
"eval_wer": 0.4456392622694592, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 21.43, |
|
"learning_rate": 7.91608339538347e-05, |
|
"loss": 0.1625, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 21.43, |
|
"eval_loss": 0.4736146628856659, |
|
"eval_runtime": 173.7886, |
|
"eval_samples_per_second": 17.418, |
|
"eval_steps_per_second": 0.547, |
|
"eval_wer": 0.44676461394185685, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 22.17, |
|
"learning_rate": 7.841697691734922e-05, |
|
"loss": 0.1611, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 22.17, |
|
"eval_loss": 0.4669973850250244, |
|
"eval_runtime": 174.8827, |
|
"eval_samples_per_second": 17.309, |
|
"eval_steps_per_second": 0.543, |
|
"eval_wer": 0.43813691778680836, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 22.91, |
|
"learning_rate": 7.767237527922562e-05, |
|
"loss": 0.1557, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 22.91, |
|
"eval_loss": 0.49789130687713623, |
|
"eval_runtime": 176.2993, |
|
"eval_samples_per_second": 17.17, |
|
"eval_steps_per_second": 0.539, |
|
"eval_wer": 0.4390121913097843, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 23.65, |
|
"learning_rate": 7.692851824274014e-05, |
|
"loss": 0.1531, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 23.65, |
|
"eval_loss": 0.49486255645751953, |
|
"eval_runtime": 178.461, |
|
"eval_samples_per_second": 16.962, |
|
"eval_steps_per_second": 0.532, |
|
"eval_wer": 0.4311347296030009, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 24.39, |
|
"learning_rate": 7.618391660461653e-05, |
|
"loss": 0.1485, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 24.39, |
|
"eval_loss": 0.5052071809768677, |
|
"eval_runtime": 197.1526, |
|
"eval_samples_per_second": 15.354, |
|
"eval_steps_per_second": 0.482, |
|
"eval_wer": 0.43432322600812756, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 25.13, |
|
"learning_rate": 7.543931496649293e-05, |
|
"loss": 0.149, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 25.13, |
|
"eval_loss": 0.4628446400165558, |
|
"eval_runtime": 191.152, |
|
"eval_samples_per_second": 15.836, |
|
"eval_steps_per_second": 0.497, |
|
"eval_wer": 0.42638324476398876, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 25.87, |
|
"learning_rate": 7.469545793000745e-05, |
|
"loss": 0.1439, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 25.87, |
|
"eval_loss": 0.4972988963127136, |
|
"eval_runtime": 199.2887, |
|
"eval_samples_per_second": 15.189, |
|
"eval_steps_per_second": 0.477, |
|
"eval_wer": 0.4331353547983745, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 26.61, |
|
"learning_rate": 7.395160089352196e-05, |
|
"loss": 0.14, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 26.61, |
|
"eval_loss": 0.5113205313682556, |
|
"eval_runtime": 177.8533, |
|
"eval_samples_per_second": 17.02, |
|
"eval_steps_per_second": 0.534, |
|
"eval_wer": 0.4366364488902782, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 27.35, |
|
"learning_rate": 7.320699925539837e-05, |
|
"loss": 0.1373, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 27.35, |
|
"eval_loss": 0.5227766036987305, |
|
"eval_runtime": 197.1721, |
|
"eval_samples_per_second": 15.352, |
|
"eval_steps_per_second": 0.482, |
|
"eval_wer": 0.4340731478587058, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 28.09, |
|
"learning_rate": 7.246239761727476e-05, |
|
"loss": 0.134, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 28.09, |
|
"eval_loss": 0.5289518237113953, |
|
"eval_runtime": 186.9942, |
|
"eval_samples_per_second": 16.188, |
|
"eval_steps_per_second": 0.508, |
|
"eval_wer": 0.42744607689903097, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 28.82, |
|
"learning_rate": 7.171779597915115e-05, |
|
"loss": 0.1309, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 28.82, |
|
"eval_loss": 0.5079456567764282, |
|
"eval_runtime": 177.9839, |
|
"eval_samples_per_second": 17.007, |
|
"eval_steps_per_second": 0.534, |
|
"eval_wer": 0.4363238512035011, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 29.56, |
|
"learning_rate": 7.097319434102755e-05, |
|
"loss": 0.129, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 29.56, |
|
"eval_loss": 0.5366144776344299, |
|
"eval_runtime": 174.4769, |
|
"eval_samples_per_second": 17.349, |
|
"eval_steps_per_second": 0.544, |
|
"eval_wer": 0.4248202563301032, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 30.3, |
|
"learning_rate": 7.02300819061802e-05, |
|
"loss": 0.126, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 30.3, |
|
"eval_loss": 0.5051801204681396, |
|
"eval_runtime": 174.1717, |
|
"eval_samples_per_second": 17.379, |
|
"eval_steps_per_second": 0.545, |
|
"eval_wer": 0.42275711159737417, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 31.04, |
|
"learning_rate": 6.94854802680566e-05, |
|
"loss": 0.1255, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 31.04, |
|
"eval_loss": 0.5653140544891357, |
|
"eval_runtime": 175.3338, |
|
"eval_samples_per_second": 17.264, |
|
"eval_steps_per_second": 0.542, |
|
"eval_wer": 0.43063457330415755, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 31.78, |
|
"learning_rate": 6.874087862993299e-05, |
|
"loss": 0.122, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 31.78, |
|
"eval_loss": 0.5453576445579529, |
|
"eval_runtime": 176.4408, |
|
"eval_samples_per_second": 17.156, |
|
"eval_steps_per_second": 0.538, |
|
"eval_wer": 0.42800875273522976, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 32.52, |
|
"learning_rate": 6.799627699180938e-05, |
|
"loss": 0.1191, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 32.52, |
|
"eval_loss": 0.6267871260643005, |
|
"eval_runtime": 172.551, |
|
"eval_samples_per_second": 17.543, |
|
"eval_steps_per_second": 0.551, |
|
"eval_wer": 0.42400750234448265, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 33.26, |
|
"learning_rate": 6.725167535368579e-05, |
|
"loss": 0.1151, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 33.26, |
|
"eval_loss": 0.5381397008895874, |
|
"eval_runtime": 174.0505, |
|
"eval_samples_per_second": 17.392, |
|
"eval_steps_per_second": 0.546, |
|
"eval_wer": 0.4221944357611754, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 34.0, |
|
"learning_rate": 6.650707371556217e-05, |
|
"loss": 0.115, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 34.0, |
|
"eval_loss": 0.5648242235183716, |
|
"eval_runtime": 174.2088, |
|
"eval_samples_per_second": 17.376, |
|
"eval_steps_per_second": 0.545, |
|
"eval_wer": 0.4198812128790247, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 34.74, |
|
"learning_rate": 6.576247207743858e-05, |
|
"loss": 0.1124, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 34.74, |
|
"eval_loss": 0.613673985004425, |
|
"eval_runtime": 193.9522, |
|
"eval_samples_per_second": 15.607, |
|
"eval_steps_per_second": 0.49, |
|
"eval_wer": 0.4295092216317599, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 35.48, |
|
"learning_rate": 6.50186150409531e-05, |
|
"loss": 0.1101, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 35.48, |
|
"eval_loss": 0.5470528602600098, |
|
"eval_runtime": 193.2438, |
|
"eval_samples_per_second": 15.664, |
|
"eval_steps_per_second": 0.492, |
|
"eval_wer": 0.42206939668646454, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 36.22, |
|
"learning_rate": 6.427475800446761e-05, |
|
"loss": 0.1061, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 36.22, |
|
"eval_loss": 0.5942440032958984, |
|
"eval_runtime": 200.9573, |
|
"eval_samples_per_second": 15.063, |
|
"eval_steps_per_second": 0.473, |
|
"eval_wer": 0.4258205689277899, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 36.95, |
|
"learning_rate": 6.353015636634401e-05, |
|
"loss": 0.1065, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 36.95, |
|
"eval_loss": 0.5583751201629639, |
|
"eval_runtime": 178.0909, |
|
"eval_samples_per_second": 16.997, |
|
"eval_steps_per_second": 0.533, |
|
"eval_wer": 0.4255704907783682, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 37.69, |
|
"learning_rate": 6.27855547282204e-05, |
|
"loss": 0.1043, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 37.69, |
|
"eval_loss": 0.7426488399505615, |
|
"eval_runtime": 188.9691, |
|
"eval_samples_per_second": 16.018, |
|
"eval_steps_per_second": 0.503, |
|
"eval_wer": 0.4210065645514223, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 38.43, |
|
"learning_rate": 6.204169769173492e-05, |
|
"loss": 0.103, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 38.43, |
|
"eval_loss": 0.7028630375862122, |
|
"eval_runtime": 182.3453, |
|
"eval_samples_per_second": 16.6, |
|
"eval_steps_per_second": 0.521, |
|
"eval_wer": 0.42338230697092843, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 39.17, |
|
"learning_rate": 6.129709605361132e-05, |
|
"loss": 0.1016, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 39.17, |
|
"eval_loss": 0.644480288028717, |
|
"eval_runtime": 172.8589, |
|
"eval_samples_per_second": 17.511, |
|
"eval_steps_per_second": 0.55, |
|
"eval_wer": 0.42319474835886217, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 39.91, |
|
"learning_rate": 6.055323901712584e-05, |
|
"loss": 0.0977, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 39.91, |
|
"eval_loss": 0.6561093926429749, |
|
"eval_runtime": 174.7926, |
|
"eval_samples_per_second": 17.318, |
|
"eval_steps_per_second": 0.544, |
|
"eval_wer": 0.41394185683025947, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 40.65, |
|
"learning_rate": 5.980863737900224e-05, |
|
"loss": 0.097, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 40.65, |
|
"eval_loss": 0.6229739785194397, |
|
"eval_runtime": 176.8753, |
|
"eval_samples_per_second": 17.114, |
|
"eval_steps_per_second": 0.537, |
|
"eval_wer": 0.41450453266645826, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 41.39, |
|
"learning_rate": 5.9064780342516755e-05, |
|
"loss": 0.0963, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 41.39, |
|
"eval_loss": 0.5873845219612122, |
|
"eval_runtime": 175.9014, |
|
"eval_samples_per_second": 17.209, |
|
"eval_steps_per_second": 0.54, |
|
"eval_wer": 0.41713035323538605, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 42.13, |
|
"learning_rate": 5.832017870439316e-05, |
|
"loss": 0.0943, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 42.13, |
|
"eval_loss": 0.5930323600769043, |
|
"eval_runtime": 174.9181, |
|
"eval_samples_per_second": 17.305, |
|
"eval_steps_per_second": 0.543, |
|
"eval_wer": 0.418005626758362, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 42.87, |
|
"learning_rate": 5.757557706626955e-05, |
|
"loss": 0.093, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 42.87, |
|
"eval_loss": 0.6292551755905151, |
|
"eval_runtime": 175.0843, |
|
"eval_samples_per_second": 17.289, |
|
"eval_steps_per_second": 0.543, |
|
"eval_wer": 0.4175054704595186, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 43.61, |
|
"learning_rate": 5.683097542814595e-05, |
|
"loss": 0.0896, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 43.61, |
|
"eval_loss": 0.6171667575836182, |
|
"eval_runtime": 175.3688, |
|
"eval_samples_per_second": 17.261, |
|
"eval_steps_per_second": 0.542, |
|
"eval_wer": 0.41319162238199436, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 44.35, |
|
"learning_rate": 5.608637379002234e-05, |
|
"loss": 0.0911, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 44.35, |
|
"eval_loss": 0.6143432259559631, |
|
"eval_runtime": 178.3103, |
|
"eval_samples_per_second": 16.976, |
|
"eval_steps_per_second": 0.533, |
|
"eval_wer": 0.4172553923100969, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 45.08, |
|
"learning_rate": 5.534326135517498e-05, |
|
"loss": 0.0885, |
|
"step": 61000 |
|
}, |
|
{ |
|
"epoch": 45.08, |
|
"eval_loss": 0.6145235300064087, |
|
"eval_runtime": 191.9563, |
|
"eval_samples_per_second": 15.769, |
|
"eval_steps_per_second": 0.495, |
|
"eval_wer": 0.4133166614567052, |
|
"step": 61000 |
|
}, |
|
{ |
|
"epoch": 45.82, |
|
"learning_rate": 5.459865971705138e-05, |
|
"loss": 0.0872, |
|
"step": 62000 |
|
}, |
|
{ |
|
"epoch": 45.82, |
|
"eval_loss": 0.8499487638473511, |
|
"eval_runtime": 177.455, |
|
"eval_samples_per_second": 17.058, |
|
"eval_steps_per_second": 0.535, |
|
"eval_wer": 0.41187871209753046, |
|
"step": 62000 |
|
}, |
|
{ |
|
"epoch": 46.56, |
|
"learning_rate": 5.385405807892777e-05, |
|
"loss": 0.0883, |
|
"step": 63000 |
|
}, |
|
{ |
|
"epoch": 46.56, |
|
"eval_loss": 0.7073235511779785, |
|
"eval_runtime": 199.205, |
|
"eval_samples_per_second": 15.195, |
|
"eval_steps_per_second": 0.477, |
|
"eval_wer": 0.4190684588934042, |
|
"step": 63000 |
|
}, |
|
{ |
|
"epoch": 47.3, |
|
"learning_rate": 5.3109456440804175e-05, |
|
"loss": 0.0844, |
|
"step": 64000 |
|
}, |
|
{ |
|
"epoch": 47.3, |
|
"eval_loss": 0.6707965135574341, |
|
"eval_runtime": 203.0283, |
|
"eval_samples_per_second": 14.909, |
|
"eval_steps_per_second": 0.468, |
|
"eval_wer": 0.4145670522038137, |
|
"step": 64000 |
|
}, |
|
{ |
|
"epoch": 48.04, |
|
"learning_rate": 5.2365599404318685e-05, |
|
"loss": 0.0851, |
|
"step": 65000 |
|
}, |
|
{ |
|
"epoch": 48.04, |
|
"eval_loss": 0.6518837213516235, |
|
"eval_runtime": 184.8621, |
|
"eval_samples_per_second": 16.374, |
|
"eval_steps_per_second": 0.514, |
|
"eval_wer": 0.4101281650515786, |
|
"step": 65000 |
|
}, |
|
{ |
|
"epoch": 48.78, |
|
"learning_rate": 5.162099776619509e-05, |
|
"loss": 0.0832, |
|
"step": 66000 |
|
}, |
|
{ |
|
"epoch": 48.78, |
|
"eval_loss": 0.6594052910804749, |
|
"eval_runtime": 196.4902, |
|
"eval_samples_per_second": 15.405, |
|
"eval_steps_per_second": 0.483, |
|
"eval_wer": 0.41387933729290405, |
|
"step": 66000 |
|
}, |
|
{ |
|
"epoch": 49.52, |
|
"learning_rate": 5.0877140729709605e-05, |
|
"loss": 0.0822, |
|
"step": 67000 |
|
}, |
|
{ |
|
"epoch": 49.52, |
|
"eval_loss": 0.653843104839325, |
|
"eval_runtime": 186.7472, |
|
"eval_samples_per_second": 16.209, |
|
"eval_steps_per_second": 0.509, |
|
"eval_wer": 0.415942482025633, |
|
"step": 67000 |
|
}, |
|
{ |
|
"epoch": 50.26, |
|
"learning_rate": 5.013253909158601e-05, |
|
"loss": 0.0833, |
|
"step": 68000 |
|
}, |
|
{ |
|
"epoch": 50.26, |
|
"eval_loss": 0.6541770100593567, |
|
"eval_runtime": 175.2691, |
|
"eval_samples_per_second": 17.271, |
|
"eval_steps_per_second": 0.542, |
|
"eval_wer": 0.41844326351984995, |
|
"step": 68000 |
|
}, |
|
{ |
|
"epoch": 51.0, |
|
"learning_rate": 4.9387937453462405e-05, |
|
"loss": 0.079, |
|
"step": 69000 |
|
}, |
|
{ |
|
"epoch": 51.0, |
|
"eval_loss": 0.6602481007575989, |
|
"eval_runtime": 176.6295, |
|
"eval_samples_per_second": 17.138, |
|
"eval_steps_per_second": 0.538, |
|
"eval_wer": 0.41212879024695215, |
|
"step": 69000 |
|
}, |
|
{ |
|
"epoch": 51.74, |
|
"learning_rate": 4.8644080416976914e-05, |
|
"loss": 0.0768, |
|
"step": 70000 |
|
}, |
|
{ |
|
"epoch": 51.74, |
|
"eval_loss": 0.6583812832832336, |
|
"eval_runtime": 175.3489, |
|
"eval_samples_per_second": 17.263, |
|
"eval_steps_per_second": 0.542, |
|
"eval_wer": 0.41556736480150047, |
|
"step": 70000 |
|
}, |
|
{ |
|
"epoch": 52.48, |
|
"learning_rate": 4.790022338049144e-05, |
|
"loss": 0.0775, |
|
"step": 71000 |
|
}, |
|
{ |
|
"epoch": 52.48, |
|
"eval_loss": 0.6354698538780212, |
|
"eval_runtime": 173.883, |
|
"eval_samples_per_second": 17.408, |
|
"eval_steps_per_second": 0.546, |
|
"eval_wer": 0.4156924038762113, |
|
"step": 71000 |
|
}, |
|
{ |
|
"epoch": 53.22, |
|
"learning_rate": 4.7155621742367834e-05, |
|
"loss": 0.0762, |
|
"step": 72000 |
|
}, |
|
{ |
|
"epoch": 53.22, |
|
"eval_loss": 0.6745539307594299, |
|
"eval_runtime": 174.1736, |
|
"eval_samples_per_second": 17.379, |
|
"eval_steps_per_second": 0.545, |
|
"eval_wer": 0.40675211003438577, |
|
"step": 72000 |
|
}, |
|
{ |
|
"epoch": 53.95, |
|
"learning_rate": 4.641176470588236e-05, |
|
"loss": 0.0761, |
|
"step": 73000 |
|
}, |
|
{ |
|
"epoch": 53.95, |
|
"eval_loss": 0.6509293913841248, |
|
"eval_runtime": 176.7298, |
|
"eval_samples_per_second": 17.128, |
|
"eval_steps_per_second": 0.538, |
|
"eval_wer": 0.4095029696780244, |
|
"step": 73000 |
|
}, |
|
{ |
|
"epoch": 54.69, |
|
"learning_rate": 4.566716306775875e-05, |
|
"loss": 0.0742, |
|
"step": 74000 |
|
}, |
|
{ |
|
"epoch": 54.69, |
|
"eval_loss": 0.6646940112113953, |
|
"eval_runtime": 177.5457, |
|
"eval_samples_per_second": 17.049, |
|
"eval_steps_per_second": 0.535, |
|
"eval_wer": 0.41312910284463894, |
|
"step": 74000 |
|
}, |
|
{ |
|
"epoch": 55.43, |
|
"learning_rate": 4.492330603127327e-05, |
|
"loss": 0.0738, |
|
"step": 75000 |
|
}, |
|
{ |
|
"epoch": 55.43, |
|
"eval_loss": 0.6746073365211487, |
|
"eval_runtime": 182.277, |
|
"eval_samples_per_second": 16.607, |
|
"eval_steps_per_second": 0.521, |
|
"eval_wer": 0.4097530478274461, |
|
"step": 75000 |
|
}, |
|
{ |
|
"epoch": 56.17, |
|
"learning_rate": 4.417870439314967e-05, |
|
"loss": 0.072, |
|
"step": 76000 |
|
}, |
|
{ |
|
"epoch": 56.17, |
|
"eval_loss": 0.6681593060493469, |
|
"eval_runtime": 189.5798, |
|
"eval_samples_per_second": 15.967, |
|
"eval_steps_per_second": 0.501, |
|
"eval_wer": 0.41081587996248825, |
|
"step": 76000 |
|
}, |
|
{ |
|
"epoch": 56.91, |
|
"learning_rate": 4.3434102755026064e-05, |
|
"loss": 0.0715, |
|
"step": 77000 |
|
}, |
|
{ |
|
"epoch": 56.91, |
|
"eval_loss": 0.6641326546669006, |
|
"eval_runtime": 199.1459, |
|
"eval_samples_per_second": 15.2, |
|
"eval_steps_per_second": 0.477, |
|
"eval_wer": 0.4114410753360425, |
|
"step": 77000 |
|
}, |
|
{ |
|
"epoch": 57.65, |
|
"learning_rate": 4.2690245718540587e-05, |
|
"loss": 0.0717, |
|
"step": 78000 |
|
}, |
|
{ |
|
"epoch": 57.65, |
|
"eval_loss": 0.6959729790687561, |
|
"eval_runtime": 198.8176, |
|
"eval_samples_per_second": 15.225, |
|
"eval_steps_per_second": 0.478, |
|
"eval_wer": 0.41106595811191, |
|
"step": 78000 |
|
}, |
|
{ |
|
"epoch": 58.39, |
|
"learning_rate": 4.194564408041698e-05, |
|
"loss": 0.0711, |
|
"step": 79000 |
|
}, |
|
{ |
|
"epoch": 58.39, |
|
"eval_loss": 0.663927435874939, |
|
"eval_runtime": 201.2001, |
|
"eval_samples_per_second": 15.045, |
|
"eval_steps_per_second": 0.472, |
|
"eval_wer": 0.4081900593935605, |
|
"step": 79000 |
|
}, |
|
{ |
|
"epoch": 59.13, |
|
"learning_rate": 4.1201042442293374e-05, |
|
"loss": 0.0706, |
|
"step": 80000 |
|
}, |
|
{ |
|
"epoch": 59.13, |
|
"eval_loss": 0.6455733180046082, |
|
"eval_runtime": 180.6704, |
|
"eval_samples_per_second": 16.754, |
|
"eval_steps_per_second": 0.526, |
|
"eval_wer": 0.41156611441075336, |
|
"step": 80000 |
|
}, |
|
{ |
|
"epoch": 59.87, |
|
"learning_rate": 4.045644080416977e-05, |
|
"loss": 0.0685, |
|
"step": 81000 |
|
}, |
|
{ |
|
"epoch": 59.87, |
|
"eval_loss": 0.6929445862770081, |
|
"eval_runtime": 176.1838, |
|
"eval_samples_per_second": 17.181, |
|
"eval_steps_per_second": 0.539, |
|
"eval_wer": 0.4169427946233198, |
|
"step": 81000 |
|
}, |
|
{ |
|
"epoch": 60.61, |
|
"learning_rate": 3.971258376768429e-05, |
|
"loss": 0.0666, |
|
"step": 82000 |
|
}, |
|
{ |
|
"epoch": 60.61, |
|
"eval_loss": 0.6526830792427063, |
|
"eval_runtime": 175.1162, |
|
"eval_samples_per_second": 17.286, |
|
"eval_steps_per_second": 0.542, |
|
"eval_wer": 0.40987808690215694, |
|
"step": 82000 |
|
}, |
|
{ |
|
"epoch": 61.35, |
|
"learning_rate": 3.896798212956069e-05, |
|
"loss": 0.0677, |
|
"step": 83000 |
|
}, |
|
{ |
|
"epoch": 61.35, |
|
"eval_loss": 0.708766758441925, |
|
"eval_runtime": 174.527, |
|
"eval_samples_per_second": 17.344, |
|
"eval_steps_per_second": 0.544, |
|
"eval_wer": 0.4113785557986871, |
|
"step": 83000 |
|
}, |
|
{ |
|
"epoch": 62.08, |
|
"learning_rate": 3.822338049143708e-05, |
|
"loss": 0.0667, |
|
"step": 84000 |
|
}, |
|
{ |
|
"epoch": 62.08, |
|
"eval_loss": 0.6928330063819885, |
|
"eval_runtime": 178.7497, |
|
"eval_samples_per_second": 16.934, |
|
"eval_steps_per_second": 0.531, |
|
"eval_wer": 0.4085026570803376, |
|
"step": 84000 |
|
}, |
|
{ |
|
"epoch": 62.82, |
|
"learning_rate": 3.74795234549516e-05, |
|
"loss": 0.065, |
|
"step": 85000 |
|
}, |
|
{ |
|
"epoch": 62.82, |
|
"eval_loss": 0.7296735048294067, |
|
"eval_runtime": 174.5663, |
|
"eval_samples_per_second": 17.34, |
|
"eval_steps_per_second": 0.544, |
|
"eval_wer": 0.40962800875273525, |
|
"step": 85000 |
|
}, |
|
{ |
|
"epoch": 63.56, |
|
"learning_rate": 3.6734921816828e-05, |
|
"loss": 0.0655, |
|
"step": 86000 |
|
}, |
|
{ |
|
"epoch": 63.56, |
|
"eval_loss": 0.7482656836509705, |
|
"eval_runtime": 173.9987, |
|
"eval_samples_per_second": 17.397, |
|
"eval_steps_per_second": 0.546, |
|
"eval_wer": 0.4106908408877774, |
|
"step": 86000 |
|
}, |
|
{ |
|
"epoch": 64.3, |
|
"learning_rate": 3.59903201787044e-05, |
|
"loss": 0.0636, |
|
"step": 87000 |
|
}, |
|
{ |
|
"epoch": 64.3, |
|
"eval_loss": 0.7502027750015259, |
|
"eval_runtime": 172.4262, |
|
"eval_samples_per_second": 17.555, |
|
"eval_steps_per_second": 0.551, |
|
"eval_wer": 0.41875586120662706, |
|
"step": 87000 |
|
}, |
|
{ |
|
"epoch": 65.04, |
|
"learning_rate": 3.5245718540580794e-05, |
|
"loss": 0.0645, |
|
"step": 88000 |
|
}, |
|
{ |
|
"epoch": 65.04, |
|
"eval_loss": 0.7395268678665161, |
|
"eval_runtime": 172.5532, |
|
"eval_samples_per_second": 17.542, |
|
"eval_steps_per_second": 0.551, |
|
"eval_wer": 0.40793998124413877, |
|
"step": 88000 |
|
}, |
|
{ |
|
"epoch": 65.78, |
|
"learning_rate": 3.450186150409531e-05, |
|
"loss": 0.0628, |
|
"step": 89000 |
|
}, |
|
{ |
|
"epoch": 65.78, |
|
"eval_loss": 0.6981734037399292, |
|
"eval_runtime": 186.2785, |
|
"eval_samples_per_second": 16.25, |
|
"eval_steps_per_second": 0.51, |
|
"eval_wer": 0.4075023444826508, |
|
"step": 89000 |
|
}, |
|
{ |
|
"epoch": 66.52, |
|
"learning_rate": 3.375725986597171e-05, |
|
"loss": 0.0618, |
|
"step": 90000 |
|
}, |
|
{ |
|
"epoch": 66.52, |
|
"eval_loss": 0.7291818261146545, |
|
"eval_runtime": 195.0277, |
|
"eval_samples_per_second": 15.521, |
|
"eval_steps_per_second": 0.487, |
|
"eval_wer": 0.4113160362613317, |
|
"step": 90000 |
|
}, |
|
{ |
|
"epoch": 67.26, |
|
"learning_rate": 3.301340282948623e-05, |
|
"loss": 0.0612, |
|
"step": 91000 |
|
}, |
|
{ |
|
"epoch": 67.26, |
|
"eval_loss": 0.7428854703903198, |
|
"eval_runtime": 201.9566, |
|
"eval_samples_per_second": 14.988, |
|
"eval_steps_per_second": 0.47, |
|
"eval_wer": 0.40994060643951236, |
|
"step": 91000 |
|
}, |
|
{ |
|
"epoch": 68.0, |
|
"learning_rate": 3.2269545793000746e-05, |
|
"loss": 0.0613, |
|
"step": 92000 |
|
}, |
|
{ |
|
"epoch": 68.0, |
|
"eval_loss": 0.7314553260803223, |
|
"eval_runtime": 181.5248, |
|
"eval_samples_per_second": 16.675, |
|
"eval_steps_per_second": 0.523, |
|
"eval_wer": 0.41150359487339794, |
|
"step": 92000 |
|
}, |
|
{ |
|
"epoch": 68.74, |
|
"learning_rate": 3.152494415487714e-05, |
|
"loss": 0.0607, |
|
"step": 93000 |
|
}, |
|
{ |
|
"epoch": 68.74, |
|
"eval_loss": 0.7307026386260986, |
|
"eval_runtime": 263.2133, |
|
"eval_samples_per_second": 11.5, |
|
"eval_steps_per_second": 0.361, |
|
"eval_wer": 0.40725226633322914, |
|
"step": 93000 |
|
}, |
|
{ |
|
"epoch": 69.48, |
|
"learning_rate": 3.078034251675354e-05, |
|
"loss": 0.0597, |
|
"step": 94000 |
|
}, |
|
{ |
|
"epoch": 69.48, |
|
"eval_loss": 0.6700393557548523, |
|
"eval_runtime": 261.1836, |
|
"eval_samples_per_second": 11.59, |
|
"eval_steps_per_second": 0.364, |
|
"eval_wer": 0.40206314473272897, |
|
"step": 94000 |
|
}, |
|
{ |
|
"epoch": 70.21, |
|
"learning_rate": 3.0035740878629936e-05, |
|
"loss": 0.0586, |
|
"step": 95000 |
|
}, |
|
{ |
|
"epoch": 70.21, |
|
"eval_loss": 0.7129278779029846, |
|
"eval_runtime": 174.8869, |
|
"eval_samples_per_second": 17.308, |
|
"eval_steps_per_second": 0.543, |
|
"eval_wer": 0.40737730540794, |
|
"step": 95000 |
|
}, |
|
{ |
|
"epoch": 70.95, |
|
"learning_rate": 2.9291883842144452e-05, |
|
"loss": 0.0585, |
|
"step": 96000 |
|
}, |
|
{ |
|
"epoch": 70.95, |
|
"eval_loss": 0.7538037896156311, |
|
"eval_runtime": 175.7698, |
|
"eval_samples_per_second": 17.221, |
|
"eval_steps_per_second": 0.54, |
|
"eval_wer": 0.4142544545170366, |
|
"step": 96000 |
|
}, |
|
{ |
|
"epoch": 71.69, |
|
"learning_rate": 2.8547282204020846e-05, |
|
"loss": 0.0573, |
|
"step": 97000 |
|
}, |
|
{ |
|
"epoch": 71.69, |
|
"eval_loss": 0.703994631767273, |
|
"eval_runtime": 176.9961, |
|
"eval_samples_per_second": 17.102, |
|
"eval_steps_per_second": 0.537, |
|
"eval_wer": 0.40487652391372303, |
|
"step": 97000 |
|
}, |
|
{ |
|
"epoch": 72.43, |
|
"learning_rate": 2.780342516753537e-05, |
|
"loss": 0.0575, |
|
"step": 98000 |
|
}, |
|
{ |
|
"epoch": 72.43, |
|
"eval_loss": 0.7418561577796936, |
|
"eval_runtime": 177.6888, |
|
"eval_samples_per_second": 17.035, |
|
"eval_steps_per_second": 0.535, |
|
"eval_wer": 0.4070647077211629, |
|
"step": 98000 |
|
}, |
|
{ |
|
"epoch": 73.17, |
|
"learning_rate": 2.7058823529411766e-05, |
|
"loss": 0.0567, |
|
"step": 99000 |
|
}, |
|
{ |
|
"epoch": 73.17, |
|
"eval_loss": 0.7433472871780396, |
|
"eval_runtime": 175.8142, |
|
"eval_samples_per_second": 17.217, |
|
"eval_steps_per_second": 0.54, |
|
"eval_wer": 0.4093154110659581, |
|
"step": 99000 |
|
}, |
|
{ |
|
"epoch": 73.91, |
|
"learning_rate": 2.631422189128816e-05, |
|
"loss": 0.054, |
|
"step": 100000 |
|
}, |
|
{ |
|
"epoch": 73.91, |
|
"eval_loss": 0.6921048760414124, |
|
"eval_runtime": 180.3464, |
|
"eval_samples_per_second": 16.784, |
|
"eval_steps_per_second": 0.527, |
|
"eval_wer": 0.40256330103157234, |
|
"step": 100000 |
|
}, |
|
{ |
|
"epoch": 74.65, |
|
"learning_rate": 2.5570364854802682e-05, |
|
"loss": 0.0571, |
|
"step": 101000 |
|
}, |
|
{ |
|
"epoch": 74.65, |
|
"eval_loss": 0.7526727318763733, |
|
"eval_runtime": 176.5656, |
|
"eval_samples_per_second": 17.144, |
|
"eval_steps_per_second": 0.538, |
|
"eval_wer": 0.4104407627383557, |
|
"step": 101000 |
|
}, |
|
{ |
|
"epoch": 75.39, |
|
"learning_rate": 2.48265078183172e-05, |
|
"loss": 0.0532, |
|
"step": 102000 |
|
}, |
|
{ |
|
"epoch": 75.39, |
|
"eval_loss": 0.717917263507843, |
|
"eval_runtime": 180.7508, |
|
"eval_samples_per_second": 16.747, |
|
"eval_steps_per_second": 0.526, |
|
"eval_wer": 0.40693966864645204, |
|
"step": 102000 |
|
}, |
|
{ |
|
"epoch": 76.13, |
|
"learning_rate": 2.40819061801936e-05, |
|
"loss": 0.0549, |
|
"step": 103000 |
|
}, |
|
{ |
|
"epoch": 76.13, |
|
"eval_loss": 0.7699146866798401, |
|
"eval_runtime": 199.8758, |
|
"eval_samples_per_second": 15.144, |
|
"eval_steps_per_second": 0.475, |
|
"eval_wer": 0.4080650203188496, |
|
"step": 103000 |
|
}, |
|
{ |
|
"epoch": 76.87, |
|
"learning_rate": 2.3338049143708118e-05, |
|
"loss": 0.0533, |
|
"step": 104000 |
|
}, |
|
{ |
|
"epoch": 76.87, |
|
"eval_loss": 0.7196021676063538, |
|
"eval_runtime": 186.8201, |
|
"eval_samples_per_second": 16.203, |
|
"eval_steps_per_second": 0.509, |
|
"eval_wer": 0.4058143169740544, |
|
"step": 104000 |
|
}, |
|
{ |
|
"epoch": 77.61, |
|
"learning_rate": 2.2593447505584515e-05, |
|
"loss": 0.0538, |
|
"step": 105000 |
|
}, |
|
{ |
|
"epoch": 77.61, |
|
"eval_loss": 0.7532729506492615, |
|
"eval_runtime": 186.7527, |
|
"eval_samples_per_second": 16.209, |
|
"eval_steps_per_second": 0.509, |
|
"eval_wer": 0.4068146295717412, |
|
"step": 105000 |
|
}, |
|
{ |
|
"epoch": 78.34, |
|
"learning_rate": 2.1848845867460908e-05, |
|
"loss": 0.0533, |
|
"step": 106000 |
|
}, |
|
{ |
|
"epoch": 78.34, |
|
"eval_loss": 0.7514676451683044, |
|
"eval_runtime": 203.9001, |
|
"eval_samples_per_second": 14.846, |
|
"eval_steps_per_second": 0.466, |
|
"eval_wer": 0.40762738355736167, |
|
"step": 106000 |
|
}, |
|
{ |
|
"epoch": 79.08, |
|
"learning_rate": 2.1104244229337305e-05, |
|
"loss": 0.0517, |
|
"step": 107000 |
|
}, |
|
{ |
|
"epoch": 79.08, |
|
"eval_loss": 0.766864001750946, |
|
"eval_runtime": 183.9434, |
|
"eval_samples_per_second": 16.456, |
|
"eval_steps_per_second": 0.516, |
|
"eval_wer": 0.4061894341981869, |
|
"step": 107000 |
|
}, |
|
{ |
|
"epoch": 79.82, |
|
"learning_rate": 2.0359642591213702e-05, |
|
"loss": 0.0521, |
|
"step": 108000 |
|
}, |
|
{ |
|
"epoch": 79.82, |
|
"eval_loss": 0.7399023771286011, |
|
"eval_runtime": 175.691, |
|
"eval_samples_per_second": 17.229, |
|
"eval_steps_per_second": 0.541, |
|
"eval_wer": 0.4015004688965302, |
|
"step": 108000 |
|
}, |
|
{ |
|
"epoch": 80.56, |
|
"learning_rate": 1.961578555472822e-05, |
|
"loss": 0.0525, |
|
"step": 109000 |
|
}, |
|
{ |
|
"epoch": 80.56, |
|
"eval_loss": 0.7517969608306885, |
|
"eval_runtime": 185.511, |
|
"eval_samples_per_second": 16.317, |
|
"eval_steps_per_second": 0.512, |
|
"eval_wer": 0.4044388871522351, |
|
"step": 109000 |
|
}, |
|
{ |
|
"epoch": 81.3, |
|
"learning_rate": 1.887118391660462e-05, |
|
"loss": 0.0505, |
|
"step": 110000 |
|
}, |
|
{ |
|
"epoch": 81.3, |
|
"eval_loss": 0.7839346528053284, |
|
"eval_runtime": 181.6126, |
|
"eval_samples_per_second": 16.667, |
|
"eval_steps_per_second": 0.523, |
|
"eval_wer": 0.4077524226320725, |
|
"step": 110000 |
|
}, |
|
{ |
|
"epoch": 82.04, |
|
"learning_rate": 1.8126582278481012e-05, |
|
"loss": 0.0508, |
|
"step": 111000 |
|
}, |
|
{ |
|
"epoch": 82.04, |
|
"eval_loss": 0.7692478895187378, |
|
"eval_runtime": 182.2532, |
|
"eval_samples_per_second": 16.609, |
|
"eval_steps_per_second": 0.521, |
|
"eval_wer": 0.4070647077211629, |
|
"step": 111000 |
|
}, |
|
{ |
|
"epoch": 82.78, |
|
"learning_rate": 1.7383469843633657e-05, |
|
"loss": 0.0496, |
|
"step": 112000 |
|
}, |
|
{ |
|
"epoch": 82.78, |
|
"eval_loss": 0.7821455597877502, |
|
"eval_runtime": 227.3787, |
|
"eval_samples_per_second": 13.313, |
|
"eval_steps_per_second": 0.418, |
|
"eval_wer": 0.40987808690215694, |
|
"step": 112000 |
|
}, |
|
{ |
|
"epoch": 83.52, |
|
"learning_rate": 1.663886820551005e-05, |
|
"loss": 0.0493, |
|
"step": 113000 |
|
}, |
|
{ |
|
"epoch": 83.52, |
|
"eval_loss": 0.7528874278068542, |
|
"eval_runtime": 217.8709, |
|
"eval_samples_per_second": 13.894, |
|
"eval_steps_per_second": 0.436, |
|
"eval_wer": 0.4055642388246327, |
|
"step": 113000 |
|
}, |
|
{ |
|
"epoch": 84.26, |
|
"learning_rate": 1.5894266567386448e-05, |
|
"loss": 0.0498, |
|
"step": 114000 |
|
}, |
|
{ |
|
"epoch": 84.26, |
|
"eval_loss": 0.7717328071594238, |
|
"eval_runtime": 216.4941, |
|
"eval_samples_per_second": 13.982, |
|
"eval_steps_per_second": 0.439, |
|
"eval_wer": 0.40493904345107845, |
|
"step": 114000 |
|
}, |
|
{ |
|
"epoch": 85.0, |
|
"learning_rate": 1.5149664929262845e-05, |
|
"loss": 0.0499, |
|
"step": 115000 |
|
}, |
|
{ |
|
"epoch": 85.0, |
|
"eval_loss": 0.757793128490448, |
|
"eval_runtime": 213.4299, |
|
"eval_samples_per_second": 14.183, |
|
"eval_steps_per_second": 0.445, |
|
"eval_wer": 0.40406376992810256, |
|
"step": 115000 |
|
}, |
|
{ |
|
"epoch": 85.74, |
|
"learning_rate": 1.4405807892777364e-05, |
|
"loss": 0.0485, |
|
"step": 116000 |
|
}, |
|
{ |
|
"epoch": 85.74, |
|
"eval_loss": 0.7553907632827759, |
|
"eval_runtime": 181.6761, |
|
"eval_samples_per_second": 16.662, |
|
"eval_steps_per_second": 0.523, |
|
"eval_wer": 0.4044388871522351, |
|
"step": 116000 |
|
}, |
|
{ |
|
"epoch": 86.47, |
|
"learning_rate": 1.3661206254653761e-05, |
|
"loss": 0.0479, |
|
"step": 117000 |
|
}, |
|
{ |
|
"epoch": 86.47, |
|
"eval_loss": 0.8072551488876343, |
|
"eval_runtime": 188.7645, |
|
"eval_samples_per_second": 16.036, |
|
"eval_steps_per_second": 0.503, |
|
"eval_wer": 0.40625195373554235, |
|
"step": 117000 |
|
}, |
|
{ |
|
"epoch": 87.21, |
|
"learning_rate": 1.2917349218168282e-05, |
|
"loss": 0.0462, |
|
"step": 118000 |
|
}, |
|
{ |
|
"epoch": 87.21, |
|
"eval_loss": 0.7592765092849731, |
|
"eval_runtime": 187.3254, |
|
"eval_samples_per_second": 16.159, |
|
"eval_steps_per_second": 0.507, |
|
"eval_wer": 0.4025007814942169, |
|
"step": 118000 |
|
}, |
|
{ |
|
"epoch": 87.95, |
|
"learning_rate": 1.2172747580044677e-05, |
|
"loss": 0.0475, |
|
"step": 119000 |
|
}, |
|
{ |
|
"epoch": 87.95, |
|
"eval_loss": 0.7812408208847046, |
|
"eval_runtime": 227.3028, |
|
"eval_samples_per_second": 13.317, |
|
"eval_steps_per_second": 0.418, |
|
"eval_wer": 0.40493904345107845, |
|
"step": 119000 |
|
}, |
|
{ |
|
"epoch": 88.69, |
|
"learning_rate": 1.1428890543559197e-05, |
|
"loss": 0.0455, |
|
"step": 120000 |
|
}, |
|
{ |
|
"epoch": 88.69, |
|
"eval_loss": 0.779983401298523, |
|
"eval_runtime": 223.0101, |
|
"eval_samples_per_second": 13.573, |
|
"eval_steps_per_second": 0.426, |
|
"eval_wer": 0.40487652391372303, |
|
"step": 120000 |
|
}, |
|
{ |
|
"epoch": 89.43, |
|
"learning_rate": 1.0684288905435592e-05, |
|
"loss": 0.0451, |
|
"step": 121000 |
|
}, |
|
{ |
|
"epoch": 89.43, |
|
"eval_loss": 0.8092254400253296, |
|
"eval_runtime": 251.7762, |
|
"eval_samples_per_second": 12.023, |
|
"eval_steps_per_second": 0.377, |
|
"eval_wer": 0.40737730540794, |
|
"step": 121000 |
|
}, |
|
{ |
|
"epoch": 90.17, |
|
"learning_rate": 9.939687267311989e-06, |
|
"loss": 0.0455, |
|
"step": 122000 |
|
}, |
|
{ |
|
"epoch": 90.17, |
|
"eval_loss": 0.7754979133605957, |
|
"eval_runtime": 271.2522, |
|
"eval_samples_per_second": 11.159, |
|
"eval_steps_per_second": 0.35, |
|
"eval_wer": 0.4074398249452954, |
|
"step": 122000 |
|
}, |
|
{ |
|
"epoch": 90.91, |
|
"learning_rate": 9.195085629188386e-06, |
|
"loss": 0.0476, |
|
"step": 123000 |
|
}, |
|
{ |
|
"epoch": 90.91, |
|
"eval_loss": 0.7846540212631226, |
|
"eval_runtime": 244.3416, |
|
"eval_samples_per_second": 12.388, |
|
"eval_steps_per_second": 0.389, |
|
"eval_wer": 0.40593935604876524, |
|
"step": 123000 |
|
}, |
|
{ |
|
"epoch": 91.65, |
|
"learning_rate": 8.451228592702905e-06, |
|
"loss": 0.0461, |
|
"step": 124000 |
|
}, |
|
{ |
|
"epoch": 91.65, |
|
"eval_loss": 0.7784237861633301, |
|
"eval_runtime": 267.8746, |
|
"eval_samples_per_second": 11.3, |
|
"eval_steps_per_second": 0.355, |
|
"eval_wer": 0.4034385745545483, |
|
"step": 124000 |
|
}, |
|
{ |
|
"epoch": 92.39, |
|
"learning_rate": 7.7066269545793e-06, |
|
"loss": 0.0454, |
|
"step": 125000 |
|
}, |
|
{ |
|
"epoch": 92.39, |
|
"eval_loss": 0.7823372483253479, |
|
"eval_runtime": 207.7458, |
|
"eval_samples_per_second": 14.571, |
|
"eval_steps_per_second": 0.457, |
|
"eval_wer": 0.403251015942482, |
|
"step": 125000 |
|
}, |
|
{ |
|
"epoch": 93.13, |
|
"learning_rate": 6.9620253164556965e-06, |
|
"loss": 0.0453, |
|
"step": 126000 |
|
}, |
|
{ |
|
"epoch": 93.13, |
|
"eval_loss": 0.7685350179672241, |
|
"eval_runtime": 180.1025, |
|
"eval_samples_per_second": 16.807, |
|
"eval_steps_per_second": 0.527, |
|
"eval_wer": 0.4015004688965302, |
|
"step": 126000 |
|
}, |
|
{ |
|
"epoch": 93.87, |
|
"learning_rate": 6.2174236783320925e-06, |
|
"loss": 0.0451, |
|
"step": 127000 |
|
}, |
|
{ |
|
"epoch": 93.87, |
|
"eval_loss": 0.7680614590644836, |
|
"eval_runtime": 177.9917, |
|
"eval_samples_per_second": 17.006, |
|
"eval_steps_per_second": 0.534, |
|
"eval_wer": 0.4, |
|
"step": 127000 |
|
}, |
|
{ |
|
"epoch": 94.6, |
|
"learning_rate": 5.473566641846612e-06, |
|
"loss": 0.0428, |
|
"step": 128000 |
|
}, |
|
{ |
|
"epoch": 94.6, |
|
"eval_loss": 0.7936143279075623, |
|
"eval_runtime": 180.9951, |
|
"eval_samples_per_second": 16.724, |
|
"eval_steps_per_second": 0.525, |
|
"eval_wer": 0.4034385745545483, |
|
"step": 128000 |
|
}, |
|
{ |
|
"epoch": 95.34, |
|
"learning_rate": 4.728965003723009e-06, |
|
"loss": 0.0445, |
|
"step": 129000 |
|
}, |
|
{ |
|
"epoch": 95.34, |
|
"eval_loss": 0.7666231989860535, |
|
"eval_runtime": 203.0139, |
|
"eval_samples_per_second": 14.91, |
|
"eval_steps_per_second": 0.468, |
|
"eval_wer": 0.39968740231322286, |
|
"step": 129000 |
|
}, |
|
{ |
|
"epoch": 96.08, |
|
"learning_rate": 3.984363365599404e-06, |
|
"loss": 0.0441, |
|
"step": 130000 |
|
}, |
|
{ |
|
"epoch": 96.08, |
|
"eval_loss": 0.7681093811988831, |
|
"eval_runtime": 206.6264, |
|
"eval_samples_per_second": 14.65, |
|
"eval_steps_per_second": 0.46, |
|
"eval_wer": 0.4008752735229759, |
|
"step": 130000 |
|
}, |
|
{ |
|
"epoch": 96.82, |
|
"learning_rate": 3.2412509307520474e-06, |
|
"loss": 0.043, |
|
"step": 131000 |
|
}, |
|
{ |
|
"epoch": 96.82, |
|
"eval_loss": 0.7704082131385803, |
|
"eval_runtime": 196.2406, |
|
"eval_samples_per_second": 15.425, |
|
"eval_steps_per_second": 0.484, |
|
"eval_wer": 0.4008752735229759, |
|
"step": 131000 |
|
}, |
|
{ |
|
"epoch": 97.56, |
|
"learning_rate": 2.496649292628444e-06, |
|
"loss": 0.043, |
|
"step": 132000 |
|
}, |
|
{ |
|
"epoch": 97.56, |
|
"eval_loss": 0.7688604593276978, |
|
"eval_runtime": 208.2948, |
|
"eval_samples_per_second": 14.532, |
|
"eval_steps_per_second": 0.456, |
|
"eval_wer": 0.3981244138793373, |
|
"step": 132000 |
|
}, |
|
{ |
|
"epoch": 98.3, |
|
"learning_rate": 1.7520476545048399e-06, |
|
"loss": 0.0429, |
|
"step": 133000 |
|
}, |
|
{ |
|
"epoch": 98.3, |
|
"eval_loss": 0.7736847996711731, |
|
"eval_runtime": 207.1539, |
|
"eval_samples_per_second": 14.612, |
|
"eval_steps_per_second": 0.459, |
|
"eval_wer": 0.3992497655517349, |
|
"step": 133000 |
|
}, |
|
{ |
|
"epoch": 99.04, |
|
"learning_rate": 1.0081906180193596e-06, |
|
"loss": 0.0434, |
|
"step": 134000 |
|
}, |
|
{ |
|
"epoch": 99.04, |
|
"eval_loss": 0.7795138955116272, |
|
"eval_runtime": 196.4909, |
|
"eval_samples_per_second": 15.405, |
|
"eval_steps_per_second": 0.483, |
|
"eval_wer": 0.4012503907471085, |
|
"step": 134000 |
|
}, |
|
{ |
|
"epoch": 99.78, |
|
"learning_rate": 2.635889798957558e-07, |
|
"loss": 0.0433, |
|
"step": 135000 |
|
}, |
|
{ |
|
"epoch": 99.78, |
|
"eval_loss": 0.7773757576942444, |
|
"eval_runtime": 198.9794, |
|
"eval_samples_per_second": 15.213, |
|
"eval_steps_per_second": 0.477, |
|
"eval_wer": 0.4008127539856205, |
|
"step": 135000 |
|
} |
|
], |
|
"max_steps": 135300, |
|
"num_train_epochs": 100, |
|
"total_flos": 1.184566523613751e+21, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|