|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 100.0, |
|
"global_step": 78400, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.9880000000000003e-05, |
|
"loss": 6.8222, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"eval_loss": 3.502612352371216, |
|
"eval_runtime": 238.4374, |
|
"eval_samples_per_second": 34.974, |
|
"eval_steps_per_second": 1.095, |
|
"eval_wer": 1.0, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 1.987240051347882e-05, |
|
"loss": 3.2136, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"eval_loss": 3.0592851638793945, |
|
"eval_runtime": 239.7957, |
|
"eval_samples_per_second": 34.775, |
|
"eval_steps_per_second": 1.088, |
|
"eval_wer": 0.9999770836675298, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 1.9744030808729143e-05, |
|
"loss": 2.8882, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"eval_loss": 2.4670279026031494, |
|
"eval_runtime": 236.3109, |
|
"eval_samples_per_second": 35.288, |
|
"eval_steps_per_second": 1.104, |
|
"eval_wer": 0.9938584228979994, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 1.9615661103979463e-05, |
|
"loss": 2.3743, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"eval_loss": 1.1843628883361816, |
|
"eval_runtime": 239.6417, |
|
"eval_samples_per_second": 34.798, |
|
"eval_steps_per_second": 1.089, |
|
"eval_wer": 0.8656873753924422, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 1.9487291399229784e-05, |
|
"loss": 1.9456, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"eval_loss": 0.8227887153625488, |
|
"eval_runtime": 238.3867, |
|
"eval_samples_per_second": 34.981, |
|
"eval_steps_per_second": 1.095, |
|
"eval_wer": 0.7397162958040195, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 3.83, |
|
"learning_rate": 1.9358921694480104e-05, |
|
"loss": 1.7781, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 3.83, |
|
"eval_loss": 0.6826250553131104, |
|
"eval_runtime": 237.1431, |
|
"eval_samples_per_second": 35.164, |
|
"eval_steps_per_second": 1.101, |
|
"eval_wer": 0.6752984852304237, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 4.46, |
|
"learning_rate": 1.9230551989730427e-05, |
|
"loss": 1.6848, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 4.46, |
|
"eval_loss": 0.5885177254676819, |
|
"eval_runtime": 235.3547, |
|
"eval_samples_per_second": 35.432, |
|
"eval_steps_per_second": 1.109, |
|
"eval_wer": 0.6139743795402983, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 5.1, |
|
"learning_rate": 1.9102182284980744e-05, |
|
"loss": 1.6228, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 5.1, |
|
"eval_loss": 0.527351975440979, |
|
"eval_runtime": 237.0762, |
|
"eval_samples_per_second": 35.174, |
|
"eval_steps_per_second": 1.101, |
|
"eval_wer": 0.5788894745284965, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 5.74, |
|
"learning_rate": 1.8973812580231068e-05, |
|
"loss": 1.5768, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 5.74, |
|
"eval_loss": 0.48999691009521484, |
|
"eval_runtime": 235.6055, |
|
"eval_samples_per_second": 35.394, |
|
"eval_steps_per_second": 1.108, |
|
"eval_wer": 0.5518711185461879, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 6.38, |
|
"learning_rate": 1.8845699614890885e-05, |
|
"loss": 1.5431, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 6.38, |
|
"eval_loss": 0.45079049468040466, |
|
"eval_runtime": 235.5935, |
|
"eval_samples_per_second": 35.396, |
|
"eval_steps_per_second": 1.108, |
|
"eval_wer": 0.5237986112702523, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 1.871732991014121e-05, |
|
"loss": 1.5019, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"eval_loss": 0.42483240365982056, |
|
"eval_runtime": 235.975, |
|
"eval_samples_per_second": 35.338, |
|
"eval_steps_per_second": 1.106, |
|
"eval_wer": 0.5021426770859592, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 7.65, |
|
"learning_rate": 1.858921694480103e-05, |
|
"loss": 1.4684, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 7.65, |
|
"eval_loss": 0.4009439945220947, |
|
"eval_runtime": 236.3695, |
|
"eval_samples_per_second": 35.28, |
|
"eval_steps_per_second": 1.104, |
|
"eval_wer": 0.4827096271512707, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 8.29, |
|
"learning_rate": 1.846084724005135e-05, |
|
"loss": 1.4635, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 8.29, |
|
"eval_loss": 0.3830302357673645, |
|
"eval_runtime": 234.8449, |
|
"eval_samples_per_second": 35.509, |
|
"eval_steps_per_second": 1.111, |
|
"eval_wer": 0.4700139789628068, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 8.93, |
|
"learning_rate": 1.833247753530167e-05, |
|
"loss": 1.4291, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 8.93, |
|
"eval_loss": 0.3707072138786316, |
|
"eval_runtime": 237.9649, |
|
"eval_samples_per_second": 35.043, |
|
"eval_steps_per_second": 1.097, |
|
"eval_wer": 0.45947246602653713, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 9.57, |
|
"learning_rate": 1.820410783055199e-05, |
|
"loss": 1.4271, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 9.57, |
|
"eval_loss": 0.3570108115673065, |
|
"eval_runtime": 236.9986, |
|
"eval_samples_per_second": 35.186, |
|
"eval_steps_per_second": 1.101, |
|
"eval_wer": 0.4514059169970438, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 10.2, |
|
"learning_rate": 1.807573812580231e-05, |
|
"loss": 1.3938, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 10.2, |
|
"eval_loss": 0.34792855381965637, |
|
"eval_runtime": 234.9447, |
|
"eval_samples_per_second": 35.493, |
|
"eval_steps_per_second": 1.111, |
|
"eval_wer": 0.4378394481747141, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 10.84, |
|
"learning_rate": 1.794762516046213e-05, |
|
"loss": 1.3914, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 10.84, |
|
"eval_loss": 0.3395732045173645, |
|
"eval_runtime": 235.1285, |
|
"eval_samples_per_second": 35.466, |
|
"eval_steps_per_second": 1.11, |
|
"eval_wer": 0.4368082132135573, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 11.48, |
|
"learning_rate": 1.7819255455712454e-05, |
|
"loss": 1.3767, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 11.48, |
|
"eval_loss": 0.3253381848335266, |
|
"eval_runtime": 238.3808, |
|
"eval_samples_per_second": 34.982, |
|
"eval_steps_per_second": 1.095, |
|
"eval_wer": 0.42622086761234734, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 12.12, |
|
"learning_rate": 1.7690885750962775e-05, |
|
"loss": 1.3641, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 12.12, |
|
"eval_loss": 0.32509171962738037, |
|
"eval_runtime": 234.3424, |
|
"eval_samples_per_second": 35.585, |
|
"eval_steps_per_second": 1.114, |
|
"eval_wer": 0.4178105735958017, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 12.76, |
|
"learning_rate": 1.7562772785622595e-05, |
|
"loss": 1.355, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 12.76, |
|
"eval_loss": 0.3138255476951599, |
|
"eval_runtime": 234.3344, |
|
"eval_samples_per_second": 35.586, |
|
"eval_steps_per_second": 1.114, |
|
"eval_wer": 0.4135710520888237, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 13.39, |
|
"learning_rate": 1.7434403080872915e-05, |
|
"loss": 1.336, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 13.39, |
|
"eval_loss": 0.31212833523750305, |
|
"eval_runtime": 238.4088, |
|
"eval_samples_per_second": 34.978, |
|
"eval_steps_per_second": 1.095, |
|
"eval_wer": 0.40694823200494995, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 14.03, |
|
"learning_rate": 1.7306033376123236e-05, |
|
"loss": 1.3292, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 14.03, |
|
"eval_loss": 0.30410972237586975, |
|
"eval_runtime": 233.9502, |
|
"eval_samples_per_second": 35.644, |
|
"eval_steps_per_second": 1.116, |
|
"eval_wer": 0.4014483122121136, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 14.67, |
|
"learning_rate": 1.7177663671373556e-05, |
|
"loss": 1.3249, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 14.67, |
|
"eval_loss": 0.3013543486595154, |
|
"eval_runtime": 234.7846, |
|
"eval_samples_per_second": 35.518, |
|
"eval_steps_per_second": 1.112, |
|
"eval_wer": 0.3931296835254486, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 15.31, |
|
"learning_rate": 1.704929396662388e-05, |
|
"loss": 1.3156, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 15.31, |
|
"eval_loss": 0.301371306180954, |
|
"eval_runtime": 237.6831, |
|
"eval_samples_per_second": 35.085, |
|
"eval_steps_per_second": 1.098, |
|
"eval_wer": 0.39294635286568735, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 15.94, |
|
"learning_rate": 1.69209242618742e-05, |
|
"loss": 1.313, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 15.94, |
|
"eval_loss": 0.29687899351119995, |
|
"eval_runtime": 234.864, |
|
"eval_samples_per_second": 35.506, |
|
"eval_steps_per_second": 1.111, |
|
"eval_wer": 0.39679629672067285, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 16.58, |
|
"learning_rate": 1.679255455712452e-05, |
|
"loss": 1.3068, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 16.58, |
|
"eval_loss": 0.2965061664581299, |
|
"eval_runtime": 239.4104, |
|
"eval_samples_per_second": 34.831, |
|
"eval_steps_per_second": 1.09, |
|
"eval_wer": 0.39663588239338177, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 17.22, |
|
"learning_rate": 1.666444159178434e-05, |
|
"loss": 1.2785, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 17.22, |
|
"eval_loss": 0.29434436559677124, |
|
"eval_runtime": 241.1756, |
|
"eval_samples_per_second": 34.576, |
|
"eval_steps_per_second": 1.082, |
|
"eval_wer": 0.38497146916607466, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 17.86, |
|
"learning_rate": 1.653607188703466e-05, |
|
"loss": 1.2867, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 17.86, |
|
"eval_loss": 0.291163831949234, |
|
"eval_runtime": 234.1726, |
|
"eval_samples_per_second": 35.61, |
|
"eval_steps_per_second": 1.115, |
|
"eval_wer": 0.37821115108737996, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 18.49, |
|
"learning_rate": 1.640770218228498e-05, |
|
"loss": 1.2714, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 18.49, |
|
"eval_loss": 0.281938374042511, |
|
"eval_runtime": 233.7302, |
|
"eval_samples_per_second": 35.678, |
|
"eval_steps_per_second": 1.117, |
|
"eval_wer": 0.3746820358869766, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 19.13, |
|
"learning_rate": 1.6279332477535305e-05, |
|
"loss": 1.2844, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 19.13, |
|
"eval_loss": 0.28400421142578125, |
|
"eval_runtime": 234.7428, |
|
"eval_samples_per_second": 35.524, |
|
"eval_steps_per_second": 1.112, |
|
"eval_wer": 0.3740403785778124, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 19.77, |
|
"learning_rate": 1.6150962772785625e-05, |
|
"loss": 1.2684, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 19.77, |
|
"eval_loss": 0.2913394272327423, |
|
"eval_runtime": 238.9443, |
|
"eval_samples_per_second": 34.899, |
|
"eval_steps_per_second": 1.092, |
|
"eval_wer": 0.38279441758141025, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 20.41, |
|
"learning_rate": 1.6022849807445445e-05, |
|
"loss": 1.26, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 20.41, |
|
"eval_loss": 0.27389487624168396, |
|
"eval_runtime": 235.3865, |
|
"eval_samples_per_second": 35.427, |
|
"eval_steps_per_second": 1.109, |
|
"eval_wer": 0.3674404748264088, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 21.05, |
|
"learning_rate": 1.5894736842105266e-05, |
|
"loss": 1.2543, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 21.05, |
|
"eval_loss": 0.2739649713039398, |
|
"eval_runtime": 234.6769, |
|
"eval_samples_per_second": 35.534, |
|
"eval_steps_per_second": 1.112, |
|
"eval_wer": 0.36909045076425967, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 21.68, |
|
"learning_rate": 1.5766367137355586e-05, |
|
"loss": 1.2532, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 21.68, |
|
"eval_loss": 0.2709480822086334, |
|
"eval_runtime": 236.2528, |
|
"eval_samples_per_second": 35.297, |
|
"eval_steps_per_second": 1.105, |
|
"eval_wer": 0.37562160551825285, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 22.32, |
|
"learning_rate": 1.5637997432605906e-05, |
|
"loss": 1.2409, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 22.32, |
|
"eval_loss": 0.266906201839447, |
|
"eval_runtime": 233.9016, |
|
"eval_samples_per_second": 35.652, |
|
"eval_steps_per_second": 1.116, |
|
"eval_wer": 0.35932809313197517, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 22.96, |
|
"learning_rate": 1.5509627727856227e-05, |
|
"loss": 1.2404, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 22.96, |
|
"eval_loss": 0.2673003077507019, |
|
"eval_runtime": 235.0288, |
|
"eval_samples_per_second": 35.481, |
|
"eval_steps_per_second": 1.111, |
|
"eval_wer": 0.35763228452918394, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 23.6, |
|
"learning_rate": 1.538125802310655e-05, |
|
"loss": 1.2347, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 23.6, |
|
"eval_loss": 0.2678009271621704, |
|
"eval_runtime": 234.3074, |
|
"eval_samples_per_second": 35.59, |
|
"eval_steps_per_second": 1.114, |
|
"eval_wer": 0.36427802094552786, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 24.23, |
|
"learning_rate": 1.5252888318356869e-05, |
|
"loss": 1.2351, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 24.23, |
|
"eval_loss": 0.2715151607990265, |
|
"eval_runtime": 239.5626, |
|
"eval_samples_per_second": 34.809, |
|
"eval_steps_per_second": 1.089, |
|
"eval_wer": 0.36498842725210257, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 24.87, |
|
"learning_rate": 1.512451861360719e-05, |
|
"loss": 1.2409, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 24.87, |
|
"eval_loss": 0.2636788487434387, |
|
"eval_runtime": 234.5142, |
|
"eval_samples_per_second": 35.559, |
|
"eval_steps_per_second": 1.113, |
|
"eval_wer": 0.3570822925499003, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 25.51, |
|
"learning_rate": 1.4996148908857511e-05, |
|
"loss": 1.2152, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 25.51, |
|
"eval_loss": 0.2785368859767914, |
|
"eval_runtime": 234.4552, |
|
"eval_samples_per_second": 35.568, |
|
"eval_steps_per_second": 1.113, |
|
"eval_wer": 0.36088640373994546, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 26.15, |
|
"learning_rate": 1.4868035943517332e-05, |
|
"loss": 1.2046, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 26.15, |
|
"eval_loss": 0.2610293924808502, |
|
"eval_runtime": 233.6859, |
|
"eval_samples_per_second": 35.685, |
|
"eval_steps_per_second": 1.117, |
|
"eval_wer": 0.3508032174530788, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 26.79, |
|
"learning_rate": 1.4739666238767652e-05, |
|
"loss": 1.2082, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 26.79, |
|
"eval_loss": 0.26189225912094116, |
|
"eval_runtime": 234.8807, |
|
"eval_samples_per_second": 35.503, |
|
"eval_steps_per_second": 1.111, |
|
"eval_wer": 0.3460595366317575, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 27.42, |
|
"learning_rate": 1.4611296534017974e-05, |
|
"loss": 1.2109, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 27.42, |
|
"eval_loss": 0.2596839666366577, |
|
"eval_runtime": 235.2549, |
|
"eval_samples_per_second": 35.447, |
|
"eval_steps_per_second": 1.109, |
|
"eval_wer": 0.3501615601439146, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 28.06, |
|
"learning_rate": 1.4482926829268294e-05, |
|
"loss": 1.2014, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 28.06, |
|
"eval_loss": 0.2608143091201782, |
|
"eval_runtime": 234.8521, |
|
"eval_samples_per_second": 35.507, |
|
"eval_steps_per_second": 1.111, |
|
"eval_wer": 0.3468386919357426, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 28.7, |
|
"learning_rate": 1.4354813863928115e-05, |
|
"loss": 1.1948, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 28.7, |
|
"eval_loss": 0.2573413550853729, |
|
"eval_runtime": 233.0487, |
|
"eval_samples_per_second": 35.782, |
|
"eval_steps_per_second": 1.12, |
|
"eval_wer": 0.3456699589797649, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 29.34, |
|
"learning_rate": 1.4226444159178435e-05, |
|
"loss": 1.205, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 29.34, |
|
"eval_loss": 0.2619355320930481, |
|
"eval_runtime": 236.4546, |
|
"eval_samples_per_second": 35.267, |
|
"eval_steps_per_second": 1.104, |
|
"eval_wer": 0.3463803652863396, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 29.97, |
|
"learning_rate": 1.4098074454428757e-05, |
|
"loss": 1.2019, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 29.97, |
|
"eval_loss": 0.2558630406856537, |
|
"eval_runtime": 235.5761, |
|
"eval_samples_per_second": 35.398, |
|
"eval_steps_per_second": 1.108, |
|
"eval_wer": 0.34738868391502625, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 30.61, |
|
"learning_rate": 1.3969704749679077e-05, |
|
"loss": 1.1917, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 30.61, |
|
"eval_loss": 0.2601245045661926, |
|
"eval_runtime": 235.3838, |
|
"eval_samples_per_second": 35.427, |
|
"eval_steps_per_second": 1.109, |
|
"eval_wer": 0.34619703462657836, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 31.25, |
|
"learning_rate": 1.3841591784338897e-05, |
|
"loss": 1.1939, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 31.25, |
|
"eval_loss": 0.2575133740901947, |
|
"eval_runtime": 235.7288, |
|
"eval_samples_per_second": 35.375, |
|
"eval_steps_per_second": 1.107, |
|
"eval_wer": 0.33874922657377915, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 31.89, |
|
"learning_rate": 1.3713478818998718e-05, |
|
"loss": 1.1882, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 31.89, |
|
"eval_loss": 0.2534877359867096, |
|
"eval_runtime": 235.7724, |
|
"eval_samples_per_second": 35.369, |
|
"eval_steps_per_second": 1.107, |
|
"eval_wer": 0.33675550564887596, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 32.53, |
|
"learning_rate": 1.3585109114249038e-05, |
|
"loss": 1.191, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 32.53, |
|
"eval_loss": 0.24889491498470306, |
|
"eval_runtime": 236.163, |
|
"eval_samples_per_second": 35.31, |
|
"eval_steps_per_second": 1.105, |
|
"eval_wer": 0.336457593326764, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 33.16, |
|
"learning_rate": 1.345673940949936e-05, |
|
"loss": 1.1767, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 33.16, |
|
"eval_loss": 0.2501368224620819, |
|
"eval_runtime": 236.8496, |
|
"eval_samples_per_second": 35.208, |
|
"eval_steps_per_second": 1.102, |
|
"eval_wer": 0.3347159520590325, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 33.8, |
|
"learning_rate": 1.332836970474968e-05, |
|
"loss": 1.167, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 33.8, |
|
"eval_loss": 0.25037434697151184, |
|
"eval_runtime": 233.4394, |
|
"eval_samples_per_second": 35.722, |
|
"eval_steps_per_second": 1.118, |
|
"eval_wer": 0.33469303572656234, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 34.44, |
|
"learning_rate": 1.3200000000000002e-05, |
|
"loss": 1.1678, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 34.44, |
|
"eval_loss": 0.24800854921340942, |
|
"eval_runtime": 233.3087, |
|
"eval_samples_per_second": 35.742, |
|
"eval_steps_per_second": 1.119, |
|
"eval_wer": 0.33783257327497307, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 35.08, |
|
"learning_rate": 1.3071630295250321e-05, |
|
"loss": 1.1803, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 35.08, |
|
"eval_loss": 0.24871966242790222, |
|
"eval_runtime": 235.3613, |
|
"eval_samples_per_second": 35.431, |
|
"eval_steps_per_second": 1.109, |
|
"eval_wer": 0.33448678873433096, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 35.71, |
|
"learning_rate": 1.2943260590500643e-05, |
|
"loss": 1.167, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 35.71, |
|
"eval_loss": 0.24418053030967712, |
|
"eval_runtime": 232.8217, |
|
"eval_samples_per_second": 35.817, |
|
"eval_steps_per_second": 1.121, |
|
"eval_wer": 0.3318514105002635, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 36.35, |
|
"learning_rate": 1.2814890885750963e-05, |
|
"loss": 1.1661, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 36.35, |
|
"eval_loss": 0.24953289330005646, |
|
"eval_runtime": 237.0041, |
|
"eval_samples_per_second": 35.185, |
|
"eval_steps_per_second": 1.101, |
|
"eval_wer": 0.3333868047757637, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 36.99, |
|
"learning_rate": 1.2686521181001285e-05, |
|
"loss": 1.164, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 36.99, |
|
"eval_loss": 0.24715770781040192, |
|
"eval_runtime": 233.2505, |
|
"eval_samples_per_second": 35.751, |
|
"eval_steps_per_second": 1.119, |
|
"eval_wer": 0.32923894859866626, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 37.63, |
|
"learning_rate": 1.2558151476251605e-05, |
|
"loss": 1.1578, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 37.63, |
|
"eval_loss": 0.24420544505119324, |
|
"eval_runtime": 239.1685, |
|
"eval_samples_per_second": 34.867, |
|
"eval_steps_per_second": 1.091, |
|
"eval_wer": 0.3242431881201732, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 38.27, |
|
"learning_rate": 1.2430038510911426e-05, |
|
"loss": 1.1584, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 38.27, |
|
"eval_loss": 0.243063285946846, |
|
"eval_runtime": 235.3653, |
|
"eval_samples_per_second": 35.43, |
|
"eval_steps_per_second": 1.109, |
|
"eval_wer": 0.3314389165158008, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 38.9, |
|
"learning_rate": 1.2301668806161746e-05, |
|
"loss": 1.1526, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 38.9, |
|
"eval_loss": 0.2440960705280304, |
|
"eval_runtime": 238.3272, |
|
"eval_samples_per_second": 34.99, |
|
"eval_steps_per_second": 1.095, |
|
"eval_wer": 0.3347159520590325, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 39.54, |
|
"learning_rate": 1.2173299101412068e-05, |
|
"loss": 1.1542, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 39.54, |
|
"eval_loss": 0.2436649352312088, |
|
"eval_runtime": 236.1705, |
|
"eval_samples_per_second": 35.309, |
|
"eval_steps_per_second": 1.105, |
|
"eval_wer": 0.3330430597887114, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 40.18, |
|
"learning_rate": 1.2044929396662388e-05, |
|
"loss": 1.1508, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 40.18, |
|
"eval_loss": 0.24334189295768738, |
|
"eval_runtime": 234.2613, |
|
"eval_samples_per_second": 35.597, |
|
"eval_steps_per_second": 1.114, |
|
"eval_wer": 0.3294222792584275, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 40.82, |
|
"learning_rate": 1.191655969191271e-05, |
|
"loss": 1.1406, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 40.82, |
|
"eval_loss": 0.24340486526489258, |
|
"eval_runtime": 233.0368, |
|
"eval_samples_per_second": 35.784, |
|
"eval_steps_per_second": 1.12, |
|
"eval_wer": 0.3270618970140019, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 41.45, |
|
"learning_rate": 1.178870346598203e-05, |
|
"loss": 1.1514, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 41.45, |
|
"eval_loss": 0.24264076352119446, |
|
"eval_runtime": 238.874, |
|
"eval_samples_per_second": 34.91, |
|
"eval_steps_per_second": 1.093, |
|
"eval_wer": 0.32552650273850176, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 42.09, |
|
"learning_rate": 1.166033376123235e-05, |
|
"loss": 1.1418, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 42.09, |
|
"eval_loss": 0.2431638389825821, |
|
"eval_runtime": 238.255, |
|
"eval_samples_per_second": 35.0, |
|
"eval_steps_per_second": 1.095, |
|
"eval_wer": 0.3232807021564269, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 42.73, |
|
"learning_rate": 1.1531964056482671e-05, |
|
"loss": 1.1365, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 42.73, |
|
"eval_loss": 0.2436181753873825, |
|
"eval_runtime": 234.1309, |
|
"eval_samples_per_second": 35.617, |
|
"eval_steps_per_second": 1.115, |
|
"eval_wer": 0.32399110846300155, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 43.37, |
|
"learning_rate": 1.1403594351732992e-05, |
|
"loss": 1.1348, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 43.37, |
|
"eval_loss": 0.2482648491859436, |
|
"eval_runtime": 238.6079, |
|
"eval_samples_per_second": 34.949, |
|
"eval_steps_per_second": 1.094, |
|
"eval_wer": 0.3257327497307331, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 44.01, |
|
"learning_rate": 1.1275224646983314e-05, |
|
"loss": 1.1301, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 44.01, |
|
"eval_loss": 0.24202857911586761, |
|
"eval_runtime": 232.8764, |
|
"eval_samples_per_second": 35.809, |
|
"eval_steps_per_second": 1.121, |
|
"eval_wer": 0.3271077296789422, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 44.64, |
|
"learning_rate": 1.1146854942233634e-05, |
|
"loss": 1.1268, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 44.64, |
|
"eval_loss": 0.24715635180473328, |
|
"eval_runtime": 233.1436, |
|
"eval_samples_per_second": 35.768, |
|
"eval_steps_per_second": 1.119, |
|
"eval_wer": 0.3224786305199716, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 45.28, |
|
"learning_rate": 1.1018485237483956e-05, |
|
"loss": 1.1224, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 45.28, |
|
"eval_loss": 0.2381853461265564, |
|
"eval_runtime": 238.9288, |
|
"eval_samples_per_second": 34.902, |
|
"eval_steps_per_second": 1.092, |
|
"eval_wer": 0.3204849095950684, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 45.92, |
|
"learning_rate": 1.0890115532734276e-05, |
|
"loss": 1.1224, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 45.92, |
|
"eval_loss": 0.23876558244228363, |
|
"eval_runtime": 243.8522, |
|
"eval_samples_per_second": 34.197, |
|
"eval_steps_per_second": 1.07, |
|
"eval_wer": 0.3184224396727548, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 46.56, |
|
"learning_rate": 1.0762002567394097e-05, |
|
"loss": 1.1198, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 46.56, |
|
"eval_loss": 0.23819147050380707, |
|
"eval_runtime": 238.5089, |
|
"eval_samples_per_second": 34.963, |
|
"eval_steps_per_second": 1.094, |
|
"eval_wer": 0.32016408094048626, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 47.19, |
|
"learning_rate": 1.0633632862644417e-05, |
|
"loss": 1.1274, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 47.19, |
|
"eval_loss": 0.24040183424949646, |
|
"eval_runtime": 237.5269, |
|
"eval_samples_per_second": 35.108, |
|
"eval_steps_per_second": 1.099, |
|
"eval_wer": 0.31720787405183676, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 47.83, |
|
"learning_rate": 1.0505263157894739e-05, |
|
"loss": 1.1147, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 47.83, |
|
"eval_loss": 0.23944656550884247, |
|
"eval_runtime": 235.5135, |
|
"eval_samples_per_second": 35.408, |
|
"eval_steps_per_second": 1.108, |
|
"eval_wer": 0.3164287187478516, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 48.47, |
|
"learning_rate": 1.0377150192554558e-05, |
|
"loss": 1.121, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 48.47, |
|
"eval_loss": 0.2405969202518463, |
|
"eval_runtime": 233.9426, |
|
"eval_samples_per_second": 35.645, |
|
"eval_steps_per_second": 1.116, |
|
"eval_wer": 0.3202099136054266, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 49.11, |
|
"learning_rate": 1.024878048780488e-05, |
|
"loss": 1.1109, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 49.11, |
|
"eval_loss": 0.23839789628982544, |
|
"eval_runtime": 240.6944, |
|
"eval_samples_per_second": 34.646, |
|
"eval_steps_per_second": 1.084, |
|
"eval_wer": 0.31537456745422465, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 49.74, |
|
"learning_rate": 1.0120410783055198e-05, |
|
"loss": 1.1164, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 49.74, |
|
"eval_loss": 0.2374895066022873, |
|
"eval_runtime": 237.3857, |
|
"eval_samples_per_second": 35.128, |
|
"eval_steps_per_second": 1.099, |
|
"eval_wer": 0.31686412906478445, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 50.38, |
|
"learning_rate": 9.992041078305522e-06, |
|
"loss": 1.1105, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 50.38, |
|
"eval_loss": 0.23865096271038055, |
|
"eval_runtime": 232.9433, |
|
"eval_samples_per_second": 35.798, |
|
"eval_steps_per_second": 1.12, |
|
"eval_wer": 0.31729953938171734, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 51.02, |
|
"learning_rate": 9.86392811296534e-06, |
|
"loss": 1.1054, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 51.02, |
|
"eval_loss": 0.23617714643478394, |
|
"eval_runtime": 233.5533, |
|
"eval_samples_per_second": 35.705, |
|
"eval_steps_per_second": 1.118, |
|
"eval_wer": 0.31200586658111235, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 51.66, |
|
"learning_rate": 9.735558408215663e-06, |
|
"loss": 1.0893, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 51.66, |
|
"eval_loss": 0.23990453779697418, |
|
"eval_runtime": 234.1326, |
|
"eval_samples_per_second": 35.617, |
|
"eval_steps_per_second": 1.115, |
|
"eval_wer": 0.31299126887732887, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 52.3, |
|
"learning_rate": 9.607188703465983e-06, |
|
"loss": 1.0913, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 52.3, |
|
"eval_loss": 0.2356649786233902, |
|
"eval_runtime": 238.2462, |
|
"eval_samples_per_second": 35.002, |
|
"eval_steps_per_second": 1.096, |
|
"eval_wer": 0.3087517473703508, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 52.93, |
|
"learning_rate": 9.478818998716303e-06, |
|
"loss": 1.1017, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 52.93, |
|
"eval_loss": 0.23450981080532074, |
|
"eval_runtime": 233.807, |
|
"eval_samples_per_second": 35.666, |
|
"eval_steps_per_second": 1.116, |
|
"eval_wer": 0.3083621697183583, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 53.57, |
|
"learning_rate": 9.350449293966625e-06, |
|
"loss": 1.0937, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 53.57, |
|
"eval_loss": 0.23296880722045898, |
|
"eval_runtime": 238.7521, |
|
"eval_samples_per_second": 34.927, |
|
"eval_steps_per_second": 1.093, |
|
"eval_wer": 0.31395375484107524, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 54.21, |
|
"learning_rate": 9.222079589216945e-06, |
|
"loss": 1.0945, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 54.21, |
|
"eval_loss": 0.23989614844322205, |
|
"eval_runtime": 233.2848, |
|
"eval_samples_per_second": 35.746, |
|
"eval_steps_per_second": 1.119, |
|
"eval_wer": 0.3106767192978436, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 54.85, |
|
"learning_rate": 9.093966623876766e-06, |
|
"loss": 1.0933, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 54.85, |
|
"eval_loss": 0.23831374943256378, |
|
"eval_runtime": 234.0959, |
|
"eval_samples_per_second": 35.622, |
|
"eval_steps_per_second": 1.115, |
|
"eval_wer": 0.31338084652932147, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 55.48, |
|
"learning_rate": 8.965596919127086e-06, |
|
"loss": 1.0912, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 55.48, |
|
"eval_loss": 0.2371690571308136, |
|
"eval_runtime": 234.0624, |
|
"eval_samples_per_second": 35.627, |
|
"eval_steps_per_second": 1.115, |
|
"eval_wer": 0.30769759607672387, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 56.12, |
|
"learning_rate": 8.837227214377408e-06, |
|
"loss": 1.0898, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 56.12, |
|
"eval_loss": 0.23387998342514038, |
|
"eval_runtime": 233.2957, |
|
"eval_samples_per_second": 35.744, |
|
"eval_steps_per_second": 1.119, |
|
"eval_wer": 0.30829342072094784, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 56.76, |
|
"learning_rate": 8.708857509627728e-06, |
|
"loss": 1.0903, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 56.76, |
|
"eval_loss": 0.23673543334007263, |
|
"eval_runtime": 236.1402, |
|
"eval_samples_per_second": 35.314, |
|
"eval_steps_per_second": 1.105, |
|
"eval_wer": 0.3065288631207462, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 57.4, |
|
"learning_rate": 8.58048780487805e-06, |
|
"loss": 1.0947, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 57.4, |
|
"eval_loss": 0.23515641689300537, |
|
"eval_runtime": 235.6322, |
|
"eval_samples_per_second": 35.39, |
|
"eval_steps_per_second": 1.108, |
|
"eval_wer": 0.31035589064326147, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 58.04, |
|
"learning_rate": 8.45211810012837e-06, |
|
"loss": 1.0751, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 58.04, |
|
"eval_loss": 0.23344124853610992, |
|
"eval_runtime": 236.1866, |
|
"eval_samples_per_second": 35.307, |
|
"eval_steps_per_second": 1.105, |
|
"eval_wer": 0.3084309187157687, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 58.67, |
|
"learning_rate": 8.32374839537869e-06, |
|
"loss": 1.09, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 58.67, |
|
"eval_loss": 0.2327876091003418, |
|
"eval_runtime": 233.0944, |
|
"eval_samples_per_second": 35.775, |
|
"eval_steps_per_second": 1.12, |
|
"eval_wer": 0.3100350619886793, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 59.31, |
|
"learning_rate": 8.195378690629013e-06, |
|
"loss": 1.0876, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 59.31, |
|
"eval_loss": 0.22762973606586456, |
|
"eval_runtime": 232.6706, |
|
"eval_samples_per_second": 35.84, |
|
"eval_steps_per_second": 1.122, |
|
"eval_wer": 0.3049705525127758, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 59.95, |
|
"learning_rate": 8.067008985879333e-06, |
|
"loss": 1.076, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 59.95, |
|
"eval_loss": 0.23087561130523682, |
|
"eval_runtime": 233.5675, |
|
"eval_samples_per_second": 35.703, |
|
"eval_steps_per_second": 1.117, |
|
"eval_wer": 0.304695556523134, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 60.59, |
|
"learning_rate": 7.938896020539153e-06, |
|
"loss": 1.086, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 60.59, |
|
"eval_loss": 0.229255810379982, |
|
"eval_runtime": 234.5783, |
|
"eval_samples_per_second": 35.549, |
|
"eval_steps_per_second": 1.113, |
|
"eval_wer": 0.304695556523134, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 61.22, |
|
"learning_rate": 7.810526315789474e-06, |
|
"loss": 1.082, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 61.22, |
|
"eval_loss": 0.232817605137825, |
|
"eval_runtime": 239.1878, |
|
"eval_samples_per_second": 34.864, |
|
"eval_steps_per_second": 1.091, |
|
"eval_wer": 0.3027476682631712, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 61.86, |
|
"learning_rate": 7.682413350449294e-06, |
|
"loss": 1.0714, |
|
"step": 48500 |
|
}, |
|
{ |
|
"epoch": 61.86, |
|
"eval_loss": 0.22896084189414978, |
|
"eval_runtime": 234.6636, |
|
"eval_samples_per_second": 35.536, |
|
"eval_steps_per_second": 1.112, |
|
"eval_wer": 0.3019914292916562, |
|
"step": 48500 |
|
}, |
|
{ |
|
"epoch": 62.5, |
|
"learning_rate": 7.554043645699616e-06, |
|
"loss": 1.0746, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 62.5, |
|
"eval_loss": 0.2313106656074524, |
|
"eval_runtime": 234.936, |
|
"eval_samples_per_second": 35.495, |
|
"eval_steps_per_second": 1.111, |
|
"eval_wer": 0.30586428947911176, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 63.14, |
|
"learning_rate": 7.425673940949937e-06, |
|
"loss": 1.076, |
|
"step": 49500 |
|
}, |
|
{ |
|
"epoch": 63.14, |
|
"eval_loss": 0.23419487476348877, |
|
"eval_runtime": 235.1421, |
|
"eval_samples_per_second": 35.464, |
|
"eval_steps_per_second": 1.11, |
|
"eval_wer": 0.3050393015101863, |
|
"step": 49500 |
|
}, |
|
{ |
|
"epoch": 63.78, |
|
"learning_rate": 7.297304236200257e-06, |
|
"loss": 1.0648, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 63.78, |
|
"eval_loss": 0.22855186462402344, |
|
"eval_runtime": 233.6212, |
|
"eval_samples_per_second": 35.695, |
|
"eval_steps_per_second": 1.117, |
|
"eval_wer": 0.30247267227352936, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 64.41, |
|
"learning_rate": 7.169191270860078e-06, |
|
"loss": 1.0586, |
|
"step": 50500 |
|
}, |
|
{ |
|
"epoch": 64.41, |
|
"eval_loss": 0.23375432193279266, |
|
"eval_runtime": 233.8132, |
|
"eval_samples_per_second": 35.665, |
|
"eval_steps_per_second": 1.116, |
|
"eval_wer": 0.30435181153608176, |
|
"step": 50500 |
|
}, |
|
{ |
|
"epoch": 65.05, |
|
"learning_rate": 7.040821566110398e-06, |
|
"loss": 1.0753, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 65.05, |
|
"eval_loss": 0.23079071938991547, |
|
"eval_runtime": 238.4184, |
|
"eval_samples_per_second": 34.976, |
|
"eval_steps_per_second": 1.095, |
|
"eval_wer": 0.30446639319843255, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 65.69, |
|
"learning_rate": 6.912451861360719e-06, |
|
"loss": 1.0664, |
|
"step": 51500 |
|
}, |
|
{ |
|
"epoch": 65.69, |
|
"eval_loss": 0.22733470797538757, |
|
"eval_runtime": 233.5628, |
|
"eval_samples_per_second": 35.703, |
|
"eval_steps_per_second": 1.117, |
|
"eval_wer": 0.3009372779980292, |
|
"step": 51500 |
|
}, |
|
{ |
|
"epoch": 66.33, |
|
"learning_rate": 6.78408215661104e-06, |
|
"loss": 1.0739, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 66.33, |
|
"eval_loss": 0.2298172414302826, |
|
"eval_runtime": 234.5885, |
|
"eval_samples_per_second": 35.547, |
|
"eval_steps_per_second": 1.113, |
|
"eval_wer": 0.30272475193070103, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 66.96, |
|
"learning_rate": 6.655712451861361e-06, |
|
"loss": 1.0695, |
|
"step": 52500 |
|
}, |
|
{ |
|
"epoch": 66.96, |
|
"eval_loss": 0.2247009128332138, |
|
"eval_runtime": 234.2623, |
|
"eval_samples_per_second": 35.597, |
|
"eval_steps_per_second": 1.114, |
|
"eval_wer": 0.29958521438229024, |
|
"step": 52500 |
|
}, |
|
{ |
|
"epoch": 67.6, |
|
"learning_rate": 6.5273427471116825e-06, |
|
"loss": 1.06, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 67.6, |
|
"eval_loss": 0.22762586176395416, |
|
"eval_runtime": 238.2043, |
|
"eval_samples_per_second": 35.008, |
|
"eval_steps_per_second": 1.096, |
|
"eval_wer": 0.30148726997731284, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 68.24, |
|
"learning_rate": 6.399229781771502e-06, |
|
"loss": 1.0742, |
|
"step": 53500 |
|
}, |
|
{ |
|
"epoch": 68.24, |
|
"eval_loss": 0.22798208892345428, |
|
"eval_runtime": 234.5841, |
|
"eval_samples_per_second": 35.548, |
|
"eval_steps_per_second": 1.113, |
|
"eval_wer": 0.2974081627976259, |
|
"step": 53500 |
|
}, |
|
{ |
|
"epoch": 68.88, |
|
"learning_rate": 6.270860077021823e-06, |
|
"loss": 1.0618, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 68.88, |
|
"eval_loss": 0.2291172742843628, |
|
"eval_runtime": 233.7512, |
|
"eval_samples_per_second": 35.675, |
|
"eval_steps_per_second": 1.117, |
|
"eval_wer": 0.2988518917432454, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 69.52, |
|
"learning_rate": 6.142490372272144e-06, |
|
"loss": 1.062, |
|
"step": 54500 |
|
}, |
|
{ |
|
"epoch": 69.52, |
|
"eval_loss": 0.23021718859672546, |
|
"eval_runtime": 236.2703, |
|
"eval_samples_per_second": 35.294, |
|
"eval_steps_per_second": 1.105, |
|
"eval_wer": 0.2970873341430437, |
|
"step": 54500 |
|
}, |
|
{ |
|
"epoch": 70.15, |
|
"learning_rate": 6.0141206675224654e-06, |
|
"loss": 1.0572, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 70.15, |
|
"eval_loss": 0.22801047563552856, |
|
"eval_runtime": 232.8765, |
|
"eval_samples_per_second": 35.809, |
|
"eval_steps_per_second": 1.121, |
|
"eval_wer": 0.2989893897380663, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 70.79, |
|
"learning_rate": 5.885750962772786e-06, |
|
"loss": 1.055, |
|
"step": 55500 |
|
}, |
|
{ |
|
"epoch": 70.79, |
|
"eval_loss": 0.22775860130786896, |
|
"eval_runtime": 234.9515, |
|
"eval_samples_per_second": 35.492, |
|
"eval_steps_per_second": 1.111, |
|
"eval_wer": 0.29834773242890206, |
|
"step": 55500 |
|
}, |
|
{ |
|
"epoch": 71.43, |
|
"learning_rate": 5.757381258023107e-06, |
|
"loss": 1.0553, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 71.43, |
|
"eval_loss": 0.22817462682724, |
|
"eval_runtime": 233.8475, |
|
"eval_samples_per_second": 35.66, |
|
"eval_steps_per_second": 1.116, |
|
"eval_wer": 0.29905813873547676, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 72.07, |
|
"learning_rate": 5.629011553273428e-06, |
|
"loss": 1.0509, |
|
"step": 56500 |
|
}, |
|
{ |
|
"epoch": 72.07, |
|
"eval_loss": 0.2261197566986084, |
|
"eval_runtime": 235.4495, |
|
"eval_samples_per_second": 35.417, |
|
"eval_steps_per_second": 1.109, |
|
"eval_wer": 0.29591860118706603, |
|
"step": 56500 |
|
}, |
|
{ |
|
"epoch": 72.7, |
|
"learning_rate": 5.500641848523749e-06, |
|
"loss": 1.0469, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 72.7, |
|
"eval_loss": 0.22159287333488464, |
|
"eval_runtime": 233.6081, |
|
"eval_samples_per_second": 35.697, |
|
"eval_steps_per_second": 1.117, |
|
"eval_wer": 0.29188532667231937, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 73.34, |
|
"learning_rate": 5.37227214377407e-06, |
|
"loss": 1.0476, |
|
"step": 57500 |
|
}, |
|
{ |
|
"epoch": 73.34, |
|
"eval_loss": 0.22673606872558594, |
|
"eval_runtime": 233.9171, |
|
"eval_samples_per_second": 35.649, |
|
"eval_steps_per_second": 1.116, |
|
"eval_wer": 0.2988518917432454, |
|
"step": 57500 |
|
}, |
|
{ |
|
"epoch": 73.98, |
|
"learning_rate": 5.24415917843389e-06, |
|
"loss": 1.0494, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 73.98, |
|
"eval_loss": 0.22603431344032288, |
|
"eval_runtime": 234.549, |
|
"eval_samples_per_second": 35.553, |
|
"eval_steps_per_second": 1.113, |
|
"eval_wer": 0.2960102665169466, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 74.62, |
|
"learning_rate": 5.115789473684211e-06, |
|
"loss": 1.0517, |
|
"step": 58500 |
|
}, |
|
{ |
|
"epoch": 74.62, |
|
"eval_loss": 0.2297457605600357, |
|
"eval_runtime": 234.3971, |
|
"eval_samples_per_second": 35.576, |
|
"eval_steps_per_second": 1.113, |
|
"eval_wer": 0.29894355707312603, |
|
"step": 58500 |
|
}, |
|
{ |
|
"epoch": 75.26, |
|
"learning_rate": 4.987419768934532e-06, |
|
"loss": 1.0458, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 75.26, |
|
"eval_loss": 0.2246260643005371, |
|
"eval_runtime": 233.0094, |
|
"eval_samples_per_second": 35.788, |
|
"eval_steps_per_second": 1.12, |
|
"eval_wer": 0.2923436533217224, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 75.89, |
|
"learning_rate": 4.859306803594352e-06, |
|
"loss": 1.0382, |
|
"step": 59500 |
|
}, |
|
{ |
|
"epoch": 75.89, |
|
"eval_loss": 0.22546544671058655, |
|
"eval_runtime": 233.7397, |
|
"eval_samples_per_second": 35.676, |
|
"eval_steps_per_second": 1.117, |
|
"eval_wer": 0.2921603226619612, |
|
"step": 59500 |
|
}, |
|
{ |
|
"epoch": 76.53, |
|
"learning_rate": 4.730937098844673e-06, |
|
"loss": 1.0462, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 76.53, |
|
"eval_loss": 0.22583839297294617, |
|
"eval_runtime": 233.483, |
|
"eval_samples_per_second": 35.716, |
|
"eval_steps_per_second": 1.118, |
|
"eval_wer": 0.29539152554025255, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 77.17, |
|
"learning_rate": 4.602567394094994e-06, |
|
"loss": 1.0375, |
|
"step": 60500 |
|
}, |
|
{ |
|
"epoch": 77.17, |
|
"eval_loss": 0.22505535185337067, |
|
"eval_runtime": 239.1762, |
|
"eval_samples_per_second": 34.866, |
|
"eval_steps_per_second": 1.091, |
|
"eval_wer": 0.2928707289685359, |
|
"step": 60500 |
|
}, |
|
{ |
|
"epoch": 77.81, |
|
"learning_rate": 4.474197689345315e-06, |
|
"loss": 1.0332, |
|
"step": 61000 |
|
}, |
|
{ |
|
"epoch": 77.81, |
|
"eval_loss": 0.22773370146751404, |
|
"eval_runtime": 237.0288, |
|
"eval_samples_per_second": 35.181, |
|
"eval_steps_per_second": 1.101, |
|
"eval_wer": 0.2939936292595733, |
|
"step": 61000 |
|
}, |
|
{ |
|
"epoch": 78.44, |
|
"learning_rate": 4.345827984595636e-06, |
|
"loss": 1.0423, |
|
"step": 61500 |
|
}, |
|
{ |
|
"epoch": 78.44, |
|
"eval_loss": 0.22426965832710266, |
|
"eval_runtime": 233.5981, |
|
"eval_samples_per_second": 35.698, |
|
"eval_steps_per_second": 1.117, |
|
"eval_wer": 0.28961660975777437, |
|
"step": 61500 |
|
}, |
|
{ |
|
"epoch": 79.08, |
|
"learning_rate": 4.217458279845957e-06, |
|
"loss": 1.0379, |
|
"step": 62000 |
|
}, |
|
{ |
|
"epoch": 79.08, |
|
"eval_loss": 0.2274049073457718, |
|
"eval_runtime": 234.6064, |
|
"eval_samples_per_second": 35.545, |
|
"eval_steps_per_second": 1.113, |
|
"eval_wer": 0.29280197997112545, |
|
"step": 62000 |
|
}, |
|
{ |
|
"epoch": 79.72, |
|
"learning_rate": 4.089345314505777e-06, |
|
"loss": 1.0398, |
|
"step": 62500 |
|
}, |
|
{ |
|
"epoch": 79.72, |
|
"eval_loss": 0.22373217344284058, |
|
"eval_runtime": 234.7196, |
|
"eval_samples_per_second": 35.527, |
|
"eval_steps_per_second": 1.112, |
|
"eval_wer": 0.29280197997112545, |
|
"step": 62500 |
|
}, |
|
{ |
|
"epoch": 80.36, |
|
"learning_rate": 3.960975609756098e-06, |
|
"loss": 1.0395, |
|
"step": 63000 |
|
}, |
|
{ |
|
"epoch": 80.36, |
|
"eval_loss": 0.22652971744537354, |
|
"eval_runtime": 233.8582, |
|
"eval_samples_per_second": 35.658, |
|
"eval_steps_per_second": 1.116, |
|
"eval_wer": 0.2956206888649541, |
|
"step": 63000 |
|
}, |
|
{ |
|
"epoch": 80.99, |
|
"learning_rate": 3.832605905006419e-06, |
|
"loss": 1.0397, |
|
"step": 63500 |
|
}, |
|
{ |
|
"epoch": 80.99, |
|
"eval_loss": 0.22396665811538696, |
|
"eval_runtime": 235.3841, |
|
"eval_samples_per_second": 35.427, |
|
"eval_steps_per_second": 1.109, |
|
"eval_wer": 0.29202282466714025, |
|
"step": 63500 |
|
}, |
|
{ |
|
"epoch": 81.63, |
|
"learning_rate": 3.7042362002567396e-06, |
|
"loss": 1.0262, |
|
"step": 64000 |
|
}, |
|
{ |
|
"epoch": 81.63, |
|
"eval_loss": 0.22437626123428345, |
|
"eval_runtime": 239.3117, |
|
"eval_samples_per_second": 34.846, |
|
"eval_steps_per_second": 1.091, |
|
"eval_wer": 0.2933748882828792, |
|
"step": 64000 |
|
}, |
|
{ |
|
"epoch": 82.27, |
|
"learning_rate": 3.5758664955070607e-06, |
|
"loss": 1.0335, |
|
"step": 64500 |
|
}, |
|
{ |
|
"epoch": 82.27, |
|
"eval_loss": 0.22651708126068115, |
|
"eval_runtime": 234.0011, |
|
"eval_samples_per_second": 35.637, |
|
"eval_steps_per_second": 1.115, |
|
"eval_wer": 0.2936269679400509, |
|
"step": 64500 |
|
}, |
|
{ |
|
"epoch": 82.91, |
|
"learning_rate": 3.4474967907573813e-06, |
|
"loss": 1.0385, |
|
"step": 65000 |
|
}, |
|
{ |
|
"epoch": 82.91, |
|
"eval_loss": 0.22380606830120087, |
|
"eval_runtime": 236.2381, |
|
"eval_samples_per_second": 35.299, |
|
"eval_steps_per_second": 1.105, |
|
"eval_wer": 0.29284781263606574, |
|
"step": 65000 |
|
}, |
|
{ |
|
"epoch": 83.55, |
|
"learning_rate": 3.3191270860077024e-06, |
|
"loss": 1.0289, |
|
"step": 65500 |
|
}, |
|
{ |
|
"epoch": 83.55, |
|
"eval_loss": 0.221877783536911, |
|
"eval_runtime": 233.6081, |
|
"eval_samples_per_second": 35.697, |
|
"eval_steps_per_second": 1.117, |
|
"eval_wer": 0.2911520040332745, |
|
"step": 65500 |
|
}, |
|
{ |
|
"epoch": 84.18, |
|
"learning_rate": 3.1910141206675225e-06, |
|
"loss": 1.0372, |
|
"step": 66000 |
|
}, |
|
{ |
|
"epoch": 84.18, |
|
"eval_loss": 0.22357773780822754, |
|
"eval_runtime": 234.3938, |
|
"eval_samples_per_second": 35.577, |
|
"eval_steps_per_second": 1.114, |
|
"eval_wer": 0.2898457730824759, |
|
"step": 66000 |
|
}, |
|
{ |
|
"epoch": 84.82, |
|
"learning_rate": 3.0626444159178436e-06, |
|
"loss": 1.0279, |
|
"step": 66500 |
|
}, |
|
{ |
|
"epoch": 84.82, |
|
"eval_loss": 0.22185130417346954, |
|
"eval_runtime": 235.3542, |
|
"eval_samples_per_second": 35.432, |
|
"eval_steps_per_second": 1.109, |
|
"eval_wer": 0.2902124344019983, |
|
"step": 66500 |
|
}, |
|
{ |
|
"epoch": 85.46, |
|
"learning_rate": 2.9342747111681647e-06, |
|
"loss": 1.0325, |
|
"step": 67000 |
|
}, |
|
{ |
|
"epoch": 85.46, |
|
"eval_loss": 0.22399279475212097, |
|
"eval_runtime": 232.7577, |
|
"eval_samples_per_second": 35.827, |
|
"eval_steps_per_second": 1.121, |
|
"eval_wer": 0.2907853427137521, |
|
"step": 67000 |
|
}, |
|
{ |
|
"epoch": 86.1, |
|
"learning_rate": 2.8059050064184858e-06, |
|
"loss": 1.0202, |
|
"step": 67500 |
|
}, |
|
{ |
|
"epoch": 86.1, |
|
"eval_loss": 0.22061078250408173, |
|
"eval_runtime": 236.8657, |
|
"eval_samples_per_second": 35.206, |
|
"eval_steps_per_second": 1.102, |
|
"eval_wer": 0.28863120746155785, |
|
"step": 67500 |
|
}, |
|
{ |
|
"epoch": 86.73, |
|
"learning_rate": 2.6775353016688064e-06, |
|
"loss": 1.0166, |
|
"step": 68000 |
|
}, |
|
{ |
|
"epoch": 86.73, |
|
"eval_loss": 0.2218720018863678, |
|
"eval_runtime": 233.7184, |
|
"eval_samples_per_second": 35.68, |
|
"eval_steps_per_second": 1.117, |
|
"eval_wer": 0.2885624584641474, |
|
"step": 68000 |
|
}, |
|
{ |
|
"epoch": 87.37, |
|
"learning_rate": 2.5491655969191275e-06, |
|
"loss": 1.0259, |
|
"step": 68500 |
|
}, |
|
{ |
|
"epoch": 87.37, |
|
"eval_loss": 0.2234880030155182, |
|
"eval_runtime": 235.0768, |
|
"eval_samples_per_second": 35.474, |
|
"eval_steps_per_second": 1.11, |
|
"eval_wer": 0.28970827508765495, |
|
"step": 68500 |
|
}, |
|
{ |
|
"epoch": 88.01, |
|
"learning_rate": 2.420795892169448e-06, |
|
"loss": 1.0337, |
|
"step": 69000 |
|
}, |
|
{ |
|
"epoch": 88.01, |
|
"eval_loss": 0.220963254570961, |
|
"eval_runtime": 233.4278, |
|
"eval_samples_per_second": 35.724, |
|
"eval_steps_per_second": 1.118, |
|
"eval_wer": 0.2873249765107592, |
|
"step": 69000 |
|
}, |
|
{ |
|
"epoch": 88.65, |
|
"learning_rate": 2.2926829268292687e-06, |
|
"loss": 1.0264, |
|
"step": 69500 |
|
}, |
|
{ |
|
"epoch": 88.65, |
|
"eval_loss": 0.22157110273838043, |
|
"eval_runtime": 233.5356, |
|
"eval_samples_per_second": 35.708, |
|
"eval_steps_per_second": 1.118, |
|
"eval_wer": 0.2882187134770951, |
|
"step": 69500 |
|
}, |
|
{ |
|
"epoch": 89.29, |
|
"learning_rate": 2.1643132220795894e-06, |
|
"loss": 1.0231, |
|
"step": 70000 |
|
}, |
|
{ |
|
"epoch": 89.29, |
|
"eval_loss": 0.2222645878791809, |
|
"eval_runtime": 238.1657, |
|
"eval_samples_per_second": 35.013, |
|
"eval_steps_per_second": 1.096, |
|
"eval_wer": 0.2899145220798863, |
|
"step": 70000 |
|
}, |
|
{ |
|
"epoch": 89.92, |
|
"learning_rate": 2.03594351732991e-06, |
|
"loss": 1.0281, |
|
"step": 70500 |
|
}, |
|
{ |
|
"epoch": 89.92, |
|
"eval_loss": 0.22140374779701233, |
|
"eval_runtime": 233.9034, |
|
"eval_samples_per_second": 35.651, |
|
"eval_steps_per_second": 1.116, |
|
"eval_wer": 0.28723331118087864, |
|
"step": 70500 |
|
}, |
|
{ |
|
"epoch": 90.56, |
|
"learning_rate": 1.907573812580231e-06, |
|
"loss": 1.0135, |
|
"step": 71000 |
|
}, |
|
{ |
|
"epoch": 90.56, |
|
"eval_loss": 0.22179700434207916, |
|
"eval_runtime": 235.221, |
|
"eval_samples_per_second": 35.452, |
|
"eval_steps_per_second": 1.11, |
|
"eval_wer": 0.28684373352888604, |
|
"step": 71000 |
|
}, |
|
{ |
|
"epoch": 91.2, |
|
"learning_rate": 1.7792041078305522e-06, |
|
"loss": 1.0291, |
|
"step": 71500 |
|
}, |
|
{ |
|
"epoch": 91.2, |
|
"eval_loss": 0.22093360126018524, |
|
"eval_runtime": 236.8894, |
|
"eval_samples_per_second": 35.202, |
|
"eval_steps_per_second": 1.102, |
|
"eval_wer": 0.2862937415496024, |
|
"step": 71500 |
|
}, |
|
{ |
|
"epoch": 91.84, |
|
"learning_rate": 1.6508344030808731e-06, |
|
"loss": 1.0321, |
|
"step": 72000 |
|
}, |
|
{ |
|
"epoch": 91.84, |
|
"eval_loss": 0.219914972782135, |
|
"eval_runtime": 233.825, |
|
"eval_samples_per_second": 35.663, |
|
"eval_steps_per_second": 1.116, |
|
"eval_wer": 0.2875770561679309, |
|
"step": 72000 |
|
}, |
|
{ |
|
"epoch": 92.47, |
|
"learning_rate": 1.5227214377406932e-06, |
|
"loss": 1.028, |
|
"step": 72500 |
|
}, |
|
{ |
|
"epoch": 92.47, |
|
"eval_loss": 0.22144030034542084, |
|
"eval_runtime": 234.5883, |
|
"eval_samples_per_second": 35.547, |
|
"eval_steps_per_second": 1.113, |
|
"eval_wer": 0.28576666590278893, |
|
"step": 72500 |
|
}, |
|
{ |
|
"epoch": 93.11, |
|
"learning_rate": 1.3943517329910141e-06, |
|
"loss": 1.0213, |
|
"step": 73000 |
|
}, |
|
{ |
|
"epoch": 93.11, |
|
"eval_loss": 0.2219470739364624, |
|
"eval_runtime": 234.1897, |
|
"eval_samples_per_second": 35.608, |
|
"eval_steps_per_second": 1.114, |
|
"eval_wer": 0.2874624745055801, |
|
"step": 73000 |
|
}, |
|
{ |
|
"epoch": 93.75, |
|
"learning_rate": 1.2659820282413352e-06, |
|
"loss": 1.0261, |
|
"step": 73500 |
|
}, |
|
{ |
|
"epoch": 93.75, |
|
"eval_loss": 0.22315841913223267, |
|
"eval_runtime": 234.7799, |
|
"eval_samples_per_second": 35.518, |
|
"eval_steps_per_second": 1.112, |
|
"eval_wer": 0.28688956619382633, |
|
"step": 73500 |
|
}, |
|
{ |
|
"epoch": 94.39, |
|
"learning_rate": 1.137612323491656e-06, |
|
"loss": 1.0197, |
|
"step": 74000 |
|
}, |
|
{ |
|
"epoch": 94.39, |
|
"eval_loss": 0.22271272540092468, |
|
"eval_runtime": 233.7941, |
|
"eval_samples_per_second": 35.668, |
|
"eval_steps_per_second": 1.116, |
|
"eval_wer": 0.2866145702041845, |
|
"step": 74000 |
|
}, |
|
{ |
|
"epoch": 95.03, |
|
"learning_rate": 1.009242618741977e-06, |
|
"loss": 1.0298, |
|
"step": 74500 |
|
}, |
|
{ |
|
"epoch": 95.03, |
|
"eval_loss": 0.22284960746765137, |
|
"eval_runtime": 236.8784, |
|
"eval_samples_per_second": 35.204, |
|
"eval_steps_per_second": 1.102, |
|
"eval_wer": 0.28675206819900545, |
|
"step": 74500 |
|
}, |
|
{ |
|
"epoch": 95.66, |
|
"learning_rate": 8.808729139922979e-07, |
|
"loss": 1.0192, |
|
"step": 75000 |
|
}, |
|
{ |
|
"epoch": 95.66, |
|
"eval_loss": 0.22300392389297485, |
|
"eval_runtime": 233.5836, |
|
"eval_samples_per_second": 35.7, |
|
"eval_steps_per_second": 1.117, |
|
"eval_wer": 0.28652290487430393, |
|
"step": 75000 |
|
}, |
|
{ |
|
"epoch": 96.3, |
|
"learning_rate": 7.525032092426188e-07, |
|
"loss": 1.0156, |
|
"step": 75500 |
|
}, |
|
{ |
|
"epoch": 96.3, |
|
"eval_loss": 0.2220136970281601, |
|
"eval_runtime": 231.7898, |
|
"eval_samples_per_second": 35.977, |
|
"eval_steps_per_second": 1.126, |
|
"eval_wer": 0.2869353988587666, |
|
"step": 75500 |
|
}, |
|
{ |
|
"epoch": 96.94, |
|
"learning_rate": 6.241335044929398e-07, |
|
"loss": 1.0075, |
|
"step": 76000 |
|
}, |
|
{ |
|
"epoch": 96.94, |
|
"eval_loss": 0.2222670316696167, |
|
"eval_runtime": 232.6992, |
|
"eval_samples_per_second": 35.836, |
|
"eval_steps_per_second": 1.122, |
|
"eval_wer": 0.2866145702041845, |
|
"step": 76000 |
|
}, |
|
{ |
|
"epoch": 97.58, |
|
"learning_rate": 4.9602053915276e-07, |
|
"loss": 1.0201, |
|
"step": 76500 |
|
}, |
|
{ |
|
"epoch": 97.58, |
|
"eval_loss": 0.2218976616859436, |
|
"eval_runtime": 233.5288, |
|
"eval_samples_per_second": 35.709, |
|
"eval_steps_per_second": 1.118, |
|
"eval_wer": 0.28659165387171437, |
|
"step": 76500 |
|
}, |
|
{ |
|
"epoch": 98.21, |
|
"learning_rate": 3.676508344030809e-07, |
|
"loss": 1.0159, |
|
"step": 77000 |
|
}, |
|
{ |
|
"epoch": 98.21, |
|
"eval_loss": 0.22194130718708038, |
|
"eval_runtime": 233.434, |
|
"eval_samples_per_second": 35.723, |
|
"eval_steps_per_second": 1.118, |
|
"eval_wer": 0.28764580516534133, |
|
"step": 77000 |
|
}, |
|
{ |
|
"epoch": 98.85, |
|
"learning_rate": 2.395378690629012e-07, |
|
"loss": 1.0087, |
|
"step": 77500 |
|
}, |
|
{ |
|
"epoch": 98.85, |
|
"eval_loss": 0.2219318002462387, |
|
"eval_runtime": 235.7197, |
|
"eval_samples_per_second": 35.377, |
|
"eval_steps_per_second": 1.107, |
|
"eval_wer": 0.2873249765107592, |
|
"step": 77500 |
|
}, |
|
{ |
|
"epoch": 99.49, |
|
"learning_rate": 1.111681643132221e-07, |
|
"loss": 1.0159, |
|
"step": 78000 |
|
}, |
|
{ |
|
"epoch": 99.49, |
|
"eval_loss": 0.2223382443189621, |
|
"eval_runtime": 234.8374, |
|
"eval_samples_per_second": 35.51, |
|
"eval_steps_per_second": 1.111, |
|
"eval_wer": 0.2867291518665353, |
|
"step": 78000 |
|
}, |
|
{ |
|
"epoch": 100.0, |
|
"step": 78400, |
|
"total_flos": 1.850628803077507e+20, |
|
"train_loss": 1.2247810986577248, |
|
"train_runtime": 113701.5353, |
|
"train_samples_per_second": 22.038, |
|
"train_steps_per_second": 0.69 |
|
} |
|
], |
|
"max_steps": 78400, |
|
"num_train_epochs": 100, |
|
"total_flos": 1.850628803077507e+20, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|