|
{ |
|
"best_metric": 0.0955534353852272, |
|
"best_model_checkpoint": "facebook/wav2vec2-xlsr-53-espeak-cv-ft-clean/checkpoint-4000", |
|
"epoch": 7.680945347119645, |
|
"global_step": 5200, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.6380000000000002e-05, |
|
"loss": 1.8358, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 3.978e-05, |
|
"loss": 1.5868, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 6.318e-05, |
|
"loss": 1.4356, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 8.658e-05, |
|
"loss": 1.2539, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00010997999999999999, |
|
"loss": 1.2239, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00013337999999999997, |
|
"loss": 0.966, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00015678, |
|
"loss": 0.9195, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00018018, |
|
"loss": 0.744, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00020358, |
|
"loss": 0.6638, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00022698, |
|
"loss": 0.6554, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00023395957551826258, |
|
"loss": 0.5807, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00023390182625863772, |
|
"loss": 0.4771, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00023384407699901283, |
|
"loss": 0.4634, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00023378632773938794, |
|
"loss": 0.5991, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00023372857847976308, |
|
"loss": 0.5512, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.0002336708292201382, |
|
"loss": 0.4616, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0002336130799605133, |
|
"loss": 0.4028, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00023355533070088845, |
|
"loss": 0.386, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00023349758144126356, |
|
"loss": 0.3513, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00023343983218163868, |
|
"loss": 0.4561, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00023338208292201382, |
|
"loss": 0.3157, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00023332433366238893, |
|
"loss": 0.2983, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00023326658440276404, |
|
"loss": 0.3078, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00023320883514313918, |
|
"loss": 0.3487, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0002331510858835143, |
|
"loss": 0.4429, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0002330933366238894, |
|
"loss": 0.3194, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00023303558736426457, |
|
"loss": 0.3121, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0002329778381046397, |
|
"loss": 0.3229, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00023292008884501483, |
|
"loss": 0.317, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00023286233958538994, |
|
"loss": 0.3877, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023280459032576505, |
|
"loss": 0.2582, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.0002327468410661402, |
|
"loss": 0.2814, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.0002326890918065153, |
|
"loss": 0.2822, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00023263134254689042, |
|
"loss": 0.3077, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00023257359328726556, |
|
"loss": 0.3531, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00023251584402764067, |
|
"loss": 0.272, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00023245809476801578, |
|
"loss": 0.3182, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.00023240034550839092, |
|
"loss": 0.2482, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 0.00023234259624876604, |
|
"loss": 0.2565, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 0.00023228484698914115, |
|
"loss": 0.3474, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"eval_cer": 0.20053818094839443, |
|
"eval_loss": 0.12517954409122467, |
|
"eval_runtime": 167.3283, |
|
"eval_samples_per_second": 14.23, |
|
"eval_steps_per_second": 0.478, |
|
"eval_wer": 0.996220075598488, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 0.0002322270977295163, |
|
"loss": 0.2553, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 0.0002321693484698914, |
|
"loss": 0.29, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 0.00023211159921026654, |
|
"loss": 0.2583, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 0.00023205384995064166, |
|
"loss": 0.2701, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 0.00023199610069101677, |
|
"loss": 0.3269, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 0.0002319383514313919, |
|
"loss": 0.256, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 0.00023188060217176702, |
|
"loss": 0.277, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 0.00023182285291214214, |
|
"loss": 0.2547, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 0.00023176510365251728, |
|
"loss": 0.2853, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 0.0002317073543928924, |
|
"loss": 0.3382, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 0.0002316496051332675, |
|
"loss": 0.2393, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 0.00023159185587364264, |
|
"loss": 0.2374, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 0.00023153410661401775, |
|
"loss": 0.2067, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 0.00023147635735439287, |
|
"loss": 0.268, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 0.000231418608094768, |
|
"loss": 0.3156, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 0.00023136085883514312, |
|
"loss": 0.223, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 0.00023130310957551823, |
|
"loss": 0.2374, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 0.0002312453603158934, |
|
"loss": 0.2144, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 0.0002311876110562685, |
|
"loss": 0.2324, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 0.00023112986179664365, |
|
"loss": 0.3466, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 0.00023107211253701877, |
|
"loss": 0.2559, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 0.00023101436327739388, |
|
"loss": 0.2261, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 0.00023095661401776902, |
|
"loss": 0.2305, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 0.00023089886475814413, |
|
"loss": 0.2357, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 0.00023084111549851924, |
|
"loss": 0.356, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 0.00023078336623889438, |
|
"loss": 0.2131, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 0.0002307256169792695, |
|
"loss": 0.2462, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 0.0002306678677196446, |
|
"loss": 0.2411, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 0.00023061011846001975, |
|
"loss": 0.209, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 0.00023055236920039486, |
|
"loss": 0.1952, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 0.00023049461994076998, |
|
"loss": 0.1918, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 0.00023043687068114512, |
|
"loss": 0.2332, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 0.00023037912142152023, |
|
"loss": 0.2301, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 0.00023032137216189537, |
|
"loss": 0.2231, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 0.00023026362290227048, |
|
"loss": 0.2176, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 0.0002302058736426456, |
|
"loss": 0.1718, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 0.00023014812438302074, |
|
"loss": 0.213, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 0.00023009037512339585, |
|
"loss": 0.2865, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 0.00023003262586377096, |
|
"loss": 0.2347, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 0.0002299748766041461, |
|
"loss": 0.2116, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"eval_cer": 0.16925790827004725, |
|
"eval_loss": 0.10652817785739899, |
|
"eval_runtime": 166.4912, |
|
"eval_samples_per_second": 14.301, |
|
"eval_steps_per_second": 0.481, |
|
"eval_wer": 0.9886602267954641, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 0.0002299229022704837, |
|
"loss": 0.1885, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 0.00022986515301085883, |
|
"loss": 0.1934, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 0.00022980740375123395, |
|
"loss": 0.305, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 0.00022974965449160906, |
|
"loss": 0.2144, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 0.0002296919052319842, |
|
"loss": 0.2157, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 0.0002296341559723593, |
|
"loss": 0.2149, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 0.00022957640671273443, |
|
"loss": 0.2186, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 0.00022951865745310957, |
|
"loss": 0.2686, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 0.00022946668311944716, |
|
"loss": 0.2136, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 0.00022940893385982233, |
|
"loss": 0.1865, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 0.00022935118460019744, |
|
"loss": 0.1724, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 0.00022929343534057255, |
|
"loss": 0.1752, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 0.0002292356860809477, |
|
"loss": 0.2561, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 0.0002291779368213228, |
|
"loss": 0.1831, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 0.00022912018756169792, |
|
"loss": 0.1773, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 0.00022906243830207306, |
|
"loss": 0.1994, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 0.00022900468904244817, |
|
"loss": 0.1943, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 0.0002289469397828233, |
|
"loss": 0.27, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 0.00022888919052319842, |
|
"loss": 0.1959, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 0.00022883144126357354, |
|
"loss": 0.1822, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 0.00022877369200394868, |
|
"loss": 0.2212, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 0.0002287159427443238, |
|
"loss": 0.2142, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 0.0002286581934846989, |
|
"loss": 0.2741, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 0.00022860044422507404, |
|
"loss": 0.1909, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 0.00022854269496544916, |
|
"loss": 0.1991, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 0.00022848494570582427, |
|
"loss": 0.1728, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 0.0002284271964461994, |
|
"loss": 0.2023, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 0.00022836944718657452, |
|
"loss": 0.2307, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 0.00022831169792694963, |
|
"loss": 0.1708, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 0.00022825394866732477, |
|
"loss": 0.2045, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 0.0002281961994076999, |
|
"loss": 0.1962, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 0.00022813845014807503, |
|
"loss": 0.2261, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 0.00022808070088845014, |
|
"loss": 0.2592, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 0.00022802295162882525, |
|
"loss": 0.1846, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 0.0002279652023692004, |
|
"loss": 0.1764, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 0.0002279074531095755, |
|
"loss": 0.173, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 0.00022784970384995062, |
|
"loss": 0.1976, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 0.00022779195459032576, |
|
"loss": 0.2587, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 0.00022773420533070087, |
|
"loss": 0.171, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 0.00022767645607107599, |
|
"loss": 0.1942, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"eval_cer": 0.14407103988518807, |
|
"eval_loss": 0.09608560800552368, |
|
"eval_runtime": 165.7261, |
|
"eval_samples_per_second": 14.367, |
|
"eval_steps_per_second": 0.483, |
|
"eval_wer": 0.9865602687946241, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 0.00022761870681145113, |
|
"loss": 0.1729, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 0.00022756095755182627, |
|
"loss": 0.1849, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 0.00022750320829220138, |
|
"loss": 0.2318, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 0.00022744545903257652, |
|
"loss": 0.1915, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 0.00022738770977295163, |
|
"loss": 0.1969, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 0.00022732996051332674, |
|
"loss": 0.1764, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 0.00022727221125370188, |
|
"loss": 0.1896, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 0.000227214461994077, |
|
"loss": 0.2015, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 0.00022715671273445214, |
|
"loss": 0.1796, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 0.00022709896347482725, |
|
"loss": 0.205, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 0.00022704121421520236, |
|
"loss": 0.1773, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 0.0002269834649555775, |
|
"loss": 0.1917, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 0.00022692571569595262, |
|
"loss": 0.233, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 0.00022686796643632773, |
|
"loss": 0.1977, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 0.00022681021717670287, |
|
"loss": 0.176, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 0.00022675246791707798, |
|
"loss": 0.2263, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 0.0002266947186574531, |
|
"loss": 0.1504, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 0.00022663696939782823, |
|
"loss": 0.1572, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 0.00022657922013820335, |
|
"loss": 0.1765, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 0.00022652147087857846, |
|
"loss": 0.1634, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 0.0002264637216189536, |
|
"loss": 0.2366, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 0.00022640597235932871, |
|
"loss": 0.1596, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 0.00022634822309970385, |
|
"loss": 0.1419, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 0.00022629047384007897, |
|
"loss": 0.1785, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 0.00022623272458045408, |
|
"loss": 0.1351, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 0.00022617497532082922, |
|
"loss": 0.2157, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 0.00022611722606120433, |
|
"loss": 0.1547, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 0.00022605947680157945, |
|
"loss": 0.1938, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 0.00022600172754195459, |
|
"loss": 0.1561, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 0.0002259439782823297, |
|
"loss": 0.2033, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 0.0002258862290227048, |
|
"loss": 0.2139, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 0.00022582847976307995, |
|
"loss": 0.1894, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 0.0002257707305034551, |
|
"loss": 0.1765, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 0.0002257129812438302, |
|
"loss": 0.1716, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 0.00022565523198420534, |
|
"loss": 0.1856, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 0.00022559748272458046, |
|
"loss": 0.2059, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 0.00022553973346495557, |
|
"loss": 0.1688, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 0.0002254819842053307, |
|
"loss": 0.1618, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 0.00022542423494570582, |
|
"loss": 0.1745, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 0.00022536648568608096, |
|
"loss": 0.1825, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"eval_cer": 0.1480535789033068, |
|
"eval_loss": 0.1054612547159195, |
|
"eval_runtime": 165.9306, |
|
"eval_samples_per_second": 14.349, |
|
"eval_steps_per_second": 0.482, |
|
"eval_wer": 0.9861402771944561, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 0.00022530873642645608, |
|
"loss": 0.2008, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 0.0002252509871668312, |
|
"loss": 0.1951, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 0.00022519323790720633, |
|
"loss": 0.1537, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 0.00022513548864758144, |
|
"loss": 0.171, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 0.00022507773938795655, |
|
"loss": 0.1947, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 0.0002250199901283317, |
|
"loss": 0.1987, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 0.0002249622408687068, |
|
"loss": 0.1625, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 0.00022490449160908192, |
|
"loss": 0.1478, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 0.00022484674234945706, |
|
"loss": 0.1909, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 0.00022478899308983217, |
|
"loss": 0.1789, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 0.0002247312438302073, |
|
"loss": 0.1984, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 0.00022467349457058243, |
|
"loss": 0.1525, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 0.00022461574531095754, |
|
"loss": 0.1623, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 0.00022455799605133268, |
|
"loss": 0.1666, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 0.0002245002467917078, |
|
"loss": 0.1731, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 0.0002244424975320829, |
|
"loss": 0.211, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 0.00022438474827245805, |
|
"loss": 0.184, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 0.00022432699901283316, |
|
"loss": 0.1317, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 0.00022426924975320827, |
|
"loss": 0.1603, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 0.0002242115004935834, |
|
"loss": 0.186, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 0.00022415375123395852, |
|
"loss": 0.2378, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 0.00022409600197433364, |
|
"loss": 0.1734, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 0.00022403825271470878, |
|
"loss": 0.1453, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 0.0002239805034550839, |
|
"loss": 0.171, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 0.00022392275419545903, |
|
"loss": 0.2007, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 0.00022386500493583417, |
|
"loss": 0.2472, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 0.00022380725567620928, |
|
"loss": 0.1497, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 0.0002237495064165844, |
|
"loss": 0.1762, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 0.00022369175715695954, |
|
"loss": 0.142, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 0.00022363400789733465, |
|
"loss": 0.176, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 0.0002235762586377098, |
|
"loss": 0.2183, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 0.0002235185093780849, |
|
"loss": 0.1707, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 0.00022346076011846001, |
|
"loss": 0.1642, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 0.00022340301085883515, |
|
"loss": 0.1454, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 0.00022334526159921027, |
|
"loss": 0.1854, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 0.00022328751233958538, |
|
"loss": 0.2147, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 0.00022322976307996052, |
|
"loss": 0.1664, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 0.00022317201382033563, |
|
"loss": 0.1495, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 0.00022311426456071075, |
|
"loss": 0.1694, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 0.0002230565153010859, |
|
"loss": 0.1664, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"eval_cer": 0.15189858279016924, |
|
"eval_loss": 0.09597988426685333, |
|
"eval_runtime": 165.3281, |
|
"eval_samples_per_second": 14.402, |
|
"eval_steps_per_second": 0.484, |
|
"eval_wer": 0.9882402351952961, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 0.000222998766041461, |
|
"loss": 0.2587, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 0.0002229410167818361, |
|
"loss": 0.1783, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 0.00022288326752221125, |
|
"loss": 0.1742, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 0.00022282551826258637, |
|
"loss": 0.1699, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 0.0002227677690029615, |
|
"loss": 0.1536, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 0.00022271001974333662, |
|
"loss": 0.155, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 0.00022265227048371173, |
|
"loss": 0.1595, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"learning_rate": 0.00022259452122408687, |
|
"loss": 0.1566, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 0.00022253677196446198, |
|
"loss": 0.2007, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"learning_rate": 0.0002224790227048371, |
|
"loss": 0.1453, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"learning_rate": 0.00022242127344521224, |
|
"loss": 0.1786, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"learning_rate": 0.00022236352418558735, |
|
"loss": 0.1646, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"learning_rate": 0.00022230577492596246, |
|
"loss": 0.2085, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"learning_rate": 0.0002222480256663376, |
|
"loss": 0.1987, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"learning_rate": 0.00022219027640671272, |
|
"loss": 0.1479, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 0.00022213252714708783, |
|
"loss": 0.1693, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 0.000222074777887463, |
|
"loss": 0.1579, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 0.0002220170286278381, |
|
"loss": 0.1905, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 0.00022195927936821322, |
|
"loss": 0.1969, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"learning_rate": 0.00022190153010858836, |
|
"loss": 0.1504, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"learning_rate": 0.00022184378084896347, |
|
"loss": 0.1784, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 0.00022178603158933861, |
|
"loss": 0.1693, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"learning_rate": 0.00022172828232971373, |
|
"loss": 0.1799, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"learning_rate": 0.00022167053307008884, |
|
"loss": 0.19, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 0.00022161278381046398, |
|
"loss": 0.1471, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"learning_rate": 0.0002215550345508391, |
|
"loss": 0.154, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"learning_rate": 0.0002214972852912142, |
|
"loss": 0.1579, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 0.00022143953603158935, |
|
"loss": 0.2001, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"learning_rate": 0.00022138178677196446, |
|
"loss": 0.186, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 0.00022132403751233957, |
|
"loss": 0.1495, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 0.0002212662882527147, |
|
"loss": 0.1514, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"learning_rate": 0.00022120853899308983, |
|
"loss": 0.1803, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 0.00022115078973346494, |
|
"loss": 0.1838, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 0.00022109304047384008, |
|
"loss": 0.2408, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 0.0002210352912142152, |
|
"loss": 0.137, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 3.49, |
|
"learning_rate": 0.00022097754195459033, |
|
"loss": 0.1435, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 0.00022091979269496544, |
|
"loss": 0.1597, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 3.52, |
|
"learning_rate": 0.00022086204343534056, |
|
"loss": 0.2343, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 3.53, |
|
"learning_rate": 0.0002208042941757157, |
|
"loss": 0.1752, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"learning_rate": 0.0002207465449160908, |
|
"loss": 0.145, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"eval_cer": 0.16321832207139866, |
|
"eval_loss": 0.10200775414705276, |
|
"eval_runtime": 165.6528, |
|
"eval_samples_per_second": 14.373, |
|
"eval_steps_per_second": 0.483, |
|
"eval_wer": 0.989920201595968, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"learning_rate": 0.00022068879565646592, |
|
"loss": 0.1755, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"learning_rate": 0.00022063104639684106, |
|
"loss": 0.1808, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 3.59, |
|
"learning_rate": 0.00022057329713721618, |
|
"loss": 0.1764, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 3.6, |
|
"learning_rate": 0.0002205155478775913, |
|
"loss": 0.1693, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 3.62, |
|
"learning_rate": 0.00022045779861796643, |
|
"loss": 0.1472, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 3.63, |
|
"learning_rate": 0.00022040004935834154, |
|
"loss": 0.1485, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 3.65, |
|
"learning_rate": 0.00022034230009871665, |
|
"loss": 0.1508, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 3.66, |
|
"learning_rate": 0.00022028455083909182, |
|
"loss": 0.1928, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 3.68, |
|
"learning_rate": 0.00022022680157946693, |
|
"loss": 0.1844, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 3.69, |
|
"learning_rate": 0.00022016905231984205, |
|
"loss": 0.1618, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"learning_rate": 0.0002201113030602172, |
|
"loss": 0.1525, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 3.72, |
|
"learning_rate": 0.0002200535538005923, |
|
"loss": 0.1799, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 3.74, |
|
"learning_rate": 0.00021999580454096744, |
|
"loss": 0.203, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"learning_rate": 0.00021993805528134255, |
|
"loss": 0.1789, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 3.77, |
|
"learning_rate": 0.00021988030602171767, |
|
"loss": 0.1593, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 3.78, |
|
"learning_rate": 0.0002198225567620928, |
|
"loss": 0.1344, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 3.8, |
|
"learning_rate": 0.00021976480750246792, |
|
"loss": 0.1924, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 3.81, |
|
"learning_rate": 0.00021970705824284303, |
|
"loss": 0.2305, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 3.83, |
|
"learning_rate": 0.00021964930898321817, |
|
"loss": 0.1798, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 3.84, |
|
"learning_rate": 0.00021959155972359329, |
|
"loss": 0.1414, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 3.86, |
|
"learning_rate": 0.0002195338104639684, |
|
"loss": 0.1392, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 3.87, |
|
"learning_rate": 0.00021947606120434354, |
|
"loss": 0.1533, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"learning_rate": 0.00021941831194471865, |
|
"loss": 0.1981, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 3.9, |
|
"learning_rate": 0.00021936056268509376, |
|
"loss": 0.179, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"learning_rate": 0.0002193028134254689, |
|
"loss": 0.1409, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 3.93, |
|
"learning_rate": 0.00021924506416584402, |
|
"loss": 0.1624, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 3.94, |
|
"learning_rate": 0.00021918731490621916, |
|
"loss": 0.1619, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 3.96, |
|
"learning_rate": 0.00021912956564659427, |
|
"loss": 0.1727, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 3.97, |
|
"learning_rate": 0.00021907181638696938, |
|
"loss": 0.1602, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 3.99, |
|
"learning_rate": 0.00021901406712734452, |
|
"loss": 0.1295, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 0.00021895631786771964, |
|
"loss": 0.1825, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 0.00021889856860809475, |
|
"loss": 0.1724, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 0.0002188408193484699, |
|
"loss": 0.1503, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 4.05, |
|
"learning_rate": 0.000218783070088845, |
|
"loss": 0.1364, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 4.06, |
|
"learning_rate": 0.00021872532082922011, |
|
"loss": 0.144, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 4.08, |
|
"learning_rate": 0.00021866757156959525, |
|
"loss": 0.178, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 4.09, |
|
"learning_rate": 0.00021860982230997037, |
|
"loss": 0.1625, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 4.11, |
|
"learning_rate": 0.00021855207305034548, |
|
"loss": 0.1344, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 4.12, |
|
"learning_rate": 0.00021849432379072062, |
|
"loss": 0.1355, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 4.14, |
|
"learning_rate": 0.00021843657453109576, |
|
"loss": 0.1534, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 4.14, |
|
"eval_cer": 0.15161155295102555, |
|
"eval_loss": 0.10632243752479553, |
|
"eval_runtime": 166.0979, |
|
"eval_samples_per_second": 14.335, |
|
"eval_steps_per_second": 0.482, |
|
"eval_wer": 0.9764804703905922, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 4.15, |
|
"learning_rate": 0.00021837882527147087, |
|
"loss": 0.1826, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 4.17, |
|
"learning_rate": 0.000218321076011846, |
|
"loss": 0.1615, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 4.18, |
|
"learning_rate": 0.00021826332675222113, |
|
"loss": 0.1218, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 4.19, |
|
"learning_rate": 0.00021820557749259627, |
|
"loss": 0.1326, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 4.21, |
|
"learning_rate": 0.00021814782823297138, |
|
"loss": 0.1623, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 4.22, |
|
"learning_rate": 0.0002180900789733465, |
|
"loss": 0.1729, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 4.24, |
|
"learning_rate": 0.00021803232971372163, |
|
"loss": 0.1435, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 4.25, |
|
"learning_rate": 0.00021797458045409675, |
|
"loss": 0.1547, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 4.27, |
|
"learning_rate": 0.00021791683119447186, |
|
"loss": 0.156, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 4.28, |
|
"learning_rate": 0.000217859081934847, |
|
"loss": 0.1387, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 4.3, |
|
"learning_rate": 0.0002178013326752221, |
|
"loss": 0.2081, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 4.31, |
|
"learning_rate": 0.00021774358341559722, |
|
"loss": 0.1335, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 4.33, |
|
"learning_rate": 0.00021768583415597236, |
|
"loss": 0.1493, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 4.34, |
|
"learning_rate": 0.00021762808489634748, |
|
"loss": 0.1436, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 4.36, |
|
"learning_rate": 0.0002175703356367226, |
|
"loss": 0.1497, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 4.37, |
|
"learning_rate": 0.00021751258637709773, |
|
"loss": 0.1468, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 4.39, |
|
"learning_rate": 0.00021745483711747284, |
|
"loss": 0.1292, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 4.4, |
|
"learning_rate": 0.00021739708785784798, |
|
"loss": 0.1211, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 4.42, |
|
"learning_rate": 0.0002173393385982231, |
|
"loss": 0.1397, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 4.43, |
|
"learning_rate": 0.0002172815893385982, |
|
"loss": 0.1508, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 4.45, |
|
"learning_rate": 0.00021722384007897335, |
|
"loss": 0.2053, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 4.46, |
|
"learning_rate": 0.00021716609081934846, |
|
"loss": 0.1531, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 4.48, |
|
"learning_rate": 0.00021710834155972357, |
|
"loss": 0.132, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 4.49, |
|
"learning_rate": 0.00021705059230009871, |
|
"loss": 0.1469, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 4.51, |
|
"learning_rate": 0.00021699284304047383, |
|
"loss": 0.1471, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 4.52, |
|
"learning_rate": 0.00021693509378084894, |
|
"loss": 0.161, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 4.53, |
|
"learning_rate": 0.00021687734452122408, |
|
"loss": 0.1464, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 4.55, |
|
"learning_rate": 0.0002168195952615992, |
|
"loss": 0.1569, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 4.56, |
|
"learning_rate": 0.0002167618460019743, |
|
"loss": 0.1392, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 4.58, |
|
"learning_rate": 0.00021670409674234945, |
|
"loss": 0.1533, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 4.59, |
|
"learning_rate": 0.00021664634748272456, |
|
"loss": 0.1944, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 4.61, |
|
"learning_rate": 0.0002165885982230997, |
|
"loss": 0.1442, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 4.62, |
|
"learning_rate": 0.00021653084896347484, |
|
"loss": 0.1265, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 4.64, |
|
"learning_rate": 0.00021647309970384995, |
|
"loss": 0.1276, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 4.65, |
|
"learning_rate": 0.0002164153504442251, |
|
"loss": 0.1633, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 4.67, |
|
"learning_rate": 0.0002163576011846002, |
|
"loss": 0.1694, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 4.68, |
|
"learning_rate": 0.00021629985192497532, |
|
"loss": 0.1232, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 4.7, |
|
"learning_rate": 0.00021624210266535046, |
|
"loss": 0.1466, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 4.71, |
|
"learning_rate": 0.00021618435340572557, |
|
"loss": 0.1246, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 4.73, |
|
"learning_rate": 0.00021612660414610068, |
|
"loss": 0.1327, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 4.73, |
|
"eval_cer": 0.12719009747054955, |
|
"eval_loss": 0.09979347139596939, |
|
"eval_runtime": 166.6582, |
|
"eval_samples_per_second": 14.287, |
|
"eval_steps_per_second": 0.48, |
|
"eval_wer": 0.9764804703905922, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 4.74, |
|
"learning_rate": 0.00021606885488647582, |
|
"loss": 0.2253, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 4.76, |
|
"learning_rate": 0.00021601110562685094, |
|
"loss": 0.1384, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 4.77, |
|
"learning_rate": 0.00021595335636722605, |
|
"loss": 0.1735, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 4.79, |
|
"learning_rate": 0.0002158956071076012, |
|
"loss": 0.121, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 4.8, |
|
"learning_rate": 0.0002158378578479763, |
|
"loss": 0.1469, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 4.82, |
|
"learning_rate": 0.00021578010858835142, |
|
"loss": 0.1975, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 4.83, |
|
"learning_rate": 0.00021572235932872656, |
|
"loss": 0.133, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 4.84, |
|
"learning_rate": 0.00021566461006910167, |
|
"loss": 0.1317, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 4.86, |
|
"learning_rate": 0.0002156068608094768, |
|
"loss": 0.1458, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 4.87, |
|
"learning_rate": 0.00021554911154985192, |
|
"loss": 0.1423, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 4.89, |
|
"learning_rate": 0.00021549136229022703, |
|
"loss": 0.1852, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 4.9, |
|
"learning_rate": 0.00021543361303060217, |
|
"loss": 0.1516, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 4.92, |
|
"learning_rate": 0.0002153758637709773, |
|
"loss": 0.1418, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 4.93, |
|
"learning_rate": 0.0002153181145113524, |
|
"loss": 0.1275, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 4.95, |
|
"learning_rate": 0.00021526036525172754, |
|
"loss": 0.1199, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 4.96, |
|
"learning_rate": 0.00021520261599210265, |
|
"loss": 0.135, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 4.98, |
|
"learning_rate": 0.00021514486673247777, |
|
"loss": 0.1277, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 4.99, |
|
"learning_rate": 0.0002150871174728529, |
|
"loss": 0.1338, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"learning_rate": 0.00021502936821322802, |
|
"loss": 0.2228, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"learning_rate": 0.00021497161895360313, |
|
"loss": 0.138, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"learning_rate": 0.00021491386969397827, |
|
"loss": 0.1168, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 5.05, |
|
"learning_rate": 0.00021485612043435339, |
|
"loss": 0.1339, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 5.07, |
|
"learning_rate": 0.00021479837117472853, |
|
"loss": 0.1233, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 5.08, |
|
"learning_rate": 0.00021474062191510367, |
|
"loss": 0.1767, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 5.1, |
|
"learning_rate": 0.00021468287265547878, |
|
"loss": 0.126, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 5.11, |
|
"learning_rate": 0.00021462512339585392, |
|
"loss": 0.1149, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 5.13, |
|
"learning_rate": 0.00021456737413622903, |
|
"loss": 0.1379, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 5.14, |
|
"learning_rate": 0.00021450962487660414, |
|
"loss": 0.1406, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 5.16, |
|
"learning_rate": 0.00021445187561697928, |
|
"loss": 0.2308, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 5.17, |
|
"learning_rate": 0.0002143941263573544, |
|
"loss": 0.1305, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 5.18, |
|
"learning_rate": 0.0002143363770977295, |
|
"loss": 0.1516, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 5.2, |
|
"learning_rate": 0.00021427862783810465, |
|
"loss": 0.1331, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 5.21, |
|
"learning_rate": 0.00021422087857847976, |
|
"loss": 0.1383, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 5.23, |
|
"learning_rate": 0.00021416312931885488, |
|
"loss": 0.1725, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 5.24, |
|
"learning_rate": 0.00021410538005923002, |
|
"loss": 0.1231, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 5.26, |
|
"learning_rate": 0.00021404763079960513, |
|
"loss": 0.1221, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 5.27, |
|
"learning_rate": 0.00021398988153998024, |
|
"loss": 0.1338, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 5.29, |
|
"learning_rate": 0.00021393213228035538, |
|
"loss": 0.149, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 5.3, |
|
"learning_rate": 0.0002138743830207305, |
|
"loss": 0.1826, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 5.32, |
|
"learning_rate": 0.00021381663376110563, |
|
"loss": 0.1227, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 5.32, |
|
"eval_cer": 0.1120133947258267, |
|
"eval_loss": 0.10438049584627151, |
|
"eval_runtime": 166.8135, |
|
"eval_samples_per_second": 14.273, |
|
"eval_steps_per_second": 0.48, |
|
"eval_wer": 0.9706005879882402, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 5.33, |
|
"learning_rate": 0.00021375888450148075, |
|
"loss": 0.1549, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 5.35, |
|
"learning_rate": 0.00021370113524185586, |
|
"loss": 0.1344, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 5.36, |
|
"learning_rate": 0.000213643385982231, |
|
"loss": 0.1332, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 5.38, |
|
"learning_rate": 0.0002135856367226061, |
|
"loss": 0.1962, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 5.39, |
|
"learning_rate": 0.00021352788746298123, |
|
"loss": 0.1639, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 5.41, |
|
"learning_rate": 0.00021347013820335637, |
|
"loss": 0.1347, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 5.42, |
|
"learning_rate": 0.00021341238894373148, |
|
"loss": 0.1247, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 5.44, |
|
"learning_rate": 0.0002133546396841066, |
|
"loss": 0.1204, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 5.45, |
|
"learning_rate": 0.00021329689042448173, |
|
"loss": 0.1719, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 5.47, |
|
"learning_rate": 0.00021323914116485685, |
|
"loss": 0.1347, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 5.48, |
|
"learning_rate": 0.00021318139190523196, |
|
"loss": 0.1365, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 5.49, |
|
"learning_rate": 0.0002131236426456071, |
|
"loss": 0.1205, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 5.51, |
|
"learning_rate": 0.0002130658933859822, |
|
"loss": 0.115, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 5.52, |
|
"learning_rate": 0.00021300814412635735, |
|
"loss": 0.185, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 5.54, |
|
"learning_rate": 0.0002129503948667325, |
|
"loss": 0.142, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 5.55, |
|
"learning_rate": 0.0002128926456071076, |
|
"loss": 0.1271, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 5.57, |
|
"learning_rate": 0.00021283489634748274, |
|
"loss": 0.1131, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 5.58, |
|
"learning_rate": 0.00021277714708785786, |
|
"loss": 0.1687, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 5.6, |
|
"learning_rate": 0.00021271939782823297, |
|
"loss": 0.1643, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 5.61, |
|
"learning_rate": 0.0002126616485686081, |
|
"loss": 0.1241, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 5.63, |
|
"learning_rate": 0.00021260389930898322, |
|
"loss": 0.1178, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 5.64, |
|
"learning_rate": 0.00021254615004935834, |
|
"loss": 0.1287, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 5.66, |
|
"learning_rate": 0.00021248840078973348, |
|
"loss": 0.1464, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 5.67, |
|
"learning_rate": 0.0002124306515301086, |
|
"loss": 0.1721, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 5.69, |
|
"learning_rate": 0.0002123729022704837, |
|
"loss": 0.1249, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 5.7, |
|
"learning_rate": 0.00021231515301085884, |
|
"loss": 0.1276, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 5.72, |
|
"learning_rate": 0.00021225740375123395, |
|
"loss": 0.1106, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 5.73, |
|
"learning_rate": 0.00021219965449160907, |
|
"loss": 0.1401, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 5.75, |
|
"learning_rate": 0.0002121419052319842, |
|
"loss": 0.1725, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 5.76, |
|
"learning_rate": 0.00021208415597235932, |
|
"loss": 0.1278, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 5.78, |
|
"learning_rate": 0.00021202640671273446, |
|
"loss": 0.122, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 5.79, |
|
"learning_rate": 0.00021196865745310957, |
|
"loss": 0.1494, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 5.81, |
|
"learning_rate": 0.00021191090819348469, |
|
"loss": 0.123, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 5.82, |
|
"learning_rate": 0.00021185315893385983, |
|
"loss": 0.1845, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 5.83, |
|
"learning_rate": 0.00021179540967423494, |
|
"loss": 0.1123, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 5.85, |
|
"learning_rate": 0.00021173766041461005, |
|
"loss": 0.1372, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 5.86, |
|
"learning_rate": 0.0002116799111549852, |
|
"loss": 0.1328, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 5.88, |
|
"learning_rate": 0.0002116221618953603, |
|
"loss": 0.1381, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 5.89, |
|
"learning_rate": 0.00021156441263573542, |
|
"loss": 0.1655, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 5.91, |
|
"learning_rate": 0.00021150666337611056, |
|
"loss": 0.1214, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 5.91, |
|
"eval_cer": 0.17154218740656582, |
|
"eval_loss": 0.0955534353852272, |
|
"eval_runtime": 165.9473, |
|
"eval_samples_per_second": 14.348, |
|
"eval_steps_per_second": 0.482, |
|
"eval_wer": 0.9815203695926081, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 5.92, |
|
"learning_rate": 0.00021144891411648567, |
|
"loss": 0.1512, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 5.94, |
|
"learning_rate": 0.00021139116485686078, |
|
"loss": 0.1352, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 5.95, |
|
"learning_rate": 0.00021133341559723592, |
|
"loss": 0.1583, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 5.97, |
|
"learning_rate": 0.00021127566633761104, |
|
"loss": 0.1663, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 5.98, |
|
"learning_rate": 0.00021121791707798618, |
|
"loss": 0.1244, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"learning_rate": 0.0002111601678183613, |
|
"loss": 0.1249, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 6.01, |
|
"learning_rate": 0.00021110241855873643, |
|
"loss": 0.1476, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"learning_rate": 0.00021104466929911157, |
|
"loss": 0.1145, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 6.04, |
|
"learning_rate": 0.00021098692003948668, |
|
"loss": 0.125, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 6.06, |
|
"learning_rate": 0.0002109291707798618, |
|
"loss": 0.1257, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 6.07, |
|
"learning_rate": 0.00021087142152023694, |
|
"loss": 0.1312, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 6.09, |
|
"learning_rate": 0.00021081367226061205, |
|
"loss": 0.1667, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 6.1, |
|
"learning_rate": 0.00021075592300098716, |
|
"loss": 0.1385, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 6.12, |
|
"learning_rate": 0.0002106981737413623, |
|
"loss": 0.1052, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 6.13, |
|
"learning_rate": 0.00021064042448173741, |
|
"loss": 0.1436, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 6.14, |
|
"learning_rate": 0.00021058267522211253, |
|
"loss": 0.1238, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 6.16, |
|
"learning_rate": 0.00021052492596248767, |
|
"loss": 0.1597, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 6.17, |
|
"learning_rate": 0.00021046717670286278, |
|
"loss": 0.1195, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 6.19, |
|
"learning_rate": 0.0002104094274432379, |
|
"loss": 0.1236, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 6.2, |
|
"learning_rate": 0.00021035167818361303, |
|
"loss": 0.131, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 6.22, |
|
"learning_rate": 0.00021029392892398815, |
|
"loss": 0.1599, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 6.23, |
|
"learning_rate": 0.00021023617966436329, |
|
"loss": 0.1528, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 6.25, |
|
"learning_rate": 0.0002101784304047384, |
|
"loss": 0.1046, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 6.26, |
|
"learning_rate": 0.0002101206811451135, |
|
"loss": 0.1251, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 6.28, |
|
"learning_rate": 0.00021006293188548865, |
|
"loss": 0.1403, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 6.29, |
|
"learning_rate": 0.00021000518262586376, |
|
"loss": 0.1351, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 6.31, |
|
"learning_rate": 0.00020994743336623888, |
|
"loss": 0.1858, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 6.32, |
|
"learning_rate": 0.00020988968410661402, |
|
"loss": 0.1426, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 6.34, |
|
"learning_rate": 0.00020983193484698913, |
|
"loss": 0.121, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 6.35, |
|
"learning_rate": 0.00020977418558736424, |
|
"loss": 0.1312, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 6.37, |
|
"learning_rate": 0.00020971643632773938, |
|
"loss": 0.1339, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 6.38, |
|
"learning_rate": 0.0002096586870681145, |
|
"loss": 0.1572, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 6.4, |
|
"learning_rate": 0.0002096009378084896, |
|
"loss": 0.1363, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 6.41, |
|
"learning_rate": 0.00020954318854886475, |
|
"loss": 0.1165, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 6.43, |
|
"learning_rate": 0.00020948543928923986, |
|
"loss": 0.1348, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 6.44, |
|
"learning_rate": 0.000209427690029615, |
|
"loss": 0.1251, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 6.45, |
|
"learning_rate": 0.00020936994076999012, |
|
"loss": 0.1619, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 6.47, |
|
"learning_rate": 0.00020931219151036526, |
|
"loss": 0.1307, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 6.48, |
|
"learning_rate": 0.0002092544422507404, |
|
"loss": 0.1258, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 6.5, |
|
"learning_rate": 0.0002091966929911155, |
|
"loss": 0.1169, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 6.5, |
|
"eval_cer": 0.1349219637624828, |
|
"eval_loss": 0.0982118621468544, |
|
"eval_runtime": 165.5517, |
|
"eval_samples_per_second": 14.382, |
|
"eval_steps_per_second": 0.483, |
|
"eval_wer": 0.9815203695926081, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 6.51, |
|
"learning_rate": 0.00020913894373149062, |
|
"loss": 0.1491, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 6.53, |
|
"learning_rate": 0.00020908119447186576, |
|
"loss": 0.1562, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 6.54, |
|
"learning_rate": 0.00020902344521224087, |
|
"loss": 0.1651, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 6.56, |
|
"learning_rate": 0.000208965695952616, |
|
"loss": 0.1447, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 6.57, |
|
"learning_rate": 0.00020890794669299113, |
|
"loss": 0.1284, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 6.59, |
|
"learning_rate": 0.00020885019743336624, |
|
"loss": 0.1374, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 6.6, |
|
"learning_rate": 0.00020879244817374135, |
|
"loss": 0.1388, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 6.62, |
|
"learning_rate": 0.0002087346989141165, |
|
"loss": 0.1255, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 6.63, |
|
"learning_rate": 0.0002086769496544916, |
|
"loss": 0.1194, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 6.65, |
|
"learning_rate": 0.00020861920039486672, |
|
"loss": 0.1333, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 6.66, |
|
"learning_rate": 0.00020856145113524186, |
|
"loss": 0.1433, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 6.68, |
|
"learning_rate": 0.00020850370187561697, |
|
"loss": 0.143, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 6.69, |
|
"learning_rate": 0.0002084459526159921, |
|
"loss": 0.1392, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 6.71, |
|
"learning_rate": 0.00020838820335636722, |
|
"loss": 0.1325, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 6.72, |
|
"learning_rate": 0.00020833045409674234, |
|
"loss": 0.1197, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 6.74, |
|
"learning_rate": 0.00020827270483711748, |
|
"loss": 0.1346, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 6.75, |
|
"learning_rate": 0.0002082149555774926, |
|
"loss": 0.1403, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 6.77, |
|
"learning_rate": 0.0002081572063178677, |
|
"loss": 0.1181, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 6.78, |
|
"learning_rate": 0.00020809945705824284, |
|
"loss": 0.1239, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 6.79, |
|
"learning_rate": 0.00020804170779861796, |
|
"loss": 0.1675, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 6.81, |
|
"learning_rate": 0.00020798395853899307, |
|
"loss": 0.1435, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 6.82, |
|
"learning_rate": 0.0002079262092793682, |
|
"loss": 0.1441, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 6.84, |
|
"learning_rate": 0.00020786846001974332, |
|
"loss": 0.1482, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 6.85, |
|
"learning_rate": 0.00020781071076011844, |
|
"loss": 0.1217, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 6.87, |
|
"learning_rate": 0.00020775296150049358, |
|
"loss": 0.1092, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 6.88, |
|
"learning_rate": 0.0002076952122408687, |
|
"loss": 0.1154, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 6.9, |
|
"learning_rate": 0.00020763746298124383, |
|
"loss": 0.1945, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 6.91, |
|
"learning_rate": 0.00020757971372161894, |
|
"loss": 0.1312, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 6.93, |
|
"learning_rate": 0.00020752196446199405, |
|
"loss": 0.1162, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 6.94, |
|
"learning_rate": 0.00020746421520236922, |
|
"loss": 0.1188, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 6.96, |
|
"learning_rate": 0.00020740646594274433, |
|
"loss": 0.1391, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 6.97, |
|
"learning_rate": 0.00020734871668311945, |
|
"loss": 0.1672, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 6.99, |
|
"learning_rate": 0.0002072909674234946, |
|
"loss": 0.1429, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"learning_rate": 0.0002072332181638697, |
|
"loss": 0.1438, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 0.0002071754689042448, |
|
"loss": 0.1274, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 7.03, |
|
"learning_rate": 0.00020711771964461995, |
|
"loss": 0.1086, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 7.05, |
|
"learning_rate": 0.00020705997038499507, |
|
"loss": 0.1207, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 7.06, |
|
"learning_rate": 0.00020700222112537018, |
|
"loss": 0.1179, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 7.08, |
|
"learning_rate": 0.00020694447186574532, |
|
"loss": 0.1464, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 7.09, |
|
"learning_rate": 0.00020688672260612043, |
|
"loss": 0.1115, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 7.09, |
|
"eval_cer": 0.15084614004664235, |
|
"eval_loss": 0.10529889166355133, |
|
"eval_runtime": 167.4152, |
|
"eval_samples_per_second": 14.222, |
|
"eval_steps_per_second": 0.478, |
|
"eval_wer": 0.9844603107937842, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 7.1, |
|
"learning_rate": 0.00020682897334649554, |
|
"loss": 0.1201, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 7.12, |
|
"learning_rate": 0.00020677122408687068, |
|
"loss": 0.1359, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 7.13, |
|
"learning_rate": 0.0002067134748272458, |
|
"loss": 0.1161, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 7.15, |
|
"learning_rate": 0.00020665572556762094, |
|
"loss": 0.1579, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 7.16, |
|
"learning_rate": 0.00020659797630799605, |
|
"loss": 0.1196, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 7.18, |
|
"learning_rate": 0.00020654022704837116, |
|
"loss": 0.1122, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 7.19, |
|
"learning_rate": 0.0002064824777887463, |
|
"loss": 0.1323, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 7.21, |
|
"learning_rate": 0.00020642472852912142, |
|
"loss": 0.14, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 7.22, |
|
"learning_rate": 0.00020636697926949653, |
|
"loss": 0.1683, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 7.24, |
|
"learning_rate": 0.00020630923000987167, |
|
"loss": 0.1203, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 7.25, |
|
"learning_rate": 0.00020625148075024678, |
|
"loss": 0.1275, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 7.27, |
|
"learning_rate": 0.0002061937314906219, |
|
"loss": 0.1206, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 7.28, |
|
"learning_rate": 0.00020613598223099704, |
|
"loss": 0.127, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 7.3, |
|
"learning_rate": 0.00020607823297137215, |
|
"loss": 0.1551, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 7.31, |
|
"learning_rate": 0.00020602048371174726, |
|
"loss": 0.1446, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 7.33, |
|
"learning_rate": 0.0002059627344521224, |
|
"loss": 0.1283, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 7.34, |
|
"learning_rate": 0.00020590498519249751, |
|
"loss": 0.1253, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 7.36, |
|
"learning_rate": 0.00020584723593287265, |
|
"loss": 0.1143, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 7.37, |
|
"learning_rate": 0.00020578948667324777, |
|
"loss": 0.1261, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 7.39, |
|
"learning_rate": 0.00020573173741362288, |
|
"loss": 0.1282, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 7.4, |
|
"learning_rate": 0.00020567398815399802, |
|
"loss": 0.1288, |
|
"step": 5010 |
|
}, |
|
{ |
|
"epoch": 7.42, |
|
"learning_rate": 0.00020561623889437316, |
|
"loss": 0.105, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 7.43, |
|
"learning_rate": 0.00020555848963474827, |
|
"loss": 0.1218, |
|
"step": 5030 |
|
}, |
|
{ |
|
"epoch": 7.44, |
|
"learning_rate": 0.0002055007403751234, |
|
"loss": 0.1393, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 7.46, |
|
"learning_rate": 0.00020544299111549853, |
|
"loss": 0.1122, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 7.47, |
|
"learning_rate": 0.00020538524185587364, |
|
"loss": 0.1103, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 7.49, |
|
"learning_rate": 0.00020532749259624878, |
|
"loss": 0.0996, |
|
"step": 5070 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"learning_rate": 0.0002052697433366239, |
|
"loss": 0.1136, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 7.52, |
|
"learning_rate": 0.000205211994076999, |
|
"loss": 0.1593, |
|
"step": 5090 |
|
}, |
|
{ |
|
"epoch": 7.53, |
|
"learning_rate": 0.00020515424481737414, |
|
"loss": 0.1223, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 7.55, |
|
"learning_rate": 0.00020509649555774926, |
|
"loss": 0.1164, |
|
"step": 5110 |
|
}, |
|
{ |
|
"epoch": 7.56, |
|
"learning_rate": 0.00020503874629812437, |
|
"loss": 0.1101, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 7.58, |
|
"learning_rate": 0.0002049809970384995, |
|
"loss": 0.1082, |
|
"step": 5130 |
|
}, |
|
{ |
|
"epoch": 7.59, |
|
"learning_rate": 0.00020492324777887462, |
|
"loss": 0.1355, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 7.61, |
|
"learning_rate": 0.00020486549851924976, |
|
"loss": 0.1187, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 7.62, |
|
"learning_rate": 0.00020480774925962488, |
|
"loss": 0.1147, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 7.64, |
|
"learning_rate": 0.00020475, |
|
"loss": 0.1072, |
|
"step": 5170 |
|
}, |
|
{ |
|
"epoch": 7.65, |
|
"learning_rate": 0.00020469225074037513, |
|
"loss": 0.1298, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 7.67, |
|
"learning_rate": 0.00020463450148075024, |
|
"loss": 0.1563, |
|
"step": 5190 |
|
}, |
|
{ |
|
"epoch": 7.68, |
|
"learning_rate": 0.00020457675222112536, |
|
"loss": 0.1066, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 7.68, |
|
"eval_cer": 0.10776774502182623, |
|
"eval_loss": 0.09926649183034897, |
|
"eval_runtime": 165.9352, |
|
"eval_samples_per_second": 14.349, |
|
"eval_steps_per_second": 0.482, |
|
"eval_wer": 0.9601007979840404, |
|
"step": 5200 |
|
} |
|
], |
|
"max_steps": 40620, |
|
"num_train_epochs": 60, |
|
"total_flos": 5.489095218329608e+19, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|