|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.0, |
|
"global_step": 15012, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.998667732480682e-05, |
|
"loss": 3.8577, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.997335464961364e-05, |
|
"loss": 1.113, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.9960031974420465e-05, |
|
"loss": 1.2262, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.994670929922729e-05, |
|
"loss": 1.5109, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.993338662403411e-05, |
|
"loss": 1.5867, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.992006394884093e-05, |
|
"loss": 1.342, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.990674127364775e-05, |
|
"loss": 1.2, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.989341859845457e-05, |
|
"loss": 1.0629, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.9880095923261392e-05, |
|
"loss": 1.515, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.9866773248068215e-05, |
|
"loss": 1.1302, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.9853450572875035e-05, |
|
"loss": 1.0941, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.9840127897681855e-05, |
|
"loss": 0.9305, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.9826805222488675e-05, |
|
"loss": 0.9049, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.98134825472955e-05, |
|
"loss": 1.3378, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.980015987210232e-05, |
|
"loss": 1.5296, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.9786837196909142e-05, |
|
"loss": 0.9969, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.9773514521715962e-05, |
|
"loss": 1.0424, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.9760191846522782e-05, |
|
"loss": 1.1835, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.9746869171329606e-05, |
|
"loss": 0.8778, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.9733546496136426e-05, |
|
"loss": 1.0943, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.9720223820943246e-05, |
|
"loss": 1.3601, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.970690114575007e-05, |
|
"loss": 1.2044, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.969357847055689e-05, |
|
"loss": 1.0574, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.968025579536371e-05, |
|
"loss": 1.1926, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.9666933120170533e-05, |
|
"loss": 1.1607, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.9653610444977353e-05, |
|
"loss": 0.9255, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.9640287769784173e-05, |
|
"loss": 1.3423, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.9626965094590996e-05, |
|
"loss": 1.2185, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.9613642419397816e-05, |
|
"loss": 0.8564, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.960031974420464e-05, |
|
"loss": 1.0111, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.958699706901146e-05, |
|
"loss": 1.1537, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.957367439381828e-05, |
|
"loss": 0.8285, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.95603517186251e-05, |
|
"loss": 0.7506, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.9547029043431923e-05, |
|
"loss": 0.9619, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.9533706368238743e-05, |
|
"loss": 1.1323, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.9520383693045567e-05, |
|
"loss": 1.1748, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.9507061017852387e-05, |
|
"loss": 1.0394, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.9493738342659207e-05, |
|
"loss": 1.4735, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.948041566746603e-05, |
|
"loss": 1.0793, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.946709299227285e-05, |
|
"loss": 1.1222, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.9453770317079673e-05, |
|
"loss": 1.2979, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.9440447641886493e-05, |
|
"loss": 1.2262, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.9427124966693313e-05, |
|
"loss": 0.7893, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.9413802291500133e-05, |
|
"loss": 1.0507, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.9400479616306957e-05, |
|
"loss": 1.0554, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.9387156941113777e-05, |
|
"loss": 0.8673, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.93738342659206e-05, |
|
"loss": 0.8038, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.936051159072742e-05, |
|
"loss": 1.1252, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.934718891553424e-05, |
|
"loss": 0.798, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.933386624034106e-05, |
|
"loss": 1.0203, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.9320543565147884e-05, |
|
"loss": 1.1765, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.9307220889954704e-05, |
|
"loss": 1.1677, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.9293898214761527e-05, |
|
"loss": 0.959, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.9280575539568347e-05, |
|
"loss": 1.2401, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.9267252864375167e-05, |
|
"loss": 1.1293, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.9253930189181987e-05, |
|
"loss": 1.0167, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.924060751398881e-05, |
|
"loss": 1.0203, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.922728483879563e-05, |
|
"loss": 1.01, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.9213962163602454e-05, |
|
"loss": 0.8721, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.9200639488409274e-05, |
|
"loss": 0.9725, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.9187316813216094e-05, |
|
"loss": 1.1241, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.9173994138022914e-05, |
|
"loss": 1.0535, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.9160671462829738e-05, |
|
"loss": 0.9679, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.914734878763656e-05, |
|
"loss": 1.1622, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.913402611244338e-05, |
|
"loss": 1.1936, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.91207034372502e-05, |
|
"loss": 0.771, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.910738076205702e-05, |
|
"loss": 0.6295, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.9094058086863845e-05, |
|
"loss": 1.0405, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.9080735411670665e-05, |
|
"loss": 1.2107, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.9067412736477488e-05, |
|
"loss": 0.9424, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.9054090061284308e-05, |
|
"loss": 1.1574, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.9040767386091128e-05, |
|
"loss": 0.7862, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.9027444710897948e-05, |
|
"loss": 1.1039, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.901412203570477e-05, |
|
"loss": 0.9475, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.900079936051159e-05, |
|
"loss": 0.7481, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.8987476685318415e-05, |
|
"loss": 0.7635, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.8974154010125235e-05, |
|
"loss": 0.9573, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.8960831334932055e-05, |
|
"loss": 0.9435, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.8947508659738878e-05, |
|
"loss": 0.8267, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.89341859845457e-05, |
|
"loss": 0.9648, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.892086330935252e-05, |
|
"loss": 0.6978, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.8907540634159342e-05, |
|
"loss": 0.7828, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.8894217958966162e-05, |
|
"loss": 1.3017, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.8880895283772985e-05, |
|
"loss": 1.0878, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.8867572608579805e-05, |
|
"loss": 0.8126, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.8854249933386625e-05, |
|
"loss": 1.1181, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.8840927258193445e-05, |
|
"loss": 1.1542, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.882760458300027e-05, |
|
"loss": 0.7235, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.881428190780709e-05, |
|
"loss": 0.9951, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.8800959232613912e-05, |
|
"loss": 0.957, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.8787636557420732e-05, |
|
"loss": 1.2426, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.8774313882227552e-05, |
|
"loss": 0.8761, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.8760991207034372e-05, |
|
"loss": 1.0079, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.8747668531841196e-05, |
|
"loss": 1.1776, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.873434585664802e-05, |
|
"loss": 1.1795, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.872102318145484e-05, |
|
"loss": 1.0256, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.870770050626166e-05, |
|
"loss": 0.9522, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.869437783106848e-05, |
|
"loss": 0.6613, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.86810551558753e-05, |
|
"loss": 1.5441, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.8667732480682123e-05, |
|
"loss": 1.221, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.8654409805488946e-05, |
|
"loss": 0.8684, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.8641087130295766e-05, |
|
"loss": 0.8281, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.8627764455102586e-05, |
|
"loss": 0.8654, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.8614441779909406e-05, |
|
"loss": 1.2252, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.8601119104716226e-05, |
|
"loss": 0.8553, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.858779642952305e-05, |
|
"loss": 0.9223, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.8574473754329873e-05, |
|
"loss": 1.1264, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.8561151079136693e-05, |
|
"loss": 1.1312, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.8547828403943513e-05, |
|
"loss": 0.9532, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.8534505728750333e-05, |
|
"loss": 1.1151, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.8521183053557156e-05, |
|
"loss": 0.8711, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.8507860378363976e-05, |
|
"loss": 1.0257, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.84945377031708e-05, |
|
"loss": 1.0568, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.848121502797762e-05, |
|
"loss": 1.0497, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.846789235278444e-05, |
|
"loss": 0.8945, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.845456967759126e-05, |
|
"loss": 1.0013, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.8441247002398083e-05, |
|
"loss": 1.3454, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.8427924327204903e-05, |
|
"loss": 0.9102, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.8414601652011727e-05, |
|
"loss": 1.1129, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.8401278976818547e-05, |
|
"loss": 1.0036, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.8387956301625367e-05, |
|
"loss": 0.9195, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.837463362643219e-05, |
|
"loss": 0.8959, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.836131095123901e-05, |
|
"loss": 1.0366, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.834798827604583e-05, |
|
"loss": 1.3046, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.8334665600852654e-05, |
|
"loss": 0.8456, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.8321342925659474e-05, |
|
"loss": 0.8431, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.8308020250466294e-05, |
|
"loss": 0.7821, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.8294697575273117e-05, |
|
"loss": 1.1036, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.8281374900079937e-05, |
|
"loss": 0.6821, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.8268052224886757e-05, |
|
"loss": 0.9491, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.825472954969358e-05, |
|
"loss": 0.9321, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.82414068745004e-05, |
|
"loss": 1.08, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.8228084199307224e-05, |
|
"loss": 0.9724, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.8214761524114044e-05, |
|
"loss": 0.9978, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.8201438848920864e-05, |
|
"loss": 1.1282, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.8188116173727687e-05, |
|
"loss": 1.3154, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.8174793498534507e-05, |
|
"loss": 0.9287, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.816147082334133e-05, |
|
"loss": 1.0567, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.814814814814815e-05, |
|
"loss": 1.0243, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.813482547295497e-05, |
|
"loss": 0.6169, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.812150279776179e-05, |
|
"loss": 0.9887, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.8108180122568614e-05, |
|
"loss": 0.9247, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.8094857447375434e-05, |
|
"loss": 0.7982, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.8081534772182258e-05, |
|
"loss": 0.9119, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.8068212096989078e-05, |
|
"loss": 0.9946, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.8054889421795898e-05, |
|
"loss": 0.732, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.8041566746602718e-05, |
|
"loss": 1.1512, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.802824407140954e-05, |
|
"loss": 1.1482, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.801492139621636e-05, |
|
"loss": 1.1026, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.8001598721023185e-05, |
|
"loss": 1.0065, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.7988276045830005e-05, |
|
"loss": 0.9553, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.7974953370636825e-05, |
|
"loss": 1.1751, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.7961630695443645e-05, |
|
"loss": 0.954, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.7948308020250468e-05, |
|
"loss": 0.9903, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.7934985345057288e-05, |
|
"loss": 1.2137, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.792166266986411e-05, |
|
"loss": 1.2528, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.790833999467093e-05, |
|
"loss": 1.2044, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.789501731947775e-05, |
|
"loss": 1.1712, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.788169464428457e-05, |
|
"loss": 0.8754, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.7868371969091395e-05, |
|
"loss": 1.2087, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.7855049293898215e-05, |
|
"loss": 0.8691, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.784172661870504e-05, |
|
"loss": 0.7455, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.782840394351186e-05, |
|
"loss": 0.7186, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.781508126831868e-05, |
|
"loss": 0.8787, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.7801758593125502e-05, |
|
"loss": 1.0877, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.7788435917932322e-05, |
|
"loss": 1.0954, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.7775113242739145e-05, |
|
"loss": 1.0374, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.7761790567545965e-05, |
|
"loss": 0.6339, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.7748467892352785e-05, |
|
"loss": 0.8952, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.7735145217159605e-05, |
|
"loss": 0.8032, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.772182254196643e-05, |
|
"loss": 0.8893, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.770849986677325e-05, |
|
"loss": 0.9514, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.7695177191580072e-05, |
|
"loss": 0.9611, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.7681854516386892e-05, |
|
"loss": 0.8059, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.7668531841193712e-05, |
|
"loss": 0.6649, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.7655209166000536e-05, |
|
"loss": 1.0976, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.7641886490807356e-05, |
|
"loss": 0.7004, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.7628563815614176e-05, |
|
"loss": 1.4345, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.7615241140421e-05, |
|
"loss": 0.8439, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.760191846522782e-05, |
|
"loss": 0.7677, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.758859579003464e-05, |
|
"loss": 0.7943, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.7575273114841463e-05, |
|
"loss": 0.9815, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.7561950439648283e-05, |
|
"loss": 1.0038, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.7548627764455103e-05, |
|
"loss": 1.4571, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.7535305089261926e-05, |
|
"loss": 0.7428, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.7521982414068746e-05, |
|
"loss": 1.1553, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.750865973887557e-05, |
|
"loss": 1.1216, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.749533706368239e-05, |
|
"loss": 0.8908, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.748201438848921e-05, |
|
"loss": 0.8248, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.746869171329603e-05, |
|
"loss": 0.893, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.7455369038102853e-05, |
|
"loss": 1.123, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.7442046362909676e-05, |
|
"loss": 0.9023, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.7428723687716496e-05, |
|
"loss": 0.8609, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.7415401012523316e-05, |
|
"loss": 1.041, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.7402078337330136e-05, |
|
"loss": 0.9551, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.7388755662136956e-05, |
|
"loss": 0.9415, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.737543298694378e-05, |
|
"loss": 1.0175, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.7362110311750603e-05, |
|
"loss": 0.9307, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.7348787636557423e-05, |
|
"loss": 1.1073, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.7335464961364243e-05, |
|
"loss": 0.8744, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.7322142286171063e-05, |
|
"loss": 0.9754, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.7308819610977883e-05, |
|
"loss": 1.2237, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.7295496935784707e-05, |
|
"loss": 1.0859, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.728217426059153e-05, |
|
"loss": 0.7246, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.726885158539835e-05, |
|
"loss": 1.1179, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.725552891020517e-05, |
|
"loss": 1.0577, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.724220623501199e-05, |
|
"loss": 0.8672, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.722888355981881e-05, |
|
"loss": 1.0886, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.7215560884625634e-05, |
|
"loss": 0.9956, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.7202238209432457e-05, |
|
"loss": 1.1679, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.7188915534239277e-05, |
|
"loss": 0.7628, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.7175592859046097e-05, |
|
"loss": 1.0862, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.7162270183852917e-05, |
|
"loss": 1.1052, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.714894750865974e-05, |
|
"loss": 0.977, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.713562483346656e-05, |
|
"loss": 1.0277, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.7122302158273384e-05, |
|
"loss": 0.8481, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.7108979483080204e-05, |
|
"loss": 0.9281, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.7095656807887024e-05, |
|
"loss": 0.8152, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.7082334132693844e-05, |
|
"loss": 0.8787, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.7069011457500668e-05, |
|
"loss": 0.6508, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.7055688782307488e-05, |
|
"loss": 1.0112, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.704236610711431e-05, |
|
"loss": 0.8638, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.702904343192113e-05, |
|
"loss": 0.7121, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.701572075672795e-05, |
|
"loss": 0.9776, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.7002398081534774e-05, |
|
"loss": 1.076, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.6989075406341594e-05, |
|
"loss": 1.0184, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.6975752731148414e-05, |
|
"loss": 1.1203, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.6962430055955238e-05, |
|
"loss": 0.6124, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.6949107380762058e-05, |
|
"loss": 1.0818, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.693578470556888e-05, |
|
"loss": 0.7722, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.69224620303757e-05, |
|
"loss": 0.8718, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.690913935518252e-05, |
|
"loss": 1.4208, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.689581667998934e-05, |
|
"loss": 0.641, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.6882494004796165e-05, |
|
"loss": 1.188, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.6869171329602985e-05, |
|
"loss": 0.6591, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.6855848654409808e-05, |
|
"loss": 1.0678, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.6842525979216628e-05, |
|
"loss": 0.8758, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.6829203304023448e-05, |
|
"loss": 1.2125, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.681588062883027e-05, |
|
"loss": 0.9793, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.6802557953637092e-05, |
|
"loss": 0.8843, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.6789235278443915e-05, |
|
"loss": 0.8358, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.6775912603250735e-05, |
|
"loss": 0.9347, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.6762589928057555e-05, |
|
"loss": 0.9711, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.6749267252864375e-05, |
|
"loss": 0.8316, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.67359445776712e-05, |
|
"loss": 0.9148, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.672262190247802e-05, |
|
"loss": 0.9615, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.6709299227284842e-05, |
|
"loss": 0.902, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.6695976552091662e-05, |
|
"loss": 0.8277, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.6682653876898482e-05, |
|
"loss": 0.9935, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.6669331201705302e-05, |
|
"loss": 1.1443, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.6656008526512126e-05, |
|
"loss": 0.8983, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.6642685851318946e-05, |
|
"loss": 0.8136, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.662936317612577e-05, |
|
"loss": 0.9253, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.661604050093259e-05, |
|
"loss": 1.2926, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.660271782573941e-05, |
|
"loss": 1.0296, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.658939515054623e-05, |
|
"loss": 0.9648, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.6576072475353052e-05, |
|
"loss": 1.1027, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.6562749800159872e-05, |
|
"loss": 1.277, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.6549427124966696e-05, |
|
"loss": 0.7621, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.6536104449773516e-05, |
|
"loss": 0.969, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.6522781774580336e-05, |
|
"loss": 1.2513, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.6509459099387156e-05, |
|
"loss": 0.9775, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.649613642419398e-05, |
|
"loss": 0.7067, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.6482813749000803e-05, |
|
"loss": 1.2218, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.6469491073807623e-05, |
|
"loss": 0.9326, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.6456168398614443e-05, |
|
"loss": 0.8712, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.6442845723421263e-05, |
|
"loss": 0.9815, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.6429523048228086e-05, |
|
"loss": 0.8492, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.6416200373034906e-05, |
|
"loss": 1.2447, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.640287769784173e-05, |
|
"loss": 0.954, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.638955502264855e-05, |
|
"loss": 0.9911, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.637623234745537e-05, |
|
"loss": 0.6388, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.636290967226219e-05, |
|
"loss": 0.8665, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.6349586997069013e-05, |
|
"loss": 1.04, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.6336264321875833e-05, |
|
"loss": 1.1303, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.6322941646682657e-05, |
|
"loss": 0.7516, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.6309618971489477e-05, |
|
"loss": 0.9283, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.6296296296296297e-05, |
|
"loss": 0.9545, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.628297362110312e-05, |
|
"loss": 0.9962, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.626965094590994e-05, |
|
"loss": 0.6158, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.625632827071676e-05, |
|
"loss": 1.098, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.6243005595523583e-05, |
|
"loss": 1.0136, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.6229682920330404e-05, |
|
"loss": 1.2423, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.6216360245137227e-05, |
|
"loss": 1.0788, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.6203037569944047e-05, |
|
"loss": 1.0617, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.6189714894750867e-05, |
|
"loss": 1.2612, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.6176392219557687e-05, |
|
"loss": 0.9274, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.616306954436451e-05, |
|
"loss": 1.3712, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.614974686917133e-05, |
|
"loss": 1.1587, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.6136424193978154e-05, |
|
"loss": 0.9279, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.6123101518784974e-05, |
|
"loss": 0.9461, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.6109778843591794e-05, |
|
"loss": 0.5225, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.6096456168398614e-05, |
|
"loss": 0.7481, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.6083133493205437e-05, |
|
"loss": 0.8047, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.606981081801226e-05, |
|
"loss": 1.2123, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.605648814281908e-05, |
|
"loss": 0.6625, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.60431654676259e-05, |
|
"loss": 0.8275, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.602984279243272e-05, |
|
"loss": 0.8503, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.601652011723954e-05, |
|
"loss": 1.1244, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.6003197442046364e-05, |
|
"loss": 1.022, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.5989874766853188e-05, |
|
"loss": 0.8131, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.5976552091660008e-05, |
|
"loss": 0.7888, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.5963229416466828e-05, |
|
"loss": 0.8245, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.5949906741273648e-05, |
|
"loss": 0.7496, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.5936584066080468e-05, |
|
"loss": 1.0171, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.592326139088729e-05, |
|
"loss": 0.8417, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.5909938715694115e-05, |
|
"loss": 0.792, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.5896616040500935e-05, |
|
"loss": 1.0214, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.5883293365307755e-05, |
|
"loss": 0.9708, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.5869970690114575e-05, |
|
"loss": 0.7127, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.5856648014921398e-05, |
|
"loss": 0.7655, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.5843325339728218e-05, |
|
"loss": 0.7671, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.583000266453504e-05, |
|
"loss": 0.7252, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.581667998934186e-05, |
|
"loss": 0.8863, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.580335731414868e-05, |
|
"loss": 1.1643, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.57900346389555e-05, |
|
"loss": 0.9615, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.5776711963762325e-05, |
|
"loss": 1.1006, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.5763389288569145e-05, |
|
"loss": 1.1006, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.575006661337597e-05, |
|
"loss": 0.7646, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.573674393818279e-05, |
|
"loss": 0.7419, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.572342126298961e-05, |
|
"loss": 0.8336, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.5710098587796432e-05, |
|
"loss": 1.0134, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.5696775912603252e-05, |
|
"loss": 0.9505, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.5683453237410072e-05, |
|
"loss": 0.6422, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.5670130562216895e-05, |
|
"loss": 0.929, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.5656807887023715e-05, |
|
"loss": 1.0643, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.5643485211830535e-05, |
|
"loss": 0.9868, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.563016253663736e-05, |
|
"loss": 1.0235, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.561683986144418e-05, |
|
"loss": 0.8494, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.5603517186251e-05, |
|
"loss": 0.5709, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.5590194511057822e-05, |
|
"loss": 0.978, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.5576871835864642e-05, |
|
"loss": 0.9563, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.5563549160671466e-05, |
|
"loss": 0.8088, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.5550226485478286e-05, |
|
"loss": 1.116, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.5536903810285106e-05, |
|
"loss": 1.0865, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.552358113509193e-05, |
|
"loss": 0.8222, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.551025845989875e-05, |
|
"loss": 0.8851, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.5496935784705573e-05, |
|
"loss": 0.7073, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.5483613109512393e-05, |
|
"loss": 0.8308, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.5470290434319213e-05, |
|
"loss": 0.9575, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.5456967759126033e-05, |
|
"loss": 1.0262, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.5443645083932856e-05, |
|
"loss": 0.7715, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.5430322408739676e-05, |
|
"loss": 0.6188, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.54169997335465e-05, |
|
"loss": 1.0241, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.540367705835332e-05, |
|
"loss": 1.2733, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.539035438316014e-05, |
|
"loss": 1.0629, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.537703170796696e-05, |
|
"loss": 1.0838, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.5363709032773783e-05, |
|
"loss": 0.9539, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.5350386357580603e-05, |
|
"loss": 0.9519, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.5337063682387426e-05, |
|
"loss": 1.0351, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.5323741007194246e-05, |
|
"loss": 1.3052, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.5310418332001066e-05, |
|
"loss": 0.7774, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.5297095656807886e-05, |
|
"loss": 0.7707, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.528377298161471e-05, |
|
"loss": 0.5777, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.527045030642153e-05, |
|
"loss": 1.0646, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.5257127631228352e-05, |
|
"loss": 0.7333, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.5243804956035173e-05, |
|
"loss": 0.854, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.5230482280841993e-05, |
|
"loss": 1.2105, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.5217159605648815e-05, |
|
"loss": 0.7533, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.5203836930455638e-05, |
|
"loss": 0.8481, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.5190514255262458e-05, |
|
"loss": 1.169, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.5177191580069278e-05, |
|
"loss": 1.2347, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.51638689048761e-05, |
|
"loss": 0.8002, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.515054622968292e-05, |
|
"loss": 0.9941, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.5137223554489742e-05, |
|
"loss": 1.1608, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.5123900879296565e-05, |
|
"loss": 1.2734, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.5110578204103385e-05, |
|
"loss": 1.1145, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.5097255528910207e-05, |
|
"loss": 0.8585, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.5083932853717027e-05, |
|
"loss": 0.7982, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.5070610178523847e-05, |
|
"loss": 0.8173, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.505728750333067e-05, |
|
"loss": 0.8473, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.5043964828137492e-05, |
|
"loss": 0.8075, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.5030642152944312e-05, |
|
"loss": 0.7314, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.5017319477751134e-05, |
|
"loss": 0.8119, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.5003996802557954e-05, |
|
"loss": 0.8956, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.4990674127364777e-05, |
|
"loss": 1.1716, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.4977351452171597e-05, |
|
"loss": 0.8228, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.496402877697842e-05, |
|
"loss": 0.7261, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.495070610178524e-05, |
|
"loss": 0.8364, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.4937383426592061e-05, |
|
"loss": 0.965, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.4924060751398881e-05, |
|
"loss": 0.8599, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.4910738076205704e-05, |
|
"loss": 1.0879, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.4897415401012524e-05, |
|
"loss": 0.9431, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.4884092725819346e-05, |
|
"loss": 1.1419, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.4870770050626166e-05, |
|
"loss": 0.8545, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.4857447375432988e-05, |
|
"loss": 0.8475, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.484412470023981e-05, |
|
"loss": 0.7707, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.4830802025046631e-05, |
|
"loss": 0.9314, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.4817479349853451e-05, |
|
"loss": 0.773, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.4804156674660273e-05, |
|
"loss": 0.7724, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.4790833999467093e-05, |
|
"loss": 1.022, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.4777511324273915e-05, |
|
"loss": 0.5351, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.4764188649080738e-05, |
|
"loss": 1.0508, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.4750865973887558e-05, |
|
"loss": 0.7651, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 1.4737543298694378e-05, |
|
"loss": 1.1721, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 1.47242206235012e-05, |
|
"loss": 0.9186, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 1.471089794830802e-05, |
|
"loss": 0.9167, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 1.4697575273114843e-05, |
|
"loss": 0.8532, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 1.4684252597921665e-05, |
|
"loss": 0.8338, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 1.4670929922728485e-05, |
|
"loss": 0.6336, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 1.4657607247535305e-05, |
|
"loss": 0.7279, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 1.4644284572342127e-05, |
|
"loss": 0.8743, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 1.463096189714895e-05, |
|
"loss": 0.8711, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 1.461763922195577e-05, |
|
"loss": 0.8079, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 1.4604316546762592e-05, |
|
"loss": 0.8619, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 1.4590993871569412e-05, |
|
"loss": 0.7443, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 1.4577671196376234e-05, |
|
"loss": 0.7558, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 1.4564348521183054e-05, |
|
"loss": 0.9821, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 1.4551025845989877e-05, |
|
"loss": 0.7893, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.4537703170796697e-05, |
|
"loss": 0.9171, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.4524380495603519e-05, |
|
"loss": 0.6518, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.4511057820410339e-05, |
|
"loss": 0.9738, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.449773514521716e-05, |
|
"loss": 0.9537, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.4484412470023982e-05, |
|
"loss": 0.8501, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.4471089794830804e-05, |
|
"loss": 0.8564, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.4457767119637624e-05, |
|
"loss": 0.7312, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.4444444444444446e-05, |
|
"loss": 0.7642, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.4431121769251266e-05, |
|
"loss": 1.0306, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.4417799094058088e-05, |
|
"loss": 0.8579, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.440447641886491e-05, |
|
"loss": 0.8951, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.4391153743671731e-05, |
|
"loss": 0.8076, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.4377831068478551e-05, |
|
"loss": 1.0002, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.4364508393285373e-05, |
|
"loss": 1.0057, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.4351185718092193e-05, |
|
"loss": 1.1086, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 1.4337863042899016e-05, |
|
"loss": 0.8413, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 1.4324540367705836e-05, |
|
"loss": 0.9097, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 1.4311217692512658e-05, |
|
"loss": 1.0422, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 1.4297895017319478e-05, |
|
"loss": 0.91, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 1.42845723421263e-05, |
|
"loss": 1.0644, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 1.4271249666933123e-05, |
|
"loss": 0.7929, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 1.4257926991739943e-05, |
|
"loss": 0.8431, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.4244604316546765e-05, |
|
"loss": 0.9138, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.4231281641353585e-05, |
|
"loss": 1.1477, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.4217958966160405e-05, |
|
"loss": 0.6862, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.4204636290967227e-05, |
|
"loss": 1.0002, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.419131361577405e-05, |
|
"loss": 1.1307, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.417799094058087e-05, |
|
"loss": 0.804, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.4164668265387692e-05, |
|
"loss": 0.8695, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.4151345590194512e-05, |
|
"loss": 0.7704, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.4138022915001333e-05, |
|
"loss": 0.7866, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.4124700239808155e-05, |
|
"loss": 0.9987, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.4111377564614977e-05, |
|
"loss": 0.6727, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.4098054889421797e-05, |
|
"loss": 0.7705, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.4084732214228619e-05, |
|
"loss": 0.8893, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.4071409539035439e-05, |
|
"loss": 0.9634, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.405808686384226e-05, |
|
"loss": 0.8511, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 1.4044764188649082e-05, |
|
"loss": 1.1291, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 1.4031441513455904e-05, |
|
"loss": 0.8015, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 1.4018118838262724e-05, |
|
"loss": 0.8958, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 1.4004796163069546e-05, |
|
"loss": 0.8095, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 1.3991473487876366e-05, |
|
"loss": 1.0297, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 1.3978150812683189e-05, |
|
"loss": 0.9276, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 1.3964828137490009e-05, |
|
"loss": 0.7591, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 1.395150546229683e-05, |
|
"loss": 0.9263, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.393818278710365e-05, |
|
"loss": 1.0132, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.3924860111910472e-05, |
|
"loss": 0.8801, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.3911537436717296e-05, |
|
"loss": 0.8833, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.3898214761524116e-05, |
|
"loss": 0.9065, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.3884892086330936e-05, |
|
"loss": 0.748, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.3871569411137758e-05, |
|
"loss": 1.0739, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.3858246735944578e-05, |
|
"loss": 1.0306, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.38449240607514e-05, |
|
"loss": 1.0399, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.3831601385558223e-05, |
|
"loss": 0.8172, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.3818278710365043e-05, |
|
"loss": 0.9688, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.3804956035171865e-05, |
|
"loss": 1.1328, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.3791633359978685e-05, |
|
"loss": 0.5959, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.3778310684785505e-05, |
|
"loss": 0.8724, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.3764988009592328e-05, |
|
"loss": 1.0068, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.375166533439915e-05, |
|
"loss": 1.1397, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 1.373834265920597e-05, |
|
"loss": 1.1376, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 1.3725019984012791e-05, |
|
"loss": 0.617, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 1.3711697308819611e-05, |
|
"loss": 0.8408, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 1.3698374633626431e-05, |
|
"loss": 0.8512, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 1.3685051958433255e-05, |
|
"loss": 0.8946, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 1.3671729283240077e-05, |
|
"loss": 0.8904, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 1.3658406608046897e-05, |
|
"loss": 0.992, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.3645083932853718e-05, |
|
"loss": 1.024, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.3631761257660538e-05, |
|
"loss": 0.9548, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.3618438582467362e-05, |
|
"loss": 0.5628, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.3605115907274182e-05, |
|
"loss": 0.924, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.3591793232081004e-05, |
|
"loss": 0.7587, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.3578470556887824e-05, |
|
"loss": 1.0084, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.3565147881694645e-05, |
|
"loss": 0.8048, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.3551825206501467e-05, |
|
"loss": 0.6874, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.3538502531308289e-05, |
|
"loss": 0.7296, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.3525179856115109e-05, |
|
"loss": 0.9631, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.351185718092193e-05, |
|
"loss": 0.9696, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.349853450572875e-05, |
|
"loss": 0.9157, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.3485211830535572e-05, |
|
"loss": 0.556, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.3471889155342394e-05, |
|
"loss": 0.8368, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.3458566480149216e-05, |
|
"loss": 0.8747, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 1.3445243804956036e-05, |
|
"loss": 1.0028, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 1.3431921129762857e-05, |
|
"loss": 0.9508, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 1.3418598454569677e-05, |
|
"loss": 0.9811, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 1.34052757793765e-05, |
|
"loss": 0.8648, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 1.3391953104183322e-05, |
|
"loss": 0.4586, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 1.3378630428990143e-05, |
|
"loss": 0.8933, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 1.3365307753796963e-05, |
|
"loss": 0.9248, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 1.3351985078603784e-05, |
|
"loss": 1.0249, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.3338662403410604e-05, |
|
"loss": 0.8078, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.3325339728217428e-05, |
|
"loss": 0.8139, |
|
"step": 5010 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.331201705302425e-05, |
|
"loss": 0.5048, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.329869437783107e-05, |
|
"loss": 0.7647, |
|
"step": 5030 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.3285371702637891e-05, |
|
"loss": 0.7331, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.3272049027444711e-05, |
|
"loss": 1.1092, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.3258726352251535e-05, |
|
"loss": 0.7277, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.3245403677058355e-05, |
|
"loss": 0.7868, |
|
"step": 5070 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.3232081001865176e-05, |
|
"loss": 0.8461, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.3218758326671996e-05, |
|
"loss": 0.7978, |
|
"step": 5090 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.3205435651478818e-05, |
|
"loss": 0.9252, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.3192112976285638e-05, |
|
"loss": 0.7401, |
|
"step": 5110 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.3178790301092461e-05, |
|
"loss": 0.7592, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.3165467625899282e-05, |
|
"loss": 1.158, |
|
"step": 5130 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.3152144950706103e-05, |
|
"loss": 0.6913, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.3138822275512923e-05, |
|
"loss": 0.6624, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.3125499600319745e-05, |
|
"loss": 0.8496, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.3112176925126567e-05, |
|
"loss": 0.7897, |
|
"step": 5170 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.3098854249933388e-05, |
|
"loss": 0.862, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.3085531574740208e-05, |
|
"loss": 0.7539, |
|
"step": 5190 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.307220889954703e-05, |
|
"loss": 0.7561, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.305888622435385e-05, |
|
"loss": 0.7537, |
|
"step": 5210 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.3045563549160674e-05, |
|
"loss": 1.1637, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.3032240873967494e-05, |
|
"loss": 0.9744, |
|
"step": 5230 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.3018918198774315e-05, |
|
"loss": 1.0448, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.3005595523581135e-05, |
|
"loss": 0.8154, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.2992272848387957e-05, |
|
"loss": 0.9379, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.2978950173194777e-05, |
|
"loss": 1.0857, |
|
"step": 5270 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.29656274980016e-05, |
|
"loss": 0.7569, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.2952304822808422e-05, |
|
"loss": 0.8747, |
|
"step": 5290 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.2938982147615242e-05, |
|
"loss": 0.8236, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.2925659472422062e-05, |
|
"loss": 0.9229, |
|
"step": 5310 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.2912336797228884e-05, |
|
"loss": 0.7699, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.2899014122035707e-05, |
|
"loss": 1.0339, |
|
"step": 5330 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.2885691446842527e-05, |
|
"loss": 0.5565, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.2872368771649349e-05, |
|
"loss": 1.1413, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.2859046096456169e-05, |
|
"loss": 0.7633, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.284572342126299e-05, |
|
"loss": 0.9841, |
|
"step": 5370 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.2832400746069811e-05, |
|
"loss": 0.8538, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.2819078070876634e-05, |
|
"loss": 0.8028, |
|
"step": 5390 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.2805755395683454e-05, |
|
"loss": 1.1235, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.2792432720490276e-05, |
|
"loss": 1.1321, |
|
"step": 5410 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.2779110045297096e-05, |
|
"loss": 0.9859, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.2765787370103918e-05, |
|
"loss": 0.987, |
|
"step": 5430 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.275246469491074e-05, |
|
"loss": 0.9717, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.2739142019717561e-05, |
|
"loss": 1.0352, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.2725819344524381e-05, |
|
"loss": 0.8472, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.2712496669331203e-05, |
|
"loss": 0.6774, |
|
"step": 5470 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.2699173994138023e-05, |
|
"loss": 0.9617, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.2685851318944846e-05, |
|
"loss": 0.7842, |
|
"step": 5490 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.2672528643751666e-05, |
|
"loss": 0.9939, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.2659205968558488e-05, |
|
"loss": 0.9137, |
|
"step": 5510 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.2645883293365308e-05, |
|
"loss": 0.7807, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.263256061817213e-05, |
|
"loss": 0.7988, |
|
"step": 5530 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.261923794297895e-05, |
|
"loss": 0.9518, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.2605915267785773e-05, |
|
"loss": 0.6778, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.2592592592592593e-05, |
|
"loss": 1.0503, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.2579269917399415e-05, |
|
"loss": 0.7174, |
|
"step": 5570 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.2565947242206235e-05, |
|
"loss": 0.7978, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.2552624567013057e-05, |
|
"loss": 0.839, |
|
"step": 5590 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.253930189181988e-05, |
|
"loss": 1.3887, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.25259792166267e-05, |
|
"loss": 0.708, |
|
"step": 5610 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.251265654143352e-05, |
|
"loss": 0.8197, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.2499333866240342e-05, |
|
"loss": 0.5705, |
|
"step": 5630 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.2486011191047162e-05, |
|
"loss": 0.8979, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.2472688515853984e-05, |
|
"loss": 0.7944, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.2459365840660807e-05, |
|
"loss": 0.8836, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.2446043165467627e-05, |
|
"loss": 0.8557, |
|
"step": 5670 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.2432720490274449e-05, |
|
"loss": 0.8535, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.2419397815081269e-05, |
|
"loss": 0.5644, |
|
"step": 5690 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.2406075139888089e-05, |
|
"loss": 0.9361, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.2392752464694912e-05, |
|
"loss": 0.9234, |
|
"step": 5710 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.2379429789501734e-05, |
|
"loss": 0.9151, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.2366107114308554e-05, |
|
"loss": 1.1469, |
|
"step": 5730 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.2352784439115376e-05, |
|
"loss": 0.7642, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.2339461763922196e-05, |
|
"loss": 0.8959, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.232613908872902e-05, |
|
"loss": 0.9791, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.231281641353584e-05, |
|
"loss": 0.7513, |
|
"step": 5770 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.2299493738342661e-05, |
|
"loss": 0.8127, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.2286171063149481e-05, |
|
"loss": 0.6569, |
|
"step": 5790 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.2272848387956303e-05, |
|
"loss": 0.9072, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.2259525712763123e-05, |
|
"loss": 0.8966, |
|
"step": 5810 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.2246203037569946e-05, |
|
"loss": 0.7972, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.2232880362376766e-05, |
|
"loss": 0.7422, |
|
"step": 5830 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.2219557687183588e-05, |
|
"loss": 0.9659, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.2206235011990408e-05, |
|
"loss": 1.1856, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.219291233679723e-05, |
|
"loss": 0.788, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.2179589661604051e-05, |
|
"loss": 0.7689, |
|
"step": 5870 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.2166266986410873e-05, |
|
"loss": 1.1042, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.2152944311217693e-05, |
|
"loss": 0.7513, |
|
"step": 5890 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.2139621636024515e-05, |
|
"loss": 0.6981, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.2126298960831335e-05, |
|
"loss": 0.6604, |
|
"step": 5910 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.2112976285638156e-05, |
|
"loss": 0.8738, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.209965361044498e-05, |
|
"loss": 0.7364, |
|
"step": 5930 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.20863309352518e-05, |
|
"loss": 0.9592, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.207300826005862e-05, |
|
"loss": 1.0524, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.2059685584865442e-05, |
|
"loss": 0.5619, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.2046362909672262e-05, |
|
"loss": 0.9171, |
|
"step": 5970 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.2033040234479085e-05, |
|
"loss": 0.8864, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.2019717559285907e-05, |
|
"loss": 0.8777, |
|
"step": 5990 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.2006394884092727e-05, |
|
"loss": 1.2906, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.1993072208899549e-05, |
|
"loss": 0.865, |
|
"step": 6010 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.1979749533706369e-05, |
|
"loss": 0.615, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.1966426858513192e-05, |
|
"loss": 0.7379, |
|
"step": 6030 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.1953104183320012e-05, |
|
"loss": 0.6808, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.1939781508126834e-05, |
|
"loss": 0.8351, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.1926458832933654e-05, |
|
"loss": 0.7282, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.1913136157740475e-05, |
|
"loss": 0.934, |
|
"step": 6070 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.1899813482547295e-05, |
|
"loss": 0.8292, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.1886490807354119e-05, |
|
"loss": 0.8739, |
|
"step": 6090 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.1873168132160939e-05, |
|
"loss": 0.7058, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.185984545696776e-05, |
|
"loss": 0.7018, |
|
"step": 6110 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.184652278177458e-05, |
|
"loss": 1.0218, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.1833200106581402e-05, |
|
"loss": 1.1008, |
|
"step": 6130 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.1819877431388224e-05, |
|
"loss": 0.8833, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.1806554756195046e-05, |
|
"loss": 0.9845, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.1793232081001866e-05, |
|
"loss": 1.2978, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.1779909405808688e-05, |
|
"loss": 1.0189, |
|
"step": 6170 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.1766586730615508e-05, |
|
"loss": 1.2776, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.175326405542233e-05, |
|
"loss": 0.7908, |
|
"step": 6190 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.1739941380229151e-05, |
|
"loss": 0.8986, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.1726618705035973e-05, |
|
"loss": 0.8974, |
|
"step": 6210 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.1713296029842793e-05, |
|
"loss": 0.4973, |
|
"step": 6220 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.1699973354649614e-05, |
|
"loss": 0.9393, |
|
"step": 6230 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.1686650679456434e-05, |
|
"loss": 1.0344, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.1673328004263258e-05, |
|
"loss": 0.8439, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.1660005329070078e-05, |
|
"loss": 0.9039, |
|
"step": 6260 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.16466826538769e-05, |
|
"loss": 0.862, |
|
"step": 6270 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.163335997868372e-05, |
|
"loss": 0.8327, |
|
"step": 6280 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.1620037303490541e-05, |
|
"loss": 0.8622, |
|
"step": 6290 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.1606714628297361e-05, |
|
"loss": 0.8518, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.1593391953104185e-05, |
|
"loss": 0.8339, |
|
"step": 6310 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.1580069277911007e-05, |
|
"loss": 1.0355, |
|
"step": 6320 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.1566746602717827e-05, |
|
"loss": 0.6166, |
|
"step": 6330 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.1553423927524647e-05, |
|
"loss": 0.7943, |
|
"step": 6340 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.1540101252331468e-05, |
|
"loss": 0.5717, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.1526778577138292e-05, |
|
"loss": 1.0663, |
|
"step": 6360 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.1513455901945112e-05, |
|
"loss": 1.1115, |
|
"step": 6370 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.1500133226751933e-05, |
|
"loss": 0.9026, |
|
"step": 6380 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.1486810551558753e-05, |
|
"loss": 1.0676, |
|
"step": 6390 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.1473487876365575e-05, |
|
"loss": 0.9403, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.1460165201172397e-05, |
|
"loss": 0.9409, |
|
"step": 6410 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.1446842525979219e-05, |
|
"loss": 0.7303, |
|
"step": 6420 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.1433519850786039e-05, |
|
"loss": 0.7126, |
|
"step": 6430 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.142019717559286e-05, |
|
"loss": 0.7029, |
|
"step": 6440 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.140687450039968e-05, |
|
"loss": 0.8466, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.1393551825206502e-05, |
|
"loss": 0.827, |
|
"step": 6460 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.1380229150013324e-05, |
|
"loss": 1.2105, |
|
"step": 6470 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.1366906474820146e-05, |
|
"loss": 1.0531, |
|
"step": 6480 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.1353583799626966e-05, |
|
"loss": 0.7658, |
|
"step": 6490 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.1340261124433787e-05, |
|
"loss": 0.7935, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.1326938449240607e-05, |
|
"loss": 1.148, |
|
"step": 6510 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.131361577404743e-05, |
|
"loss": 1.3408, |
|
"step": 6520 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.130029309885425e-05, |
|
"loss": 0.9053, |
|
"step": 6530 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.1286970423661072e-05, |
|
"loss": 0.4991, |
|
"step": 6540 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.1273647748467892e-05, |
|
"loss": 1.0191, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.1260325073274714e-05, |
|
"loss": 0.8286, |
|
"step": 6560 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.1247002398081534e-05, |
|
"loss": 0.8756, |
|
"step": 6570 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.1233679722888358e-05, |
|
"loss": 0.8976, |
|
"step": 6580 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.1220357047695178e-05, |
|
"loss": 0.7993, |
|
"step": 6590 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.1207034372502e-05, |
|
"loss": 0.8111, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.119371169730882e-05, |
|
"loss": 0.8845, |
|
"step": 6610 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.1180389022115641e-05, |
|
"loss": 1.0341, |
|
"step": 6620 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.1167066346922464e-05, |
|
"loss": 0.8975, |
|
"step": 6630 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.1153743671729285e-05, |
|
"loss": 0.7685, |
|
"step": 6640 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.1140420996536106e-05, |
|
"loss": 1.2886, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.1127098321342926e-05, |
|
"loss": 0.646, |
|
"step": 6660 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.1113775646149746e-05, |
|
"loss": 1.0076, |
|
"step": 6670 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.110045297095657e-05, |
|
"loss": 0.7607, |
|
"step": 6680 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.1087130295763391e-05, |
|
"loss": 0.7927, |
|
"step": 6690 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.1073807620570211e-05, |
|
"loss": 0.7231, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.1060484945377033e-05, |
|
"loss": 0.7715, |
|
"step": 6710 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.1047162270183853e-05, |
|
"loss": 0.8849, |
|
"step": 6720 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.1033839594990673e-05, |
|
"loss": 1.0985, |
|
"step": 6730 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.1020516919797497e-05, |
|
"loss": 0.8239, |
|
"step": 6740 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.1007194244604318e-05, |
|
"loss": 0.8532, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.0993871569411138e-05, |
|
"loss": 0.9088, |
|
"step": 6760 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.098054889421796e-05, |
|
"loss": 0.7076, |
|
"step": 6770 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.096722621902478e-05, |
|
"loss": 0.8952, |
|
"step": 6780 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.0953903543831603e-05, |
|
"loss": 0.9333, |
|
"step": 6790 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.0940580868638424e-05, |
|
"loss": 1.0662, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.0927258193445245e-05, |
|
"loss": 0.9924, |
|
"step": 6810 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.0913935518252065e-05, |
|
"loss": 0.6257, |
|
"step": 6820 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.0900612843058887e-05, |
|
"loss": 0.9909, |
|
"step": 6830 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.0887290167865707e-05, |
|
"loss": 0.6456, |
|
"step": 6840 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.087396749267253e-05, |
|
"loss": 0.955, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.086064481747935e-05, |
|
"loss": 0.9396, |
|
"step": 6860 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.0847322142286172e-05, |
|
"loss": 1.0635, |
|
"step": 6870 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.0833999467092992e-05, |
|
"loss": 0.8497, |
|
"step": 6880 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.0820676791899814e-05, |
|
"loss": 0.6953, |
|
"step": 6890 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.0807354116706637e-05, |
|
"loss": 0.8164, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.0794031441513457e-05, |
|
"loss": 0.8728, |
|
"step": 6910 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.0780708766320277e-05, |
|
"loss": 0.8216, |
|
"step": 6920 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.0767386091127099e-05, |
|
"loss": 0.8249, |
|
"step": 6930 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.0754063415933919e-05, |
|
"loss": 0.9111, |
|
"step": 6940 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.0740740740740742e-05, |
|
"loss": 0.8234, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.0727418065547564e-05, |
|
"loss": 0.7209, |
|
"step": 6960 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.0714095390354384e-05, |
|
"loss": 0.6214, |
|
"step": 6970 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.0700772715161204e-05, |
|
"loss": 0.8723, |
|
"step": 6980 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.0687450039968026e-05, |
|
"loss": 0.8791, |
|
"step": 6990 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.0674127364774846e-05, |
|
"loss": 1.0159, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.066080468958167e-05, |
|
"loss": 1.1373, |
|
"step": 7010 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.0647482014388491e-05, |
|
"loss": 0.8018, |
|
"step": 7020 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.0634159339195311e-05, |
|
"loss": 0.7598, |
|
"step": 7030 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.0620836664002133e-05, |
|
"loss": 0.8811, |
|
"step": 7040 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.0607513988808953e-05, |
|
"loss": 0.5787, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.0594191313615776e-05, |
|
"loss": 0.6979, |
|
"step": 7060 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.0580868638422596e-05, |
|
"loss": 0.8698, |
|
"step": 7070 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.0567545963229418e-05, |
|
"loss": 0.7918, |
|
"step": 7080 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.0554223288036238e-05, |
|
"loss": 0.791, |
|
"step": 7090 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.054090061284306e-05, |
|
"loss": 0.9026, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.052757793764988e-05, |
|
"loss": 0.7427, |
|
"step": 7110 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.0514255262456703e-05, |
|
"loss": 0.8489, |
|
"step": 7120 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.0500932587263523e-05, |
|
"loss": 0.5992, |
|
"step": 7130 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.0487609912070345e-05, |
|
"loss": 0.948, |
|
"step": 7140 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.0474287236877165e-05, |
|
"loss": 0.7375, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.0460964561683987e-05, |
|
"loss": 0.8703, |
|
"step": 7160 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.0447641886490808e-05, |
|
"loss": 0.588, |
|
"step": 7170 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.043431921129763e-05, |
|
"loss": 0.9549, |
|
"step": 7180 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.042099653610445e-05, |
|
"loss": 0.7655, |
|
"step": 7190 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.0407673860911272e-05, |
|
"loss": 0.9914, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.0394351185718092e-05, |
|
"loss": 0.716, |
|
"step": 7210 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.0381028510524915e-05, |
|
"loss": 0.8625, |
|
"step": 7220 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.0367705835331735e-05, |
|
"loss": 0.546, |
|
"step": 7230 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.0354383160138557e-05, |
|
"loss": 0.9452, |
|
"step": 7240 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.0341060484945377e-05, |
|
"loss": 0.7429, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.0327737809752199e-05, |
|
"loss": 0.8323, |
|
"step": 7260 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.0314415134559019e-05, |
|
"loss": 0.8861, |
|
"step": 7270 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.0301092459365842e-05, |
|
"loss": 0.5798, |
|
"step": 7280 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.0287769784172664e-05, |
|
"loss": 0.6907, |
|
"step": 7290 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.0274447108979484e-05, |
|
"loss": 1.1387, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.0261124433786304e-05, |
|
"loss": 0.851, |
|
"step": 7310 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.0247801758593126e-05, |
|
"loss": 0.7026, |
|
"step": 7320 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.0234479083399949e-05, |
|
"loss": 1.0397, |
|
"step": 7330 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.0221156408206769e-05, |
|
"loss": 0.8539, |
|
"step": 7340 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.020783373301359e-05, |
|
"loss": 0.7249, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.0194511057820411e-05, |
|
"loss": 0.6053, |
|
"step": 7360 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.0181188382627233e-05, |
|
"loss": 0.9206, |
|
"step": 7370 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.0167865707434053e-05, |
|
"loss": 0.72, |
|
"step": 7380 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.0154543032240876e-05, |
|
"loss": 0.7217, |
|
"step": 7390 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.0141220357047696e-05, |
|
"loss": 0.9969, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.0127897681854518e-05, |
|
"loss": 0.9146, |
|
"step": 7410 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.0114575006661338e-05, |
|
"loss": 0.8879, |
|
"step": 7420 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.010125233146816e-05, |
|
"loss": 0.6825, |
|
"step": 7430 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.0087929656274981e-05, |
|
"loss": 1.2015, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.0074606981081803e-05, |
|
"loss": 0.8784, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.0061284305888623e-05, |
|
"loss": 0.7296, |
|
"step": 7460 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 1.0047961630695445e-05, |
|
"loss": 0.5832, |
|
"step": 7470 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 1.0034638955502265e-05, |
|
"loss": 0.9754, |
|
"step": 7480 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 1.0021316280309088e-05, |
|
"loss": 0.8457, |
|
"step": 7490 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 1.0007993605115908e-05, |
|
"loss": 0.8754, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.99467092992273e-06, |
|
"loss": 0.6651, |
|
"step": 7510 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.98134825472955e-06, |
|
"loss": 0.5887, |
|
"step": 7520 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.968025579536372e-06, |
|
"loss": 0.5617, |
|
"step": 7530 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.954702904343193e-06, |
|
"loss": 0.6945, |
|
"step": 7540 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.941380229150013e-06, |
|
"loss": 0.6885, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.928057553956835e-06, |
|
"loss": 0.6411, |
|
"step": 7560 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.914734878763657e-06, |
|
"loss": 0.6029, |
|
"step": 7570 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.901412203570477e-06, |
|
"loss": 0.479, |
|
"step": 7580 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.888089528377298e-06, |
|
"loss": 0.4233, |
|
"step": 7590 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.87476685318412e-06, |
|
"loss": 0.7149, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.86144417799094e-06, |
|
"loss": 0.6039, |
|
"step": 7610 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.848121502797764e-06, |
|
"loss": 0.3904, |
|
"step": 7620 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.834798827604584e-06, |
|
"loss": 0.554, |
|
"step": 7630 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.821476152411405e-06, |
|
"loss": 0.6223, |
|
"step": 7640 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.808153477218227e-06, |
|
"loss": 0.6661, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.794830802025047e-06, |
|
"loss": 0.6223, |
|
"step": 7660 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.781508126831869e-06, |
|
"loss": 0.4074, |
|
"step": 7670 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.76818545163869e-06, |
|
"loss": 0.5239, |
|
"step": 7680 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.75486277644551e-06, |
|
"loss": 0.6673, |
|
"step": 7690 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.741540101252332e-06, |
|
"loss": 0.4243, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.728217426059154e-06, |
|
"loss": 0.5662, |
|
"step": 7710 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.714894750865976e-06, |
|
"loss": 0.5088, |
|
"step": 7720 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.701572075672796e-06, |
|
"loss": 0.6867, |
|
"step": 7730 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.688249400479617e-06, |
|
"loss": 0.6084, |
|
"step": 7740 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.67492672528644e-06, |
|
"loss": 0.5469, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.66160405009326e-06, |
|
"loss": 0.4736, |
|
"step": 7760 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.648281374900081e-06, |
|
"loss": 0.5609, |
|
"step": 7770 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.634958699706903e-06, |
|
"loss": 0.4694, |
|
"step": 7780 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.621636024513723e-06, |
|
"loss": 0.6748, |
|
"step": 7790 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.608313349320544e-06, |
|
"loss": 0.6605, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.594990674127366e-06, |
|
"loss": 0.5726, |
|
"step": 7810 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.581667998934186e-06, |
|
"loss": 0.8413, |
|
"step": 7820 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.568345323741008e-06, |
|
"loss": 0.4299, |
|
"step": 7830 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.55502264854783e-06, |
|
"loss": 0.6246, |
|
"step": 7840 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 9.54169997335465e-06, |
|
"loss": 0.7487, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 9.528377298161471e-06, |
|
"loss": 0.7082, |
|
"step": 7860 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 9.515054622968293e-06, |
|
"loss": 0.5832, |
|
"step": 7870 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 9.501731947775113e-06, |
|
"loss": 0.5197, |
|
"step": 7880 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 9.488409272581935e-06, |
|
"loss": 0.4889, |
|
"step": 7890 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 9.475086597388756e-06, |
|
"loss": 0.3017, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 9.461763922195578e-06, |
|
"loss": 0.5872, |
|
"step": 7910 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 9.448441247002398e-06, |
|
"loss": 0.6167, |
|
"step": 7920 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 9.43511857180922e-06, |
|
"loss": 0.5968, |
|
"step": 7930 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 9.421795896616042e-06, |
|
"loss": 0.7126, |
|
"step": 7940 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 9.408473221422862e-06, |
|
"loss": 0.6875, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 9.395150546229683e-06, |
|
"loss": 0.4871, |
|
"step": 7960 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 9.381827871036505e-06, |
|
"loss": 0.6521, |
|
"step": 7970 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 9.368505195843325e-06, |
|
"loss": 0.5513, |
|
"step": 7980 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 9.355182520650149e-06, |
|
"loss": 0.438, |
|
"step": 7990 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 9.341859845456969e-06, |
|
"loss": 0.6675, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 9.32853717026379e-06, |
|
"loss": 0.7628, |
|
"step": 8010 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 9.315214495070612e-06, |
|
"loss": 0.48, |
|
"step": 8020 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 9.301891819877432e-06, |
|
"loss": 0.5852, |
|
"step": 8030 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 9.288569144684254e-06, |
|
"loss": 0.5102, |
|
"step": 8040 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 9.275246469491075e-06, |
|
"loss": 0.3851, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 9.261923794297895e-06, |
|
"loss": 0.7704, |
|
"step": 8060 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 9.248601119104717e-06, |
|
"loss": 0.7697, |
|
"step": 8070 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 9.235278443911539e-06, |
|
"loss": 0.4341, |
|
"step": 8080 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 9.221955768718359e-06, |
|
"loss": 0.4089, |
|
"step": 8090 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 9.20863309352518e-06, |
|
"loss": 0.825, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 9.195310418332002e-06, |
|
"loss": 0.667, |
|
"step": 8110 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 9.181987743138822e-06, |
|
"loss": 0.6688, |
|
"step": 8120 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 9.168665067945644e-06, |
|
"loss": 0.426, |
|
"step": 8130 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 9.155342392752466e-06, |
|
"loss": 0.6672, |
|
"step": 8140 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 9.142019717559286e-06, |
|
"loss": 0.4767, |
|
"step": 8150 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 9.128697042366108e-06, |
|
"loss": 0.6029, |
|
"step": 8160 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 9.11537436717293e-06, |
|
"loss": 0.6964, |
|
"step": 8170 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 9.102051691979751e-06, |
|
"loss": 0.7647, |
|
"step": 8180 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 9.088729016786571e-06, |
|
"loss": 0.5858, |
|
"step": 8190 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 9.075406341593393e-06, |
|
"loss": 0.6002, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 9.062083666400214e-06, |
|
"loss": 0.6143, |
|
"step": 8210 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 9.048760991207034e-06, |
|
"loss": 0.4929, |
|
"step": 8220 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 9.035438316013856e-06, |
|
"loss": 0.6795, |
|
"step": 8230 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 9.022115640820678e-06, |
|
"loss": 0.5844, |
|
"step": 8240 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 9.008792965627498e-06, |
|
"loss": 0.597, |
|
"step": 8250 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 8.99547029043432e-06, |
|
"loss": 0.5097, |
|
"step": 8260 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 8.982147615241141e-06, |
|
"loss": 0.8274, |
|
"step": 8270 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 8.968824940047961e-06, |
|
"loss": 0.4721, |
|
"step": 8280 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 8.955502264854785e-06, |
|
"loss": 0.7477, |
|
"step": 8290 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 8.942179589661605e-06, |
|
"loss": 0.515, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 8.928856914468425e-06, |
|
"loss": 0.6419, |
|
"step": 8310 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 8.915534239275248e-06, |
|
"loss": 0.548, |
|
"step": 8320 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 8.902211564082068e-06, |
|
"loss": 0.6217, |
|
"step": 8330 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 8.888888888888888e-06, |
|
"loss": 0.515, |
|
"step": 8340 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 8.875566213695712e-06, |
|
"loss": 0.5288, |
|
"step": 8350 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 8.862243538502532e-06, |
|
"loss": 0.6349, |
|
"step": 8360 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 8.848920863309353e-06, |
|
"loss": 0.6642, |
|
"step": 8370 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 8.835598188116175e-06, |
|
"loss": 0.6257, |
|
"step": 8380 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 8.822275512922995e-06, |
|
"loss": 0.5945, |
|
"step": 8390 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 8.808952837729817e-06, |
|
"loss": 0.4554, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 8.795630162536639e-06, |
|
"loss": 0.5942, |
|
"step": 8410 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 8.782307487343459e-06, |
|
"loss": 0.5292, |
|
"step": 8420 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 8.76898481215028e-06, |
|
"loss": 0.517, |
|
"step": 8430 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 8.755662136957102e-06, |
|
"loss": 0.4412, |
|
"step": 8440 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 8.742339461763924e-06, |
|
"loss": 0.6529, |
|
"step": 8450 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 8.729016786570744e-06, |
|
"loss": 0.3173, |
|
"step": 8460 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 8.715694111377566e-06, |
|
"loss": 0.6011, |
|
"step": 8470 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 8.702371436184387e-06, |
|
"loss": 0.511, |
|
"step": 8480 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 8.689048760991207e-06, |
|
"loss": 0.5281, |
|
"step": 8490 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 8.675726085798029e-06, |
|
"loss": 0.5546, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 8.66240341060485e-06, |
|
"loss": 0.4424, |
|
"step": 8510 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 8.64908073541167e-06, |
|
"loss": 0.5005, |
|
"step": 8520 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 8.635758060218492e-06, |
|
"loss": 0.5843, |
|
"step": 8530 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 8.622435385025314e-06, |
|
"loss": 0.5101, |
|
"step": 8540 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 8.609112709832134e-06, |
|
"loss": 0.5558, |
|
"step": 8550 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 8.595790034638956e-06, |
|
"loss": 0.5574, |
|
"step": 8560 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 8.582467359445778e-06, |
|
"loss": 0.5099, |
|
"step": 8570 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 8.569144684252598e-06, |
|
"loss": 0.5025, |
|
"step": 8580 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 8.55582200905942e-06, |
|
"loss": 0.646, |
|
"step": 8590 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 8.542499333866241e-06, |
|
"loss": 0.511, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 8.529176658673061e-06, |
|
"loss": 0.5669, |
|
"step": 8610 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 8.515853983479885e-06, |
|
"loss": 0.8414, |
|
"step": 8620 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 8.502531308286705e-06, |
|
"loss": 0.5391, |
|
"step": 8630 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 8.489208633093526e-06, |
|
"loss": 0.841, |
|
"step": 8640 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 8.475885957900348e-06, |
|
"loss": 0.8282, |
|
"step": 8650 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 8.462563282707168e-06, |
|
"loss": 0.4807, |
|
"step": 8660 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 8.44924060751399e-06, |
|
"loss": 0.5238, |
|
"step": 8670 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 8.435917932320811e-06, |
|
"loss": 0.7307, |
|
"step": 8680 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 8.422595257127631e-06, |
|
"loss": 0.6926, |
|
"step": 8690 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 8.409272581934453e-06, |
|
"loss": 0.6897, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 8.395949906741275e-06, |
|
"loss": 0.6377, |
|
"step": 8710 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 8.382627231548095e-06, |
|
"loss": 0.6711, |
|
"step": 8720 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 8.369304556354917e-06, |
|
"loss": 0.4469, |
|
"step": 8730 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 8.355981881161738e-06, |
|
"loss": 0.4468, |
|
"step": 8740 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 8.34265920596856e-06, |
|
"loss": 0.5768, |
|
"step": 8750 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 8.32933653077538e-06, |
|
"loss": 0.5792, |
|
"step": 8760 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 8.316013855582202e-06, |
|
"loss": 0.6167, |
|
"step": 8770 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 8.302691180389024e-06, |
|
"loss": 0.5216, |
|
"step": 8780 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 8.289368505195844e-06, |
|
"loss": 0.706, |
|
"step": 8790 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 8.276045830002665e-06, |
|
"loss": 0.5537, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 8.262723154809487e-06, |
|
"loss": 0.6349, |
|
"step": 8810 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 8.249400479616307e-06, |
|
"loss": 0.4659, |
|
"step": 8820 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 8.236077804423129e-06, |
|
"loss": 0.4473, |
|
"step": 8830 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 8.22275512922995e-06, |
|
"loss": 0.5256, |
|
"step": 8840 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 8.20943245403677e-06, |
|
"loss": 0.5484, |
|
"step": 8850 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 8.196109778843592e-06, |
|
"loss": 0.6666, |
|
"step": 8860 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 8.182787103650414e-06, |
|
"loss": 0.9443, |
|
"step": 8870 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 8.169464428457234e-06, |
|
"loss": 0.4171, |
|
"step": 8880 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 8.156141753264056e-06, |
|
"loss": 0.5668, |
|
"step": 8890 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 8.142819078070877e-06, |
|
"loss": 0.7022, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 8.129496402877699e-06, |
|
"loss": 0.5253, |
|
"step": 8910 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 8.116173727684519e-06, |
|
"loss": 0.8242, |
|
"step": 8920 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 8.10285105249134e-06, |
|
"loss": 0.6424, |
|
"step": 8930 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 8.089528377298163e-06, |
|
"loss": 0.4108, |
|
"step": 8940 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 8.076205702104983e-06, |
|
"loss": 0.7285, |
|
"step": 8950 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 8.062883026911804e-06, |
|
"loss": 0.6103, |
|
"step": 8960 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 8.049560351718626e-06, |
|
"loss": 0.4268, |
|
"step": 8970 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 8.036237676525448e-06, |
|
"loss": 0.6346, |
|
"step": 8980 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 8.022915001332268e-06, |
|
"loss": 0.5144, |
|
"step": 8990 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 8.00959232613909e-06, |
|
"loss": 0.622, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 7.996269650945911e-06, |
|
"loss": 0.6551, |
|
"step": 9010 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 7.982946975752733e-06, |
|
"loss": 0.5867, |
|
"step": 9020 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 7.969624300559553e-06, |
|
"loss": 0.5306, |
|
"step": 9030 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 7.956301625366375e-06, |
|
"loss": 0.7457, |
|
"step": 9040 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 7.942978950173196e-06, |
|
"loss": 0.4717, |
|
"step": 9050 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 7.929656274980016e-06, |
|
"loss": 0.7841, |
|
"step": 9060 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 7.916333599786838e-06, |
|
"loss": 0.4112, |
|
"step": 9070 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 7.90301092459366e-06, |
|
"loss": 0.6972, |
|
"step": 9080 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 7.88968824940048e-06, |
|
"loss": 0.7216, |
|
"step": 9090 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 7.876365574207302e-06, |
|
"loss": 0.5126, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 7.863042899014123e-06, |
|
"loss": 0.7825, |
|
"step": 9110 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 7.849720223820943e-06, |
|
"loss": 0.5064, |
|
"step": 9120 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 7.836397548627765e-06, |
|
"loss": 0.5202, |
|
"step": 9130 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 7.823074873434587e-06, |
|
"loss": 0.7074, |
|
"step": 9140 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 7.809752198241407e-06, |
|
"loss": 0.6743, |
|
"step": 9150 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 7.796429523048228e-06, |
|
"loss": 0.5713, |
|
"step": 9160 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 7.78310684785505e-06, |
|
"loss": 0.5938, |
|
"step": 9170 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 7.769784172661872e-06, |
|
"loss": 0.5986, |
|
"step": 9180 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 7.756461497468692e-06, |
|
"loss": 0.6494, |
|
"step": 9190 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 7.743138822275514e-06, |
|
"loss": 0.7668, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 7.729816147082335e-06, |
|
"loss": 0.5961, |
|
"step": 9210 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 7.716493471889155e-06, |
|
"loss": 0.5373, |
|
"step": 9220 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 7.703170796695977e-06, |
|
"loss": 0.5084, |
|
"step": 9230 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 7.689848121502799e-06, |
|
"loss": 0.4927, |
|
"step": 9240 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 7.676525446309619e-06, |
|
"loss": 0.5338, |
|
"step": 9250 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 7.66320277111644e-06, |
|
"loss": 0.5202, |
|
"step": 9260 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 7.649880095923262e-06, |
|
"loss": 0.7087, |
|
"step": 9270 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 7.636557420730082e-06, |
|
"loss": 0.4755, |
|
"step": 9280 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 7.623234745536905e-06, |
|
"loss": 0.6859, |
|
"step": 9290 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 7.609912070343726e-06, |
|
"loss": 0.5637, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 7.5965893951505466e-06, |
|
"loss": 0.4301, |
|
"step": 9310 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 7.583266719957368e-06, |
|
"loss": 0.555, |
|
"step": 9320 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 7.569944044764189e-06, |
|
"loss": 0.5708, |
|
"step": 9330 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 7.55662136957101e-06, |
|
"loss": 0.3748, |
|
"step": 9340 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 7.543298694377832e-06, |
|
"loss": 0.3104, |
|
"step": 9350 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 7.529976019184653e-06, |
|
"loss": 0.4811, |
|
"step": 9360 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 7.516653343991474e-06, |
|
"loss": 0.7626, |
|
"step": 9370 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 7.503330668798295e-06, |
|
"loss": 0.7432, |
|
"step": 9380 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 7.490007993605116e-06, |
|
"loss": 0.4564, |
|
"step": 9390 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 7.476685318411938e-06, |
|
"loss": 0.4931, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 7.463362643218759e-06, |
|
"loss": 0.5394, |
|
"step": 9410 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 7.4500399680255795e-06, |
|
"loss": 0.6123, |
|
"step": 9420 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 7.436717292832401e-06, |
|
"loss": 0.4299, |
|
"step": 9430 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 7.423394617639222e-06, |
|
"loss": 0.5161, |
|
"step": 9440 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 7.410071942446043e-06, |
|
"loss": 0.5089, |
|
"step": 9450 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 7.3967492672528655e-06, |
|
"loss": 0.6182, |
|
"step": 9460 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 7.3834265920596856e-06, |
|
"loss": 0.5111, |
|
"step": 9470 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 7.370103916866508e-06, |
|
"loss": 0.6544, |
|
"step": 9480 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 7.356781241673329e-06, |
|
"loss": 0.5086, |
|
"step": 9490 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 7.343458566480149e-06, |
|
"loss": 0.5772, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 7.3301358912869716e-06, |
|
"loss": 0.7216, |
|
"step": 9510 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 7.3168132160937924e-06, |
|
"loss": 0.3456, |
|
"step": 9520 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 7.303490540900613e-06, |
|
"loss": 0.5831, |
|
"step": 9530 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 7.290167865707435e-06, |
|
"loss": 0.4155, |
|
"step": 9540 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 7.276845190514256e-06, |
|
"loss": 0.4813, |
|
"step": 9550 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 7.263522515321078e-06, |
|
"loss": 0.4247, |
|
"step": 9560 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 7.2501998401278985e-06, |
|
"loss": 0.3583, |
|
"step": 9570 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 7.236877164934719e-06, |
|
"loss": 0.5876, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 7.223554489741541e-06, |
|
"loss": 0.4301, |
|
"step": 9590 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 7.210231814548362e-06, |
|
"loss": 0.6511, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 7.196909139355183e-06, |
|
"loss": 0.3719, |
|
"step": 9610 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 7.1835864641620045e-06, |
|
"loss": 0.8085, |
|
"step": 9620 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 7.170263788968825e-06, |
|
"loss": 0.495, |
|
"step": 9630 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 7.156941113775647e-06, |
|
"loss": 0.4479, |
|
"step": 9640 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 7.143618438582468e-06, |
|
"loss": 0.5258, |
|
"step": 9650 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 7.130295763389289e-06, |
|
"loss": 0.5833, |
|
"step": 9660 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 7.1169730881961106e-06, |
|
"loss": 0.6848, |
|
"step": 9670 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 7.1036504130029314e-06, |
|
"loss": 0.681, |
|
"step": 9680 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 7.090327737809752e-06, |
|
"loss": 0.7938, |
|
"step": 9690 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 7.077005062616574e-06, |
|
"loss": 0.7305, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 7.063682387423395e-06, |
|
"loss": 0.6472, |
|
"step": 9710 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 7.050359712230216e-06, |
|
"loss": 0.6139, |
|
"step": 9720 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 7.0370370370370375e-06, |
|
"loss": 0.6646, |
|
"step": 9730 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 7.023714361843858e-06, |
|
"loss": 0.6314, |
|
"step": 9740 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 7.01039168665068e-06, |
|
"loss": 0.6679, |
|
"step": 9750 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 6.997069011457501e-06, |
|
"loss": 0.3637, |
|
"step": 9760 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 6.983746336264322e-06, |
|
"loss": 0.5498, |
|
"step": 9770 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 6.970423661071144e-06, |
|
"loss": 0.5565, |
|
"step": 9780 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 6.957100985877964e-06, |
|
"loss": 0.7444, |
|
"step": 9790 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 6.943778310684785e-06, |
|
"loss": 0.4358, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 6.930455635491608e-06, |
|
"loss": 0.3845, |
|
"step": 9810 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 6.917132960298429e-06, |
|
"loss": 0.7262, |
|
"step": 9820 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 6.90381028510525e-06, |
|
"loss": 0.5637, |
|
"step": 9830 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 6.890487609912071e-06, |
|
"loss": 0.5843, |
|
"step": 9840 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 6.877164934718892e-06, |
|
"loss": 0.6286, |
|
"step": 9850 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 6.863842259525714e-06, |
|
"loss": 0.4262, |
|
"step": 9860 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 6.850519584332535e-06, |
|
"loss": 0.6395, |
|
"step": 9870 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 6.837196909139356e-06, |
|
"loss": 0.618, |
|
"step": 9880 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 6.823874233946177e-06, |
|
"loss": 0.4814, |
|
"step": 9890 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 6.810551558752998e-06, |
|
"loss": 0.474, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 6.79722888355982e-06, |
|
"loss": 0.5267, |
|
"step": 9910 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 6.783906208366641e-06, |
|
"loss": 0.6874, |
|
"step": 9920 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 6.770583533173462e-06, |
|
"loss": 0.5276, |
|
"step": 9930 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 6.757260857980283e-06, |
|
"loss": 0.5303, |
|
"step": 9940 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 6.743938182787104e-06, |
|
"loss": 0.4643, |
|
"step": 9950 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 6.730615507593925e-06, |
|
"loss": 0.4328, |
|
"step": 9960 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 6.717292832400747e-06, |
|
"loss": 0.542, |
|
"step": 9970 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 6.703970157207568e-06, |
|
"loss": 0.502, |
|
"step": 9980 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 6.6906474820143886e-06, |
|
"loss": 0.603, |
|
"step": 9990 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 6.67732480682121e-06, |
|
"loss": 0.7646, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 6.664002131628031e-06, |
|
"loss": 0.6183, |
|
"step": 10010 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 6.650679456434853e-06, |
|
"loss": 0.462, |
|
"step": 10020 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 6.637356781241674e-06, |
|
"loss": 0.4754, |
|
"step": 10030 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 6.624034106048495e-06, |
|
"loss": 0.632, |
|
"step": 10040 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 6.610711430855316e-06, |
|
"loss": 0.5148, |
|
"step": 10050 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 6.597388755662137e-06, |
|
"loss": 0.4957, |
|
"step": 10060 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 6.584066080468958e-06, |
|
"loss": 0.5495, |
|
"step": 10070 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 6.57074340527578e-06, |
|
"loss": 0.5744, |
|
"step": 10080 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 6.557420730082601e-06, |
|
"loss": 0.8109, |
|
"step": 10090 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 6.544098054889423e-06, |
|
"loss": 0.4862, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 6.530775379696243e-06, |
|
"loss": 0.4743, |
|
"step": 10110 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 6.517452704503064e-06, |
|
"loss": 0.6421, |
|
"step": 10120 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 6.504130029309887e-06, |
|
"loss": 0.376, |
|
"step": 10130 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 6.4908073541167076e-06, |
|
"loss": 0.5927, |
|
"step": 10140 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 6.4774846789235276e-06, |
|
"loss": 0.805, |
|
"step": 10150 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 6.46416200373035e-06, |
|
"loss": 0.6305, |
|
"step": 10160 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 6.450839328537171e-06, |
|
"loss": 0.5092, |
|
"step": 10170 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 6.437516653343991e-06, |
|
"loss": 0.7112, |
|
"step": 10180 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 6.424193978150814e-06, |
|
"loss": 0.7156, |
|
"step": 10190 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 6.4108713029576345e-06, |
|
"loss": 0.4235, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 6.397548627764456e-06, |
|
"loss": 0.3841, |
|
"step": 10210 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 6.384225952571277e-06, |
|
"loss": 0.4779, |
|
"step": 10220 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 6.370903277378098e-06, |
|
"loss": 0.5718, |
|
"step": 10230 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 6.35758060218492e-06, |
|
"loss": 0.5171, |
|
"step": 10240 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 6.3442579269917405e-06, |
|
"loss": 0.5425, |
|
"step": 10250 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 6.330935251798561e-06, |
|
"loss": 0.6426, |
|
"step": 10260 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 6.317612576605383e-06, |
|
"loss": 0.5412, |
|
"step": 10270 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 6.304289901412204e-06, |
|
"loss": 0.6446, |
|
"step": 10280 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 6.290967226219026e-06, |
|
"loss": 0.4798, |
|
"step": 10290 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 6.2776445510258465e-06, |
|
"loss": 0.601, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 6.264321875832667e-06, |
|
"loss": 0.481, |
|
"step": 10310 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 6.250999200639489e-06, |
|
"loss": 0.3424, |
|
"step": 10320 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 6.23767652544631e-06, |
|
"loss": 0.6193, |
|
"step": 10330 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 6.224353850253131e-06, |
|
"loss": 0.462, |
|
"step": 10340 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 6.211031175059953e-06, |
|
"loss": 0.518, |
|
"step": 10350 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 6.1977084998667735e-06, |
|
"loss": 0.6449, |
|
"step": 10360 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 6.184385824673595e-06, |
|
"loss": 0.6115, |
|
"step": 10370 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 6.171063149480416e-06, |
|
"loss": 0.5942, |
|
"step": 10380 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 6.157740474287237e-06, |
|
"loss": 0.7223, |
|
"step": 10390 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 6.144417799094059e-06, |
|
"loss": 0.5872, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 6.1310951239008795e-06, |
|
"loss": 0.8311, |
|
"step": 10410 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 6.1177724487077e-06, |
|
"loss": 0.4722, |
|
"step": 10420 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 6.104449773514522e-06, |
|
"loss": 0.6166, |
|
"step": 10430 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 6.091127098321343e-06, |
|
"loss": 0.5281, |
|
"step": 10440 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 6.077804423128164e-06, |
|
"loss": 0.5282, |
|
"step": 10450 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 6.064481747934986e-06, |
|
"loss": 0.4552, |
|
"step": 10460 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 6.051159072741806e-06, |
|
"loss": 0.486, |
|
"step": 10470 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 6.037836397548629e-06, |
|
"loss": 0.7482, |
|
"step": 10480 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 6.02451372235545e-06, |
|
"loss": 0.6316, |
|
"step": 10490 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 6.011191047162271e-06, |
|
"loss": 0.6591, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 5.9978683719690924e-06, |
|
"loss": 0.5151, |
|
"step": 10510 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 5.984545696775913e-06, |
|
"loss": 0.5123, |
|
"step": 10520 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 5.971223021582734e-06, |
|
"loss": 0.3546, |
|
"step": 10530 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 5.957900346389556e-06, |
|
"loss": 0.5524, |
|
"step": 10540 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 5.944577671196377e-06, |
|
"loss": 0.4409, |
|
"step": 10550 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 5.9312549960031985e-06, |
|
"loss": 0.5293, |
|
"step": 10560 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 5.917932320810019e-06, |
|
"loss": 0.437, |
|
"step": 10570 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 5.90460964561684e-06, |
|
"loss": 0.5775, |
|
"step": 10580 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 5.891286970423662e-06, |
|
"loss": 0.5641, |
|
"step": 10590 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 5.877964295230483e-06, |
|
"loss": 0.4839, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 5.864641620037304e-06, |
|
"loss": 0.4325, |
|
"step": 10610 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 5.851318944844125e-06, |
|
"loss": 0.5035, |
|
"step": 10620 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 5.837996269650946e-06, |
|
"loss": 0.6819, |
|
"step": 10630 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 5.824673594457768e-06, |
|
"loss": 0.6849, |
|
"step": 10640 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 5.811350919264589e-06, |
|
"loss": 0.4236, |
|
"step": 10650 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 5.79802824407141e-06, |
|
"loss": 0.3773, |
|
"step": 10660 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 5.7847055688782314e-06, |
|
"loss": 0.6302, |
|
"step": 10670 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 5.771382893685052e-06, |
|
"loss": 0.4993, |
|
"step": 10680 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 5.758060218491873e-06, |
|
"loss": 0.5893, |
|
"step": 10690 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 5.744737543298695e-06, |
|
"loss": 0.4807, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 5.731414868105516e-06, |
|
"loss": 0.5702, |
|
"step": 10710 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 5.718092192912337e-06, |
|
"loss": 0.512, |
|
"step": 10720 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 5.704769517719158e-06, |
|
"loss": 0.5136, |
|
"step": 10730 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 5.691446842525979e-06, |
|
"loss": 0.2975, |
|
"step": 10740 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 5.678124167332801e-06, |
|
"loss": 0.4943, |
|
"step": 10750 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 5.664801492139622e-06, |
|
"loss": 0.4108, |
|
"step": 10760 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 5.651478816946443e-06, |
|
"loss": 0.647, |
|
"step": 10770 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 5.638156141753265e-06, |
|
"loss": 0.4033, |
|
"step": 10780 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 5.624833466560085e-06, |
|
"loss": 0.5327, |
|
"step": 10790 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 5.611510791366906e-06, |
|
"loss": 0.659, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 5.598188116173729e-06, |
|
"loss": 0.582, |
|
"step": 10810 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 5.5848654409805496e-06, |
|
"loss": 0.5686, |
|
"step": 10820 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 5.571542765787371e-06, |
|
"loss": 0.774, |
|
"step": 10830 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 5.558220090594192e-06, |
|
"loss": 0.4268, |
|
"step": 10840 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 5.544897415401013e-06, |
|
"loss": 0.6283, |
|
"step": 10850 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 5.531574740207835e-06, |
|
"loss": 0.7076, |
|
"step": 10860 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 5.518252065014656e-06, |
|
"loss": 0.4266, |
|
"step": 10870 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 5.5049293898214765e-06, |
|
"loss": 0.615, |
|
"step": 10880 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 5.491606714628298e-06, |
|
"loss": 0.5533, |
|
"step": 10890 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 5.478284039435119e-06, |
|
"loss": 0.4839, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 5.46496136424194e-06, |
|
"loss": 0.5785, |
|
"step": 10910 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 5.451638689048762e-06, |
|
"loss": 0.52, |
|
"step": 10920 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 5.4383160138555825e-06, |
|
"loss": 0.4498, |
|
"step": 10930 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 5.424993338662404e-06, |
|
"loss": 0.5643, |
|
"step": 10940 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 5.411670663469225e-06, |
|
"loss": 0.6726, |
|
"step": 10950 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 5.398347988276046e-06, |
|
"loss": 0.6215, |
|
"step": 10960 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 5.385025313082868e-06, |
|
"loss": 0.4321, |
|
"step": 10970 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 5.3717026378896886e-06, |
|
"loss": 0.5597, |
|
"step": 10980 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 5.3583799626965094e-06, |
|
"loss": 0.4928, |
|
"step": 10990 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 5.345057287503331e-06, |
|
"loss": 0.5622, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 5.331734612310152e-06, |
|
"loss": 0.5592, |
|
"step": 11010 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 5.318411937116974e-06, |
|
"loss": 0.4732, |
|
"step": 11020 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 5.305089261923795e-06, |
|
"loss": 0.5939, |
|
"step": 11030 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 5.2917665867306155e-06, |
|
"loss": 0.5467, |
|
"step": 11040 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 5.278443911537437e-06, |
|
"loss": 0.6886, |
|
"step": 11050 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 5.265121236344258e-06, |
|
"loss": 0.492, |
|
"step": 11060 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 5.251798561151079e-06, |
|
"loss": 0.3911, |
|
"step": 11070 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 5.238475885957901e-06, |
|
"loss": 0.7715, |
|
"step": 11080 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 5.2251532107647215e-06, |
|
"loss": 0.5695, |
|
"step": 11090 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 5.211830535571544e-06, |
|
"loss": 0.5881, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 5.198507860378364e-06, |
|
"loss": 0.5352, |
|
"step": 11110 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 5.185185185185185e-06, |
|
"loss": 0.6824, |
|
"step": 11120 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 5.1718625099920075e-06, |
|
"loss": 0.6598, |
|
"step": 11130 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 5.158539834798828e-06, |
|
"loss": 0.5027, |
|
"step": 11140 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 5.1452171596056484e-06, |
|
"loss": 0.4774, |
|
"step": 11150 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 5.131894484412471e-06, |
|
"loss": 0.309, |
|
"step": 11160 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 5.118571809219292e-06, |
|
"loss": 0.5187, |
|
"step": 11170 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 5.105249134026113e-06, |
|
"loss": 0.5329, |
|
"step": 11180 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 5.0919264588329345e-06, |
|
"loss": 0.4912, |
|
"step": 11190 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 5.078603783639755e-06, |
|
"loss": 0.5734, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 5.065281108446577e-06, |
|
"loss": 0.695, |
|
"step": 11210 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 5.051958433253398e-06, |
|
"loss": 0.4631, |
|
"step": 11220 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 5.038635758060219e-06, |
|
"loss": 0.9621, |
|
"step": 11230 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 5.0253130828670405e-06, |
|
"loss": 0.6088, |
|
"step": 11240 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 5.011990407673861e-06, |
|
"loss": 0.4807, |
|
"step": 11250 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 4.998667732480682e-06, |
|
"loss": 0.4774, |
|
"step": 11260 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 4.985345057287504e-06, |
|
"loss": 0.3755, |
|
"step": 11270 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 4.972022382094325e-06, |
|
"loss": 0.4934, |
|
"step": 11280 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 4.958699706901146e-06, |
|
"loss": 0.3653, |
|
"step": 11290 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 4.945377031707967e-06, |
|
"loss": 0.5374, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 4.932054356514789e-06, |
|
"loss": 0.6684, |
|
"step": 11310 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 4.91873168132161e-06, |
|
"loss": 0.3613, |
|
"step": 11320 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 4.905409006128431e-06, |
|
"loss": 0.5459, |
|
"step": 11330 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 4.892086330935253e-06, |
|
"loss": 0.4648, |
|
"step": 11340 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 4.8787636557420735e-06, |
|
"loss": 0.6087, |
|
"step": 11350 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 4.865440980548894e-06, |
|
"loss": 0.4454, |
|
"step": 11360 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 4.852118305355716e-06, |
|
"loss": 0.6797, |
|
"step": 11370 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 4.838795630162537e-06, |
|
"loss": 0.444, |
|
"step": 11380 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 4.825472954969358e-06, |
|
"loss": 0.4996, |
|
"step": 11390 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 4.8121502797761795e-06, |
|
"loss": 0.5194, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 4.798827604583e-06, |
|
"loss": 0.7734, |
|
"step": 11410 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 4.785504929389822e-06, |
|
"loss": 0.406, |
|
"step": 11420 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 4.772182254196643e-06, |
|
"loss": 0.5761, |
|
"step": 11430 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 4.758859579003464e-06, |
|
"loss": 0.4261, |
|
"step": 11440 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 4.7455369038102855e-06, |
|
"loss": 0.5295, |
|
"step": 11450 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 4.732214228617107e-06, |
|
"loss": 0.5268, |
|
"step": 11460 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 4.718891553423927e-06, |
|
"loss": 0.4441, |
|
"step": 11470 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 4.705568878230749e-06, |
|
"loss": 0.5511, |
|
"step": 11480 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 4.692246203037571e-06, |
|
"loss": 0.4482, |
|
"step": 11490 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 4.678923527844392e-06, |
|
"loss": 0.3605, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 4.6656008526512125e-06, |
|
"loss": 0.828, |
|
"step": 11510 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 4.652278177458034e-06, |
|
"loss": 0.4273, |
|
"step": 11520 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 4.638955502264855e-06, |
|
"loss": 0.6486, |
|
"step": 11530 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 4.625632827071677e-06, |
|
"loss": 0.5631, |
|
"step": 11540 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 4.612310151878498e-06, |
|
"loss": 0.5487, |
|
"step": 11550 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 4.5989874766853185e-06, |
|
"loss": 0.4099, |
|
"step": 11560 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 4.58566480149214e-06, |
|
"loss": 0.6225, |
|
"step": 11570 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 4.572342126298961e-06, |
|
"loss": 0.5682, |
|
"step": 11580 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 4.559019451105782e-06, |
|
"loss": 0.3695, |
|
"step": 11590 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 4.545696775912604e-06, |
|
"loss": 0.5973, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 4.5323741007194245e-06, |
|
"loss": 0.5033, |
|
"step": 11610 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 4.519051425526246e-06, |
|
"loss": 0.6484, |
|
"step": 11620 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 4.505728750333067e-06, |
|
"loss": 0.516, |
|
"step": 11630 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 4.492406075139889e-06, |
|
"loss": 0.6044, |
|
"step": 11640 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 4.47908339994671e-06, |
|
"loss": 0.6888, |
|
"step": 11650 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 4.465760724753531e-06, |
|
"loss": 0.6246, |
|
"step": 11660 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 4.452438049560352e-06, |
|
"loss": 0.4872, |
|
"step": 11670 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 4.439115374367173e-06, |
|
"loss": 0.4675, |
|
"step": 11680 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 4.425792699173995e-06, |
|
"loss": 0.594, |
|
"step": 11690 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 4.412470023980816e-06, |
|
"loss": 0.7868, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 4.399147348787637e-06, |
|
"loss": 0.5357, |
|
"step": 11710 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 4.385824673594458e-06, |
|
"loss": 0.4265, |
|
"step": 11720 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 4.372501998401279e-06, |
|
"loss": 0.4934, |
|
"step": 11730 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 4.3591793232081e-06, |
|
"loss": 0.5334, |
|
"step": 11740 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 4.345856648014922e-06, |
|
"loss": 0.5469, |
|
"step": 11750 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 4.332533972821743e-06, |
|
"loss": 0.5974, |
|
"step": 11760 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 4.319211297628564e-06, |
|
"loss": 0.6087, |
|
"step": 11770 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 4.305888622435385e-06, |
|
"loss": 0.4159, |
|
"step": 11780 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 4.292565947242206e-06, |
|
"loss": 0.4755, |
|
"step": 11790 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 4.279243272049028e-06, |
|
"loss": 0.5726, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 4.2659205968558496e-06, |
|
"loss": 0.4904, |
|
"step": 11810 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 4.2525979216626704e-06, |
|
"loss": 0.6391, |
|
"step": 11820 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 4.239275246469491e-06, |
|
"loss": 0.5228, |
|
"step": 11830 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 4.225952571276313e-06, |
|
"loss": 0.7083, |
|
"step": 11840 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 4.212629896083134e-06, |
|
"loss": 0.5254, |
|
"step": 11850 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 4.199307220889955e-06, |
|
"loss": 0.416, |
|
"step": 11860 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 4.1859845456967765e-06, |
|
"loss": 0.6754, |
|
"step": 11870 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 4.172661870503597e-06, |
|
"loss": 0.3771, |
|
"step": 11880 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 4.159339195310418e-06, |
|
"loss": 0.6722, |
|
"step": 11890 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 4.14601652011724e-06, |
|
"loss": 0.5206, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 4.132693844924061e-06, |
|
"loss": 0.6298, |
|
"step": 11910 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 4.1193711697308825e-06, |
|
"loss": 0.5899, |
|
"step": 11920 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 4.106048494537703e-06, |
|
"loss": 0.4625, |
|
"step": 11930 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 4.092725819344524e-06, |
|
"loss": 0.5672, |
|
"step": 11940 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 4.079403144151346e-06, |
|
"loss": 0.5623, |
|
"step": 11950 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 4.066080468958168e-06, |
|
"loss": 0.3423, |
|
"step": 11960 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 4.052757793764988e-06, |
|
"loss": 0.4891, |
|
"step": 11970 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 4.0394351185718094e-06, |
|
"loss": 0.3861, |
|
"step": 11980 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 4.026112443378631e-06, |
|
"loss": 0.4937, |
|
"step": 11990 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 4.012789768185452e-06, |
|
"loss": 0.5285, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 3.999467092992273e-06, |
|
"loss": 0.3314, |
|
"step": 12010 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 3.986144417799095e-06, |
|
"loss": 0.5348, |
|
"step": 12020 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 3.9728217426059155e-06, |
|
"loss": 0.3142, |
|
"step": 12030 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 3.959499067412737e-06, |
|
"loss": 0.644, |
|
"step": 12040 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 3.946176392219558e-06, |
|
"loss": 0.6136, |
|
"step": 12050 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 3.932853717026379e-06, |
|
"loss": 0.6213, |
|
"step": 12060 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 3.919531041833201e-06, |
|
"loss": 0.5401, |
|
"step": 12070 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 3.9062083666400215e-06, |
|
"loss": 0.4765, |
|
"step": 12080 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 3.892885691446842e-06, |
|
"loss": 0.6232, |
|
"step": 12090 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 3.879563016253664e-06, |
|
"loss": 0.5262, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 3.866240341060486e-06, |
|
"loss": 0.5294, |
|
"step": 12110 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 3.852917665867306e-06, |
|
"loss": 0.6858, |
|
"step": 12120 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 3.8395949906741276e-06, |
|
"loss": 0.5539, |
|
"step": 12130 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 3.826272315480949e-06, |
|
"loss": 0.5337, |
|
"step": 12140 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 3.81294964028777e-06, |
|
"loss": 0.4814, |
|
"step": 12150 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 3.799626965094591e-06, |
|
"loss": 0.5019, |
|
"step": 12160 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 3.7863042899014123e-06, |
|
"loss": 0.4682, |
|
"step": 12170 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 3.7729816147082336e-06, |
|
"loss": 0.6597, |
|
"step": 12180 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 3.7596589395150553e-06, |
|
"loss": 0.5827, |
|
"step": 12190 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 3.7463362643218758e-06, |
|
"loss": 0.4335, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 3.733013589128697e-06, |
|
"loss": 0.4961, |
|
"step": 12210 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 3.7196909139355188e-06, |
|
"loss": 0.5998, |
|
"step": 12220 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 3.70636823874234e-06, |
|
"loss": 0.5215, |
|
"step": 12230 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 3.693045563549161e-06, |
|
"loss": 0.6032, |
|
"step": 12240 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 3.6797228883559822e-06, |
|
"loss": 0.6204, |
|
"step": 12250 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 3.6664002131628035e-06, |
|
"loss": 0.5401, |
|
"step": 12260 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 3.653077537969625e-06, |
|
"loss": 0.7566, |
|
"step": 12270 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 3.6397548627764457e-06, |
|
"loss": 0.59, |
|
"step": 12280 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 3.626432187583267e-06, |
|
"loss": 0.5552, |
|
"step": 12290 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 3.6131095123900883e-06, |
|
"loss": 0.4806, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 3.5997868371969096e-06, |
|
"loss": 0.4438, |
|
"step": 12310 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 3.5864641620037304e-06, |
|
"loss": 0.3571, |
|
"step": 12320 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 3.5731414868105517e-06, |
|
"loss": 0.5311, |
|
"step": 12330 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 3.559818811617373e-06, |
|
"loss": 0.7225, |
|
"step": 12340 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 3.546496136424194e-06, |
|
"loss": 0.5775, |
|
"step": 12350 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 3.533173461231015e-06, |
|
"loss": 0.6592, |
|
"step": 12360 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 3.5198507860378365e-06, |
|
"loss": 0.5143, |
|
"step": 12370 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 3.506528110844658e-06, |
|
"loss": 0.6855, |
|
"step": 12380 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 3.4932054356514787e-06, |
|
"loss": 0.3721, |
|
"step": 12390 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 3.4798827604583004e-06, |
|
"loss": 0.5709, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 3.4665600852651217e-06, |
|
"loss": 0.5464, |
|
"step": 12410 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 3.453237410071943e-06, |
|
"loss": 0.5, |
|
"step": 12420 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 3.439914734878764e-06, |
|
"loss": 0.3392, |
|
"step": 12430 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 3.426592059685585e-06, |
|
"loss": 0.7597, |
|
"step": 12440 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 3.4132693844924064e-06, |
|
"loss": 0.6143, |
|
"step": 12450 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 3.3999467092992277e-06, |
|
"loss": 0.5592, |
|
"step": 12460 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 3.3866240341060486e-06, |
|
"loss": 0.6935, |
|
"step": 12470 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 3.37330135891287e-06, |
|
"loss": 0.6525, |
|
"step": 12480 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 3.359978683719691e-06, |
|
"loss": 0.4908, |
|
"step": 12490 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 3.3466560085265125e-06, |
|
"loss": 0.5525, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 3.3333333333333333e-06, |
|
"loss": 0.7124, |
|
"step": 12510 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 3.3200106581401546e-06, |
|
"loss": 0.7283, |
|
"step": 12520 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 3.3066879829469763e-06, |
|
"loss": 0.4084, |
|
"step": 12530 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 3.2933653077537976e-06, |
|
"loss": 0.5301, |
|
"step": 12540 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 3.280042632560618e-06, |
|
"loss": 0.6944, |
|
"step": 12550 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 3.26671995736744e-06, |
|
"loss": 0.545, |
|
"step": 12560 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 3.253397282174261e-06, |
|
"loss": 0.5842, |
|
"step": 12570 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 3.2400746069810824e-06, |
|
"loss": 0.3925, |
|
"step": 12580 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 3.2267519317879032e-06, |
|
"loss": 0.4888, |
|
"step": 12590 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 3.2134292565947245e-06, |
|
"loss": 0.777, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 3.200106581401546e-06, |
|
"loss": 0.7409, |
|
"step": 12610 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 3.1867839062083667e-06, |
|
"loss": 0.4189, |
|
"step": 12620 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 3.173461231015188e-06, |
|
"loss": 0.4592, |
|
"step": 12630 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 3.1601385558220093e-06, |
|
"loss": 0.5899, |
|
"step": 12640 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 3.1468158806288306e-06, |
|
"loss": 0.5961, |
|
"step": 12650 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 3.1334932054356515e-06, |
|
"loss": 0.4069, |
|
"step": 12660 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 3.1201705302424727e-06, |
|
"loss": 0.4548, |
|
"step": 12670 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 3.106847855049294e-06, |
|
"loss": 0.4679, |
|
"step": 12680 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 3.0935251798561158e-06, |
|
"loss": 0.3883, |
|
"step": 12690 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 3.080202504662936e-06, |
|
"loss": 0.3436, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 3.0668798294697575e-06, |
|
"loss": 0.6312, |
|
"step": 12710 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 3.0535571542765792e-06, |
|
"loss": 0.6167, |
|
"step": 12720 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 3.0402344790834005e-06, |
|
"loss": 0.4497, |
|
"step": 12730 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 3.0269118038902214e-06, |
|
"loss": 0.4056, |
|
"step": 12740 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 3.0135891286970427e-06, |
|
"loss": 0.6267, |
|
"step": 12750 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 3.000266453503864e-06, |
|
"loss": 0.5406, |
|
"step": 12760 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 2.9869437783106853e-06, |
|
"loss": 0.5243, |
|
"step": 12770 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 2.973621103117506e-06, |
|
"loss": 0.5, |
|
"step": 12780 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 2.9602984279243274e-06, |
|
"loss": 0.4584, |
|
"step": 12790 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 2.9469757527311487e-06, |
|
"loss": 0.758, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 2.93365307753797e-06, |
|
"loss": 0.3937, |
|
"step": 12810 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 2.920330402344791e-06, |
|
"loss": 0.4183, |
|
"step": 12820 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 2.907007727151612e-06, |
|
"loss": 0.518, |
|
"step": 12830 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 2.8936850519584335e-06, |
|
"loss": 0.4435, |
|
"step": 12840 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 2.8803623767652543e-06, |
|
"loss": 0.6761, |
|
"step": 12850 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 2.8670397015720756e-06, |
|
"loss": 0.5532, |
|
"step": 12860 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 2.8537170263788973e-06, |
|
"loss": 0.3782, |
|
"step": 12870 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 2.8403943511857186e-06, |
|
"loss": 0.7742, |
|
"step": 12880 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 2.827071675992539e-06, |
|
"loss": 0.4839, |
|
"step": 12890 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 2.813749000799361e-06, |
|
"loss": 0.4243, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 2.800426325606182e-06, |
|
"loss": 0.5923, |
|
"step": 12910 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 2.7871036504130034e-06, |
|
"loss": 0.5949, |
|
"step": 12920 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 2.7737809752198243e-06, |
|
"loss": 0.5117, |
|
"step": 12930 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 2.7604583000266455e-06, |
|
"loss": 0.7791, |
|
"step": 12940 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 2.747135624833467e-06, |
|
"loss": 0.5184, |
|
"step": 12950 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 2.733812949640288e-06, |
|
"loss": 0.5903, |
|
"step": 12960 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 2.720490274447109e-06, |
|
"loss": 0.6456, |
|
"step": 12970 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 2.7071675992539303e-06, |
|
"loss": 0.4857, |
|
"step": 12980 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 2.6938449240607516e-06, |
|
"loss": 0.5727, |
|
"step": 12990 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 2.680522248867573e-06, |
|
"loss": 0.4204, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 2.6671995736743938e-06, |
|
"loss": 0.3454, |
|
"step": 13010 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 2.653876898481215e-06, |
|
"loss": 0.6435, |
|
"step": 13020 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 2.6405542232880368e-06, |
|
"loss": 0.6825, |
|
"step": 13030 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 2.627231548094858e-06, |
|
"loss": 0.7152, |
|
"step": 13040 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 2.6139088729016785e-06, |
|
"loss": 0.448, |
|
"step": 13050 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 2.6005861977085002e-06, |
|
"loss": 0.4351, |
|
"step": 13060 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 2.5872635225153215e-06, |
|
"loss": 0.7678, |
|
"step": 13070 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 2.5739408473221424e-06, |
|
"loss": 0.3533, |
|
"step": 13080 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 2.5606181721289637e-06, |
|
"loss": 0.6413, |
|
"step": 13090 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 2.547295496935785e-06, |
|
"loss": 0.5917, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 2.5339728217426063e-06, |
|
"loss": 0.5904, |
|
"step": 13110 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 2.520650146549427e-06, |
|
"loss": 0.5905, |
|
"step": 13120 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 2.5073274713562484e-06, |
|
"loss": 0.3289, |
|
"step": 13130 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 2.4940047961630697e-06, |
|
"loss": 0.4646, |
|
"step": 13140 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 2.480682120969891e-06, |
|
"loss": 0.5391, |
|
"step": 13150 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 2.4673594457767123e-06, |
|
"loss": 0.4321, |
|
"step": 13160 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 2.454036770583533e-06, |
|
"loss": 0.7442, |
|
"step": 13170 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 2.4407140953903545e-06, |
|
"loss": 0.5778, |
|
"step": 13180 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 2.4273914201971758e-06, |
|
"loss": 0.4422, |
|
"step": 13190 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 2.414068745003997e-06, |
|
"loss": 0.6558, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 2.4007460698108184e-06, |
|
"loss": 0.7679, |
|
"step": 13210 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 2.3874233946176396e-06, |
|
"loss": 0.4981, |
|
"step": 13220 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 2.3741007194244605e-06, |
|
"loss": 0.6829, |
|
"step": 13230 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 2.360778044231282e-06, |
|
"loss": 0.3649, |
|
"step": 13240 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 2.347455369038103e-06, |
|
"loss": 0.5827, |
|
"step": 13250 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 2.3341326938449244e-06, |
|
"loss": 0.4577, |
|
"step": 13260 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 2.3208100186517453e-06, |
|
"loss": 0.3576, |
|
"step": 13270 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 2.307487343458567e-06, |
|
"loss": 0.566, |
|
"step": 13280 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 2.294164668265388e-06, |
|
"loss": 0.4685, |
|
"step": 13290 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 2.280841993072209e-06, |
|
"loss": 0.5552, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 2.2675193178790304e-06, |
|
"loss": 0.4454, |
|
"step": 13310 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 2.2541966426858513e-06, |
|
"loss": 0.6684, |
|
"step": 13320 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 2.2408739674926726e-06, |
|
"loss": 0.5333, |
|
"step": 13330 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 2.227551292299494e-06, |
|
"loss": 0.5584, |
|
"step": 13340 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 2.214228617106315e-06, |
|
"loss": 0.592, |
|
"step": 13350 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 2.200905941913136e-06, |
|
"loss": 0.5376, |
|
"step": 13360 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 2.1875832667199578e-06, |
|
"loss": 0.3944, |
|
"step": 13370 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 2.1742605915267786e-06, |
|
"loss": 0.5578, |
|
"step": 13380 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 2.1609379163336e-06, |
|
"loss": 0.3291, |
|
"step": 13390 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 2.1476152411404212e-06, |
|
"loss": 0.6107, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 2.1342925659472425e-06, |
|
"loss": 0.5605, |
|
"step": 13410 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 2.1209698907540634e-06, |
|
"loss": 0.7272, |
|
"step": 13420 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 2.1076472155608847e-06, |
|
"loss": 0.6814, |
|
"step": 13430 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 2.094324540367706e-06, |
|
"loss": 0.5591, |
|
"step": 13440 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 2.0810018651745273e-06, |
|
"loss": 0.4281, |
|
"step": 13450 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 2.0676791899813486e-06, |
|
"loss": 0.4085, |
|
"step": 13460 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 2.05435651478817e-06, |
|
"loss": 0.4977, |
|
"step": 13470 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 2.0410338395949907e-06, |
|
"loss": 0.403, |
|
"step": 13480 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 2.027711164401812e-06, |
|
"loss": 0.5554, |
|
"step": 13490 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 2.0143884892086333e-06, |
|
"loss": 0.563, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 2.0010658140154546e-06, |
|
"loss": 0.5317, |
|
"step": 13510 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 1.9877431388222755e-06, |
|
"loss": 0.5416, |
|
"step": 13520 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 1.974420463629097e-06, |
|
"loss": 0.59, |
|
"step": 13530 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 1.961097788435918e-06, |
|
"loss": 0.4728, |
|
"step": 13540 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 1.9477751132427394e-06, |
|
"loss": 0.6021, |
|
"step": 13550 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 1.9344524380495607e-06, |
|
"loss": 0.5801, |
|
"step": 13560 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 1.9211297628563815e-06, |
|
"loss": 0.5634, |
|
"step": 13570 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 1.907807087663203e-06, |
|
"loss": 0.5021, |
|
"step": 13580 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 1.8944844124700241e-06, |
|
"loss": 0.5824, |
|
"step": 13590 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 1.8811617372768454e-06, |
|
"loss": 0.6598, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 1.8678390620836665e-06, |
|
"loss": 0.5233, |
|
"step": 13610 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 1.8545163868904878e-06, |
|
"loss": 0.7899, |
|
"step": 13620 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 1.8411937116973089e-06, |
|
"loss": 0.4576, |
|
"step": 13630 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 1.8278710365041302e-06, |
|
"loss": 0.4726, |
|
"step": 13640 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 1.8145483613109512e-06, |
|
"loss": 0.6278, |
|
"step": 13650 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 1.8012256861177727e-06, |
|
"loss": 0.4879, |
|
"step": 13660 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 1.7879030109245938e-06, |
|
"loss": 0.5082, |
|
"step": 13670 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 1.7745803357314151e-06, |
|
"loss": 0.5767, |
|
"step": 13680 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 1.7612576605382362e-06, |
|
"loss": 0.4782, |
|
"step": 13690 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 1.7479349853450575e-06, |
|
"loss": 0.5911, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 1.7346123101518786e-06, |
|
"loss": 0.4495, |
|
"step": 13710 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 1.7212896349586999e-06, |
|
"loss": 0.6191, |
|
"step": 13720 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 1.707966959765521e-06, |
|
"loss": 0.5123, |
|
"step": 13730 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 1.6946442845723425e-06, |
|
"loss": 0.6759, |
|
"step": 13740 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 1.6813216093791635e-06, |
|
"loss": 0.6946, |
|
"step": 13750 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 1.6679989341859848e-06, |
|
"loss": 0.8245, |
|
"step": 13760 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 1.654676258992806e-06, |
|
"loss": 0.4792, |
|
"step": 13770 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 1.6413535837996272e-06, |
|
"loss": 0.4847, |
|
"step": 13780 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 1.6280309086064483e-06, |
|
"loss": 0.5823, |
|
"step": 13790 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 1.6147082334132696e-06, |
|
"loss": 0.6636, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 1.6013855582200907e-06, |
|
"loss": 0.4208, |
|
"step": 13810 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 1.5880628830269117e-06, |
|
"loss": 0.4308, |
|
"step": 13820 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 1.5747402078337332e-06, |
|
"loss": 0.6352, |
|
"step": 13830 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 1.5614175326405543e-06, |
|
"loss": 0.4812, |
|
"step": 13840 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 1.5480948574473756e-06, |
|
"loss": 0.281, |
|
"step": 13850 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 1.5347721822541967e-06, |
|
"loss": 0.6616, |
|
"step": 13860 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 1.521449507061018e-06, |
|
"loss": 0.6039, |
|
"step": 13870 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 1.508126831867839e-06, |
|
"loss": 0.5124, |
|
"step": 13880 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 1.4948041566746604e-06, |
|
"loss": 0.5828, |
|
"step": 13890 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 1.4814814814814815e-06, |
|
"loss": 0.4895, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 1.468158806288303e-06, |
|
"loss": 0.5507, |
|
"step": 13910 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 1.454836131095124e-06, |
|
"loss": 0.3509, |
|
"step": 13920 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 1.4415134559019453e-06, |
|
"loss": 0.6344, |
|
"step": 13930 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 1.4281907807087664e-06, |
|
"loss": 0.435, |
|
"step": 13940 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 1.4148681055155877e-06, |
|
"loss": 0.5045, |
|
"step": 13950 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 1.4015454303224088e-06, |
|
"loss": 0.6366, |
|
"step": 13960 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 1.38822275512923e-06, |
|
"loss": 0.5821, |
|
"step": 13970 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 1.3749000799360512e-06, |
|
"loss": 0.3629, |
|
"step": 13980 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 1.3615774047428727e-06, |
|
"loss": 0.4365, |
|
"step": 13990 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.3482547295496938e-06, |
|
"loss": 0.5178, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.334932054356515e-06, |
|
"loss": 0.7401, |
|
"step": 14010 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.3216093791633361e-06, |
|
"loss": 0.59, |
|
"step": 14020 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.3082867039701574e-06, |
|
"loss": 0.5952, |
|
"step": 14030 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.2949640287769785e-06, |
|
"loss": 0.7081, |
|
"step": 14040 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.2816413535837996e-06, |
|
"loss": 0.3996, |
|
"step": 14050 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.2683186783906209e-06, |
|
"loss": 0.537, |
|
"step": 14060 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.254996003197442e-06, |
|
"loss": 0.4258, |
|
"step": 14070 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.2416733280042635e-06, |
|
"loss": 0.3974, |
|
"step": 14080 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.2283506528110845e-06, |
|
"loss": 0.5042, |
|
"step": 14090 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.2150279776179058e-06, |
|
"loss": 0.5661, |
|
"step": 14100 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.201705302424727e-06, |
|
"loss": 0.457, |
|
"step": 14110 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.1883826272315482e-06, |
|
"loss": 0.6158, |
|
"step": 14120 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.1750599520383695e-06, |
|
"loss": 0.5112, |
|
"step": 14130 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.1617372768451906e-06, |
|
"loss": 0.7427, |
|
"step": 14140 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.1484146016520119e-06, |
|
"loss": 0.548, |
|
"step": 14150 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.1350919264588332e-06, |
|
"loss": 0.4385, |
|
"step": 14160 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.1217692512656543e-06, |
|
"loss": 0.6332, |
|
"step": 14170 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.1084465760724753e-06, |
|
"loss": 0.6839, |
|
"step": 14180 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.0951239008792966e-06, |
|
"loss": 0.5741, |
|
"step": 14190 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.0818012256861177e-06, |
|
"loss": 0.4525, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.068478550492939e-06, |
|
"loss": 0.4109, |
|
"step": 14210 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.0551558752997603e-06, |
|
"loss": 0.4207, |
|
"step": 14220 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 1.0418332001065814e-06, |
|
"loss": 0.5708, |
|
"step": 14230 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 1.0285105249134027e-06, |
|
"loss": 0.5641, |
|
"step": 14240 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 1.015187849720224e-06, |
|
"loss": 0.5639, |
|
"step": 14250 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 1.001865174527045e-06, |
|
"loss": 0.5215, |
|
"step": 14260 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 9.885424993338663e-07, |
|
"loss": 0.5582, |
|
"step": 14270 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 9.752198241406874e-07, |
|
"loss": 0.6286, |
|
"step": 14280 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 9.618971489475087e-07, |
|
"loss": 0.4788, |
|
"step": 14290 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 9.485744737543299e-07, |
|
"loss": 0.5799, |
|
"step": 14300 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 9.352517985611512e-07, |
|
"loss": 0.7389, |
|
"step": 14310 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 9.219291233679724e-07, |
|
"loss": 0.6456, |
|
"step": 14320 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 9.086064481747936e-07, |
|
"loss": 0.4753, |
|
"step": 14330 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 8.952837729816148e-07, |
|
"loss": 0.5516, |
|
"step": 14340 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 8.819610977884361e-07, |
|
"loss": 0.6144, |
|
"step": 14350 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 8.686384225952572e-07, |
|
"loss": 0.5399, |
|
"step": 14360 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 8.553157474020784e-07, |
|
"loss": 0.4359, |
|
"step": 14370 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 8.419930722088996e-07, |
|
"loss": 0.6362, |
|
"step": 14380 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 8.286703970157209e-07, |
|
"loss": 0.6311, |
|
"step": 14390 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 8.153477218225421e-07, |
|
"loss": 0.528, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 8.020250466293632e-07, |
|
"loss": 0.4934, |
|
"step": 14410 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 7.887023714361844e-07, |
|
"loss": 0.3794, |
|
"step": 14420 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 7.753796962430056e-07, |
|
"loss": 0.622, |
|
"step": 14430 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 7.620570210498268e-07, |
|
"loss": 0.542, |
|
"step": 14440 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 7.48734345856648e-07, |
|
"loss": 0.5141, |
|
"step": 14450 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 7.354116706634692e-07, |
|
"loss": 0.6752, |
|
"step": 14460 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 7.220889954702904e-07, |
|
"loss": 0.5327, |
|
"step": 14470 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 7.087663202771117e-07, |
|
"loss": 0.4935, |
|
"step": 14480 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 6.954436450839329e-07, |
|
"loss": 0.53, |
|
"step": 14490 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 6.821209698907541e-07, |
|
"loss": 0.5226, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 6.687982946975753e-07, |
|
"loss": 0.4801, |
|
"step": 14510 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 6.554756195043966e-07, |
|
"loss": 0.5048, |
|
"step": 14520 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 6.421529443112177e-07, |
|
"loss": 0.5826, |
|
"step": 14530 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 6.288302691180389e-07, |
|
"loss": 0.553, |
|
"step": 14540 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 6.155075939248602e-07, |
|
"loss": 0.592, |
|
"step": 14550 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 6.021849187316814e-07, |
|
"loss": 0.5556, |
|
"step": 14560 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 5.888622435385026e-07, |
|
"loss": 0.5731, |
|
"step": 14570 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 5.755395683453238e-07, |
|
"loss": 0.5511, |
|
"step": 14580 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 5.62216893152145e-07, |
|
"loss": 0.6793, |
|
"step": 14590 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 5.488942179589662e-07, |
|
"loss": 0.4778, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 5.355715427657874e-07, |
|
"loss": 0.5263, |
|
"step": 14610 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 5.222488675726086e-07, |
|
"loss": 0.5329, |
|
"step": 14620 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 5.089261923794298e-07, |
|
"loss": 0.4903, |
|
"step": 14630 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 4.95603517186251e-07, |
|
"loss": 0.5053, |
|
"step": 14640 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 4.822808419930722e-07, |
|
"loss": 0.6906, |
|
"step": 14650 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 4.6895816679989345e-07, |
|
"loss": 0.4313, |
|
"step": 14660 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 4.556354916067147e-07, |
|
"loss": 0.3458, |
|
"step": 14670 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 4.423128164135359e-07, |
|
"loss": 0.4378, |
|
"step": 14680 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 4.289901412203571e-07, |
|
"loss": 0.6225, |
|
"step": 14690 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 4.156674660271783e-07, |
|
"loss": 0.4633, |
|
"step": 14700 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 4.0234479083399944e-07, |
|
"loss": 0.5472, |
|
"step": 14710 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 3.890221156408207e-07, |
|
"loss": 0.4989, |
|
"step": 14720 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 3.756994404476419e-07, |
|
"loss": 0.6646, |
|
"step": 14730 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 3.623767652544631e-07, |
|
"loss": 0.5993, |
|
"step": 14740 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 3.4905409006128435e-07, |
|
"loss": 0.5009, |
|
"step": 14750 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 3.3573141486810554e-07, |
|
"loss": 0.4229, |
|
"step": 14760 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 3.224087396749268e-07, |
|
"loss": 0.7085, |
|
"step": 14770 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 3.0908606448174796e-07, |
|
"loss": 0.4241, |
|
"step": 14780 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 2.957633892885692e-07, |
|
"loss": 0.5699, |
|
"step": 14790 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 2.824407140953904e-07, |
|
"loss": 0.5307, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 2.691180389022116e-07, |
|
"loss": 0.4612, |
|
"step": 14810 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 2.5579536370903277e-07, |
|
"loss": 0.5194, |
|
"step": 14820 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 2.42472688515854e-07, |
|
"loss": 0.4469, |
|
"step": 14830 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 2.2915001332267522e-07, |
|
"loss": 0.3941, |
|
"step": 14840 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 2.1582733812949643e-07, |
|
"loss": 0.4714, |
|
"step": 14850 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 2.0250466293631765e-07, |
|
"loss": 0.5364, |
|
"step": 14860 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 1.8918198774313884e-07, |
|
"loss": 0.4054, |
|
"step": 14870 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 1.7585931254996005e-07, |
|
"loss": 0.4351, |
|
"step": 14880 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 1.6253663735678126e-07, |
|
"loss": 0.4373, |
|
"step": 14890 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 1.4921396216360248e-07, |
|
"loss": 0.5169, |
|
"step": 14900 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 1.3589128697042367e-07, |
|
"loss": 0.4695, |
|
"step": 14910 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 1.2256861177724488e-07, |
|
"loss": 0.6149, |
|
"step": 14920 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 1.0924593658406609e-07, |
|
"loss": 0.6501, |
|
"step": 14930 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 9.59232613908873e-08, |
|
"loss": 0.6889, |
|
"step": 14940 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 8.260058619770851e-08, |
|
"loss": 0.5285, |
|
"step": 14950 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 6.927791100452971e-08, |
|
"loss": 0.5547, |
|
"step": 14960 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 5.595523581135092e-08, |
|
"loss": 0.3699, |
|
"step": 14970 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 4.263256061817213e-08, |
|
"loss": 0.4134, |
|
"step": 14980 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 2.930988542499334e-08, |
|
"loss": 0.7625, |
|
"step": 14990 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 1.5987210231814548e-08, |
|
"loss": 0.546, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 2.664535038635758e-09, |
|
"loss": 0.4524, |
|
"step": 15010 |
|
} |
|
], |
|
"max_steps": 15012, |
|
"num_train_epochs": 2, |
|
"total_flos": 1.1146687779176448e+17, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|