|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 3.0, |
|
"eval_steps": 500, |
|
"global_step": 18384, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 3.2608695652173915e-07, |
|
"loss": 1.057, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 6.884057971014494e-07, |
|
"loss": 1.0404, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.0144927536231885e-06, |
|
"loss": 1.0415, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.3405797101449278e-06, |
|
"loss": 1.0467, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.7028985507246378e-06, |
|
"loss": 1.0394, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.065217391304348e-06, |
|
"loss": 1.0325, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.4275362318840583e-06, |
|
"loss": 1.0236, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.7898550724637686e-06, |
|
"loss": 1.0099, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 3.152173913043479e-06, |
|
"loss": 1.024, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.5144927536231887e-06, |
|
"loss": 0.9877, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.8768115942028985e-06, |
|
"loss": 0.9921, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.239130434782609e-06, |
|
"loss": 0.982, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.601449275362319e-06, |
|
"loss": 0.9623, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.963768115942029e-06, |
|
"loss": 0.9441, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 5.3260869565217395e-06, |
|
"loss": 0.9225, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 5.688405797101449e-06, |
|
"loss": 0.9129, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 6.05072463768116e-06, |
|
"loss": 0.8737, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 6.41304347826087e-06, |
|
"loss": 0.8554, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 6.7753623188405805e-06, |
|
"loss": 0.8249, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 7.13768115942029e-06, |
|
"loss": 0.796, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 7.500000000000001e-06, |
|
"loss": 0.7771, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 7.862318840579712e-06, |
|
"loss": 0.7432, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.22463768115942e-06, |
|
"loss": 0.7279, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.586956521739131e-06, |
|
"loss": 0.7114, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.94927536231884e-06, |
|
"loss": 0.6932, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.275362318840581e-06, |
|
"loss": 0.726, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.63768115942029e-06, |
|
"loss": 0.6796, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1e-05, |
|
"loss": 0.6754, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.036231884057971e-05, |
|
"loss": 0.6558, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.0724637681159422e-05, |
|
"loss": 0.6553, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.1086956521739131e-05, |
|
"loss": 0.6355, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.1449275362318842e-05, |
|
"loss": 0.6364, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.181159420289855e-05, |
|
"loss": 0.6412, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.2173913043478263e-05, |
|
"loss": 0.6304, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.2536231884057972e-05, |
|
"loss": 0.6308, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.2898550724637681e-05, |
|
"loss": 0.6139, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.3260869565217392e-05, |
|
"loss": 0.6292, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.3623188405797103e-05, |
|
"loss": 0.6259, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.3985507246376813e-05, |
|
"loss": 0.6165, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.4347826086956522e-05, |
|
"loss": 0.6156, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.4710144927536235e-05, |
|
"loss": 0.622, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.5072463768115944e-05, |
|
"loss": 0.6195, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.5434782608695654e-05, |
|
"loss": 0.612, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.5797101449275363e-05, |
|
"loss": 0.6231, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.6159420289855076e-05, |
|
"loss": 0.6172, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.6521739130434785e-05, |
|
"loss": 0.6111, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.6884057971014494e-05, |
|
"loss": 0.5962, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.7246376811594206e-05, |
|
"loss": 0.621, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.7608695652173915e-05, |
|
"loss": 0.5874, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.7971014492753624e-05, |
|
"loss": 0.5981, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.8333333333333333e-05, |
|
"loss": 0.6084, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.8695652173913045e-05, |
|
"loss": 0.6028, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.9057971014492754e-05, |
|
"loss": 0.6004, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.9420289855072467e-05, |
|
"loss": 0.5926, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.9782608695652176e-05, |
|
"loss": 0.5886, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.9995513683266043e-05, |
|
"loss": 0.5958, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.9984297891431138e-05, |
|
"loss": 0.596, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.9973082099596232e-05, |
|
"loss": 0.5946, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.996186630776133e-05, |
|
"loss": 0.6011, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.9950650515926425e-05, |
|
"loss": 0.5956, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.9939434724091523e-05, |
|
"loss": 0.5838, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.992821893225662e-05, |
|
"loss": 0.5925, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.9917003140421716e-05, |
|
"loss": 0.5697, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.990578734858681e-05, |
|
"loss": 0.5766, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.989457155675191e-05, |
|
"loss": 0.5821, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.9883355764917006e-05, |
|
"loss": 0.5945, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.98721399730821e-05, |
|
"loss": 0.5798, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.98609241812472e-05, |
|
"loss": 0.5778, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.9849708389412294e-05, |
|
"loss": 0.5762, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.983849259757739e-05, |
|
"loss": 0.5755, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.9827276805742487e-05, |
|
"loss": 0.5721, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.9816061013907585e-05, |
|
"loss": 0.5892, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.980484522207268e-05, |
|
"loss": 0.5829, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.9793629430237777e-05, |
|
"loss": 0.5775, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.9782413638402872e-05, |
|
"loss": 0.5792, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.977119784656797e-05, |
|
"loss": 0.579, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.9759982054733068e-05, |
|
"loss": 0.5767, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.9748766262898163e-05, |
|
"loss": 0.5719, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.9737550471063257e-05, |
|
"loss": 0.5804, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.9726334679228355e-05, |
|
"loss": 0.5908, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.971511888739345e-05, |
|
"loss": 0.5737, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.9703903095558548e-05, |
|
"loss": 0.5715, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.9692687303723646e-05, |
|
"loss": 0.5866, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.968147151188874e-05, |
|
"loss": 0.5663, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.9670255720053836e-05, |
|
"loss": 0.5799, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.9659039928218934e-05, |
|
"loss": 0.5682, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.964782413638403e-05, |
|
"loss": 0.5795, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.9636608344549126e-05, |
|
"loss": 0.5683, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.9625392552714224e-05, |
|
"loss": 0.5683, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.961417676087932e-05, |
|
"loss": 0.5645, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.9602960969044417e-05, |
|
"loss": 0.572, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.9591745177209515e-05, |
|
"loss": 0.5552, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.958052938537461e-05, |
|
"loss": 0.5713, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.9569313593539704e-05, |
|
"loss": 0.5688, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.9558097801704803e-05, |
|
"loss": 0.5666, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.9546882009869897e-05, |
|
"loss": 0.566, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.9535666218034995e-05, |
|
"loss": 0.5658, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.9524450426200093e-05, |
|
"loss": 0.5718, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.9513234634365188e-05, |
|
"loss": 0.5559, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.9502018842530283e-05, |
|
"loss": 0.5644, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.949080305069538e-05, |
|
"loss": 0.5648, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.947958725886048e-05, |
|
"loss": 0.562, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.9468371467025573e-05, |
|
"loss": 0.5652, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.945715567519067e-05, |
|
"loss": 0.5679, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.9445939883355766e-05, |
|
"loss": 0.5643, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.943472409152086e-05, |
|
"loss": 0.5566, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.942350829968596e-05, |
|
"loss": 0.566, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.9412292507851057e-05, |
|
"loss": 0.5593, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.940107671601615e-05, |
|
"loss": 0.5719, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.938986092418125e-05, |
|
"loss": 0.5686, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.9378645132346344e-05, |
|
"loss": 0.5606, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.9367429340511442e-05, |
|
"loss": 0.5578, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.935621354867654e-05, |
|
"loss": 0.5519, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.9344997756841635e-05, |
|
"loss": 0.5563, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.933378196500673e-05, |
|
"loss": 0.5595, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.9322566173171828e-05, |
|
"loss": 0.55, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.9311350381336926e-05, |
|
"loss": 0.5602, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.930013458950202e-05, |
|
"loss": 0.5553, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.928891879766712e-05, |
|
"loss": 0.5724, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.9277703005832213e-05, |
|
"loss": 0.5535, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.9266487213997308e-05, |
|
"loss": 0.5615, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.9255271422162406e-05, |
|
"loss": 0.5656, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.9244055630327504e-05, |
|
"loss": 0.5501, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.92328398384926e-05, |
|
"loss": 0.5659, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.9221624046657697e-05, |
|
"loss": 0.5625, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.921040825482279e-05, |
|
"loss": 0.5602, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.919919246298789e-05, |
|
"loss": 0.5519, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.9187976671152987e-05, |
|
"loss": 0.549, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.9176760879318082e-05, |
|
"loss": 0.5636, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.9165545087483177e-05, |
|
"loss": 0.553, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.9154329295648275e-05, |
|
"loss": 0.5597, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.914311350381337e-05, |
|
"loss": 0.5419, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.9131897711978467e-05, |
|
"loss": 0.5368, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.9120681920143566e-05, |
|
"loss": 0.555, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.910946612830866e-05, |
|
"loss": 0.5579, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.9098250336473755e-05, |
|
"loss": 0.5479, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.9087034544638853e-05, |
|
"loss": 0.563, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.907581875280395e-05, |
|
"loss": 0.5491, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.9064602960969046e-05, |
|
"loss": 0.5491, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.9053387169134144e-05, |
|
"loss": 0.5511, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.904217137729924e-05, |
|
"loss": 0.5562, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.9030955585464333e-05, |
|
"loss": 0.5568, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.901973979362943e-05, |
|
"loss": 0.5497, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.900852400179453e-05, |
|
"loss": 0.5571, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.8997308209959624e-05, |
|
"loss": 0.5538, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.8986092418124722e-05, |
|
"loss": 0.5479, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.8974876626289816e-05, |
|
"loss": 0.556, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.8963660834454915e-05, |
|
"loss": 0.5588, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.8952445042620013e-05, |
|
"loss": 0.5358, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.8941229250785107e-05, |
|
"loss": 0.5424, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.8930013458950202e-05, |
|
"loss": 0.5487, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.89187976671153e-05, |
|
"loss": 0.5383, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.8907581875280398e-05, |
|
"loss": 0.5493, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.8896366083445493e-05, |
|
"loss": 0.5604, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.888515029161059e-05, |
|
"loss": 0.5501, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.8873934499775685e-05, |
|
"loss": 0.5419, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.886271870794078e-05, |
|
"loss": 0.5444, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.8851502916105878e-05, |
|
"loss": 0.5508, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.8840287124270976e-05, |
|
"loss": 0.5488, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.882907133243607e-05, |
|
"loss": 0.5653, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.881785554060117e-05, |
|
"loss": 0.5458, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.8806639748766264e-05, |
|
"loss": 0.5548, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.879542395693136e-05, |
|
"loss": 0.5379, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.878420816509646e-05, |
|
"loss": 0.5567, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.8772992373261554e-05, |
|
"loss": 0.5523, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.876177658142665e-05, |
|
"loss": 0.5509, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.8750560789591747e-05, |
|
"loss": 0.5523, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.873934499775684e-05, |
|
"loss": 0.5454, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.872812920592194e-05, |
|
"loss": 0.5513, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.8716913414087038e-05, |
|
"loss": 0.5462, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.8705697622252132e-05, |
|
"loss": 0.5468, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.8694481830417227e-05, |
|
"loss": 0.5398, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.8683266038582325e-05, |
|
"loss": 0.5456, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.8672050246747423e-05, |
|
"loss": 0.5361, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.8660834454912518e-05, |
|
"loss": 0.5508, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.8649618663077616e-05, |
|
"loss": 0.5339, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.863840287124271e-05, |
|
"loss": 0.5403, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.8627187079407805e-05, |
|
"loss": 0.548, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.8615971287572903e-05, |
|
"loss": 0.5424, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.8604755495738e-05, |
|
"loss": 0.5378, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.8593539703903096e-05, |
|
"loss": 0.5549, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.8582323912068194e-05, |
|
"loss": 0.5467, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.857110812023329e-05, |
|
"loss": 0.5449, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.8559892328398387e-05, |
|
"loss": 0.545, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.8548676536563485e-05, |
|
"loss": 0.5476, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.853746074472858e-05, |
|
"loss": 0.5323, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.8526244952893674e-05, |
|
"loss": 0.5532, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.8515029161058772e-05, |
|
"loss": 0.5405, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.850381336922387e-05, |
|
"loss": 0.5494, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.8492597577388965e-05, |
|
"loss": 0.5471, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.8481381785554063e-05, |
|
"loss": 0.5387, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.8470165993719158e-05, |
|
"loss": 0.5436, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.8458950201884252e-05, |
|
"loss": 0.5469, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.844773441004935e-05, |
|
"loss": 0.529, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.843651861821445e-05, |
|
"loss": 0.5439, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.8425302826379543e-05, |
|
"loss": 0.5408, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.841408703454464e-05, |
|
"loss": 0.5361, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.8402871242709736e-05, |
|
"loss": 0.5388, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.8391655450874834e-05, |
|
"loss": 0.5369, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.8380439659039932e-05, |
|
"loss": 0.5366, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.8369223867205027e-05, |
|
"loss": 0.5493, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.835800807537012e-05, |
|
"loss": 0.5346, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.834679228353522e-05, |
|
"loss": 0.5455, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.8335576491700314e-05, |
|
"loss": 0.5452, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.8324360699865412e-05, |
|
"loss": 0.5398, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.831314490803051e-05, |
|
"loss": 0.5558, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.8301929116195605e-05, |
|
"loss": 0.5367, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.82907133243607e-05, |
|
"loss": 0.5377, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.8279497532525797e-05, |
|
"loss": 0.5487, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.8268281740690895e-05, |
|
"loss": 0.5455, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.825706594885599e-05, |
|
"loss": 0.5506, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.8245850157021088e-05, |
|
"loss": 0.533, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.8234634365186183e-05, |
|
"loss": 0.5375, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.8223418573351277e-05, |
|
"loss": 0.5261, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.8212202781516376e-05, |
|
"loss": 0.5478, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.8200986989681474e-05, |
|
"loss": 0.5414, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.8189771197846568e-05, |
|
"loss": 0.5471, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.8178555406011666e-05, |
|
"loss": 0.53, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.816733961417676e-05, |
|
"loss": 0.5265, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.815612382234186e-05, |
|
"loss": 0.5309, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.8144908030506957e-05, |
|
"loss": 0.5479, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.8133692238672052e-05, |
|
"loss": 0.546, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.8122476446837146e-05, |
|
"loss": 0.5355, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.8111260655002244e-05, |
|
"loss": 0.5308, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.8100044863167343e-05, |
|
"loss": 0.5306, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.8088829071332437e-05, |
|
"loss": 0.54, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.8077613279497535e-05, |
|
"loss": 0.5279, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.806639748766263e-05, |
|
"loss": 0.5311, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.8055181695827725e-05, |
|
"loss": 0.5315, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.8043965903992823e-05, |
|
"loss": 0.5456, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.803275011215792e-05, |
|
"loss": 0.5342, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.8021534320323015e-05, |
|
"loss": 0.5445, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.8010318528488113e-05, |
|
"loss": 0.5295, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.7999102736653208e-05, |
|
"loss": 0.5337, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.7987886944818306e-05, |
|
"loss": 0.5375, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.7976671152983404e-05, |
|
"loss": 0.5292, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.79654553611485e-05, |
|
"loss": 0.5365, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.7954239569313593e-05, |
|
"loss": 0.5428, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.794302377747869e-05, |
|
"loss": 0.5489, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.7931807985643786e-05, |
|
"loss": 0.5344, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.7920592193808884e-05, |
|
"loss": 0.5293, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.7909376401973982e-05, |
|
"loss": 0.5282, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.7898160610139077e-05, |
|
"loss": 0.5259, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.788694481830417e-05, |
|
"loss": 0.5316, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.787572902646927e-05, |
|
"loss": 0.5329, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.7864513234634368e-05, |
|
"loss": 0.515, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.7853297442799462e-05, |
|
"loss": 0.5393, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.784208165096456e-05, |
|
"loss": 0.5301, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.7830865859129655e-05, |
|
"loss": 0.5246, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.7819650067294753e-05, |
|
"loss": 0.5313, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.7808434275459848e-05, |
|
"loss": 0.5329, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.7797218483624946e-05, |
|
"loss": 0.5419, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.778600269179004e-05, |
|
"loss": 0.5322, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.777478689995514e-05, |
|
"loss": 0.5385, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.7763571108120233e-05, |
|
"loss": 0.5218, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.775235531628533e-05, |
|
"loss": 0.5205, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.774113952445043e-05, |
|
"loss": 0.5293, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.7729923732615524e-05, |
|
"loss": 0.5289, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.771870794078062e-05, |
|
"loss": 0.5492, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.7707492148945717e-05, |
|
"loss": 0.5246, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.7696276357110815e-05, |
|
"loss": 0.5331, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.768506056527591e-05, |
|
"loss": 0.5423, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.7673844773441007e-05, |
|
"loss": 0.533, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.7662628981606102e-05, |
|
"loss": 0.538, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.7651413189771197e-05, |
|
"loss": 0.5277, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.7640197397936295e-05, |
|
"loss": 0.5307, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.7628981606101393e-05, |
|
"loss": 0.5279, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.7617765814266488e-05, |
|
"loss": 0.5358, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.7606550022431586e-05, |
|
"loss": 0.5265, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.759533423059668e-05, |
|
"loss": 0.5285, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.758411843876178e-05, |
|
"loss": 0.5303, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.7572902646926876e-05, |
|
"loss": 0.5431, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.756168685509197e-05, |
|
"loss": 0.5312, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.7550471063257066e-05, |
|
"loss": 0.5296, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.7539255271422164e-05, |
|
"loss": 0.5282, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.752803947958726e-05, |
|
"loss": 0.5378, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.7516823687752356e-05, |
|
"loss": 0.5372, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.7505607895917455e-05, |
|
"loss": 0.5235, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.749439210408255e-05, |
|
"loss": 0.5319, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.7483176312247647e-05, |
|
"loss": 0.522, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.7471960520412742e-05, |
|
"loss": 0.5346, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.746074472857784e-05, |
|
"loss": 0.5313, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.7449528936742938e-05, |
|
"loss": 0.5459, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.7438313144908033e-05, |
|
"loss": 0.5291, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.7427097353073127e-05, |
|
"loss": 0.5398, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.7415881561238225e-05, |
|
"loss": 0.5225, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.740466576940332e-05, |
|
"loss": 0.5232, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.7393449977568418e-05, |
|
"loss": 0.5315, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.7382234185733516e-05, |
|
"loss": 0.5323, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.737101839389861e-05, |
|
"loss": 0.5278, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.7359802602063705e-05, |
|
"loss": 0.5367, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.7348586810228804e-05, |
|
"loss": 0.5203, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.73373710183939e-05, |
|
"loss": 0.5267, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.7326155226558996e-05, |
|
"loss": 0.5389, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.7314939434724094e-05, |
|
"loss": 0.5327, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.730372364288919e-05, |
|
"loss": 0.5294, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.7292507851054287e-05, |
|
"loss": 0.5287, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.728129205921938e-05, |
|
"loss": 0.5358, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.727007626738448e-05, |
|
"loss": 0.5345, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.7258860475549574e-05, |
|
"loss": 0.5265, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.7247644683714672e-05, |
|
"loss": 0.525, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.7236428891879767e-05, |
|
"loss": 0.5299, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.7225213100044865e-05, |
|
"loss": 0.5295, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.7213997308209963e-05, |
|
"loss": 0.5317, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.7202781516375058e-05, |
|
"loss": 0.5275, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.7191565724540153e-05, |
|
"loss": 0.5304, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.718034993270525e-05, |
|
"loss": 0.5288, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.716913414087035e-05, |
|
"loss": 0.5197, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.7157918349035443e-05, |
|
"loss": 0.5305, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.714670255720054e-05, |
|
"loss": 0.5205, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.7135486765365636e-05, |
|
"loss": 0.5306, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.712427097353073e-05, |
|
"loss": 0.5228, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.711305518169583e-05, |
|
"loss": 0.5241, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.7101839389860927e-05, |
|
"loss": 0.5302, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.709062359802602e-05, |
|
"loss": 0.5254, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.707940780619112e-05, |
|
"loss": 0.529, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.7068192014356214e-05, |
|
"loss": 0.5224, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.7056976222521312e-05, |
|
"loss": 0.5328, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.704576043068641e-05, |
|
"loss": 0.5248, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.7034544638851505e-05, |
|
"loss": 0.5255, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.70233288470166e-05, |
|
"loss": 0.5227, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 1.7012113055181698e-05, |
|
"loss": 0.5266, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 1.7000897263346792e-05, |
|
"loss": 0.5202, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 1.698968147151189e-05, |
|
"loss": 0.5281, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 1.697846567967699e-05, |
|
"loss": 0.5326, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 1.6967249887842083e-05, |
|
"loss": 0.5226, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 1.6956034096007178e-05, |
|
"loss": 0.5169, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 1.6944818304172276e-05, |
|
"loss": 0.5308, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 1.6933602512337374e-05, |
|
"loss": 0.5189, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 1.692238672050247e-05, |
|
"loss": 0.5162, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 1.6911170928667567e-05, |
|
"loss": 0.5273, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 1.689995513683266e-05, |
|
"loss": 0.5184, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 1.688873934499776e-05, |
|
"loss": 0.5327, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.6877523553162857e-05, |
|
"loss": 0.5266, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.6866307761327952e-05, |
|
"loss": 0.5298, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.6855091969493047e-05, |
|
"loss": 0.529, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.6843876177658145e-05, |
|
"loss": 0.5149, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.683266038582324e-05, |
|
"loss": 0.5194, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.6821444593988337e-05, |
|
"loss": 0.5307, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.6810228802153435e-05, |
|
"loss": 0.5248, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.679901301031853e-05, |
|
"loss": 0.5214, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.6787797218483625e-05, |
|
"loss": 0.5306, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.6776581426648723e-05, |
|
"loss": 0.5154, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.676536563481382e-05, |
|
"loss": 0.5237, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.6754149842978916e-05, |
|
"loss": 0.5081, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.6742934051144014e-05, |
|
"loss": 0.5244, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 1.6731718259309108e-05, |
|
"loss": 0.5235, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 1.6720502467474203e-05, |
|
"loss": 0.5234, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 1.67092866756393e-05, |
|
"loss": 0.5189, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 1.66980708838044e-05, |
|
"loss": 0.5247, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 1.6686855091969494e-05, |
|
"loss": 0.5196, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 1.6675639300134592e-05, |
|
"loss": 0.5155, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.6664423508299686e-05, |
|
"loss": 0.5308, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.6653207716464784e-05, |
|
"loss": 0.5132, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.6641991924629883e-05, |
|
"loss": 0.5259, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.6630776132794977e-05, |
|
"loss": 0.5314, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.6619560340960072e-05, |
|
"loss": 0.5243, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.660834454912517e-05, |
|
"loss": 0.5169, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.6597128757290265e-05, |
|
"loss": 0.5337, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.6585912965455363e-05, |
|
"loss": 0.5289, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.657469717362046e-05, |
|
"loss": 0.5277, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.6563481381785555e-05, |
|
"loss": 0.5248, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.655226558995065e-05, |
|
"loss": 0.5183, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.6541049798115748e-05, |
|
"loss": 0.5115, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 1.6529834006280846e-05, |
|
"loss": 0.5304, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 1.651861821444594e-05, |
|
"loss": 0.521, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 1.650740242261104e-05, |
|
"loss": 0.5214, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 1.6496186630776133e-05, |
|
"loss": 0.5096, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 1.648497083894123e-05, |
|
"loss": 0.5259, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 1.647375504710633e-05, |
|
"loss": 0.522, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.6462539255271424e-05, |
|
"loss": 0.5248, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.645132346343652e-05, |
|
"loss": 0.5199, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.6440107671601617e-05, |
|
"loss": 0.5139, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.642889187976671e-05, |
|
"loss": 0.5187, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.641767608793181e-05, |
|
"loss": 0.5287, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.6406460296096908e-05, |
|
"loss": 0.5186, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.6395244504262002e-05, |
|
"loss": 0.5163, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.6384028712427097e-05, |
|
"loss": 0.5234, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.6372812920592195e-05, |
|
"loss": 0.5194, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.6361597128757293e-05, |
|
"loss": 0.5202, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.6350381336922388e-05, |
|
"loss": 0.5079, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.6339165545087486e-05, |
|
"loss": 0.5285, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.632794975325258e-05, |
|
"loss": 0.5293, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 1.6316733961417675e-05, |
|
"loss": 0.5281, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 1.6305518169582773e-05, |
|
"loss": 0.5256, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 1.629430237774787e-05, |
|
"loss": 0.5185, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 1.6283086585912966e-05, |
|
"loss": 0.5191, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 1.6271870794078064e-05, |
|
"loss": 0.5282, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 1.626065500224316e-05, |
|
"loss": 0.5249, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.6249439210408257e-05, |
|
"loss": 0.5072, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.6238223418573355e-05, |
|
"loss": 0.5154, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.622700762673845e-05, |
|
"loss": 0.5154, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.6215791834903544e-05, |
|
"loss": 0.5181, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.6204576043068642e-05, |
|
"loss": 0.516, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.6193360251233737e-05, |
|
"loss": 0.519, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.6182144459398835e-05, |
|
"loss": 0.5253, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.6170928667563933e-05, |
|
"loss": 0.5235, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.6159712875729028e-05, |
|
"loss": 0.5187, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.6148497083894122e-05, |
|
"loss": 0.5098, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.613728129205922e-05, |
|
"loss": 0.5143, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.612606550022432e-05, |
|
"loss": 0.5114, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 1.6114849708389413e-05, |
|
"loss": 0.5152, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 1.610363391655451e-05, |
|
"loss": 0.508, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 1.6092418124719606e-05, |
|
"loss": 0.5062, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 1.6081202332884704e-05, |
|
"loss": 0.5186, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 1.6069986541049802e-05, |
|
"loss": 0.5312, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 1.6058770749214896e-05, |
|
"loss": 0.5238, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.604755495737999e-05, |
|
"loss": 0.5026, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.603633916554509e-05, |
|
"loss": 0.5239, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.6025123373710184e-05, |
|
"loss": 0.5185, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.6013907581875282e-05, |
|
"loss": 0.5162, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.600269179004038e-05, |
|
"loss": 0.5212, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.5991475998205475e-05, |
|
"loss": 0.5128, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.598026020637057e-05, |
|
"loss": 0.5188, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.5969044414535667e-05, |
|
"loss": 0.506, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.5957828622700765e-05, |
|
"loss": 0.5128, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.594661283086586e-05, |
|
"loss": 0.5244, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.5935397039030958e-05, |
|
"loss": 0.5078, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.5924181247196053e-05, |
|
"loss": 0.5319, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.5912965455361147e-05, |
|
"loss": 0.5186, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.5901749663526245e-05, |
|
"loss": 0.5107, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.5890533871691344e-05, |
|
"loss": 0.5131, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.5879318079856438e-05, |
|
"loss": 0.5136, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.5868102288021536e-05, |
|
"loss": 0.5059, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.585688649618663e-05, |
|
"loss": 0.5064, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.584567070435173e-05, |
|
"loss": 0.5063, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.5834454912516827e-05, |
|
"loss": 0.5301, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.582323912068192e-05, |
|
"loss": 0.5228, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.5812023328847016e-05, |
|
"loss": 0.5207, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.5800807537012114e-05, |
|
"loss": 0.5064, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.578959174517721e-05, |
|
"loss": 0.5148, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.5778375953342307e-05, |
|
"loss": 0.5181, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.5767160161507405e-05, |
|
"loss": 0.5093, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.57559443696725e-05, |
|
"loss": 0.5074, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.5744728577837594e-05, |
|
"loss": 0.5226, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.5733512786002693e-05, |
|
"loss": 0.5122, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.572229699416779e-05, |
|
"loss": 0.5197, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.5711081202332885e-05, |
|
"loss": 0.5143, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.5699865410497983e-05, |
|
"loss": 0.5138, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.5688649618663078e-05, |
|
"loss": 0.5326, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.5677433826828176e-05, |
|
"loss": 0.5139, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.5666218034993274e-05, |
|
"loss": 0.5208, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.565500224315837e-05, |
|
"loss": 0.5114, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.5643786451323463e-05, |
|
"loss": 0.5302, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.563257065948856e-05, |
|
"loss": 0.5207, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.5621354867653656e-05, |
|
"loss": 0.4986, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.5610139075818754e-05, |
|
"loss": 0.5175, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.5598923283983852e-05, |
|
"loss": 0.5168, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.5587707492148947e-05, |
|
"loss": 0.5169, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.557649170031404e-05, |
|
"loss": 0.5121, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.556527590847914e-05, |
|
"loss": 0.5094, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.5554060116644238e-05, |
|
"loss": 0.5169, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.5542844324809332e-05, |
|
"loss": 0.5004, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.553162853297443e-05, |
|
"loss": 0.5154, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.5520412741139525e-05, |
|
"loss": 0.5143, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.550919694930462e-05, |
|
"loss": 0.5234, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.5497981157469718e-05, |
|
"loss": 0.5072, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.5486765365634816e-05, |
|
"loss": 0.5196, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.547554957379991e-05, |
|
"loss": 0.5093, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.546433378196501e-05, |
|
"loss": 0.508, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.5453117990130103e-05, |
|
"loss": 0.5118, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.54419021982952e-05, |
|
"loss": 0.5169, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.54306864064603e-05, |
|
"loss": 0.5196, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.5419470614625394e-05, |
|
"loss": 0.5169, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.540825482279049e-05, |
|
"loss": 0.5043, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.5397039030955587e-05, |
|
"loss": 0.5099, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.5385823239120685e-05, |
|
"loss": 0.5285, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.537460744728578e-05, |
|
"loss": 0.5192, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.5363391655450877e-05, |
|
"loss": 0.521, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.5352175863615972e-05, |
|
"loss": 0.5209, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.5340960071781067e-05, |
|
"loss": 0.5177, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.5329744279946165e-05, |
|
"loss": 0.504, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.5318528488111263e-05, |
|
"loss": 0.5031, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.5307312696276357e-05, |
|
"loss": 0.5065, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.5296096904441456e-05, |
|
"loss": 0.5102, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.528488111260655e-05, |
|
"loss": 0.5086, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.5273665320771648e-05, |
|
"loss": 0.5045, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.5262449528936746e-05, |
|
"loss": 0.503, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.5251233737101841e-05, |
|
"loss": 0.5178, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.5240017945266936e-05, |
|
"loss": 0.5174, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.5228802153432034e-05, |
|
"loss": 0.5183, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.521758636159713e-05, |
|
"loss": 0.5003, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.5206370569762225e-05, |
|
"loss": 0.5163, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.5195154777927323e-05, |
|
"loss": 0.5069, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.5183938986092419e-05, |
|
"loss": 0.5132, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.5172723194257515e-05, |
|
"loss": 0.5199, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.5161507402422614e-05, |
|
"loss": 0.5211, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.5150291610587708e-05, |
|
"loss": 0.5, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.5139075818752805e-05, |
|
"loss": 0.5073, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.5127860026917903e-05, |
|
"loss": 0.5107, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.5116644235082997e-05, |
|
"loss": 0.5222, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.5105428443248094e-05, |
|
"loss": 0.5136, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.5094212651413192e-05, |
|
"loss": 0.5058, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.5082996859578288e-05, |
|
"loss": 0.514, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.5071781067743383e-05, |
|
"loss": 0.5073, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.506056527590848e-05, |
|
"loss": 0.5185, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.5049349484073577e-05, |
|
"loss": 0.5202, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.5038133692238672e-05, |
|
"loss": 0.5137, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.502691790040377e-05, |
|
"loss": 0.5035, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.5015702108568866e-05, |
|
"loss": 0.5166, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.500448631673396e-05, |
|
"loss": 0.5113, |
|
"step": 5010 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.4993270524899059e-05, |
|
"loss": 0.5061, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.4982054733064155e-05, |
|
"loss": 0.4983, |
|
"step": 5030 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.4970838941229252e-05, |
|
"loss": 0.5184, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.495962314939435e-05, |
|
"loss": 0.5146, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.4948407357559444e-05, |
|
"loss": 0.5032, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.493719156572454e-05, |
|
"loss": 0.5022, |
|
"step": 5070 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.4925975773889639e-05, |
|
"loss": 0.518, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.4914759982054733e-05, |
|
"loss": 0.5044, |
|
"step": 5090 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.490354419021983e-05, |
|
"loss": 0.5065, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.4892328398384928e-05, |
|
"loss": 0.506, |
|
"step": 5110 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.4881112606550024e-05, |
|
"loss": 0.5093, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.4869896814715119e-05, |
|
"loss": 0.5114, |
|
"step": 5130 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.4858681022880217e-05, |
|
"loss": 0.5172, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.4847465231045313e-05, |
|
"loss": 0.5236, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.4836249439210408e-05, |
|
"loss": 0.5068, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.4825033647375506e-05, |
|
"loss": 0.5078, |
|
"step": 5170 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.4813817855540602e-05, |
|
"loss": 0.5083, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.4802602063705697e-05, |
|
"loss": 0.51, |
|
"step": 5190 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.4791386271870795e-05, |
|
"loss": 0.5009, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.4780170480035891e-05, |
|
"loss": 0.5067, |
|
"step": 5210 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.4768954688200988e-05, |
|
"loss": 0.5092, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.4757738896366086e-05, |
|
"loss": 0.4984, |
|
"step": 5230 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.474652310453118e-05, |
|
"loss": 0.5114, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.4735307312696277e-05, |
|
"loss": 0.5106, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.4724091520861375e-05, |
|
"loss": 0.5154, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.471287572902647e-05, |
|
"loss": 0.5065, |
|
"step": 5270 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.4701659937191568e-05, |
|
"loss": 0.5049, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.4690444145356664e-05, |
|
"loss": 0.5126, |
|
"step": 5290 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.467922835352176e-05, |
|
"loss": 0.5086, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.4668012561686857e-05, |
|
"loss": 0.5081, |
|
"step": 5310 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.4656796769851953e-05, |
|
"loss": 0.5125, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.464558097801705e-05, |
|
"loss": 0.5097, |
|
"step": 5330 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.4634365186182147e-05, |
|
"loss": 0.498, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.4623149394347242e-05, |
|
"loss": 0.5011, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.4611933602512338e-05, |
|
"loss": 0.519, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.4600717810677436e-05, |
|
"loss": 0.4988, |
|
"step": 5370 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.4589502018842531e-05, |
|
"loss": 0.5005, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.4578286227007627e-05, |
|
"loss": 0.5051, |
|
"step": 5390 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.4567070435172726e-05, |
|
"loss": 0.495, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.4555854643337822e-05, |
|
"loss": 0.5107, |
|
"step": 5410 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.4544638851502917e-05, |
|
"loss": 0.5104, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.4533423059668015e-05, |
|
"loss": 0.5104, |
|
"step": 5430 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.4522207267833111e-05, |
|
"loss": 0.5118, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.4510991475998206e-05, |
|
"loss": 0.5085, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.4499775684163304e-05, |
|
"loss": 0.4994, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.44885598923284e-05, |
|
"loss": 0.4985, |
|
"step": 5470 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.4477344100493496e-05, |
|
"loss": 0.4968, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.4466128308658593e-05, |
|
"loss": 0.5028, |
|
"step": 5490 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.4454912516823689e-05, |
|
"loss": 0.5092, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.4443696724988785e-05, |
|
"loss": 0.5083, |
|
"step": 5510 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.4432480933153884e-05, |
|
"loss": 0.4946, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.4421265141318978e-05, |
|
"loss": 0.5098, |
|
"step": 5530 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.4410049349484075e-05, |
|
"loss": 0.5119, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.4398833557649173e-05, |
|
"loss": 0.499, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.4387617765814267e-05, |
|
"loss": 0.5077, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.4376401973979364e-05, |
|
"loss": 0.5167, |
|
"step": 5570 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.4365186182144462e-05, |
|
"loss": 0.5103, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.4353970390309558e-05, |
|
"loss": 0.5073, |
|
"step": 5590 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.4342754598474653e-05, |
|
"loss": 0.5183, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.433153880663975e-05, |
|
"loss": 0.4945, |
|
"step": 5610 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.4320323014804847e-05, |
|
"loss": 0.5049, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.4309107222969942e-05, |
|
"loss": 0.4994, |
|
"step": 5630 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.429789143113504e-05, |
|
"loss": 0.4904, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.4286675639300136e-05, |
|
"loss": 0.5215, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.4275459847465233e-05, |
|
"loss": 0.5026, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.4264244055630329e-05, |
|
"loss": 0.5094, |
|
"step": 5670 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.4253028263795425e-05, |
|
"loss": 0.5058, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.4241812471960522e-05, |
|
"loss": 0.5149, |
|
"step": 5690 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.423059668012562e-05, |
|
"loss": 0.5156, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.4219380888290714e-05, |
|
"loss": 0.5042, |
|
"step": 5710 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.420816509645581e-05, |
|
"loss": 0.4958, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.4196949304620909e-05, |
|
"loss": 0.5071, |
|
"step": 5730 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.4185733512786003e-05, |
|
"loss": 0.4999, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.41745177209511e-05, |
|
"loss": 0.5089, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.4163301929116198e-05, |
|
"loss": 0.4997, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.4152086137281294e-05, |
|
"loss": 0.5034, |
|
"step": 5770 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.4140870345446389e-05, |
|
"loss": 0.4994, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.4129654553611487e-05, |
|
"loss": 0.5106, |
|
"step": 5790 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.4118438761776583e-05, |
|
"loss": 0.5032, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.4107222969941678e-05, |
|
"loss": 0.5118, |
|
"step": 5810 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.4096007178106776e-05, |
|
"loss": 0.5173, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.4084791386271872e-05, |
|
"loss": 0.5107, |
|
"step": 5830 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.4073575594436969e-05, |
|
"loss": 0.5117, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.4062359802602065e-05, |
|
"loss": 0.5038, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.4051144010767161e-05, |
|
"loss": 0.5058, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.4039928218932258e-05, |
|
"loss": 0.5131, |
|
"step": 5870 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.4028712427097356e-05, |
|
"loss": 0.492, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.401749663526245e-05, |
|
"loss": 0.5011, |
|
"step": 5890 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.4006280843427547e-05, |
|
"loss": 0.5045, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.3995065051592645e-05, |
|
"loss": 0.5069, |
|
"step": 5910 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.398384925975774e-05, |
|
"loss": 0.4989, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.3972633467922836e-05, |
|
"loss": 0.4968, |
|
"step": 5930 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.3961417676087934e-05, |
|
"loss": 0.4994, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.395020188425303e-05, |
|
"loss": 0.4964, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.3938986092418125e-05, |
|
"loss": 0.5107, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.3927770300583223e-05, |
|
"loss": 0.503, |
|
"step": 5970 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.391655450874832e-05, |
|
"loss": 0.5122, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.3905338716913414e-05, |
|
"loss": 0.4971, |
|
"step": 5990 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.3894122925078512e-05, |
|
"loss": 0.5108, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.3882907133243608e-05, |
|
"loss": 0.4972, |
|
"step": 6010 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.3871691341408705e-05, |
|
"loss": 0.5065, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.3860475549573801e-05, |
|
"loss": 0.5199, |
|
"step": 6030 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.3849259757738897e-05, |
|
"loss": 0.498, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.3838043965903994e-05, |
|
"loss": 0.5128, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.3826828174069092e-05, |
|
"loss": 0.4961, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.3815612382234187e-05, |
|
"loss": 0.4963, |
|
"step": 6070 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.3804396590399283e-05, |
|
"loss": 0.5079, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.3793180798564381e-05, |
|
"loss": 0.5068, |
|
"step": 6090 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 1.3781965006729476e-05, |
|
"loss": 0.4963, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 1.3770749214894572e-05, |
|
"loss": 0.5092, |
|
"step": 6110 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 1.375953342305967e-05, |
|
"loss": 0.5008, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 1.3748317631224766e-05, |
|
"loss": 0.5106, |
|
"step": 6130 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 1.3737101839389861e-05, |
|
"loss": 0.4955, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 1.3725886047554959e-05, |
|
"loss": 0.4905, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 1.3714670255720055e-05, |
|
"loss": 0.4966, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 1.370345446388515e-05, |
|
"loss": 0.4885, |
|
"step": 6170 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 1.3692238672050248e-05, |
|
"loss": 0.5109, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 1.3681022880215345e-05, |
|
"loss": 0.5029, |
|
"step": 6190 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 1.3669807088380441e-05, |
|
"loss": 0.4944, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 1.3658591296545537e-05, |
|
"loss": 0.494, |
|
"step": 6210 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 1.3647375504710634e-05, |
|
"loss": 0.5075, |
|
"step": 6220 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 1.363615971287573e-05, |
|
"loss": 0.5079, |
|
"step": 6230 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 1.3624943921040828e-05, |
|
"loss": 0.495, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 1.3613728129205923e-05, |
|
"loss": 0.506, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 1.3602512337371019e-05, |
|
"loss": 0.4918, |
|
"step": 6260 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 1.3591296545536117e-05, |
|
"loss": 0.4981, |
|
"step": 6270 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 1.3580080753701212e-05, |
|
"loss": 0.4978, |
|
"step": 6280 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 1.3568864961866308e-05, |
|
"loss": 0.4942, |
|
"step": 6290 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 1.3557649170031406e-05, |
|
"loss": 0.5062, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 1.3546433378196503e-05, |
|
"loss": 0.5008, |
|
"step": 6310 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 1.3535217586361597e-05, |
|
"loss": 0.5065, |
|
"step": 6320 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 1.3524001794526695e-05, |
|
"loss": 0.5129, |
|
"step": 6330 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 1.3512786002691792e-05, |
|
"loss": 0.5002, |
|
"step": 6340 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 1.3501570210856886e-05, |
|
"loss": 0.5028, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 1.3490354419021984e-05, |
|
"loss": 0.4931, |
|
"step": 6360 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 1.347913862718708e-05, |
|
"loss": 0.4971, |
|
"step": 6370 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 1.3467922835352177e-05, |
|
"loss": 0.4965, |
|
"step": 6380 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 1.3456707043517273e-05, |
|
"loss": 0.5136, |
|
"step": 6390 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 1.344549125168237e-05, |
|
"loss": 0.4982, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 1.3434275459847466e-05, |
|
"loss": 0.4885, |
|
"step": 6410 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 1.3423059668012564e-05, |
|
"loss": 0.4949, |
|
"step": 6420 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 1.3411843876177659e-05, |
|
"loss": 0.5015, |
|
"step": 6430 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 1.3400628084342755e-05, |
|
"loss": 0.4977, |
|
"step": 6440 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 1.3389412292507853e-05, |
|
"loss": 0.5005, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 1.3378196500672948e-05, |
|
"loss": 0.4907, |
|
"step": 6460 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 1.3366980708838044e-05, |
|
"loss": 0.509, |
|
"step": 6470 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 1.3355764917003142e-05, |
|
"loss": 0.4904, |
|
"step": 6480 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 1.3344549125168239e-05, |
|
"loss": 0.488, |
|
"step": 6490 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 1.3333333333333333e-05, |
|
"loss": 0.4863, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 1.3322117541498431e-05, |
|
"loss": 0.5017, |
|
"step": 6510 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 1.3310901749663528e-05, |
|
"loss": 0.5029, |
|
"step": 6520 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.3299685957828622e-05, |
|
"loss": 0.507, |
|
"step": 6530 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.328847016599372e-05, |
|
"loss": 0.4966, |
|
"step": 6540 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.3277254374158817e-05, |
|
"loss": 0.5035, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.3266038582323913e-05, |
|
"loss": 0.4877, |
|
"step": 6560 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.325482279048901e-05, |
|
"loss": 0.4896, |
|
"step": 6570 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.3243606998654106e-05, |
|
"loss": 0.5011, |
|
"step": 6580 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.3232391206819202e-05, |
|
"loss": 0.5059, |
|
"step": 6590 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.32211754149843e-05, |
|
"loss": 0.5017, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.3209959623149395e-05, |
|
"loss": 0.4982, |
|
"step": 6610 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.3198743831314491e-05, |
|
"loss": 0.5009, |
|
"step": 6620 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.318752803947959e-05, |
|
"loss": 0.4909, |
|
"step": 6630 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.3176312247644684e-05, |
|
"loss": 0.5027, |
|
"step": 6640 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.316509645580978e-05, |
|
"loss": 0.4997, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.3153880663974878e-05, |
|
"loss": 0.5015, |
|
"step": 6660 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.3142664872139975e-05, |
|
"loss": 0.4974, |
|
"step": 6670 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.313144908030507e-05, |
|
"loss": 0.4913, |
|
"step": 6680 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.3120233288470167e-05, |
|
"loss": 0.4968, |
|
"step": 6690 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.3109017496635264e-05, |
|
"loss": 0.4932, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.3097801704800358e-05, |
|
"loss": 0.5002, |
|
"step": 6710 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.3086585912965457e-05, |
|
"loss": 0.5007, |
|
"step": 6720 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.3075370121130553e-05, |
|
"loss": 0.5003, |
|
"step": 6730 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.306415432929565e-05, |
|
"loss": 0.4896, |
|
"step": 6740 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.3052938537460747e-05, |
|
"loss": 0.5006, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.3041722745625842e-05, |
|
"loss": 0.4972, |
|
"step": 6760 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.3030506953790938e-05, |
|
"loss": 0.4971, |
|
"step": 6770 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.3019291161956036e-05, |
|
"loss": 0.4925, |
|
"step": 6780 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.3008075370121131e-05, |
|
"loss": 0.5005, |
|
"step": 6790 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.2996859578286227e-05, |
|
"loss": 0.4981, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.2985643786451325e-05, |
|
"loss": 0.5063, |
|
"step": 6810 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.297442799461642e-05, |
|
"loss": 0.4991, |
|
"step": 6820 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.2963212202781516e-05, |
|
"loss": 0.4969, |
|
"step": 6830 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.2951996410946615e-05, |
|
"loss": 0.491, |
|
"step": 6840 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.2940780619111711e-05, |
|
"loss": 0.5001, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.2929564827276806e-05, |
|
"loss": 0.501, |
|
"step": 6860 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.2918349035441904e-05, |
|
"loss": 0.5133, |
|
"step": 6870 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.2907133243607e-05, |
|
"loss": 0.4921, |
|
"step": 6880 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.2895917451772095e-05, |
|
"loss": 0.5047, |
|
"step": 6890 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.2884701659937193e-05, |
|
"loss": 0.4917, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.2873485868102289e-05, |
|
"loss": 0.5017, |
|
"step": 6910 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.2862270076267385e-05, |
|
"loss": 0.5029, |
|
"step": 6920 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.2851054284432483e-05, |
|
"loss": 0.4866, |
|
"step": 6930 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.2839838492597578e-05, |
|
"loss": 0.4978, |
|
"step": 6940 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.2828622700762674e-05, |
|
"loss": 0.4744, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.2817406908927773e-05, |
|
"loss": 0.486, |
|
"step": 6960 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.2806191117092867e-05, |
|
"loss": 0.4956, |
|
"step": 6970 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.2794975325257964e-05, |
|
"loss": 0.4935, |
|
"step": 6980 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.2783759533423062e-05, |
|
"loss": 0.5023, |
|
"step": 6990 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.2772543741588156e-05, |
|
"loss": 0.4937, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.2761327949753253e-05, |
|
"loss": 0.4873, |
|
"step": 7010 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.275011215791835e-05, |
|
"loss": 0.5001, |
|
"step": 7020 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.2738896366083447e-05, |
|
"loss": 0.4909, |
|
"step": 7030 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.2727680574248542e-05, |
|
"loss": 0.49, |
|
"step": 7040 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.271646478241364e-05, |
|
"loss": 0.501, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.2705248990578736e-05, |
|
"loss": 0.4924, |
|
"step": 7060 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.269403319874383e-05, |
|
"loss": 0.4927, |
|
"step": 7070 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.2682817406908929e-05, |
|
"loss": 0.5003, |
|
"step": 7080 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.2671601615074025e-05, |
|
"loss": 0.4894, |
|
"step": 7090 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.2660385823239122e-05, |
|
"loss": 0.4939, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.264917003140422e-05, |
|
"loss": 0.4937, |
|
"step": 7110 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.2637954239569314e-05, |
|
"loss": 0.4839, |
|
"step": 7120 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.262673844773441e-05, |
|
"loss": 0.5071, |
|
"step": 7130 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 1.2615522655899509e-05, |
|
"loss": 0.4893, |
|
"step": 7140 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 1.2604306864064603e-05, |
|
"loss": 0.4954, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 1.25930910722297e-05, |
|
"loss": 0.4876, |
|
"step": 7160 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 1.2581875280394798e-05, |
|
"loss": 0.5056, |
|
"step": 7170 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 1.2570659488559892e-05, |
|
"loss": 0.4937, |
|
"step": 7180 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 1.2559443696724989e-05, |
|
"loss": 0.4979, |
|
"step": 7190 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 1.2548227904890087e-05, |
|
"loss": 0.4967, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 1.2537012113055183e-05, |
|
"loss": 0.4812, |
|
"step": 7210 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 1.2525796321220278e-05, |
|
"loss": 0.4944, |
|
"step": 7220 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 1.2514580529385376e-05, |
|
"loss": 0.4977, |
|
"step": 7230 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 1.2503364737550472e-05, |
|
"loss": 0.4907, |
|
"step": 7240 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 1.2492148945715567e-05, |
|
"loss": 0.4902, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 1.2480933153880665e-05, |
|
"loss": 0.4908, |
|
"step": 7260 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 1.2469717362045761e-05, |
|
"loss": 0.5047, |
|
"step": 7270 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 1.2458501570210858e-05, |
|
"loss": 0.4909, |
|
"step": 7280 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 1.2447285778375956e-05, |
|
"loss": 0.4877, |
|
"step": 7290 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 1.243606998654105e-05, |
|
"loss": 0.4972, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 1.2424854194706147e-05, |
|
"loss": 0.4952, |
|
"step": 7310 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 1.2413638402871245e-05, |
|
"loss": 0.4881, |
|
"step": 7320 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 1.240242261103634e-05, |
|
"loss": 0.4944, |
|
"step": 7330 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 1.2391206819201436e-05, |
|
"loss": 0.4959, |
|
"step": 7340 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 1.2379991027366534e-05, |
|
"loss": 0.4963, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 1.2368775235531628e-05, |
|
"loss": 0.5035, |
|
"step": 7360 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 1.2357559443696725e-05, |
|
"loss": 0.5018, |
|
"step": 7370 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 1.2346343651861823e-05, |
|
"loss": 0.4973, |
|
"step": 7380 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 1.233512786002692e-05, |
|
"loss": 0.4998, |
|
"step": 7390 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 1.2323912068192014e-05, |
|
"loss": 0.5015, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 1.2312696276357112e-05, |
|
"loss": 0.4906, |
|
"step": 7410 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 1.2301480484522208e-05, |
|
"loss": 0.4865, |
|
"step": 7420 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 1.2290264692687303e-05, |
|
"loss": 0.4918, |
|
"step": 7430 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 1.2279048900852401e-05, |
|
"loss": 0.4913, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 1.2267833109017497e-05, |
|
"loss": 0.4915, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 1.2256617317182594e-05, |
|
"loss": 0.5034, |
|
"step": 7460 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 1.2245401525347692e-05, |
|
"loss": 0.478, |
|
"step": 7470 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 1.2234185733512786e-05, |
|
"loss": 0.4911, |
|
"step": 7480 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 1.2222969941677883e-05, |
|
"loss": 0.4918, |
|
"step": 7490 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 1.2211754149842981e-05, |
|
"loss": 0.4844, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 1.2200538358008076e-05, |
|
"loss": 0.488, |
|
"step": 7510 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 1.2189322566173172e-05, |
|
"loss": 0.4928, |
|
"step": 7520 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 1.217810677433827e-05, |
|
"loss": 0.4913, |
|
"step": 7530 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 1.2166890982503365e-05, |
|
"loss": 0.4841, |
|
"step": 7540 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 1.2155675190668461e-05, |
|
"loss": 0.4758, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 1.2144459398833559e-05, |
|
"loss": 0.4977, |
|
"step": 7560 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.2133243606998655e-05, |
|
"loss": 0.4991, |
|
"step": 7570 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.212202781516375e-05, |
|
"loss": 0.5023, |
|
"step": 7580 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.2110812023328848e-05, |
|
"loss": 0.4849, |
|
"step": 7590 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.2099596231493944e-05, |
|
"loss": 0.4867, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.2088380439659039e-05, |
|
"loss": 0.4962, |
|
"step": 7610 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.2077164647824137e-05, |
|
"loss": 0.4983, |
|
"step": 7620 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.2065948855989234e-05, |
|
"loss": 0.4894, |
|
"step": 7630 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.205473306415433e-05, |
|
"loss": 0.4896, |
|
"step": 7640 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.2043517272319428e-05, |
|
"loss": 0.4961, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.2032301480484523e-05, |
|
"loss": 0.4857, |
|
"step": 7660 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.2021085688649619e-05, |
|
"loss": 0.491, |
|
"step": 7670 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.2009869896814717e-05, |
|
"loss": 0.4829, |
|
"step": 7680 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.1998654104979812e-05, |
|
"loss": 0.4778, |
|
"step": 7690 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 1.1987438313144908e-05, |
|
"loss": 0.4795, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 1.1976222521310006e-05, |
|
"loss": 0.4923, |
|
"step": 7710 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 1.19650067294751e-05, |
|
"loss": 0.4903, |
|
"step": 7720 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 1.1953790937640197e-05, |
|
"loss": 0.5006, |
|
"step": 7730 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 1.1942575145805295e-05, |
|
"loss": 0.4916, |
|
"step": 7740 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 1.1931359353970392e-05, |
|
"loss": 0.496, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 1.192014356213549e-05, |
|
"loss": 0.4908, |
|
"step": 7760 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 1.1908927770300584e-05, |
|
"loss": 0.4964, |
|
"step": 7770 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 1.189771197846568e-05, |
|
"loss": 0.4954, |
|
"step": 7780 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 1.1886496186630779e-05, |
|
"loss": 0.4817, |
|
"step": 7790 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 1.1875280394795873e-05, |
|
"loss": 0.4917, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 1.186406460296097e-05, |
|
"loss": 0.4877, |
|
"step": 7810 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 1.1852848811126068e-05, |
|
"loss": 0.488, |
|
"step": 7820 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 1.1841633019291164e-05, |
|
"loss": 0.4982, |
|
"step": 7830 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 1.1830417227456259e-05, |
|
"loss": 0.4874, |
|
"step": 7840 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 1.1819201435621357e-05, |
|
"loss": 0.4925, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 1.1807985643786453e-05, |
|
"loss": 0.499, |
|
"step": 7860 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 1.1796769851951548e-05, |
|
"loss": 0.4916, |
|
"step": 7870 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 1.1785554060116646e-05, |
|
"loss": 0.4835, |
|
"step": 7880 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 1.1774338268281742e-05, |
|
"loss": 0.4978, |
|
"step": 7890 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 1.1763122476446837e-05, |
|
"loss": 0.4952, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 1.1751906684611935e-05, |
|
"loss": 0.4839, |
|
"step": 7910 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 1.1740690892777031e-05, |
|
"loss": 0.4841, |
|
"step": 7920 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 1.1729475100942128e-05, |
|
"loss": 0.4884, |
|
"step": 7930 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 1.1718259309107226e-05, |
|
"loss": 0.4948, |
|
"step": 7940 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 1.170704351727232e-05, |
|
"loss": 0.4888, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 1.1695827725437417e-05, |
|
"loss": 0.4922, |
|
"step": 7960 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 1.1684611933602515e-05, |
|
"loss": 0.4939, |
|
"step": 7970 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 1.167339614176761e-05, |
|
"loss": 0.4905, |
|
"step": 7980 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 1.1662180349932706e-05, |
|
"loss": 0.4913, |
|
"step": 7990 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 1.1650964558097804e-05, |
|
"loss": 0.4926, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 1.16397487662629e-05, |
|
"loss": 0.496, |
|
"step": 8010 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 1.1628532974427995e-05, |
|
"loss": 0.4878, |
|
"step": 8020 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 1.1617317182593093e-05, |
|
"loss": 0.4944, |
|
"step": 8030 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 1.160610139075819e-05, |
|
"loss": 0.4954, |
|
"step": 8040 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 1.1594885598923284e-05, |
|
"loss": 0.502, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 1.1583669807088382e-05, |
|
"loss": 0.4789, |
|
"step": 8060 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 1.1572454015253478e-05, |
|
"loss": 0.4862, |
|
"step": 8070 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 1.1561238223418573e-05, |
|
"loss": 0.5012, |
|
"step": 8080 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 1.1550022431583671e-05, |
|
"loss": 0.4831, |
|
"step": 8090 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 1.1538806639748767e-05, |
|
"loss": 0.4823, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 1.1527590847913864e-05, |
|
"loss": 0.4948, |
|
"step": 8110 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 1.1516375056078962e-05, |
|
"loss": 0.4874, |
|
"step": 8120 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 1.1505159264244056e-05, |
|
"loss": 0.4847, |
|
"step": 8130 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 1.1493943472409153e-05, |
|
"loss": 0.4985, |
|
"step": 8140 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 1.1482727680574251e-05, |
|
"loss": 0.4889, |
|
"step": 8150 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 1.1471511888739346e-05, |
|
"loss": 0.493, |
|
"step": 8160 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 1.1460296096904442e-05, |
|
"loss": 0.4945, |
|
"step": 8170 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 1.144908030506954e-05, |
|
"loss": 0.4813, |
|
"step": 8180 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 1.1437864513234636e-05, |
|
"loss": 0.487, |
|
"step": 8190 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 1.1426648721399731e-05, |
|
"loss": 0.4875, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 1.1415432929564829e-05, |
|
"loss": 0.4893, |
|
"step": 8210 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 1.1404217137729925e-05, |
|
"loss": 0.4984, |
|
"step": 8220 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 1.139300134589502e-05, |
|
"loss": 0.5001, |
|
"step": 8230 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 1.1381785554060118e-05, |
|
"loss": 0.4866, |
|
"step": 8240 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 1.1370569762225214e-05, |
|
"loss": 0.4904, |
|
"step": 8250 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 1.135935397039031e-05, |
|
"loss": 0.4808, |
|
"step": 8260 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 1.1348138178555407e-05, |
|
"loss": 0.4902, |
|
"step": 8270 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 1.1336922386720504e-05, |
|
"loss": 0.4877, |
|
"step": 8280 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 1.13257065948856e-05, |
|
"loss": 0.4904, |
|
"step": 8290 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 1.1314490803050698e-05, |
|
"loss": 0.4923, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 1.1303275011215793e-05, |
|
"loss": 0.4984, |
|
"step": 8310 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 1.1292059219380889e-05, |
|
"loss": 0.487, |
|
"step": 8320 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 1.1280843427545987e-05, |
|
"loss": 0.4958, |
|
"step": 8330 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 1.1269627635711082e-05, |
|
"loss": 0.4884, |
|
"step": 8340 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 1.1258411843876178e-05, |
|
"loss": 0.4885, |
|
"step": 8350 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 1.1247196052041276e-05, |
|
"loss": 0.4871, |
|
"step": 8360 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 1.1235980260206372e-05, |
|
"loss": 0.4961, |
|
"step": 8370 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 1.1224764468371467e-05, |
|
"loss": 0.4842, |
|
"step": 8380 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 1.1213548676536565e-05, |
|
"loss": 0.4891, |
|
"step": 8390 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 1.1202332884701661e-05, |
|
"loss": 0.4836, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 1.1191117092866756e-05, |
|
"loss": 0.4893, |
|
"step": 8410 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 1.1179901301031854e-05, |
|
"loss": 0.4785, |
|
"step": 8420 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 1.116868550919695e-05, |
|
"loss": 0.4926, |
|
"step": 8430 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 1.1157469717362047e-05, |
|
"loss": 0.4804, |
|
"step": 8440 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 1.1146253925527143e-05, |
|
"loss": 0.4888, |
|
"step": 8450 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 1.113503813369224e-05, |
|
"loss": 0.5088, |
|
"step": 8460 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 1.1123822341857336e-05, |
|
"loss": 0.4904, |
|
"step": 8470 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 1.1112606550022434e-05, |
|
"loss": 0.4913, |
|
"step": 8480 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 1.1101390758187529e-05, |
|
"loss": 0.4901, |
|
"step": 8490 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 1.1090174966352625e-05, |
|
"loss": 0.4887, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 1.1078959174517723e-05, |
|
"loss": 0.4816, |
|
"step": 8510 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 1.1067743382682818e-05, |
|
"loss": 0.4989, |
|
"step": 8520 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 1.1056527590847914e-05, |
|
"loss": 0.4807, |
|
"step": 8530 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 1.1045311799013012e-05, |
|
"loss": 0.4929, |
|
"step": 8540 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 1.1034096007178109e-05, |
|
"loss": 0.4752, |
|
"step": 8550 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 1.1022880215343203e-05, |
|
"loss": 0.4822, |
|
"step": 8560 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 1.1011664423508301e-05, |
|
"loss": 0.491, |
|
"step": 8570 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 1.1000448631673398e-05, |
|
"loss": 0.4987, |
|
"step": 8580 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 1.0989232839838492e-05, |
|
"loss": 0.4867, |
|
"step": 8590 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 1.097801704800359e-05, |
|
"loss": 0.493, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 1.0966801256168687e-05, |
|
"loss": 0.5023, |
|
"step": 8610 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 1.0955585464333783e-05, |
|
"loss": 0.4931, |
|
"step": 8620 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 1.094436967249888e-05, |
|
"loss": 0.4875, |
|
"step": 8630 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 1.0933153880663976e-05, |
|
"loss": 0.4818, |
|
"step": 8640 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 1.0921938088829072e-05, |
|
"loss": 0.4828, |
|
"step": 8650 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 1.091072229699417e-05, |
|
"loss": 0.4897, |
|
"step": 8660 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 1.0899506505159265e-05, |
|
"loss": 0.48, |
|
"step": 8670 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 1.0888290713324361e-05, |
|
"loss": 0.4865, |
|
"step": 8680 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 1.087707492148946e-05, |
|
"loss": 0.4941, |
|
"step": 8690 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 1.0865859129654554e-05, |
|
"loss": 0.4858, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 1.085464333781965e-05, |
|
"loss": 0.4864, |
|
"step": 8710 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 1.0843427545984748e-05, |
|
"loss": 0.4864, |
|
"step": 8720 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 1.0832211754149845e-05, |
|
"loss": 0.5011, |
|
"step": 8730 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 1.082099596231494e-05, |
|
"loss": 0.4841, |
|
"step": 8740 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 1.0809780170480037e-05, |
|
"loss": 0.4966, |
|
"step": 8750 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 1.0798564378645134e-05, |
|
"loss": 0.5004, |
|
"step": 8760 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 1.0787348586810228e-05, |
|
"loss": 0.4976, |
|
"step": 8770 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 1.0776132794975326e-05, |
|
"loss": 0.481, |
|
"step": 8780 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 1.0764917003140423e-05, |
|
"loss": 0.4866, |
|
"step": 8790 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 1.075370121130552e-05, |
|
"loss": 0.496, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 1.0742485419470616e-05, |
|
"loss": 0.492, |
|
"step": 8810 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 1.0731269627635712e-05, |
|
"loss": 0.4819, |
|
"step": 8820 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 1.0720053835800808e-05, |
|
"loss": 0.492, |
|
"step": 8830 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 1.0708838043965906e-05, |
|
"loss": 0.4818, |
|
"step": 8840 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 1.0697622252131001e-05, |
|
"loss": 0.4845, |
|
"step": 8850 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 1.0686406460296097e-05, |
|
"loss": 0.5005, |
|
"step": 8860 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 1.0675190668461195e-05, |
|
"loss": 0.4761, |
|
"step": 8870 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 1.066397487662629e-05, |
|
"loss": 0.4906, |
|
"step": 8880 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 1.0652759084791386e-05, |
|
"loss": 0.4873, |
|
"step": 8890 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 1.0641543292956484e-05, |
|
"loss": 0.4991, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 1.063032750112158e-05, |
|
"loss": 0.491, |
|
"step": 8910 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 1.0619111709286675e-05, |
|
"loss": 0.4937, |
|
"step": 8920 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 1.0607895917451774e-05, |
|
"loss": 0.482, |
|
"step": 8930 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 1.059668012561687e-05, |
|
"loss": 0.4956, |
|
"step": 8940 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 1.0585464333781965e-05, |
|
"loss": 0.4852, |
|
"step": 8950 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 1.0574248541947063e-05, |
|
"loss": 0.4891, |
|
"step": 8960 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 1.0563032750112159e-05, |
|
"loss": 0.4873, |
|
"step": 8970 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 1.0551816958277255e-05, |
|
"loss": 0.4929, |
|
"step": 8980 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 1.0540601166442352e-05, |
|
"loss": 0.4868, |
|
"step": 8990 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 1.0529385374607448e-05, |
|
"loss": 0.4931, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 1.0518169582772544e-05, |
|
"loss": 0.4922, |
|
"step": 9010 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 1.0506953790937642e-05, |
|
"loss": 0.4791, |
|
"step": 9020 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 1.0495737999102737e-05, |
|
"loss": 0.4915, |
|
"step": 9030 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 1.0484522207267833e-05, |
|
"loss": 0.4853, |
|
"step": 9040 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 1.0473306415432931e-05, |
|
"loss": 0.4826, |
|
"step": 9050 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 1.0462090623598026e-05, |
|
"loss": 0.4948, |
|
"step": 9060 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 1.0450874831763123e-05, |
|
"loss": 0.4932, |
|
"step": 9070 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 1.043965903992822e-05, |
|
"loss": 0.4855, |
|
"step": 9080 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 1.0428443248093317e-05, |
|
"loss": 0.4912, |
|
"step": 9090 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 1.0417227456258412e-05, |
|
"loss": 0.4895, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 1.040601166442351e-05, |
|
"loss": 0.4907, |
|
"step": 9110 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 1.0394795872588606e-05, |
|
"loss": 0.486, |
|
"step": 9120 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 1.03835800807537e-05, |
|
"loss": 0.4858, |
|
"step": 9130 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 1.0372364288918799e-05, |
|
"loss": 0.4949, |
|
"step": 9140 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 1.0361148497083895e-05, |
|
"loss": 0.4826, |
|
"step": 9150 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 1.0349932705248991e-05, |
|
"loss": 0.4847, |
|
"step": 9160 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 1.0338716913414088e-05, |
|
"loss": 0.4926, |
|
"step": 9170 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 1.0327501121579184e-05, |
|
"loss": 0.4815, |
|
"step": 9180 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 1.031628532974428e-05, |
|
"loss": 0.489, |
|
"step": 9190 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 1.0305069537909379e-05, |
|
"loss": 0.4925, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 1.0293853746074473e-05, |
|
"loss": 0.4765, |
|
"step": 9210 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 1.028263795423957e-05, |
|
"loss": 0.4857, |
|
"step": 9220 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 1.0271422162404668e-05, |
|
"loss": 0.4905, |
|
"step": 9230 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 1.0260206370569762e-05, |
|
"loss": 0.4887, |
|
"step": 9240 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 1.0248990578734859e-05, |
|
"loss": 0.4823, |
|
"step": 9250 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 1.0237774786899957e-05, |
|
"loss": 0.4873, |
|
"step": 9260 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 1.0226558995065053e-05, |
|
"loss": 0.4931, |
|
"step": 9270 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 1.0215343203230148e-05, |
|
"loss": 0.4917, |
|
"step": 9280 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 1.0204127411395246e-05, |
|
"loss": 0.4895, |
|
"step": 9290 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 1.0192911619560342e-05, |
|
"loss": 0.4763, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 1.0181695827725437e-05, |
|
"loss": 0.4917, |
|
"step": 9310 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 1.0170480035890535e-05, |
|
"loss": 0.4755, |
|
"step": 9320 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 1.0159264244055631e-05, |
|
"loss": 0.503, |
|
"step": 9330 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 1.0148048452220728e-05, |
|
"loss": 0.4761, |
|
"step": 9340 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 1.0136832660385824e-05, |
|
"loss": 0.479, |
|
"step": 9350 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 1.012561686855092e-05, |
|
"loss": 0.4863, |
|
"step": 9360 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 1.0114401076716017e-05, |
|
"loss": 0.491, |
|
"step": 9370 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 1.0103185284881115e-05, |
|
"loss": 0.4828, |
|
"step": 9380 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 1.009196949304621e-05, |
|
"loss": 0.4839, |
|
"step": 9390 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 1.0080753701211306e-05, |
|
"loss": 0.4931, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 1.0069537909376404e-05, |
|
"loss": 0.4743, |
|
"step": 9410 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 1.0058322117541498e-05, |
|
"loss": 0.4825, |
|
"step": 9420 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 1.0047106325706595e-05, |
|
"loss": 0.4883, |
|
"step": 9430 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 1.0035890533871693e-05, |
|
"loss": 0.4772, |
|
"step": 9440 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 1.002467474203679e-05, |
|
"loss": 0.4868, |
|
"step": 9450 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 1.0013458950201884e-05, |
|
"loss": 0.4832, |
|
"step": 9460 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 1.0002243158366982e-05, |
|
"loss": 0.4858, |
|
"step": 9470 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 9.991027366532078e-06, |
|
"loss": 0.4846, |
|
"step": 9480 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 9.979811574697175e-06, |
|
"loss": 0.486, |
|
"step": 9490 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 9.968595782862271e-06, |
|
"loss": 0.487, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 9.957379991027367e-06, |
|
"loss": 0.4783, |
|
"step": 9510 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 9.946164199192464e-06, |
|
"loss": 0.4837, |
|
"step": 9520 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 9.93494840735756e-06, |
|
"loss": 0.4794, |
|
"step": 9530 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 9.923732615522656e-06, |
|
"loss": 0.4787, |
|
"step": 9540 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 9.912516823687753e-06, |
|
"loss": 0.4906, |
|
"step": 9550 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 9.901301031852849e-06, |
|
"loss": 0.4844, |
|
"step": 9560 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 9.890085240017945e-06, |
|
"loss": 0.4828, |
|
"step": 9570 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 9.878869448183044e-06, |
|
"loss": 0.4763, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 9.867653656348138e-06, |
|
"loss": 0.4879, |
|
"step": 9590 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 9.856437864513235e-06, |
|
"loss": 0.4978, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 9.845222072678333e-06, |
|
"loss": 0.4805, |
|
"step": 9610 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 9.834006280843429e-06, |
|
"loss": 0.4924, |
|
"step": 9620 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 9.822790489008525e-06, |
|
"loss": 0.4909, |
|
"step": 9630 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 9.811574697173622e-06, |
|
"loss": 0.4831, |
|
"step": 9640 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 9.800358905338718e-06, |
|
"loss": 0.485, |
|
"step": 9650 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 9.789143113503814e-06, |
|
"loss": 0.4802, |
|
"step": 9660 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 9.77792732166891e-06, |
|
"loss": 0.4951, |
|
"step": 9670 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 9.766711529834007e-06, |
|
"loss": 0.4784, |
|
"step": 9680 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 9.755495737999103e-06, |
|
"loss": 0.4788, |
|
"step": 9690 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 9.7442799461642e-06, |
|
"loss": 0.4798, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 9.733064154329296e-06, |
|
"loss": 0.4858, |
|
"step": 9710 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 9.721848362494393e-06, |
|
"loss": 0.4845, |
|
"step": 9720 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 9.710632570659489e-06, |
|
"loss": 0.4871, |
|
"step": 9730 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 9.699416778824587e-06, |
|
"loss": 0.4809, |
|
"step": 9740 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 9.688200986989682e-06, |
|
"loss": 0.4902, |
|
"step": 9750 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 9.67698519515478e-06, |
|
"loss": 0.4961, |
|
"step": 9760 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 9.665769403319876e-06, |
|
"loss": 0.4749, |
|
"step": 9770 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 9.65455361148497e-06, |
|
"loss": 0.4922, |
|
"step": 9780 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 9.643337819650069e-06, |
|
"loss": 0.4824, |
|
"step": 9790 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 9.632122027815165e-06, |
|
"loss": 0.4853, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 9.620906235980261e-06, |
|
"loss": 0.4808, |
|
"step": 9810 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 9.609690444145358e-06, |
|
"loss": 0.4878, |
|
"step": 9820 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 9.598474652310454e-06, |
|
"loss": 0.4847, |
|
"step": 9830 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 9.58725886047555e-06, |
|
"loss": 0.4722, |
|
"step": 9840 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 9.576043068640647e-06, |
|
"loss": 0.492, |
|
"step": 9850 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 9.564827276805743e-06, |
|
"loss": 0.4868, |
|
"step": 9860 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 9.55361148497084e-06, |
|
"loss": 0.4877, |
|
"step": 9870 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 9.542395693135936e-06, |
|
"loss": 0.4925, |
|
"step": 9880 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 9.531179901301032e-06, |
|
"loss": 0.487, |
|
"step": 9890 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 9.519964109466129e-06, |
|
"loss": 0.477, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 9.508748317631225e-06, |
|
"loss": 0.4781, |
|
"step": 9910 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 9.497532525796323e-06, |
|
"loss": 0.4943, |
|
"step": 9920 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 9.486316733961418e-06, |
|
"loss": 0.4877, |
|
"step": 9930 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 9.475100942126516e-06, |
|
"loss": 0.4861, |
|
"step": 9940 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 9.463885150291612e-06, |
|
"loss": 0.4761, |
|
"step": 9950 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 9.452669358456707e-06, |
|
"loss": 0.4803, |
|
"step": 9960 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 9.441453566621805e-06, |
|
"loss": 0.4799, |
|
"step": 9970 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 9.430237774786901e-06, |
|
"loss": 0.4706, |
|
"step": 9980 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 9.419021982951998e-06, |
|
"loss": 0.4929, |
|
"step": 9990 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 9.407806191117094e-06, |
|
"loss": 0.4891, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 9.39659039928219e-06, |
|
"loss": 0.4708, |
|
"step": 10010 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 9.385374607447287e-06, |
|
"loss": 0.4793, |
|
"step": 10020 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 9.374158815612383e-06, |
|
"loss": 0.4819, |
|
"step": 10030 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 9.36294302377748e-06, |
|
"loss": 0.4805, |
|
"step": 10040 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 9.351727231942576e-06, |
|
"loss": 0.4882, |
|
"step": 10050 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 9.340511440107672e-06, |
|
"loss": 0.4785, |
|
"step": 10060 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 9.329295648272768e-06, |
|
"loss": 0.4856, |
|
"step": 10070 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 9.318079856437865e-06, |
|
"loss": 0.4885, |
|
"step": 10080 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 9.306864064602961e-06, |
|
"loss": 0.4737, |
|
"step": 10090 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 9.29564827276806e-06, |
|
"loss": 0.4882, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 9.284432480933154e-06, |
|
"loss": 0.4874, |
|
"step": 10110 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 9.273216689098252e-06, |
|
"loss": 0.4861, |
|
"step": 10120 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 9.262000897263348e-06, |
|
"loss": 0.4882, |
|
"step": 10130 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 9.250785105428443e-06, |
|
"loss": 0.4722, |
|
"step": 10140 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 9.239569313593541e-06, |
|
"loss": 0.4803, |
|
"step": 10150 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 9.228353521758637e-06, |
|
"loss": 0.467, |
|
"step": 10160 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 9.217137729923734e-06, |
|
"loss": 0.4861, |
|
"step": 10170 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 9.20592193808883e-06, |
|
"loss": 0.4878, |
|
"step": 10180 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 9.194706146253926e-06, |
|
"loss": 0.4742, |
|
"step": 10190 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 9.183490354419023e-06, |
|
"loss": 0.4746, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 9.172274562584119e-06, |
|
"loss": 0.4853, |
|
"step": 10210 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 9.161058770749215e-06, |
|
"loss": 0.4875, |
|
"step": 10220 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 9.149842978914312e-06, |
|
"loss": 0.4843, |
|
"step": 10230 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 9.138627187079408e-06, |
|
"loss": 0.4984, |
|
"step": 10240 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 9.127411395244505e-06, |
|
"loss": 0.485, |
|
"step": 10250 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 9.116195603409601e-06, |
|
"loss": 0.4859, |
|
"step": 10260 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 9.104979811574697e-06, |
|
"loss": 0.4807, |
|
"step": 10270 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 9.093764019739795e-06, |
|
"loss": 0.4842, |
|
"step": 10280 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 9.08254822790489e-06, |
|
"loss": 0.4626, |
|
"step": 10290 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 9.071332436069988e-06, |
|
"loss": 0.4593, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 9.060116644235084e-06, |
|
"loss": 0.4902, |
|
"step": 10310 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 9.048900852400179e-06, |
|
"loss": 0.4693, |
|
"step": 10320 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 9.037685060565277e-06, |
|
"loss": 0.4949, |
|
"step": 10330 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 9.026469268730373e-06, |
|
"loss": 0.4833, |
|
"step": 10340 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 9.01525347689547e-06, |
|
"loss": 0.4874, |
|
"step": 10350 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 9.004037685060566e-06, |
|
"loss": 0.4766, |
|
"step": 10360 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 8.992821893225663e-06, |
|
"loss": 0.4739, |
|
"step": 10370 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 8.981606101390759e-06, |
|
"loss": 0.4683, |
|
"step": 10380 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 8.970390309555855e-06, |
|
"loss": 0.4887, |
|
"step": 10390 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 8.959174517720952e-06, |
|
"loss": 0.4944, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 8.947958725886048e-06, |
|
"loss": 0.4843, |
|
"step": 10410 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 8.936742934051144e-06, |
|
"loss": 0.4733, |
|
"step": 10420 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 8.925527142216242e-06, |
|
"loss": 0.476, |
|
"step": 10430 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 8.914311350381337e-06, |
|
"loss": 0.4806, |
|
"step": 10440 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 8.903095558546433e-06, |
|
"loss": 0.49, |
|
"step": 10450 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 8.891879766711531e-06, |
|
"loss": 0.472, |
|
"step": 10460 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 8.880663974876626e-06, |
|
"loss": 0.4827, |
|
"step": 10470 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 8.869448183041724e-06, |
|
"loss": 0.4752, |
|
"step": 10480 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 8.85823239120682e-06, |
|
"loss": 0.4922, |
|
"step": 10490 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 8.847016599371915e-06, |
|
"loss": 0.487, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 8.835800807537013e-06, |
|
"loss": 0.4892, |
|
"step": 10510 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 8.82458501570211e-06, |
|
"loss": 0.4788, |
|
"step": 10520 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 8.813369223867206e-06, |
|
"loss": 0.473, |
|
"step": 10530 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 8.802153432032302e-06, |
|
"loss": 0.4749, |
|
"step": 10540 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 8.790937640197399e-06, |
|
"loss": 0.4792, |
|
"step": 10550 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 8.779721848362495e-06, |
|
"loss": 0.4705, |
|
"step": 10560 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 8.768506056527591e-06, |
|
"loss": 0.4854, |
|
"step": 10570 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 8.757290264692688e-06, |
|
"loss": 0.49, |
|
"step": 10580 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 8.746074472857784e-06, |
|
"loss": 0.4821, |
|
"step": 10590 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 8.73485868102288e-06, |
|
"loss": 0.4762, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 8.723642889187978e-06, |
|
"loss": 0.4871, |
|
"step": 10610 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 8.712427097353073e-06, |
|
"loss": 0.4743, |
|
"step": 10620 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 8.70121130551817e-06, |
|
"loss": 0.4749, |
|
"step": 10630 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 8.689995513683268e-06, |
|
"loss": 0.4855, |
|
"step": 10640 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 8.678779721848362e-06, |
|
"loss": 0.4769, |
|
"step": 10650 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 8.66756393001346e-06, |
|
"loss": 0.478, |
|
"step": 10660 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 8.656348138178557e-06, |
|
"loss": 0.4818, |
|
"step": 10670 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 8.645132346343651e-06, |
|
"loss": 0.469, |
|
"step": 10680 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 8.63391655450875e-06, |
|
"loss": 0.4792, |
|
"step": 10690 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 8.622700762673846e-06, |
|
"loss": 0.4864, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 8.611484970838942e-06, |
|
"loss": 0.4849, |
|
"step": 10710 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 8.600269179004038e-06, |
|
"loss": 0.4867, |
|
"step": 10720 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 8.589053387169135e-06, |
|
"loss": 0.4735, |
|
"step": 10730 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 8.577837595334231e-06, |
|
"loss": 0.49, |
|
"step": 10740 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 8.566621803499327e-06, |
|
"loss": 0.4854, |
|
"step": 10750 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 8.555406011664424e-06, |
|
"loss": 0.4781, |
|
"step": 10760 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 8.54419021982952e-06, |
|
"loss": 0.4723, |
|
"step": 10770 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 8.532974427994617e-06, |
|
"loss": 0.4835, |
|
"step": 10780 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 8.521758636159715e-06, |
|
"loss": 0.4752, |
|
"step": 10790 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 8.51054284432481e-06, |
|
"loss": 0.4745, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 8.499327052489906e-06, |
|
"loss": 0.4869, |
|
"step": 10810 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 8.488111260655004e-06, |
|
"loss": 0.4875, |
|
"step": 10820 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 8.476895468820098e-06, |
|
"loss": 0.4837, |
|
"step": 10830 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 8.465679676985196e-06, |
|
"loss": 0.477, |
|
"step": 10840 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 8.454463885150293e-06, |
|
"loss": 0.4963, |
|
"step": 10850 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 8.443248093315389e-06, |
|
"loss": 0.486, |
|
"step": 10860 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 8.432032301480485e-06, |
|
"loss": 0.4843, |
|
"step": 10870 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 8.420816509645582e-06, |
|
"loss": 0.4866, |
|
"step": 10880 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 8.409600717810678e-06, |
|
"loss": 0.4873, |
|
"step": 10890 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 8.398384925975775e-06, |
|
"loss": 0.4797, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 8.387169134140871e-06, |
|
"loss": 0.4702, |
|
"step": 10910 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 8.375953342305967e-06, |
|
"loss": 0.4971, |
|
"step": 10920 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 8.364737550471064e-06, |
|
"loss": 0.4778, |
|
"step": 10930 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 8.35352175863616e-06, |
|
"loss": 0.4836, |
|
"step": 10940 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 8.342305966801258e-06, |
|
"loss": 0.4871, |
|
"step": 10950 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 8.331090174966353e-06, |
|
"loss": 0.4664, |
|
"step": 10960 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 8.31987438313145e-06, |
|
"loss": 0.4833, |
|
"step": 10970 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 8.308658591296547e-06, |
|
"loss": 0.4874, |
|
"step": 10980 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 8.297442799461642e-06, |
|
"loss": 0.4779, |
|
"step": 10990 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 8.28622700762674e-06, |
|
"loss": 0.4789, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 8.275011215791836e-06, |
|
"loss": 0.4619, |
|
"step": 11010 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 8.263795423956933e-06, |
|
"loss": 0.4737, |
|
"step": 11020 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 8.252579632122029e-06, |
|
"loss": 0.4787, |
|
"step": 11030 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 8.241363840287125e-06, |
|
"loss": 0.4809, |
|
"step": 11040 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 8.230148048452222e-06, |
|
"loss": 0.479, |
|
"step": 11050 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 8.218932256617318e-06, |
|
"loss": 0.4817, |
|
"step": 11060 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 8.207716464782414e-06, |
|
"loss": 0.4586, |
|
"step": 11070 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 8.19650067294751e-06, |
|
"loss": 0.488, |
|
"step": 11080 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 8.185284881112607e-06, |
|
"loss": 0.4714, |
|
"step": 11090 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 8.174069089277703e-06, |
|
"loss": 0.4828, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 8.1628532974428e-06, |
|
"loss": 0.4935, |
|
"step": 11110 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 8.151637505607896e-06, |
|
"loss": 0.4799, |
|
"step": 11120 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 8.140421713772994e-06, |
|
"loss": 0.483, |
|
"step": 11130 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 8.129205921938089e-06, |
|
"loss": 0.4798, |
|
"step": 11140 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 8.117990130103187e-06, |
|
"loss": 0.4727, |
|
"step": 11150 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 8.106774338268283e-06, |
|
"loss": 0.4746, |
|
"step": 11160 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 8.095558546433378e-06, |
|
"loss": 0.4844, |
|
"step": 11170 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 8.084342754598476e-06, |
|
"loss": 0.4718, |
|
"step": 11180 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 8.073126962763572e-06, |
|
"loss": 0.4758, |
|
"step": 11190 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 8.061911170928669e-06, |
|
"loss": 0.4762, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 8.050695379093765e-06, |
|
"loss": 0.4765, |
|
"step": 11210 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 8.039479587258861e-06, |
|
"loss": 0.4748, |
|
"step": 11220 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 8.028263795423958e-06, |
|
"loss": 0.481, |
|
"step": 11230 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 8.017048003589054e-06, |
|
"loss": 0.4715, |
|
"step": 11240 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 8.00583221175415e-06, |
|
"loss": 0.4802, |
|
"step": 11250 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 7.994616419919247e-06, |
|
"loss": 0.4812, |
|
"step": 11260 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 7.983400628084343e-06, |
|
"loss": 0.4891, |
|
"step": 11270 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 7.97218483624944e-06, |
|
"loss": 0.4914, |
|
"step": 11280 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 7.960969044414536e-06, |
|
"loss": 0.4749, |
|
"step": 11290 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 7.949753252579632e-06, |
|
"loss": 0.4697, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 7.93853746074473e-06, |
|
"loss": 0.4848, |
|
"step": 11310 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 7.927321668909825e-06, |
|
"loss": 0.4861, |
|
"step": 11320 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 7.916105877074923e-06, |
|
"loss": 0.4813, |
|
"step": 11330 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 7.90489008524002e-06, |
|
"loss": 0.4799, |
|
"step": 11340 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 7.893674293405114e-06, |
|
"loss": 0.4856, |
|
"step": 11350 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 7.882458501570212e-06, |
|
"loss": 0.4823, |
|
"step": 11360 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 7.871242709735308e-06, |
|
"loss": 0.4806, |
|
"step": 11370 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 7.860026917900405e-06, |
|
"loss": 0.4802, |
|
"step": 11380 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 7.848811126065501e-06, |
|
"loss": 0.489, |
|
"step": 11390 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 7.837595334230597e-06, |
|
"loss": 0.4811, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 7.826379542395694e-06, |
|
"loss": 0.4777, |
|
"step": 11410 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 7.81516375056079e-06, |
|
"loss": 0.4803, |
|
"step": 11420 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 7.803947958725887e-06, |
|
"loss": 0.4757, |
|
"step": 11430 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 7.792732166890983e-06, |
|
"loss": 0.4696, |
|
"step": 11440 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 7.78151637505608e-06, |
|
"loss": 0.4754, |
|
"step": 11450 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 7.770300583221176e-06, |
|
"loss": 0.4764, |
|
"step": 11460 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 7.759084791386272e-06, |
|
"loss": 0.4717, |
|
"step": 11470 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 7.747868999551368e-06, |
|
"loss": 0.4794, |
|
"step": 11480 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 7.736653207716466e-06, |
|
"loss": 0.4758, |
|
"step": 11490 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 7.725437415881561e-06, |
|
"loss": 0.4767, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 7.714221624046659e-06, |
|
"loss": 0.4745, |
|
"step": 11510 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 7.703005832211755e-06, |
|
"loss": 0.475, |
|
"step": 11520 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 7.69179004037685e-06, |
|
"loss": 0.4748, |
|
"step": 11530 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 7.680574248541948e-06, |
|
"loss": 0.4808, |
|
"step": 11540 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 7.669358456707045e-06, |
|
"loss": 0.4655, |
|
"step": 11550 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 7.658142664872141e-06, |
|
"loss": 0.4846, |
|
"step": 11560 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 7.646926873037237e-06, |
|
"loss": 0.4688, |
|
"step": 11570 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 7.635711081202334e-06, |
|
"loss": 0.4665, |
|
"step": 11580 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 7.62449528936743e-06, |
|
"loss": 0.4782, |
|
"step": 11590 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 7.613279497532526e-06, |
|
"loss": 0.4785, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 7.6020637056976235e-06, |
|
"loss": 0.4821, |
|
"step": 11610 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 7.590847913862719e-06, |
|
"loss": 0.4835, |
|
"step": 11620 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 7.579632122027815e-06, |
|
"loss": 0.4875, |
|
"step": 11630 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 7.568416330192913e-06, |
|
"loss": 0.4837, |
|
"step": 11640 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 7.557200538358008e-06, |
|
"loss": 0.4819, |
|
"step": 11650 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 7.545984746523105e-06, |
|
"loss": 0.4704, |
|
"step": 11660 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 7.534768954688202e-06, |
|
"loss": 0.477, |
|
"step": 11670 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 7.523553162853298e-06, |
|
"loss": 0.4685, |
|
"step": 11680 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 7.512337371018394e-06, |
|
"loss": 0.4822, |
|
"step": 11690 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 7.5011215791834916e-06, |
|
"loss": 0.4747, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 7.489905787348587e-06, |
|
"loss": 0.4768, |
|
"step": 11710 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 7.4786899955136834e-06, |
|
"loss": 0.4722, |
|
"step": 11720 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 7.467474203678781e-06, |
|
"loss": 0.4909, |
|
"step": 11730 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 7.456258411843876e-06, |
|
"loss": 0.4749, |
|
"step": 11740 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 7.445042620008973e-06, |
|
"loss": 0.4766, |
|
"step": 11750 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 7.43382682817407e-06, |
|
"loss": 0.4767, |
|
"step": 11760 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 7.422611036339166e-06, |
|
"loss": 0.4737, |
|
"step": 11770 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 7.4113952445042624e-06, |
|
"loss": 0.4827, |
|
"step": 11780 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 7.40017945266936e-06, |
|
"loss": 0.4819, |
|
"step": 11790 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 7.388963660834455e-06, |
|
"loss": 0.4827, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 7.3777478689995515e-06, |
|
"loss": 0.4795, |
|
"step": 11810 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 7.366532077164649e-06, |
|
"loss": 0.4636, |
|
"step": 11820 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 7.355316285329744e-06, |
|
"loss": 0.4824, |
|
"step": 11830 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 7.344100493494841e-06, |
|
"loss": 0.4732, |
|
"step": 11840 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 7.332884701659938e-06, |
|
"loss": 0.4798, |
|
"step": 11850 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 7.321668909825034e-06, |
|
"loss": 0.4813, |
|
"step": 11860 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 7.3104531179901305e-06, |
|
"loss": 0.4709, |
|
"step": 11870 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 7.299237326155228e-06, |
|
"loss": 0.4833, |
|
"step": 11880 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 7.288021534320323e-06, |
|
"loss": 0.4772, |
|
"step": 11890 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 7.2768057424854196e-06, |
|
"loss": 0.4654, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 7.265589950650517e-06, |
|
"loss": 0.4779, |
|
"step": 11910 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 7.254374158815612e-06, |
|
"loss": 0.4738, |
|
"step": 11920 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 7.2431583669807095e-06, |
|
"loss": 0.4783, |
|
"step": 11930 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 7.231942575145806e-06, |
|
"loss": 0.4798, |
|
"step": 11940 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 7.220726783310902e-06, |
|
"loss": 0.4845, |
|
"step": 11950 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 7.2095109914759986e-06, |
|
"loss": 0.4701, |
|
"step": 11960 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 7.198295199641096e-06, |
|
"loss": 0.4759, |
|
"step": 11970 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 7.187079407806191e-06, |
|
"loss": 0.4781, |
|
"step": 11980 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 7.175863615971288e-06, |
|
"loss": 0.4689, |
|
"step": 11990 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 7.164647824136385e-06, |
|
"loss": 0.4796, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 7.15343203230148e-06, |
|
"loss": 0.4801, |
|
"step": 12010 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 7.1422162404665775e-06, |
|
"loss": 0.4685, |
|
"step": 12020 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 7.131000448631674e-06, |
|
"loss": 0.4834, |
|
"step": 12030 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 7.11978465679677e-06, |
|
"loss": 0.474, |
|
"step": 12040 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 7.108568864961867e-06, |
|
"loss": 0.4703, |
|
"step": 12050 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 7.097353073126964e-06, |
|
"loss": 0.4895, |
|
"step": 12060 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 7.086137281292059e-06, |
|
"loss": 0.4765, |
|
"step": 12070 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 7.074921489457156e-06, |
|
"loss": 0.4661, |
|
"step": 12080 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 7.063705697622253e-06, |
|
"loss": 0.4753, |
|
"step": 12090 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 7.05248990578735e-06, |
|
"loss": 0.4835, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 7.041274113952446e-06, |
|
"loss": 0.4747, |
|
"step": 12110 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 7.030058322117542e-06, |
|
"loss": 0.4751, |
|
"step": 12120 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 7.018842530282639e-06, |
|
"loss": 0.4702, |
|
"step": 12130 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 7.007626738447735e-06, |
|
"loss": 0.4744, |
|
"step": 12140 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 6.996410946612832e-06, |
|
"loss": 0.4729, |
|
"step": 12150 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 6.985195154777928e-06, |
|
"loss": 0.4689, |
|
"step": 12160 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 6.973979362943024e-06, |
|
"loss": 0.4664, |
|
"step": 12170 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 6.962763571108121e-06, |
|
"loss": 0.4741, |
|
"step": 12180 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 6.951547779273218e-06, |
|
"loss": 0.4676, |
|
"step": 12190 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 6.940331987438314e-06, |
|
"loss": 0.4841, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 6.92911619560341e-06, |
|
"loss": 0.4848, |
|
"step": 12210 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 6.917900403768507e-06, |
|
"loss": 0.4792, |
|
"step": 12220 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 6.906684611933603e-06, |
|
"loss": 0.4863, |
|
"step": 12230 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 6.8954688200987e-06, |
|
"loss": 0.4789, |
|
"step": 12240 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 6.884253028263796e-06, |
|
"loss": 0.4726, |
|
"step": 12250 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 6.873037236428892e-06, |
|
"loss": 0.4762, |
|
"step": 12260 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 6.861821444593989e-06, |
|
"loss": 0.4746, |
|
"step": 12270 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 6.850605652759086e-06, |
|
"loss": 0.4637, |
|
"step": 12280 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 6.839389860924182e-06, |
|
"loss": 0.4619, |
|
"step": 12290 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 6.828174069089278e-06, |
|
"loss": 0.4725, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 6.816958277254375e-06, |
|
"loss": 0.4737, |
|
"step": 12310 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 6.805742485419471e-06, |
|
"loss": 0.4792, |
|
"step": 12320 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 6.794526693584568e-06, |
|
"loss": 0.4797, |
|
"step": 12330 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 6.783310901749664e-06, |
|
"loss": 0.4707, |
|
"step": 12340 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 6.77209510991476e-06, |
|
"loss": 0.4752, |
|
"step": 12350 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 6.760879318079857e-06, |
|
"loss": 0.4748, |
|
"step": 12360 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 6.749663526244954e-06, |
|
"loss": 0.471, |
|
"step": 12370 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 6.73844773441005e-06, |
|
"loss": 0.475, |
|
"step": 12380 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 6.727231942575146e-06, |
|
"loss": 0.4686, |
|
"step": 12390 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 6.716016150740243e-06, |
|
"loss": 0.4706, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 6.704800358905339e-06, |
|
"loss": 0.4615, |
|
"step": 12410 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 6.693584567070436e-06, |
|
"loss": 0.4762, |
|
"step": 12420 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 6.6823687752355324e-06, |
|
"loss": 0.4815, |
|
"step": 12430 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 6.671152983400628e-06, |
|
"loss": 0.477, |
|
"step": 12440 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 6.659937191565725e-06, |
|
"loss": 0.4645, |
|
"step": 12450 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 6.648721399730822e-06, |
|
"loss": 0.4692, |
|
"step": 12460 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 6.637505607895918e-06, |
|
"loss": 0.4739, |
|
"step": 12470 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 6.626289816061014e-06, |
|
"loss": 0.4704, |
|
"step": 12480 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 6.615074024226111e-06, |
|
"loss": 0.4788, |
|
"step": 12490 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 6.603858232391207e-06, |
|
"loss": 0.4717, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 6.592642440556304e-06, |
|
"loss": 0.4595, |
|
"step": 12510 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 6.5814266487214005e-06, |
|
"loss": 0.4764, |
|
"step": 12520 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 6.570210856886496e-06, |
|
"loss": 0.4757, |
|
"step": 12530 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 6.558995065051593e-06, |
|
"loss": 0.4757, |
|
"step": 12540 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 6.54777927321669e-06, |
|
"loss": 0.474, |
|
"step": 12550 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 6.536563481381786e-06, |
|
"loss": 0.4786, |
|
"step": 12560 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 6.525347689546882e-06, |
|
"loss": 0.4662, |
|
"step": 12570 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 6.5141318977119795e-06, |
|
"loss": 0.4684, |
|
"step": 12580 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 6.502916105877075e-06, |
|
"loss": 0.4727, |
|
"step": 12590 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 6.491700314042172e-06, |
|
"loss": 0.4913, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 6.4804845222072685e-06, |
|
"loss": 0.4618, |
|
"step": 12610 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 6.469268730372364e-06, |
|
"loss": 0.4695, |
|
"step": 12620 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 6.458052938537461e-06, |
|
"loss": 0.4704, |
|
"step": 12630 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 6.4468371467025585e-06, |
|
"loss": 0.4812, |
|
"step": 12640 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 6.435621354867654e-06, |
|
"loss": 0.4707, |
|
"step": 12650 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 6.42440556303275e-06, |
|
"loss": 0.4845, |
|
"step": 12660 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 6.4131897711978475e-06, |
|
"loss": 0.469, |
|
"step": 12670 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 6.401973979362943e-06, |
|
"loss": 0.4649, |
|
"step": 12680 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 6.39075818752804e-06, |
|
"loss": 0.4684, |
|
"step": 12690 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 6.379542395693137e-06, |
|
"loss": 0.4688, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 6.368326603858232e-06, |
|
"loss": 0.4748, |
|
"step": 12710 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 6.357110812023329e-06, |
|
"loss": 0.4631, |
|
"step": 12720 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 6.3458950201884265e-06, |
|
"loss": 0.4814, |
|
"step": 12730 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 6.334679228353522e-06, |
|
"loss": 0.4598, |
|
"step": 12740 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 6.323463436518618e-06, |
|
"loss": 0.4807, |
|
"step": 12750 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 6.312247644683716e-06, |
|
"loss": 0.4775, |
|
"step": 12760 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 6.301031852848811e-06, |
|
"loss": 0.464, |
|
"step": 12770 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 6.289816061013908e-06, |
|
"loss": 0.4724, |
|
"step": 12780 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 6.278600269179005e-06, |
|
"loss": 0.4737, |
|
"step": 12790 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 6.2673844773441e-06, |
|
"loss": 0.4737, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 6.256168685509197e-06, |
|
"loss": 0.4748, |
|
"step": 12810 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 6.244952893674295e-06, |
|
"loss": 0.4736, |
|
"step": 12820 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 6.23373710183939e-06, |
|
"loss": 0.483, |
|
"step": 12830 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 6.2225213100044865e-06, |
|
"loss": 0.4687, |
|
"step": 12840 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 6.211305518169584e-06, |
|
"loss": 0.4716, |
|
"step": 12850 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 6.200089726334679e-06, |
|
"loss": 0.4723, |
|
"step": 12860 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 6.188873934499776e-06, |
|
"loss": 0.4714, |
|
"step": 12870 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 6.177658142664873e-06, |
|
"loss": 0.4782, |
|
"step": 12880 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 6.166442350829969e-06, |
|
"loss": 0.4804, |
|
"step": 12890 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 6.1552265589950654e-06, |
|
"loss": 0.4793, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 6.144010767160163e-06, |
|
"loss": 0.4639, |
|
"step": 12910 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 6.132794975325258e-06, |
|
"loss": 0.4628, |
|
"step": 12920 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 6.1215791834903545e-06, |
|
"loss": 0.4746, |
|
"step": 12930 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 6.110363391655452e-06, |
|
"loss": 0.4625, |
|
"step": 12940 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 6.099147599820547e-06, |
|
"loss": 0.4649, |
|
"step": 12950 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 6.0879318079856444e-06, |
|
"loss": 0.4696, |
|
"step": 12960 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 6.076716016150741e-06, |
|
"loss": 0.4631, |
|
"step": 12970 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 6.065500224315837e-06, |
|
"loss": 0.4634, |
|
"step": 12980 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 6.0542844324809335e-06, |
|
"loss": 0.4702, |
|
"step": 12990 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 6.043068640646031e-06, |
|
"loss": 0.4643, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 6.031852848811126e-06, |
|
"loss": 0.4724, |
|
"step": 13010 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 6.020637056976223e-06, |
|
"loss": 0.4738, |
|
"step": 13020 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 6.00942126514132e-06, |
|
"loss": 0.4656, |
|
"step": 13030 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 5.998205473306415e-06, |
|
"loss": 0.4733, |
|
"step": 13040 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 5.9869896814715125e-06, |
|
"loss": 0.4786, |
|
"step": 13050 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 5.975773889636609e-06, |
|
"loss": 0.4657, |
|
"step": 13060 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 5.964558097801705e-06, |
|
"loss": 0.4794, |
|
"step": 13070 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 5.9533423059668016e-06, |
|
"loss": 0.4775, |
|
"step": 13080 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 5.942126514131899e-06, |
|
"loss": 0.4779, |
|
"step": 13090 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 5.930910722296994e-06, |
|
"loss": 0.4712, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 5.919694930462091e-06, |
|
"loss": 0.467, |
|
"step": 13110 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 5.908479138627188e-06, |
|
"loss": 0.4607, |
|
"step": 13120 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 5.897263346792283e-06, |
|
"loss": 0.4673, |
|
"step": 13130 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 5.8860475549573806e-06, |
|
"loss": 0.459, |
|
"step": 13140 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 5.874831763122477e-06, |
|
"loss": 0.4771, |
|
"step": 13150 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 5.863615971287573e-06, |
|
"loss": 0.4767, |
|
"step": 13160 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 5.85240017945267e-06, |
|
"loss": 0.464, |
|
"step": 13170 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 5.841184387617767e-06, |
|
"loss": 0.4724, |
|
"step": 13180 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 5.829968595782862e-06, |
|
"loss": 0.464, |
|
"step": 13190 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 5.818752803947959e-06, |
|
"loss": 0.4793, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 5.807537012113056e-06, |
|
"loss": 0.4777, |
|
"step": 13210 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 5.796321220278151e-06, |
|
"loss": 0.4756, |
|
"step": 13220 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 5.785105428443249e-06, |
|
"loss": 0.4695, |
|
"step": 13230 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 5.773889636608345e-06, |
|
"loss": 0.4699, |
|
"step": 13240 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 5.762673844773441e-06, |
|
"loss": 0.4676, |
|
"step": 13250 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 5.751458052938538e-06, |
|
"loss": 0.4714, |
|
"step": 13260 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 5.740242261103635e-06, |
|
"loss": 0.475, |
|
"step": 13270 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 5.72902646926873e-06, |
|
"loss": 0.4803, |
|
"step": 13280 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 5.717810677433827e-06, |
|
"loss": 0.4757, |
|
"step": 13290 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 5.706594885598924e-06, |
|
"loss": 0.4742, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 5.6953790937640195e-06, |
|
"loss": 0.4717, |
|
"step": 13310 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 5.684163301929117e-06, |
|
"loss": 0.4766, |
|
"step": 13320 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 5.672947510094213e-06, |
|
"loss": 0.4793, |
|
"step": 13330 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 5.66173171825931e-06, |
|
"loss": 0.4717, |
|
"step": 13340 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 5.650515926424406e-06, |
|
"loss": 0.4591, |
|
"step": 13350 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 5.639300134589503e-06, |
|
"loss": 0.463, |
|
"step": 13360 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 5.628084342754599e-06, |
|
"loss": 0.4669, |
|
"step": 13370 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 5.616868550919695e-06, |
|
"loss": 0.4773, |
|
"step": 13380 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 5.605652759084792e-06, |
|
"loss": 0.4642, |
|
"step": 13390 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 5.594436967249889e-06, |
|
"loss": 0.4689, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 5.583221175414985e-06, |
|
"loss": 0.4697, |
|
"step": 13410 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 5.572005383580081e-06, |
|
"loss": 0.4627, |
|
"step": 13420 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 5.560789591745178e-06, |
|
"loss": 0.4595, |
|
"step": 13430 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 5.549573799910274e-06, |
|
"loss": 0.4661, |
|
"step": 13440 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 5.538358008075371e-06, |
|
"loss": 0.4616, |
|
"step": 13450 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 5.527142216240467e-06, |
|
"loss": 0.4599, |
|
"step": 13460 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 5.515926424405563e-06, |
|
"loss": 0.4659, |
|
"step": 13470 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 5.50471063257066e-06, |
|
"loss": 0.4645, |
|
"step": 13480 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 5.493494840735757e-06, |
|
"loss": 0.4646, |
|
"step": 13490 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 5.482279048900853e-06, |
|
"loss": 0.4703, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 5.471063257065949e-06, |
|
"loss": 0.4717, |
|
"step": 13510 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 5.459847465231046e-06, |
|
"loss": 0.4822, |
|
"step": 13520 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 5.448631673396142e-06, |
|
"loss": 0.4635, |
|
"step": 13530 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 5.437415881561239e-06, |
|
"loss": 0.4709, |
|
"step": 13540 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 5.4262000897263354e-06, |
|
"loss": 0.4734, |
|
"step": 13550 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 5.414984297891431e-06, |
|
"loss": 0.4664, |
|
"step": 13560 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 5.403768506056528e-06, |
|
"loss": 0.472, |
|
"step": 13570 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 5.392552714221625e-06, |
|
"loss": 0.4685, |
|
"step": 13580 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 5.381336922386721e-06, |
|
"loss": 0.4605, |
|
"step": 13590 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 5.370121130551817e-06, |
|
"loss": 0.4594, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 5.3589053387169144e-06, |
|
"loss": 0.4852, |
|
"step": 13610 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 5.34768954688201e-06, |
|
"loss": 0.4696, |
|
"step": 13620 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 5.336473755047107e-06, |
|
"loss": 0.4669, |
|
"step": 13630 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 5.3252579632122035e-06, |
|
"loss": 0.4792, |
|
"step": 13640 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 5.314042171377299e-06, |
|
"loss": 0.461, |
|
"step": 13650 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 5.302826379542396e-06, |
|
"loss": 0.4772, |
|
"step": 13660 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 5.2916105877074934e-06, |
|
"loss": 0.4671, |
|
"step": 13670 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 5.280394795872589e-06, |
|
"loss": 0.4792, |
|
"step": 13680 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 5.269179004037685e-06, |
|
"loss": 0.4604, |
|
"step": 13690 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 5.2579632122027825e-06, |
|
"loss": 0.4778, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 5.246747420367878e-06, |
|
"loss": 0.4726, |
|
"step": 13710 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 5.235531628532975e-06, |
|
"loss": 0.4681, |
|
"step": 13720 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 5.2243158366980716e-06, |
|
"loss": 0.4716, |
|
"step": 13730 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 5.213100044863167e-06, |
|
"loss": 0.4725, |
|
"step": 13740 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 5.201884253028264e-06, |
|
"loss": 0.4678, |
|
"step": 13750 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 5.1906684611933615e-06, |
|
"loss": 0.4672, |
|
"step": 13760 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 5.179452669358457e-06, |
|
"loss": 0.4768, |
|
"step": 13770 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 5.168236877523553e-06, |
|
"loss": 0.4675, |
|
"step": 13780 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 5.1570210856886506e-06, |
|
"loss": 0.4757, |
|
"step": 13790 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 5.145805293853746e-06, |
|
"loss": 0.472, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 5.134589502018843e-06, |
|
"loss": 0.4828, |
|
"step": 13810 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 5.12337371018394e-06, |
|
"loss": 0.4626, |
|
"step": 13820 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 5.112157918349035e-06, |
|
"loss": 0.4668, |
|
"step": 13830 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 5.100942126514132e-06, |
|
"loss": 0.4738, |
|
"step": 13840 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 5.0897263346792296e-06, |
|
"loss": 0.4797, |
|
"step": 13850 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 5.078510542844325e-06, |
|
"loss": 0.4635, |
|
"step": 13860 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 5.067294751009421e-06, |
|
"loss": 0.482, |
|
"step": 13870 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 5.056078959174519e-06, |
|
"loss": 0.4805, |
|
"step": 13880 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 5.044863167339614e-06, |
|
"loss": 0.4657, |
|
"step": 13890 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 5.033647375504711e-06, |
|
"loss": 0.4603, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 5.022431583669808e-06, |
|
"loss": 0.46, |
|
"step": 13910 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 5.011215791834903e-06, |
|
"loss": 0.4665, |
|
"step": 13920 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 5e-06, |
|
"loss": 0.467, |
|
"step": 13930 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 4.988784208165097e-06, |
|
"loss": 0.4726, |
|
"step": 13940 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 4.977568416330193e-06, |
|
"loss": 0.4608, |
|
"step": 13950 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 4.9663526244952895e-06, |
|
"loss": 0.4735, |
|
"step": 13960 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 4.955136832660386e-06, |
|
"loss": 0.4627, |
|
"step": 13970 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 4.943921040825483e-06, |
|
"loss": 0.4797, |
|
"step": 13980 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 4.932705248990579e-06, |
|
"loss": 0.4708, |
|
"step": 13990 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 4.921489457155676e-06, |
|
"loss": 0.4673, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 4.910273665320772e-06, |
|
"loss": 0.4703, |
|
"step": 14010 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 4.8990578734858685e-06, |
|
"loss": 0.4644, |
|
"step": 14020 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 4.887842081650965e-06, |
|
"loss": 0.4712, |
|
"step": 14030 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 4.876626289816061e-06, |
|
"loss": 0.4643, |
|
"step": 14040 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 4.8654104979811575e-06, |
|
"loss": 0.4795, |
|
"step": 14050 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 4.854194706146254e-06, |
|
"loss": 0.4697, |
|
"step": 14060 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 4.842978914311351e-06, |
|
"loss": 0.4651, |
|
"step": 14070 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 4.8317631224764475e-06, |
|
"loss": 0.4607, |
|
"step": 14080 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 4.820547330641544e-06, |
|
"loss": 0.4746, |
|
"step": 14090 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 4.80933153880664e-06, |
|
"loss": 0.4804, |
|
"step": 14100 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 4.7981157469717365e-06, |
|
"loss": 0.4538, |
|
"step": 14110 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 4.786899955136833e-06, |
|
"loss": 0.4744, |
|
"step": 14120 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 4.775684163301929e-06, |
|
"loss": 0.4747, |
|
"step": 14130 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 4.764468371467026e-06, |
|
"loss": 0.4603, |
|
"step": 14140 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 4.753252579632122e-06, |
|
"loss": 0.4742, |
|
"step": 14150 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 4.742036787797219e-06, |
|
"loss": 0.4678, |
|
"step": 14160 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 4.7308209959623155e-06, |
|
"loss": 0.4707, |
|
"step": 14170 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 4.719605204127412e-06, |
|
"loss": 0.4724, |
|
"step": 14180 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 4.708389412292508e-06, |
|
"loss": 0.4657, |
|
"step": 14190 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 4.697173620457605e-06, |
|
"loss": 0.4748, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 4.685957828622701e-06, |
|
"loss": 0.4663, |
|
"step": 14210 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 4.674742036787797e-06, |
|
"loss": 0.4656, |
|
"step": 14220 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 4.663526244952894e-06, |
|
"loss": 0.4611, |
|
"step": 14230 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 4.652310453117991e-06, |
|
"loss": 0.4658, |
|
"step": 14240 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 4.641094661283087e-06, |
|
"loss": 0.4741, |
|
"step": 14250 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 4.629878869448184e-06, |
|
"loss": 0.4728, |
|
"step": 14260 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 4.61866307761328e-06, |
|
"loss": 0.4627, |
|
"step": 14270 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 4.607447285778376e-06, |
|
"loss": 0.4702, |
|
"step": 14280 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 4.596231493943473e-06, |
|
"loss": 0.4744, |
|
"step": 14290 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 4.58501570210857e-06, |
|
"loss": 0.4606, |
|
"step": 14300 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 4.573799910273665e-06, |
|
"loss": 0.467, |
|
"step": 14310 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 4.562584118438762e-06, |
|
"loss": 0.4593, |
|
"step": 14320 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 4.551368326603859e-06, |
|
"loss": 0.4724, |
|
"step": 14330 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 4.540152534768955e-06, |
|
"loss": 0.4659, |
|
"step": 14340 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 4.528936742934052e-06, |
|
"loss": 0.4653, |
|
"step": 14350 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 4.517720951099148e-06, |
|
"loss": 0.4571, |
|
"step": 14360 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 4.506505159264244e-06, |
|
"loss": 0.4711, |
|
"step": 14370 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 4.495289367429341e-06, |
|
"loss": 0.4657, |
|
"step": 14380 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 4.484073575594438e-06, |
|
"loss": 0.4671, |
|
"step": 14390 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 4.4728577837595334e-06, |
|
"loss": 0.4677, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 4.46164199192463e-06, |
|
"loss": 0.4751, |
|
"step": 14410 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 4.450426200089727e-06, |
|
"loss": 0.4614, |
|
"step": 14420 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 4.439210408254823e-06, |
|
"loss": 0.4771, |
|
"step": 14430 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 4.42799461641992e-06, |
|
"loss": 0.4573, |
|
"step": 14440 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 4.416778824585016e-06, |
|
"loss": 0.4673, |
|
"step": 14450 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 4.4055630327501124e-06, |
|
"loss": 0.4746, |
|
"step": 14460 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 4.394347240915209e-06, |
|
"loss": 0.4792, |
|
"step": 14470 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 4.383131449080306e-06, |
|
"loss": 0.4728, |
|
"step": 14480 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 4.3719156572454015e-06, |
|
"loss": 0.4674, |
|
"step": 14490 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 4.360699865410498e-06, |
|
"loss": 0.4759, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 4.349484073575595e-06, |
|
"loss": 0.478, |
|
"step": 14510 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 4.338268281740691e-06, |
|
"loss": 0.464, |
|
"step": 14520 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 4.327052489905788e-06, |
|
"loss": 0.4801, |
|
"step": 14530 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 4.315836698070884e-06, |
|
"loss": 0.4704, |
|
"step": 14540 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 4.3046209062359805e-06, |
|
"loss": 0.4778, |
|
"step": 14550 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 4.293405114401077e-06, |
|
"loss": 0.4783, |
|
"step": 14560 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 4.282189322566174e-06, |
|
"loss": 0.4821, |
|
"step": 14570 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 4.2709735307312696e-06, |
|
"loss": 0.4607, |
|
"step": 14580 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 4.259757738896366e-06, |
|
"loss": 0.4752, |
|
"step": 14590 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 4.248541947061463e-06, |
|
"loss": 0.4672, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 4.2373261552265595e-06, |
|
"loss": 0.466, |
|
"step": 14610 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 4.226110363391656e-06, |
|
"loss": 0.4711, |
|
"step": 14620 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 4.214894571556752e-06, |
|
"loss": 0.4813, |
|
"step": 14630 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 4.2036787797218485e-06, |
|
"loss": 0.4731, |
|
"step": 14640 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 4.192462987886945e-06, |
|
"loss": 0.4656, |
|
"step": 14650 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 4.181247196052042e-06, |
|
"loss": 0.4605, |
|
"step": 14660 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 4.170031404217138e-06, |
|
"loss": 0.4535, |
|
"step": 14670 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 4.158815612382234e-06, |
|
"loss": 0.4601, |
|
"step": 14680 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 4.147599820547331e-06, |
|
"loss": 0.4617, |
|
"step": 14690 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 4.1363840287124275e-06, |
|
"loss": 0.4691, |
|
"step": 14700 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 4.125168236877524e-06, |
|
"loss": 0.4707, |
|
"step": 14710 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 4.11395244504262e-06, |
|
"loss": 0.4678, |
|
"step": 14720 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 4.102736653207717e-06, |
|
"loss": 0.468, |
|
"step": 14730 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 4.091520861372813e-06, |
|
"loss": 0.4675, |
|
"step": 14740 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 4.08030506953791e-06, |
|
"loss": 0.4746, |
|
"step": 14750 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 4.069089277703006e-06, |
|
"loss": 0.4729, |
|
"step": 14760 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 4.057873485868102e-06, |
|
"loss": 0.4681, |
|
"step": 14770 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 4.046657694033199e-06, |
|
"loss": 0.4726, |
|
"step": 14780 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 4.035441902198296e-06, |
|
"loss": 0.472, |
|
"step": 14790 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 4.024226110363392e-06, |
|
"loss": 0.4532, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 4.013010318528488e-06, |
|
"loss": 0.475, |
|
"step": 14810 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 4.001794526693585e-06, |
|
"loss": 0.4659, |
|
"step": 14820 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 3.990578734858681e-06, |
|
"loss": 0.4747, |
|
"step": 14830 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 3.979362943023778e-06, |
|
"loss": 0.4697, |
|
"step": 14840 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 3.968147151188875e-06, |
|
"loss": 0.4709, |
|
"step": 14850 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 3.956931359353971e-06, |
|
"loss": 0.4738, |
|
"step": 14860 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 3.945715567519067e-06, |
|
"loss": 0.4628, |
|
"step": 14870 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 3.934499775684164e-06, |
|
"loss": 0.4695, |
|
"step": 14880 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 3.92328398384926e-06, |
|
"loss": 0.4789, |
|
"step": 14890 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 3.912068192014356e-06, |
|
"loss": 0.462, |
|
"step": 14900 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 3.900852400179453e-06, |
|
"loss": 0.4543, |
|
"step": 14910 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 3.88963660834455e-06, |
|
"loss": 0.4577, |
|
"step": 14920 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 3.878420816509646e-06, |
|
"loss": 0.4619, |
|
"step": 14930 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 3.867205024674743e-06, |
|
"loss": 0.4723, |
|
"step": 14940 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 3.855989232839839e-06, |
|
"loss": 0.4687, |
|
"step": 14950 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 3.844773441004935e-06, |
|
"loss": 0.4745, |
|
"step": 14960 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 3.833557649170032e-06, |
|
"loss": 0.4795, |
|
"step": 14970 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 3.822341857335128e-06, |
|
"loss": 0.4591, |
|
"step": 14980 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 3.811126065500225e-06, |
|
"loss": 0.4607, |
|
"step": 14990 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 3.799910273665321e-06, |
|
"loss": 0.4693, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 3.7886944818304176e-06, |
|
"loss": 0.4798, |
|
"step": 15010 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 3.777478689995514e-06, |
|
"loss": 0.4655, |
|
"step": 15020 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 3.7662628981606103e-06, |
|
"loss": 0.4543, |
|
"step": 15030 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 3.755047106325707e-06, |
|
"loss": 0.4617, |
|
"step": 15040 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 3.7438313144908034e-06, |
|
"loss": 0.4731, |
|
"step": 15050 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 3.7337371018393903e-06, |
|
"loss": 0.4608, |
|
"step": 15060 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 3.7225213100044867e-06, |
|
"loss": 0.4767, |
|
"step": 15070 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 3.711305518169583e-06, |
|
"loss": 0.4679, |
|
"step": 15080 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 3.70008972633468e-06, |
|
"loss": 0.4646, |
|
"step": 15090 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 3.6888739344997758e-06, |
|
"loss": 0.4713, |
|
"step": 15100 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 3.677658142664872e-06, |
|
"loss": 0.4628, |
|
"step": 15110 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 3.666442350829969e-06, |
|
"loss": 0.4609, |
|
"step": 15120 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 3.6552265589950652e-06, |
|
"loss": 0.4664, |
|
"step": 15130 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 3.6440107671601616e-06, |
|
"loss": 0.4698, |
|
"step": 15140 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 3.6327949753252584e-06, |
|
"loss": 0.4703, |
|
"step": 15150 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 3.6215791834903547e-06, |
|
"loss": 0.4652, |
|
"step": 15160 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 3.610363391655451e-06, |
|
"loss": 0.4676, |
|
"step": 15170 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 3.599147599820548e-06, |
|
"loss": 0.4586, |
|
"step": 15180 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 3.587931807985644e-06, |
|
"loss": 0.4713, |
|
"step": 15190 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 3.57671601615074e-06, |
|
"loss": 0.4608, |
|
"step": 15200 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 3.565500224315837e-06, |
|
"loss": 0.4606, |
|
"step": 15210 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 3.5542844324809333e-06, |
|
"loss": 0.4805, |
|
"step": 15220 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 3.5430686406460297e-06, |
|
"loss": 0.4572, |
|
"step": 15230 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 3.5318528488111264e-06, |
|
"loss": 0.4591, |
|
"step": 15240 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 3.520637056976223e-06, |
|
"loss": 0.4635, |
|
"step": 15250 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 3.5094212651413196e-06, |
|
"loss": 0.4734, |
|
"step": 15260 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 3.498205473306416e-06, |
|
"loss": 0.4624, |
|
"step": 15270 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 3.486989681471512e-06, |
|
"loss": 0.4719, |
|
"step": 15280 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 3.475773889636609e-06, |
|
"loss": 0.4617, |
|
"step": 15290 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 3.464558097801705e-06, |
|
"loss": 0.4742, |
|
"step": 15300 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 3.4533423059668014e-06, |
|
"loss": 0.47, |
|
"step": 15310 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 3.442126514131898e-06, |
|
"loss": 0.4725, |
|
"step": 15320 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 3.4309107222969945e-06, |
|
"loss": 0.4618, |
|
"step": 15330 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 3.419694930462091e-06, |
|
"loss": 0.4651, |
|
"step": 15340 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 3.4084791386271876e-06, |
|
"loss": 0.4662, |
|
"step": 15350 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 3.397263346792284e-06, |
|
"loss": 0.4773, |
|
"step": 15360 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 3.38604755495738e-06, |
|
"loss": 0.457, |
|
"step": 15370 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 3.374831763122477e-06, |
|
"loss": 0.4621, |
|
"step": 15380 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 3.363615971287573e-06, |
|
"loss": 0.472, |
|
"step": 15390 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 3.3524001794526694e-06, |
|
"loss": 0.4784, |
|
"step": 15400 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 3.3411843876177662e-06, |
|
"loss": 0.4743, |
|
"step": 15410 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 3.3299685957828626e-06, |
|
"loss": 0.4677, |
|
"step": 15420 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 3.318752803947959e-06, |
|
"loss": 0.4707, |
|
"step": 15430 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 3.3075370121130557e-06, |
|
"loss": 0.4629, |
|
"step": 15440 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 3.296321220278152e-06, |
|
"loss": 0.4771, |
|
"step": 15450 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 3.285105428443248e-06, |
|
"loss": 0.4681, |
|
"step": 15460 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 3.273889636608345e-06, |
|
"loss": 0.4653, |
|
"step": 15470 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 3.262673844773441e-06, |
|
"loss": 0.459, |
|
"step": 15480 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 3.2514580529385375e-06, |
|
"loss": 0.471, |
|
"step": 15490 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 3.2402422611036343e-06, |
|
"loss": 0.464, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 3.2290264692687306e-06, |
|
"loss": 0.4774, |
|
"step": 15510 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 3.217810677433827e-06, |
|
"loss": 0.4736, |
|
"step": 15520 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 3.2065948855989238e-06, |
|
"loss": 0.4702, |
|
"step": 15530 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 3.19537909376402e-06, |
|
"loss": 0.4598, |
|
"step": 15540 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 3.184163301929116e-06, |
|
"loss": 0.4565, |
|
"step": 15550 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 3.1729475100942133e-06, |
|
"loss": 0.4617, |
|
"step": 15560 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 3.161731718259309e-06, |
|
"loss": 0.4555, |
|
"step": 15570 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 3.1505159264244056e-06, |
|
"loss": 0.4594, |
|
"step": 15580 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 3.1393001345895023e-06, |
|
"loss": 0.4625, |
|
"step": 15590 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 3.1280843427545987e-06, |
|
"loss": 0.4703, |
|
"step": 15600 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 3.116868550919695e-06, |
|
"loss": 0.4745, |
|
"step": 15610 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 3.105652759084792e-06, |
|
"loss": 0.477, |
|
"step": 15620 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 3.094436967249888e-06, |
|
"loss": 0.4619, |
|
"step": 15630 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 3.0832211754149845e-06, |
|
"loss": 0.4658, |
|
"step": 15640 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 3.0720053835800813e-06, |
|
"loss": 0.4684, |
|
"step": 15650 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 3.0607895917451773e-06, |
|
"loss": 0.461, |
|
"step": 15660 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 3.0495737999102736e-06, |
|
"loss": 0.4716, |
|
"step": 15670 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 3.0383580080753704e-06, |
|
"loss": 0.4632, |
|
"step": 15680 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 3.0271422162404668e-06, |
|
"loss": 0.4726, |
|
"step": 15690 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 3.015926424405563e-06, |
|
"loss": 0.4733, |
|
"step": 15700 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 3.00471063257066e-06, |
|
"loss": 0.4668, |
|
"step": 15710 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 2.9934948407357563e-06, |
|
"loss": 0.466, |
|
"step": 15720 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 2.9822790489008526e-06, |
|
"loss": 0.4708, |
|
"step": 15730 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 2.9710632570659494e-06, |
|
"loss": 0.4719, |
|
"step": 15740 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 2.9598474652310453e-06, |
|
"loss": 0.4757, |
|
"step": 15750 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 2.9486316733961417e-06, |
|
"loss": 0.4775, |
|
"step": 15760 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 2.9374158815612385e-06, |
|
"loss": 0.4569, |
|
"step": 15770 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 2.926200089726335e-06, |
|
"loss": 0.4758, |
|
"step": 15780 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 2.914984297891431e-06, |
|
"loss": 0.4654, |
|
"step": 15790 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 2.903768506056528e-06, |
|
"loss": 0.4641, |
|
"step": 15800 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 2.8925527142216243e-06, |
|
"loss": 0.4675, |
|
"step": 15810 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 2.8813369223867207e-06, |
|
"loss": 0.4771, |
|
"step": 15820 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 2.8701211305518175e-06, |
|
"loss": 0.4653, |
|
"step": 15830 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 2.8589053387169134e-06, |
|
"loss": 0.4646, |
|
"step": 15840 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 2.8476895468820097e-06, |
|
"loss": 0.4671, |
|
"step": 15850 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 2.8364737550471065e-06, |
|
"loss": 0.4733, |
|
"step": 15860 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 2.825257963212203e-06, |
|
"loss": 0.4629, |
|
"step": 15870 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 2.8140421713772997e-06, |
|
"loss": 0.4803, |
|
"step": 15880 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 2.802826379542396e-06, |
|
"loss": 0.465, |
|
"step": 15890 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 2.7916105877074924e-06, |
|
"loss": 0.466, |
|
"step": 15900 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 2.780394795872589e-06, |
|
"loss": 0.4667, |
|
"step": 15910 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 2.7691790040376855e-06, |
|
"loss": 0.4703, |
|
"step": 15920 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 2.7579632122027814e-06, |
|
"loss": 0.4772, |
|
"step": 15930 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 2.7467474203678787e-06, |
|
"loss": 0.4728, |
|
"step": 15940 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 2.7355316285329746e-06, |
|
"loss": 0.4724, |
|
"step": 15950 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 2.724315836698071e-06, |
|
"loss": 0.4666, |
|
"step": 15960 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 2.7131000448631677e-06, |
|
"loss": 0.4648, |
|
"step": 15970 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 2.701884253028264e-06, |
|
"loss": 0.4647, |
|
"step": 15980 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 2.6906684611933604e-06, |
|
"loss": 0.4679, |
|
"step": 15990 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 2.6794526693584572e-06, |
|
"loss": 0.4604, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 2.6682368775235536e-06, |
|
"loss": 0.4632, |
|
"step": 16010 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 2.6570210856886495e-06, |
|
"loss": 0.4593, |
|
"step": 16020 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 2.6458052938537467e-06, |
|
"loss": 0.4646, |
|
"step": 16030 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 2.6345895020188426e-06, |
|
"loss": 0.459, |
|
"step": 16040 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 2.623373710183939e-06, |
|
"loss": 0.4707, |
|
"step": 16050 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 2.6121579183490358e-06, |
|
"loss": 0.462, |
|
"step": 16060 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 2.600942126514132e-06, |
|
"loss": 0.4696, |
|
"step": 16070 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 2.5897263346792285e-06, |
|
"loss": 0.4678, |
|
"step": 16080 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 2.5785105428443253e-06, |
|
"loss": 0.4669, |
|
"step": 16090 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 2.5672947510094216e-06, |
|
"loss": 0.4619, |
|
"step": 16100 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 2.5560789591745176e-06, |
|
"loss": 0.4588, |
|
"step": 16110 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 2.5448631673396148e-06, |
|
"loss": 0.4608, |
|
"step": 16120 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 2.5336473755047107e-06, |
|
"loss": 0.4564, |
|
"step": 16130 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 2.522431583669807e-06, |
|
"loss": 0.4682, |
|
"step": 16140 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 2.511215791834904e-06, |
|
"loss": 0.4623, |
|
"step": 16150 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 2.5e-06, |
|
"loss": 0.4652, |
|
"step": 16160 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 2.4887842081650966e-06, |
|
"loss": 0.4679, |
|
"step": 16170 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 2.477568416330193e-06, |
|
"loss": 0.4719, |
|
"step": 16180 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 2.4663526244952897e-06, |
|
"loss": 0.4835, |
|
"step": 16190 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 2.455136832660386e-06, |
|
"loss": 0.4721, |
|
"step": 16200 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 2.4439210408254824e-06, |
|
"loss": 0.4616, |
|
"step": 16210 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 2.4327052489905788e-06, |
|
"loss": 0.4592, |
|
"step": 16220 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 2.4214894571556756e-06, |
|
"loss": 0.4607, |
|
"step": 16230 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 2.410273665320772e-06, |
|
"loss": 0.4716, |
|
"step": 16240 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 2.3990578734858683e-06, |
|
"loss": 0.4689, |
|
"step": 16250 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 2.3878420816509646e-06, |
|
"loss": 0.4712, |
|
"step": 16260 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 2.376626289816061e-06, |
|
"loss": 0.4771, |
|
"step": 16270 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 2.3654104979811578e-06, |
|
"loss": 0.4687, |
|
"step": 16280 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 2.354194706146254e-06, |
|
"loss": 0.4683, |
|
"step": 16290 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 2.3429789143113505e-06, |
|
"loss": 0.4672, |
|
"step": 16300 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 2.331763122476447e-06, |
|
"loss": 0.4772, |
|
"step": 16310 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 2.3205473306415436e-06, |
|
"loss": 0.4696, |
|
"step": 16320 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 2.30933153880664e-06, |
|
"loss": 0.4691, |
|
"step": 16330 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 2.2981157469717363e-06, |
|
"loss": 0.4646, |
|
"step": 16340 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 2.2868999551368327e-06, |
|
"loss": 0.4736, |
|
"step": 16350 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 2.2756841633019295e-06, |
|
"loss": 0.4831, |
|
"step": 16360 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 2.264468371467026e-06, |
|
"loss": 0.4555, |
|
"step": 16370 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 2.253252579632122e-06, |
|
"loss": 0.4751, |
|
"step": 16380 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 2.242036787797219e-06, |
|
"loss": 0.472, |
|
"step": 16390 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 2.230820995962315e-06, |
|
"loss": 0.4743, |
|
"step": 16400 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 2.2196052041274117e-06, |
|
"loss": 0.4657, |
|
"step": 16410 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 2.208389412292508e-06, |
|
"loss": 0.459, |
|
"step": 16420 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 2.1971736204576044e-06, |
|
"loss": 0.4586, |
|
"step": 16430 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 2.1859578286227007e-06, |
|
"loss": 0.4627, |
|
"step": 16440 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 2.1747420367877975e-06, |
|
"loss": 0.4665, |
|
"step": 16450 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 2.163526244952894e-06, |
|
"loss": 0.4625, |
|
"step": 16460 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 2.1523104531179902e-06, |
|
"loss": 0.4576, |
|
"step": 16470 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 2.141094661283087e-06, |
|
"loss": 0.465, |
|
"step": 16480 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 2.129878869448183e-06, |
|
"loss": 0.4622, |
|
"step": 16490 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 2.1186630776132797e-06, |
|
"loss": 0.4624, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 2.107447285778376e-06, |
|
"loss": 0.4644, |
|
"step": 16510 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 2.0962314939434725e-06, |
|
"loss": 0.472, |
|
"step": 16520 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 2.085015702108569e-06, |
|
"loss": 0.4694, |
|
"step": 16530 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 2.0737999102736656e-06, |
|
"loss": 0.4766, |
|
"step": 16540 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 2.062584118438762e-06, |
|
"loss": 0.467, |
|
"step": 16550 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 2.0513683266038583e-06, |
|
"loss": 0.4701, |
|
"step": 16560 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 2.040152534768955e-06, |
|
"loss": 0.4611, |
|
"step": 16570 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 2.028936742934051e-06, |
|
"loss": 0.4631, |
|
"step": 16580 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 2.017720951099148e-06, |
|
"loss": 0.4672, |
|
"step": 16590 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 2.006505159264244e-06, |
|
"loss": 0.4649, |
|
"step": 16600 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 1.9952893674293405e-06, |
|
"loss": 0.4703, |
|
"step": 16610 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 1.9840735755944373e-06, |
|
"loss": 0.4685, |
|
"step": 16620 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 1.9728577837595337e-06, |
|
"loss": 0.4678, |
|
"step": 16630 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 1.96164199192463e-06, |
|
"loss": 0.4651, |
|
"step": 16640 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 1.9504262000897264e-06, |
|
"loss": 0.4585, |
|
"step": 16650 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 1.939210408254823e-06, |
|
"loss": 0.4764, |
|
"step": 16660 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 1.9279946164199195e-06, |
|
"loss": 0.4656, |
|
"step": 16670 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 1.916778824585016e-06, |
|
"loss": 0.4627, |
|
"step": 16680 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 1.9055630327501124e-06, |
|
"loss": 0.4646, |
|
"step": 16690 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 1.8943472409152088e-06, |
|
"loss": 0.4678, |
|
"step": 16700 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 1.8831314490803051e-06, |
|
"loss": 0.4778, |
|
"step": 16710 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 1.8719156572454017e-06, |
|
"loss": 0.4632, |
|
"step": 16720 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 1.8606998654104983e-06, |
|
"loss": 0.474, |
|
"step": 16730 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 1.8494840735755944e-06, |
|
"loss": 0.4667, |
|
"step": 16740 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 1.838268281740691e-06, |
|
"loss": 0.4698, |
|
"step": 16750 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 1.8270524899057876e-06, |
|
"loss": 0.4575, |
|
"step": 16760 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 1.815836698070884e-06, |
|
"loss": 0.4755, |
|
"step": 16770 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 1.8046209062359805e-06, |
|
"loss": 0.4809, |
|
"step": 16780 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 1.7934051144010769e-06, |
|
"loss": 0.4655, |
|
"step": 16790 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 1.7821893225661732e-06, |
|
"loss": 0.4685, |
|
"step": 16800 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 1.7709735307312698e-06, |
|
"loss": 0.4659, |
|
"step": 16810 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 1.7597577388963663e-06, |
|
"loss": 0.4676, |
|
"step": 16820 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 1.7485419470614625e-06, |
|
"loss": 0.4582, |
|
"step": 16830 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 1.737326155226559e-06, |
|
"loss": 0.4692, |
|
"step": 16840 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 1.7261103633916556e-06, |
|
"loss": 0.4674, |
|
"step": 16850 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 1.714894571556752e-06, |
|
"loss": 0.4644, |
|
"step": 16860 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 1.7036787797218486e-06, |
|
"loss": 0.4727, |
|
"step": 16870 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 1.6924629878869451e-06, |
|
"loss": 0.4748, |
|
"step": 16880 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 1.6812471960520413e-06, |
|
"loss": 0.4653, |
|
"step": 16890 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 1.6700314042171378e-06, |
|
"loss": 0.4676, |
|
"step": 16900 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 1.6588156123822344e-06, |
|
"loss": 0.4725, |
|
"step": 16910 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 1.6475998205473306e-06, |
|
"loss": 0.4715, |
|
"step": 16920 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 1.6363840287124271e-06, |
|
"loss": 0.4602, |
|
"step": 16930 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 1.6251682368775237e-06, |
|
"loss": 0.4657, |
|
"step": 16940 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 1.6139524450426203e-06, |
|
"loss": 0.4735, |
|
"step": 16950 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 1.6027366532077166e-06, |
|
"loss": 0.4625, |
|
"step": 16960 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 1.5915208613728132e-06, |
|
"loss": 0.4575, |
|
"step": 16970 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 1.5803050695379095e-06, |
|
"loss": 0.4619, |
|
"step": 16980 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 1.569089277703006e-06, |
|
"loss": 0.4673, |
|
"step": 16990 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 1.5578734858681025e-06, |
|
"loss": 0.4607, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 1.546657694033199e-06, |
|
"loss": 0.4702, |
|
"step": 17010 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 1.5354419021982952e-06, |
|
"loss": 0.4622, |
|
"step": 17020 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 1.5242261103633918e-06, |
|
"loss": 0.46, |
|
"step": 17030 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 1.5130103185284883e-06, |
|
"loss": 0.4759, |
|
"step": 17040 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 1.5017945266935847e-06, |
|
"loss": 0.4663, |
|
"step": 17050 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 1.4905787348586812e-06, |
|
"loss": 0.4773, |
|
"step": 17060 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 1.4793629430237776e-06, |
|
"loss": 0.4588, |
|
"step": 17070 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 1.468147151188874e-06, |
|
"loss": 0.4705, |
|
"step": 17080 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 1.4569313593539705e-06, |
|
"loss": 0.466, |
|
"step": 17090 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 1.445715567519067e-06, |
|
"loss": 0.4704, |
|
"step": 17100 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 1.4344997756841632e-06, |
|
"loss": 0.4631, |
|
"step": 17110 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 1.4232839838492598e-06, |
|
"loss": 0.4658, |
|
"step": 17120 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 1.4120681920143564e-06, |
|
"loss": 0.4629, |
|
"step": 17130 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 1.4008524001794527e-06, |
|
"loss": 0.4614, |
|
"step": 17140 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 1.3896366083445493e-06, |
|
"loss": 0.4742, |
|
"step": 17150 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 1.3784208165096457e-06, |
|
"loss": 0.4547, |
|
"step": 17160 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 1.367205024674742e-06, |
|
"loss": 0.4481, |
|
"step": 17170 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 1.3559892328398386e-06, |
|
"loss": 0.4628, |
|
"step": 17180 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 1.3447734410049352e-06, |
|
"loss": 0.47, |
|
"step": 17190 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 1.3335576491700313e-06, |
|
"loss": 0.4543, |
|
"step": 17200 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 1.3223418573351279e-06, |
|
"loss": 0.4646, |
|
"step": 17210 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 1.3111260655002244e-06, |
|
"loss": 0.4705, |
|
"step": 17220 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 1.2999102736653208e-06, |
|
"loss": 0.4649, |
|
"step": 17230 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 1.2886944818304174e-06, |
|
"loss": 0.4637, |
|
"step": 17240 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 1.2774786899955137e-06, |
|
"loss": 0.4602, |
|
"step": 17250 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 1.2662628981606103e-06, |
|
"loss": 0.4743, |
|
"step": 17260 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 1.2550471063257067e-06, |
|
"loss": 0.4649, |
|
"step": 17270 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 1.2438313144908032e-06, |
|
"loss": 0.4541, |
|
"step": 17280 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 1.2326155226558996e-06, |
|
"loss": 0.4619, |
|
"step": 17290 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 1.221399730820996e-06, |
|
"loss": 0.4783, |
|
"step": 17300 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 1.2101839389860925e-06, |
|
"loss": 0.451, |
|
"step": 17310 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 1.1989681471511889e-06, |
|
"loss": 0.4639, |
|
"step": 17320 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 1.1877523553162854e-06, |
|
"loss": 0.4599, |
|
"step": 17330 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 1.176536563481382e-06, |
|
"loss": 0.4702, |
|
"step": 17340 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 1.1653207716464784e-06, |
|
"loss": 0.4689, |
|
"step": 17350 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 1.154104979811575e-06, |
|
"loss": 0.4683, |
|
"step": 17360 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 1.1428891879766713e-06, |
|
"loss": 0.4672, |
|
"step": 17370 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 1.1316733961417676e-06, |
|
"loss": 0.4728, |
|
"step": 17380 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 1.1204576043068642e-06, |
|
"loss": 0.4736, |
|
"step": 17390 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 1.1092418124719606e-06, |
|
"loss": 0.461, |
|
"step": 17400 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 1.098026020637057e-06, |
|
"loss": 0.4564, |
|
"step": 17410 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 1.0868102288021535e-06, |
|
"loss": 0.4573, |
|
"step": 17420 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 1.07559443696725e-06, |
|
"loss": 0.4646, |
|
"step": 17430 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 1.0643786451323464e-06, |
|
"loss": 0.4579, |
|
"step": 17440 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 1.053162853297443e-06, |
|
"loss": 0.4557, |
|
"step": 17450 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 1.0419470614625394e-06, |
|
"loss": 0.4654, |
|
"step": 17460 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 1.0307312696276357e-06, |
|
"loss": 0.4673, |
|
"step": 17470 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 1.0195154777927323e-06, |
|
"loss": 0.4602, |
|
"step": 17480 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 1.0082996859578286e-06, |
|
"loss": 0.4536, |
|
"step": 17490 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 9.970838941229252e-07, |
|
"loss": 0.4669, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 9.858681022880216e-07, |
|
"loss": 0.4633, |
|
"step": 17510 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 9.746523104531181e-07, |
|
"loss": 0.468, |
|
"step": 17520 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 9.634365186182145e-07, |
|
"loss": 0.4733, |
|
"step": 17530 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 9.52220726783311e-07, |
|
"loss": 0.476, |
|
"step": 17540 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 9.410049349484074e-07, |
|
"loss": 0.4666, |
|
"step": 17550 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 9.29789143113504e-07, |
|
"loss": 0.4684, |
|
"step": 17560 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 9.185733512786003e-07, |
|
"loss": 0.4673, |
|
"step": 17570 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 9.073575594436967e-07, |
|
"loss": 0.4658, |
|
"step": 17580 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 8.961417676087933e-07, |
|
"loss": 0.4626, |
|
"step": 17590 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 8.849259757738897e-07, |
|
"loss": 0.4619, |
|
"step": 17600 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 8.737101839389861e-07, |
|
"loss": 0.4563, |
|
"step": 17610 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 8.624943921040827e-07, |
|
"loss": 0.4705, |
|
"step": 17620 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 8.51278600269179e-07, |
|
"loss": 0.4573, |
|
"step": 17630 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 8.400628084342756e-07, |
|
"loss": 0.4791, |
|
"step": 17640 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 8.28847016599372e-07, |
|
"loss": 0.458, |
|
"step": 17650 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 8.176312247644684e-07, |
|
"loss": 0.4609, |
|
"step": 17660 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 8.06415432929565e-07, |
|
"loss": 0.4632, |
|
"step": 17670 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 7.951996410946613e-07, |
|
"loss": 0.458, |
|
"step": 17680 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 7.839838492597578e-07, |
|
"loss": 0.4806, |
|
"step": 17690 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 7.727680574248544e-07, |
|
"loss": 0.467, |
|
"step": 17700 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 7.615522655899507e-07, |
|
"loss": 0.4666, |
|
"step": 17710 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 7.503364737550471e-07, |
|
"loss": 0.4675, |
|
"step": 17720 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 7.391206819201436e-07, |
|
"loss": 0.4607, |
|
"step": 17730 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 7.279048900852401e-07, |
|
"loss": 0.4656, |
|
"step": 17740 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 7.166890982503365e-07, |
|
"loss": 0.462, |
|
"step": 17750 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 7.05473306415433e-07, |
|
"loss": 0.4694, |
|
"step": 17760 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 6.942575145805294e-07, |
|
"loss": 0.4644, |
|
"step": 17770 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 6.830417227456259e-07, |
|
"loss": 0.4688, |
|
"step": 17780 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 6.718259309107224e-07, |
|
"loss": 0.4758, |
|
"step": 17790 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 6.606101390758188e-07, |
|
"loss": 0.4581, |
|
"step": 17800 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 6.493943472409153e-07, |
|
"loss": 0.4693, |
|
"step": 17810 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 6.381785554060117e-07, |
|
"loss": 0.4669, |
|
"step": 17820 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 6.269627635711082e-07, |
|
"loss": 0.4576, |
|
"step": 17830 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 6.157469717362046e-07, |
|
"loss": 0.4635, |
|
"step": 17840 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 6.045311799013011e-07, |
|
"loss": 0.4788, |
|
"step": 17850 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 5.933153880663976e-07, |
|
"loss": 0.4627, |
|
"step": 17860 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 5.820995962314939e-07, |
|
"loss": 0.4625, |
|
"step": 17870 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 5.708838043965905e-07, |
|
"loss": 0.4626, |
|
"step": 17880 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 5.596680125616869e-07, |
|
"loss": 0.4663, |
|
"step": 17890 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 5.484522207267833e-07, |
|
"loss": 0.4782, |
|
"step": 17900 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 5.372364288918798e-07, |
|
"loss": 0.4608, |
|
"step": 17910 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 5.260206370569762e-07, |
|
"loss": 0.4672, |
|
"step": 17920 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 5.148048452220728e-07, |
|
"loss": 0.4614, |
|
"step": 17930 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 5.035890533871692e-07, |
|
"loss": 0.4591, |
|
"step": 17940 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 4.923732615522656e-07, |
|
"loss": 0.4624, |
|
"step": 17950 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 4.811574697173621e-07, |
|
"loss": 0.4589, |
|
"step": 17960 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 4.699416778824585e-07, |
|
"loss": 0.4598, |
|
"step": 17970 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 4.58725886047555e-07, |
|
"loss": 0.4641, |
|
"step": 17980 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 4.4751009421265147e-07, |
|
"loss": 0.4718, |
|
"step": 17990 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 4.3629430237774793e-07, |
|
"loss": 0.4635, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 4.2507851054284434e-07, |
|
"loss": 0.4641, |
|
"step": 18010 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 4.138627187079408e-07, |
|
"loss": 0.48, |
|
"step": 18020 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 4.0264692687303727e-07, |
|
"loss": 0.4724, |
|
"step": 18030 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 3.914311350381337e-07, |
|
"loss": 0.4706, |
|
"step": 18040 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 3.8021534320323014e-07, |
|
"loss": 0.4717, |
|
"step": 18050 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 3.6899955136832666e-07, |
|
"loss": 0.4662, |
|
"step": 18060 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 3.577837595334231e-07, |
|
"loss": 0.4572, |
|
"step": 18070 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 3.4656796769851953e-07, |
|
"loss": 0.466, |
|
"step": 18080 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 3.35352175863616e-07, |
|
"loss": 0.47, |
|
"step": 18090 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 3.2413638402871246e-07, |
|
"loss": 0.4666, |
|
"step": 18100 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 3.1292059219380887e-07, |
|
"loss": 0.4632, |
|
"step": 18110 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 3.0170480035890533e-07, |
|
"loss": 0.4671, |
|
"step": 18120 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 2.9048900852400185e-07, |
|
"loss": 0.4745, |
|
"step": 18130 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 2.7927321668909826e-07, |
|
"loss": 0.4666, |
|
"step": 18140 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 2.680574248541947e-07, |
|
"loss": 0.4636, |
|
"step": 18150 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 2.568416330192912e-07, |
|
"loss": 0.4618, |
|
"step": 18160 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 2.4562584118438765e-07, |
|
"loss": 0.4618, |
|
"step": 18170 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 2.344100493494841e-07, |
|
"loss": 0.479, |
|
"step": 18180 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 2.2319425751458055e-07, |
|
"loss": 0.4626, |
|
"step": 18190 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 2.11978465679677e-07, |
|
"loss": 0.4701, |
|
"step": 18200 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 2.0076267384477345e-07, |
|
"loss": 0.4585, |
|
"step": 18210 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 1.895468820098699e-07, |
|
"loss": 0.4665, |
|
"step": 18220 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 1.7833109017496637e-07, |
|
"loss": 0.4693, |
|
"step": 18230 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 1.671152983400628e-07, |
|
"loss": 0.4653, |
|
"step": 18240 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 1.5589950650515927e-07, |
|
"loss": 0.4653, |
|
"step": 18250 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 1.4468371467025574e-07, |
|
"loss": 0.4652, |
|
"step": 18260 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 1.3346792283535217e-07, |
|
"loss": 0.4618, |
|
"step": 18270 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 1.2225213100044864e-07, |
|
"loss": 0.4677, |
|
"step": 18280 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 1.110363391655451e-07, |
|
"loss": 0.4629, |
|
"step": 18290 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 9.982054733064156e-08, |
|
"loss": 0.4664, |
|
"step": 18300 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 8.8604755495738e-08, |
|
"loss": 0.4599, |
|
"step": 18310 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 7.738896366083446e-08, |
|
"loss": 0.4586, |
|
"step": 18320 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 6.617317182593091e-08, |
|
"loss": 0.4506, |
|
"step": 18330 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 5.4957379991027376e-08, |
|
"loss": 0.467, |
|
"step": 18340 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 4.374158815612382e-08, |
|
"loss": 0.4655, |
|
"step": 18350 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 3.252579632122028e-08, |
|
"loss": 0.4675, |
|
"step": 18360 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 2.131000448631674e-08, |
|
"loss": 0.4673, |
|
"step": 18370 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 1.009421265141319e-08, |
|
"loss": 0.4584, |
|
"step": 18380 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 18384, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"total_flos": 1.2862587547397652e+19, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|