|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 8.994283294516197, |
|
"eval_steps": 500, |
|
"global_step": 5310, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.01693838661867457, |
|
"grad_norm": 5.0625, |
|
"learning_rate": 1.8832391713747647e-07, |
|
"loss": 2.0482, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.03387677323734914, |
|
"grad_norm": 5.59375, |
|
"learning_rate": 3.7664783427495294e-07, |
|
"loss": 2.069, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.050815159856023716, |
|
"grad_norm": 5.96875, |
|
"learning_rate": 5.649717514124295e-07, |
|
"loss": 2.0861, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.06775354647469828, |
|
"grad_norm": 5.65625, |
|
"learning_rate": 7.532956685499059e-07, |
|
"loss": 2.0413, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.08469193309337286, |
|
"grad_norm": 16.125, |
|
"learning_rate": 9.416195856873824e-07, |
|
"loss": 2.0095, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.10163031971204743, |
|
"grad_norm": 17.75, |
|
"learning_rate": 1.129943502824859e-06, |
|
"loss": 2.0333, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.118568706330722, |
|
"grad_norm": 14.3125, |
|
"learning_rate": 1.3182674199623351e-06, |
|
"loss": 2.0062, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.13550709294939656, |
|
"grad_norm": 15.5625, |
|
"learning_rate": 1.5065913370998117e-06, |
|
"loss": 1.9655, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.15244547956807114, |
|
"grad_norm": 14.4375, |
|
"learning_rate": 1.6949152542372882e-06, |
|
"loss": 1.9698, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.16938386618674572, |
|
"grad_norm": 3.4375, |
|
"learning_rate": 1.8832391713747648e-06, |
|
"loss": 1.9211, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.18632225280542028, |
|
"grad_norm": 2.703125, |
|
"learning_rate": 2.071563088512241e-06, |
|
"loss": 1.9071, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.20326063942409486, |
|
"grad_norm": 3.453125, |
|
"learning_rate": 2.259887005649718e-06, |
|
"loss": 1.8766, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.22019902604276942, |
|
"grad_norm": 2.9375, |
|
"learning_rate": 2.448210922787194e-06, |
|
"loss": 1.8211, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.237137412661444, |
|
"grad_norm": 3.859375, |
|
"learning_rate": 2.6365348399246702e-06, |
|
"loss": 1.7678, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.2540757992801186, |
|
"grad_norm": 2.515625, |
|
"learning_rate": 2.8248587570621473e-06, |
|
"loss": 1.6984, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.2710141858987931, |
|
"grad_norm": 2.671875, |
|
"learning_rate": 3.0131826741996235e-06, |
|
"loss": 1.6916, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.2879525725174677, |
|
"grad_norm": 2.515625, |
|
"learning_rate": 3.2015065913371e-06, |
|
"loss": 1.7083, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.3048909591361423, |
|
"grad_norm": 2.609375, |
|
"learning_rate": 3.3898305084745763e-06, |
|
"loss": 1.7021, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.32182934575481686, |
|
"grad_norm": 3.109375, |
|
"learning_rate": 3.5781544256120534e-06, |
|
"loss": 1.654, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.33876773237349145, |
|
"grad_norm": 1.875, |
|
"learning_rate": 3.7664783427495296e-06, |
|
"loss": 1.6125, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.355706118992166, |
|
"grad_norm": 1.9765625, |
|
"learning_rate": 3.954802259887006e-06, |
|
"loss": 1.6243, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.37264450561084056, |
|
"grad_norm": 2.046875, |
|
"learning_rate": 4.143126177024482e-06, |
|
"loss": 1.6311, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.38958289222951514, |
|
"grad_norm": 2.375, |
|
"learning_rate": 4.331450094161959e-06, |
|
"loss": 1.6088, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.4065212788481897, |
|
"grad_norm": 1.9765625, |
|
"learning_rate": 4.519774011299436e-06, |
|
"loss": 1.6126, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.42345966546686425, |
|
"grad_norm": 2.0, |
|
"learning_rate": 4.7080979284369114e-06, |
|
"loss": 1.5862, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.44039805208553884, |
|
"grad_norm": 1.9296875, |
|
"learning_rate": 4.896421845574388e-06, |
|
"loss": 1.5869, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.4573364387042134, |
|
"grad_norm": 2.03125, |
|
"learning_rate": 5.084745762711865e-06, |
|
"loss": 1.568, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.474274825322888, |
|
"grad_norm": 2.234375, |
|
"learning_rate": 5.2730696798493405e-06, |
|
"loss": 1.4917, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.4912132119415626, |
|
"grad_norm": 3.0, |
|
"learning_rate": 5.461393596986818e-06, |
|
"loss": 1.5614, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.5081515985602372, |
|
"grad_norm": 2.09375, |
|
"learning_rate": 5.6497175141242946e-06, |
|
"loss": 1.5684, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.5250899851789117, |
|
"grad_norm": 1.765625, |
|
"learning_rate": 5.83804143126177e-06, |
|
"loss": 1.5933, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.5420283717975862, |
|
"grad_norm": 1.953125, |
|
"learning_rate": 6.026365348399247e-06, |
|
"loss": 1.5864, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.5589667584162609, |
|
"grad_norm": 1.9296875, |
|
"learning_rate": 6.2146892655367244e-06, |
|
"loss": 1.539, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.5759051450349354, |
|
"grad_norm": 2.125, |
|
"learning_rate": 6.4030131826742e-06, |
|
"loss": 1.4972, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.59284353165361, |
|
"grad_norm": 1.765625, |
|
"learning_rate": 6.591337099811677e-06, |
|
"loss": 1.5136, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.6097819182722846, |
|
"grad_norm": 1.921875, |
|
"learning_rate": 6.779661016949153e-06, |
|
"loss": 1.5159, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.6267203048909591, |
|
"grad_norm": 1.8125, |
|
"learning_rate": 6.96798493408663e-06, |
|
"loss": 1.4829, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.6436586915096337, |
|
"grad_norm": 1.9296875, |
|
"learning_rate": 7.156308851224107e-06, |
|
"loss": 1.4972, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.6605970781283083, |
|
"grad_norm": 2.140625, |
|
"learning_rate": 7.3446327683615825e-06, |
|
"loss": 1.4115, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.6775354647469829, |
|
"grad_norm": 2.296875, |
|
"learning_rate": 7.532956685499059e-06, |
|
"loss": 1.5167, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.6944738513656574, |
|
"grad_norm": 1.96875, |
|
"learning_rate": 7.721280602636536e-06, |
|
"loss": 1.4496, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.711412237984332, |
|
"grad_norm": 1.7734375, |
|
"learning_rate": 7.909604519774012e-06, |
|
"loss": 1.4258, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.7283506246030066, |
|
"grad_norm": 3.28125, |
|
"learning_rate": 8.097928436911489e-06, |
|
"loss": 1.498, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.7452890112216811, |
|
"grad_norm": 2.078125, |
|
"learning_rate": 8.286252354048965e-06, |
|
"loss": 1.4588, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.7622273978403558, |
|
"grad_norm": 2.03125, |
|
"learning_rate": 8.47457627118644e-06, |
|
"loss": 1.4085, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.7791657844590303, |
|
"grad_norm": 1.9765625, |
|
"learning_rate": 8.662900188323918e-06, |
|
"loss": 1.4479, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.7961041710777048, |
|
"grad_norm": 2.25, |
|
"learning_rate": 8.851224105461394e-06, |
|
"loss": 1.4711, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.8130425576963795, |
|
"grad_norm": 2.078125, |
|
"learning_rate": 9.039548022598871e-06, |
|
"loss": 1.4004, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.829980944315054, |
|
"grad_norm": 2.09375, |
|
"learning_rate": 9.227871939736347e-06, |
|
"loss": 1.3785, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.8469193309337285, |
|
"grad_norm": 1.765625, |
|
"learning_rate": 9.416195856873823e-06, |
|
"loss": 1.3784, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.8469193309337285, |
|
"eval_loss": 1.401244044303894, |
|
"eval_runtime": 21.2893, |
|
"eval_samples_per_second": 24.66, |
|
"eval_steps_per_second": 24.66, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.8638577175524031, |
|
"grad_norm": 1.96875, |
|
"learning_rate": 9.6045197740113e-06, |
|
"loss": 1.4401, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.8807961041710777, |
|
"grad_norm": 1.6640625, |
|
"learning_rate": 9.792843691148776e-06, |
|
"loss": 1.4191, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.8977344907897523, |
|
"grad_norm": 1.6015625, |
|
"learning_rate": 9.981167608286254e-06, |
|
"loss": 1.3972, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.9146728774084268, |
|
"grad_norm": 1.7421875, |
|
"learning_rate": 9.999912491660088e-06, |
|
"loss": 1.4051, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.9316112640271014, |
|
"grad_norm": 1.8984375, |
|
"learning_rate": 9.999609997627497e-06, |
|
"loss": 1.3855, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.948549650645776, |
|
"grad_norm": 1.765625, |
|
"learning_rate": 9.999091450621204e-06, |
|
"loss": 1.4033, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.9654880372644505, |
|
"grad_norm": 1.640625, |
|
"learning_rate": 9.99835687304968e-06, |
|
"loss": 1.4429, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.9824264238831252, |
|
"grad_norm": 1.8515625, |
|
"learning_rate": 9.997406296656937e-06, |
|
"loss": 1.3925, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.9993648105017997, |
|
"grad_norm": 1.78125, |
|
"learning_rate": 9.996239762521152e-06, |
|
"loss": 1.4055, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 1.0163031971204743, |
|
"grad_norm": 2.203125, |
|
"learning_rate": 9.994857321052887e-06, |
|
"loss": 1.2929, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.0332415837391489, |
|
"grad_norm": 1.8203125, |
|
"learning_rate": 9.993259031992921e-06, |
|
"loss": 1.2752, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 1.0501799703578234, |
|
"grad_norm": 1.59375, |
|
"learning_rate": 9.991444964409665e-06, |
|
"loss": 1.3491, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 1.067118356976498, |
|
"grad_norm": 1.8515625, |
|
"learning_rate": 9.989415196696169e-06, |
|
"loss": 1.3204, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 1.0840567435951725, |
|
"grad_norm": 1.625, |
|
"learning_rate": 9.98716981656675e-06, |
|
"loss": 1.3425, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 1.1009951302138472, |
|
"grad_norm": 1.921875, |
|
"learning_rate": 9.98470892105318e-06, |
|
"loss": 1.3116, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 1.1179335168325217, |
|
"grad_norm": 1.8203125, |
|
"learning_rate": 9.982032616500516e-06, |
|
"loss": 1.307, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 1.1348719034511963, |
|
"grad_norm": 1.875, |
|
"learning_rate": 9.979141018562494e-06, |
|
"loss": 1.3132, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 1.1518102900698708, |
|
"grad_norm": 1.84375, |
|
"learning_rate": 9.976034252196524e-06, |
|
"loss": 1.3304, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 1.1687486766885455, |
|
"grad_norm": 1.6796875, |
|
"learning_rate": 9.9727124516583e-06, |
|
"loss": 1.347, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 1.18568706330722, |
|
"grad_norm": 1.734375, |
|
"learning_rate": 9.969175760495997e-06, |
|
"loss": 1.2844, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.2026254499258946, |
|
"grad_norm": 1.6953125, |
|
"learning_rate": 9.965424331544064e-06, |
|
"loss": 1.2891, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 1.2195638365445691, |
|
"grad_norm": 1.828125, |
|
"learning_rate": 9.961458326916624e-06, |
|
"loss": 1.3258, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 1.2365022231632437, |
|
"grad_norm": 1.609375, |
|
"learning_rate": 9.95727791800046e-06, |
|
"loss": 1.2944, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 1.2534406097819182, |
|
"grad_norm": 1.6171875, |
|
"learning_rate": 9.952883285447622e-06, |
|
"loss": 1.2602, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 1.2703789964005927, |
|
"grad_norm": 1.796875, |
|
"learning_rate": 9.948274619167608e-06, |
|
"loss": 1.2353, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 1.2873173830192675, |
|
"grad_norm": 1.6953125, |
|
"learning_rate": 9.943452118319165e-06, |
|
"loss": 1.2749, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 1.304255769637942, |
|
"grad_norm": 1.765625, |
|
"learning_rate": 9.938415991301675e-06, |
|
"loss": 1.2432, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 1.3211941562566165, |
|
"grad_norm": 1.625, |
|
"learning_rate": 9.933166455746159e-06, |
|
"loss": 1.2824, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 1.3381325428752913, |
|
"grad_norm": 1.5703125, |
|
"learning_rate": 9.927703738505863e-06, |
|
"loss": 1.2335, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 1.3550709294939658, |
|
"grad_norm": 1.5625, |
|
"learning_rate": 9.922028075646465e-06, |
|
"loss": 1.3065, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.3720093161126403, |
|
"grad_norm": 1.5703125, |
|
"learning_rate": 9.91613971243586e-06, |
|
"loss": 1.2443, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 1.3889477027313148, |
|
"grad_norm": 1.578125, |
|
"learning_rate": 9.910038903333574e-06, |
|
"loss": 1.2303, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 1.4058860893499894, |
|
"grad_norm": 1.90625, |
|
"learning_rate": 9.90372591197976e-06, |
|
"loss": 1.1844, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 1.422824475968664, |
|
"grad_norm": 1.53125, |
|
"learning_rate": 9.89720101118381e-06, |
|
"loss": 1.2035, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 1.4397628625873384, |
|
"grad_norm": 1.421875, |
|
"learning_rate": 9.890464482912561e-06, |
|
"loss": 1.2399, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 1.4567012492060132, |
|
"grad_norm": 1.578125, |
|
"learning_rate": 9.88351661827811e-06, |
|
"loss": 1.227, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 1.4736396358246877, |
|
"grad_norm": 1.484375, |
|
"learning_rate": 9.876357717525245e-06, |
|
"loss": 1.2589, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 1.4905780224433622, |
|
"grad_norm": 1.84375, |
|
"learning_rate": 9.86898809001845e-06, |
|
"loss": 1.1935, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 1.507516409062037, |
|
"grad_norm": 1.5, |
|
"learning_rate": 9.861408054228556e-06, |
|
"loss": 1.2177, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 1.5244547956807115, |
|
"grad_norm": 1.3828125, |
|
"learning_rate": 9.853617937718966e-06, |
|
"loss": 1.2285, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 1.541393182299386, |
|
"grad_norm": 1.3515625, |
|
"learning_rate": 9.845618077131508e-06, |
|
"loss": 1.226, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 1.5583315689180606, |
|
"grad_norm": 1.421875, |
|
"learning_rate": 9.837408818171884e-06, |
|
"loss": 1.2536, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 1.575269955536735, |
|
"grad_norm": 1.2578125, |
|
"learning_rate": 9.82899051559472e-06, |
|
"loss": 1.2041, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 1.5922083421554096, |
|
"grad_norm": 1.515625, |
|
"learning_rate": 9.820363533188257e-06, |
|
"loss": 1.2279, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 1.6091467287740842, |
|
"grad_norm": 1.2421875, |
|
"learning_rate": 9.811528243758612e-06, |
|
"loss": 1.1829, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 1.6260851153927587, |
|
"grad_norm": 1.6015625, |
|
"learning_rate": 9.802485029113682e-06, |
|
"loss": 1.1698, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 1.6430235020114334, |
|
"grad_norm": 1.234375, |
|
"learning_rate": 9.793234280046627e-06, |
|
"loss": 1.2212, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 1.659961888630108, |
|
"grad_norm": 1.234375, |
|
"learning_rate": 9.783776396318998e-06, |
|
"loss": 1.1377, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 1.6769002752487827, |
|
"grad_norm": 1.3203125, |
|
"learning_rate": 9.77411178664346e-06, |
|
"loss": 1.2086, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 1.6938386618674572, |
|
"grad_norm": 1.3125, |
|
"learning_rate": 9.76424086866612e-06, |
|
"loss": 1.1764, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.6938386618674572, |
|
"eval_loss": 1.2227357625961304, |
|
"eval_runtime": 20.8763, |
|
"eval_samples_per_second": 25.148, |
|
"eval_steps_per_second": 25.148, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.7107770484861318, |
|
"grad_norm": 1.2421875, |
|
"learning_rate": 9.754164068948482e-06, |
|
"loss": 1.1537, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 1.7277154351048063, |
|
"grad_norm": 1.453125, |
|
"learning_rate": 9.743881822949025e-06, |
|
"loss": 1.2405, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 1.7446538217234808, |
|
"grad_norm": 1.1875, |
|
"learning_rate": 9.733394575004373e-06, |
|
"loss": 1.2403, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 1.7615922083421554, |
|
"grad_norm": 1.40625, |
|
"learning_rate": 9.722702778310096e-06, |
|
"loss": 1.152, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 1.7785305949608299, |
|
"grad_norm": 1.265625, |
|
"learning_rate": 9.711806894901124e-06, |
|
"loss": 1.1596, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 1.7954689815795044, |
|
"grad_norm": 1.25, |
|
"learning_rate": 9.700707395631787e-06, |
|
"loss": 1.1555, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 1.8124073681981792, |
|
"grad_norm": 1.171875, |
|
"learning_rate": 9.689404760155465e-06, |
|
"loss": 1.1611, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 1.8293457548168537, |
|
"grad_norm": 1.640625, |
|
"learning_rate": 9.677899476903857e-06, |
|
"loss": 1.1727, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 1.8462841414355282, |
|
"grad_norm": 1.2265625, |
|
"learning_rate": 9.666192043065875e-06, |
|
"loss": 1.1526, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 1.863222528054203, |
|
"grad_norm": 1.265625, |
|
"learning_rate": 9.654282964566159e-06, |
|
"loss": 1.1677, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 1.8801609146728775, |
|
"grad_norm": 1.1953125, |
|
"learning_rate": 9.642172756043217e-06, |
|
"loss": 1.1363, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 1.897099301291552, |
|
"grad_norm": 1.1953125, |
|
"learning_rate": 9.629861940827181e-06, |
|
"loss": 1.1216, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 1.9140376879102265, |
|
"grad_norm": 1.2109375, |
|
"learning_rate": 9.617351050917197e-06, |
|
"loss": 1.1432, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 1.930976074528901, |
|
"grad_norm": 1.0859375, |
|
"learning_rate": 9.604640626958428e-06, |
|
"loss": 1.1714, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 1.9479144611475756, |
|
"grad_norm": 1.21875, |
|
"learning_rate": 9.591731218218694e-06, |
|
"loss": 1.1999, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 1.9648528477662501, |
|
"grad_norm": 1.078125, |
|
"learning_rate": 9.57862338256474e-06, |
|
"loss": 1.1692, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 1.9817912343849249, |
|
"grad_norm": 1.125, |
|
"learning_rate": 9.565317686438122e-06, |
|
"loss": 1.1193, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 1.9987296210035994, |
|
"grad_norm": 0.92578125, |
|
"learning_rate": 9.551814704830734e-06, |
|
"loss": 1.1596, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 2.015668007622274, |
|
"grad_norm": 0.96875, |
|
"learning_rate": 9.538115021259956e-06, |
|
"loss": 1.0715, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 2.0326063942409487, |
|
"grad_norm": 0.9375, |
|
"learning_rate": 9.52421922774344e-06, |
|
"loss": 1.0312, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 2.049544780859623, |
|
"grad_norm": 0.9609375, |
|
"learning_rate": 9.510127924773529e-06, |
|
"loss": 1.0574, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 2.0664831674782977, |
|
"grad_norm": 0.95703125, |
|
"learning_rate": 9.4958417212913e-06, |
|
"loss": 1.0862, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 2.0834215540969723, |
|
"grad_norm": 0.96875, |
|
"learning_rate": 9.481361234660258e-06, |
|
"loss": 1.0776, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 2.100359940715647, |
|
"grad_norm": 0.95703125, |
|
"learning_rate": 9.466687090639652e-06, |
|
"loss": 1.088, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 2.1172983273343213, |
|
"grad_norm": 0.9296875, |
|
"learning_rate": 9.451819923357435e-06, |
|
"loss": 1.1108, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 2.134236713952996, |
|
"grad_norm": 0.9921875, |
|
"learning_rate": 9.436760375282858e-06, |
|
"loss": 1.0786, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 2.1511751005716704, |
|
"grad_norm": 0.87890625, |
|
"learning_rate": 9.421509097198715e-06, |
|
"loss": 1.0269, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 2.168113487190345, |
|
"grad_norm": 1.171875, |
|
"learning_rate": 9.406066748173208e-06, |
|
"loss": 1.1046, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 2.18505187380902, |
|
"grad_norm": 0.9140625, |
|
"learning_rate": 9.390433995531473e-06, |
|
"loss": 1.0651, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 2.2019902604276944, |
|
"grad_norm": 0.91796875, |
|
"learning_rate": 9.374611514826746e-06, |
|
"loss": 1.0212, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 2.218928647046369, |
|
"grad_norm": 1.03125, |
|
"learning_rate": 9.358599989811158e-06, |
|
"loss": 1.0786, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 2.2358670336650435, |
|
"grad_norm": 1.046875, |
|
"learning_rate": 9.3424001124062e-06, |
|
"loss": 1.0125, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 2.252805420283718, |
|
"grad_norm": 0.90234375, |
|
"learning_rate": 9.326012582672814e-06, |
|
"loss": 1.0756, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 2.2697438069023925, |
|
"grad_norm": 1.015625, |
|
"learning_rate": 9.309438108781143e-06, |
|
"loss": 1.1165, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 2.286682193521067, |
|
"grad_norm": 0.9609375, |
|
"learning_rate": 9.29267740697993e-06, |
|
"loss": 1.037, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 2.3036205801397416, |
|
"grad_norm": 1.25, |
|
"learning_rate": 9.275731201565564e-06, |
|
"loss": 1.03, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 2.320558966758416, |
|
"grad_norm": 0.93359375, |
|
"learning_rate": 9.258600224850777e-06, |
|
"loss": 1.0461, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 2.337497353377091, |
|
"grad_norm": 1.2421875, |
|
"learning_rate": 9.24128521713301e-06, |
|
"loss": 1.0065, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 2.3544357399957656, |
|
"grad_norm": 1.0234375, |
|
"learning_rate": 9.223786926662404e-06, |
|
"loss": 1.0661, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 2.37137412661444, |
|
"grad_norm": 1.0078125, |
|
"learning_rate": 9.20610610960948e-06, |
|
"loss": 1.0764, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 2.3883125132331147, |
|
"grad_norm": 1.015625, |
|
"learning_rate": 9.188243530032458e-06, |
|
"loss": 1.0448, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 2.405250899851789, |
|
"grad_norm": 0.984375, |
|
"learning_rate": 9.170199959844232e-06, |
|
"loss": 1.0695, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 2.4221892864704637, |
|
"grad_norm": 1.015625, |
|
"learning_rate": 9.15197617877902e-06, |
|
"loss": 0.9995, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 2.4391276730891382, |
|
"grad_norm": 1.015625, |
|
"learning_rate": 9.133572974358669e-06, |
|
"loss": 1.0307, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 2.4560660597078128, |
|
"grad_norm": 1.1328125, |
|
"learning_rate": 9.114991141858619e-06, |
|
"loss": 1.0004, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 2.4730044463264873, |
|
"grad_norm": 1.390625, |
|
"learning_rate": 9.096231484273535e-06, |
|
"loss": 1.0553, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 2.489942832945162, |
|
"grad_norm": 1.1328125, |
|
"learning_rate": 9.077294812282615e-06, |
|
"loss": 1.0481, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 2.5068812195638364, |
|
"grad_norm": 1.203125, |
|
"learning_rate": 9.058181944214547e-06, |
|
"loss": 1.0438, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 2.523819606182511, |
|
"grad_norm": 1.0234375, |
|
"learning_rate": 9.038893706012153e-06, |
|
"loss": 1.0458, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 2.5407579928011854, |
|
"grad_norm": 1.4765625, |
|
"learning_rate": 9.01943093119669e-06, |
|
"loss": 0.9808, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 2.5407579928011854, |
|
"eval_loss": 1.1499994993209839, |
|
"eval_runtime": 21.1383, |
|
"eval_samples_per_second": 24.836, |
|
"eval_steps_per_second": 24.836, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 2.5576963794198604, |
|
"grad_norm": 1.4609375, |
|
"learning_rate": 8.999794460831836e-06, |
|
"loss": 1.0539, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 2.574634766038535, |
|
"grad_norm": 1.6640625, |
|
"learning_rate": 8.979985143487343e-06, |
|
"loss": 1.046, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 2.5915731526572094, |
|
"grad_norm": 1.4140625, |
|
"learning_rate": 8.960003835202369e-06, |
|
"loss": 1.0188, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 2.608511539275884, |
|
"grad_norm": 1.453125, |
|
"learning_rate": 8.939851399448478e-06, |
|
"loss": 1.0423, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 2.6254499258945585, |
|
"grad_norm": 2.921875, |
|
"learning_rate": 8.919528707092333e-06, |
|
"loss": 1.085, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 2.642388312513233, |
|
"grad_norm": 2.859375, |
|
"learning_rate": 8.899036636358058e-06, |
|
"loss": 1.0252, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 2.6593266991319076, |
|
"grad_norm": 3.125, |
|
"learning_rate": 8.878376072789292e-06, |
|
"loss": 1.0246, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 2.6762650857505825, |
|
"grad_norm": 2.78125, |
|
"learning_rate": 8.857547909210913e-06, |
|
"loss": 1.0358, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 2.693203472369257, |
|
"grad_norm": 2.59375, |
|
"learning_rate": 8.836553045690462e-06, |
|
"loss": 1.0746, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 2.7101418589879316, |
|
"grad_norm": 3.96875, |
|
"learning_rate": 8.81539238949925e-06, |
|
"loss": 1.0806, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 2.727080245606606, |
|
"grad_norm": 4.15625, |
|
"learning_rate": 8.794066855073136e-06, |
|
"loss": 0.9917, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 2.7440186322252806, |
|
"grad_norm": 4.5625, |
|
"learning_rate": 8.772577363973032e-06, |
|
"loss": 1.0071, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 2.760957018843955, |
|
"grad_norm": 3.25, |
|
"learning_rate": 8.750924844845065e-06, |
|
"loss": 1.0139, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 2.7778954054626297, |
|
"grad_norm": 4.25, |
|
"learning_rate": 8.729110233380445e-06, |
|
"loss": 1.0684, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 2.7948337920813042, |
|
"grad_norm": 9.125, |
|
"learning_rate": 8.707134472275042e-06, |
|
"loss": 0.9606, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 2.8117721786999788, |
|
"grad_norm": 7.78125, |
|
"learning_rate": 8.684998511188634e-06, |
|
"loss": 1.0529, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 2.8287105653186533, |
|
"grad_norm": 7.34375, |
|
"learning_rate": 8.662703306703878e-06, |
|
"loss": 1.0939, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 2.845648951937328, |
|
"grad_norm": 8.375, |
|
"learning_rate": 8.640249822284975e-06, |
|
"loss": 1.0322, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 2.8625873385560023, |
|
"grad_norm": 7.40625, |
|
"learning_rate": 8.617639028236021e-06, |
|
"loss": 1.0303, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 2.879525725174677, |
|
"grad_norm": 2.625, |
|
"learning_rate": 8.59487190165909e-06, |
|
"loss": 1.0173, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 2.896464111793352, |
|
"grad_norm": 2.328125, |
|
"learning_rate": 8.571949426412002e-06, |
|
"loss": 1.0218, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 2.9134024984120264, |
|
"grad_norm": 2.140625, |
|
"learning_rate": 8.54887259306581e-06, |
|
"loss": 1.0029, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 2.930340885030701, |
|
"grad_norm": 2.28125, |
|
"learning_rate": 8.525642398861997e-06, |
|
"loss": 1.0295, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 2.9472792716493754, |
|
"grad_norm": 2.5, |
|
"learning_rate": 8.50225984766937e-06, |
|
"loss": 1.0432, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 2.96421765826805, |
|
"grad_norm": 2.0625, |
|
"learning_rate": 8.478725949940687e-06, |
|
"loss": 1.026, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 2.9811560448867245, |
|
"grad_norm": 2.265625, |
|
"learning_rate": 8.455041722668993e-06, |
|
"loss": 1.0523, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 2.998094431505399, |
|
"grad_norm": 1.8671875, |
|
"learning_rate": 8.43120818934367e-06, |
|
"loss": 1.0287, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 3.0150328181240735, |
|
"grad_norm": 2.109375, |
|
"learning_rate": 8.407226379906197e-06, |
|
"loss": 0.9998, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 3.031971204742748, |
|
"grad_norm": 2.140625, |
|
"learning_rate": 8.383097330705664e-06, |
|
"loss": 1.0157, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 3.048909591361423, |
|
"grad_norm": 1.8828125, |
|
"learning_rate": 8.358822084453964e-06, |
|
"loss": 0.9514, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 3.0658479779800976, |
|
"grad_norm": 1.796875, |
|
"learning_rate": 8.334401690180755e-06, |
|
"loss": 1.0163, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 3.082786364598772, |
|
"grad_norm": 1.8828125, |
|
"learning_rate": 8.30983720318811e-06, |
|
"loss": 0.969, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 3.0997247512174466, |
|
"grad_norm": 1.8203125, |
|
"learning_rate": 8.285129685004922e-06, |
|
"loss": 0.9769, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 3.116663137836121, |
|
"grad_norm": 2.09375, |
|
"learning_rate": 8.260280203341026e-06, |
|
"loss": 0.9997, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 3.1336015244547957, |
|
"grad_norm": 1.828125, |
|
"learning_rate": 8.235289832041066e-06, |
|
"loss": 0.9577, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 3.15053991107347, |
|
"grad_norm": 2.4375, |
|
"learning_rate": 8.210159651038081e-06, |
|
"loss": 0.9457, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 3.1674782976921447, |
|
"grad_norm": 1.7734375, |
|
"learning_rate": 8.184890746306848e-06, |
|
"loss": 0.9381, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 3.1844166843108193, |
|
"grad_norm": 1.96875, |
|
"learning_rate": 8.15948420981694e-06, |
|
"loss": 0.9709, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 3.201355070929494, |
|
"grad_norm": 1.8125, |
|
"learning_rate": 8.133941139485551e-06, |
|
"loss": 0.9769, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 3.2182934575481683, |
|
"grad_norm": 1.875, |
|
"learning_rate": 8.108262639130038e-06, |
|
"loss": 0.9064, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 3.2352318441668433, |
|
"grad_norm": 1.71875, |
|
"learning_rate": 8.082449818420227e-06, |
|
"loss": 0.9944, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 3.252170230785518, |
|
"grad_norm": 2.125, |
|
"learning_rate": 8.056503792830458e-06, |
|
"loss": 0.9484, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 3.2691086174041923, |
|
"grad_norm": 1.8046875, |
|
"learning_rate": 8.030425683591391e-06, |
|
"loss": 0.9666, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 3.286047004022867, |
|
"grad_norm": 1.890625, |
|
"learning_rate": 8.004216617641532e-06, |
|
"loss": 0.9871, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 3.3029853906415414, |
|
"grad_norm": 1.765625, |
|
"learning_rate": 7.977877727578556e-06, |
|
"loss": 0.9462, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 3.319923777260216, |
|
"grad_norm": 2.09375, |
|
"learning_rate": 7.951410151610343e-06, |
|
"loss": 0.9509, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 3.3368621638788905, |
|
"grad_norm": 1.7734375, |
|
"learning_rate": 7.92481503350581e-06, |
|
"loss": 0.9299, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 3.353800550497565, |
|
"grad_norm": 1.734375, |
|
"learning_rate": 7.898093522545471e-06, |
|
"loss": 0.9824, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 3.3707389371162395, |
|
"grad_norm": 1.78125, |
|
"learning_rate": 7.871246773471779e-06, |
|
"loss": 0.9336, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 3.3876773237349145, |
|
"grad_norm": 2.015625, |
|
"learning_rate": 7.844275946439216e-06, |
|
"loss": 0.9778, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 3.3876773237349145, |
|
"eval_loss": 1.1204777956008911, |
|
"eval_runtime": 20.947, |
|
"eval_samples_per_second": 25.063, |
|
"eval_steps_per_second": 25.063, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 3.404615710353589, |
|
"grad_norm": 2.359375, |
|
"learning_rate": 7.817182206964177e-06, |
|
"loss": 0.9817, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 3.4215540969722635, |
|
"grad_norm": 2.0625, |
|
"learning_rate": 7.789966725874578e-06, |
|
"loss": 0.9942, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 3.438492483590938, |
|
"grad_norm": 1.734375, |
|
"learning_rate": 7.76263067925928e-06, |
|
"loss": 0.9783, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 3.4554308702096126, |
|
"grad_norm": 1.796875, |
|
"learning_rate": 7.735175248417255e-06, |
|
"loss": 0.9551, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 3.472369256828287, |
|
"grad_norm": 2.046875, |
|
"learning_rate": 7.707601619806547e-06, |
|
"loss": 0.9592, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 3.4893076434469616, |
|
"grad_norm": 2.40625, |
|
"learning_rate": 7.679910984992988e-06, |
|
"loss": 0.902, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 3.506246030065636, |
|
"grad_norm": 2.203125, |
|
"learning_rate": 7.652104540598712e-06, |
|
"loss": 0.9763, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 3.5231844166843107, |
|
"grad_norm": 2.0625, |
|
"learning_rate": 7.6241834882504516e-06, |
|
"loss": 0.9522, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 3.5401228033029852, |
|
"grad_norm": 2.140625, |
|
"learning_rate": 7.596149034527594e-06, |
|
"loss": 0.992, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 3.5570611899216598, |
|
"grad_norm": 1.859375, |
|
"learning_rate": 7.5680023909100545e-06, |
|
"loss": 0.9366, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 3.5739995765403343, |
|
"grad_norm": 1.7421875, |
|
"learning_rate": 7.539744773725916e-06, |
|
"loss": 0.9888, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 3.5909379631590093, |
|
"grad_norm": 1.71875, |
|
"learning_rate": 7.511377404098876e-06, |
|
"loss": 0.9801, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 3.607876349777684, |
|
"grad_norm": 2.0, |
|
"learning_rate": 7.482901507895463e-06, |
|
"loss": 0.9589, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 3.6248147363963583, |
|
"grad_norm": 3.65625, |
|
"learning_rate": 7.454318315672074e-06, |
|
"loss": 0.9926, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 3.641753123015033, |
|
"grad_norm": 1.859375, |
|
"learning_rate": 7.425629062621784e-06, |
|
"loss": 0.9448, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 3.6586915096337074, |
|
"grad_norm": 2.140625, |
|
"learning_rate": 7.396834988520989e-06, |
|
"loss": 0.9507, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 3.675629896252382, |
|
"grad_norm": 2.125, |
|
"learning_rate": 7.36793733767581e-06, |
|
"loss": 0.9719, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 3.6925682828710564, |
|
"grad_norm": 2.34375, |
|
"learning_rate": 7.338937358868334e-06, |
|
"loss": 0.976, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 3.709506669489731, |
|
"grad_norm": 2.28125, |
|
"learning_rate": 7.309836305302638e-06, |
|
"loss": 0.9431, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 3.726445056108406, |
|
"grad_norm": 1.734375, |
|
"learning_rate": 7.2806354345506525e-06, |
|
"loss": 0.9833, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 3.7433834427270805, |
|
"grad_norm": 1.859375, |
|
"learning_rate": 7.251336008497793e-06, |
|
"loss": 0.9356, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 3.760321829345755, |
|
"grad_norm": 1.8203125, |
|
"learning_rate": 7.221939293288445e-06, |
|
"loss": 0.9911, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 3.7772602159644295, |
|
"grad_norm": 1.8671875, |
|
"learning_rate": 7.192446559271247e-06, |
|
"loss": 0.9569, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 3.794198602583104, |
|
"grad_norm": 1.875, |
|
"learning_rate": 7.162859080944182e-06, |
|
"loss": 0.9907, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 3.8111369892017786, |
|
"grad_norm": 1.6484375, |
|
"learning_rate": 7.133178136899522e-06, |
|
"loss": 0.9672, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 3.828075375820453, |
|
"grad_norm": 2.359375, |
|
"learning_rate": 7.103405009768554e-06, |
|
"loss": 0.9754, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 3.8450137624391276, |
|
"grad_norm": 1.7421875, |
|
"learning_rate": 7.073540986166166e-06, |
|
"loss": 0.9548, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 3.861952149057802, |
|
"grad_norm": 1.7421875, |
|
"learning_rate": 7.0435873566352434e-06, |
|
"loss": 0.9682, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 3.8788905356764767, |
|
"grad_norm": 1.8515625, |
|
"learning_rate": 7.013545415590897e-06, |
|
"loss": 0.8956, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 3.895828922295151, |
|
"grad_norm": 1.96875, |
|
"learning_rate": 6.983416461264527e-06, |
|
"loss": 0.9314, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 3.9127673089138257, |
|
"grad_norm": 1.890625, |
|
"learning_rate": 6.953201795647728e-06, |
|
"loss": 0.9101, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 3.9297056955325003, |
|
"grad_norm": 1.921875, |
|
"learning_rate": 6.922902724436016e-06, |
|
"loss": 0.9463, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 3.9466440821511752, |
|
"grad_norm": 1.8984375, |
|
"learning_rate": 6.892520556972404e-06, |
|
"loss": 0.9443, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 3.9635824687698498, |
|
"grad_norm": 1.8125, |
|
"learning_rate": 6.86205660619083e-06, |
|
"loss": 0.9371, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 3.9805208553885243, |
|
"grad_norm": 1.9375, |
|
"learning_rate": 6.8315121885594105e-06, |
|
"loss": 0.95, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 3.997459242007199, |
|
"grad_norm": 1.9140625, |
|
"learning_rate": 6.800888624023552e-06, |
|
"loss": 0.9412, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 4.014397628625874, |
|
"grad_norm": 1.59375, |
|
"learning_rate": 6.770187235948916e-06, |
|
"loss": 0.8627, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 4.031336015244548, |
|
"grad_norm": 1.671875, |
|
"learning_rate": 6.739409351064225e-06, |
|
"loss": 0.8663, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 4.048274401863223, |
|
"grad_norm": 1.8046875, |
|
"learning_rate": 6.708556299403935e-06, |
|
"loss": 0.8087, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 4.065212788481897, |
|
"grad_norm": 1.7578125, |
|
"learning_rate": 6.6776294142507535e-06, |
|
"loss": 0.8501, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 4.082151175100572, |
|
"grad_norm": 1.6328125, |
|
"learning_rate": 6.646630032078024e-06, |
|
"loss": 0.854, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 4.099089561719246, |
|
"grad_norm": 1.5, |
|
"learning_rate": 6.615559492491981e-06, |
|
"loss": 0.884, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 4.116027948337921, |
|
"grad_norm": 1.890625, |
|
"learning_rate": 6.584419138173846e-06, |
|
"loss": 0.8693, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 4.1329663349565955, |
|
"grad_norm": 1.625, |
|
"learning_rate": 6.553210314821814e-06, |
|
"loss": 0.8397, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 4.14990472157527, |
|
"grad_norm": 1.7265625, |
|
"learning_rate": 6.521934371092901e-06, |
|
"loss": 0.8355, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 4.1668431081939445, |
|
"grad_norm": 1.5, |
|
"learning_rate": 6.490592658544661e-06, |
|
"loss": 0.8784, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 4.183781494812619, |
|
"grad_norm": 1.484375, |
|
"learning_rate": 6.459186531576772e-06, |
|
"loss": 0.8902, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 4.200719881431294, |
|
"grad_norm": 1.640625, |
|
"learning_rate": 6.427717347372523e-06, |
|
"loss": 0.8175, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 4.217658268049968, |
|
"grad_norm": 1.7109375, |
|
"learning_rate": 6.396186465840152e-06, |
|
"loss": 0.8717, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 4.234596654668643, |
|
"grad_norm": 1.5234375, |
|
"learning_rate": 6.364595249554084e-06, |
|
"loss": 0.8815, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 4.234596654668643, |
|
"eval_loss": 1.09398353099823, |
|
"eval_runtime": 20.9341, |
|
"eval_samples_per_second": 25.079, |
|
"eval_steps_per_second": 25.079, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 4.251535041287317, |
|
"grad_norm": 1.4296875, |
|
"learning_rate": 6.332945063696049e-06, |
|
"loss": 0.8658, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 4.268473427905992, |
|
"grad_norm": 1.3828125, |
|
"learning_rate": 6.3012372759960825e-06, |
|
"loss": 0.8465, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 4.285411814524666, |
|
"grad_norm": 1.5234375, |
|
"learning_rate": 6.26947325667343e-06, |
|
"loss": 0.8707, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 4.302350201143341, |
|
"grad_norm": 1.4140625, |
|
"learning_rate": 6.237654378377324e-06, |
|
"loss": 0.8921, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 4.319288587762015, |
|
"grad_norm": 1.5390625, |
|
"learning_rate": 6.2057820161276715e-06, |
|
"loss": 0.8703, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 4.33622697438069, |
|
"grad_norm": 1.2734375, |
|
"learning_rate": 6.173857547255634e-06, |
|
"loss": 0.8467, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 4.353165360999364, |
|
"grad_norm": 1.578125, |
|
"learning_rate": 6.141882351344106e-06, |
|
"loss": 0.8581, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 4.37010374761804, |
|
"grad_norm": 1.4609375, |
|
"learning_rate": 6.1098578101680986e-06, |
|
"loss": 0.8617, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 4.387042134236714, |
|
"grad_norm": 1.3046875, |
|
"learning_rate": 6.077785307635027e-06, |
|
"loss": 0.8082, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 4.403980520855389, |
|
"grad_norm": 1.34375, |
|
"learning_rate": 6.045666229724909e-06, |
|
"loss": 0.8048, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 4.420918907474063, |
|
"grad_norm": 1.4453125, |
|
"learning_rate": 6.013501964430468e-06, |
|
"loss": 0.8537, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 4.437857294092738, |
|
"grad_norm": 1.3359375, |
|
"learning_rate": 5.981293901697153e-06, |
|
"loss": 0.8683, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 4.454795680711412, |
|
"grad_norm": 1.1953125, |
|
"learning_rate": 5.9490434333630735e-06, |
|
"loss": 0.8599, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 4.471734067330087, |
|
"grad_norm": 1.3203125, |
|
"learning_rate": 5.916751953098854e-06, |
|
"loss": 0.8811, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 4.4886724539487615, |
|
"grad_norm": 1.21875, |
|
"learning_rate": 5.884420856347406e-06, |
|
"loss": 0.8891, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 4.505610840567436, |
|
"grad_norm": 1.3515625, |
|
"learning_rate": 5.852051540263628e-06, |
|
"loss": 0.8634, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 4.5225492271861105, |
|
"grad_norm": 1.28125, |
|
"learning_rate": 5.819645403654027e-06, |
|
"loss": 0.842, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 4.539487613804785, |
|
"grad_norm": 1.6484375, |
|
"learning_rate": 5.787203846916273e-06, |
|
"loss": 0.8485, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 4.55642600042346, |
|
"grad_norm": 1.265625, |
|
"learning_rate": 5.754728271978675e-06, |
|
"loss": 0.883, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 4.573364387042134, |
|
"grad_norm": 1.2265625, |
|
"learning_rate": 5.722220082239608e-06, |
|
"loss": 0.8457, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 4.590302773660809, |
|
"grad_norm": 1.1640625, |
|
"learning_rate": 5.689680682506861e-06, |
|
"loss": 0.8999, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 4.607241160279483, |
|
"grad_norm": 1.171875, |
|
"learning_rate": 5.657111478936926e-06, |
|
"loss": 0.8391, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 4.624179546898158, |
|
"grad_norm": 1.2265625, |
|
"learning_rate": 5.6245138789742444e-06, |
|
"loss": 0.8313, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 4.641117933516832, |
|
"grad_norm": 1.1640625, |
|
"learning_rate": 5.591889291290373e-06, |
|
"loss": 0.8709, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 4.658056320135507, |
|
"grad_norm": 0.9921875, |
|
"learning_rate": 5.559239125723119e-06, |
|
"loss": 0.8507, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 4.674994706754182, |
|
"grad_norm": 0.96484375, |
|
"learning_rate": 5.5265647932156066e-06, |
|
"loss": 0.9057, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 4.691933093372857, |
|
"grad_norm": 1.1171875, |
|
"learning_rate": 5.493867705755312e-06, |
|
"loss": 0.8386, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 4.708871479991531, |
|
"grad_norm": 1.0, |
|
"learning_rate": 5.461149276313046e-06, |
|
"loss": 0.9338, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 4.725809866610206, |
|
"grad_norm": 1.0234375, |
|
"learning_rate": 5.428410918781885e-06, |
|
"loss": 0.8176, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 4.74274825322888, |
|
"grad_norm": 1.046875, |
|
"learning_rate": 5.395654047916083e-06, |
|
"loss": 0.8609, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 4.759686639847555, |
|
"grad_norm": 1.09375, |
|
"learning_rate": 5.362880079269921e-06, |
|
"loss": 0.8494, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 4.776625026466229, |
|
"grad_norm": 1.1953125, |
|
"learning_rate": 5.330090429136553e-06, |
|
"loss": 0.8254, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 4.793563413084904, |
|
"grad_norm": 1.2890625, |
|
"learning_rate": 5.297286514486785e-06, |
|
"loss": 0.8689, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 4.810501799703578, |
|
"grad_norm": 1.125, |
|
"learning_rate": 5.26446975290785e-06, |
|
"loss": 0.8483, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 4.827440186322253, |
|
"grad_norm": 0.98046875, |
|
"learning_rate": 5.23164156254215e-06, |
|
"loss": 0.8436, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 4.844378572940927, |
|
"grad_norm": 0.91796875, |
|
"learning_rate": 5.19880336202597e-06, |
|
"loss": 0.8584, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 4.861316959559602, |
|
"grad_norm": 1.2265625, |
|
"learning_rate": 5.165956570428173e-06, |
|
"loss": 0.8122, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 4.8782553461782765, |
|
"grad_norm": 1.0703125, |
|
"learning_rate": 5.133102607188875e-06, |
|
"loss": 0.8705, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 4.895193732796951, |
|
"grad_norm": 1.34375, |
|
"learning_rate": 5.100242892058108e-06, |
|
"loss": 0.8788, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 4.9121321194156256, |
|
"grad_norm": 1.015625, |
|
"learning_rate": 5.067378845034468e-06, |
|
"loss": 0.8663, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 4.9290705060343, |
|
"grad_norm": 1.046875, |
|
"learning_rate": 5.034511886303746e-06, |
|
"loss": 0.8706, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 4.946008892652975, |
|
"grad_norm": 1.1171875, |
|
"learning_rate": 5.00164343617756e-06, |
|
"loss": 0.845, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 4.962947279271649, |
|
"grad_norm": 1.0078125, |
|
"learning_rate": 4.968774915031982e-06, |
|
"loss": 0.8165, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 4.979885665890324, |
|
"grad_norm": 1.3125, |
|
"learning_rate": 4.9359077432461455e-06, |
|
"loss": 0.8719, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 4.996824052508998, |
|
"grad_norm": 0.88671875, |
|
"learning_rate": 4.903043341140879e-06, |
|
"loss": 0.8579, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 5.013762439127673, |
|
"grad_norm": 0.921875, |
|
"learning_rate": 4.870183128917318e-06, |
|
"loss": 0.7986, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 5.030700825746347, |
|
"grad_norm": 1.140625, |
|
"learning_rate": 4.837328526595539e-06, |
|
"loss": 0.8119, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 5.047639212365023, |
|
"grad_norm": 0.97265625, |
|
"learning_rate": 4.804480953953193e-06, |
|
"loss": 0.8013, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 5.064577598983697, |
|
"grad_norm": 0.9296875, |
|
"learning_rate": 4.771641830464146e-06, |
|
"loss": 0.7954, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 5.081515985602372, |
|
"grad_norm": 1.046875, |
|
"learning_rate": 4.738812575237148e-06, |
|
"loss": 0.8159, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 5.081515985602372, |
|
"eval_loss": 1.0747902393341064, |
|
"eval_runtime": 20.9737, |
|
"eval_samples_per_second": 25.031, |
|
"eval_steps_per_second": 25.031, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 5.098454372221046, |
|
"grad_norm": 1.0625, |
|
"learning_rate": 4.705994606954497e-06, |
|
"loss": 0.7888, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 5.115392758839721, |
|
"grad_norm": 1.09375, |
|
"learning_rate": 4.673189343810747e-06, |
|
"loss": 0.8088, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 5.132331145458395, |
|
"grad_norm": 1.0, |
|
"learning_rate": 4.640398203451402e-06, |
|
"loss": 0.7762, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 5.14926953207707, |
|
"grad_norm": 1.078125, |
|
"learning_rate": 4.607622602911668e-06, |
|
"loss": 0.8108, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 5.166207918695744, |
|
"grad_norm": 1.1171875, |
|
"learning_rate": 4.574863958555215e-06, |
|
"loss": 0.7981, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 5.183146305314419, |
|
"grad_norm": 1.1953125, |
|
"learning_rate": 4.542123686012969e-06, |
|
"loss": 0.8295, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 5.200084691933093, |
|
"grad_norm": 1.1875, |
|
"learning_rate": 4.509403200121936e-06, |
|
"loss": 0.7986, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 5.217023078551768, |
|
"grad_norm": 1.1875, |
|
"learning_rate": 4.47670391486406e-06, |
|
"loss": 0.792, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 5.2339614651704425, |
|
"grad_norm": 1.1484375, |
|
"learning_rate": 4.44402724330513e-06, |
|
"loss": 0.8072, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 5.250899851789117, |
|
"grad_norm": 1.3203125, |
|
"learning_rate": 4.4113745975337e-06, |
|
"loss": 0.7877, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 5.2678382384077915, |
|
"grad_norm": 1.4921875, |
|
"learning_rate": 4.378747388600075e-06, |
|
"loss": 0.7823, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 5.284776625026466, |
|
"grad_norm": 1.40625, |
|
"learning_rate": 4.346147026455334e-06, |
|
"loss": 0.7957, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 5.301715011645141, |
|
"grad_norm": 1.390625, |
|
"learning_rate": 4.313574919890402e-06, |
|
"loss": 0.7562, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 5.318653398263815, |
|
"grad_norm": 1.3671875, |
|
"learning_rate": 4.281032476475165e-06, |
|
"loss": 0.7981, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 5.33559178488249, |
|
"grad_norm": 2.265625, |
|
"learning_rate": 4.248521102497649e-06, |
|
"loss": 0.7806, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 5.352530171501164, |
|
"grad_norm": 2.578125, |
|
"learning_rate": 4.216042202903246e-06, |
|
"loss": 0.848, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 5.369468558119839, |
|
"grad_norm": 2.171875, |
|
"learning_rate": 4.183597181234001e-06, |
|
"loss": 0.7903, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 5.386406944738514, |
|
"grad_norm": 2.515625, |
|
"learning_rate": 4.151187439567961e-06, |
|
"loss": 0.7914, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 5.403345331357189, |
|
"grad_norm": 2.25, |
|
"learning_rate": 4.118814378458582e-06, |
|
"loss": 0.7983, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 5.420283717975863, |
|
"grad_norm": 3.09375, |
|
"learning_rate": 4.086479396874206e-06, |
|
"loss": 0.7606, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 5.437222104594538, |
|
"grad_norm": 2.828125, |
|
"learning_rate": 4.054183892137611e-06, |
|
"loss": 0.7947, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 5.454160491213212, |
|
"grad_norm": 4.375, |
|
"learning_rate": 4.021929259865624e-06, |
|
"loss": 0.7764, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 5.471098877831887, |
|
"grad_norm": 4.46875, |
|
"learning_rate": 3.989716893908807e-06, |
|
"loss": 0.7829, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 5.488037264450561, |
|
"grad_norm": 3.5625, |
|
"learning_rate": 3.9575481862912335e-06, |
|
"loss": 0.7811, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 5.504975651069236, |
|
"grad_norm": 8.125, |
|
"learning_rate": 3.925424527150321e-06, |
|
"loss": 0.804, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 5.52191403768791, |
|
"grad_norm": 7.96875, |
|
"learning_rate": 3.893347304676766e-06, |
|
"loss": 0.7975, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 5.538852424306585, |
|
"grad_norm": 7.6875, |
|
"learning_rate": 3.861317905054551e-06, |
|
"loss": 0.7812, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 5.555790810925259, |
|
"grad_norm": 8.0625, |
|
"learning_rate": 3.8293377124010476e-06, |
|
"loss": 0.8092, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 5.572729197543934, |
|
"grad_norm": 8.0625, |
|
"learning_rate": 3.797408108707193e-06, |
|
"loss": 0.8141, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 5.5896675841626084, |
|
"grad_norm": 2.125, |
|
"learning_rate": 3.7655304737777775e-06, |
|
"loss": 0.815, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 5.606605970781283, |
|
"grad_norm": 2.109375, |
|
"learning_rate": 3.733706185171819e-06, |
|
"loss": 0.7505, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 5.6235443573999575, |
|
"grad_norm": 1.96875, |
|
"learning_rate": 3.701936618143025e-06, |
|
"loss": 0.7535, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 5.640482744018632, |
|
"grad_norm": 2.46875, |
|
"learning_rate": 3.6702231455803662e-06, |
|
"loss": 0.7703, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 5.657421130637307, |
|
"grad_norm": 2.15625, |
|
"learning_rate": 3.6385671379487496e-06, |
|
"loss": 0.7493, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 5.674359517255981, |
|
"grad_norm": 1.875, |
|
"learning_rate": 3.606969963229798e-06, |
|
"loss": 0.8281, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 5.691297903874656, |
|
"grad_norm": 1.9140625, |
|
"learning_rate": 3.5754329868627256e-06, |
|
"loss": 0.8207, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 5.70823629049333, |
|
"grad_norm": 2.046875, |
|
"learning_rate": 3.5439575716853414e-06, |
|
"loss": 0.8263, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 5.725174677112005, |
|
"grad_norm": 2.015625, |
|
"learning_rate": 3.5125450778751453e-06, |
|
"loss": 0.8106, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 5.742113063730679, |
|
"grad_norm": 2.125, |
|
"learning_rate": 3.4811968628905646e-06, |
|
"loss": 0.7993, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 5.759051450349354, |
|
"grad_norm": 1.84375, |
|
"learning_rate": 3.4499142814122766e-06, |
|
"loss": 0.7784, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 5.775989836968029, |
|
"grad_norm": 1.7890625, |
|
"learning_rate": 3.418698685284676e-06, |
|
"loss": 0.7762, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 5.792928223586704, |
|
"grad_norm": 1.8984375, |
|
"learning_rate": 3.387551423457456e-06, |
|
"loss": 0.7905, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 5.809866610205378, |
|
"grad_norm": 1.9765625, |
|
"learning_rate": 3.356473841927313e-06, |
|
"loss": 0.8273, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 5.826804996824053, |
|
"grad_norm": 2.046875, |
|
"learning_rate": 3.3254672836797814e-06, |
|
"loss": 0.7962, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 5.843743383442727, |
|
"grad_norm": 2.03125, |
|
"learning_rate": 3.2945330886311964e-06, |
|
"loss": 0.8146, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 5.860681770061402, |
|
"grad_norm": 1.9765625, |
|
"learning_rate": 3.263672593570797e-06, |
|
"loss": 0.8032, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 5.877620156680076, |
|
"grad_norm": 1.84375, |
|
"learning_rate": 3.23288713210295e-06, |
|
"loss": 0.7997, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 5.894558543298751, |
|
"grad_norm": 1.8828125, |
|
"learning_rate": 3.2021780345895216e-06, |
|
"loss": 0.8009, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 5.911496929917425, |
|
"grad_norm": 1.953125, |
|
"learning_rate": 3.1715466280923924e-06, |
|
"loss": 0.8302, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 5.9284353165361, |
|
"grad_norm": 2.359375, |
|
"learning_rate": 3.1409942363161037e-06, |
|
"loss": 0.8317, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 5.9284353165361, |
|
"eval_loss": 1.0828534364700317, |
|
"eval_runtime": 21.0141, |
|
"eval_samples_per_second": 24.983, |
|
"eval_steps_per_second": 24.983, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 5.945373703154774, |
|
"grad_norm": 2.125, |
|
"learning_rate": 3.1105221795506584e-06, |
|
"loss": 0.8026, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 5.962312089773449, |
|
"grad_norm": 2.046875, |
|
"learning_rate": 3.080131774614464e-06, |
|
"loss": 0.8209, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 5.9792504763921235, |
|
"grad_norm": 1.8515625, |
|
"learning_rate": 3.049824334797431e-06, |
|
"loss": 0.8086, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 5.996188863010798, |
|
"grad_norm": 2.015625, |
|
"learning_rate": 3.019601169804216e-06, |
|
"loss": 0.7651, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 6.0131272496294725, |
|
"grad_norm": 1.703125, |
|
"learning_rate": 2.9894635856976268e-06, |
|
"loss": 0.7734, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 6.030065636248147, |
|
"grad_norm": 1.8671875, |
|
"learning_rate": 2.959412884842181e-06, |
|
"loss": 0.7799, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 6.047004022866822, |
|
"grad_norm": 1.921875, |
|
"learning_rate": 2.929450365847828e-06, |
|
"loss": 0.7448, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 6.063942409485496, |
|
"grad_norm": 2.0, |
|
"learning_rate": 2.899577323513829e-06, |
|
"loss": 0.7354, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 6.080880796104171, |
|
"grad_norm": 2.109375, |
|
"learning_rate": 2.8697950487728e-06, |
|
"loss": 0.7585, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 6.097819182722846, |
|
"grad_norm": 1.828125, |
|
"learning_rate": 2.8401048286349353e-06, |
|
"loss": 0.7635, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 6.114757569341521, |
|
"grad_norm": 1.84375, |
|
"learning_rate": 2.8105079461323818e-06, |
|
"loss": 0.7781, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 6.131695955960195, |
|
"grad_norm": 1.8359375, |
|
"learning_rate": 2.7810056802637995e-06, |
|
"loss": 0.7646, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 6.14863434257887, |
|
"grad_norm": 1.96875, |
|
"learning_rate": 2.751599305939082e-06, |
|
"loss": 0.7657, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 6.165572729197544, |
|
"grad_norm": 2.109375, |
|
"learning_rate": 2.7222900939242787e-06, |
|
"loss": 0.7932, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 6.182511115816219, |
|
"grad_norm": 2.203125, |
|
"learning_rate": 2.693079310786661e-06, |
|
"loss": 0.7611, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 6.199449502434893, |
|
"grad_norm": 2.1875, |
|
"learning_rate": 2.6639682188400017e-06, |
|
"loss": 0.7631, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 6.216387889053568, |
|
"grad_norm": 2.140625, |
|
"learning_rate": 2.6349580760900232e-06, |
|
"loss": 0.7784, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 6.233326275672242, |
|
"grad_norm": 2.234375, |
|
"learning_rate": 2.6060501361800333e-06, |
|
"loss": 0.7479, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 6.250264662290917, |
|
"grad_norm": 2.171875, |
|
"learning_rate": 2.57724564833675e-06, |
|
"loss": 0.7367, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 6.267203048909591, |
|
"grad_norm": 1.875, |
|
"learning_rate": 2.5485458573163114e-06, |
|
"loss": 0.7463, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 6.284141435528266, |
|
"grad_norm": 1.953125, |
|
"learning_rate": 2.519952003350499e-06, |
|
"loss": 0.7557, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 6.30107982214694, |
|
"grad_norm": 2.0, |
|
"learning_rate": 2.4914653220931283e-06, |
|
"loss": 0.8084, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 6.318018208765615, |
|
"grad_norm": 1.875, |
|
"learning_rate": 2.4630870445666576e-06, |
|
"loss": 0.7493, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 6.3349565953842895, |
|
"grad_norm": 2.15625, |
|
"learning_rate": 2.434818397108993e-06, |
|
"loss": 0.7959, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 6.351894982002964, |
|
"grad_norm": 2.21875, |
|
"learning_rate": 2.406660601320489e-06, |
|
"loss": 0.7245, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 6.3688333686216385, |
|
"grad_norm": 2.25, |
|
"learning_rate": 2.378614874011162e-06, |
|
"loss": 0.7637, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 6.385771755240313, |
|
"grad_norm": 2.15625, |
|
"learning_rate": 2.3506824271480987e-06, |
|
"loss": 0.7702, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 6.402710141858988, |
|
"grad_norm": 2.09375, |
|
"learning_rate": 2.322864467803097e-06, |
|
"loss": 0.7155, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 6.419648528477662, |
|
"grad_norm": 2.015625, |
|
"learning_rate": 2.2951621981004873e-06, |
|
"loss": 0.7925, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 6.436586915096337, |
|
"grad_norm": 1.96875, |
|
"learning_rate": 2.2675768151651993e-06, |
|
"loss": 0.7901, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 6.453525301715012, |
|
"grad_norm": 2.09375, |
|
"learning_rate": 2.240109511071014e-06, |
|
"loss": 0.729, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 6.470463688333687, |
|
"grad_norm": 2.109375, |
|
"learning_rate": 2.2127614727890627e-06, |
|
"loss": 0.7481, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 6.487402074952361, |
|
"grad_norm": 1.90625, |
|
"learning_rate": 2.1855338821365302e-06, |
|
"loss": 0.7148, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 6.504340461571036, |
|
"grad_norm": 1.8125, |
|
"learning_rate": 2.158427915725573e-06, |
|
"loss": 0.7748, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 6.52127884818971, |
|
"grad_norm": 2.125, |
|
"learning_rate": 2.131444744912492e-06, |
|
"loss": 0.7601, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 6.538217234808385, |
|
"grad_norm": 2.0625, |
|
"learning_rate": 2.1045855357470944e-06, |
|
"loss": 0.7397, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 6.555155621427059, |
|
"grad_norm": 2.578125, |
|
"learning_rate": 2.0778514489223207e-06, |
|
"loss": 0.7072, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 6.572094008045734, |
|
"grad_norm": 2.21875, |
|
"learning_rate": 2.051243639724071e-06, |
|
"loss": 0.7204, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 6.589032394664408, |
|
"grad_norm": 2.15625, |
|
"learning_rate": 2.024763257981296e-06, |
|
"loss": 0.7579, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 6.605970781283083, |
|
"grad_norm": 1.9765625, |
|
"learning_rate": 1.9984114480162976e-06, |
|
"loss": 0.7411, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 6.622909167901757, |
|
"grad_norm": 2.03125, |
|
"learning_rate": 1.9721893485952785e-06, |
|
"loss": 0.7513, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 6.639847554520432, |
|
"grad_norm": 2.0625, |
|
"learning_rate": 1.9460980928791407e-06, |
|
"loss": 0.7629, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 6.656785941139106, |
|
"grad_norm": 1.96875, |
|
"learning_rate": 1.9201388083745036e-06, |
|
"loss": 0.767, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 6.673724327757781, |
|
"grad_norm": 2.0625, |
|
"learning_rate": 1.8943126168849947e-06, |
|
"loss": 0.731, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 6.690662714376455, |
|
"grad_norm": 1.78125, |
|
"learning_rate": 1.868620634462756e-06, |
|
"loss": 0.7335, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 6.70760110099513, |
|
"grad_norm": 2.109375, |
|
"learning_rate": 1.8430639713602317e-06, |
|
"loss": 0.7791, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 6.7245394876138045, |
|
"grad_norm": 1.9921875, |
|
"learning_rate": 1.8176437319821789e-06, |
|
"loss": 0.7231, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 6.741477874232479, |
|
"grad_norm": 1.9375, |
|
"learning_rate": 1.79236101483794e-06, |
|
"loss": 0.7355, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 6.7584162608511535, |
|
"grad_norm": 1.734375, |
|
"learning_rate": 1.7672169124939843e-06, |
|
"loss": 0.7737, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 6.775354647469829, |
|
"grad_norm": 1.96875, |
|
"learning_rate": 1.7422125115266785e-06, |
|
"loss": 0.7269, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 6.775354647469829, |
|
"eval_loss": 1.0812190771102905, |
|
"eval_runtime": 21.7625, |
|
"eval_samples_per_second": 24.124, |
|
"eval_steps_per_second": 24.124, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 6.7922930340885035, |
|
"grad_norm": 1.7890625, |
|
"learning_rate": 1.7173488924753451e-06, |
|
"loss": 0.7756, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 6.809231420707178, |
|
"grad_norm": 1.6953125, |
|
"learning_rate": 1.6926271297955549e-06, |
|
"loss": 0.7597, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 6.8261698073258525, |
|
"grad_norm": 1.90625, |
|
"learning_rate": 1.668048291812709e-06, |
|
"loss": 0.7463, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 6.843108193944527, |
|
"grad_norm": 3.25, |
|
"learning_rate": 1.6436134406758635e-06, |
|
"loss": 0.7202, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 6.860046580563202, |
|
"grad_norm": 1.6875, |
|
"learning_rate": 1.6193236323118283e-06, |
|
"loss": 0.7765, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 6.876984967181876, |
|
"grad_norm": 1.84375, |
|
"learning_rate": 1.5951799163795461e-06, |
|
"loss": 0.7552, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 6.893923353800551, |
|
"grad_norm": 1.6484375, |
|
"learning_rate": 1.5711833362247204e-06, |
|
"loss": 0.7565, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 6.910861740419225, |
|
"grad_norm": 1.7265625, |
|
"learning_rate": 1.547334928834739e-06, |
|
"loss": 0.7342, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 6.9278001270379, |
|
"grad_norm": 1.578125, |
|
"learning_rate": 1.523635724793852e-06, |
|
"loss": 0.7602, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 6.944738513656574, |
|
"grad_norm": 1.6015625, |
|
"learning_rate": 1.5000867482386478e-06, |
|
"loss": 0.7839, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 6.961676900275249, |
|
"grad_norm": 1.6171875, |
|
"learning_rate": 1.4766890168137816e-06, |
|
"loss": 0.7512, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 6.978615286893923, |
|
"grad_norm": 1.6484375, |
|
"learning_rate": 1.4534435416280135e-06, |
|
"loss": 0.7715, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 6.995553673512598, |
|
"grad_norm": 1.546875, |
|
"learning_rate": 1.4303513272105057e-06, |
|
"loss": 0.7924, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 7.012492060131272, |
|
"grad_norm": 1.4453125, |
|
"learning_rate": 1.4074133714674116e-06, |
|
"loss": 0.7509, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 7.029430446749947, |
|
"grad_norm": 1.3984375, |
|
"learning_rate": 1.3846306656387603e-06, |
|
"loss": 0.7349, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 7.046368833368621, |
|
"grad_norm": 1.4765625, |
|
"learning_rate": 1.3620041942556127e-06, |
|
"loss": 0.7525, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 7.063307219987296, |
|
"grad_norm": 1.34375, |
|
"learning_rate": 1.339534935097524e-06, |
|
"loss": 0.7515, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 7.0802456066059705, |
|
"grad_norm": 1.4296875, |
|
"learning_rate": 1.3172238591502794e-06, |
|
"loss": 0.7433, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 7.097183993224645, |
|
"grad_norm": 1.265625, |
|
"learning_rate": 1.2950719305639464e-06, |
|
"loss": 0.7416, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 7.1141223798433195, |
|
"grad_norm": 1.3984375, |
|
"learning_rate": 1.2730801066112024e-06, |
|
"loss": 0.7115, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 7.131060766461994, |
|
"grad_norm": 1.3203125, |
|
"learning_rate": 1.251249337645965e-06, |
|
"loss": 0.6892, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 7.1479991530806695, |
|
"grad_norm": 1.3125, |
|
"learning_rate": 1.229580567062334e-06, |
|
"loss": 0.704, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 7.164937539699344, |
|
"grad_norm": 1.3203125, |
|
"learning_rate": 1.2080747312538082e-06, |
|
"loss": 0.7021, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 7.1818759263180185, |
|
"grad_norm": 1.3828125, |
|
"learning_rate": 1.1867327595728373e-06, |
|
"loss": 0.6669, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 7.198814312936693, |
|
"grad_norm": 1.2890625, |
|
"learning_rate": 1.1655555742906438e-06, |
|
"loss": 0.7433, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 7.215752699555368, |
|
"grad_norm": 1.3046875, |
|
"learning_rate": 1.1445440905573857e-06, |
|
"loss": 0.7294, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 7.232691086174042, |
|
"grad_norm": 1.359375, |
|
"learning_rate": 1.1236992163625933e-06, |
|
"loss": 0.7121, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 7.249629472792717, |
|
"grad_norm": 1.3046875, |
|
"learning_rate": 1.1030218524959375e-06, |
|
"loss": 0.6407, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 7.266567859411391, |
|
"grad_norm": 1.515625, |
|
"learning_rate": 1.082512892508309e-06, |
|
"loss": 0.7426, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 7.283506246030066, |
|
"grad_norm": 1.265625, |
|
"learning_rate": 1.062173222673194e-06, |
|
"loss": 0.7256, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 7.30044463264874, |
|
"grad_norm": 1.140625, |
|
"learning_rate": 1.0420037219483842e-06, |
|
"loss": 0.7149, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 7.317383019267415, |
|
"grad_norm": 1.171875, |
|
"learning_rate": 1.0220052619379844e-06, |
|
"loss": 0.7327, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 7.334321405886089, |
|
"grad_norm": 1.2734375, |
|
"learning_rate": 1.0021787068547568e-06, |
|
"loss": 0.7728, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 7.351259792504764, |
|
"grad_norm": 1.2890625, |
|
"learning_rate": 9.825249134827664e-07, |
|
"loss": 0.745, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 7.368198179123438, |
|
"grad_norm": 1.046875, |
|
"learning_rate": 9.630447311403623e-07, |
|
"loss": 0.7625, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 7.385136565742113, |
|
"grad_norm": 1.0703125, |
|
"learning_rate": 9.437390016434711e-07, |
|
"loss": 0.7297, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 7.402074952360787, |
|
"grad_norm": 1.1015625, |
|
"learning_rate": 9.246085592692183e-07, |
|
"loss": 0.6938, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 7.419013338979462, |
|
"grad_norm": 1.1015625, |
|
"learning_rate": 9.056542307198823e-07, |
|
"loss": 0.721, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 7.435951725598136, |
|
"grad_norm": 1.0703125, |
|
"learning_rate": 8.868768350871592e-07, |
|
"loss": 0.712, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 7.452890112216811, |
|
"grad_norm": 1.1640625, |
|
"learning_rate": 8.682771838167775e-07, |
|
"loss": 0.7684, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 7.469828498835486, |
|
"grad_norm": 1.109375, |
|
"learning_rate": 8.49856080673423e-07, |
|
"loss": 0.723, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 7.486766885454161, |
|
"grad_norm": 1.046875, |
|
"learning_rate": 8.316143217060129e-07, |
|
"loss": 0.7165, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 7.503705272072835, |
|
"grad_norm": 1.0625, |
|
"learning_rate": 8.135526952132877e-07, |
|
"loss": 0.7471, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 7.52064365869151, |
|
"grad_norm": 1.109375, |
|
"learning_rate": 7.956719817097491e-07, |
|
"loss": 0.7388, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 7.5375820453101845, |
|
"grad_norm": 1.0078125, |
|
"learning_rate": 7.779729538919351e-07, |
|
"loss": 0.7165, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 7.554520431928859, |
|
"grad_norm": 0.9609375, |
|
"learning_rate": 7.6045637660502e-07, |
|
"loss": 0.7349, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 7.5714588185475336, |
|
"grad_norm": 1.171875, |
|
"learning_rate": 7.4312300680977e-07, |
|
"loss": 0.6813, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 7.588397205166208, |
|
"grad_norm": 1.09375, |
|
"learning_rate": 7.259735935498297e-07, |
|
"loss": 0.7447, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 7.605335591784883, |
|
"grad_norm": 0.9609375, |
|
"learning_rate": 7.090088779193516e-07, |
|
"loss": 0.6976, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 7.622273978403557, |
|
"grad_norm": 0.98046875, |
|
"learning_rate": 6.922295930309691e-07, |
|
"loss": 0.7372, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 7.622273978403557, |
|
"eval_loss": 1.0816632509231567, |
|
"eval_runtime": 21.0317, |
|
"eval_samples_per_second": 24.962, |
|
"eval_steps_per_second": 24.962, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 7.639212365022232, |
|
"grad_norm": 1.078125, |
|
"learning_rate": 6.756364639841217e-07, |
|
"loss": 0.703, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 7.656150751640906, |
|
"grad_norm": 1.0390625, |
|
"learning_rate": 6.592302078337132e-07, |
|
"loss": 0.7121, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 7.673089138259581, |
|
"grad_norm": 1.0078125, |
|
"learning_rate": 6.430115335591291e-07, |
|
"loss": 0.7669, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 7.690027524878255, |
|
"grad_norm": 1.0234375, |
|
"learning_rate": 6.26981142033602e-07, |
|
"loss": 0.7015, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 7.70696591149693, |
|
"grad_norm": 1.0390625, |
|
"learning_rate": 6.111397259939128e-07, |
|
"loss": 0.736, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 7.723904298115604, |
|
"grad_norm": 1.0390625, |
|
"learning_rate": 5.954879700104732e-07, |
|
"loss": 0.7379, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 7.740842684734279, |
|
"grad_norm": 1.09375, |
|
"learning_rate": 5.800265504577202e-07, |
|
"loss": 0.7687, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 7.757781071352953, |
|
"grad_norm": 0.96875, |
|
"learning_rate": 5.647561354849079e-07, |
|
"loss": 0.7316, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 7.774719457971628, |
|
"grad_norm": 1.03125, |
|
"learning_rate": 5.496773849872183e-07, |
|
"loss": 0.7466, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 7.791657844590302, |
|
"grad_norm": 1.09375, |
|
"learning_rate": 5.347909505772569e-07, |
|
"loss": 0.734, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 7.808596231208977, |
|
"grad_norm": 1.0546875, |
|
"learning_rate": 5.200974755568849e-07, |
|
"loss": 0.7497, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 7.8255346178276515, |
|
"grad_norm": 1.046875, |
|
"learning_rate": 5.055975948894226e-07, |
|
"loss": 0.7562, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 7.842473004446326, |
|
"grad_norm": 1.125, |
|
"learning_rate": 4.912919351722151e-07, |
|
"loss": 0.7017, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 7.859411391065001, |
|
"grad_norm": 1.1328125, |
|
"learning_rate": 4.771811146095451e-07, |
|
"loss": 0.6976, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 7.876349777683676, |
|
"grad_norm": 1.046875, |
|
"learning_rate": 4.6326574298592666e-07, |
|
"loss": 0.6964, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 7.8932881643023505, |
|
"grad_norm": 1.203125, |
|
"learning_rate": 4.49546421639746e-07, |
|
"loss": 0.7239, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 7.910226550921025, |
|
"grad_norm": 1.0625, |
|
"learning_rate": 4.360237434372838e-07, |
|
"loss": 0.7632, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 7.9271649375396995, |
|
"grad_norm": 1.1640625, |
|
"learning_rate": 4.226982927470874e-07, |
|
"loss": 0.7184, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 7.944103324158374, |
|
"grad_norm": 1.1484375, |
|
"learning_rate": 4.0957064541472305e-07, |
|
"loss": 0.7238, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 7.961041710777049, |
|
"grad_norm": 1.421875, |
|
"learning_rate": 3.966413687378895e-07, |
|
"loss": 0.7361, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 7.977980097395723, |
|
"grad_norm": 1.390625, |
|
"learning_rate": 3.8391102144190175e-07, |
|
"loss": 0.723, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 7.994918484014398, |
|
"grad_norm": 1.4296875, |
|
"learning_rate": 3.7138015365554834e-07, |
|
"loss": 0.7064, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 8.011856870633073, |
|
"grad_norm": 1.3125, |
|
"learning_rate": 3.590493068873163e-07, |
|
"loss": 0.7317, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 8.028795257251748, |
|
"grad_norm": 1.3359375, |
|
"learning_rate": 3.4691901400199234e-07, |
|
"loss": 0.7102, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 8.045733643870422, |
|
"grad_norm": 2.4375, |
|
"learning_rate": 3.349897991976336e-07, |
|
"loss": 0.681, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 8.062672030489097, |
|
"grad_norm": 2.171875, |
|
"learning_rate": 3.232621779829176e-07, |
|
"loss": 0.6939, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 8.079610417107771, |
|
"grad_norm": 2.140625, |
|
"learning_rate": 3.1173665715486076e-07, |
|
"loss": 0.7276, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 8.096548803726446, |
|
"grad_norm": 2.09375, |
|
"learning_rate": 3.0041373477692315e-07, |
|
"loss": 0.6762, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 8.11348719034512, |
|
"grad_norm": 2.171875, |
|
"learning_rate": 2.8929390015748124e-07, |
|
"loss": 0.7547, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 8.130425576963795, |
|
"grad_norm": 3.71875, |
|
"learning_rate": 2.7837763382868253e-07, |
|
"loss": 0.716, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 8.14736396358247, |
|
"grad_norm": 3.15625, |
|
"learning_rate": 2.6766540752568514e-07, |
|
"loss": 0.7306, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 8.164302350201144, |
|
"grad_norm": 3.46875, |
|
"learning_rate": 2.5715768416626486e-07, |
|
"loss": 0.6856, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 8.181240736819818, |
|
"grad_norm": 3.859375, |
|
"learning_rate": 2.4685491783081715e-07, |
|
"loss": 0.6798, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 8.198179123438493, |
|
"grad_norm": 3.46875, |
|
"learning_rate": 2.3675755374272857e-07, |
|
"loss": 0.7652, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 8.215117510057167, |
|
"grad_norm": 8.0, |
|
"learning_rate": 2.268660282491436e-07, |
|
"loss": 0.6881, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 8.232055896675842, |
|
"grad_norm": 7.21875, |
|
"learning_rate": 2.1718076880210325e-07, |
|
"loss": 0.7532, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 8.248994283294516, |
|
"grad_norm": 7.0625, |
|
"learning_rate": 2.0770219394007362e-07, |
|
"loss": 0.6872, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 8.265932669913191, |
|
"grad_norm": 7.8125, |
|
"learning_rate": 1.9843071326986264e-07, |
|
"loss": 0.7416, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 8.282871056531866, |
|
"grad_norm": 8.3125, |
|
"learning_rate": 1.893667274489136e-07, |
|
"loss": 0.703, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 8.29980944315054, |
|
"grad_norm": 2.34375, |
|
"learning_rate": 1.8051062816799913e-07, |
|
"loss": 0.6915, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 8.316747829769215, |
|
"grad_norm": 2.109375, |
|
"learning_rate": 1.7186279813428518e-07, |
|
"loss": 0.6801, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 8.333686216387889, |
|
"grad_norm": 2.03125, |
|
"learning_rate": 1.6342361105480098e-07, |
|
"loss": 0.7441, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 8.350624603006564, |
|
"grad_norm": 2.140625, |
|
"learning_rate": 1.551934316202869e-07, |
|
"loss": 0.7246, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 8.367562989625238, |
|
"grad_norm": 2.140625, |
|
"learning_rate": 1.471726154894304e-07, |
|
"loss": 0.6718, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 8.384501376243913, |
|
"grad_norm": 1.65625, |
|
"learning_rate": 1.39361509273504e-07, |
|
"loss": 0.709, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 8.401439762862587, |
|
"grad_norm": 1.875, |
|
"learning_rate": 1.3176045052138097e-07, |
|
"loss": 0.7661, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 8.418378149481262, |
|
"grad_norm": 2.359375, |
|
"learning_rate": 1.2436976770495267e-07, |
|
"loss": 0.7727, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 8.435316536099936, |
|
"grad_norm": 1.703125, |
|
"learning_rate": 1.1718978020492988e-07, |
|
"loss": 0.802, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 8.45225492271861, |
|
"grad_norm": 1.8046875, |
|
"learning_rate": 1.1022079829704436e-07, |
|
"loss": 0.7435, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 8.469193309337285, |
|
"grad_norm": 1.703125, |
|
"learning_rate": 1.0346312313864127e-07, |
|
"loss": 0.7366, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 8.469193309337285, |
|
"eval_loss": 1.0841984748840332, |
|
"eval_runtime": 21.0727, |
|
"eval_samples_per_second": 24.914, |
|
"eval_steps_per_second": 24.914, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 8.48613169595596, |
|
"grad_norm": 2.265625, |
|
"learning_rate": 9.691704675565904e-08, |
|
"loss": 0.7348, |
|
"step": 5010 |
|
}, |
|
{ |
|
"epoch": 8.503070082574634, |
|
"grad_norm": 1.953125, |
|
"learning_rate": 9.058285203001837e-08, |
|
"loss": 0.703, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 8.520008469193309, |
|
"grad_norm": 1.796875, |
|
"learning_rate": 8.446081268738971e-08, |
|
"loss": 0.6796, |
|
"step": 5030 |
|
}, |
|
{ |
|
"epoch": 8.536946855811983, |
|
"grad_norm": 1.9140625, |
|
"learning_rate": 7.855119328537109e-08, |
|
"loss": 0.7424, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 8.553885242430658, |
|
"grad_norm": 1.7890625, |
|
"learning_rate": 7.285424920204953e-08, |
|
"loss": 0.7379, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 8.570823629049332, |
|
"grad_norm": 1.90625, |
|
"learning_rate": 6.737022662497095e-08, |
|
"loss": 0.7113, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 8.587762015668007, |
|
"grad_norm": 1.9140625, |
|
"learning_rate": 6.209936254049808e-08, |
|
"loss": 0.698, |
|
"step": 5070 |
|
}, |
|
{ |
|
"epoch": 8.604700402286682, |
|
"grad_norm": 2.25, |
|
"learning_rate": 5.704188472356764e-08, |
|
"loss": 0.7441, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 8.621638788905356, |
|
"grad_norm": 2.375, |
|
"learning_rate": 5.219801172785255e-08, |
|
"loss": 0.6628, |
|
"step": 5090 |
|
}, |
|
{ |
|
"epoch": 8.63857717552403, |
|
"grad_norm": 2.09375, |
|
"learning_rate": 4.756795287631288e-08, |
|
"loss": 0.7012, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 8.655515562142705, |
|
"grad_norm": 2.375, |
|
"learning_rate": 4.315190825215143e-08, |
|
"loss": 0.7372, |
|
"step": 5110 |
|
}, |
|
{ |
|
"epoch": 8.67245394876138, |
|
"grad_norm": 1.8125, |
|
"learning_rate": 3.895006869016837e-08, |
|
"loss": 0.7353, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 8.689392335380056, |
|
"grad_norm": 1.6484375, |
|
"learning_rate": 3.496261576851345e-08, |
|
"loss": 0.7127, |
|
"step": 5130 |
|
}, |
|
{ |
|
"epoch": 8.706330721998729, |
|
"grad_norm": 2.15625, |
|
"learning_rate": 3.1189721800839453e-08, |
|
"loss": 0.7462, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 8.723269108617405, |
|
"grad_norm": 2.140625, |
|
"learning_rate": 2.763154982885541e-08, |
|
"loss": 0.7486, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 8.74020749523608, |
|
"grad_norm": 1.9296875, |
|
"learning_rate": 2.4288253615282754e-08, |
|
"loss": 0.7136, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 8.757145881854754, |
|
"grad_norm": 2.046875, |
|
"learning_rate": 2.1159977637208452e-08, |
|
"loss": 0.746, |
|
"step": 5170 |
|
}, |
|
{ |
|
"epoch": 8.774084268473429, |
|
"grad_norm": 1.90625, |
|
"learning_rate": 1.824685707984164e-08, |
|
"loss": 0.7379, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 8.791022655092103, |
|
"grad_norm": 1.9375, |
|
"learning_rate": 1.5549017830674416e-08, |
|
"loss": 0.7561, |
|
"step": 5190 |
|
}, |
|
{ |
|
"epoch": 8.807961041710778, |
|
"grad_norm": 1.96875, |
|
"learning_rate": 1.3066576474038416e-08, |
|
"loss": 0.7722, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 8.824899428329452, |
|
"grad_norm": 2.03125, |
|
"learning_rate": 1.0799640286068836e-08, |
|
"loss": 0.7318, |
|
"step": 5210 |
|
}, |
|
{ |
|
"epoch": 8.841837814948127, |
|
"grad_norm": 1.7890625, |
|
"learning_rate": 8.748307230067587e-09, |
|
"loss": 0.7358, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 8.858776201566801, |
|
"grad_norm": 1.7734375, |
|
"learning_rate": 6.912665952270581e-09, |
|
"loss": 0.7051, |
|
"step": 5230 |
|
}, |
|
{ |
|
"epoch": 8.875714588185476, |
|
"grad_norm": 1.8203125, |
|
"learning_rate": 5.29279577801689e-09, |
|
"loss": 0.7542, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 8.89265297480415, |
|
"grad_norm": 2.0625, |
|
"learning_rate": 3.888766708319836e-09, |
|
"loss": 0.7354, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 8.909591361422825, |
|
"grad_norm": 2.1875, |
|
"learning_rate": 2.700639416843287e-09, |
|
"loss": 0.7226, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 8.9265297480415, |
|
"grad_norm": 2.171875, |
|
"learning_rate": 1.7284652472787655e-09, |
|
"loss": 0.7387, |
|
"step": 5270 |
|
}, |
|
{ |
|
"epoch": 8.943468134660174, |
|
"grad_norm": 2.125, |
|
"learning_rate": 9.722862111277708e-10, |
|
"loss": 0.6839, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 8.960406521278848, |
|
"grad_norm": 2.34375, |
|
"learning_rate": 4.321349858843471e-10, |
|
"loss": 0.7071, |
|
"step": 5290 |
|
}, |
|
{ |
|
"epoch": 8.977344907897523, |
|
"grad_norm": 1.9765625, |
|
"learning_rate": 1.080349136250991e-10, |
|
"loss": 0.7484, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 8.994283294516197, |
|
"grad_norm": 1.9375, |
|
"learning_rate": 0.0, |
|
"loss": 0.7226, |
|
"step": 5310 |
|
}, |
|
{ |
|
"epoch": 8.994283294516197, |
|
"step": 5310, |
|
"total_flos": 5.095460800932741e+17, |
|
"train_loss": 0.9683849982173653, |
|
"train_runtime": 5676.759, |
|
"train_samples_per_second": 7.488, |
|
"train_steps_per_second": 0.935 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 5310, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 9, |
|
"save_steps": 0, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 5.095460800932741e+17, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|