|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 500, |
|
"global_step": 922, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.010845986984815618, |
|
"grad_norm": 5.733648324455126, |
|
"learning_rate": 5.405405405405406e-07, |
|
"loss": 0.9884, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.021691973969631236, |
|
"grad_norm": 3.143018703050177, |
|
"learning_rate": 1.0810810810810812e-06, |
|
"loss": 1.003, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.03253796095444685, |
|
"grad_norm": 1.95130960523758, |
|
"learning_rate": 1.6216216216216219e-06, |
|
"loss": 0.949, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.04338394793926247, |
|
"grad_norm": 1.309889103250474, |
|
"learning_rate": 2.1621621621621623e-06, |
|
"loss": 0.9141, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.05422993492407809, |
|
"grad_norm": 0.964663580942374, |
|
"learning_rate": 2.702702702702703e-06, |
|
"loss": 0.8705, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.0650759219088937, |
|
"grad_norm": 0.9285010836421135, |
|
"learning_rate": 3.2432432432432437e-06, |
|
"loss": 0.8527, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.07592190889370933, |
|
"grad_norm": 0.858798188360779, |
|
"learning_rate": 3.7837837837837844e-06, |
|
"loss": 0.8526, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.08676789587852494, |
|
"grad_norm": 0.9904029778258275, |
|
"learning_rate": 4.324324324324325e-06, |
|
"loss": 0.8533, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.09761388286334056, |
|
"grad_norm": 1.016007763190019, |
|
"learning_rate": 4.864864864864866e-06, |
|
"loss": 0.8337, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.10845986984815618, |
|
"grad_norm": 1.0144931809319673, |
|
"learning_rate": 5.405405405405406e-06, |
|
"loss": 0.8571, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.1193058568329718, |
|
"grad_norm": 1.0215217769302534, |
|
"learning_rate": 5.945945945945947e-06, |
|
"loss": 0.838, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.1301518438177874, |
|
"grad_norm": 0.9196298812766283, |
|
"learning_rate": 6.486486486486487e-06, |
|
"loss": 0.8642, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.14099783080260303, |
|
"grad_norm": 0.9071017283188243, |
|
"learning_rate": 7.027027027027028e-06, |
|
"loss": 0.8455, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.15184381778741865, |
|
"grad_norm": 0.8998523509417926, |
|
"learning_rate": 7.567567567567569e-06, |
|
"loss": 0.8173, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.16268980477223427, |
|
"grad_norm": 0.8967758141211395, |
|
"learning_rate": 8.108108108108109e-06, |
|
"loss": 0.8207, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.1735357917570499, |
|
"grad_norm": 0.854081117546703, |
|
"learning_rate": 8.64864864864865e-06, |
|
"loss": 0.826, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.1843817787418655, |
|
"grad_norm": 0.7791141573115371, |
|
"learning_rate": 9.189189189189191e-06, |
|
"loss": 0.8408, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.19522776572668113, |
|
"grad_norm": 0.931666128574369, |
|
"learning_rate": 9.729729729729732e-06, |
|
"loss": 0.8572, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.20607375271149675, |
|
"grad_norm": 1.0655645445027464, |
|
"learning_rate": 9.999775878383519e-06, |
|
"loss": 0.8449, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.21691973969631237, |
|
"grad_norm": 0.8552506900147497, |
|
"learning_rate": 9.997983026003064e-06, |
|
"loss": 0.8338, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.227765726681128, |
|
"grad_norm": 0.7811721176850553, |
|
"learning_rate": 9.9943979641349e-06, |
|
"loss": 0.8419, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.2386117136659436, |
|
"grad_norm": 0.9612589991038837, |
|
"learning_rate": 9.989021978333996e-06, |
|
"loss": 0.8447, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.24945770065075923, |
|
"grad_norm": 0.8713636930259827, |
|
"learning_rate": 9.981856996356548e-06, |
|
"loss": 0.826, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.2603036876355748, |
|
"grad_norm": 1.0230589449620575, |
|
"learning_rate": 9.972905587468719e-06, |
|
"loss": 0.8509, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.27114967462039047, |
|
"grad_norm": 0.9960895654836321, |
|
"learning_rate": 9.962170961525338e-06, |
|
"loss": 0.8278, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.28199566160520606, |
|
"grad_norm": 0.9708416426642364, |
|
"learning_rate": 9.949656967818882e-06, |
|
"loss": 0.8239, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.2928416485900217, |
|
"grad_norm": 0.9863073274043072, |
|
"learning_rate": 9.935368093699171e-06, |
|
"loss": 0.8729, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.3036876355748373, |
|
"grad_norm": 1.0619350079027392, |
|
"learning_rate": 9.919309462964277e-06, |
|
"loss": 0.8336, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.31453362255965295, |
|
"grad_norm": 0.9536874477902112, |
|
"learning_rate": 9.901486834023182e-06, |
|
"loss": 0.8371, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.32537960954446854, |
|
"grad_norm": 1.0060463939033413, |
|
"learning_rate": 9.8819065978309e-06, |
|
"loss": 0.864, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.3362255965292842, |
|
"grad_norm": 0.792063518682088, |
|
"learning_rate": 9.860575775596767e-06, |
|
"loss": 0.8313, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.3470715835140998, |
|
"grad_norm": 0.9546416116474393, |
|
"learning_rate": 9.837502016266725e-06, |
|
"loss": 0.8218, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.3579175704989154, |
|
"grad_norm": 0.9802614938135474, |
|
"learning_rate": 9.812693593780515e-06, |
|
"loss": 0.8721, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.368763557483731, |
|
"grad_norm": 0.854378979529178, |
|
"learning_rate": 9.786159404104758e-06, |
|
"loss": 0.8371, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.3796095444685466, |
|
"grad_norm": 0.8717109867216107, |
|
"learning_rate": 9.757908962042968e-06, |
|
"loss": 0.8339, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.39045553145336226, |
|
"grad_norm": 0.8877006786963313, |
|
"learning_rate": 9.72795239782369e-06, |
|
"loss": 0.8547, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.40130151843817785, |
|
"grad_norm": 1.0126192151398974, |
|
"learning_rate": 9.696300453467922e-06, |
|
"loss": 0.8438, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.4121475054229935, |
|
"grad_norm": 0.8577472807238208, |
|
"learning_rate": 9.66296447893717e-06, |
|
"loss": 0.872, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.4229934924078091, |
|
"grad_norm": 0.8412488678641884, |
|
"learning_rate": 9.627956428063522e-06, |
|
"loss": 0.8408, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.43383947939262474, |
|
"grad_norm": 0.7588179294196125, |
|
"learning_rate": 9.59128885426314e-06, |
|
"loss": 0.8451, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.44468546637744033, |
|
"grad_norm": 0.8703037224398377, |
|
"learning_rate": 9.552974906034796e-06, |
|
"loss": 0.8336, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.455531453362256, |
|
"grad_norm": 0.8699706833983841, |
|
"learning_rate": 9.513028322244977e-06, |
|
"loss": 0.8153, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.46637744034707157, |
|
"grad_norm": 0.847977363828918, |
|
"learning_rate": 9.47146342720133e-06, |
|
"loss": 0.857, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.4772234273318872, |
|
"grad_norm": 0.8984826481514769, |
|
"learning_rate": 9.428295125516151e-06, |
|
"loss": 0.8467, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.4880694143167028, |
|
"grad_norm": 0.8165556682574098, |
|
"learning_rate": 9.383538896761787e-06, |
|
"loss": 0.8311, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.49891540130151846, |
|
"grad_norm": 0.8007389807149831, |
|
"learning_rate": 9.337210789919875e-06, |
|
"loss": 0.8648, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.5097613882863341, |
|
"grad_norm": 0.769668675462935, |
|
"learning_rate": 9.289327417626393e-06, |
|
"loss": 0.8342, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.5206073752711496, |
|
"grad_norm": 0.9160701884545429, |
|
"learning_rate": 9.239905950214587e-06, |
|
"loss": 0.8509, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.5314533622559653, |
|
"grad_norm": 0.8467668226954682, |
|
"learning_rate": 9.18896410955793e-06, |
|
"loss": 0.8405, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.5422993492407809, |
|
"grad_norm": 0.8109237435952316, |
|
"learning_rate": 9.136520162715288e-06, |
|
"loss": 0.8454, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.5422993492407809, |
|
"eval_loss": 0.7866095304489136, |
|
"eval_runtime": 2581.2899, |
|
"eval_samples_per_second": 1.904, |
|
"eval_steps_per_second": 0.476, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.5531453362255966, |
|
"grad_norm": 0.9798708544137009, |
|
"learning_rate": 9.082592915380596e-06, |
|
"loss": 0.8255, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.5639913232104121, |
|
"grad_norm": 0.9165811375712184, |
|
"learning_rate": 9.027201705139406e-06, |
|
"loss": 0.8663, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.5748373101952278, |
|
"grad_norm": 0.9060399071688227, |
|
"learning_rate": 8.970366394534667e-06, |
|
"loss": 0.8144, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.5856832971800434, |
|
"grad_norm": 0.8253353508928236, |
|
"learning_rate": 8.912107363944297e-06, |
|
"loss": 0.8129, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.596529284164859, |
|
"grad_norm": 0.8996220079581437, |
|
"learning_rate": 8.852445504273056e-06, |
|
"loss": 0.8493, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.6073752711496746, |
|
"grad_norm": 0.7975347083538386, |
|
"learning_rate": 8.791402209461333e-06, |
|
"loss": 0.8602, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.6182212581344902, |
|
"grad_norm": 0.7263963682022704, |
|
"learning_rate": 8.728999368813591e-06, |
|
"loss": 0.835, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.6290672451193059, |
|
"grad_norm": 0.9605105643436394, |
|
"learning_rate": 8.665259359149132e-06, |
|
"loss": 0.8362, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.6399132321041214, |
|
"grad_norm": 0.8209007974348012, |
|
"learning_rate": 8.600205036778089e-06, |
|
"loss": 0.8233, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.6507592190889371, |
|
"grad_norm": 0.911985371229915, |
|
"learning_rate": 8.533859729305447e-06, |
|
"loss": 0.8375, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.6616052060737527, |
|
"grad_norm": 0.6985325225275438, |
|
"learning_rate": 8.466247227266091e-06, |
|
"loss": 0.8225, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.6724511930585684, |
|
"grad_norm": 0.8132034730555108, |
|
"learning_rate": 8.39739177559383e-06, |
|
"loss": 0.836, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.6832971800433839, |
|
"grad_norm": 0.8360457612694335, |
|
"learning_rate": 8.327318064927488e-06, |
|
"loss": 0.8491, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.6941431670281996, |
|
"grad_norm": 0.8189142007610347, |
|
"learning_rate": 8.256051222757188e-06, |
|
"loss": 0.8486, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.7049891540130152, |
|
"grad_norm": 0.8530912616563548, |
|
"learning_rate": 8.183616804413954e-06, |
|
"loss": 0.8489, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.7158351409978309, |
|
"grad_norm": 0.9149414345864662, |
|
"learning_rate": 8.110040783905924e-06, |
|
"loss": 0.8244, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.7266811279826464, |
|
"grad_norm": 0.8342820136081186, |
|
"learning_rate": 8.035349544604419e-06, |
|
"loss": 0.8201, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.737527114967462, |
|
"grad_norm": 0.7652272869820805, |
|
"learning_rate": 7.959569869783216e-06, |
|
"loss": 0.8287, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.7483731019522777, |
|
"grad_norm": 0.8697789473135982, |
|
"learning_rate": 7.882728933014431e-06, |
|
"loss": 0.8565, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.7592190889370932, |
|
"grad_norm": 0.8289580942636415, |
|
"learning_rate": 7.80485428842444e-06, |
|
"loss": 0.8354, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.7700650759219089, |
|
"grad_norm": 0.82218666332152, |
|
"learning_rate": 7.725973860813338e-06, |
|
"loss": 0.8275, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.7809110629067245, |
|
"grad_norm": 0.8328912197470162, |
|
"learning_rate": 7.646115935641488e-06, |
|
"loss": 0.8554, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.7917570498915402, |
|
"grad_norm": 0.9144219376531081, |
|
"learning_rate": 7.5653091488867215e-06, |
|
"loss": 0.7935, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.8026030368763557, |
|
"grad_norm": 0.8432999710569549, |
|
"learning_rate": 7.48358247677588e-06, |
|
"loss": 0.8343, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.8134490238611713, |
|
"grad_norm": 0.9959358723449406, |
|
"learning_rate": 7.400965225394316e-06, |
|
"loss": 0.8215, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.824295010845987, |
|
"grad_norm": 0.7781247788376849, |
|
"learning_rate": 7.31748702017713e-06, |
|
"loss": 0.7865, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.8351409978308026, |
|
"grad_norm": 0.7268868727283686, |
|
"learning_rate": 7.23317779528589e-06, |
|
"loss": 0.8554, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.8459869848156182, |
|
"grad_norm": 0.8769959745106497, |
|
"learning_rate": 7.14806778287464e-06, |
|
"loss": 0.8556, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.8568329718004338, |
|
"grad_norm": 0.8083886562171313, |
|
"learning_rate": 7.062187502249056e-06, |
|
"loss": 0.8538, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.8676789587852495, |
|
"grad_norm": 0.8253588275102612, |
|
"learning_rate": 6.975567748922639e-06, |
|
"loss": 0.8483, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.8785249457700651, |
|
"grad_norm": 0.8419247557676373, |
|
"learning_rate": 6.888239583573852e-06, |
|
"loss": 0.8383, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.8893709327548807, |
|
"grad_norm": 0.8261807774132319, |
|
"learning_rate": 6.8002343209081766e-06, |
|
"loss": 0.8344, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.9002169197396963, |
|
"grad_norm": 0.9081092978343738, |
|
"learning_rate": 6.711583518429093e-06, |
|
"loss": 0.8614, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.911062906724512, |
|
"grad_norm": 0.8081110590736196, |
|
"learning_rate": 6.622318965121972e-06, |
|
"loss": 0.8283, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.9219088937093276, |
|
"grad_norm": 0.8961074992740756, |
|
"learning_rate": 6.532472670054975e-06, |
|
"loss": 0.8555, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.9327548806941431, |
|
"grad_norm": 0.855697485520701, |
|
"learning_rate": 6.442076850901033e-06, |
|
"loss": 0.805, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.9436008676789588, |
|
"grad_norm": 0.9715823055879019, |
|
"learning_rate": 6.351163922385026e-06, |
|
"loss": 0.8746, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.9544468546637744, |
|
"grad_norm": 0.8558421168141579, |
|
"learning_rate": 6.259766484660297e-06, |
|
"loss": 0.8194, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.96529284164859, |
|
"grad_norm": 0.9307662219253259, |
|
"learning_rate": 6.1679173116186674e-06, |
|
"loss": 0.8234, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.9761388286334056, |
|
"grad_norm": 0.8797281549707557, |
|
"learning_rate": 6.075649339138174e-06, |
|
"loss": 0.8336, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.9869848156182213, |
|
"grad_norm": 0.7892356935050042, |
|
"learning_rate": 5.982995653272699e-06, |
|
"loss": 0.8471, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.9978308026030369, |
|
"grad_norm": 0.7635682452713507, |
|
"learning_rate": 5.8899894783877536e-06, |
|
"loss": 0.8248, |
|
"step": 920 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 1844, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 49624320344064.0, |
|
"train_batch_size": 6, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|