{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 922, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.010845986984815618, "grad_norm": 5.733648324455126, "learning_rate": 5.405405405405406e-07, "loss": 0.9884, "step": 10 }, { "epoch": 0.021691973969631236, "grad_norm": 3.143018703050177, "learning_rate": 1.0810810810810812e-06, "loss": 1.003, "step": 20 }, { "epoch": 0.03253796095444685, "grad_norm": 1.95130960523758, "learning_rate": 1.6216216216216219e-06, "loss": 0.949, "step": 30 }, { "epoch": 0.04338394793926247, "grad_norm": 1.309889103250474, "learning_rate": 2.1621621621621623e-06, "loss": 0.9141, "step": 40 }, { "epoch": 0.05422993492407809, "grad_norm": 0.964663580942374, "learning_rate": 2.702702702702703e-06, "loss": 0.8705, "step": 50 }, { "epoch": 0.0650759219088937, "grad_norm": 0.9285010836421135, "learning_rate": 3.2432432432432437e-06, "loss": 0.8527, "step": 60 }, { "epoch": 0.07592190889370933, "grad_norm": 0.858798188360779, "learning_rate": 3.7837837837837844e-06, "loss": 0.8526, "step": 70 }, { "epoch": 0.08676789587852494, "grad_norm": 0.9904029778258275, "learning_rate": 4.324324324324325e-06, "loss": 0.8533, "step": 80 }, { "epoch": 0.09761388286334056, "grad_norm": 1.016007763190019, "learning_rate": 4.864864864864866e-06, "loss": 0.8337, "step": 90 }, { "epoch": 0.10845986984815618, "grad_norm": 1.0144931809319673, "learning_rate": 5.405405405405406e-06, "loss": 0.8571, "step": 100 }, { "epoch": 0.1193058568329718, "grad_norm": 1.0215217769302534, "learning_rate": 5.945945945945947e-06, "loss": 0.838, "step": 110 }, { "epoch": 0.1301518438177874, "grad_norm": 0.9196298812766283, "learning_rate": 6.486486486486487e-06, "loss": 0.8642, "step": 120 }, { "epoch": 0.14099783080260303, "grad_norm": 0.9071017283188243, "learning_rate": 7.027027027027028e-06, "loss": 0.8455, "step": 130 }, { "epoch": 0.15184381778741865, "grad_norm": 0.8998523509417926, "learning_rate": 7.567567567567569e-06, "loss": 0.8173, "step": 140 }, { "epoch": 0.16268980477223427, "grad_norm": 0.8967758141211395, "learning_rate": 8.108108108108109e-06, "loss": 0.8207, "step": 150 }, { "epoch": 0.1735357917570499, "grad_norm": 0.854081117546703, "learning_rate": 8.64864864864865e-06, "loss": 0.826, "step": 160 }, { "epoch": 0.1843817787418655, "grad_norm": 0.7791141573115371, "learning_rate": 9.189189189189191e-06, "loss": 0.8408, "step": 170 }, { "epoch": 0.19522776572668113, "grad_norm": 0.931666128574369, "learning_rate": 9.729729729729732e-06, "loss": 0.8572, "step": 180 }, { "epoch": 0.20607375271149675, "grad_norm": 1.0655645445027464, "learning_rate": 9.999775878383519e-06, "loss": 0.8449, "step": 190 }, { "epoch": 0.21691973969631237, "grad_norm": 0.8552506900147497, "learning_rate": 9.997983026003064e-06, "loss": 0.8338, "step": 200 }, { "epoch": 0.227765726681128, "grad_norm": 0.7811721176850553, "learning_rate": 9.9943979641349e-06, "loss": 0.8419, "step": 210 }, { "epoch": 0.2386117136659436, "grad_norm": 0.9612589991038837, "learning_rate": 9.989021978333996e-06, "loss": 0.8447, "step": 220 }, { "epoch": 0.24945770065075923, "grad_norm": 0.8713636930259827, "learning_rate": 9.981856996356548e-06, "loss": 0.826, "step": 230 }, { "epoch": 0.2603036876355748, "grad_norm": 1.0230589449620575, "learning_rate": 9.972905587468719e-06, "loss": 0.8509, "step": 240 }, { "epoch": 0.27114967462039047, "grad_norm": 0.9960895654836321, "learning_rate": 9.962170961525338e-06, "loss": 0.8278, "step": 250 }, { "epoch": 0.28199566160520606, "grad_norm": 0.9708416426642364, "learning_rate": 9.949656967818882e-06, "loss": 0.8239, "step": 260 }, { "epoch": 0.2928416485900217, "grad_norm": 0.9863073274043072, "learning_rate": 9.935368093699171e-06, "loss": 0.8729, "step": 270 }, { "epoch": 0.3036876355748373, "grad_norm": 1.0619350079027392, "learning_rate": 9.919309462964277e-06, "loss": 0.8336, "step": 280 }, { "epoch": 0.31453362255965295, "grad_norm": 0.9536874477902112, "learning_rate": 9.901486834023182e-06, "loss": 0.8371, "step": 290 }, { "epoch": 0.32537960954446854, "grad_norm": 1.0060463939033413, "learning_rate": 9.8819065978309e-06, "loss": 0.864, "step": 300 }, { "epoch": 0.3362255965292842, "grad_norm": 0.792063518682088, "learning_rate": 9.860575775596767e-06, "loss": 0.8313, "step": 310 }, { "epoch": 0.3470715835140998, "grad_norm": 0.9546416116474393, "learning_rate": 9.837502016266725e-06, "loss": 0.8218, "step": 320 }, { "epoch": 0.3579175704989154, "grad_norm": 0.9802614938135474, "learning_rate": 9.812693593780515e-06, "loss": 0.8721, "step": 330 }, { "epoch": 0.368763557483731, "grad_norm": 0.854378979529178, "learning_rate": 9.786159404104758e-06, "loss": 0.8371, "step": 340 }, { "epoch": 0.3796095444685466, "grad_norm": 0.8717109867216107, "learning_rate": 9.757908962042968e-06, "loss": 0.8339, "step": 350 }, { "epoch": 0.39045553145336226, "grad_norm": 0.8877006786963313, "learning_rate": 9.72795239782369e-06, "loss": 0.8547, "step": 360 }, { "epoch": 0.40130151843817785, "grad_norm": 1.0126192151398974, "learning_rate": 9.696300453467922e-06, "loss": 0.8438, "step": 370 }, { "epoch": 0.4121475054229935, "grad_norm": 0.8577472807238208, "learning_rate": 9.66296447893717e-06, "loss": 0.872, "step": 380 }, { "epoch": 0.4229934924078091, "grad_norm": 0.8412488678641884, "learning_rate": 9.627956428063522e-06, "loss": 0.8408, "step": 390 }, { "epoch": 0.43383947939262474, "grad_norm": 0.7588179294196125, "learning_rate": 9.59128885426314e-06, "loss": 0.8451, "step": 400 }, { "epoch": 0.44468546637744033, "grad_norm": 0.8703037224398377, "learning_rate": 9.552974906034796e-06, "loss": 0.8336, "step": 410 }, { "epoch": 0.455531453362256, "grad_norm": 0.8699706833983841, "learning_rate": 9.513028322244977e-06, "loss": 0.8153, "step": 420 }, { "epoch": 0.46637744034707157, "grad_norm": 0.847977363828918, "learning_rate": 9.47146342720133e-06, "loss": 0.857, "step": 430 }, { "epoch": 0.4772234273318872, "grad_norm": 0.8984826481514769, "learning_rate": 9.428295125516151e-06, "loss": 0.8467, "step": 440 }, { "epoch": 0.4880694143167028, "grad_norm": 0.8165556682574098, "learning_rate": 9.383538896761787e-06, "loss": 0.8311, "step": 450 }, { "epoch": 0.49891540130151846, "grad_norm": 0.8007389807149831, "learning_rate": 9.337210789919875e-06, "loss": 0.8648, "step": 460 }, { "epoch": 0.5097613882863341, "grad_norm": 0.769668675462935, "learning_rate": 9.289327417626393e-06, "loss": 0.8342, "step": 470 }, { "epoch": 0.5206073752711496, "grad_norm": 0.9160701884545429, "learning_rate": 9.239905950214587e-06, "loss": 0.8509, "step": 480 }, { "epoch": 0.5314533622559653, "grad_norm": 0.8467668226954682, "learning_rate": 9.18896410955793e-06, "loss": 0.8405, "step": 490 }, { "epoch": 0.5422993492407809, "grad_norm": 0.8109237435952316, "learning_rate": 9.136520162715288e-06, "loss": 0.8454, "step": 500 }, { "epoch": 0.5422993492407809, "eval_loss": 0.7866095304489136, "eval_runtime": 2581.2899, "eval_samples_per_second": 1.904, "eval_steps_per_second": 0.476, "step": 500 }, { "epoch": 0.5531453362255966, "grad_norm": 0.9798708544137009, "learning_rate": 9.082592915380596e-06, "loss": 0.8255, "step": 510 }, { "epoch": 0.5639913232104121, "grad_norm": 0.9165811375712184, "learning_rate": 9.027201705139406e-06, "loss": 0.8663, "step": 520 }, { "epoch": 0.5748373101952278, "grad_norm": 0.9060399071688227, "learning_rate": 8.970366394534667e-06, "loss": 0.8144, "step": 530 }, { "epoch": 0.5856832971800434, "grad_norm": 0.8253353508928236, "learning_rate": 8.912107363944297e-06, "loss": 0.8129, "step": 540 }, { "epoch": 0.596529284164859, "grad_norm": 0.8996220079581437, "learning_rate": 8.852445504273056e-06, "loss": 0.8493, "step": 550 }, { "epoch": 0.6073752711496746, "grad_norm": 0.7975347083538386, "learning_rate": 8.791402209461333e-06, "loss": 0.8602, "step": 560 }, { "epoch": 0.6182212581344902, "grad_norm": 0.7263963682022704, "learning_rate": 8.728999368813591e-06, "loss": 0.835, "step": 570 }, { "epoch": 0.6290672451193059, "grad_norm": 0.9605105643436394, "learning_rate": 8.665259359149132e-06, "loss": 0.8362, "step": 580 }, { "epoch": 0.6399132321041214, "grad_norm": 0.8209007974348012, "learning_rate": 8.600205036778089e-06, "loss": 0.8233, "step": 590 }, { "epoch": 0.6507592190889371, "grad_norm": 0.911985371229915, "learning_rate": 8.533859729305447e-06, "loss": 0.8375, "step": 600 }, { "epoch": 0.6616052060737527, "grad_norm": 0.6985325225275438, "learning_rate": 8.466247227266091e-06, "loss": 0.8225, "step": 610 }, { "epoch": 0.6724511930585684, "grad_norm": 0.8132034730555108, "learning_rate": 8.39739177559383e-06, "loss": 0.836, "step": 620 }, { "epoch": 0.6832971800433839, "grad_norm": 0.8360457612694335, "learning_rate": 8.327318064927488e-06, "loss": 0.8491, "step": 630 }, { "epoch": 0.6941431670281996, "grad_norm": 0.8189142007610347, "learning_rate": 8.256051222757188e-06, "loss": 0.8486, "step": 640 }, { "epoch": 0.7049891540130152, "grad_norm": 0.8530912616563548, "learning_rate": 8.183616804413954e-06, "loss": 0.8489, "step": 650 }, { "epoch": 0.7158351409978309, "grad_norm": 0.9149414345864662, "learning_rate": 8.110040783905924e-06, "loss": 0.8244, "step": 660 }, { "epoch": 0.7266811279826464, "grad_norm": 0.8342820136081186, "learning_rate": 8.035349544604419e-06, "loss": 0.8201, "step": 670 }, { "epoch": 0.737527114967462, "grad_norm": 0.7652272869820805, "learning_rate": 7.959569869783216e-06, "loss": 0.8287, "step": 680 }, { "epoch": 0.7483731019522777, "grad_norm": 0.8697789473135982, "learning_rate": 7.882728933014431e-06, "loss": 0.8565, "step": 690 }, { "epoch": 0.7592190889370932, "grad_norm": 0.8289580942636415, "learning_rate": 7.80485428842444e-06, "loss": 0.8354, "step": 700 }, { "epoch": 0.7700650759219089, "grad_norm": 0.82218666332152, "learning_rate": 7.725973860813338e-06, "loss": 0.8275, "step": 710 }, { "epoch": 0.7809110629067245, "grad_norm": 0.8328912197470162, "learning_rate": 7.646115935641488e-06, "loss": 0.8554, "step": 720 }, { "epoch": 0.7917570498915402, "grad_norm": 0.9144219376531081, "learning_rate": 7.5653091488867215e-06, "loss": 0.7935, "step": 730 }, { "epoch": 0.8026030368763557, "grad_norm": 0.8432999710569549, "learning_rate": 7.48358247677588e-06, "loss": 0.8343, "step": 740 }, { "epoch": 0.8134490238611713, "grad_norm": 0.9959358723449406, "learning_rate": 7.400965225394316e-06, "loss": 0.8215, "step": 750 }, { "epoch": 0.824295010845987, "grad_norm": 0.7781247788376849, "learning_rate": 7.31748702017713e-06, "loss": 0.7865, "step": 760 }, { "epoch": 0.8351409978308026, "grad_norm": 0.7268868727283686, "learning_rate": 7.23317779528589e-06, "loss": 0.8554, "step": 770 }, { "epoch": 0.8459869848156182, "grad_norm": 0.8769959745106497, "learning_rate": 7.14806778287464e-06, "loss": 0.8556, "step": 780 }, { "epoch": 0.8568329718004338, "grad_norm": 0.8083886562171313, "learning_rate": 7.062187502249056e-06, "loss": 0.8538, "step": 790 }, { "epoch": 0.8676789587852495, "grad_norm": 0.8253588275102612, "learning_rate": 6.975567748922639e-06, "loss": 0.8483, "step": 800 }, { "epoch": 0.8785249457700651, "grad_norm": 0.8419247557676373, "learning_rate": 6.888239583573852e-06, "loss": 0.8383, "step": 810 }, { "epoch": 0.8893709327548807, "grad_norm": 0.8261807774132319, "learning_rate": 6.8002343209081766e-06, "loss": 0.8344, "step": 820 }, { "epoch": 0.9002169197396963, "grad_norm": 0.9081092978343738, "learning_rate": 6.711583518429093e-06, "loss": 0.8614, "step": 830 }, { "epoch": 0.911062906724512, "grad_norm": 0.8081110590736196, "learning_rate": 6.622318965121972e-06, "loss": 0.8283, "step": 840 }, { "epoch": 0.9219088937093276, "grad_norm": 0.8961074992740756, "learning_rate": 6.532472670054975e-06, "loss": 0.8555, "step": 850 }, { "epoch": 0.9327548806941431, "grad_norm": 0.855697485520701, "learning_rate": 6.442076850901033e-06, "loss": 0.805, "step": 860 }, { "epoch": 0.9436008676789588, "grad_norm": 0.9715823055879019, "learning_rate": 6.351163922385026e-06, "loss": 0.8746, "step": 870 }, { "epoch": 0.9544468546637744, "grad_norm": 0.8558421168141579, "learning_rate": 6.259766484660297e-06, "loss": 0.8194, "step": 880 }, { "epoch": 0.96529284164859, "grad_norm": 0.9307662219253259, "learning_rate": 6.1679173116186674e-06, "loss": 0.8234, "step": 890 }, { "epoch": 0.9761388286334056, "grad_norm": 0.8797281549707557, "learning_rate": 6.075649339138174e-06, "loss": 0.8336, "step": 900 }, { "epoch": 0.9869848156182213, "grad_norm": 0.7892356935050042, "learning_rate": 5.982995653272699e-06, "loss": 0.8471, "step": 910 }, { "epoch": 0.9978308026030369, "grad_norm": 0.7635682452713507, "learning_rate": 5.8899894783877536e-06, "loss": 0.8248, "step": 920 } ], "logging_steps": 10, "max_steps": 1844, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 49624320344064.0, "train_batch_size": 6, "trial_name": null, "trial_params": null }