{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0213386832026263, "eval_steps": 500, "global_step": 1400, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.007295276308590188, "grad_norm": 0.9992304444313049, "learning_rate": 9.999671349822887e-06, "loss": 0.087, "step": 10 }, { "epoch": 0.014590552617180376, "grad_norm": 1.0456160306930542, "learning_rate": 9.998685442495921e-06, "loss": 0.0757, "step": 20 }, { "epoch": 0.021885828925770564, "grad_norm": 1.014310598373413, "learning_rate": 9.99704240762655e-06, "loss": 0.0901, "step": 30 }, { "epoch": 0.029181105234360752, "grad_norm": 0.9391605257987976, "learning_rate": 9.994742461208251e-06, "loss": 0.0811, "step": 40 }, { "epoch": 0.036476381542950936, "grad_norm": 0.7982305884361267, "learning_rate": 9.991785905592149e-06, "loss": 0.0818, "step": 50 }, { "epoch": 0.04377165785154113, "grad_norm": 0.9946795701980591, "learning_rate": 9.988173129447251e-06, "loss": 0.0905, "step": 60 }, { "epoch": 0.05106693416013131, "grad_norm": 0.8562415242195129, "learning_rate": 9.983904607709365e-06, "loss": 0.0831, "step": 70 }, { "epoch": 0.058362210468721504, "grad_norm": 1.054521918296814, "learning_rate": 9.978980901518663e-06, "loss": 0.083, "step": 80 }, { "epoch": 0.06565748677731169, "grad_norm": 1.0189253091812134, "learning_rate": 9.973402658145908e-06, "loss": 0.0891, "step": 90 }, { "epoch": 0.07295276308590187, "grad_norm": 0.7745731472969055, "learning_rate": 9.96717061090737e-06, "loss": 0.0846, "step": 100 }, { "epoch": 0.08024803939449207, "grad_norm": 0.8214731812477112, "learning_rate": 9.960285579068419e-06, "loss": 0.0895, "step": 110 }, { "epoch": 0.08754331570308226, "grad_norm": 0.9114850759506226, "learning_rate": 9.95274846773583e-06, "loss": 0.0862, "step": 120 }, { "epoch": 0.09483859201167244, "grad_norm": 1.216098427772522, "learning_rate": 9.944560267738792e-06, "loss": 0.0896, "step": 130 }, { "epoch": 0.10213386832026262, "grad_norm": 0.7587252855300903, "learning_rate": 9.935722055498655e-06, "loss": 0.0888, "step": 140 }, { "epoch": 0.10942914462885282, "grad_norm": 0.8791028261184692, "learning_rate": 9.92623499288743e-06, "loss": 0.0832, "step": 150 }, { "epoch": 0.11672442093744301, "grad_norm": 1.133716344833374, "learning_rate": 9.916100327075038e-06, "loss": 0.0854, "step": 160 }, { "epoch": 0.12401969724603319, "grad_norm": 0.7774285674095154, "learning_rate": 9.905319390365364e-06, "loss": 0.0873, "step": 170 }, { "epoch": 0.13131497355462338, "grad_norm": 0.7029784321784973, "learning_rate": 9.893893600021112e-06, "loss": 0.0786, "step": 180 }, { "epoch": 0.13861024986321357, "grad_norm": 1.0894945859909058, "learning_rate": 9.881824458077491e-06, "loss": 0.089, "step": 190 }, { "epoch": 0.14590552617180375, "grad_norm": 0.7053947448730469, "learning_rate": 9.869113551144754e-06, "loss": 0.0843, "step": 200 }, { "epoch": 0.15320080248039394, "grad_norm": 0.9423328638076782, "learning_rate": 9.85576255019963e-06, "loss": 0.0887, "step": 210 }, { "epoch": 0.16049607878898414, "grad_norm": 1.1778377294540405, "learning_rate": 9.841773210365646e-06, "loss": 0.0939, "step": 220 }, { "epoch": 0.1677913550975743, "grad_norm": 0.8132289052009583, "learning_rate": 9.82714737068241e-06, "loss": 0.0812, "step": 230 }, { "epoch": 0.1750866314061645, "grad_norm": 0.925722062587738, "learning_rate": 9.811886953863841e-06, "loss": 0.091, "step": 240 }, { "epoch": 0.1823819077147547, "grad_norm": 0.8174653649330139, "learning_rate": 9.795993966045418e-06, "loss": 0.0868, "step": 250 }, { "epoch": 0.18967718402334488, "grad_norm": 1.086427092552185, "learning_rate": 9.779470496520442e-06, "loss": 0.0882, "step": 260 }, { "epoch": 0.19697246033193508, "grad_norm": 0.8431033492088318, "learning_rate": 9.76231871746539e-06, "loss": 0.0819, "step": 270 }, { "epoch": 0.20426773664052525, "grad_norm": 0.9793428778648376, "learning_rate": 9.744540883654348e-06, "loss": 0.0838, "step": 280 }, { "epoch": 0.21156301294911545, "grad_norm": 1.1424306631088257, "learning_rate": 9.726139332162613e-06, "loss": 0.0923, "step": 290 }, { "epoch": 0.21885828925770565, "grad_norm": 0.8779186010360718, "learning_rate": 9.707116482059447e-06, "loss": 0.0909, "step": 300 }, { "epoch": 0.22615356556629582, "grad_norm": 0.862575113773346, "learning_rate": 9.68747483409007e-06, "loss": 0.094, "step": 310 }, { "epoch": 0.23344884187488602, "grad_norm": 0.8033697009086609, "learning_rate": 9.667216970346916e-06, "loss": 0.0821, "step": 320 }, { "epoch": 0.24074411818347619, "grad_norm": 1.136168360710144, "learning_rate": 9.646345553930187e-06, "loss": 0.084, "step": 330 }, { "epoch": 0.24803939449206638, "grad_norm": 1.227287769317627, "learning_rate": 9.624863328597767e-06, "loss": 0.0867, "step": 340 }, { "epoch": 0.2553346708006566, "grad_norm": 0.8906837105751038, "learning_rate": 9.602773118404518e-06, "loss": 0.0845, "step": 350 }, { "epoch": 0.26262994710924675, "grad_norm": 0.9557612538337708, "learning_rate": 9.580077827331038e-06, "loss": 0.0896, "step": 360 }, { "epoch": 0.2699252234178369, "grad_norm": 0.8804728984832764, "learning_rate": 9.556780438901899e-06, "loss": 0.0799, "step": 370 }, { "epoch": 0.27722049972642715, "grad_norm": 0.8825680017471313, "learning_rate": 9.532884015793432e-06, "loss": 0.0879, "step": 380 }, { "epoch": 0.2845157760350173, "grad_norm": 1.216217041015625, "learning_rate": 9.508391699431114e-06, "loss": 0.0878, "step": 390 }, { "epoch": 0.2918110523436075, "grad_norm": 1.0918773412704468, "learning_rate": 9.48330670957659e-06, "loss": 0.0842, "step": 400 }, { "epoch": 0.2991063286521977, "grad_norm": 0.8797096014022827, "learning_rate": 9.457632343904404e-06, "loss": 0.075, "step": 410 }, { "epoch": 0.3064016049607879, "grad_norm": 0.9638001322746277, "learning_rate": 9.431371977568483e-06, "loss": 0.0834, "step": 420 }, { "epoch": 0.31369688126937806, "grad_norm": 0.7447642683982849, "learning_rate": 9.404529062758447e-06, "loss": 0.0909, "step": 430 }, { "epoch": 0.3209921575779683, "grad_norm": 1.179291844367981, "learning_rate": 9.377107128245782e-06, "loss": 0.0832, "step": 440 }, { "epoch": 0.32828743388655846, "grad_norm": 0.6798911094665527, "learning_rate": 9.349109778919938e-06, "loss": 0.0842, "step": 450 }, { "epoch": 0.3355827101951486, "grad_norm": 0.9468401670455933, "learning_rate": 9.32054069531444e-06, "loss": 0.0937, "step": 460 }, { "epoch": 0.34287798650373885, "grad_norm": 0.9600223898887634, "learning_rate": 9.291403633123046e-06, "loss": 0.0909, "step": 470 }, { "epoch": 0.350173262812329, "grad_norm": 0.7957492470741272, "learning_rate": 9.261702422706014e-06, "loss": 0.0924, "step": 480 }, { "epoch": 0.3574685391209192, "grad_norm": 0.9197902083396912, "learning_rate": 9.231440968586572e-06, "loss": 0.0852, "step": 490 }, { "execution_accuracy": 37.0, "selected_difficulty": "challenging", "step": 500 }, { "epoch": 0.3647638154295094, "grad_norm": 1.0779783725738525, "learning_rate": 9.200623248937619e-06, "loss": 0.0872, "step": 500 }, { "epoch": 0.3720590917380996, "grad_norm": 1.0330190658569336, "learning_rate": 9.169253315058764e-06, "loss": 0.0884, "step": 510 }, { "epoch": 0.37935436804668976, "grad_norm": 0.8854564428329468, "learning_rate": 9.13733529084374e-06, "loss": 0.0888, "step": 520 }, { "epoch": 0.38664964435527993, "grad_norm": 0.7928789258003235, "learning_rate": 9.104873372238269e-06, "loss": 0.0858, "step": 530 }, { "epoch": 0.39394492066387016, "grad_norm": 1.1074901819229126, "learning_rate": 9.071871826688472e-06, "loss": 0.0866, "step": 540 }, { "epoch": 0.40124019697246033, "grad_norm": 0.9030365347862244, "learning_rate": 9.038334992579863e-06, "loss": 0.084, "step": 550 }, { "epoch": 0.4085354732810505, "grad_norm": 0.9121033549308777, "learning_rate": 9.004267278667032e-06, "loss": 0.0841, "step": 560 }, { "epoch": 0.4158307495896407, "grad_norm": 0.7888039350509644, "learning_rate": 8.969673163494063e-06, "loss": 0.0844, "step": 570 }, { "epoch": 0.4231260258982309, "grad_norm": 0.8982441425323486, "learning_rate": 8.934557194805787e-06, "loss": 0.0881, "step": 580 }, { "epoch": 0.43042130220682107, "grad_norm": 0.9907477498054504, "learning_rate": 8.898923988949936e-06, "loss": 0.0829, "step": 590 }, { "epoch": 0.4377165785154113, "grad_norm": 0.7510169744491577, "learning_rate": 8.862778230270276e-06, "loss": 0.0812, "step": 600 }, { "epoch": 0.44501185482400146, "grad_norm": 0.8274700045585632, "learning_rate": 8.826124670490804e-06, "loss": 0.0872, "step": 610 }, { "epoch": 0.45230713113259163, "grad_norm": 1.118033766746521, "learning_rate": 8.788968128091084e-06, "loss": 0.0934, "step": 620 }, { "epoch": 0.45960240744118186, "grad_norm": 0.7703680992126465, "learning_rate": 8.751313487672815e-06, "loss": 0.0891, "step": 630 }, { "epoch": 0.46689768374977203, "grad_norm": 1.0380890369415283, "learning_rate": 8.71316569931769e-06, "loss": 0.0931, "step": 640 }, { "epoch": 0.4741929600583622, "grad_norm": 0.9548109769821167, "learning_rate": 8.674529777936674e-06, "loss": 0.0816, "step": 650 }, { "epoch": 0.48148823636695237, "grad_norm": 1.2412773370742798, "learning_rate": 8.635410802610724e-06, "loss": 0.0872, "step": 660 }, { "epoch": 0.4887835126755426, "grad_norm": 0.8023186922073364, "learning_rate": 8.595813915923113e-06, "loss": 0.0857, "step": 670 }, { "epoch": 0.49607878898413277, "grad_norm": 0.829302966594696, "learning_rate": 8.555744323283364e-06, "loss": 0.0932, "step": 680 }, { "epoch": 0.503374065292723, "grad_norm": 1.0218778848648071, "learning_rate": 8.515207292242969e-06, "loss": 0.0791, "step": 690 }, { "epoch": 0.5106693416013132, "grad_norm": 0.9110598564147949, "learning_rate": 8.474208151802898e-06, "loss": 0.0917, "step": 700 }, { "epoch": 0.5179646179099033, "grad_norm": 0.8709658980369568, "learning_rate": 8.432752291713058e-06, "loss": 0.0805, "step": 710 }, { "epoch": 0.5252598942184935, "grad_norm": 0.9375218749046326, "learning_rate": 8.390845161763756e-06, "loss": 0.0887, "step": 720 }, { "epoch": 0.5325551705270837, "grad_norm": 0.819020688533783, "learning_rate": 8.34849227106926e-06, "loss": 0.0823, "step": 730 }, { "epoch": 0.5398504468356738, "grad_norm": 0.7377147078514099, "learning_rate": 8.305699187343586e-06, "loss": 0.0867, "step": 740 }, { "epoch": 0.5471457231442641, "grad_norm": 0.9633333086967468, "learning_rate": 8.262471536168547e-06, "loss": 0.0893, "step": 750 }, { "epoch": 0.5544409994528543, "grad_norm": 0.894578218460083, "learning_rate": 8.218815000254233e-06, "loss": 0.0874, "step": 760 }, { "epoch": 0.5617362757614445, "grad_norm": 0.9977262616157532, "learning_rate": 8.174735318691946e-06, "loss": 0.0822, "step": 770 }, { "epoch": 0.5690315520700346, "grad_norm": 0.8501657247543335, "learning_rate": 8.130238286199747e-06, "loss": 0.0874, "step": 780 }, { "epoch": 0.5763268283786248, "grad_norm": 0.7603849172592163, "learning_rate": 8.085329752360683e-06, "loss": 0.0784, "step": 790 }, { "epoch": 0.583622104687215, "grad_norm": 0.8487511277198792, "learning_rate": 8.04001562085379e-06, "loss": 0.0878, "step": 800 }, { "epoch": 0.5909173809958053, "grad_norm": 0.9253877401351929, "learning_rate": 7.994301848678006e-06, "loss": 0.0815, "step": 810 }, { "epoch": 0.5982126573043954, "grad_norm": 0.9946874976158142, "learning_rate": 7.948194445369065e-06, "loss": 0.0901, "step": 820 }, { "epoch": 0.6055079336129856, "grad_norm": 1.006040334701538, "learning_rate": 7.901699472209467e-06, "loss": 0.0792, "step": 830 }, { "epoch": 0.6128032099215758, "grad_norm": 0.8797623515129089, "learning_rate": 7.85482304143168e-06, "loss": 0.0816, "step": 840 }, { "epoch": 0.620098486230166, "grad_norm": 0.9626962542533875, "learning_rate": 7.807571315414616e-06, "loss": 0.08, "step": 850 }, { "epoch": 0.6273937625387561, "grad_norm": 1.075498104095459, "learning_rate": 7.759950505873523e-06, "loss": 0.0802, "step": 860 }, { "epoch": 0.6346890388473464, "grad_norm": 0.90218585729599, "learning_rate": 7.711966873043396e-06, "loss": 0.0824, "step": 870 }, { "epoch": 0.6419843151559366, "grad_norm": 0.8692035675048828, "learning_rate": 7.66362672485601e-06, "loss": 0.0769, "step": 880 }, { "epoch": 0.6492795914645267, "grad_norm": 1.062687635421753, "learning_rate": 7.614936416110668e-06, "loss": 0.0834, "step": 890 }, { "epoch": 0.6565748677731169, "grad_norm": 0.9357954859733582, "learning_rate": 7.565902347638806e-06, "loss": 0.0833, "step": 900 }, { "epoch": 0.6638701440817071, "grad_norm": 0.873466432094574, "learning_rate": 7.5165309654625405e-06, "loss": 0.0803, "step": 910 }, { "epoch": 0.6711654203902973, "grad_norm": 0.8454645276069641, "learning_rate": 7.466828759947271e-06, "loss": 0.0793, "step": 920 }, { "epoch": 0.6784606966988874, "grad_norm": 1.1247007846832275, "learning_rate": 7.416802264948455e-06, "loss": 0.0757, "step": 930 }, { "epoch": 0.6857559730074777, "grad_norm": 0.9007195234298706, "learning_rate": 7.366458056952668e-06, "loss": 0.0855, "step": 940 }, { "epoch": 0.6930512493160679, "grad_norm": 0.775996208190918, "learning_rate": 7.315802754213062e-06, "loss": 0.079, "step": 950 }, { "epoch": 0.700346525624658, "grad_norm": 1.1148663759231567, "learning_rate": 7.264843015879321e-06, "loss": 0.0881, "step": 960 }, { "epoch": 0.7076418019332482, "grad_norm": 0.8462682366371155, "learning_rate": 7.213585541122261e-06, "loss": 0.0848, "step": 970 }, { "epoch": 0.7149370782418384, "grad_norm": 0.8557429313659668, "learning_rate": 7.162037068253141e-06, "loss": 0.0839, "step": 980 }, { "epoch": 0.7222323545504286, "grad_norm": 0.8856000900268555, "learning_rate": 7.110204373837857e-06, "loss": 0.0804, "step": 990 }, { "execution_accuracy": 47.0, "selected_difficulty": "challenging", "step": 1000 }, { "epoch": 0.7295276308590188, "grad_norm": 0.7584331631660461, "learning_rate": 7.058094271806091e-06, "loss": 0.0818, "step": 1000 }, { "epoch": 0.736822907167609, "grad_norm": 0.9286295175552368, "learning_rate": 7.0057136125555456e-06, "loss": 0.0766, "step": 1010 }, { "epoch": 0.7441181834761992, "grad_norm": 1.1081056594848633, "learning_rate": 6.953069282051397e-06, "loss": 0.0835, "step": 1020 }, { "epoch": 0.7514134597847894, "grad_norm": 0.9020804762840271, "learning_rate": 6.900168200921065e-06, "loss": 0.0791, "step": 1030 }, { "epoch": 0.7587087360933795, "grad_norm": 1.0643606185913086, "learning_rate": 6.84701732354442e-06, "loss": 0.0866, "step": 1040 }, { "epoch": 0.7660040124019697, "grad_norm": 0.9293026924133301, "learning_rate": 6.79362363713957e-06, "loss": 0.0828, "step": 1050 }, { "epoch": 0.7732992887105599, "grad_norm": 0.9997085332870483, "learning_rate": 6.7399941608443096e-06, "loss": 0.0737, "step": 1060 }, { "epoch": 0.7805945650191501, "grad_norm": 0.8557327389717102, "learning_rate": 6.686135944793395e-06, "loss": 0.0836, "step": 1070 }, { "epoch": 0.7878898413277403, "grad_norm": 0.8752471208572388, "learning_rate": 6.632056069191723e-06, "loss": 0.0897, "step": 1080 }, { "epoch": 0.7951851176363305, "grad_norm": 0.9777392148971558, "learning_rate": 6.57776164338357e-06, "loss": 0.0805, "step": 1090 }, { "epoch": 0.8024803939449207, "grad_norm": 0.9013363122940063, "learning_rate": 6.523259804918001e-06, "loss": 0.0764, "step": 1100 }, { "epoch": 0.8097756702535108, "grad_norm": 0.930316686630249, "learning_rate": 6.4685577186105595e-06, "loss": 0.0858, "step": 1110 }, { "epoch": 0.817070946562101, "grad_norm": 1.2016055583953857, "learning_rate": 6.413662575601391e-06, "loss": 0.0809, "step": 1120 }, { "epoch": 0.8243662228706913, "grad_norm": 0.8230682611465454, "learning_rate": 6.358581592409881e-06, "loss": 0.0771, "step": 1130 }, { "epoch": 0.8316614991792814, "grad_norm": 0.9575796127319336, "learning_rate": 6.303322009985984e-06, "loss": 0.0893, "step": 1140 }, { "epoch": 0.8389567754878716, "grad_norm": 0.7613864541053772, "learning_rate": 6.247891092758319e-06, "loss": 0.0802, "step": 1150 }, { "epoch": 0.8462520517964618, "grad_norm": 0.8168739080429077, "learning_rate": 6.1922961276791925e-06, "loss": 0.0718, "step": 1160 }, { "epoch": 0.853547328105052, "grad_norm": 0.719241201877594, "learning_rate": 6.136544423266651e-06, "loss": 0.073, "step": 1170 }, { "epoch": 0.8608426044136421, "grad_norm": 1.0418319702148438, "learning_rate": 6.08064330864371e-06, "loss": 0.0816, "step": 1180 }, { "epoch": 0.8681378807222323, "grad_norm": 0.7831118106842041, "learning_rate": 6.024600132574855e-06, "loss": 0.0775, "step": 1190 }, { "epoch": 0.8754331570308226, "grad_norm": 0.9321058988571167, "learning_rate": 5.968422262499983e-06, "loss": 0.0777, "step": 1200 }, { "epoch": 0.8827284333394128, "grad_norm": 0.9959325194358826, "learning_rate": 5.912117083565874e-06, "loss": 0.0743, "step": 1210 }, { "epoch": 0.8900237096480029, "grad_norm": 0.7858604192733765, "learning_rate": 5.85569199765534e-06, "loss": 0.0742, "step": 1220 }, { "epoch": 0.8973189859565931, "grad_norm": 1.0771974325180054, "learning_rate": 5.799154422414174e-06, "loss": 0.0784, "step": 1230 }, { "epoch": 0.9046142622651833, "grad_norm": 1.0624542236328125, "learning_rate": 5.7425117902760195e-06, "loss": 0.0784, "step": 1240 }, { "epoch": 0.9119095385737734, "grad_norm": 0.7207911014556885, "learning_rate": 5.685771547485312e-06, "loss": 0.0694, "step": 1250 }, { "epoch": 0.9192048148823637, "grad_norm": 1.0225133895874023, "learning_rate": 5.628941153118388e-06, "loss": 0.0747, "step": 1260 }, { "epoch": 0.9265000911909539, "grad_norm": 1.0003268718719482, "learning_rate": 5.572028078102917e-06, "loss": 0.077, "step": 1270 }, { "epoch": 0.9337953674995441, "grad_norm": 0.9205290079116821, "learning_rate": 5.515039804235772e-06, "loss": 0.0764, "step": 1280 }, { "epoch": 0.9410906438081342, "grad_norm": 1.0423191785812378, "learning_rate": 5.457983823199475e-06, "loss": 0.0801, "step": 1290 }, { "epoch": 0.9483859201167244, "grad_norm": 0.8814120888710022, "learning_rate": 5.400867635577335e-06, "loss": 0.0801, "step": 1300 }, { "epoch": 0.9556811964253146, "grad_norm": 0.9548910856246948, "learning_rate": 5.343698749867421e-06, "loss": 0.0802, "step": 1310 }, { "epoch": 0.9629764727339047, "grad_norm": 0.7712908983230591, "learning_rate": 5.2864846814955e-06, "loss": 0.077, "step": 1320 }, { "epoch": 0.970271749042495, "grad_norm": 0.9300876259803772, "learning_rate": 5.229232951827054e-06, "loss": 0.081, "step": 1330 }, { "epoch": 0.9775670253510852, "grad_norm": 0.932421863079071, "learning_rate": 5.17195108717852e-06, "loss": 0.0841, "step": 1340 }, { "epoch": 0.9848623016596754, "grad_norm": 0.9440054297447205, "learning_rate": 5.114646617827884e-06, "loss": 0.0714, "step": 1350 }, { "epoch": 0.9921575779682655, "grad_norm": 0.8986610174179077, "learning_rate": 5.057327077024745e-06, "loss": 0.0781, "step": 1360 }, { "epoch": 0.9994528542768557, "grad_norm": 0.9697067141532898, "learning_rate": 5e-06, "loss": 0.0863, "step": 1370 }, { "epoch": 1.006748130585446, "grad_norm": 0.5880870223045349, "learning_rate": 4.942672922975255e-06, "loss": 0.0443, "step": 1380 }, { "epoch": 1.014043406894036, "grad_norm": 0.8269129395484924, "learning_rate": 4.8853533821721175e-06, "loss": 0.0464, "step": 1390 }, { "epoch": 1.0213386832026263, "grad_norm": 0.9354344606399536, "learning_rate": 4.82804891282148e-06, "loss": 0.0436, "step": 1400 } ], "logging_steps": 10, "max_steps": 2740, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 200, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.849931290145456e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }