{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 1861, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0005373455131649651, "grad_norm": 23.072626458069166, "learning_rate": 5.3475935828877005e-08, "loss": 1.329, "step": 1 }, { "epoch": 0.0026867275658248252, "grad_norm": 23.380544253320735, "learning_rate": 2.6737967914438503e-07, "loss": 1.3238, "step": 5 }, { "epoch": 0.0053734551316496505, "grad_norm": 14.233532908496148, "learning_rate": 5.347593582887701e-07, "loss": 1.2647, "step": 10 }, { "epoch": 0.008060182697474477, "grad_norm": 11.791057381257374, "learning_rate": 8.021390374331551e-07, "loss": 1.1476, "step": 15 }, { "epoch": 0.010746910263299301, "grad_norm": 7.369104708436159, "learning_rate": 1.0695187165775401e-06, "loss": 1.0118, "step": 20 }, { "epoch": 0.013433637829124127, "grad_norm": 3.790654764997102, "learning_rate": 1.3368983957219254e-06, "loss": 0.9345, "step": 25 }, { "epoch": 0.016120365394948953, "grad_norm": 3.447275461870041, "learning_rate": 1.6042780748663103e-06, "loss": 0.8918, "step": 30 }, { "epoch": 0.018807092960773777, "grad_norm": 2.9186352407291847, "learning_rate": 1.8716577540106954e-06, "loss": 0.8518, "step": 35 }, { "epoch": 0.021493820526598602, "grad_norm": 2.9408795060935695, "learning_rate": 2.1390374331550802e-06, "loss": 0.844, "step": 40 }, { "epoch": 0.02418054809242343, "grad_norm": 2.8790720251278565, "learning_rate": 2.4064171122994653e-06, "loss": 0.8318, "step": 45 }, { "epoch": 0.026867275658248254, "grad_norm": 2.963731808583039, "learning_rate": 2.673796791443851e-06, "loss": 0.8099, "step": 50 }, { "epoch": 0.02955400322407308, "grad_norm": 2.8588219846284297, "learning_rate": 2.9411764705882355e-06, "loss": 0.8054, "step": 55 }, { "epoch": 0.032240730789897906, "grad_norm": 3.105229156653952, "learning_rate": 3.2085561497326205e-06, "loss": 0.7913, "step": 60 }, { "epoch": 0.03492745835572273, "grad_norm": 3.1413414626518943, "learning_rate": 3.4759358288770056e-06, "loss": 0.7867, "step": 65 }, { "epoch": 0.037614185921547555, "grad_norm": 3.132841956959321, "learning_rate": 3.7433155080213907e-06, "loss": 0.7893, "step": 70 }, { "epoch": 0.04030091348737238, "grad_norm": 3.0623425732500142, "learning_rate": 4.010695187165775e-06, "loss": 0.7717, "step": 75 }, { "epoch": 0.042987641053197204, "grad_norm": 9.233814569420407, "learning_rate": 4.2780748663101604e-06, "loss": 0.7715, "step": 80 }, { "epoch": 0.04567436861902203, "grad_norm": 3.3097521575259368, "learning_rate": 4.5454545454545455e-06, "loss": 0.761, "step": 85 }, { "epoch": 0.04836109618484686, "grad_norm": 3.0349283160634597, "learning_rate": 4.812834224598931e-06, "loss": 0.7621, "step": 90 }, { "epoch": 0.051047823750671684, "grad_norm": 3.1760527715984153, "learning_rate": 5.0802139037433165e-06, "loss": 0.7591, "step": 95 }, { "epoch": 0.05373455131649651, "grad_norm": 2.928755685153135, "learning_rate": 5.347593582887702e-06, "loss": 0.7433, "step": 100 }, { "epoch": 0.05642127888232133, "grad_norm": 3.0202006224727174, "learning_rate": 5.614973262032086e-06, "loss": 0.7314, "step": 105 }, { "epoch": 0.05910800644814616, "grad_norm": 3.1295278900578682, "learning_rate": 5.882352941176471e-06, "loss": 0.7483, "step": 110 }, { "epoch": 0.06179473401397098, "grad_norm": 3.0863071896945167, "learning_rate": 6.149732620320856e-06, "loss": 0.7276, "step": 115 }, { "epoch": 0.06448146157979581, "grad_norm": 3.0744127119070526, "learning_rate": 6.417112299465241e-06, "loss": 0.7183, "step": 120 }, { "epoch": 0.06716818914562063, "grad_norm": 3.22926416708401, "learning_rate": 6.684491978609626e-06, "loss": 0.7242, "step": 125 }, { "epoch": 0.06985491671144546, "grad_norm": 2.9238184305563033, "learning_rate": 6.951871657754011e-06, "loss": 0.7194, "step": 130 }, { "epoch": 0.07254164427727028, "grad_norm": 2.907399933470749, "learning_rate": 7.219251336898396e-06, "loss": 0.7204, "step": 135 }, { "epoch": 0.07522837184309511, "grad_norm": 2.8941698993873137, "learning_rate": 7.486631016042781e-06, "loss": 0.6986, "step": 140 }, { "epoch": 0.07791509940891994, "grad_norm": 2.899562495835514, "learning_rate": 7.754010695187166e-06, "loss": 0.7021, "step": 145 }, { "epoch": 0.08060182697474476, "grad_norm": 2.8740571869341687, "learning_rate": 8.02139037433155e-06, "loss": 0.7096, "step": 150 }, { "epoch": 0.08328855454056959, "grad_norm": 2.912059141124366, "learning_rate": 8.288770053475937e-06, "loss": 0.7012, "step": 155 }, { "epoch": 0.08597528210639441, "grad_norm": 2.933852566630193, "learning_rate": 8.556149732620321e-06, "loss": 0.6909, "step": 160 }, { "epoch": 0.08866200967221924, "grad_norm": 2.954361982135904, "learning_rate": 8.823529411764707e-06, "loss": 0.6947, "step": 165 }, { "epoch": 0.09134873723804406, "grad_norm": 2.914429965048127, "learning_rate": 9.090909090909091e-06, "loss": 0.7064, "step": 170 }, { "epoch": 0.09403546480386889, "grad_norm": 2.824165175450085, "learning_rate": 9.358288770053477e-06, "loss": 0.6962, "step": 175 }, { "epoch": 0.09672219236969372, "grad_norm": 2.7629699563001724, "learning_rate": 9.625668449197861e-06, "loss": 0.7093, "step": 180 }, { "epoch": 0.09940891993551854, "grad_norm": 2.9700788007957484, "learning_rate": 9.893048128342247e-06, "loss": 0.7001, "step": 185 }, { "epoch": 0.10209564750134337, "grad_norm": 2.720187963907155, "learning_rate": 9.999920755303033e-06, "loss": 0.6972, "step": 190 }, { "epoch": 0.10478237506716818, "grad_norm": 2.579060861199631, "learning_rate": 9.999436491251425e-06, "loss": 0.7011, "step": 195 }, { "epoch": 0.10746910263299302, "grad_norm": 2.650831290761427, "learning_rate": 9.998512030567253e-06, "loss": 0.7045, "step": 200 }, { "epoch": 0.11015583019881783, "grad_norm": 3.3197406459470673, "learning_rate": 9.99714745464859e-06, "loss": 0.6992, "step": 205 }, { "epoch": 0.11284255776464266, "grad_norm": 3.0709444563569126, "learning_rate": 9.995342883645325e-06, "loss": 0.708, "step": 210 }, { "epoch": 0.1155292853304675, "grad_norm": 2.815374931010047, "learning_rate": 9.993098476448576e-06, "loss": 0.7011, "step": 215 }, { "epoch": 0.11821601289629231, "grad_norm": 2.9613320208688005, "learning_rate": 9.990414430676716e-06, "loss": 0.6821, "step": 220 }, { "epoch": 0.12090274046211714, "grad_norm": 2.690351994232397, "learning_rate": 9.987290982657961e-06, "loss": 0.6848, "step": 225 }, { "epoch": 0.12358946802794196, "grad_norm": 3.5601408393747445, "learning_rate": 9.983728407409565e-06, "loss": 0.6832, "step": 230 }, { "epoch": 0.1262761955937668, "grad_norm": 2.4701820214975014, "learning_rate": 9.979727018613607e-06, "loss": 0.6934, "step": 235 }, { "epoch": 0.12896292315959162, "grad_norm": 2.4626384928709983, "learning_rate": 9.975287168589369e-06, "loss": 0.6785, "step": 240 }, { "epoch": 0.13164965072541646, "grad_norm": 2.588440598744194, "learning_rate": 9.970409248262314e-06, "loss": 0.6978, "step": 245 }, { "epoch": 0.13433637829124126, "grad_norm": 2.6880919710650413, "learning_rate": 9.965093687129669e-06, "loss": 0.6808, "step": 250 }, { "epoch": 0.1370231058570661, "grad_norm": 2.9087730515287955, "learning_rate": 9.959340953222602e-06, "loss": 0.6758, "step": 255 }, { "epoch": 0.13970983342289092, "grad_norm": 2.6772738824086075, "learning_rate": 9.953151553065019e-06, "loss": 0.6698, "step": 260 }, { "epoch": 0.14239656098871575, "grad_norm": 2.669218732871304, "learning_rate": 9.94652603162896e-06, "loss": 0.6939, "step": 265 }, { "epoch": 0.14508328855454056, "grad_norm": 2.4902462796339457, "learning_rate": 9.939464972286618e-06, "loss": 0.6702, "step": 270 }, { "epoch": 0.1477700161203654, "grad_norm": 2.4766182391586975, "learning_rate": 9.931968996758972e-06, "loss": 0.6697, "step": 275 }, { "epoch": 0.15045674368619022, "grad_norm": 2.5255485155595228, "learning_rate": 9.924038765061042e-06, "loss": 0.6665, "step": 280 }, { "epoch": 0.15314347125201505, "grad_norm": 2.5656888228811794, "learning_rate": 9.915674975443779e-06, "loss": 0.6565, "step": 285 }, { "epoch": 0.15583019881783988, "grad_norm": 2.825801360492887, "learning_rate": 9.906878364332586e-06, "loss": 0.6642, "step": 290 }, { "epoch": 0.15851692638366469, "grad_norm": 2.6497249215881253, "learning_rate": 9.897649706262474e-06, "loss": 0.6624, "step": 295 }, { "epoch": 0.16120365394948952, "grad_norm": 2.6523672302785575, "learning_rate": 9.88798981380986e-06, "loss": 0.6659, "step": 300 }, { "epoch": 0.16389038151531435, "grad_norm": 2.6257923259309495, "learning_rate": 9.877899537521028e-06, "loss": 0.6533, "step": 305 }, { "epoch": 0.16657710908113918, "grad_norm": 2.465767578464491, "learning_rate": 9.867379765837237e-06, "loss": 0.6518, "step": 310 }, { "epoch": 0.169263836646964, "grad_norm": 2.556282218650964, "learning_rate": 9.85643142501649e-06, "loss": 0.6559, "step": 315 }, { "epoch": 0.17195056421278881, "grad_norm": 2.6286087389340227, "learning_rate": 9.845055479051986e-06, "loss": 0.6469, "step": 320 }, { "epoch": 0.17463729177861365, "grad_norm": 2.419535618698552, "learning_rate": 9.833252929587231e-06, "loss": 0.6631, "step": 325 }, { "epoch": 0.17732401934443848, "grad_norm": 2.4542391410254956, "learning_rate": 9.821024815827854e-06, "loss": 0.6591, "step": 330 }, { "epoch": 0.1800107469102633, "grad_norm": 2.4908873755968997, "learning_rate": 9.808372214450093e-06, "loss": 0.6658, "step": 335 }, { "epoch": 0.1826974744760881, "grad_norm": 2.3733848426387176, "learning_rate": 9.795296239506011e-06, "loss": 0.662, "step": 340 }, { "epoch": 0.18538420204191294, "grad_norm": 2.865296798119886, "learning_rate": 9.781798042325392e-06, "loss": 0.6526, "step": 345 }, { "epoch": 0.18807092960773777, "grad_norm": 2.487445934484789, "learning_rate": 9.767878811414373e-06, "loss": 0.6525, "step": 350 }, { "epoch": 0.1907576571735626, "grad_norm": 2.388699669948045, "learning_rate": 9.753539772350792e-06, "loss": 0.647, "step": 355 }, { "epoch": 0.19344438473938744, "grad_norm": 2.445554737991476, "learning_rate": 9.738782187676282e-06, "loss": 0.6425, "step": 360 }, { "epoch": 0.19613111230521224, "grad_norm": 2.435104584478995, "learning_rate": 9.723607356785103e-06, "loss": 0.6443, "step": 365 }, { "epoch": 0.19881783987103707, "grad_norm": 2.479158175926532, "learning_rate": 9.70801661580973e-06, "loss": 0.6575, "step": 370 }, { "epoch": 0.2015045674368619, "grad_norm": 2.637755494092567, "learning_rate": 9.692011337503212e-06, "loss": 0.6421, "step": 375 }, { "epoch": 0.20419129500268673, "grad_norm": 2.520737872507786, "learning_rate": 9.675592931118293e-06, "loss": 0.6358, "step": 380 }, { "epoch": 0.20687802256851157, "grad_norm": 2.3907603747141675, "learning_rate": 9.658762842283343e-06, "loss": 0.6285, "step": 385 }, { "epoch": 0.20956475013433637, "grad_norm": 2.3651549722988023, "learning_rate": 9.641522552875055e-06, "loss": 0.6119, "step": 390 }, { "epoch": 0.2122514777001612, "grad_norm": 2.5208115737245604, "learning_rate": 9.62387358088798e-06, "loss": 0.6333, "step": 395 }, { "epoch": 0.21493820526598603, "grad_norm": 2.6819070319357037, "learning_rate": 9.605817480300863e-06, "loss": 0.6205, "step": 400 }, { "epoch": 0.21762493283181086, "grad_norm": 2.356493009815594, "learning_rate": 9.587355840939813e-06, "loss": 0.64, "step": 405 }, { "epoch": 0.22031166039763567, "grad_norm": 2.405371194871644, "learning_rate": 9.568490288338324e-06, "loss": 0.6245, "step": 410 }, { "epoch": 0.2229983879634605, "grad_norm": 2.6012482105829706, "learning_rate": 9.549222483594154e-06, "loss": 0.6451, "step": 415 }, { "epoch": 0.22568511552928533, "grad_norm": 2.3956259391703196, "learning_rate": 9.529554123223053e-06, "loss": 0.6223, "step": 420 }, { "epoch": 0.22837184309511016, "grad_norm": 2.398583860789219, "learning_rate": 9.5094869390094e-06, "loss": 0.6427, "step": 425 }, { "epoch": 0.231058570660935, "grad_norm": 2.5232663534614863, "learning_rate": 9.48902269785371e-06, "loss": 0.6264, "step": 430 }, { "epoch": 0.2337452982267598, "grad_norm": 2.4227785562286033, "learning_rate": 9.468163201617063e-06, "loss": 0.6172, "step": 435 }, { "epoch": 0.23643202579258463, "grad_norm": 2.4898638907719586, "learning_rate": 9.446910286962453e-06, "loss": 0.6186, "step": 440 }, { "epoch": 0.23911875335840946, "grad_norm": 2.3600937624812524, "learning_rate": 9.425265825193077e-06, "loss": 0.6009, "step": 445 }, { "epoch": 0.2418054809242343, "grad_norm": 2.479426617963906, "learning_rate": 9.403231722087554e-06, "loss": 0.6373, "step": 450 }, { "epoch": 0.24449220849005912, "grad_norm": 2.4357419818938117, "learning_rate": 9.380809917732132e-06, "loss": 0.6076, "step": 455 }, { "epoch": 0.24717893605588392, "grad_norm": 2.47734706096724, "learning_rate": 9.358002386349862e-06, "loss": 0.6052, "step": 460 }, { "epoch": 0.24986566362170876, "grad_norm": 2.545179042557431, "learning_rate": 9.334811136126778e-06, "loss": 0.6067, "step": 465 }, { "epoch": 0.2525523911875336, "grad_norm": 2.576572600102202, "learning_rate": 9.31123820903506e-06, "loss": 0.6035, "step": 470 }, { "epoch": 0.2552391187533584, "grad_norm": 2.4612256004699145, "learning_rate": 9.287285680653254e-06, "loss": 0.606, "step": 475 }, { "epoch": 0.25792584631918325, "grad_norm": 2.495237864289767, "learning_rate": 9.262955659983522e-06, "loss": 0.5942, "step": 480 }, { "epoch": 0.2606125738850081, "grad_norm": 2.65866167773882, "learning_rate": 9.238250289265921e-06, "loss": 0.599, "step": 485 }, { "epoch": 0.2632993014508329, "grad_norm": 2.473870495050297, "learning_rate": 9.21317174378982e-06, "loss": 0.612, "step": 490 }, { "epoch": 0.2659860290166577, "grad_norm": 2.3022323203953, "learning_rate": 9.187722231702326e-06, "loss": 0.5794, "step": 495 }, { "epoch": 0.2686727565824825, "grad_norm": 2.3360448253147643, "learning_rate": 9.161903993813892e-06, "loss": 0.5821, "step": 500 }, { "epoch": 0.27135948414830735, "grad_norm": 2.4400501057412454, "learning_rate": 9.135719303400995e-06, "loss": 0.5969, "step": 505 }, { "epoch": 0.2740462117141322, "grad_norm": 2.324402047254847, "learning_rate": 9.10917046600598e-06, "loss": 0.5914, "step": 510 }, { "epoch": 0.276732939279957, "grad_norm": 2.4328149953899296, "learning_rate": 9.082259819234063e-06, "loss": 0.5871, "step": 515 }, { "epoch": 0.27941966684578184, "grad_norm": 2.326659710915374, "learning_rate": 9.054989732547507e-06, "loss": 0.5874, "step": 520 }, { "epoch": 0.2821063944116067, "grad_norm": 2.4409777215420942, "learning_rate": 9.027362607056986e-06, "loss": 0.5964, "step": 525 }, { "epoch": 0.2847931219774315, "grad_norm": 2.412955363836415, "learning_rate": 8.999380875310176e-06, "loss": 0.5877, "step": 530 }, { "epoch": 0.28747984954325634, "grad_norm": 2.8194077899428516, "learning_rate": 8.971047001077561e-06, "loss": 0.5883, "step": 535 }, { "epoch": 0.2901665771090811, "grad_norm": 2.469304660089161, "learning_rate": 8.942363479135516e-06, "loss": 0.5801, "step": 540 }, { "epoch": 0.29285330467490595, "grad_norm": 2.3564533376477192, "learning_rate": 8.913332835046629e-06, "loss": 0.576, "step": 545 }, { "epoch": 0.2955400322407308, "grad_norm": 2.4491464187209067, "learning_rate": 8.883957624937333e-06, "loss": 0.5792, "step": 550 }, { "epoch": 0.2982267598065556, "grad_norm": 2.3159629602291276, "learning_rate": 8.854240435272842e-06, "loss": 0.5617, "step": 555 }, { "epoch": 0.30091348737238044, "grad_norm": 2.5317785480271584, "learning_rate": 8.824183882629411e-06, "loss": 0.5669, "step": 560 }, { "epoch": 0.30360021493820527, "grad_norm": 2.3852078136194135, "learning_rate": 8.793790613463956e-06, "loss": 0.5697, "step": 565 }, { "epoch": 0.3062869425040301, "grad_norm": 2.4921927583022168, "learning_rate": 8.763063303881021e-06, "loss": 0.5796, "step": 570 }, { "epoch": 0.30897367006985493, "grad_norm": 9.308745310175, "learning_rate": 8.73200465939717e-06, "loss": 0.5735, "step": 575 }, { "epoch": 0.31166039763567976, "grad_norm": 2.4517036683344458, "learning_rate": 8.700617414702746e-06, "loss": 0.5671, "step": 580 }, { "epoch": 0.31434712520150454, "grad_norm": 2.2949061651822635, "learning_rate": 8.668904333421098e-06, "loss": 0.5696, "step": 585 }, { "epoch": 0.31703385276732937, "grad_norm": 2.2916177476898096, "learning_rate": 8.636868207865244e-06, "loss": 0.5756, "step": 590 }, { "epoch": 0.3197205803331542, "grad_norm": 2.5518854172821297, "learning_rate": 8.604511858792006e-06, "loss": 0.5622, "step": 595 }, { "epoch": 0.32240730789897903, "grad_norm": 2.4475160043733797, "learning_rate": 8.571838135153645e-06, "loss": 0.5525, "step": 600 }, { "epoch": 0.32509403546480387, "grad_norm": 2.3488225651794497, "learning_rate": 8.538849913847019e-06, "loss": 0.569, "step": 605 }, { "epoch": 0.3277807630306287, "grad_norm": 2.353785375547004, "learning_rate": 8.505550099460264e-06, "loss": 0.5575, "step": 610 }, { "epoch": 0.33046749059645353, "grad_norm": 2.5058455114214166, "learning_rate": 8.471941624017058e-06, "loss": 0.5497, "step": 615 }, { "epoch": 0.33315421816227836, "grad_norm": 2.4037582668437674, "learning_rate": 8.43802744671845e-06, "loss": 0.5592, "step": 620 }, { "epoch": 0.3358409457281032, "grad_norm": 2.3703138788584206, "learning_rate": 8.403810553682307e-06, "loss": 0.55, "step": 625 }, { "epoch": 0.338527673293928, "grad_norm": 2.4422133082637605, "learning_rate": 8.369293957680397e-06, "loss": 0.5451, "step": 630 }, { "epoch": 0.3412144008597528, "grad_norm": 2.372256863029062, "learning_rate": 8.334480697873101e-06, "loss": 0.5616, "step": 635 }, { "epoch": 0.34390112842557763, "grad_norm": 2.30267735280899, "learning_rate": 8.299373839541829e-06, "loss": 0.547, "step": 640 }, { "epoch": 0.34658785599140246, "grad_norm": 2.312324577995851, "learning_rate": 8.26397647381912e-06, "loss": 0.5406, "step": 645 }, { "epoch": 0.3492745835572273, "grad_norm": 2.4374915255270175, "learning_rate": 8.228291717416472e-06, "loss": 0.5388, "step": 650 }, { "epoch": 0.3519613111230521, "grad_norm": 2.3871133470130674, "learning_rate": 8.192322712349917e-06, "loss": 0.5432, "step": 655 }, { "epoch": 0.35464803868887695, "grad_norm": 2.519969391683205, "learning_rate": 8.15607262566337e-06, "loss": 0.5348, "step": 660 }, { "epoch": 0.3573347662547018, "grad_norm": 2.4866718815954623, "learning_rate": 8.119544649149762e-06, "loss": 0.5567, "step": 665 }, { "epoch": 0.3600214938205266, "grad_norm": 2.401099368264956, "learning_rate": 8.08274199907003e-06, "loss": 0.5409, "step": 670 }, { "epoch": 0.36270822138635145, "grad_norm": 2.320605485178993, "learning_rate": 8.0456679158699e-06, "loss": 0.5283, "step": 675 }, { "epoch": 0.3653949489521762, "grad_norm": 2.3005736041616642, "learning_rate": 8.008325663894586e-06, "loss": 0.5368, "step": 680 }, { "epoch": 0.36808167651800106, "grad_norm": 2.305647834819274, "learning_rate": 7.970718531101365e-06, "loss": 0.5398, "step": 685 }, { "epoch": 0.3707684040838259, "grad_norm": 2.3268228923372263, "learning_rate": 7.932849828770062e-06, "loss": 0.5348, "step": 690 }, { "epoch": 0.3734551316496507, "grad_norm": 2.2813438553059053, "learning_rate": 7.89472289121151e-06, "loss": 0.5602, "step": 695 }, { "epoch": 0.37614185921547555, "grad_norm": 2.2064861994852816, "learning_rate": 7.856341075473963e-06, "loss": 0.5227, "step": 700 }, { "epoch": 0.3788285867813004, "grad_norm": 2.385886752618868, "learning_rate": 7.817707761047498e-06, "loss": 0.5292, "step": 705 }, { "epoch": 0.3815153143471252, "grad_norm": 2.3717077533920867, "learning_rate": 7.77882634956647e-06, "loss": 0.5332, "step": 710 }, { "epoch": 0.38420204191295004, "grad_norm": 2.389159454524817, "learning_rate": 7.739700264509993e-06, "loss": 0.5236, "step": 715 }, { "epoch": 0.3868887694787749, "grad_norm": 2.314814719990629, "learning_rate": 7.700332950900504e-06, "loss": 0.5217, "step": 720 }, { "epoch": 0.38957549704459965, "grad_norm": 2.207891685389126, "learning_rate": 7.660727875000432e-06, "loss": 0.5078, "step": 725 }, { "epoch": 0.3922622246104245, "grad_norm": 2.3264819673930432, "learning_rate": 7.6208885240069995e-06, "loss": 0.5267, "step": 730 }, { "epoch": 0.3949489521762493, "grad_norm": 2.288819911752258, "learning_rate": 7.5808184057451765e-06, "loss": 0.5067, "step": 735 }, { "epoch": 0.39763567974207414, "grad_norm": 2.3653593555899857, "learning_rate": 7.540521048358814e-06, "loss": 0.5344, "step": 740 }, { "epoch": 0.400322407307899, "grad_norm": 2.4648359924535743, "learning_rate": 7.500000000000001e-06, "loss": 0.5227, "step": 745 }, { "epoch": 0.4030091348737238, "grad_norm": 2.409473996165666, "learning_rate": 7.459258828516645e-06, "loss": 0.5096, "step": 750 }, { "epoch": 0.40569586243954864, "grad_norm": 2.3668323179746236, "learning_rate": 7.418301121138335e-06, "loss": 0.5171, "step": 755 }, { "epoch": 0.40838259000537347, "grad_norm": 2.4782926067309825, "learning_rate": 7.3771304841604764e-06, "loss": 0.51, "step": 760 }, { "epoch": 0.4110693175711983, "grad_norm": 2.267794928141394, "learning_rate": 7.335750542626772e-06, "loss": 0.5245, "step": 765 }, { "epoch": 0.41375604513702313, "grad_norm": 2.318079683575523, "learning_rate": 7.294164940010031e-06, "loss": 0.5124, "step": 770 }, { "epoch": 0.4164427727028479, "grad_norm": 2.1909268149999988, "learning_rate": 7.2523773378913655e-06, "loss": 0.5168, "step": 775 }, { "epoch": 0.41912950026867274, "grad_norm": 2.2456539972858773, "learning_rate": 7.210391415637797e-06, "loss": 0.4946, "step": 780 }, { "epoch": 0.42181622783449757, "grad_norm": 2.4974603199079133, "learning_rate": 7.168210870078277e-06, "loss": 0.5064, "step": 785 }, { "epoch": 0.4245029554003224, "grad_norm": 2.334444583921718, "learning_rate": 7.125839415178204e-06, "loss": 0.5116, "step": 790 }, { "epoch": 0.42718968296614723, "grad_norm": 2.397330148256394, "learning_rate": 7.083280781712394e-06, "loss": 0.5081, "step": 795 }, { "epoch": 0.42987641053197206, "grad_norm": 2.5416270893996273, "learning_rate": 7.0405387169365965e-06, "loss": 0.499, "step": 800 }, { "epoch": 0.4325631380977969, "grad_norm": 2.3363347223952657, "learning_rate": 6.9976169842575526e-06, "loss": 0.5049, "step": 805 }, { "epoch": 0.4352498656636217, "grad_norm": 2.296216857638828, "learning_rate": 6.9545193629016215e-06, "loss": 0.5168, "step": 810 }, { "epoch": 0.43793659322944656, "grad_norm": 2.4900891751826237, "learning_rate": 6.911249647582036e-06, "loss": 0.4939, "step": 815 }, { "epoch": 0.44062332079527133, "grad_norm": 2.3708479871646797, "learning_rate": 6.867811648164769e-06, "loss": 0.4968, "step": 820 }, { "epoch": 0.44331004836109617, "grad_norm": 2.2942739105778007, "learning_rate": 6.824209189333082e-06, "loss": 0.4892, "step": 825 }, { "epoch": 0.445996775926921, "grad_norm": 2.298491349290504, "learning_rate": 6.780446110250766e-06, "loss": 0.5064, "step": 830 }, { "epoch": 0.4486835034927458, "grad_norm": 2.3571392373027624, "learning_rate": 6.736526264224101e-06, "loss": 0.5005, "step": 835 }, { "epoch": 0.45137023105857066, "grad_norm": 2.459885767558715, "learning_rate": 6.692453518362587e-06, "loss": 0.4951, "step": 840 }, { "epoch": 0.4540569586243955, "grad_norm": 2.411449421942393, "learning_rate": 6.648231753238431e-06, "loss": 0.4828, "step": 845 }, { "epoch": 0.4567436861902203, "grad_norm": 2.404363576863236, "learning_rate": 6.603864862544879e-06, "loss": 0.493, "step": 850 }, { "epoch": 0.45943041375604515, "grad_norm": 2.2737487815721478, "learning_rate": 6.5593567527533715e-06, "loss": 0.4744, "step": 855 }, { "epoch": 0.46211714132187, "grad_norm": 2.3678978985955186, "learning_rate": 6.514711342769588e-06, "loss": 0.5012, "step": 860 }, { "epoch": 0.4648038688876948, "grad_norm": 2.356303245264604, "learning_rate": 6.469932563588386e-06, "loss": 0.4973, "step": 865 }, { "epoch": 0.4674905964535196, "grad_norm": 2.7020262831984527, "learning_rate": 6.425024357947677e-06, "loss": 0.4918, "step": 870 }, { "epoch": 0.4701773240193444, "grad_norm": 2.3340485214143745, "learning_rate": 6.3799906799812805e-06, "loss": 0.4982, "step": 875 }, { "epoch": 0.47286405158516925, "grad_norm": 2.4191389152178373, "learning_rate": 6.334835494870759e-06, "loss": 0.4751, "step": 880 }, { "epoch": 0.4755507791509941, "grad_norm": 2.272119239990629, "learning_rate": 6.289562778496285e-06, "loss": 0.4872, "step": 885 }, { "epoch": 0.4782375067168189, "grad_norm": 2.300477262266863, "learning_rate": 6.244176517086573e-06, "loss": 0.4821, "step": 890 }, { "epoch": 0.48092423428264375, "grad_norm": 2.446204046472319, "learning_rate": 6.1986807068678926e-06, "loss": 0.4823, "step": 895 }, { "epoch": 0.4836109618484686, "grad_norm": 2.290872783259082, "learning_rate": 6.153079353712201e-06, "loss": 0.4683, "step": 900 }, { "epoch": 0.4862976894142934, "grad_norm": 2.311264668993125, "learning_rate": 6.107376472784438e-06, "loss": 0.4759, "step": 905 }, { "epoch": 0.48898441698011824, "grad_norm": 2.1766053144740805, "learning_rate": 6.061576088188981e-06, "loss": 0.4541, "step": 910 }, { "epoch": 0.491671144545943, "grad_norm": 2.2863598186899563, "learning_rate": 6.015682232615336e-06, "loss": 0.4751, "step": 915 }, { "epoch": 0.49435787211176785, "grad_norm": 2.3157344313618453, "learning_rate": 5.969698946983055e-06, "loss": 0.4579, "step": 920 }, { "epoch": 0.4970445996775927, "grad_norm": 2.4556667211388308, "learning_rate": 5.923630280085948e-06, "loss": 0.4667, "step": 925 }, { "epoch": 0.4997313272434175, "grad_norm": 2.2693647056660575, "learning_rate": 5.877480288235569e-06, "loss": 0.4642, "step": 930 }, { "epoch": 0.5024180548092424, "grad_norm": 2.299235513307737, "learning_rate": 5.831253034904083e-06, "loss": 0.4625, "step": 935 }, { "epoch": 0.5051047823750672, "grad_norm": 2.2594758653976603, "learning_rate": 5.7849525903664636e-06, "loss": 0.4741, "step": 940 }, { "epoch": 0.507791509940892, "grad_norm": 2.346200404836091, "learning_rate": 5.738583031342123e-06, "loss": 0.4533, "step": 945 }, { "epoch": 0.5104782375067168, "grad_norm": 2.198398726512657, "learning_rate": 5.692148440635946e-06, "loss": 0.4561, "step": 950 }, { "epoch": 0.5131649650725416, "grad_norm": 2.3681691462263466, "learning_rate": 5.645652906778808e-06, "loss": 0.4621, "step": 955 }, { "epoch": 0.5158516926383665, "grad_norm": 2.344944047074419, "learning_rate": 5.599100523667586e-06, "loss": 0.4599, "step": 960 }, { "epoch": 0.5185384202041913, "grad_norm": 2.2092289217419085, "learning_rate": 5.552495390204691e-06, "loss": 0.4722, "step": 965 }, { "epoch": 0.5212251477700162, "grad_norm": 2.3297576895231606, "learning_rate": 5.505841609937162e-06, "loss": 0.4532, "step": 970 }, { "epoch": 0.5239118753358409, "grad_norm": 2.267324308428592, "learning_rate": 5.4591432906953515e-06, "loss": 0.4527, "step": 975 }, { "epoch": 0.5265986029016658, "grad_norm": 2.4550058840280884, "learning_rate": 5.412404544231235e-06, "loss": 0.4562, "step": 980 }, { "epoch": 0.5292853304674906, "grad_norm": 2.268535616599798, "learning_rate": 5.365629485856381e-06, "loss": 0.4454, "step": 985 }, { "epoch": 0.5319720580333154, "grad_norm": 2.598323262451839, "learning_rate": 5.318822234079584e-06, "loss": 0.4614, "step": 990 }, { "epoch": 0.5346587855991403, "grad_norm": 6.381482924500483, "learning_rate": 5.271986910244254e-06, "loss": 0.4366, "step": 995 }, { "epoch": 0.537345513164965, "grad_norm": 2.359236695134924, "learning_rate": 5.225127638165514e-06, "loss": 0.4535, "step": 1000 }, { "epoch": 0.5400322407307899, "grad_norm": 2.268448533208311, "learning_rate": 5.178248543767122e-06, "loss": 0.4342, "step": 1005 }, { "epoch": 0.5427189682966147, "grad_norm": 2.358403761646346, "learning_rate": 5.1313537547181716e-06, "loss": 0.4405, "step": 1010 }, { "epoch": 0.5454056958624396, "grad_norm": 2.2710106949262805, "learning_rate": 5.084447400069656e-06, "loss": 0.4419, "step": 1015 }, { "epoch": 0.5480924234282644, "grad_norm": 2.368278941060436, "learning_rate": 5.037533609890917e-06, "loss": 0.4364, "step": 1020 }, { "epoch": 0.5507791509940893, "grad_norm": 2.284175829190933, "learning_rate": 4.990616514905982e-06, "loss": 0.4452, "step": 1025 }, { "epoch": 0.553465878559914, "grad_norm": 2.41532321709073, "learning_rate": 4.943700246129871e-06, "loss": 0.436, "step": 1030 }, { "epoch": 0.5561526061257388, "grad_norm": 2.3211025544242787, "learning_rate": 4.896788934504853e-06, "loss": 0.4467, "step": 1035 }, { "epoch": 0.5588393336915637, "grad_norm": 2.347444291429596, "learning_rate": 4.849886710536725e-06, "loss": 0.4288, "step": 1040 }, { "epoch": 0.5615260612573885, "grad_norm": 2.4127644606277014, "learning_rate": 4.802997703931124e-06, "loss": 0.4355, "step": 1045 }, { "epoch": 0.5642127888232134, "grad_norm": 2.4325621488004776, "learning_rate": 4.7561260432299015e-06, "loss": 0.4551, "step": 1050 }, { "epoch": 0.5668995163890381, "grad_norm": 2.242665015732091, "learning_rate": 4.7092758554476215e-06, "loss": 0.4364, "step": 1055 }, { "epoch": 0.569586243954863, "grad_norm": 2.4289606347980692, "learning_rate": 4.662451265708174e-06, "loss": 0.4316, "step": 1060 }, { "epoch": 0.5722729715206878, "grad_norm": 2.2828904861500257, "learning_rate": 4.6156563968815575e-06, "loss": 0.4298, "step": 1065 }, { "epoch": 0.5749596990865127, "grad_norm": 2.320278023559682, "learning_rate": 4.568895369220868e-06, "loss": 0.4332, "step": 1070 }, { "epoch": 0.5776464266523375, "grad_norm": 2.228672536344676, "learning_rate": 4.52217229999951e-06, "loss": 0.4291, "step": 1075 }, { "epoch": 0.5803331542181622, "grad_norm": 2.2818442860821664, "learning_rate": 4.47549130314868e-06, "loss": 0.4254, "step": 1080 }, { "epoch": 0.5830198817839871, "grad_norm": 2.218722467460458, "learning_rate": 4.428856488895128e-06, "loss": 0.441, "step": 1085 }, { "epoch": 0.5857066093498119, "grad_norm": 2.2852430401184907, "learning_rate": 4.382271963399268e-06, "loss": 0.4201, "step": 1090 }, { "epoch": 0.5883933369156368, "grad_norm": 2.2551743568535416, "learning_rate": 4.33574182839362e-06, "loss": 0.4284, "step": 1095 }, { "epoch": 0.5910800644814616, "grad_norm": 2.161938070818427, "learning_rate": 4.28927018082167e-06, "loss": 0.4274, "step": 1100 }, { "epoch": 0.5937667920472864, "grad_norm": 2.4061344797719753, "learning_rate": 4.2428611124771184e-06, "loss": 0.4141, "step": 1105 }, { "epoch": 0.5964535196131112, "grad_norm": 2.325443484824538, "learning_rate": 4.19651870964362e-06, "loss": 0.4286, "step": 1110 }, { "epoch": 0.5991402471789361, "grad_norm": 2.3440359985504133, "learning_rate": 4.150247052734979e-06, "loss": 0.4302, "step": 1115 }, { "epoch": 0.6018269747447609, "grad_norm": 2.4615226071875567, "learning_rate": 4.104050215935875e-06, "loss": 0.4111, "step": 1120 }, { "epoch": 0.6045137023105857, "grad_norm": 2.4429242324006983, "learning_rate": 4.0579322668431295e-06, "loss": 0.4193, "step": 1125 }, { "epoch": 0.6072004298764105, "grad_norm": 2.3015769237979677, "learning_rate": 4.011897266107567e-06, "loss": 0.42, "step": 1130 }, { "epoch": 0.6098871574422353, "grad_norm": 2.1719410742796, "learning_rate": 3.965949267076465e-06, "loss": 0.4027, "step": 1135 }, { "epoch": 0.6125738850080602, "grad_norm": 2.192156669745234, "learning_rate": 3.9200923154366685e-06, "loss": 0.423, "step": 1140 }, { "epoch": 0.615260612573885, "grad_norm": 2.2811472307718414, "learning_rate": 3.874330448858369e-06, "loss": 0.4142, "step": 1145 }, { "epoch": 0.6179473401397099, "grad_norm": 2.116499798174077, "learning_rate": 3.8286676966395895e-06, "loss": 0.4107, "step": 1150 }, { "epoch": 0.6206340677055346, "grad_norm": 2.244666553521093, "learning_rate": 3.7831080793514065e-06, "loss": 0.4017, "step": 1155 }, { "epoch": 0.6233207952713595, "grad_norm": 2.4675679743826295, "learning_rate": 3.7376556084839465e-06, "loss": 0.4059, "step": 1160 }, { "epoch": 0.6260075228371843, "grad_norm": 2.443117409755618, "learning_rate": 3.692314286093167e-06, "loss": 0.4162, "step": 1165 }, { "epoch": 0.6286942504030091, "grad_norm": 2.2029356738105363, "learning_rate": 3.647088104448494e-06, "loss": 0.4045, "step": 1170 }, { "epoch": 0.631380977968834, "grad_norm": 2.2250689532361045, "learning_rate": 3.601981045681292e-06, "loss": 0.4035, "step": 1175 }, { "epoch": 0.6340677055346587, "grad_norm": 2.3017262536884395, "learning_rate": 3.556997081434248e-06, "loss": 0.4043, "step": 1180 }, { "epoch": 0.6367544331004836, "grad_norm": 2.242688700397434, "learning_rate": 3.5121401725116653e-06, "loss": 0.405, "step": 1185 }, { "epoch": 0.6394411606663084, "grad_norm": 2.0746209840311747, "learning_rate": 3.4674142685307264e-06, "loss": 0.4086, "step": 1190 }, { "epoch": 0.6421278882321333, "grad_norm": 2.2459358965322993, "learning_rate": 3.4228233075737225e-06, "loss": 0.4121, "step": 1195 }, { "epoch": 0.6448146157979581, "grad_norm": 2.1930816457650235, "learning_rate": 3.3783712158413163e-06, "loss": 0.395, "step": 1200 }, { "epoch": 0.647501343363783, "grad_norm": 2.0405182523892145, "learning_rate": 3.3340619073068347e-06, "loss": 0.3865, "step": 1205 }, { "epoch": 0.6501880709296077, "grad_norm": 2.280921131796164, "learning_rate": 3.289899283371657e-06, "loss": 0.3905, "step": 1210 }, { "epoch": 0.6528747984954326, "grad_norm": 2.297801391720408, "learning_rate": 3.2458872325216893e-06, "loss": 0.3992, "step": 1215 }, { "epoch": 0.6555615260612574, "grad_norm": 2.2255966269222363, "learning_rate": 3.202029629984991e-06, "loss": 0.4012, "step": 1220 }, { "epoch": 0.6582482536270822, "grad_norm": 2.159969955127635, "learning_rate": 3.158330337390565e-06, "loss": 0.4008, "step": 1225 }, { "epoch": 0.6609349811929071, "grad_norm": 2.2717513219011636, "learning_rate": 3.1147932024283424e-06, "loss": 0.3915, "step": 1230 }, { "epoch": 0.6636217087587318, "grad_norm": 2.2223909448906247, "learning_rate": 3.071422058510394e-06, "loss": 0.3918, "step": 1235 }, { "epoch": 0.6663084363245567, "grad_norm": 2.3054337829096294, "learning_rate": 3.0282207244334084e-06, "loss": 0.4009, "step": 1240 }, { "epoch": 0.6689951638903815, "grad_norm": 2.2290900583766677, "learning_rate": 2.9851930040424383e-06, "loss": 0.4068, "step": 1245 }, { "epoch": 0.6716818914562064, "grad_norm": 2.1757281052534303, "learning_rate": 2.9423426858959892e-06, "loss": 0.3882, "step": 1250 }, { "epoch": 0.6743686190220312, "grad_norm": 2.1885464432673314, "learning_rate": 2.8996735429324256e-06, "loss": 0.389, "step": 1255 }, { "epoch": 0.677055346587856, "grad_norm": 2.0896855258083464, "learning_rate": 2.8571893321377773e-06, "loss": 0.3856, "step": 1260 }, { "epoch": 0.6797420741536808, "grad_norm": 2.313353541234272, "learning_rate": 2.8148937942149347e-06, "loss": 0.392, "step": 1265 }, { "epoch": 0.6824288017195056, "grad_norm": 2.3338524034127346, "learning_rate": 2.7727906532542783e-06, "loss": 0.3869, "step": 1270 }, { "epoch": 0.6851155292853305, "grad_norm": 2.084691037407018, "learning_rate": 2.7308836164057913e-06, "loss": 0.3826, "step": 1275 }, { "epoch": 0.6878022568511553, "grad_norm": 2.324796342452183, "learning_rate": 2.6891763735526223e-06, "loss": 0.3871, "step": 1280 }, { "epoch": 0.6904889844169801, "grad_norm": 2.160101574703143, "learning_rate": 2.6476725969862227e-06, "loss": 0.3728, "step": 1285 }, { "epoch": 0.6931757119828049, "grad_norm": 2.067721186262042, "learning_rate": 2.6063759410829813e-06, "loss": 0.3811, "step": 1290 }, { "epoch": 0.6958624395486298, "grad_norm": 2.4251135654528646, "learning_rate": 2.565290041982471e-06, "loss": 0.3884, "step": 1295 }, { "epoch": 0.6985491671144546, "grad_norm": 2.1893196908992123, "learning_rate": 2.524418517267283e-06, "loss": 0.3915, "step": 1300 }, { "epoch": 0.7012358946802795, "grad_norm": 2.1923809598940602, "learning_rate": 2.4837649656445117e-06, "loss": 0.3853, "step": 1305 }, { "epoch": 0.7039226222461042, "grad_norm": 2.2626797878823712, "learning_rate": 2.4433329666288774e-06, "loss": 0.3771, "step": 1310 }, { "epoch": 0.706609349811929, "grad_norm": 2.208206845682125, "learning_rate": 2.4031260802275623e-06, "loss": 0.3689, "step": 1315 }, { "epoch": 0.7092960773777539, "grad_norm": 2.18574695336372, "learning_rate": 2.3631478466267498e-06, "loss": 0.3728, "step": 1320 }, { "epoch": 0.7119828049435787, "grad_norm": 2.207451564888781, "learning_rate": 2.323401785879921e-06, "loss": 0.3758, "step": 1325 }, { "epoch": 0.7146695325094036, "grad_norm": 2.168655924857549, "learning_rate": 2.283891397597908e-06, "loss": 0.3672, "step": 1330 }, { "epoch": 0.7173562600752283, "grad_norm": 2.255193862061058, "learning_rate": 2.2446201606407537e-06, "loss": 0.3794, "step": 1335 }, { "epoch": 0.7200429876410532, "grad_norm": 2.2636128693117166, "learning_rate": 2.205591532811416e-06, "loss": 0.3742, "step": 1340 }, { "epoch": 0.722729715206878, "grad_norm": 2.1328893959069117, "learning_rate": 2.166808950551296e-06, "loss": 0.3778, "step": 1345 }, { "epoch": 0.7254164427727029, "grad_norm": 2.1602209025177705, "learning_rate": 2.128275828637664e-06, "loss": 0.3788, "step": 1350 }, { "epoch": 0.7281031703385277, "grad_norm": 2.28424029482692, "learning_rate": 2.089995559883004e-06, "loss": 0.371, "step": 1355 }, { "epoch": 0.7307898979043524, "grad_norm": 2.15439710916817, "learning_rate": 2.0519715148362585e-06, "loss": 0.3724, "step": 1360 }, { "epoch": 0.7334766254701773, "grad_norm": 2.1370354226282355, "learning_rate": 2.0142070414860704e-06, "loss": 0.3551, "step": 1365 }, { "epoch": 0.7361633530360021, "grad_norm": 2.1423156686501836, "learning_rate": 1.976705464965985e-06, "loss": 0.3817, "step": 1370 }, { "epoch": 0.738850080601827, "grad_norm": 2.1767372641376976, "learning_rate": 1.9394700872616856e-06, "loss": 0.3692, "step": 1375 }, { "epoch": 0.7415368081676518, "grad_norm": 2.1148705992310552, "learning_rate": 1.902504186920245e-06, "loss": 0.3701, "step": 1380 }, { "epoch": 0.7442235357334767, "grad_norm": 2.127192293566224, "learning_rate": 1.8658110187614538e-06, "loss": 0.3716, "step": 1385 }, { "epoch": 0.7469102632993014, "grad_norm": 2.1527774862898394, "learning_rate": 1.8293938135912475e-06, "loss": 0.3663, "step": 1390 }, { "epoch": 0.7495969908651263, "grad_norm": 2.3762653649020553, "learning_rate": 1.793255777917217e-06, "loss": 0.356, "step": 1395 }, { "epoch": 0.7522837184309511, "grad_norm": 2.0672220526784444, "learning_rate": 1.7574000936662928e-06, "loss": 0.3666, "step": 1400 }, { "epoch": 0.7549704459967759, "grad_norm": 2.131662760991029, "learning_rate": 1.7218299179045789e-06, "loss": 0.3777, "step": 1405 }, { "epoch": 0.7576571735626008, "grad_norm": 2.3447659932151423, "learning_rate": 1.6865483825593643e-06, "loss": 0.3718, "step": 1410 }, { "epoch": 0.7603439011284255, "grad_norm": 2.393514042224357, "learning_rate": 1.6515585941433694e-06, "loss": 0.3642, "step": 1415 }, { "epoch": 0.7630306286942504, "grad_norm": 2.3166220749115194, "learning_rate": 1.6168636334812126e-06, "loss": 0.3712, "step": 1420 }, { "epoch": 0.7657173562600752, "grad_norm": 2.1832693900107745, "learning_rate": 1.5824665554381579e-06, "loss": 0.3658, "step": 1425 }, { "epoch": 0.7684040838259001, "grad_norm": 2.0779160960958594, "learning_rate": 1.5483703886511191e-06, "loss": 0.3615, "step": 1430 }, { "epoch": 0.7710908113917249, "grad_norm": 2.098593414629146, "learning_rate": 1.5145781352620054e-06, "loss": 0.353, "step": 1435 }, { "epoch": 0.7737775389575497, "grad_norm": 2.1861074294439224, "learning_rate": 1.481092770653374e-06, "loss": 0.3599, "step": 1440 }, { "epoch": 0.7764642665233745, "grad_norm": 2.2448007733369923, "learning_rate": 1.4479172431864647e-06, "loss": 0.3548, "step": 1445 }, { "epoch": 0.7791509940891993, "grad_norm": 2.307957775652367, "learning_rate": 1.4150544739415755e-06, "loss": 0.3549, "step": 1450 }, { "epoch": 0.7818377216550242, "grad_norm": 2.1872047936836903, "learning_rate": 1.382507356460891e-06, "loss": 0.3533, "step": 1455 }, { "epoch": 0.784524449220849, "grad_norm": 2.1557683435766304, "learning_rate": 1.3502787564936875e-06, "loss": 0.3678, "step": 1460 }, { "epoch": 0.7872111767866738, "grad_norm": 2.088183401090359, "learning_rate": 1.3183715117440143e-06, "loss": 0.3452, "step": 1465 }, { "epoch": 0.7898979043524986, "grad_norm": 2.1559096637691195, "learning_rate": 1.2867884316208345e-06, "loss": 0.3605, "step": 1470 }, { "epoch": 0.7925846319183235, "grad_norm": 1.8737663503332451, "learning_rate": 1.255532296990662e-06, "loss": 0.3514, "step": 1475 }, { "epoch": 0.7952713594841483, "grad_norm": 2.196223751578153, "learning_rate": 1.2246058599327021e-06, "loss": 0.3511, "step": 1480 }, { "epoch": 0.7979580870499732, "grad_norm": 2.0822224737026294, "learning_rate": 1.194011843496537e-06, "loss": 0.3653, "step": 1485 }, { "epoch": 0.800644814615798, "grad_norm": 2.2216664507383457, "learning_rate": 1.163752941462362e-06, "loss": 0.3491, "step": 1490 }, { "epoch": 0.8033315421816228, "grad_norm": 2.259893004106475, "learning_rate": 1.1338318181038037e-06, "loss": 0.3495, "step": 1495 }, { "epoch": 0.8060182697474476, "grad_norm": 2.1666650973142687, "learning_rate": 1.1042511079533275e-06, "loss": 0.3611, "step": 1500 }, { "epoch": 0.8087049973132724, "grad_norm": 2.2659364719554898, "learning_rate": 1.0750134155702674e-06, "loss": 0.3626, "step": 1505 }, { "epoch": 0.8113917248790973, "grad_norm": 2.108890981344199, "learning_rate": 1.046121315311508e-06, "loss": 0.3522, "step": 1510 }, { "epoch": 0.814078452444922, "grad_norm": 1.9553754507307322, "learning_rate": 1.017577351104801e-06, "loss": 0.3491, "step": 1515 }, { "epoch": 0.8167651800107469, "grad_norm": 2.042354110902273, "learning_rate": 9.893840362247809e-07, "loss": 0.3489, "step": 1520 }, { "epoch": 0.8194519075765717, "grad_norm": 2.1701934651234613, "learning_rate": 9.615438530716753e-07, "loss": 0.3556, "step": 1525 }, { "epoch": 0.8221386351423966, "grad_norm": 2.298551493046633, "learning_rate": 9.340592529527237e-07, "loss": 0.3533, "step": 1530 }, { "epoch": 0.8248253627082214, "grad_norm": 2.0661919944679656, "learning_rate": 9.069326558663488e-07, "loss": 0.3504, "step": 1535 }, { "epoch": 0.8275120902740463, "grad_norm": 2.238182054477219, "learning_rate": 8.801664502890722e-07, "loss": 0.3583, "step": 1540 }, { "epoch": 0.830198817839871, "grad_norm": 2.206073109503085, "learning_rate": 8.537629929652186e-07, "loss": 0.3493, "step": 1545 }, { "epoch": 0.8328855454056958, "grad_norm": 2.112920269926838, "learning_rate": 8.277246086993962e-07, "loss": 0.3485, "step": 1550 }, { "epoch": 0.8355722729715207, "grad_norm": 2.1091946693423824, "learning_rate": 8.02053590151805e-07, "loss": 0.3602, "step": 1555 }, { "epoch": 0.8382590005373455, "grad_norm": 2.0674911667909592, "learning_rate": 7.767521976363735e-07, "loss": 0.344, "step": 1560 }, { "epoch": 0.8409457281031704, "grad_norm": 2.094630956022192, "learning_rate": 7.518226589217286e-07, "loss": 0.3555, "step": 1565 }, { "epoch": 0.8436324556689951, "grad_norm": 1.9871731688863412, "learning_rate": 7.27267169035053e-07, "loss": 0.3446, "step": 1570 }, { "epoch": 0.84631918323482, "grad_norm": 1.9968082439337067, "learning_rate": 7.030878900688115e-07, "loss": 0.3453, "step": 1575 }, { "epoch": 0.8490059108006448, "grad_norm": 2.0534061294688435, "learning_rate": 6.792869509903777e-07, "loss": 0.3471, "step": 1580 }, { "epoch": 0.8516926383664697, "grad_norm": 2.06210125589499, "learning_rate": 6.558664474545817e-07, "loss": 0.3395, "step": 1585 }, { "epoch": 0.8543793659322945, "grad_norm": 1.9567313896707874, "learning_rate": 6.328284416191893e-07, "loss": 0.3401, "step": 1590 }, { "epoch": 0.8570660934981192, "grad_norm": 2.060413092979503, "learning_rate": 6.101749619633346e-07, "loss": 0.3394, "step": 1595 }, { "epoch": 0.8597528210639441, "grad_norm": 2.1887059434963603, "learning_rate": 5.879080031089047e-07, "loss": 0.3543, "step": 1600 }, { "epoch": 0.8624395486297689, "grad_norm": 1.9845082909122282, "learning_rate": 5.660295256449233e-07, "loss": 0.3389, "step": 1605 }, { "epoch": 0.8651262761955938, "grad_norm": 2.3356240543268414, "learning_rate": 5.445414559549167e-07, "loss": 0.3403, "step": 1610 }, { "epoch": 0.8678130037614186, "grad_norm": 2.2880237510434123, "learning_rate": 5.234456860473042e-07, "loss": 0.333, "step": 1615 }, { "epoch": 0.8704997313272435, "grad_norm": 1.987792837037979, "learning_rate": 5.027440733887973e-07, "loss": 0.3413, "step": 1620 }, { "epoch": 0.8731864588930682, "grad_norm": 2.145351742840335, "learning_rate": 4.824384407408622e-07, "loss": 0.3419, "step": 1625 }, { "epoch": 0.8758731864588931, "grad_norm": 2.0454741447896034, "learning_rate": 4.625305759992205e-07, "loss": 0.3446, "step": 1630 }, { "epoch": 0.8785599140247179, "grad_norm": 2.3653559873352754, "learning_rate": 4.4302223203642803e-07, "loss": 0.3456, "step": 1635 }, { "epoch": 0.8812466415905427, "grad_norm": 1.9465586799987966, "learning_rate": 4.2391512654753443e-07, "loss": 0.3401, "step": 1640 }, { "epoch": 0.8839333691563676, "grad_norm": 2.0641483788200166, "learning_rate": 4.05210941898847e-07, "loss": 0.3444, "step": 1645 }, { "epoch": 0.8866200967221923, "grad_norm": 1.9634013461126347, "learning_rate": 3.8691132497979064e-07, "loss": 0.3262, "step": 1650 }, { "epoch": 0.8893068242880172, "grad_norm": 1.9771681166485953, "learning_rate": 3.6901788705790753e-07, "loss": 0.3339, "step": 1655 }, { "epoch": 0.891993551853842, "grad_norm": 2.0517920199325363, "learning_rate": 3.5153220363698225e-07, "loss": 0.3505, "step": 1660 }, { "epoch": 0.8946802794196669, "grad_norm": 1.9243621922759124, "learning_rate": 3.344558143183246e-07, "loss": 0.342, "step": 1665 }, { "epoch": 0.8973670069854917, "grad_norm": 1.9232396735880564, "learning_rate": 3.1779022266520245e-07, "loss": 0.347, "step": 1670 }, { "epoch": 0.9000537345513165, "grad_norm": 2.0661817058830305, "learning_rate": 3.015368960704584e-07, "loss": 0.3293, "step": 1675 }, { "epoch": 0.9027404621171413, "grad_norm": 2.1100704132509094, "learning_rate": 2.856972656273066e-07, "loss": 0.3331, "step": 1680 }, { "epoch": 0.9054271896829661, "grad_norm": 2.0718216201058444, "learning_rate": 2.7027272600332443e-07, "loss": 0.3422, "step": 1685 }, { "epoch": 0.908113917248791, "grad_norm": 2.1084735618419246, "learning_rate": 2.5526463531765467e-07, "loss": 0.3453, "step": 1690 }, { "epoch": 0.9108006448146158, "grad_norm": 1.8503057057006569, "learning_rate": 2.4067431502142414e-07, "loss": 0.3369, "step": 1695 }, { "epoch": 0.9134873723804406, "grad_norm": 2.2410702930225437, "learning_rate": 2.2650304978138916e-07, "loss": 0.3422, "step": 1700 }, { "epoch": 0.9161740999462654, "grad_norm": 1.988010259435427, "learning_rate": 2.1275208736682262e-07, "loss": 0.3273, "step": 1705 }, { "epoch": 0.9188608275120903, "grad_norm": 2.162508541847929, "learning_rate": 1.9942263853964917e-07, "loss": 0.348, "step": 1710 }, { "epoch": 0.9215475550779151, "grad_norm": 2.173499113968412, "learning_rate": 1.8651587694783924e-07, "loss": 0.3366, "step": 1715 }, { "epoch": 0.92423428264374, "grad_norm": 1.9783121068716565, "learning_rate": 1.7403293902206851e-07, "loss": 0.3282, "step": 1720 }, { "epoch": 0.9269210102095647, "grad_norm": 1.8953184142427286, "learning_rate": 1.6197492387565629e-07, "loss": 0.3346, "step": 1725 }, { "epoch": 0.9296077377753896, "grad_norm": 2.0149463904024474, "learning_rate": 1.503428932077916e-07, "loss": 0.3457, "step": 1730 }, { "epoch": 0.9322944653412144, "grad_norm": 2.0287919597007447, "learning_rate": 1.3913787121004717e-07, "loss": 0.3301, "step": 1735 }, { "epoch": 0.9349811929070392, "grad_norm": 2.049509308073804, "learning_rate": 1.2836084447620466e-07, "loss": 0.3483, "step": 1740 }, { "epoch": 0.9376679204728641, "grad_norm": 2.0687381489986714, "learning_rate": 1.180127619153837e-07, "loss": 0.3296, "step": 1745 }, { "epoch": 0.9403546480386888, "grad_norm": 2.129632171046833, "learning_rate": 1.0809453466849029e-07, "loss": 0.3416, "step": 1750 }, { "epoch": 0.9430413756045137, "grad_norm": 1.940321140330739, "learning_rate": 9.860703602799281e-08, "loss": 0.344, "step": 1755 }, { "epoch": 0.9457281031703385, "grad_norm": 2.2300828043810434, "learning_rate": 8.955110136102952e-08, "loss": 0.3332, "step": 1760 }, { "epoch": 0.9484148307361634, "grad_norm": 2.039077625849431, "learning_rate": 8.092752803585513e-08, "loss": 0.3398, "step": 1765 }, { "epoch": 0.9511015583019882, "grad_norm": 1.9309466845356567, "learning_rate": 7.273707535162988e-08, "loss": 0.3345, "step": 1770 }, { "epoch": 0.9537882858678131, "grad_norm": 2.085038229808163, "learning_rate": 6.498046447156958e-08, "loss": 0.3413, "step": 1775 }, { "epoch": 0.9564750134336378, "grad_norm": 2.073268982177981, "learning_rate": 5.7658378359443104e-08, "loss": 0.3389, "step": 1780 }, { "epoch": 0.9591617409994626, "grad_norm": 2.1191914780783185, "learning_rate": 5.077146171943936e-08, "loss": 0.3301, "step": 1785 }, { "epoch": 0.9618484685652875, "grad_norm": 2.0695293690655494, "learning_rate": 4.432032093940219e-08, "loss": 0.3387, "step": 1790 }, { "epoch": 0.9645351961311123, "grad_norm": 2.181866235925443, "learning_rate": 3.8305524037438035e-08, "loss": 0.3327, "step": 1795 }, { "epoch": 0.9672219236969372, "grad_norm": 2.0829327323896742, "learning_rate": 3.27276006119015e-08, "loss": 0.3269, "step": 1800 }, { "epoch": 0.9699086512627619, "grad_norm": 2.0402878486865084, "learning_rate": 2.7587041794766012e-08, "loss": 0.3428, "step": 1805 }, { "epoch": 0.9725953788285868, "grad_norm": 2.1613154681131386, "learning_rate": 2.2884300208378395e-08, "loss": 0.3389, "step": 1810 }, { "epoch": 0.9752821063944116, "grad_norm": 2.190787226227149, "learning_rate": 1.8619789925608534e-08, "loss": 0.3487, "step": 1815 }, { "epoch": 0.9779688339602365, "grad_norm": 1.9274721887684194, "learning_rate": 1.4793886433387417e-08, "loss": 0.3248, "step": 1820 }, { "epoch": 0.9806555615260613, "grad_norm": 2.0401602016541296, "learning_rate": 1.1406926599646373e-08, "loss": 0.3378, "step": 1825 }, { "epoch": 0.983342289091886, "grad_norm": 1.9245890826575305, "learning_rate": 8.459208643659122e-09, "loss": 0.3301, "step": 1830 }, { "epoch": 0.9860290166577109, "grad_norm": 2.145271841454969, "learning_rate": 5.950992109779452e-09, "loss": 0.3337, "step": 1835 }, { "epoch": 0.9887157442235357, "grad_norm": 2.102840117283832, "learning_rate": 3.88249784459227e-09, "loss": 0.3436, "step": 1840 }, { "epoch": 0.9914024717893606, "grad_norm": 1.8642178563070946, "learning_rate": 2.2539079774658303e-09, "loss": 0.3346, "step": 1845 }, { "epoch": 0.9940891993551854, "grad_norm": 2.0754130978586462, "learning_rate": 1.0653659045156695e-09, "loss": 0.323, "step": 1850 }, { "epoch": 0.9967759269210102, "grad_norm": 2.1177194452551293, "learning_rate": 3.1697627597970794e-10, "loss": 0.3273, "step": 1855 }, { "epoch": 0.999462654486835, "grad_norm": 2.036609175873343, "learning_rate": 8.804987003951937e-12, "loss": 0.3344, "step": 1860 }, { "epoch": 1.0, "eval_runtime": 3.3729, "eval_samples_per_second": 2.965, "eval_steps_per_second": 0.889, "step": 1861 }, { "epoch": 1.0, "step": 1861, "total_flos": 194827769610240.0, "train_loss": 0.5009698675307838, "train_runtime": 16706.09, "train_samples_per_second": 1.782, "train_steps_per_second": 0.111 } ], "logging_steps": 5, "max_steps": 1861, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 194827769610240.0, "train_batch_size": 4, "trial_name": null, "trial_params": null }