{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9993174061433447, "eval_steps": 183, "global_step": 549, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0018202502844141069, "grad_norm": 7.828993836368125, "learning_rate": 4.705882352941176e-07, "loss": 0.7927, "step": 1 }, { "epoch": 0.0036405005688282138, "grad_norm": 6.714284407371177, "learning_rate": 9.411764705882352e-07, "loss": 0.7939, "step": 2 }, { "epoch": 0.005460750853242321, "grad_norm": 6.364056661574126, "learning_rate": 1.411764705882353e-06, "loss": 0.8186, "step": 3 }, { "epoch": 0.0072810011376564275, "grad_norm": 6.001686432641966, "learning_rate": 1.8823529411764705e-06, "loss": 0.7232, "step": 4 }, { "epoch": 0.009101251422070534, "grad_norm": 5.534730246504558, "learning_rate": 2.352941176470588e-06, "loss": 0.7891, "step": 5 }, { "epoch": 0.010921501706484642, "grad_norm": 3.4121020699424713, "learning_rate": 2.823529411764706e-06, "loss": 0.7351, "step": 6 }, { "epoch": 0.01274175199089875, "grad_norm": 3.9324785806724476, "learning_rate": 3.294117647058823e-06, "loss": 0.6899, "step": 7 }, { "epoch": 0.014562002275312855, "grad_norm": 2.5947504275059496, "learning_rate": 3.764705882352941e-06, "loss": 0.6103, "step": 8 }, { "epoch": 0.016382252559726963, "grad_norm": 2.5403086872594955, "learning_rate": 4.235294117647058e-06, "loss": 0.6377, "step": 9 }, { "epoch": 0.01820250284414107, "grad_norm": 2.339440216808695, "learning_rate": 4.705882352941176e-06, "loss": 0.6723, "step": 10 }, { "epoch": 0.020022753128555178, "grad_norm": 2.1983237574211922, "learning_rate": 5.176470588235294e-06, "loss": 0.684, "step": 11 }, { "epoch": 0.021843003412969283, "grad_norm": 1.6424427461297102, "learning_rate": 5.647058823529412e-06, "loss": 0.6027, "step": 12 }, { "epoch": 0.02366325369738339, "grad_norm": 1.6005922788724476, "learning_rate": 6.1176470588235285e-06, "loss": 0.7195, "step": 13 }, { "epoch": 0.0254835039817975, "grad_norm": 1.8438931230594375, "learning_rate": 6.588235294117646e-06, "loss": 0.6329, "step": 14 }, { "epoch": 0.027303754266211604, "grad_norm": 1.8086010126416687, "learning_rate": 7.058823529411764e-06, "loss": 0.6812, "step": 15 }, { "epoch": 0.02912400455062571, "grad_norm": 1.6390713129791323, "learning_rate": 7.529411764705882e-06, "loss": 0.6848, "step": 16 }, { "epoch": 0.03094425483503982, "grad_norm": 1.5725949727834614, "learning_rate": 8e-06, "loss": 0.6415, "step": 17 }, { "epoch": 0.032764505119453925, "grad_norm": 1.5078400807519021, "learning_rate": 7.999930256262932e-06, "loss": 0.6284, "step": 18 }, { "epoch": 0.03458475540386803, "grad_norm": 1.5292651701417765, "learning_rate": 7.999721027483818e-06, "loss": 0.6503, "step": 19 }, { "epoch": 0.03640500568828214, "grad_norm": 1.4983054037314452, "learning_rate": 7.999372320958861e-06, "loss": 0.6167, "step": 20 }, { "epoch": 0.03822525597269624, "grad_norm": 1.4412298959675836, "learning_rate": 7.998884148848109e-06, "loss": 0.6245, "step": 21 }, { "epoch": 0.040045506257110355, "grad_norm": 1.2365167423542114, "learning_rate": 7.998256528175033e-06, "loss": 0.5953, "step": 22 }, { "epoch": 0.04186575654152446, "grad_norm": 1.5357829692377438, "learning_rate": 7.997489480825941e-06, "loss": 0.6367, "step": 23 }, { "epoch": 0.04368600682593857, "grad_norm": 1.3171822289121358, "learning_rate": 7.996583033549204e-06, "loss": 0.5577, "step": 24 }, { "epoch": 0.04550625711035267, "grad_norm": 1.3245922461587984, "learning_rate": 7.995537217954335e-06, "loss": 0.5706, "step": 25 }, { "epoch": 0.04732650739476678, "grad_norm": 1.4166210756319584, "learning_rate": 7.994352070510876e-06, "loss": 0.6612, "step": 26 }, { "epoch": 0.049146757679180884, "grad_norm": 1.4367785667776511, "learning_rate": 7.993027632547137e-06, "loss": 0.5766, "step": 27 }, { "epoch": 0.050967007963595, "grad_norm": 1.2803978303957413, "learning_rate": 7.991563950248739e-06, "loss": 0.6023, "step": 28 }, { "epoch": 0.0527872582480091, "grad_norm": 1.3201964359281728, "learning_rate": 7.989961074657023e-06, "loss": 0.6026, "step": 29 }, { "epoch": 0.05460750853242321, "grad_norm": 1.5466402476684098, "learning_rate": 7.988219061667252e-06, "loss": 0.5979, "step": 30 }, { "epoch": 0.056427758816837315, "grad_norm": 1.372807482234666, "learning_rate": 7.986337972026678e-06, "loss": 0.5928, "step": 31 }, { "epoch": 0.05824800910125142, "grad_norm": 1.2476758007512014, "learning_rate": 7.98431787133241e-06, "loss": 0.5506, "step": 32 }, { "epoch": 0.060068259385665526, "grad_norm": 1.3332407362456573, "learning_rate": 7.982158830029133e-06, "loss": 0.5252, "step": 33 }, { "epoch": 0.06188850967007964, "grad_norm": 1.2956829613552345, "learning_rate": 7.979860923406654e-06, "loss": 0.6162, "step": 34 }, { "epoch": 0.06370875995449374, "grad_norm": 1.3358171325973744, "learning_rate": 7.977424231597266e-06, "loss": 0.6323, "step": 35 }, { "epoch": 0.06552901023890785, "grad_norm": 1.2668022187917536, "learning_rate": 7.97484883957297e-06, "loss": 0.5481, "step": 36 }, { "epoch": 0.06734926052332196, "grad_norm": 1.3032578120954865, "learning_rate": 7.972134837142497e-06, "loss": 0.6982, "step": 37 }, { "epoch": 0.06916951080773606, "grad_norm": 1.438077454657845, "learning_rate": 7.969282318948179e-06, "loss": 0.6386, "step": 38 }, { "epoch": 0.07098976109215017, "grad_norm": 1.2348261155030926, "learning_rate": 7.966291384462662e-06, "loss": 0.5691, "step": 39 }, { "epoch": 0.07281001137656427, "grad_norm": 1.7781144758356542, "learning_rate": 7.963162137985416e-06, "loss": 0.6133, "step": 40 }, { "epoch": 0.07463026166097839, "grad_norm": 1.3915614778891998, "learning_rate": 7.959894688639114e-06, "loss": 0.6097, "step": 41 }, { "epoch": 0.07645051194539249, "grad_norm": 1.3714026109891253, "learning_rate": 7.956489150365818e-06, "loss": 0.7127, "step": 42 }, { "epoch": 0.0782707622298066, "grad_norm": 1.2639378240340353, "learning_rate": 7.952945641923014e-06, "loss": 0.5649, "step": 43 }, { "epoch": 0.08009101251422071, "grad_norm": 1.3658987432003644, "learning_rate": 7.949264286879461e-06, "loss": 0.5975, "step": 44 }, { "epoch": 0.08191126279863481, "grad_norm": 1.3869528512684368, "learning_rate": 7.94544521361089e-06, "loss": 0.5851, "step": 45 }, { "epoch": 0.08373151308304892, "grad_norm": 1.3457362580768018, "learning_rate": 7.941488555295519e-06, "loss": 0.6241, "step": 46 }, { "epoch": 0.08555176336746302, "grad_norm": 1.3016443215121218, "learning_rate": 7.937394449909417e-06, "loss": 0.5603, "step": 47 }, { "epoch": 0.08737201365187713, "grad_norm": 1.4891468132527028, "learning_rate": 7.933163040221691e-06, "loss": 0.6103, "step": 48 }, { "epoch": 0.08919226393629125, "grad_norm": 1.3296810848522813, "learning_rate": 7.928794473789502e-06, "loss": 0.5823, "step": 49 }, { "epoch": 0.09101251422070535, "grad_norm": 1.5647412617424314, "learning_rate": 7.924288902952924e-06, "loss": 0.6222, "step": 50 }, { "epoch": 0.09283276450511946, "grad_norm": 1.224983623587384, "learning_rate": 7.91964648482963e-06, "loss": 0.5779, "step": 51 }, { "epoch": 0.09465301478953356, "grad_norm": 1.3384527596977494, "learning_rate": 7.914867381309417e-06, "loss": 0.5721, "step": 52 }, { "epoch": 0.09647326507394767, "grad_norm": 1.3725597146198896, "learning_rate": 7.909951759048553e-06, "loss": 0.6531, "step": 53 }, { "epoch": 0.09829351535836177, "grad_norm": 1.4825184286499173, "learning_rate": 7.904899789463974e-06, "loss": 0.5767, "step": 54 }, { "epoch": 0.10011376564277588, "grad_norm": 1.2710681313032695, "learning_rate": 7.899711648727295e-06, "loss": 0.5447, "step": 55 }, { "epoch": 0.10193401592719, "grad_norm": 1.468311538305991, "learning_rate": 7.894387517758679e-06, "loss": 0.6303, "step": 56 }, { "epoch": 0.1037542662116041, "grad_norm": 1.1661278644804365, "learning_rate": 7.888927582220521e-06, "loss": 0.606, "step": 57 }, { "epoch": 0.1055745164960182, "grad_norm": 1.3217868795788283, "learning_rate": 7.883332032510978e-06, "loss": 0.5329, "step": 58 }, { "epoch": 0.1073947667804323, "grad_norm": 1.33515774996865, "learning_rate": 7.877601063757322e-06, "loss": 0.5335, "step": 59 }, { "epoch": 0.10921501706484642, "grad_norm": 1.4916296734659389, "learning_rate": 7.871734875809141e-06, "loss": 0.5705, "step": 60 }, { "epoch": 0.11103526734926053, "grad_norm": 1.3630503455015932, "learning_rate": 7.86573367323137e-06, "loss": 0.6279, "step": 61 }, { "epoch": 0.11285551763367463, "grad_norm": 1.2409472798363426, "learning_rate": 7.859597665297158e-06, "loss": 0.5096, "step": 62 }, { "epoch": 0.11467576791808874, "grad_norm": 1.241131166674189, "learning_rate": 7.853327065980567e-06, "loss": 0.5792, "step": 63 }, { "epoch": 0.11649601820250284, "grad_norm": 1.2488483644853932, "learning_rate": 7.84692209394911e-06, "loss": 0.5191, "step": 64 }, { "epoch": 0.11831626848691695, "grad_norm": 1.4362376863954367, "learning_rate": 7.84038297255613e-06, "loss": 0.5749, "step": 65 }, { "epoch": 0.12013651877133105, "grad_norm": 1.301660597552013, "learning_rate": 7.83370992983301e-06, "loss": 0.5598, "step": 66 }, { "epoch": 0.12195676905574517, "grad_norm": 1.2871188302606258, "learning_rate": 7.826903198481218e-06, "loss": 0.6357, "step": 67 }, { "epoch": 0.12377701934015928, "grad_norm": 1.2412889792540172, "learning_rate": 7.819963015864195e-06, "loss": 0.6025, "step": 68 }, { "epoch": 0.12559726962457338, "grad_norm": 1.2417123308251397, "learning_rate": 7.812889623999077e-06, "loss": 0.5973, "step": 69 }, { "epoch": 0.12741751990898748, "grad_norm": 1.334500041209377, "learning_rate": 7.805683269548253e-06, "loss": 0.5339, "step": 70 }, { "epoch": 0.1292377701934016, "grad_norm": 1.2800707133087328, "learning_rate": 7.798344203810772e-06, "loss": 0.5506, "step": 71 }, { "epoch": 0.1310580204778157, "grad_norm": 1.2634004680746123, "learning_rate": 7.790872682713567e-06, "loss": 0.554, "step": 72 }, { "epoch": 0.1328782707622298, "grad_norm": 1.3715962042311087, "learning_rate": 7.783268966802538e-06, "loss": 0.5949, "step": 73 }, { "epoch": 0.13469852104664393, "grad_norm": 1.4002060242886838, "learning_rate": 7.77553332123347e-06, "loss": 0.6422, "step": 74 }, { "epoch": 0.13651877133105803, "grad_norm": 1.2759192431077615, "learning_rate": 7.767666015762775e-06, "loss": 0.607, "step": 75 }, { "epoch": 0.13833902161547212, "grad_norm": 1.6921865669723448, "learning_rate": 7.7596673247381e-06, "loss": 0.6002, "step": 76 }, { "epoch": 0.14015927189988622, "grad_norm": 1.3571851968954738, "learning_rate": 7.751537527088742e-06, "loss": 0.5215, "step": 77 }, { "epoch": 0.14197952218430035, "grad_norm": 1.4785928356534102, "learning_rate": 7.743276906315936e-06, "loss": 0.6101, "step": 78 }, { "epoch": 0.14379977246871445, "grad_norm": 1.465222696303414, "learning_rate": 7.734885750482967e-06, "loss": 0.6187, "step": 79 }, { "epoch": 0.14562002275312855, "grad_norm": 1.264573511241066, "learning_rate": 7.726364352205116e-06, "loss": 0.5673, "step": 80 }, { "epoch": 0.14744027303754267, "grad_norm": 1.238555624330946, "learning_rate": 7.717713008639463e-06, "loss": 0.6066, "step": 81 }, { "epoch": 0.14926052332195677, "grad_norm": 1.20038139697854, "learning_rate": 7.708932021474524e-06, "loss": 0.5678, "step": 82 }, { "epoch": 0.15108077360637087, "grad_norm": 1.3190323958334018, "learning_rate": 7.70002169691973e-06, "loss": 0.5544, "step": 83 }, { "epoch": 0.15290102389078497, "grad_norm": 1.3333392166861238, "learning_rate": 7.690982345694746e-06, "loss": 0.5212, "step": 84 }, { "epoch": 0.1547212741751991, "grad_norm": 1.5189079377057624, "learning_rate": 7.68181428301864e-06, "loss": 0.5411, "step": 85 }, { "epoch": 0.1565415244596132, "grad_norm": 1.2341512401643826, "learning_rate": 7.67251782859889e-06, "loss": 0.5, "step": 86 }, { "epoch": 0.1583617747440273, "grad_norm": 1.3710782832898465, "learning_rate": 7.663093306620228e-06, "loss": 0.567, "step": 87 }, { "epoch": 0.16018202502844142, "grad_norm": 1.3224943188093254, "learning_rate": 7.653541045733351e-06, "loss": 0.6514, "step": 88 }, { "epoch": 0.16200227531285552, "grad_norm": 1.3825136099871158, "learning_rate": 7.643861379043442e-06, "loss": 0.49, "step": 89 }, { "epoch": 0.16382252559726962, "grad_norm": 1.6656667334345212, "learning_rate": 7.634054644098566e-06, "loss": 0.649, "step": 90 }, { "epoch": 0.16564277588168372, "grad_norm": 1.3683202937271444, "learning_rate": 7.624121182877892e-06, "loss": 0.497, "step": 91 }, { "epoch": 0.16746302616609784, "grad_norm": 1.5512864267072353, "learning_rate": 7.614061341779777e-06, "loss": 0.6176, "step": 92 }, { "epoch": 0.16928327645051194, "grad_norm": 1.5790193819370095, "learning_rate": 7.6038754716096755e-06, "loss": 0.5634, "step": 93 }, { "epoch": 0.17110352673492604, "grad_norm": 1.4344008312589909, "learning_rate": 7.593563927567915e-06, "loss": 0.5932, "step": 94 }, { "epoch": 0.17292377701934017, "grad_norm": 1.2510278162330568, "learning_rate": 7.583127069237302e-06, "loss": 0.5604, "step": 95 }, { "epoch": 0.17474402730375427, "grad_norm": 1.1926891094591303, "learning_rate": 7.5725652605705876e-06, "loss": 0.5746, "step": 96 }, { "epoch": 0.17656427758816837, "grad_norm": 1.3458065119541616, "learning_rate": 7.561878869877778e-06, "loss": 0.5, "step": 97 }, { "epoch": 0.1783845278725825, "grad_norm": 1.311426359460282, "learning_rate": 7.551068269813282e-06, "loss": 0.503, "step": 98 }, { "epoch": 0.1802047781569966, "grad_norm": 1.2792152183438508, "learning_rate": 7.540133837362924e-06, "loss": 0.5279, "step": 99 }, { "epoch": 0.1820250284414107, "grad_norm": 1.2349765362905594, "learning_rate": 7.5290759538307944e-06, "loss": 0.5159, "step": 100 }, { "epoch": 0.1838452787258248, "grad_norm": 1.3587018583909733, "learning_rate": 7.517895004825955e-06, "loss": 0.573, "step": 101 }, { "epoch": 0.18566552901023892, "grad_norm": 1.3554993825796526, "learning_rate": 7.506591380248991e-06, "loss": 0.5801, "step": 102 }, { "epoch": 0.18748577929465302, "grad_norm": 1.2364653944665345, "learning_rate": 7.495165474278411e-06, "loss": 0.5618, "step": 103 }, { "epoch": 0.18930602957906711, "grad_norm": 1.226975873209754, "learning_rate": 7.483617685356906e-06, "loss": 0.6663, "step": 104 }, { "epoch": 0.19112627986348124, "grad_norm": 1.4312274290636884, "learning_rate": 7.471948416177452e-06, "loss": 0.5473, "step": 105 }, { "epoch": 0.19294653014789534, "grad_norm": 1.467104665014613, "learning_rate": 7.460158073669271e-06, "loss": 0.5418, "step": 106 }, { "epoch": 0.19476678043230944, "grad_norm": 1.1804815586636788, "learning_rate": 7.448247068983638e-06, "loss": 0.5378, "step": 107 }, { "epoch": 0.19658703071672354, "grad_norm": 1.3602412614708939, "learning_rate": 7.43621581747954e-06, "loss": 0.5026, "step": 108 }, { "epoch": 0.19840728100113766, "grad_norm": 1.2691524680339796, "learning_rate": 7.4240647387092e-06, "loss": 0.591, "step": 109 }, { "epoch": 0.20022753128555176, "grad_norm": 1.2783869708675566, "learning_rate": 7.411794256403439e-06, "loss": 0.5085, "step": 110 }, { "epoch": 0.20204778156996586, "grad_norm": 1.2804361153159327, "learning_rate": 7.399404798456901e-06, "loss": 0.6244, "step": 111 }, { "epoch": 0.20386803185438, "grad_norm": 1.2444898246213776, "learning_rate": 7.3868967969131364e-06, "loss": 0.5313, "step": 112 }, { "epoch": 0.2056882821387941, "grad_norm": 1.3015010555018793, "learning_rate": 7.374270687949531e-06, "loss": 0.5512, "step": 113 }, { "epoch": 0.2075085324232082, "grad_norm": 1.2989535634921763, "learning_rate": 7.3615269118620945e-06, "loss": 0.5612, "step": 114 }, { "epoch": 0.20932878270762229, "grad_norm": 1.2893299441070913, "learning_rate": 7.348665913050114e-06, "loss": 0.4779, "step": 115 }, { "epoch": 0.2111490329920364, "grad_norm": 1.4374586177315487, "learning_rate": 7.3356881400006485e-06, "loss": 0.6057, "step": 116 }, { "epoch": 0.2129692832764505, "grad_norm": 1.2587242021503462, "learning_rate": 7.3225940452728915e-06, "loss": 0.5679, "step": 117 }, { "epoch": 0.2147895335608646, "grad_norm": 1.4250046519243573, "learning_rate": 7.309384085482396e-06, "loss": 0.5, "step": 118 }, { "epoch": 0.21660978384527874, "grad_norm": 1.222873795275555, "learning_rate": 7.29605872128514e-06, "loss": 0.6714, "step": 119 }, { "epoch": 0.21843003412969283, "grad_norm": 1.4243471522403268, "learning_rate": 7.282618417361476e-06, "loss": 0.6238, "step": 120 }, { "epoch": 0.22025028441410693, "grad_norm": 1.3065783409043068, "learning_rate": 7.269063642399912e-06, "loss": 0.5464, "step": 121 }, { "epoch": 0.22207053469852106, "grad_norm": 1.4517652323683794, "learning_rate": 7.25539486908078e-06, "loss": 0.4985, "step": 122 }, { "epoch": 0.22389078498293516, "grad_norm": 1.1483510986901082, "learning_rate": 7.241612574059745e-06, "loss": 0.4978, "step": 123 }, { "epoch": 0.22571103526734926, "grad_norm": 1.4389217565677268, "learning_rate": 7.227717237951189e-06, "loss": 0.6112, "step": 124 }, { "epoch": 0.22753128555176336, "grad_norm": 1.3389550682905482, "learning_rate": 7.213709345311444e-06, "loss": 0.6476, "step": 125 }, { "epoch": 0.22935153583617748, "grad_norm": 1.4169245520418259, "learning_rate": 7.1995893846219035e-06, "loss": 0.5354, "step": 126 }, { "epoch": 0.23117178612059158, "grad_norm": 1.2613504469980097, "learning_rate": 7.185357848271977e-06, "loss": 0.5467, "step": 127 }, { "epoch": 0.23299203640500568, "grad_norm": 1.1666125829814091, "learning_rate": 7.17101523254193e-06, "loss": 0.4698, "step": 128 }, { "epoch": 0.2348122866894198, "grad_norm": 1.352110005355786, "learning_rate": 7.156562037585575e-06, "loss": 0.6109, "step": 129 }, { "epoch": 0.2366325369738339, "grad_norm": 1.2180780408157383, "learning_rate": 7.1419987674128225e-06, "loss": 0.5332, "step": 130 }, { "epoch": 0.238452787258248, "grad_norm": 1.3933377570677665, "learning_rate": 7.127325929872119e-06, "loss": 0.6671, "step": 131 }, { "epoch": 0.2402730375426621, "grad_norm": 1.5246008222442193, "learning_rate": 7.1125440366327245e-06, "loss": 0.6212, "step": 132 }, { "epoch": 0.24209328782707623, "grad_norm": 1.2745899656103845, "learning_rate": 7.0976536031668775e-06, "loss": 0.6395, "step": 133 }, { "epoch": 0.24391353811149033, "grad_norm": 1.226190343569202, "learning_rate": 7.082655148731815e-06, "loss": 0.5761, "step": 134 }, { "epoch": 0.24573378839590443, "grad_norm": 1.1905155038972701, "learning_rate": 7.067549196351669e-06, "loss": 0.5418, "step": 135 }, { "epoch": 0.24755403868031856, "grad_norm": 1.2116012422540454, "learning_rate": 7.052336272799226e-06, "loss": 0.5273, "step": 136 }, { "epoch": 0.24937428896473265, "grad_norm": 1.282002116739219, "learning_rate": 7.037016908577555e-06, "loss": 0.4506, "step": 137 }, { "epoch": 0.25119453924914675, "grad_norm": 1.2499521919340497, "learning_rate": 7.02159163790151e-06, "loss": 0.5606, "step": 138 }, { "epoch": 0.2530147895335609, "grad_norm": 1.7373606199315674, "learning_rate": 7.006060998679105e-06, "loss": 0.559, "step": 139 }, { "epoch": 0.25483503981797495, "grad_norm": 1.4650449441633262, "learning_rate": 6.990425532492747e-06, "loss": 0.5135, "step": 140 }, { "epoch": 0.2566552901023891, "grad_norm": 1.154885986165648, "learning_rate": 6.974685784580359e-06, "loss": 0.5039, "step": 141 }, { "epoch": 0.2584755403868032, "grad_norm": 1.2096650723076037, "learning_rate": 6.958842303816359e-06, "loss": 0.5079, "step": 142 }, { "epoch": 0.2602957906712173, "grad_norm": 1.1620810226211598, "learning_rate": 6.942895642692527e-06, "loss": 0.5245, "step": 143 }, { "epoch": 0.2621160409556314, "grad_norm": 1.3200231462656904, "learning_rate": 6.926846357298732e-06, "loss": 0.5935, "step": 144 }, { "epoch": 0.26393629124004553, "grad_norm": 1.2355460824632627, "learning_rate": 6.910695007303544e-06, "loss": 0.5543, "step": 145 }, { "epoch": 0.2657565415244596, "grad_norm": 1.2033075815432748, "learning_rate": 6.894442155934719e-06, "loss": 0.4831, "step": 146 }, { "epoch": 0.2675767918088737, "grad_norm": 1.1907917135137838, "learning_rate": 6.878088369959553e-06, "loss": 0.5221, "step": 147 }, { "epoch": 0.26939704209328785, "grad_norm": 1.3255509879738674, "learning_rate": 6.861634219665117e-06, "loss": 0.6086, "step": 148 }, { "epoch": 0.2712172923777019, "grad_norm": 1.2271219658785495, "learning_rate": 6.845080278838381e-06, "loss": 0.5825, "step": 149 }, { "epoch": 0.27303754266211605, "grad_norm": 1.1690589610793065, "learning_rate": 6.82842712474619e-06, "loss": 0.5807, "step": 150 }, { "epoch": 0.2748577929465301, "grad_norm": 1.2888207762623227, "learning_rate": 6.811675338115146e-06, "loss": 0.6188, "step": 151 }, { "epoch": 0.27667804323094425, "grad_norm": 1.2514369430789496, "learning_rate": 6.7948255031113505e-06, "loss": 0.5913, "step": 152 }, { "epoch": 0.2784982935153584, "grad_norm": 1.3076941421175066, "learning_rate": 6.777878207320034e-06, "loss": 0.5054, "step": 153 }, { "epoch": 0.28031854379977245, "grad_norm": 1.2943985087075844, "learning_rate": 6.760834041725068e-06, "loss": 0.4915, "step": 154 }, { "epoch": 0.2821387940841866, "grad_norm": 1.2887581327542428, "learning_rate": 6.743693600688352e-06, "loss": 0.5538, "step": 155 }, { "epoch": 0.2839590443686007, "grad_norm": 1.157113796843012, "learning_rate": 6.726457481929095e-06, "loss": 0.537, "step": 156 }, { "epoch": 0.28577929465301477, "grad_norm": 1.1766314672266696, "learning_rate": 6.7091262865029645e-06, "loss": 0.5896, "step": 157 }, { "epoch": 0.2875995449374289, "grad_norm": 1.2292202798354899, "learning_rate": 6.691700618781126e-06, "loss": 0.6347, "step": 158 }, { "epoch": 0.289419795221843, "grad_norm": 1.1513406588801496, "learning_rate": 6.674181086429177e-06, "loss": 0.4663, "step": 159 }, { "epoch": 0.2912400455062571, "grad_norm": 1.2946840584006447, "learning_rate": 6.656568300385944e-06, "loss": 0.6247, "step": 160 }, { "epoch": 0.2930602957906712, "grad_norm": 1.2952632442735728, "learning_rate": 6.6388628748421895e-06, "loss": 0.4728, "step": 161 }, { "epoch": 0.29488054607508535, "grad_norm": 1.2684543190366842, "learning_rate": 6.62106542721918e-06, "loss": 0.5, "step": 162 }, { "epoch": 0.2967007963594994, "grad_norm": 1.2434590334770437, "learning_rate": 6.603176578147174e-06, "loss": 0.552, "step": 163 }, { "epoch": 0.29852104664391355, "grad_norm": 1.2853162665121605, "learning_rate": 6.585196951443763e-06, "loss": 0.5311, "step": 164 }, { "epoch": 0.3003412969283277, "grad_norm": 1.3910181828529422, "learning_rate": 6.5671271740921266e-06, "loss": 0.5595, "step": 165 }, { "epoch": 0.30216154721274174, "grad_norm": 1.318698152941268, "learning_rate": 6.548967876219163e-06, "loss": 0.5323, "step": 166 }, { "epoch": 0.30398179749715587, "grad_norm": 1.259567167046916, "learning_rate": 6.530719691073521e-06, "loss": 0.5773, "step": 167 }, { "epoch": 0.30580204778156994, "grad_norm": 1.3201679730014977, "learning_rate": 6.5123832550035165e-06, "loss": 0.5143, "step": 168 }, { "epoch": 0.30762229806598407, "grad_norm": 1.3232034824966301, "learning_rate": 6.493959207434934e-06, "loss": 0.553, "step": 169 }, { "epoch": 0.3094425483503982, "grad_norm": 1.3960220649200046, "learning_rate": 6.47544819084874e-06, "loss": 0.561, "step": 170 }, { "epoch": 0.31126279863481227, "grad_norm": 1.3348060079340793, "learning_rate": 6.4568508507586715e-06, "loss": 0.5047, "step": 171 }, { "epoch": 0.3130830489192264, "grad_norm": 1.2868651237482562, "learning_rate": 6.438167835688725e-06, "loss": 0.5094, "step": 172 }, { "epoch": 0.3149032992036405, "grad_norm": 1.2603952904899627, "learning_rate": 6.41939979715055e-06, "loss": 0.5323, "step": 173 }, { "epoch": 0.3167235494880546, "grad_norm": 1.2921556438401538, "learning_rate": 6.400547389620716e-06, "loss": 0.5554, "step": 174 }, { "epoch": 0.3185437997724687, "grad_norm": 1.2590564886848532, "learning_rate": 6.3816112705178984e-06, "loss": 0.5288, "step": 175 }, { "epoch": 0.32036405005688284, "grad_norm": 1.3084436554782835, "learning_rate": 6.362592100179958e-06, "loss": 0.6402, "step": 176 }, { "epoch": 0.3221843003412969, "grad_norm": 1.4261334165831296, "learning_rate": 6.343490541840899e-06, "loss": 0.489, "step": 177 }, { "epoch": 0.32400455062571104, "grad_norm": 1.4986503257367303, "learning_rate": 6.3243072616077535e-06, "loss": 0.5957, "step": 178 }, { "epoch": 0.32582480091012517, "grad_norm": 1.2030242787629297, "learning_rate": 6.3050429284373465e-06, "loss": 0.4974, "step": 179 }, { "epoch": 0.32764505119453924, "grad_norm": 1.3717989144113625, "learning_rate": 6.285698214112974e-06, "loss": 0.593, "step": 180 }, { "epoch": 0.32946530147895337, "grad_norm": 1.2021644407962897, "learning_rate": 6.2662737932209695e-06, "loss": 0.616, "step": 181 }, { "epoch": 0.33128555176336744, "grad_norm": 1.281410126238882, "learning_rate": 6.246770343127185e-06, "loss": 0.5598, "step": 182 }, { "epoch": 0.33310580204778156, "grad_norm": 1.39438795848328, "learning_rate": 6.227188543953368e-06, "loss": 0.5932, "step": 183 }, { "epoch": 0.33310580204778156, "eval_accuracy": 0.8092656088844726, "eval_accuracy_first_token": 0.7541679610645128, "eval_accuracy_first_token_<": 0.9296587926509187, "eval_accuracy_first_token_<_total": 1905, "eval_accuracy_first_token_<|python_tag|>": 0.8752515090543259, "eval_accuracy_first_token_<|python_tag|>_total": 994, "eval_accuracy_first_token_Certainly": 0.7024793388429752, "eval_accuracy_first_token_Certainly_total": 363, "eval_accuracy_first_token_The": 0.9059161873459326, "eval_accuracy_first_token_The_total": 2434, "eval_accuracy_first_token_To": 0.8237179487179487, "eval_accuracy_first_token_To_total": 936, "eval_loss": 0.5801064372062683, "eval_perplexity": 1.1141803737974993, "eval_runtime": 508.2626, "eval_samples_per_second": 1.371, "eval_steps_per_second": 0.687, "eval_total_number_first_token": 9657, "step": 183 }, { "epoch": 0.3349260523321957, "grad_norm": 1.3081524231573554, "learning_rate": 6.207529078553444e-06, "loss": 0.5457, "step": 184 }, { "epoch": 0.33674630261660976, "grad_norm": 1.3716078335539046, "learning_rate": 6.1877926324897085e-06, "loss": 0.5473, "step": 185 }, { "epoch": 0.3385665529010239, "grad_norm": 1.4145939624062198, "learning_rate": 6.16797989400891e-06, "loss": 0.5786, "step": 186 }, { "epoch": 0.340386803185438, "grad_norm": 1.2294721126569037, "learning_rate": 6.148091554018264e-06, "loss": 0.5902, "step": 187 }, { "epoch": 0.3422070534698521, "grad_norm": 1.2925378088030424, "learning_rate": 6.128128306061346e-06, "loss": 0.5142, "step": 188 }, { "epoch": 0.3440273037542662, "grad_norm": 1.2279588518524418, "learning_rate": 6.108090846293915e-06, "loss": 0.5135, "step": 189 }, { "epoch": 0.34584755403868034, "grad_norm": 1.268146835786646, "learning_rate": 6.087979873459634e-06, "loss": 0.5447, "step": 190 }, { "epoch": 0.3476678043230944, "grad_norm": 1.318995573559777, "learning_rate": 6.0677960888657015e-06, "loss": 0.6744, "step": 191 }, { "epoch": 0.34948805460750854, "grad_norm": 1.231758642404661, "learning_rate": 6.047540196358404e-06, "loss": 0.5809, "step": 192 }, { "epoch": 0.35130830489192266, "grad_norm": 1.2372891673165372, "learning_rate": 6.02721290229856e-06, "loss": 0.5807, "step": 193 }, { "epoch": 0.35312855517633673, "grad_norm": 1.481210652387573, "learning_rate": 6.006814915536894e-06, "loss": 0.5936, "step": 194 }, { "epoch": 0.35494880546075086, "grad_norm": 1.2305803524181071, "learning_rate": 5.9863469473893225e-06, "loss": 0.5438, "step": 195 }, { "epoch": 0.356769055745165, "grad_norm": 1.2206928256434937, "learning_rate": 5.965809711612137e-06, "loss": 0.5005, "step": 196 }, { "epoch": 0.35858930602957906, "grad_norm": 1.086573502847394, "learning_rate": 5.945203924377125e-06, "loss": 0.4889, "step": 197 }, { "epoch": 0.3604095563139932, "grad_norm": 1.3513470624112347, "learning_rate": 5.92453030424659e-06, "loss": 0.5599, "step": 198 }, { "epoch": 0.36222980659840726, "grad_norm": 1.1113956980921844, "learning_rate": 5.903789572148294e-06, "loss": 0.5182, "step": 199 }, { "epoch": 0.3640500568828214, "grad_norm": 1.4891052629080104, "learning_rate": 5.88298245135032e-06, "loss": 0.5716, "step": 200 }, { "epoch": 0.3658703071672355, "grad_norm": 1.3005774833983796, "learning_rate": 5.862109667435853e-06, "loss": 0.5665, "step": 201 }, { "epoch": 0.3676905574516496, "grad_norm": 1.2151067893045482, "learning_rate": 5.8411719482778645e-06, "loss": 0.4965, "step": 202 }, { "epoch": 0.3695108077360637, "grad_norm": 1.5031392413729012, "learning_rate": 5.820170024013746e-06, "loss": 0.5398, "step": 203 }, { "epoch": 0.37133105802047783, "grad_norm": 1.1627104663425107, "learning_rate": 5.79910462701984e-06, "loss": 0.4461, "step": 204 }, { "epoch": 0.3731513083048919, "grad_norm": 1.3656640622390992, "learning_rate": 5.777976491885903e-06, "loss": 0.6048, "step": 205 }, { "epoch": 0.37497155858930603, "grad_norm": 1.2327820864728312, "learning_rate": 5.756786355389481e-06, "loss": 0.5052, "step": 206 }, { "epoch": 0.37679180887372016, "grad_norm": 1.3098660955959893, "learning_rate": 5.735534956470232e-06, "loss": 0.5507, "step": 207 }, { "epoch": 0.37861205915813423, "grad_norm": 1.2828934352712993, "learning_rate": 5.714223036204144e-06, "loss": 0.5973, "step": 208 }, { "epoch": 0.38043230944254836, "grad_norm": 1.1860097743128348, "learning_rate": 5.6928513377777e-06, "loss": 0.4965, "step": 209 }, { "epoch": 0.3822525597269625, "grad_norm": 1.2517621268060033, "learning_rate": 5.671420606461956e-06, "loss": 0.5487, "step": 210 }, { "epoch": 0.38407281001137655, "grad_norm": 1.271873202647325, "learning_rate": 5.649931589586557e-06, "loss": 0.5979, "step": 211 }, { "epoch": 0.3858930602957907, "grad_norm": 1.2859574150365818, "learning_rate": 5.628385036513676e-06, "loss": 0.4776, "step": 212 }, { "epoch": 0.38771331058020475, "grad_norm": 1.354537404525919, "learning_rate": 5.606781698611878e-06, "loss": 0.4877, "step": 213 }, { "epoch": 0.3895335608646189, "grad_norm": 1.3284236870109494, "learning_rate": 5.585122329229923e-06, "loss": 0.5859, "step": 214 }, { "epoch": 0.391353811149033, "grad_norm": 1.2340130409038237, "learning_rate": 5.56340768367049e-06, "loss": 0.5305, "step": 215 }, { "epoch": 0.3931740614334471, "grad_norm": 1.453487991090255, "learning_rate": 5.541638519163849e-06, "loss": 0.55, "step": 216 }, { "epoch": 0.3949943117178612, "grad_norm": 1.298062739902415, "learning_rate": 5.51981559484144e-06, "loss": 0.5169, "step": 217 }, { "epoch": 0.39681456200227533, "grad_norm": 1.311968473611326, "learning_rate": 5.49793967170941e-06, "loss": 0.5751, "step": 218 }, { "epoch": 0.3986348122866894, "grad_norm": 1.2412899795871963, "learning_rate": 5.476011512622076e-06, "loss": 0.6166, "step": 219 }, { "epoch": 0.4004550625711035, "grad_norm": 1.242116646999028, "learning_rate": 5.454031882255319e-06, "loss": 0.5578, "step": 220 }, { "epoch": 0.40227531285551765, "grad_norm": 1.4168825327120473, "learning_rate": 5.43200154707992e-06, "loss": 0.5662, "step": 221 }, { "epoch": 0.4040955631399317, "grad_norm": 1.426587220049501, "learning_rate": 5.4099212753348294e-06, "loss": 0.5169, "step": 222 }, { "epoch": 0.40591581342434585, "grad_norm": 1.3014887756398712, "learning_rate": 5.3877918370003806e-06, "loss": 0.5117, "step": 223 }, { "epoch": 0.40773606370876, "grad_norm": 1.3406748606110184, "learning_rate": 5.365614003771439e-06, "loss": 0.5549, "step": 224 }, { "epoch": 0.40955631399317405, "grad_norm": 1.2661574936204552, "learning_rate": 5.343388549030491e-06, "loss": 0.5163, "step": 225 }, { "epoch": 0.4113765642775882, "grad_norm": 1.085801335463159, "learning_rate": 5.321116247820669e-06, "loss": 0.5244, "step": 226 }, { "epoch": 0.4131968145620023, "grad_norm": 1.1238570077454868, "learning_rate": 5.298797876818734e-06, "loss": 0.4877, "step": 227 }, { "epoch": 0.4150170648464164, "grad_norm": 1.2501588033198834, "learning_rate": 5.276434214307986e-06, "loss": 0.5175, "step": 228 }, { "epoch": 0.4168373151308305, "grad_norm": 1.1104132675236253, "learning_rate": 5.2540260401511255e-06, "loss": 0.4912, "step": 229 }, { "epoch": 0.41865756541524457, "grad_norm": 1.3469476360006967, "learning_rate": 5.231574135763052e-06, "loss": 0.5119, "step": 230 }, { "epoch": 0.4204778156996587, "grad_norm": 1.3917881004299013, "learning_rate": 5.209079284083626e-06, "loss": 0.5893, "step": 231 }, { "epoch": 0.4222980659840728, "grad_norm": 1.3049155919134754, "learning_rate": 5.186542269550359e-06, "loss": 0.5863, "step": 232 }, { "epoch": 0.4241183162684869, "grad_norm": 1.4688437185729748, "learning_rate": 5.163963878071058e-06, "loss": 0.6134, "step": 233 }, { "epoch": 0.425938566552901, "grad_norm": 1.334594097803803, "learning_rate": 5.141344896996421e-06, "loss": 0.4778, "step": 234 }, { "epoch": 0.42775881683731515, "grad_norm": 1.349733087487026, "learning_rate": 5.1186861150925844e-06, "loss": 0.5989, "step": 235 }, { "epoch": 0.4295790671217292, "grad_norm": 1.1945891754612503, "learning_rate": 5.09598832251361e-06, "loss": 0.4466, "step": 236 }, { "epoch": 0.43139931740614335, "grad_norm": 1.388728814777883, "learning_rate": 5.073252310773939e-06, "loss": 0.6193, "step": 237 }, { "epoch": 0.43321956769055747, "grad_norm": 1.4284168159961905, "learning_rate": 5.050478872720782e-06, "loss": 0.5535, "step": 238 }, { "epoch": 0.43503981797497154, "grad_norm": 1.3786811940208537, "learning_rate": 5.027668802506476e-06, "loss": 0.4974, "step": 239 }, { "epoch": 0.43686006825938567, "grad_norm": 1.2927644778322318, "learning_rate": 5.004822895560794e-06, "loss": 0.5029, "step": 240 }, { "epoch": 0.4386803185437998, "grad_norm": 1.252387489439096, "learning_rate": 4.981941948563196e-06, "loss": 0.5278, "step": 241 }, { "epoch": 0.44050056882821387, "grad_norm": 1.339410297479304, "learning_rate": 4.959026759415061e-06, "loss": 0.4939, "step": 242 }, { "epoch": 0.442320819112628, "grad_norm": 1.364314280772671, "learning_rate": 4.936078127211849e-06, "loss": 0.5951, "step": 243 }, { "epoch": 0.4441410693970421, "grad_norm": 1.3075257340817037, "learning_rate": 4.913096852215248e-06, "loss": 0.6049, "step": 244 }, { "epoch": 0.4459613196814562, "grad_norm": 1.4449459168578944, "learning_rate": 4.890083735825257e-06, "loss": 0.6495, "step": 245 }, { "epoch": 0.4477815699658703, "grad_norm": 1.250520396487667, "learning_rate": 4.867039580552247e-06, "loss": 0.5094, "step": 246 }, { "epoch": 0.4496018202502844, "grad_norm": 1.3729253299629682, "learning_rate": 4.843965189988969e-06, "loss": 0.601, "step": 247 }, { "epoch": 0.4514220705346985, "grad_norm": 1.3588267674574899, "learning_rate": 4.820861368782537e-06, "loss": 0.6282, "step": 248 }, { "epoch": 0.45324232081911264, "grad_norm": 1.2605268123722788, "learning_rate": 4.79772892260637e-06, "loss": 0.5305, "step": 249 }, { "epoch": 0.4550625711035267, "grad_norm": 1.2681522997617236, "learning_rate": 4.774568658132086e-06, "loss": 0.5748, "step": 250 }, { "epoch": 0.45688282138794084, "grad_norm": 1.3533665486552986, "learning_rate": 4.751381383001386e-06, "loss": 0.4689, "step": 251 }, { "epoch": 0.45870307167235497, "grad_norm": 1.267434752321521, "learning_rate": 4.728167905797877e-06, "loss": 0.534, "step": 252 }, { "epoch": 0.46052332195676904, "grad_norm": 1.2225430366963792, "learning_rate": 4.7049290360188875e-06, "loss": 0.5003, "step": 253 }, { "epoch": 0.46234357224118316, "grad_norm": 1.3388181688742944, "learning_rate": 4.681665584047227e-06, "loss": 0.5219, "step": 254 }, { "epoch": 0.4641638225255973, "grad_norm": 1.3028544028541067, "learning_rate": 4.658378361122936e-06, "loss": 0.5452, "step": 255 }, { "epoch": 0.46598407281001136, "grad_norm": 1.157965088527389, "learning_rate": 4.6350681793149884e-06, "loss": 0.5229, "step": 256 }, { "epoch": 0.4678043230944255, "grad_norm": 1.3045701677810966, "learning_rate": 4.611735851492984e-06, "loss": 0.5728, "step": 257 }, { "epoch": 0.4696245733788396, "grad_norm": 1.2850295730378811, "learning_rate": 4.588382191298787e-06, "loss": 0.5537, "step": 258 }, { "epoch": 0.4714448236632537, "grad_norm": 1.1541773263319788, "learning_rate": 4.5650080131181675e-06, "loss": 0.538, "step": 259 }, { "epoch": 0.4732650739476678, "grad_norm": 1.2910982881953428, "learning_rate": 4.541614132052393e-06, "loss": 0.5612, "step": 260 }, { "epoch": 0.4750853242320819, "grad_norm": 1.3654547247389965, "learning_rate": 4.51820136388981e-06, "loss": 0.4475, "step": 261 }, { "epoch": 0.476905574516496, "grad_norm": 1.2698111645568326, "learning_rate": 4.494770525077391e-06, "loss": 0.5621, "step": 262 }, { "epoch": 0.47872582480091014, "grad_norm": 1.3935553561279155, "learning_rate": 4.4713224326922655e-06, "loss": 0.599, "step": 263 }, { "epoch": 0.4805460750853242, "grad_norm": 1.3156202456183232, "learning_rate": 4.447857904413231e-06, "loss": 0.532, "step": 264 }, { "epoch": 0.48236632536973834, "grad_norm": 1.1337315757674442, "learning_rate": 4.424377758492232e-06, "loss": 0.5353, "step": 265 }, { "epoch": 0.48418657565415246, "grad_norm": 1.4101220722381445, "learning_rate": 4.40088281372583e-06, "loss": 0.5006, "step": 266 }, { "epoch": 0.48600682593856653, "grad_norm": 1.2300783759396936, "learning_rate": 4.377373889426649e-06, "loss": 0.5438, "step": 267 }, { "epoch": 0.48782707622298066, "grad_norm": 1.17585511931677, "learning_rate": 4.353851805394808e-06, "loss": 0.5369, "step": 268 }, { "epoch": 0.4896473265073948, "grad_norm": 1.2115284188121456, "learning_rate": 4.33031738188933e-06, "loss": 0.4524, "step": 269 }, { "epoch": 0.49146757679180886, "grad_norm": 1.4317839486717954, "learning_rate": 4.306771439599534e-06, "loss": 0.6436, "step": 270 }, { "epoch": 0.493287827076223, "grad_norm": 1.4621543028546877, "learning_rate": 4.283214799616428e-06, "loss": 0.4368, "step": 271 }, { "epoch": 0.4951080773606371, "grad_norm": 1.2936498719418335, "learning_rate": 4.259648283404062e-06, "loss": 0.5541, "step": 272 }, { "epoch": 0.4969283276450512, "grad_norm": 1.120893056901481, "learning_rate": 4.236072712770891e-06, "loss": 0.5822, "step": 273 }, { "epoch": 0.4987485779294653, "grad_norm": 1.429920580926913, "learning_rate": 4.2124889098411175e-06, "loss": 0.5302, "step": 274 }, { "epoch": 0.5005688282138794, "grad_norm": 1.312546385183068, "learning_rate": 4.1888976970260135e-06, "loss": 0.4835, "step": 275 }, { "epoch": 0.5023890784982935, "grad_norm": 1.3107231912852029, "learning_rate": 4.165299896995252e-06, "loss": 0.5421, "step": 276 }, { "epoch": 0.5042093287827076, "grad_norm": 1.232720241327702, "learning_rate": 4.141696332648216e-06, "loss": 0.5012, "step": 277 }, { "epoch": 0.5060295790671218, "grad_norm": 1.2132932647482422, "learning_rate": 4.118087827085294e-06, "loss": 0.5463, "step": 278 }, { "epoch": 0.5078498293515359, "grad_norm": 1.270716742837696, "learning_rate": 4.094475203579191e-06, "loss": 0.5383, "step": 279 }, { "epoch": 0.5096700796359499, "grad_norm": 1.2438093689244545, "learning_rate": 4.070859285546209e-06, "loss": 0.5556, "step": 280 }, { "epoch": 0.511490329920364, "grad_norm": 1.3516997697288733, "learning_rate": 4.047240896517539e-06, "loss": 0.6018, "step": 281 }, { "epoch": 0.5133105802047782, "grad_norm": 1.2439174788361766, "learning_rate": 4.023620860110533e-06, "loss": 0.4133, "step": 282 }, { "epoch": 0.5151308304891923, "grad_norm": 1.3778918541026397, "learning_rate": 4e-06, "loss": 0.6007, "step": 283 }, { "epoch": 0.5169510807736064, "grad_norm": 1.2268835899477202, "learning_rate": 3.976379139889467e-06, "loss": 0.5331, "step": 284 }, { "epoch": 0.5187713310580204, "grad_norm": 1.3429317349537357, "learning_rate": 3.9527591034824616e-06, "loss": 0.5311, "step": 285 }, { "epoch": 0.5205915813424346, "grad_norm": 1.2134030361634403, "learning_rate": 3.929140714453791e-06, "loss": 0.481, "step": 286 }, { "epoch": 0.5224118316268487, "grad_norm": 1.2462618006711519, "learning_rate": 3.9055247964208075e-06, "loss": 0.5273, "step": 287 }, { "epoch": 0.5242320819112628, "grad_norm": 1.5559356945744065, "learning_rate": 3.8819121729147055e-06, "loss": 0.6021, "step": 288 }, { "epoch": 0.5260523321956769, "grad_norm": 1.3832902722589653, "learning_rate": 3.8583036673517845e-06, "loss": 0.4454, "step": 289 }, { "epoch": 0.5278725824800911, "grad_norm": 1.3978629990846738, "learning_rate": 3.834700103004747e-06, "loss": 0.5124, "step": 290 }, { "epoch": 0.5296928327645051, "grad_norm": 1.2666994035299775, "learning_rate": 3.8111023029739866e-06, "loss": 0.4667, "step": 291 }, { "epoch": 0.5315130830489192, "grad_norm": 1.3305625183535323, "learning_rate": 3.787511090158884e-06, "loss": 0.5368, "step": 292 }, { "epoch": 0.5333333333333333, "grad_norm": 1.2554353067602742, "learning_rate": 3.763927287229109e-06, "loss": 0.5499, "step": 293 }, { "epoch": 0.5351535836177475, "grad_norm": 1.307288214215709, "learning_rate": 3.740351716595939e-06, "loss": 0.5055, "step": 294 }, { "epoch": 0.5369738339021616, "grad_norm": 1.3279514539943822, "learning_rate": 3.7167852003835723e-06, "loss": 0.511, "step": 295 }, { "epoch": 0.5387940841865757, "grad_norm": 1.223828271060312, "learning_rate": 3.6932285604004656e-06, "loss": 0.4595, "step": 296 }, { "epoch": 0.5406143344709897, "grad_norm": 1.332815030257366, "learning_rate": 3.669682618110671e-06, "loss": 0.6227, "step": 297 }, { "epoch": 0.5424345847554038, "grad_norm": 1.2271598713957907, "learning_rate": 3.646148194605191e-06, "loss": 0.5925, "step": 298 }, { "epoch": 0.544254835039818, "grad_norm": 1.2856112828358344, "learning_rate": 3.622626110573351e-06, "loss": 0.4888, "step": 299 }, { "epoch": 0.5460750853242321, "grad_norm": 1.2933732045646906, "learning_rate": 3.5991171862741713e-06, "loss": 0.5072, "step": 300 }, { "epoch": 0.5478953356086462, "grad_norm": 1.928545669129003, "learning_rate": 3.575622241507768e-06, "loss": 0.525, "step": 301 }, { "epoch": 0.5497155858930602, "grad_norm": 1.1113802518291283, "learning_rate": 3.5521420955867683e-06, "loss": 0.5977, "step": 302 }, { "epoch": 0.5515358361774744, "grad_norm": 1.3617289948905469, "learning_rate": 3.5286775673077332e-06, "loss": 0.5839, "step": 303 }, { "epoch": 0.5533560864618885, "grad_norm": 1.391130826033813, "learning_rate": 3.505229474922609e-06, "loss": 0.5181, "step": 304 }, { "epoch": 0.5551763367463026, "grad_norm": 1.2316652646361441, "learning_rate": 3.481798636110191e-06, "loss": 0.4945, "step": 305 }, { "epoch": 0.5569965870307167, "grad_norm": 1.2873086430494702, "learning_rate": 3.458385867947607e-06, "loss": 0.4924, "step": 306 }, { "epoch": 0.5588168373151309, "grad_norm": 1.430369532857129, "learning_rate": 3.434991986881833e-06, "loss": 0.4821, "step": 307 }, { "epoch": 0.5606370875995449, "grad_norm": 1.1646260708946579, "learning_rate": 3.4116178087012136e-06, "loss": 0.5052, "step": 308 }, { "epoch": 0.562457337883959, "grad_norm": 1.2327288356772756, "learning_rate": 3.388264148507016e-06, "loss": 0.5057, "step": 309 }, { "epoch": 0.5642775881683731, "grad_norm": 1.3536008520463172, "learning_rate": 3.3649318206850116e-06, "loss": 0.5178, "step": 310 }, { "epoch": 0.5660978384527873, "grad_norm": 1.1896041830424324, "learning_rate": 3.3416216388770635e-06, "loss": 0.5417, "step": 311 }, { "epoch": 0.5679180887372014, "grad_norm": 1.2975919761075365, "learning_rate": 3.3183344159527736e-06, "loss": 0.5234, "step": 312 }, { "epoch": 0.5697383390216155, "grad_norm": 1.1688323545338841, "learning_rate": 3.2950709639811134e-06, "loss": 0.4888, "step": 313 }, { "epoch": 0.5715585893060295, "grad_norm": 1.3939258642019638, "learning_rate": 3.271832094202123e-06, "loss": 0.5183, "step": 314 }, { "epoch": 0.5733788395904437, "grad_norm": 1.1897624823605304, "learning_rate": 3.2486186169986153e-06, "loss": 0.5454, "step": 315 }, { "epoch": 0.5751990898748578, "grad_norm": 1.1944714828344472, "learning_rate": 3.2254313418679154e-06, "loss": 0.4807, "step": 316 }, { "epoch": 0.5770193401592719, "grad_norm": 1.2256094296723554, "learning_rate": 3.2022710773936304e-06, "loss": 0.5223, "step": 317 }, { "epoch": 0.578839590443686, "grad_norm": 1.2296819521179183, "learning_rate": 3.1791386312174633e-06, "loss": 0.4951, "step": 318 }, { "epoch": 0.5806598407281001, "grad_norm": 1.4525163093513003, "learning_rate": 3.1560348100110315e-06, "loss": 0.4874, "step": 319 }, { "epoch": 0.5824800910125142, "grad_norm": 1.4766150215295402, "learning_rate": 3.1329604194477535e-06, "loss": 0.5186, "step": 320 }, { "epoch": 0.5843003412969283, "grad_norm": 1.5620410502792905, "learning_rate": 3.1099162641747427e-06, "loss": 0.5542, "step": 321 }, { "epoch": 0.5861205915813424, "grad_norm": 1.1792003162185065, "learning_rate": 3.0869031477847507e-06, "loss": 0.4751, "step": 322 }, { "epoch": 0.5879408418657566, "grad_norm": 1.4371388899245734, "learning_rate": 3.0639218727881508e-06, "loss": 0.5066, "step": 323 }, { "epoch": 0.5897610921501707, "grad_norm": 1.3321828422036859, "learning_rate": 3.04097324058494e-06, "loss": 0.418, "step": 324 }, { "epoch": 0.5915813424345847, "grad_norm": 1.2642329608748821, "learning_rate": 3.0180580514368034e-06, "loss": 0.6167, "step": 325 }, { "epoch": 0.5934015927189988, "grad_norm": 1.1538255682096556, "learning_rate": 2.9951771044392066e-06, "loss": 0.5297, "step": 326 }, { "epoch": 0.595221843003413, "grad_norm": 1.2453988044078719, "learning_rate": 2.972331197493523e-06, "loss": 0.4552, "step": 327 }, { "epoch": 0.5970420932878271, "grad_norm": 1.3576907607149231, "learning_rate": 2.949521127279218e-06, "loss": 0.5003, "step": 328 }, { "epoch": 0.5988623435722412, "grad_norm": 1.3497348777364608, "learning_rate": 2.926747689226062e-06, "loss": 0.5561, "step": 329 }, { "epoch": 0.6006825938566553, "grad_norm": 1.259164100959422, "learning_rate": 2.9040116774863896e-06, "loss": 0.4856, "step": 330 }, { "epoch": 0.6025028441410694, "grad_norm": 1.2529485220686642, "learning_rate": 2.881313884907416e-06, "loss": 0.5575, "step": 331 }, { "epoch": 0.6043230944254835, "grad_norm": 1.0669208953569564, "learning_rate": 2.8586551030035797e-06, "loss": 0.4644, "step": 332 }, { "epoch": 0.6061433447098976, "grad_norm": 1.3366563458096783, "learning_rate": 2.836036121928942e-06, "loss": 0.453, "step": 333 }, { "epoch": 0.6079635949943117, "grad_norm": 1.1476061284968695, "learning_rate": 2.813457730449641e-06, "loss": 0.4207, "step": 334 }, { "epoch": 0.6097838452787259, "grad_norm": 1.3411031407126155, "learning_rate": 2.790920715916372e-06, "loss": 0.5404, "step": 335 }, { "epoch": 0.6116040955631399, "grad_norm": 1.1482077310699785, "learning_rate": 2.7684258642369484e-06, "loss": 0.5205, "step": 336 }, { "epoch": 0.613424345847554, "grad_norm": 1.42649538957105, "learning_rate": 2.7459739598488762e-06, "loss": 0.5013, "step": 337 }, { "epoch": 0.6152445961319681, "grad_norm": 1.2243282105614175, "learning_rate": 2.723565785692013e-06, "loss": 0.5464, "step": 338 }, { "epoch": 0.6170648464163823, "grad_norm": 1.2484628027395077, "learning_rate": 2.701202123181266e-06, "loss": 0.5519, "step": 339 }, { "epoch": 0.6188850967007964, "grad_norm": 1.3240902412697022, "learning_rate": 2.6788837521793328e-06, "loss": 0.5205, "step": 340 }, { "epoch": 0.6207053469852105, "grad_norm": 1.2873575493742448, "learning_rate": 2.6566114509695096e-06, "loss": 0.4761, "step": 341 }, { "epoch": 0.6225255972696245, "grad_norm": 1.2093203753299095, "learning_rate": 2.634385996228561e-06, "loss": 0.4753, "step": 342 }, { "epoch": 0.6243458475540387, "grad_norm": 1.300179853101682, "learning_rate": 2.6122081629996195e-06, "loss": 0.4934, "step": 343 }, { "epoch": 0.6261660978384528, "grad_norm": 1.2890047749069995, "learning_rate": 2.5900787246651715e-06, "loss": 0.4873, "step": 344 }, { "epoch": 0.6279863481228669, "grad_norm": 1.4341436462807016, "learning_rate": 2.567998452920081e-06, "loss": 0.5213, "step": 345 }, { "epoch": 0.629806598407281, "grad_norm": 1.299103431190263, "learning_rate": 2.5459681177446797e-06, "loss": 0.4783, "step": 346 }, { "epoch": 0.631626848691695, "grad_norm": 1.331377030608932, "learning_rate": 2.523988487377924e-06, "loss": 0.5045, "step": 347 }, { "epoch": 0.6334470989761092, "grad_norm": 1.3571947037370755, "learning_rate": 2.50206032829059e-06, "loss": 0.5005, "step": 348 }, { "epoch": 0.6352673492605233, "grad_norm": 1.5380704302051296, "learning_rate": 2.4801844051585604e-06, "loss": 0.5238, "step": 349 }, { "epoch": 0.6370875995449374, "grad_norm": 1.3042046637167102, "learning_rate": 2.4583614808361508e-06, "loss": 0.4785, "step": 350 }, { "epoch": 0.6389078498293516, "grad_norm": 1.1803375558623432, "learning_rate": 2.4365923163295083e-06, "loss": 0.5518, "step": 351 }, { "epoch": 0.6407281001137657, "grad_norm": 1.1305498748692666, "learning_rate": 2.4148776707700775e-06, "loss": 0.4627, "step": 352 }, { "epoch": 0.6425483503981797, "grad_norm": 1.3696351562191598, "learning_rate": 2.393218301388123e-06, "loss": 0.4714, "step": 353 }, { "epoch": 0.6443686006825938, "grad_norm": 1.1646766669675297, "learning_rate": 2.3716149634863244e-06, "loss": 0.461, "step": 354 }, { "epoch": 0.646188850967008, "grad_norm": 1.432326181786707, "learning_rate": 2.3500684104134433e-06, "loss": 0.4775, "step": 355 }, { "epoch": 0.6480091012514221, "grad_norm": 1.2811226649238618, "learning_rate": 2.328579393538046e-06, "loss": 0.4473, "step": 356 }, { "epoch": 0.6498293515358362, "grad_norm": 1.2847418645420832, "learning_rate": 2.3071486622223e-06, "loss": 0.473, "step": 357 }, { "epoch": 0.6516496018202503, "grad_norm": 1.1991535525500763, "learning_rate": 2.2857769637958554e-06, "loss": 0.4548, "step": 358 }, { "epoch": 0.6534698521046644, "grad_norm": 1.3510869929117142, "learning_rate": 2.2644650435297675e-06, "loss": 0.474, "step": 359 }, { "epoch": 0.6552901023890785, "grad_norm": 1.2247454838152558, "learning_rate": 2.243213644610519e-06, "loss": 0.4063, "step": 360 }, { "epoch": 0.6571103526734926, "grad_norm": 1.224682187747472, "learning_rate": 2.2220235081140985e-06, "loss": 0.5137, "step": 361 }, { "epoch": 0.6589306029579067, "grad_norm": 1.5257557938450914, "learning_rate": 2.2008953729801583e-06, "loss": 0.4591, "step": 362 }, { "epoch": 0.6607508532423209, "grad_norm": 1.221121328273825, "learning_rate": 2.1798299759862545e-06, "loss": 0.5614, "step": 363 }, { "epoch": 0.6625711035267349, "grad_norm": 1.2199344648755224, "learning_rate": 2.158828051722137e-06, "loss": 0.5104, "step": 364 }, { "epoch": 0.664391353811149, "grad_norm": 1.3052741269665118, "learning_rate": 2.137890332564147e-06, "loss": 0.4732, "step": 365 }, { "epoch": 0.6662116040955631, "grad_norm": 1.1647023893060888, "learning_rate": 2.117017548649678e-06, "loss": 0.5229, "step": 366 }, { "epoch": 0.6662116040955631, "eval_accuracy": 0.8129295561130228, "eval_accuracy_first_token": 0.7684581132856995, "eval_accuracy_first_token_<": 0.9595800524934384, "eval_accuracy_first_token_<_total": 1905, "eval_accuracy_first_token_<|python_tag|>": 0.9094567404426559, "eval_accuracy_first_token_<|python_tag|>_total": 994, "eval_accuracy_first_token_Certainly": 0.7741046831955923, "eval_accuracy_first_token_Certainly_total": 363, "eval_accuracy_first_token_The": 0.8948233360723089, "eval_accuracy_first_token_The_total": 2434, "eval_accuracy_first_token_To": 0.8044871794871795, "eval_accuracy_first_token_To_total": 936, "eval_loss": 0.5655013918876648, "eval_perplexity": 1.1114110979501997, "eval_runtime": 507.2948, "eval_samples_per_second": 1.374, "eval_steps_per_second": 0.688, "eval_total_number_first_token": 9657, "step": 366 }, { "epoch": 0.6680318543799773, "grad_norm": 1.2605323109478153, "learning_rate": 2.0962104278517058e-06, "loss": 0.4634, "step": 367 }, { "epoch": 0.6698521046643914, "grad_norm": 1.153711484102447, "learning_rate": 2.0754696957534105e-06, "loss": 0.4578, "step": 368 }, { "epoch": 0.6716723549488055, "grad_norm": 1.4112272127644152, "learning_rate": 2.0547960756228746e-06, "loss": 0.5903, "step": 369 }, { "epoch": 0.6734926052332195, "grad_norm": 1.3058143917601592, "learning_rate": 2.0341902883878626e-06, "loss": 0.4261, "step": 370 }, { "epoch": 0.6753128555176336, "grad_norm": 1.241032329122879, "learning_rate": 2.013653052610678e-06, "loss": 0.4901, "step": 371 }, { "epoch": 0.6771331058020478, "grad_norm": 1.134116834066691, "learning_rate": 1.993185084463106e-06, "loss": 0.5478, "step": 372 }, { "epoch": 0.6789533560864619, "grad_norm": 1.2621524843864569, "learning_rate": 1.97278709770144e-06, "loss": 0.4521, "step": 373 }, { "epoch": 0.680773606370876, "grad_norm": 1.2445963074217277, "learning_rate": 1.952459803641597e-06, "loss": 0.5048, "step": 374 }, { "epoch": 0.6825938566552902, "grad_norm": 1.3367185945909759, "learning_rate": 1.9322039111342977e-06, "loss": 0.4859, "step": 375 }, { "epoch": 0.6844141069397042, "grad_norm": 1.502198228897516, "learning_rate": 1.912020126540366e-06, "loss": 0.5483, "step": 376 }, { "epoch": 0.6862343572241183, "grad_norm": 1.5682296957615942, "learning_rate": 1.8919091537060847e-06, "loss": 0.5403, "step": 377 }, { "epoch": 0.6880546075085324, "grad_norm": 1.2186683041461865, "learning_rate": 1.8718716939386541e-06, "loss": 0.4953, "step": 378 }, { "epoch": 0.6898748577929465, "grad_norm": 1.2104649746142353, "learning_rate": 1.8519084459817362e-06, "loss": 0.4599, "step": 379 }, { "epoch": 0.6916951080773607, "grad_norm": 1.1390426306451955, "learning_rate": 1.83202010599109e-06, "loss": 0.4164, "step": 380 }, { "epoch": 0.6935153583617747, "grad_norm": 1.2956325376708957, "learning_rate": 1.8122073675102932e-06, "loss": 0.5417, "step": 381 }, { "epoch": 0.6953356086461888, "grad_norm": 1.1586136644085798, "learning_rate": 1.792470921446557e-06, "loss": 0.4365, "step": 382 }, { "epoch": 0.697155858930603, "grad_norm": 1.1975210529143365, "learning_rate": 1.7728114560466324e-06, "loss": 0.4956, "step": 383 }, { "epoch": 0.6989761092150171, "grad_norm": 1.5675227569116297, "learning_rate": 1.753229656872815e-06, "loss": 0.4646, "step": 384 }, { "epoch": 0.7007963594994312, "grad_norm": 1.1981622083221466, "learning_rate": 1.7337262067790319e-06, "loss": 0.5042, "step": 385 }, { "epoch": 0.7026166097838453, "grad_norm": 1.291822326824022, "learning_rate": 1.7143017858870259e-06, "loss": 0.5786, "step": 386 }, { "epoch": 0.7044368600682593, "grad_norm": 1.3381873610330526, "learning_rate": 1.6949570715626532e-06, "loss": 0.3987, "step": 387 }, { "epoch": 0.7062571103526735, "grad_norm": 1.5233756050791378, "learning_rate": 1.675692738392247e-06, "loss": 0.5373, "step": 388 }, { "epoch": 0.7080773606370876, "grad_norm": 1.405379762218711, "learning_rate": 1.6565094581591015e-06, "loss": 0.5151, "step": 389 }, { "epoch": 0.7098976109215017, "grad_norm": 1.3827588130238773, "learning_rate": 1.6374078998200424e-06, "loss": 0.4868, "step": 390 }, { "epoch": 0.7117178612059158, "grad_norm": 1.3281467896725871, "learning_rate": 1.6183887294820995e-06, "loss": 0.4892, "step": 391 }, { "epoch": 0.71353811149033, "grad_norm": 1.5562464103926885, "learning_rate": 1.5994526103792852e-06, "loss": 0.5977, "step": 392 }, { "epoch": 0.715358361774744, "grad_norm": 1.2645130650718202, "learning_rate": 1.5806002028494509e-06, "loss": 0.4245, "step": 393 }, { "epoch": 0.7171786120591581, "grad_norm": 1.3281593922925885, "learning_rate": 1.5618321643112738e-06, "loss": 0.5813, "step": 394 }, { "epoch": 0.7189988623435722, "grad_norm": 1.1215366227811656, "learning_rate": 1.5431491492413286e-06, "loss": 0.4276, "step": 395 }, { "epoch": 0.7208191126279864, "grad_norm": 1.3212838118308114, "learning_rate": 1.52455180915126e-06, "loss": 0.5774, "step": 396 }, { "epoch": 0.7226393629124005, "grad_norm": 1.2852914600481689, "learning_rate": 1.506040792565066e-06, "loss": 0.5057, "step": 397 }, { "epoch": 0.7244596131968145, "grad_norm": 1.280275275618163, "learning_rate": 1.487616744996484e-06, "loss": 0.444, "step": 398 }, { "epoch": 0.7262798634812286, "grad_norm": 1.1583238977099228, "learning_rate": 1.4692803089264772e-06, "loss": 0.5377, "step": 399 }, { "epoch": 0.7281001137656428, "grad_norm": 1.435157708312753, "learning_rate": 1.4510321237808377e-06, "loss": 0.5444, "step": 400 }, { "epoch": 0.7299203640500569, "grad_norm": 1.3208185752900872, "learning_rate": 1.4328728259078746e-06, "loss": 0.5566, "step": 401 }, { "epoch": 0.731740614334471, "grad_norm": 1.2130339190915678, "learning_rate": 1.414803048556236e-06, "loss": 0.4988, "step": 402 }, { "epoch": 0.7335608646188851, "grad_norm": 1.1363530661008532, "learning_rate": 1.396823421852825e-06, "loss": 0.6129, "step": 403 }, { "epoch": 0.7353811149032992, "grad_norm": 1.3222588910481998, "learning_rate": 1.3789345727808207e-06, "loss": 0.546, "step": 404 }, { "epoch": 0.7372013651877133, "grad_norm": 1.3949194783709729, "learning_rate": 1.3611371251578114e-06, "loss": 0.5583, "step": 405 }, { "epoch": 0.7390216154721274, "grad_norm": 1.2917335175784925, "learning_rate": 1.3434316996140553e-06, "loss": 0.5151, "step": 406 }, { "epoch": 0.7408418657565415, "grad_norm": 1.2895735708732046, "learning_rate": 1.3258189135708229e-06, "loss": 0.5098, "step": 407 }, { "epoch": 0.7426621160409557, "grad_norm": 1.2978294874532978, "learning_rate": 1.3082993812188735e-06, "loss": 0.5414, "step": 408 }, { "epoch": 0.7444823663253698, "grad_norm": 1.2095221030821062, "learning_rate": 1.2908737134970364e-06, "loss": 0.5268, "step": 409 }, { "epoch": 0.7463026166097838, "grad_norm": 1.3840563503977592, "learning_rate": 1.2735425180709039e-06, "loss": 0.479, "step": 410 }, { "epoch": 0.7481228668941979, "grad_norm": 1.2789076883026242, "learning_rate": 1.2563063993116482e-06, "loss": 0.5503, "step": 411 }, { "epoch": 0.7499431171786121, "grad_norm": 1.283587802005637, "learning_rate": 1.239165958274933e-06, "loss": 0.4113, "step": 412 }, { "epoch": 0.7517633674630262, "grad_norm": 1.2909165266250262, "learning_rate": 1.2221217926799652e-06, "loss": 0.535, "step": 413 }, { "epoch": 0.7535836177474403, "grad_norm": 1.3531455484884616, "learning_rate": 1.2051744968886489e-06, "loss": 0.5052, "step": 414 }, { "epoch": 0.7554038680318543, "grad_norm": 1.2730404093480168, "learning_rate": 1.1883246618848533e-06, "loss": 0.4566, "step": 415 }, { "epoch": 0.7572241183162685, "grad_norm": 1.37616764437592, "learning_rate": 1.1715728752538101e-06, "loss": 0.566, "step": 416 }, { "epoch": 0.7590443686006826, "grad_norm": 1.1512441975212944, "learning_rate": 1.1549197211616203e-06, "loss": 0.5044, "step": 417 }, { "epoch": 0.7608646188850967, "grad_norm": 1.2438970988598956, "learning_rate": 1.1383657803348835e-06, "loss": 0.5109, "step": 418 }, { "epoch": 0.7626848691695108, "grad_norm": 1.5233735431446764, "learning_rate": 1.1219116300404486e-06, "loss": 0.507, "step": 419 }, { "epoch": 0.764505119453925, "grad_norm": 1.3253161212074762, "learning_rate": 1.10555784406528e-06, "loss": 0.5082, "step": 420 }, { "epoch": 0.766325369738339, "grad_norm": 1.1775521474516462, "learning_rate": 1.089304992696455e-06, "loss": 0.46, "step": 421 }, { "epoch": 0.7681456200227531, "grad_norm": 1.2462962157301152, "learning_rate": 1.0731536427012695e-06, "loss": 0.5253, "step": 422 }, { "epoch": 0.7699658703071672, "grad_norm": 1.3347631673740097, "learning_rate": 1.0571043573074736e-06, "loss": 0.4449, "step": 423 }, { "epoch": 0.7717861205915814, "grad_norm": 1.292727758187721, "learning_rate": 1.041157696183641e-06, "loss": 0.441, "step": 424 }, { "epoch": 0.7736063708759955, "grad_norm": 1.293278742294603, "learning_rate": 1.0253142154196415e-06, "loss": 0.4867, "step": 425 }, { "epoch": 0.7754266211604095, "grad_norm": 1.2102494852297525, "learning_rate": 1.0095744675072525e-06, "loss": 0.4898, "step": 426 }, { "epoch": 0.7772468714448236, "grad_norm": 1.224313028246693, "learning_rate": 9.93939001320895e-07, "loss": 0.4686, "step": 427 }, { "epoch": 0.7790671217292378, "grad_norm": 1.3632517015375165, "learning_rate": 9.784083620984884e-07, "loss": 0.4639, "step": 428 }, { "epoch": 0.7808873720136519, "grad_norm": 1.3987002856426751, "learning_rate": 9.62983091422446e-07, "loss": 0.4528, "step": 429 }, { "epoch": 0.782707622298066, "grad_norm": 1.329331750067852, "learning_rate": 9.476637272007746e-07, "loss": 0.4562, "step": 430 }, { "epoch": 0.7845278725824801, "grad_norm": 1.4216744583623766, "learning_rate": 9.324508036483303e-07, "loss": 0.4622, "step": 431 }, { "epoch": 0.7863481228668942, "grad_norm": 1.3060911776176307, "learning_rate": 9.173448512681848e-07, "loss": 0.5405, "step": 432 }, { "epoch": 0.7881683731513083, "grad_norm": 1.3971532684012182, "learning_rate": 9.023463968331238e-07, "loss": 0.4642, "step": 433 }, { "epoch": 0.7899886234357224, "grad_norm": 1.351332971443725, "learning_rate": 8.874559633672754e-07, "loss": 0.4146, "step": 434 }, { "epoch": 0.7918088737201365, "grad_norm": 1.2506853747891504, "learning_rate": 8.726740701278808e-07, "loss": 0.5233, "step": 435 }, { "epoch": 0.7936291240045507, "grad_norm": 1.2588296359051319, "learning_rate": 8.580012325871773e-07, "loss": 0.5196, "step": 436 }, { "epoch": 0.7954493742889648, "grad_norm": 1.3656683873360818, "learning_rate": 8.434379624144261e-07, "loss": 0.4426, "step": 437 }, { "epoch": 0.7972696245733788, "grad_norm": 1.3415371986074633, "learning_rate": 8.289847674580702e-07, "loss": 0.5025, "step": 438 }, { "epoch": 0.7990898748577929, "grad_norm": 1.210310044679145, "learning_rate": 8.146421517280226e-07, "loss": 0.4922, "step": 439 }, { "epoch": 0.800910125142207, "grad_norm": 1.675036054936253, "learning_rate": 8.004106153780967e-07, "loss": 0.4396, "step": 440 }, { "epoch": 0.8027303754266212, "grad_norm": 1.1849449434556916, "learning_rate": 7.862906546885559e-07, "loss": 0.5348, "step": 441 }, { "epoch": 0.8045506257110353, "grad_norm": 1.3294402423567042, "learning_rate": 7.722827620488108e-07, "loss": 0.4472, "step": 442 }, { "epoch": 0.8063708759954493, "grad_norm": 1.2341888202472633, "learning_rate": 7.583874259402545e-07, "loss": 0.4926, "step": 443 }, { "epoch": 0.8081911262798634, "grad_norm": 1.3727750069417188, "learning_rate": 7.446051309192203e-07, "loss": 0.5187, "step": 444 }, { "epoch": 0.8100113765642776, "grad_norm": 1.1665673148184286, "learning_rate": 7.30936357600088e-07, "loss": 0.4459, "step": 445 }, { "epoch": 0.8118316268486917, "grad_norm": 1.4461908262228584, "learning_rate": 7.173815826385246e-07, "loss": 0.5931, "step": 446 }, { "epoch": 0.8136518771331058, "grad_norm": 1.2164762112018974, "learning_rate": 7.039412787148586e-07, "loss": 0.5769, "step": 447 }, { "epoch": 0.81547212741752, "grad_norm": 1.3268169931538385, "learning_rate": 6.906159145176049e-07, "loss": 0.4962, "step": 448 }, { "epoch": 0.817292377701934, "grad_norm": 1.3585533527783662, "learning_rate": 6.774059547271087e-07, "loss": 0.5011, "step": 449 }, { "epoch": 0.8191126279863481, "grad_norm": 1.2715237655057547, "learning_rate": 6.643118599993518e-07, "loss": 0.4591, "step": 450 }, { "epoch": 0.8209328782707622, "grad_norm": 1.1129340141314334, "learning_rate": 6.513340869498858e-07, "loss": 0.4818, "step": 451 }, { "epoch": 0.8227531285551763, "grad_norm": 1.1993408265317198, "learning_rate": 6.384730881379048e-07, "loss": 0.4826, "step": 452 }, { "epoch": 0.8245733788395905, "grad_norm": 1.305009025174831, "learning_rate": 6.257293120504692e-07, "loss": 0.4824, "step": 453 }, { "epoch": 0.8263936291240046, "grad_norm": 1.2620160974509798, "learning_rate": 6.131032030868635e-07, "loss": 0.4479, "step": 454 }, { "epoch": 0.8282138794084186, "grad_norm": 1.2693469945741236, "learning_rate": 6.005952015430993e-07, "loss": 0.5286, "step": 455 }, { "epoch": 0.8300341296928327, "grad_norm": 1.2953927032105943, "learning_rate": 5.882057435965619e-07, "loss": 0.5802, "step": 456 }, { "epoch": 0.8318543799772469, "grad_norm": 1.3055790274997285, "learning_rate": 5.759352612907999e-07, "loss": 0.5273, "step": 457 }, { "epoch": 0.833674630261661, "grad_norm": 1.3009913306704852, "learning_rate": 5.637841825204588e-07, "loss": 0.4434, "step": 458 }, { "epoch": 0.8354948805460751, "grad_norm": 1.3010149850935786, "learning_rate": 5.517529310163627e-07, "loss": 0.5302, "step": 459 }, { "epoch": 0.8373151308304891, "grad_norm": 1.1588504398899486, "learning_rate": 5.398419263307281e-07, "loss": 0.4898, "step": 460 }, { "epoch": 0.8391353811149033, "grad_norm": 1.427556447905731, "learning_rate": 5.280515838225477e-07, "loss": 0.4583, "step": 461 }, { "epoch": 0.8409556313993174, "grad_norm": 1.3382828189315212, "learning_rate": 5.163823146430944e-07, "loss": 0.4544, "step": 462 }, { "epoch": 0.8427758816837315, "grad_norm": 1.3048820751365628, "learning_rate": 5.048345257215892e-07, "loss": 0.5348, "step": 463 }, { "epoch": 0.8445961319681456, "grad_norm": 1.3464339683482869, "learning_rate": 4.934086197510088e-07, "loss": 0.4866, "step": 464 }, { "epoch": 0.8464163822525598, "grad_norm": 1.3076973707605393, "learning_rate": 4.821049951740441e-07, "loss": 0.4374, "step": 465 }, { "epoch": 0.8482366325369738, "grad_norm": 1.207783472984328, "learning_rate": 4.7092404616920547e-07, "loss": 0.5268, "step": 466 }, { "epoch": 0.8500568828213879, "grad_norm": 1.3340034898150066, "learning_rate": 4.59866162637077e-07, "loss": 0.5163, "step": 467 }, { "epoch": 0.851877133105802, "grad_norm": 1.2793323359204207, "learning_rate": 4.4893173018671816e-07, "loss": 0.464, "step": 468 }, { "epoch": 0.8536973833902162, "grad_norm": 1.3875887367624027, "learning_rate": 4.3812113012222164e-07, "loss": 0.5605, "step": 469 }, { "epoch": 0.8555176336746303, "grad_norm": 1.2752397131609516, "learning_rate": 4.2743473942941177e-07, "loss": 0.5166, "step": 470 }, { "epoch": 0.8573378839590444, "grad_norm": 1.361832548391048, "learning_rate": 4.168729307626977e-07, "loss": 0.4494, "step": 471 }, { "epoch": 0.8591581342434584, "grad_norm": 1.3313280628055624, "learning_rate": 4.0643607243208455e-07, "loss": 0.4531, "step": 472 }, { "epoch": 0.8609783845278726, "grad_norm": 1.238927541446331, "learning_rate": 3.9612452839032384e-07, "loss": 0.4629, "step": 473 }, { "epoch": 0.8627986348122867, "grad_norm": 1.440299941933543, "learning_rate": 3.859386582202231e-07, "loss": 0.5238, "step": 474 }, { "epoch": 0.8646188850967008, "grad_norm": 1.2998009060977955, "learning_rate": 3.758788171221079e-07, "loss": 0.4126, "step": 475 }, { "epoch": 0.8664391353811149, "grad_norm": 1.2344313543035759, "learning_rate": 3.659453559014345e-07, "loss": 0.3997, "step": 476 }, { "epoch": 0.868259385665529, "grad_norm": 1.3123945291493502, "learning_rate": 3.561386209565582e-07, "loss": 0.4354, "step": 477 }, { "epoch": 0.8700796359499431, "grad_norm": 1.3385863981096489, "learning_rate": 3.464589542666485e-07, "loss": 0.5423, "step": 478 }, { "epoch": 0.8718998862343572, "grad_norm": 1.4693361278099728, "learning_rate": 3.3690669337976996e-07, "loss": 0.5439, "step": 479 }, { "epoch": 0.8737201365187713, "grad_norm": 1.219115488818529, "learning_rate": 3.2748217140111e-07, "loss": 0.55, "step": 480 }, { "epoch": 0.8755403868031855, "grad_norm": 1.3943875753971013, "learning_rate": 3.1818571698135976e-07, "loss": 0.479, "step": 481 }, { "epoch": 0.8773606370875996, "grad_norm": 1.31283354874802, "learning_rate": 3.0901765430525337e-07, "loss": 0.4546, "step": 482 }, { "epoch": 0.8791808873720136, "grad_norm": 1.2760697675194013, "learning_rate": 2.9997830308027003e-07, "loss": 0.5241, "step": 483 }, { "epoch": 0.8810011376564277, "grad_norm": 1.3057898765814404, "learning_rate": 2.9106797852547483e-07, "loss": 0.5045, "step": 484 }, { "epoch": 0.8828213879408419, "grad_norm": 1.228721100779524, "learning_rate": 2.8228699136053726e-07, "loss": 0.4588, "step": 485 }, { "epoch": 0.884641638225256, "grad_norm": 1.4327825477254865, "learning_rate": 2.7363564779488446e-07, "loss": 0.4911, "step": 486 }, { "epoch": 0.8864618885096701, "grad_norm": 1.1675697744027835, "learning_rate": 2.6511424951703244e-07, "loss": 0.4503, "step": 487 }, { "epoch": 0.8882821387940842, "grad_norm": 1.3573793521283821, "learning_rate": 2.567230936840632e-07, "loss": 0.5537, "step": 488 }, { "epoch": 0.8901023890784983, "grad_norm": 1.2385857779190943, "learning_rate": 2.4846247291125897e-07, "loss": 0.5261, "step": 489 }, { "epoch": 0.8919226393629124, "grad_norm": 1.3747886513978498, "learning_rate": 2.4033267526190057e-07, "loss": 0.5116, "step": 490 }, { "epoch": 0.8937428896473265, "grad_norm": 1.3015002806547666, "learning_rate": 2.323339842372234e-07, "loss": 0.501, "step": 491 }, { "epoch": 0.8955631399317406, "grad_norm": 1.2282471393147485, "learning_rate": 2.2446667876652968e-07, "loss": 0.5615, "step": 492 }, { "epoch": 0.8973833902161548, "grad_norm": 1.2246787300329813, "learning_rate": 2.1673103319746146e-07, "loss": 0.5847, "step": 493 }, { "epoch": 0.8992036405005688, "grad_norm": 1.381507003520726, "learning_rate": 2.0912731728643362e-07, "loss": 0.4593, "step": 494 }, { "epoch": 0.9010238907849829, "grad_norm": 1.2236872730147548, "learning_rate": 2.0165579618922757e-07, "loss": 0.426, "step": 495 }, { "epoch": 0.902844141069397, "grad_norm": 1.5642555968533283, "learning_rate": 1.943167304517459e-07, "loss": 0.4669, "step": 496 }, { "epoch": 0.9046643913538112, "grad_norm": 1.4113035349877263, "learning_rate": 1.871103760009234e-07, "loss": 0.5189, "step": 497 }, { "epoch": 0.9064846416382253, "grad_norm": 1.2945664446971985, "learning_rate": 1.8003698413580427e-07, "loss": 0.5331, "step": 498 }, { "epoch": 0.9083048919226394, "grad_norm": 1.1216172834522593, "learning_rate": 1.7309680151878126e-07, "loss": 0.4596, "step": 499 }, { "epoch": 0.9101251422070534, "grad_norm": 1.4490178400997769, "learning_rate": 1.6629007016698916e-07, "loss": 0.5719, "step": 500 }, { "epoch": 0.9119453924914676, "grad_norm": 1.4927164965040023, "learning_rate": 1.5961702744386973e-07, "loss": 0.4637, "step": 501 }, { "epoch": 0.9137656427758817, "grad_norm": 1.2926779903672145, "learning_rate": 1.5307790605089045e-07, "loss": 0.4931, "step": 502 }, { "epoch": 0.9155858930602958, "grad_norm": 1.3434407972538571, "learning_rate": 1.4667293401943393e-07, "loss": 0.4843, "step": 503 }, { "epoch": 0.9174061433447099, "grad_norm": 1.2627460036138376, "learning_rate": 1.404023347028418e-07, "loss": 0.4628, "step": 504 }, { "epoch": 0.919226393629124, "grad_norm": 1.1980109325087624, "learning_rate": 1.342663267686297e-07, "loss": 0.547, "step": 505 }, { "epoch": 0.9210466439135381, "grad_norm": 1.4394748326258473, "learning_rate": 1.2826512419085922e-07, "loss": 0.4852, "step": 506 }, { "epoch": 0.9228668941979522, "grad_norm": 1.1712799414971835, "learning_rate": 1.223989362426785e-07, "loss": 0.5027, "step": 507 }, { "epoch": 0.9246871444823663, "grad_norm": 1.2917639503148088, "learning_rate": 1.1666796748902142e-07, "loss": 0.4318, "step": 508 }, { "epoch": 0.9265073947667805, "grad_norm": 1.407559329871179, "learning_rate": 1.1107241777947774e-07, "loss": 0.452, "step": 509 }, { "epoch": 0.9283276450511946, "grad_norm": 1.4395176866301798, "learning_rate": 1.0561248224132091e-07, "loss": 0.5792, "step": 510 }, { "epoch": 0.9301478953356086, "grad_norm": 1.3107228117658043, "learning_rate": 1.0028835127270552e-07, "loss": 0.523, "step": 511 }, { "epoch": 0.9319681456200227, "grad_norm": 1.319280624009732, "learning_rate": 9.510021053602679e-08, "loss": 0.4903, "step": 512 }, { "epoch": 0.9337883959044369, "grad_norm": 1.2825750147020196, "learning_rate": 9.004824095144581e-08, "loss": 0.486, "step": 513 }, { "epoch": 0.935608646188851, "grad_norm": 1.3550036994824897, "learning_rate": 8.513261869058209e-08, "loss": 0.4342, "step": 514 }, { "epoch": 0.9374288964732651, "grad_norm": 1.2912511428181583, "learning_rate": 8.035351517036914e-08, "loss": 0.4975, "step": 515 }, { "epoch": 0.9392491467576792, "grad_norm": 1.2630516224119532, "learning_rate": 7.571109704707623e-08, "loss": 0.4942, "step": 516 }, { "epoch": 0.9410693970420932, "grad_norm": 1.2630983628627157, "learning_rate": 7.120552621049825e-08, "loss": 0.4581, "step": 517 }, { "epoch": 0.9428896473265074, "grad_norm": 1.184276479260659, "learning_rate": 6.68369597783096e-08, "loss": 0.4245, "step": 518 }, { "epoch": 0.9447098976109215, "grad_norm": 1.3479750123046965, "learning_rate": 6.260555009058288e-08, "loss": 0.4734, "step": 519 }, { "epoch": 0.9465301478953356, "grad_norm": 1.184265059530281, "learning_rate": 5.851144470448144e-08, "loss": 0.5263, "step": 520 }, { "epoch": 0.9483503981797498, "grad_norm": 1.3131542129196199, "learning_rate": 5.455478638911071e-08, "loss": 0.369, "step": 521 }, { "epoch": 0.9501706484641638, "grad_norm": 1.3396828056059393, "learning_rate": 5.073571312053815e-08, "loss": 0.5098, "step": 522 }, { "epoch": 0.9519908987485779, "grad_norm": 1.3620509437765531, "learning_rate": 4.705435807698555e-08, "loss": 0.5595, "step": 523 }, { "epoch": 0.953811149032992, "grad_norm": 1.3476395824069989, "learning_rate": 4.351084963418117e-08, "loss": 0.5332, "step": 524 }, { "epoch": 0.9556313993174061, "grad_norm": 1.4056028428746756, "learning_rate": 4.010531136088691e-08, "loss": 0.5135, "step": 525 }, { "epoch": 0.9574516496018203, "grad_norm": 1.2931973314368226, "learning_rate": 3.683786201458439e-08, "loss": 0.4869, "step": 526 }, { "epoch": 0.9592718998862344, "grad_norm": 1.3728587745363008, "learning_rate": 3.370861553733784e-08, "loss": 0.544, "step": 527 }, { "epoch": 0.9610921501706484, "grad_norm": 1.379130929011516, "learning_rate": 3.071768105181993e-08, "loss": 0.4312, "step": 528 }, { "epoch": 0.9629124004550625, "grad_norm": 1.4162454724368647, "learning_rate": 2.786516285750373e-08, "loss": 0.4464, "step": 529 }, { "epoch": 0.9647326507394767, "grad_norm": 1.310107669303508, "learning_rate": 2.5151160427029582e-08, "loss": 0.4641, "step": 530 }, { "epoch": 0.9665529010238908, "grad_norm": 1.3049449814100964, "learning_rate": 2.2575768402733232e-08, "loss": 0.5079, "step": 531 }, { "epoch": 0.9683731513083049, "grad_norm": 1.301610299072927, "learning_rate": 2.013907659334624e-08, "loss": 0.4798, "step": 532 }, { "epoch": 0.970193401592719, "grad_norm": 1.2912511438851022, "learning_rate": 1.7841169970866042e-08, "loss": 0.4962, "step": 533 }, { "epoch": 0.9720136518771331, "grad_norm": 1.3741948558886383, "learning_rate": 1.5682128667589e-08, "loss": 0.4556, "step": 534 }, { "epoch": 0.9738339021615472, "grad_norm": 1.4394930710163565, "learning_rate": 1.3662027973320612e-08, "loss": 0.4808, "step": 535 }, { "epoch": 0.9756541524459613, "grad_norm": 1.3189189784853037, "learning_rate": 1.1780938332746515e-08, "loss": 0.4601, "step": 536 }, { "epoch": 0.9774744027303754, "grad_norm": 1.3675135301050803, "learning_rate": 1.0038925342977122e-08, "loss": 0.4862, "step": 537 }, { "epoch": 0.9792946530147896, "grad_norm": 1.2496142970199702, "learning_rate": 8.43604975126011e-09, "loss": 0.4972, "step": 538 }, { "epoch": 0.9811149032992036, "grad_norm": 1.2590220357743287, "learning_rate": 6.972367452863004e-09, "loss": 0.5048, "step": 539 }, { "epoch": 0.9829351535836177, "grad_norm": 1.4321645089041766, "learning_rate": 5.647929489122738e-09, "loss": 0.5688, "step": 540 }, { "epoch": 0.9847554038680318, "grad_norm": 1.3266758203446563, "learning_rate": 4.462782045664859e-09, "loss": 0.4745, "step": 541 }, { "epoch": 0.986575654152446, "grad_norm": 1.1860109568892805, "learning_rate": 3.4169664507959216e-09, "loss": 0.4616, "step": 542 }, { "epoch": 0.9883959044368601, "grad_norm": 1.2475217679660848, "learning_rate": 2.5105191740597553e-09, "loss": 0.6489, "step": 543 }, { "epoch": 0.9902161547212742, "grad_norm": 1.4270596684886099, "learning_rate": 1.7434718249664803e-09, "loss": 0.4712, "step": 544 }, { "epoch": 0.9920364050056882, "grad_norm": 1.2999695109285117, "learning_rate": 1.1158511518902791e-09, "loss": 0.5143, "step": 545 }, { "epoch": 0.9938566552901024, "grad_norm": 1.5041014788909566, "learning_rate": 6.276790411372524e-10, "loss": 0.4971, "step": 546 }, { "epoch": 0.9956769055745165, "grad_norm": 1.151430673100721, "learning_rate": 2.789725161806977e-10, "loss": 0.5446, "step": 547 }, { "epoch": 0.9974971558589306, "grad_norm": 1.2513093960410882, "learning_rate": 6.974373706869486e-11, "loss": 0.5494, "step": 548 }, { "epoch": 0.9993174061433447, "grad_norm": 1.3143110421818924, "learning_rate": 0.0, "loss": 0.4345, "step": 549 }, { "epoch": 0.9993174061433447, "eval_accuracy": 0.8141491715694426, "eval_accuracy_first_token": 0.7775706741223982, "eval_accuracy_first_token_<": 0.9648293963254593, "eval_accuracy_first_token_<_total": 1905, "eval_accuracy_first_token_<|python_tag|>": 0.9014084507042254, "eval_accuracy_first_token_<|python_tag|>_total": 994, "eval_accuracy_first_token_Certainly": 0.743801652892562, "eval_accuracy_first_token_Certainly_total": 363, "eval_accuracy_first_token_The": 0.9030402629416598, "eval_accuracy_first_token_The_total": 2434, "eval_accuracy_first_token_To": 0.8076923076923077, "eval_accuracy_first_token_To_total": 936, "eval_loss": 0.5610479116439819, "eval_perplexity": 1.110590475782418, "eval_runtime": 507.5321, "eval_samples_per_second": 1.373, "eval_steps_per_second": 0.688, "eval_total_number_first_token": 9657, "step": 549 }, { "epoch": 0.9993174061433447, "step": 549, "total_flos": 229846517022720.0, "train_loss": 0.5347839987994544, "train_runtime": 35925.1648, "train_samples_per_second": 0.245, "train_steps_per_second": 0.015 } ], "logging_steps": 1.0, "max_steps": 549, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 5.0, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": false, "should_training_stop": false }, "attributes": {} } }, "total_flos": 229846517022720.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }