{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.9993181043300376, "eval_steps": 500, "global_step": 1466, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0013637913399249914, "grad_norm": 1.598020721574378, "learning_rate": 6.8027210884353745e-06, "loss": 1.3936, "step": 1 }, { "epoch": 0.006818956699624957, "grad_norm": 1.582179170492486, "learning_rate": 3.4013605442176877e-05, "loss": 1.3732, "step": 5 }, { "epoch": 0.013637913399249914, "grad_norm": 0.5247209097240487, "learning_rate": 6.802721088435375e-05, "loss": 1.3443, "step": 10 }, { "epoch": 0.020456870098874872, "grad_norm": 0.8017522199019398, "learning_rate": 0.00010204081632653062, "loss": 1.2617, "step": 15 }, { "epoch": 0.02727582679849983, "grad_norm": 0.4014040737926046, "learning_rate": 0.0001360544217687075, "loss": 1.1857, "step": 20 }, { "epoch": 0.03409478349812479, "grad_norm": 0.32790437534680206, "learning_rate": 0.00017006802721088434, "loss": 1.1436, "step": 25 }, { "epoch": 0.040913740197749744, "grad_norm": 0.2966698205141879, "learning_rate": 0.00020408163265306123, "loss": 1.0996, "step": 30 }, { "epoch": 0.0477326968973747, "grad_norm": 0.17808447662676022, "learning_rate": 0.0002380952380952381, "loss": 1.0757, "step": 35 }, { "epoch": 0.05455165359699966, "grad_norm": 0.14553030270593192, "learning_rate": 0.000272108843537415, "loss": 1.07, "step": 40 }, { "epoch": 0.06137061029662462, "grad_norm": 0.12365194167763814, "learning_rate": 0.0003061224489795919, "loss": 1.0494, "step": 45 }, { "epoch": 0.06818956699624958, "grad_norm": 0.19450168087415454, "learning_rate": 0.0003401360544217687, "loss": 1.0515, "step": 50 }, { "epoch": 0.07500852369587453, "grad_norm": 0.1207148840427689, "learning_rate": 0.0003741496598639456, "loss": 1.0253, "step": 55 }, { "epoch": 0.08182748039549949, "grad_norm": 0.10854420757501072, "learning_rate": 0.00040816326530612246, "loss": 1.0353, "step": 60 }, { "epoch": 0.08864643709512444, "grad_norm": 0.16561709445658757, "learning_rate": 0.0004421768707482993, "loss": 1.0285, "step": 65 }, { "epoch": 0.0954653937947494, "grad_norm": 0.524524123408497, "learning_rate": 0.0004761904761904762, "loss": 1.0379, "step": 70 }, { "epoch": 0.10228435049437436, "grad_norm": 0.12853108179786138, "learning_rate": 0.0005102040816326531, "loss": 1.0186, "step": 75 }, { "epoch": 0.10910330719399931, "grad_norm": 0.2734451972437396, "learning_rate": 0.00054421768707483, "loss": 1.0063, "step": 80 }, { "epoch": 0.11592226389362427, "grad_norm": 0.14711225612929515, "learning_rate": 0.0005782312925170068, "loss": 1.0132, "step": 85 }, { "epoch": 0.12274122059324924, "grad_norm": 0.11857109812836017, "learning_rate": 0.0006122448979591838, "loss": 1.0131, "step": 90 }, { "epoch": 0.1295601772928742, "grad_norm": 0.16843211920179874, "learning_rate": 0.0006462585034013606, "loss": 0.9969, "step": 95 }, { "epoch": 0.13637913399249915, "grad_norm": 0.15975057567245052, "learning_rate": 0.0006802721088435374, "loss": 0.9909, "step": 100 }, { "epoch": 0.1431980906921241, "grad_norm": 0.15632985298714983, "learning_rate": 0.0007142857142857143, "loss": 0.9966, "step": 105 }, { "epoch": 0.15001704739174906, "grad_norm": 0.1621684383980749, "learning_rate": 0.0007482993197278912, "loss": 0.9915, "step": 110 }, { "epoch": 0.15683600409137402, "grad_norm": 0.12303095362829028, "learning_rate": 0.000782312925170068, "loss": 0.9787, "step": 115 }, { "epoch": 0.16365496079099898, "grad_norm": 0.14599562372712946, "learning_rate": 0.0008163265306122449, "loss": 0.9882, "step": 120 }, { "epoch": 0.17047391749062393, "grad_norm": 0.14693955034453152, "learning_rate": 0.0008503401360544217, "loss": 0.9801, "step": 125 }, { "epoch": 0.1772928741902489, "grad_norm": 0.22554930118958344, "learning_rate": 0.0008843537414965987, "loss": 0.9767, "step": 130 }, { "epoch": 0.18411183088987385, "grad_norm": 0.15209662483639966, "learning_rate": 0.0009183673469387756, "loss": 0.9827, "step": 135 }, { "epoch": 0.1909307875894988, "grad_norm": 0.1338153825791751, "learning_rate": 0.0009523809523809524, "loss": 0.9725, "step": 140 }, { "epoch": 0.19774974428912376, "grad_norm": 0.16873273473563985, "learning_rate": 0.0009863945578231293, "loss": 0.9747, "step": 145 }, { "epoch": 0.20456870098874871, "grad_norm": 0.18631348995548455, "learning_rate": 0.000999987235881584, "loss": 0.9737, "step": 150 }, { "epoch": 0.21138765768837367, "grad_norm": 0.1398209387998964, "learning_rate": 0.0009999092352957284, "loss": 0.9848, "step": 155 }, { "epoch": 0.21820661438799863, "grad_norm": 0.13947559562857936, "learning_rate": 0.0009997603363497414, "loss": 0.9643, "step": 160 }, { "epoch": 0.22502557108762358, "grad_norm": 0.14313053783616825, "learning_rate": 0.000999540560160838, "loss": 0.974, "step": 165 }, { "epoch": 0.23184452778724854, "grad_norm": 0.13580380953886942, "learning_rate": 0.0009992499378982194, "loss": 0.9761, "step": 170 }, { "epoch": 0.2386634844868735, "grad_norm": 0.14428497093865783, "learning_rate": 0.0009988885107786517, "loss": 0.9652, "step": 175 }, { "epoch": 0.24548244118649848, "grad_norm": 0.12382971206501955, "learning_rate": 0.0009984563300606192, "loss": 0.9678, "step": 180 }, { "epoch": 0.25230139788612344, "grad_norm": 0.14065780649342097, "learning_rate": 0.0009979534570370575, "loss": 0.969, "step": 185 }, { "epoch": 0.2591203545857484, "grad_norm": 0.11357268655774008, "learning_rate": 0.000997379963026658, "loss": 0.9692, "step": 190 }, { "epoch": 0.26593931128537335, "grad_norm": 0.1622202809625719, "learning_rate": 0.0009967359293637553, "loss": 0.9588, "step": 195 }, { "epoch": 0.2727582679849983, "grad_norm": 0.1776637907272292, "learning_rate": 0.0009960214473867907, "loss": 0.9613, "step": 200 }, { "epoch": 0.27957722468462326, "grad_norm": 0.1215741064620054, "learning_rate": 0.0009952366184253602, "loss": 0.963, "step": 205 }, { "epoch": 0.2863961813842482, "grad_norm": 0.1302207261176898, "learning_rate": 0.0009943815537858415, "loss": 0.9551, "step": 210 }, { "epoch": 0.2932151380838732, "grad_norm": 0.11682895510290592, "learning_rate": 0.00099345637473561, "loss": 0.9598, "step": 215 }, { "epoch": 0.30003409478349813, "grad_norm": 0.15188852916347045, "learning_rate": 0.0009924612124858389, "loss": 0.9583, "step": 220 }, { "epoch": 0.3068530514831231, "grad_norm": 0.13240495035670563, "learning_rate": 0.0009913962081728918, "loss": 0.9538, "step": 225 }, { "epoch": 0.31367200818274804, "grad_norm": 0.16120589037533914, "learning_rate": 0.0009902615128383062, "loss": 0.9464, "step": 230 }, { "epoch": 0.320490964882373, "grad_norm": 0.1382061477313828, "learning_rate": 0.0009890572874073713, "loss": 0.9434, "step": 235 }, { "epoch": 0.32730992158199795, "grad_norm": 0.11549986617558586, "learning_rate": 0.0009877837026663068, "loss": 0.9441, "step": 240 }, { "epoch": 0.3341288782816229, "grad_norm": 0.10980499757109458, "learning_rate": 0.00098644093923804, "loss": 0.96, "step": 245 }, { "epoch": 0.34094783498124787, "grad_norm": 0.19062656435818273, "learning_rate": 0.0009850291875565908, "loss": 0.9577, "step": 250 }, { "epoch": 0.3477667916808728, "grad_norm": 0.12141161965377588, "learning_rate": 0.0009835486478400625, "loss": 0.9456, "step": 255 }, { "epoch": 0.3545857483804978, "grad_norm": 0.11932394138168441, "learning_rate": 0.000981999530062248, "loss": 0.9546, "step": 260 }, { "epoch": 0.36140470508012273, "grad_norm": 0.1277558444188088, "learning_rate": 0.0009803820539228492, "loss": 0.9375, "step": 265 }, { "epoch": 0.3682236617797477, "grad_norm": 0.11354346745136285, "learning_rate": 0.0009786964488163194, "loss": 0.9473, "step": 270 }, { "epoch": 0.37504261847937265, "grad_norm": 0.10871503403228854, "learning_rate": 0.000976942953799331, "loss": 0.9321, "step": 275 }, { "epoch": 0.3818615751789976, "grad_norm": 0.14107095714136358, "learning_rate": 0.0009751218175568688, "loss": 0.9346, "step": 280 }, { "epoch": 0.38868053187862256, "grad_norm": 0.11468378232954199, "learning_rate": 0.0009732332983669651, "loss": 0.9333, "step": 285 }, { "epoch": 0.3954994885782475, "grad_norm": 0.11614073571579422, "learning_rate": 0.0009712776640640671, "loss": 0.933, "step": 290 }, { "epoch": 0.40231844527787247, "grad_norm": 0.14233058701254916, "learning_rate": 0.0009692551920010519, "loss": 0.9365, "step": 295 }, { "epoch": 0.40913740197749743, "grad_norm": 0.16365059288233255, "learning_rate": 0.0009671661690098941, "loss": 0.9345, "step": 300 }, { "epoch": 0.4159563586771224, "grad_norm": 0.2778458797011963, "learning_rate": 0.0009650108913609837, "loss": 0.9243, "step": 305 }, { "epoch": 0.42277531537674734, "grad_norm": 0.12497969488106986, "learning_rate": 0.0009627896647211103, "loss": 0.915, "step": 310 }, { "epoch": 0.4295942720763723, "grad_norm": 0.12132320992585822, "learning_rate": 0.0009605028041101116, "loss": 0.9306, "step": 315 }, { "epoch": 0.43641322877599725, "grad_norm": 0.15422956629129714, "learning_rate": 0.0009581506338561974, "loss": 0.9229, "step": 320 }, { "epoch": 0.4432321854756222, "grad_norm": 0.10857379661878619, "learning_rate": 0.0009557334875499513, "loss": 0.9295, "step": 325 }, { "epoch": 0.45005114217524717, "grad_norm": 0.09700456702304185, "learning_rate": 0.0009532517079970214, "loss": 0.9144, "step": 330 }, { "epoch": 0.4568700988748721, "grad_norm": 0.1574301623396698, "learning_rate": 0.000950705647169502, "loss": 0.9303, "step": 335 }, { "epoch": 0.4636890555744971, "grad_norm": 0.12653500024100792, "learning_rate": 0.000948095666156016, "loss": 0.9145, "step": 340 }, { "epoch": 0.47050801227412203, "grad_norm": 0.12348356194845742, "learning_rate": 0.0009454221351105055, "loss": 0.9115, "step": 345 }, { "epoch": 0.477326968973747, "grad_norm": 0.1307740739484115, "learning_rate": 0.0009426854331997334, "loss": 0.9188, "step": 350 }, { "epoch": 0.484145925673372, "grad_norm": 0.10835447562048035, "learning_rate": 0.0009398859485495119, "loss": 0.9247, "step": 355 }, { "epoch": 0.49096488237299696, "grad_norm": 0.11863788647430744, "learning_rate": 0.0009370240781896553, "loss": 0.9102, "step": 360 }, { "epoch": 0.4977838390726219, "grad_norm": 0.14783785955650325, "learning_rate": 0.0009341002279976728, "loss": 0.9136, "step": 365 }, { "epoch": 0.5046027957722469, "grad_norm": 0.11185935178347504, "learning_rate": 0.0009311148126412067, "loss": 0.9108, "step": 370 }, { "epoch": 0.5114217524718718, "grad_norm": 0.11360824911697796, "learning_rate": 0.0009280682555192229, "loss": 0.9167, "step": 375 }, { "epoch": 0.5182407091714968, "grad_norm": 0.10219440421235579, "learning_rate": 0.0009249609887019624, "loss": 0.9125, "step": 380 }, { "epoch": 0.5250596658711217, "grad_norm": 0.1063437764241102, "learning_rate": 0.0009217934528696652, "loss": 0.9135, "step": 385 }, { "epoch": 0.5318786225707467, "grad_norm": 0.12924256438154388, "learning_rate": 0.000918566097250072, "loss": 0.9044, "step": 390 }, { "epoch": 0.5386975792703717, "grad_norm": 0.1672822885194737, "learning_rate": 0.0009152793795547129, "loss": 0.905, "step": 395 }, { "epoch": 0.5455165359699966, "grad_norm": 0.18170913356641202, "learning_rate": 0.0009119337659139939, "loss": 0.9077, "step": 400 }, { "epoch": 0.5523354926696216, "grad_norm": 0.14583299571955888, "learning_rate": 0.0009085297308110889, "loss": 0.9013, "step": 405 }, { "epoch": 0.5591544493692465, "grad_norm": 0.1328861894472585, "learning_rate": 0.0009050677570146482, "loss": 0.8972, "step": 410 }, { "epoch": 0.5659734060688715, "grad_norm": 0.12732498687417756, "learning_rate": 0.0009015483355103298, "loss": 0.8948, "step": 415 }, { "epoch": 0.5727923627684964, "grad_norm": 0.10225924635313659, "learning_rate": 0.0008979719654311677, "loss": 0.8977, "step": 420 }, { "epoch": 0.5796113194681214, "grad_norm": 0.11222734733135577, "learning_rate": 0.0008943391539867831, "loss": 0.9086, "step": 425 }, { "epoch": 0.5864302761677463, "grad_norm": 0.10007024319396716, "learning_rate": 0.0008906504163914506, "loss": 0.9047, "step": 430 }, { "epoch": 0.5932492328673713, "grad_norm": 0.11019466700605089, "learning_rate": 0.0008869062757910296, "loss": 0.8998, "step": 435 }, { "epoch": 0.6000681895669963, "grad_norm": 0.12131434093089721, "learning_rate": 0.00088310726318877, "loss": 0.9073, "step": 440 }, { "epoch": 0.6068871462666212, "grad_norm": 0.09939916214847802, "learning_rate": 0.0008792539173700046, "loss": 0.9056, "step": 445 }, { "epoch": 0.6137061029662462, "grad_norm": 0.11240353319322192, "learning_rate": 0.0008753467848257366, "loss": 0.9023, "step": 450 }, { "epoch": 0.6205250596658711, "grad_norm": 0.12884203346821665, "learning_rate": 0.0008713864196751353, "loss": 0.891, "step": 455 }, { "epoch": 0.6273440163654961, "grad_norm": 0.13107538778471978, "learning_rate": 0.0008673733835869496, "loss": 0.9053, "step": 460 }, { "epoch": 0.634162973065121, "grad_norm": 0.1265128902066715, "learning_rate": 0.0008633082456998505, "loss": 0.8765, "step": 465 }, { "epoch": 0.640981929764746, "grad_norm": 0.11168557466408092, "learning_rate": 0.0008591915825417144, "loss": 0.8937, "step": 470 }, { "epoch": 0.647800886464371, "grad_norm": 0.10749488343940884, "learning_rate": 0.0008550239779478592, "loss": 0.8936, "step": 475 }, { "epoch": 0.6546198431639959, "grad_norm": 0.12116636988129866, "learning_rate": 0.0008508060229782422, "loss": 0.8985, "step": 480 }, { "epoch": 0.6614387998636209, "grad_norm": 0.11248409191993927, "learning_rate": 0.0008465383158336352, "loss": 0.9068, "step": 485 }, { "epoch": 0.6682577565632458, "grad_norm": 0.09616110942207395, "learning_rate": 0.0008422214617707864, "loss": 0.893, "step": 490 }, { "epoch": 0.6750767132628708, "grad_norm": 0.10944978412158525, "learning_rate": 0.000837856073016581, "loss": 0.9015, "step": 495 }, { "epoch": 0.6818956699624957, "grad_norm": 0.12373353259009305, "learning_rate": 0.0008334427686812137, "loss": 0.8805, "step": 500 }, { "epoch": 0.6887146266621207, "grad_norm": 0.10998182720982563, "learning_rate": 0.000828982174670385, "loss": 0.8764, "step": 505 }, { "epoch": 0.6955335833617456, "grad_norm": 0.11981879919775634, "learning_rate": 0.0008244749235965338, "loss": 0.8972, "step": 510 }, { "epoch": 0.7023525400613706, "grad_norm": 0.11009515701872061, "learning_rate": 0.000819921654689119, "loss": 0.883, "step": 515 }, { "epoch": 0.7091714967609956, "grad_norm": 0.1206592530642094, "learning_rate": 0.0008153230137039615, "loss": 0.8887, "step": 520 }, { "epoch": 0.7159904534606205, "grad_norm": 0.10679229804484212, "learning_rate": 0.0008106796528316626, "loss": 0.8894, "step": 525 }, { "epoch": 0.7228094101602455, "grad_norm": 0.11030470444996518, "learning_rate": 0.000805992230605108, "loss": 0.8881, "step": 530 }, { "epoch": 0.7296283668598704, "grad_norm": 0.14989020257644473, "learning_rate": 0.0008012614118060733, "loss": 0.8758, "step": 535 }, { "epoch": 0.7364473235594954, "grad_norm": 0.19608291727965096, "learning_rate": 0.0007964878673709432, "loss": 0.873, "step": 540 }, { "epoch": 0.7432662802591203, "grad_norm": 0.1357305661381077, "learning_rate": 0.0007916722742955573, "loss": 0.878, "step": 545 }, { "epoch": 0.7500852369587453, "grad_norm": 0.09279906563803725, "learning_rate": 0.0007868153155391968, "loss": 0.8844, "step": 550 }, { "epoch": 0.7569041936583703, "grad_norm": 0.1003975725083102, "learning_rate": 0.0007819176799277262, "loss": 0.8875, "step": 555 }, { "epoch": 0.7637231503579952, "grad_norm": 0.09978871199012411, "learning_rate": 0.0007769800620559015, "loss": 0.8866, "step": 560 }, { "epoch": 0.7705421070576202, "grad_norm": 0.13532770031312716, "learning_rate": 0.0007720031621888615, "loss": 0.879, "step": 565 }, { "epoch": 0.7773610637572451, "grad_norm": 0.10118664495666357, "learning_rate": 0.0007669876861628144, "loss": 0.87, "step": 570 }, { "epoch": 0.7841800204568701, "grad_norm": 0.09994831940420743, "learning_rate": 0.0007619343452849349, "loss": 0.8759, "step": 575 }, { "epoch": 0.790998977156495, "grad_norm": 0.12994411471721243, "learning_rate": 0.0007568438562324833, "loss": 0.8783, "step": 580 }, { "epoch": 0.79781793385612, "grad_norm": 0.1169831966799223, "learning_rate": 0.0007517169409511664, "loss": 0.8672, "step": 585 }, { "epoch": 0.8046368905557449, "grad_norm": 0.10141671278849922, "learning_rate": 0.0007465543265527482, "loss": 0.8695, "step": 590 }, { "epoch": 0.8114558472553699, "grad_norm": 0.11783176873723326, "learning_rate": 0.0007413567452119298, "loss": 0.8689, "step": 595 }, { "epoch": 0.8182748039549949, "grad_norm": 0.12718960852527547, "learning_rate": 0.00073612493406251, "loss": 0.8708, "step": 600 }, { "epoch": 0.8250937606546198, "grad_norm": 0.12649370297868867, "learning_rate": 0.0007308596350928434, "loss": 0.8759, "step": 605 }, { "epoch": 0.8319127173542448, "grad_norm": 0.1009142565676403, "learning_rate": 0.0007255615950406102, "loss": 0.862, "step": 610 }, { "epoch": 0.8387316740538697, "grad_norm": 0.10490423262148889, "learning_rate": 0.0007202315652869112, "loss": 0.87, "step": 615 }, { "epoch": 0.8455506307534947, "grad_norm": 0.09903668718402683, "learning_rate": 0.0007148703017497058, "loss": 0.8705, "step": 620 }, { "epoch": 0.8523695874531196, "grad_norm": 0.14442585966687965, "learning_rate": 0.0007094785647766055, "loss": 0.8681, "step": 625 }, { "epoch": 0.8591885441527446, "grad_norm": 0.16498073796059864, "learning_rate": 0.0007040571190370397, "loss": 0.8656, "step": 630 }, { "epoch": 0.8660075008523695, "grad_norm": 0.1129999366359928, "learning_rate": 0.0006986067334138079, "loss": 0.8686, "step": 635 }, { "epoch": 0.8728264575519945, "grad_norm": 0.09929311073474267, "learning_rate": 0.0006931281808940361, "loss": 0.8687, "step": 640 }, { "epoch": 0.8796454142516195, "grad_norm": 0.11972041401629113, "learning_rate": 0.0006876222384595477, "loss": 0.8652, "step": 645 }, { "epoch": 0.8864643709512444, "grad_norm": 0.09934940714328987, "learning_rate": 0.0006820896869766725, "loss": 0.8709, "step": 650 }, { "epoch": 0.8932833276508694, "grad_norm": 0.10973507676224258, "learning_rate": 0.0006765313110855009, "loss": 0.8652, "step": 655 }, { "epoch": 0.9001022843504943, "grad_norm": 0.09165992876326395, "learning_rate": 0.0006709478990886039, "loss": 0.8611, "step": 660 }, { "epoch": 0.9069212410501193, "grad_norm": 0.09988220618681196, "learning_rate": 0.0006653402428392354, "loss": 0.8556, "step": 665 }, { "epoch": 0.9137401977497442, "grad_norm": 0.09446350263562757, "learning_rate": 0.0006597091376290288, "loss": 0.8528, "step": 670 }, { "epoch": 0.9205591544493692, "grad_norm": 0.08902820350532409, "learning_rate": 0.0006540553820752069, "loss": 0.8609, "step": 675 }, { "epoch": 0.9273781111489942, "grad_norm": 0.10011816954581992, "learning_rate": 0.00064837977800732, "loss": 0.8625, "step": 680 }, { "epoch": 0.9341970678486191, "grad_norm": 0.08775726631732685, "learning_rate": 0.0006426831303535284, "loss": 0.8576, "step": 685 }, { "epoch": 0.9410160245482441, "grad_norm": 0.11117730920395297, "learning_rate": 0.0006369662470264462, "loss": 0.8704, "step": 690 }, { "epoch": 0.947834981247869, "grad_norm": 0.139154974071725, "learning_rate": 0.0006312299388085596, "loss": 0.8584, "step": 695 }, { "epoch": 0.954653937947494, "grad_norm": 0.10346972946518294, "learning_rate": 0.0006254750192372418, "loss": 0.8709, "step": 700 }, { "epoch": 0.9614728946471189, "grad_norm": 0.09900042387283375, "learning_rate": 0.0006197023044893734, "loss": 0.8482, "step": 705 }, { "epoch": 0.968291851346744, "grad_norm": 0.09908280186750126, "learning_rate": 0.0006139126132655905, "loss": 0.8597, "step": 710 }, { "epoch": 0.975110808046369, "grad_norm": 0.13127044402154367, "learning_rate": 0.0006081067666741757, "loss": 0.8483, "step": 715 }, { "epoch": 0.9819297647459939, "grad_norm": 0.09623829465725477, "learning_rate": 0.0006022855881146053, "loss": 0.8589, "step": 720 }, { "epoch": 0.9887487214456189, "grad_norm": 0.09762948252187646, "learning_rate": 0.0005964499031607727, "loss": 0.8479, "step": 725 }, { "epoch": 0.9955676781452438, "grad_norm": 0.11109345413298818, "learning_rate": 0.0005906005394439044, "loss": 0.8625, "step": 730 }, { "epoch": 0.9996590521650187, "eval_loss": 0.8650394678115845, "eval_runtime": 60.297, "eval_samples_per_second": 184.404, "eval_steps_per_second": 5.771, "step": 733 }, { "epoch": 1.0023866348448687, "grad_norm": 0.150003136868036, "learning_rate": 0.0005847383265351828, "loss": 0.8218, "step": 735 }, { "epoch": 1.0092055915444937, "grad_norm": 0.1668337683466409, "learning_rate": 0.0005788640958280941, "loss": 0.7516, "step": 740 }, { "epoch": 1.0160245482441186, "grad_norm": 0.14142825658305677, "learning_rate": 0.0005729786804205181, "loss": 0.7548, "step": 745 }, { "epoch": 1.0228435049437437, "grad_norm": 0.12528810569521878, "learning_rate": 0.0005670829149965773, "loss": 0.7473, "step": 750 }, { "epoch": 1.0296624616433685, "grad_norm": 0.11736651721866655, "learning_rate": 0.0005611776357082579, "loss": 0.747, "step": 755 }, { "epoch": 1.0364814183429936, "grad_norm": 0.09403020504249461, "learning_rate": 0.0005552636800568266, "loss": 0.7621, "step": 760 }, { "epoch": 1.0433003750426184, "grad_norm": 0.11315198038951943, "learning_rate": 0.0005493418867740529, "loss": 0.745, "step": 765 }, { "epoch": 1.0501193317422435, "grad_norm": 0.1022560757959738, "learning_rate": 0.0005434130957032589, "loss": 0.7459, "step": 770 }, { "epoch": 1.0569382884418683, "grad_norm": 0.10922375995449658, "learning_rate": 0.0005374781476802096, "loss": 0.7466, "step": 775 }, { "epoch": 1.0637572451414934, "grad_norm": 0.5193070680732547, "learning_rate": 0.0005315378844138647, "loss": 0.7627, "step": 780 }, { "epoch": 1.0705762018411182, "grad_norm": 0.12177144687559627, "learning_rate": 0.0005255931483670049, "loss": 0.7546, "step": 785 }, { "epoch": 1.0773951585407433, "grad_norm": 0.7694014802591276, "learning_rate": 0.000519644782636751, "loss": 0.7528, "step": 790 }, { "epoch": 1.0842141152403681, "grad_norm": 0.11357197300498553, "learning_rate": 0.000513693630834995, "loss": 0.7533, "step": 795 }, { "epoch": 1.0910330719399932, "grad_norm": 0.09386251189985013, "learning_rate": 0.0005077405369687564, "loss": 0.7542, "step": 800 }, { "epoch": 1.097852028639618, "grad_norm": 0.09656492984059917, "learning_rate": 0.0005017863453204828, "loss": 0.7453, "step": 805 }, { "epoch": 1.1046709853392431, "grad_norm": 0.09988629196548339, "learning_rate": 0.0004958319003283121, "loss": 0.7567, "step": 810 }, { "epoch": 1.111489942038868, "grad_norm": 0.1340291294865019, "learning_rate": 0.0004898780464663119, "loss": 0.7572, "step": 815 }, { "epoch": 1.118308898738493, "grad_norm": 0.12216818040509048, "learning_rate": 0.00048392562812471485, "loss": 0.7386, "step": 820 }, { "epoch": 1.1251278554381179, "grad_norm": 0.10511324731067179, "learning_rate": 0.0004779754894901638, "loss": 0.7551, "step": 825 }, { "epoch": 1.131946812137743, "grad_norm": 0.08878475250721538, "learning_rate": 0.00047202847442598845, "loss": 0.7476, "step": 830 }, { "epoch": 1.1387657688373678, "grad_norm": 0.10961153167460981, "learning_rate": 0.0004660854263525255, "loss": 0.7501, "step": 835 }, { "epoch": 1.1455847255369929, "grad_norm": 0.11227235770228143, "learning_rate": 0.0004601471881275041, "loss": 0.7468, "step": 840 }, { "epoch": 1.1524036822366177, "grad_norm": 0.14319195588727593, "learning_rate": 0.00045421460192650786, "loss": 0.7459, "step": 845 }, { "epoch": 1.1592226389362428, "grad_norm": 0.12581674820255845, "learning_rate": 0.00044828850912353703, "loss": 0.7496, "step": 850 }, { "epoch": 1.1660415956358676, "grad_norm": 0.10879564209007402, "learning_rate": 0.0004423697501716823, "loss": 0.7377, "step": 855 }, { "epoch": 1.1728605523354927, "grad_norm": 0.10910700864864685, "learning_rate": 0.00043645916448392885, "loss": 0.7569, "step": 860 }, { "epoch": 1.1796795090351175, "grad_norm": 0.12233584779335835, "learning_rate": 0.00043055759031411007, "loss": 0.7451, "step": 865 }, { "epoch": 1.1864984657347426, "grad_norm": 0.09336996907684925, "learning_rate": 0.0004246658646380229, "loss": 0.7486, "step": 870 }, { "epoch": 1.1933174224343674, "grad_norm": 0.09297895821672336, "learning_rate": 0.00041878482303472745, "loss": 0.7522, "step": 875 }, { "epoch": 1.2001363791339925, "grad_norm": 0.09998221831882348, "learning_rate": 0.00041291529956804195, "loss": 0.756, "step": 880 }, { "epoch": 1.2069553358336174, "grad_norm": 0.09246595578930603, "learning_rate": 0.0004070581266682539, "loss": 0.7564, "step": 885 }, { "epoch": 1.2137742925332424, "grad_norm": 0.09888677663554202, "learning_rate": 0.000401214135014063, "loss": 0.7406, "step": 890 }, { "epoch": 1.2205932492328673, "grad_norm": 0.09840879351277865, "learning_rate": 0.0003953841534147725, "loss": 0.7444, "step": 895 }, { "epoch": 1.2274122059324923, "grad_norm": 0.09608186181313223, "learning_rate": 0.000389569008692745, "loss": 0.7434, "step": 900 }, { "epoch": 1.2342311626321174, "grad_norm": 0.12027155555266326, "learning_rate": 0.0003837695255661403, "loss": 0.7505, "step": 905 }, { "epoch": 1.2410501193317423, "grad_norm": 0.09938398385533825, "learning_rate": 0.00037798652653195266, "loss": 0.7524, "step": 910 }, { "epoch": 1.247869076031367, "grad_norm": 0.0963922155626505, "learning_rate": 0.00037222083174936137, "loss": 0.7481, "step": 915 }, { "epoch": 1.2546880327309922, "grad_norm": 0.0959346234671189, "learning_rate": 0.00036647325892341393, "loss": 0.7398, "step": 920 }, { "epoch": 1.2615069894306172, "grad_norm": 0.08820345622198668, "learning_rate": 0.0003607446231890575, "loss": 0.7405, "step": 925 }, { "epoch": 1.268325946130242, "grad_norm": 0.11589956732636802, "learning_rate": 0.0003550357369955347, "loss": 0.7509, "step": 930 }, { "epoch": 1.275144902829867, "grad_norm": 0.10547612256066802, "learning_rate": 0.00034934740999115866, "loss": 0.7502, "step": 935 }, { "epoch": 1.281963859529492, "grad_norm": 0.11823191257134831, "learning_rate": 0.00034368044890848814, "loss": 0.7661, "step": 940 }, { "epoch": 1.288782816229117, "grad_norm": 0.09346920096074791, "learning_rate": 0.0003380356574499141, "loss": 0.7409, "step": 945 }, { "epoch": 1.295601772928742, "grad_norm": 0.09247106405202875, "learning_rate": 0.00033241383617367706, "loss": 0.7448, "step": 950 }, { "epoch": 1.3024207296283667, "grad_norm": 0.10126097744075767, "learning_rate": 0.00032681578238032914, "loss": 0.7455, "step": 955 }, { "epoch": 1.3092396863279918, "grad_norm": 0.10957664402150283, "learning_rate": 0.0003212422899996599, "loss": 0.7442, "step": 960 }, { "epoch": 1.3160586430276169, "grad_norm": 0.08997598908036177, "learning_rate": 0.0003156941494780983, "loss": 0.7399, "step": 965 }, { "epoch": 1.3228775997272417, "grad_norm": 0.09339106504051306, "learning_rate": 0.0003101721476666106, "loss": 0.7473, "step": 970 }, { "epoch": 1.3296965564268666, "grad_norm": 0.09094074237158431, "learning_rate": 0.00030467706770910687, "loss": 0.7349, "step": 975 }, { "epoch": 1.3365155131264916, "grad_norm": 0.08949096519435662, "learning_rate": 0.00029920968893137277, "loss": 0.7335, "step": 980 }, { "epoch": 1.3433344698261167, "grad_norm": 0.10666408169466289, "learning_rate": 0.00029377078673054524, "loss": 0.7431, "step": 985 }, { "epoch": 1.3501534265257416, "grad_norm": 0.09360456030533196, "learning_rate": 0.00028836113246514215, "loss": 0.7504, "step": 990 }, { "epoch": 1.3569723832253664, "grad_norm": 0.0857322076398207, "learning_rate": 0.00028298149334566745, "loss": 0.7461, "step": 995 }, { "epoch": 1.3637913399249915, "grad_norm": 0.08664863370331233, "learning_rate": 0.0002776326323258029, "loss": 0.7347, "step": 1000 }, { "epoch": 1.3706102966246165, "grad_norm": 0.09553160941014031, "learning_rate": 0.0002723153079942047, "loss": 0.7432, "step": 1005 }, { "epoch": 1.3774292533242414, "grad_norm": 0.09757917450665214, "learning_rate": 0.00026703027446691753, "loss": 0.7319, "step": 1010 }, { "epoch": 1.3842482100238662, "grad_norm": 0.13001681090880932, "learning_rate": 0.0002617782812804252, "loss": 0.731, "step": 1015 }, { "epoch": 1.3910671667234913, "grad_norm": 0.09842205689754281, "learning_rate": 0.00025656007328534857, "loss": 0.7377, "step": 1020 }, { "epoch": 1.3978861234231164, "grad_norm": 0.08966471368100484, "learning_rate": 0.00025137639054080975, "loss": 0.7393, "step": 1025 }, { "epoch": 1.4047050801227412, "grad_norm": 0.09544505949561016, "learning_rate": 0.0002462279682094742, "loss": 0.7448, "step": 1030 }, { "epoch": 1.411524036822366, "grad_norm": 0.09721892448809809, "learning_rate": 0.00024111553645328872, "loss": 0.7383, "step": 1035 }, { "epoch": 1.4183429935219911, "grad_norm": 0.09528944859478844, "learning_rate": 0.00023603982032992861, "loss": 0.743, "step": 1040 }, { "epoch": 1.4251619502216162, "grad_norm": 0.09295756990644403, "learning_rate": 0.00023100153968996678, "loss": 0.7354, "step": 1045 }, { "epoch": 1.431980906921241, "grad_norm": 0.09085933894422743, "learning_rate": 0.0002260014090747845, "loss": 0.7363, "step": 1050 }, { "epoch": 1.4387998636208659, "grad_norm": 0.09685032026741523, "learning_rate": 0.00022104013761523156, "loss": 0.7457, "step": 1055 }, { "epoch": 1.445618820320491, "grad_norm": 0.08857516230100924, "learning_rate": 0.00021611842893105726, "loss": 0.742, "step": 1060 }, { "epoch": 1.452437777020116, "grad_norm": 0.08773056464470375, "learning_rate": 0.0002112369810311201, "loss": 0.7332, "step": 1065 }, { "epoch": 1.4592567337197409, "grad_norm": 0.0900763405573559, "learning_rate": 0.00020639648621439488, "loss": 0.7449, "step": 1070 }, { "epoch": 1.466075690419366, "grad_norm": 0.09851993938416971, "learning_rate": 0.00020159763097178952, "loss": 0.7402, "step": 1075 }, { "epoch": 1.4728946471189908, "grad_norm": 0.08759892676999576, "learning_rate": 0.0001968410958887849, "loss": 0.7445, "step": 1080 }, { "epoch": 1.4797136038186158, "grad_norm": 0.0907347079072196, "learning_rate": 0.0001921275555489122, "loss": 0.7399, "step": 1085 }, { "epoch": 1.4865325605182407, "grad_norm": 0.08112607532285956, "learning_rate": 0.00018745767843808209, "loss": 0.7304, "step": 1090 }, { "epoch": 1.4933515172178657, "grad_norm": 0.09161753107862378, "learning_rate": 0.000182832126849779, "loss": 0.7285, "step": 1095 }, { "epoch": 1.5001704739174906, "grad_norm": 0.08358605650538908, "learning_rate": 0.00017825155679113204, "loss": 0.7348, "step": 1100 }, { "epoch": 1.5069894306171157, "grad_norm": 0.08675686822052969, "learning_rate": 0.00017371661788987875, "loss": 0.7409, "step": 1105 }, { "epoch": 1.5138083873167405, "grad_norm": 0.08810226317282835, "learning_rate": 0.0001692279533022339, "loss": 0.7384, "step": 1110 }, { "epoch": 1.5206273440163653, "grad_norm": 0.08950130348184156, "learning_rate": 0.00016478619962167495, "loss": 0.7492, "step": 1115 }, { "epoch": 1.5274463007159904, "grad_norm": 0.08034358957544274, "learning_rate": 0.00016039198678865861, "loss": 0.7332, "step": 1120 }, { "epoch": 1.5342652574156155, "grad_norm": 0.09875564116278145, "learning_rate": 0.00015604593800128157, "loss": 0.7245, "step": 1125 }, { "epoch": 1.5410842141152403, "grad_norm": 0.08654215732134887, "learning_rate": 0.00015174866962689655, "loss": 0.735, "step": 1130 }, { "epoch": 1.5479031708148652, "grad_norm": 0.08592984519220763, "learning_rate": 0.00014750079111469844, "loss": 0.7409, "step": 1135 }, { "epoch": 1.5547221275144902, "grad_norm": 0.09146622488124098, "learning_rate": 0.00014330290490928936, "loss": 0.7307, "step": 1140 }, { "epoch": 1.5615410842141153, "grad_norm": 0.09107009139517197, "learning_rate": 0.00013915560636524005, "loss": 0.7298, "step": 1145 }, { "epoch": 1.5683600409137401, "grad_norm": 0.08290920543587224, "learning_rate": 0.0001350594836626537, "loss": 0.7288, "step": 1150 }, { "epoch": 1.575178997613365, "grad_norm": 0.089017743866575, "learning_rate": 0.00013101511772375002, "loss": 0.721, "step": 1155 }, { "epoch": 1.58199795431299, "grad_norm": 0.08395496088906577, "learning_rate": 0.00012702308213047653, "loss": 0.7369, "step": 1160 }, { "epoch": 1.5888169110126151, "grad_norm": 0.09332825258379267, "learning_rate": 0.00012308394304316224, "loss": 0.7331, "step": 1165 }, { "epoch": 1.59563586771224, "grad_norm": 0.08663551849922511, "learning_rate": 0.00011919825912022336, "loss": 0.7284, "step": 1170 }, { "epoch": 1.6024548244118648, "grad_norm": 0.08467643435872389, "learning_rate": 0.00011536658143893309, "loss": 0.73, "step": 1175 }, { "epoch": 1.6092737811114899, "grad_norm": 0.10792296188935317, "learning_rate": 0.0001115894534172659, "loss": 0.7304, "step": 1180 }, { "epoch": 1.616092737811115, "grad_norm": 0.08665506882217053, "learning_rate": 0.0001078674107368291, "loss": 0.7365, "step": 1185 }, { "epoch": 1.6229116945107398, "grad_norm": 0.0869258591144236, "learning_rate": 0.00010420098126689159, "loss": 0.7271, "step": 1190 }, { "epoch": 1.6297306512103649, "grad_norm": 0.08082381008828005, "learning_rate": 0.00010059068498951912, "loss": 0.7334, "step": 1195 }, { "epoch": 1.63654960790999, "grad_norm": 0.09653014658305435, "learning_rate": 9.70370339258298e-05, "loss": 0.733, "step": 1200 }, { "epoch": 1.6433685646096148, "grad_norm": 0.08209631476731684, "learning_rate": 9.354053206337803e-05, "loss": 0.7295, "step": 1205 }, { "epoch": 1.6501875213092396, "grad_norm": 0.08281543379309102, "learning_rate": 9.010167528467784e-05, "loss": 0.7282, "step": 1210 }, { "epoch": 1.6570064780088647, "grad_norm": 0.09079242657548177, "learning_rate": 8.672095129687485e-05, "loss": 0.7332, "step": 1215 }, { "epoch": 1.6638254347084898, "grad_norm": 0.08732562104521625, "learning_rate": 8.33988395625791e-05, "loss": 0.7196, "step": 1220 }, { "epoch": 1.6706443914081146, "grad_norm": 0.09634799774642026, "learning_rate": 8.013581123186675e-05, "loss": 0.7243, "step": 1225 }, { "epoch": 1.6774633481077394, "grad_norm": 0.0892053596550817, "learning_rate": 7.693232907545955e-05, "loss": 0.7284, "step": 1230 }, { "epoch": 1.6842823048073645, "grad_norm": 0.08080152740232713, "learning_rate": 7.378884741909409e-05, "loss": 0.7226, "step": 1235 }, { "epoch": 1.6911012615069896, "grad_norm": 0.08211682147824328, "learning_rate": 7.070581207908832e-05, "loss": 0.7293, "step": 1240 }, { "epoch": 1.6979202182066144, "grad_norm": 0.08125416074026208, "learning_rate": 6.76836602991146e-05, "loss": 0.7241, "step": 1245 }, { "epoch": 1.7047391749062393, "grad_norm": 0.08118597843483698, "learning_rate": 6.472282068818857e-05, "loss": 0.7176, "step": 1250 }, { "epoch": 1.7115581316058643, "grad_norm": 0.08444459255476687, "learning_rate": 6.182371315988283e-05, "loss": 0.7345, "step": 1255 }, { "epoch": 1.7183770883054894, "grad_norm": 0.08993862852560496, "learning_rate": 5.898674887277394e-05, "loss": 0.7287, "step": 1260 }, { "epoch": 1.7251960450051143, "grad_norm": 0.09172625446804755, "learning_rate": 5.621233017213095e-05, "loss": 0.728, "step": 1265 }, { "epoch": 1.732015001704739, "grad_norm": 0.09728192339349417, "learning_rate": 5.3500850532853477e-05, "loss": 0.7252, "step": 1270 }, { "epoch": 1.7388339584043642, "grad_norm": 0.08183637004716483, "learning_rate": 5.085269450366836e-05, "loss": 0.7249, "step": 1275 }, { "epoch": 1.7456529151039892, "grad_norm": 0.08417398467645139, "learning_rate": 4.8268237652591805e-05, "loss": 0.7263, "step": 1280 }, { "epoch": 1.752471871803614, "grad_norm": 0.08559525343966955, "learning_rate": 4.574784651366581e-05, "loss": 0.7307, "step": 1285 }, { "epoch": 1.759290828503239, "grad_norm": 0.08652462817281027, "learning_rate": 4.329187853497491e-05, "loss": 0.7321, "step": 1290 }, { "epoch": 1.766109785202864, "grad_norm": 0.07780149732014437, "learning_rate": 4.0900682027952274e-05, "loss": 0.7259, "step": 1295 }, { "epoch": 1.772928741902489, "grad_norm": 0.08073420987197251, "learning_rate": 3.8574596117981367e-05, "loss": 0.7318, "step": 1300 }, { "epoch": 1.779747698602114, "grad_norm": 0.09399125884464814, "learning_rate": 3.631395069630039e-05, "loss": 0.7286, "step": 1305 }, { "epoch": 1.7865666553017387, "grad_norm": 0.08683728371693161, "learning_rate": 3.411906637321588e-05, "loss": 0.7281, "step": 1310 }, { "epoch": 1.7933856120013638, "grad_norm": 0.08220865889050938, "learning_rate": 3.199025443263331e-05, "loss": 0.7214, "step": 1315 }, { "epoch": 1.8002045687009889, "grad_norm": 0.08394991114226853, "learning_rate": 2.9927816787910233e-05, "loss": 0.7274, "step": 1320 }, { "epoch": 1.8070235254006137, "grad_norm": 0.08648979166756451, "learning_rate": 2.79320459390377e-05, "loss": 0.728, "step": 1325 }, { "epoch": 1.8138424821002386, "grad_norm": 0.0809919030945408, "learning_rate": 2.600322493115742e-05, "loss": 0.7207, "step": 1330 }, { "epoch": 1.8206614387998636, "grad_norm": 0.08146420956821185, "learning_rate": 2.414162731441971e-05, "loss": 0.7279, "step": 1335 }, { "epoch": 1.8274803954994887, "grad_norm": 0.08991080627514368, "learning_rate": 2.2347517105187952e-05, "loss": 0.7285, "step": 1340 }, { "epoch": 1.8342993521991136, "grad_norm": 0.08113174604843235, "learning_rate": 2.062114874859483e-05, "loss": 0.7212, "step": 1345 }, { "epoch": 1.8411183088987384, "grad_norm": 0.0890991046050432, "learning_rate": 1.8962767082456368e-05, "loss": 0.7238, "step": 1350 }, { "epoch": 1.8479372655983635, "grad_norm": 0.08194468587127682, "learning_rate": 1.7372607302548916e-05, "loss": 0.7278, "step": 1355 }, { "epoch": 1.8547562222979885, "grad_norm": 0.0796060532873072, "learning_rate": 1.58508949292524e-05, "loss": 0.7303, "step": 1360 }, { "epoch": 1.8615751789976134, "grad_norm": 0.08325978355636353, "learning_rate": 1.439784577556702e-05, "loss": 0.7259, "step": 1365 }, { "epoch": 1.8683941356972382, "grad_norm": 0.08101951703165321, "learning_rate": 1.3013665916505824e-05, "loss": 0.7213, "step": 1370 }, { "epoch": 1.8752130923968633, "grad_norm": 0.07474258489762015, "learning_rate": 1.1698551659868716e-05, "loss": 0.7251, "step": 1375 }, { "epoch": 1.8820320490964884, "grad_norm": 0.07988354824659524, "learning_rate": 1.0452689518401615e-05, "loss": 0.7336, "step": 1380 }, { "epoch": 1.8888510057961132, "grad_norm": 0.0869204045893242, "learning_rate": 9.276256183344767e-06, "loss": 0.7225, "step": 1385 }, { "epoch": 1.895669962495738, "grad_norm": 0.08073129640276419, "learning_rate": 8.169418499373749e-06, "loss": 0.719, "step": 1390 }, { "epoch": 1.9024889191953631, "grad_norm": 0.09277363778878074, "learning_rate": 7.132333440937666e-06, "loss": 0.7163, "step": 1395 }, { "epoch": 1.9093078758949882, "grad_norm": 0.08156654600969289, "learning_rate": 6.165148089996075e-06, "loss": 0.7294, "step": 1400 }, { "epoch": 1.916126832594613, "grad_norm": 0.07931208656000137, "learning_rate": 5.267999615159724e-06, "loss": 0.7188, "step": 1405 }, { "epoch": 1.9229457892942379, "grad_norm": 0.08734085147118956, "learning_rate": 4.441015252237113e-06, "loss": 0.7239, "step": 1410 }, { "epoch": 1.929764745993863, "grad_norm": 0.07927625348774799, "learning_rate": 3.684312286189151e-06, "loss": 0.737, "step": 1415 }, { "epoch": 1.936583702693488, "grad_norm": 0.08278514720419748, "learning_rate": 2.997998034495908e-06, "loss": 0.7232, "step": 1420 }, { "epoch": 1.9434026593931129, "grad_norm": 0.09337537685419582, "learning_rate": 2.382169831936565e-06, "loss": 0.7327, "step": 1425 }, { "epoch": 1.9502216160927377, "grad_norm": 0.07802348568191127, "learning_rate": 1.8369150167848459e-06, "loss": 0.7168, "step": 1430 }, { "epoch": 1.9570405727923628, "grad_norm": 0.07582457656438514, "learning_rate": 1.3623109184228711e-06, "loss": 0.7111, "step": 1435 }, { "epoch": 1.9638595294919878, "grad_norm": 0.07999706762970123, "learning_rate": 9.584248463739288e-07, "loss": 0.7251, "step": 1440 }, { "epoch": 1.9706784861916127, "grad_norm": 0.08421307868632791, "learning_rate": 6.253140807562785e-07, "loss": 0.7172, "step": 1445 }, { "epoch": 1.9774974428912375, "grad_norm": 0.08148471965897129, "learning_rate": 3.630258641600381e-07, "loss": 0.7198, "step": 1450 }, { "epoch": 1.9843163995908626, "grad_norm": 0.0748791624620726, "learning_rate": 1.7159739494654325e-07, "loss": 0.7267, "step": 1455 }, { "epoch": 1.9911353562904877, "grad_norm": 0.08613985665976748, "learning_rate": 5.1055821973289286e-08, "loss": 0.7258, "step": 1460 }, { "epoch": 1.9979543129901125, "grad_norm": 0.07749130518414991, "learning_rate": 1.418240743289445e-09, "loss": 0.7335, "step": 1465 }, { "epoch": 1.9993181043300376, "eval_loss": 0.8305559158325195, "eval_runtime": 60.2883, "eval_samples_per_second": 184.43, "eval_steps_per_second": 5.772, "step": 1466 }, { "epoch": 1.9993181043300376, "step": 1466, "total_flos": 143858545459200.0, "train_loss": 0.8359313647060732, "train_runtime": 3324.125, "train_samples_per_second": 56.452, "train_steps_per_second": 0.441 } ], "logging_steps": 5, "max_steps": 1466, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": false, "should_training_stop": false }, "attributes": {} } }, "total_flos": 143858545459200.0, "train_batch_size": 4, "trial_name": null, "trial_params": null }