{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 3653, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.001368738023542294, "grad_norm": 1.1110930744661907, "learning_rate": 2.73224043715847e-07, "loss": 1.71, "step": 5 }, { "epoch": 0.002737476047084588, "grad_norm": 1.0450218593816867, "learning_rate": 5.46448087431694e-07, "loss": 1.7206, "step": 10 }, { "epoch": 0.004106214070626882, "grad_norm": 1.2104539142499893, "learning_rate": 8.196721311475409e-07, "loss": 1.8131, "step": 15 }, { "epoch": 0.005474952094169176, "grad_norm": 1.1014228621000113, "learning_rate": 1.092896174863388e-06, "loss": 1.7571, "step": 20 }, { "epoch": 0.00684369011771147, "grad_norm": 1.0213348170440788, "learning_rate": 1.3661202185792352e-06, "loss": 1.7504, "step": 25 }, { "epoch": 0.008212428141253765, "grad_norm": 0.907763844914787, "learning_rate": 1.6393442622950819e-06, "loss": 1.726, "step": 30 }, { "epoch": 0.009581166164796057, "grad_norm": 0.9051570945453739, "learning_rate": 1.912568306010929e-06, "loss": 1.7158, "step": 35 }, { "epoch": 0.010949904188338352, "grad_norm": 0.7396508509648158, "learning_rate": 2.185792349726776e-06, "loss": 1.7692, "step": 40 }, { "epoch": 0.012318642211880646, "grad_norm": 0.6316750205627214, "learning_rate": 2.459016393442623e-06, "loss": 1.7118, "step": 45 }, { "epoch": 0.01368738023542294, "grad_norm": 0.6071126407750962, "learning_rate": 2.7322404371584705e-06, "loss": 1.6533, "step": 50 }, { "epoch": 0.015056118258965233, "grad_norm": 0.6129917793535419, "learning_rate": 3.0054644808743173e-06, "loss": 1.6471, "step": 55 }, { "epoch": 0.01642485628250753, "grad_norm": 0.6342687394025227, "learning_rate": 3.2786885245901638e-06, "loss": 1.6505, "step": 60 }, { "epoch": 0.017793594306049824, "grad_norm": 0.5732154266113143, "learning_rate": 3.551912568306011e-06, "loss": 1.6261, "step": 65 }, { "epoch": 0.019162332329592115, "grad_norm": 0.5200300967595174, "learning_rate": 3.825136612021858e-06, "loss": 1.5861, "step": 70 }, { "epoch": 0.02053107035313441, "grad_norm": 0.5356030837166187, "learning_rate": 4.098360655737705e-06, "loss": 1.6002, "step": 75 }, { "epoch": 0.021899808376676703, "grad_norm": 0.4893081141758957, "learning_rate": 4.371584699453552e-06, "loss": 1.5388, "step": 80 }, { "epoch": 0.023268546400218998, "grad_norm": 0.5042585034665032, "learning_rate": 4.6448087431694e-06, "loss": 1.5788, "step": 85 }, { "epoch": 0.024637284423761292, "grad_norm": 0.4756627065397568, "learning_rate": 4.918032786885246e-06, "loss": 1.5832, "step": 90 }, { "epoch": 0.026006022447303587, "grad_norm": 0.4712228645477175, "learning_rate": 5.191256830601094e-06, "loss": 1.5221, "step": 95 }, { "epoch": 0.02737476047084588, "grad_norm": 0.47712827241082845, "learning_rate": 5.464480874316941e-06, "loss": 1.5554, "step": 100 }, { "epoch": 0.028743498494388176, "grad_norm": 0.4747245775681728, "learning_rate": 5.737704918032787e-06, "loss": 1.5051, "step": 105 }, { "epoch": 0.030112236517930466, "grad_norm": 0.4619152125132191, "learning_rate": 6.010928961748635e-06, "loss": 1.5731, "step": 110 }, { "epoch": 0.031480974541472764, "grad_norm": 0.46656513563983254, "learning_rate": 6.284153005464482e-06, "loss": 1.5719, "step": 115 }, { "epoch": 0.03284971256501506, "grad_norm": 0.45441304688271805, "learning_rate": 6.5573770491803276e-06, "loss": 1.5359, "step": 120 }, { "epoch": 0.03421845058855735, "grad_norm": 0.4492251581811985, "learning_rate": 6.830601092896175e-06, "loss": 1.5254, "step": 125 }, { "epoch": 0.03558718861209965, "grad_norm": 0.4901256909260675, "learning_rate": 7.103825136612022e-06, "loss": 1.5531, "step": 130 }, { "epoch": 0.036955926635641935, "grad_norm": 0.4521871797566739, "learning_rate": 7.3770491803278695e-06, "loss": 1.5203, "step": 135 }, { "epoch": 0.03832466465918423, "grad_norm": 0.46301007561132373, "learning_rate": 7.650273224043716e-06, "loss": 1.5597, "step": 140 }, { "epoch": 0.039693402682726524, "grad_norm": 0.4443455697452246, "learning_rate": 7.923497267759564e-06, "loss": 1.4751, "step": 145 }, { "epoch": 0.04106214070626882, "grad_norm": 0.4529255518299648, "learning_rate": 8.19672131147541e-06, "loss": 1.5645, "step": 150 }, { "epoch": 0.04243087872981111, "grad_norm": 0.45575204993786755, "learning_rate": 8.469945355191259e-06, "loss": 1.5297, "step": 155 }, { "epoch": 0.04379961675335341, "grad_norm": 0.4619195990067295, "learning_rate": 8.743169398907103e-06, "loss": 1.4945, "step": 160 }, { "epoch": 0.0451683547768957, "grad_norm": 0.43620516601045234, "learning_rate": 9.016393442622952e-06, "loss": 1.5064, "step": 165 }, { "epoch": 0.046537092800437996, "grad_norm": 0.48507057683395305, "learning_rate": 9.2896174863388e-06, "loss": 1.4873, "step": 170 }, { "epoch": 0.04790583082398029, "grad_norm": 0.4591779126577447, "learning_rate": 9.562841530054644e-06, "loss": 1.4989, "step": 175 }, { "epoch": 0.049274568847522585, "grad_norm": 0.4477837953889451, "learning_rate": 9.836065573770493e-06, "loss": 1.4755, "step": 180 }, { "epoch": 0.05064330687106488, "grad_norm": 0.4580627295319435, "learning_rate": 1.0109289617486339e-05, "loss": 1.4545, "step": 185 }, { "epoch": 0.05201204489460717, "grad_norm": 0.45313033315292356, "learning_rate": 1.0382513661202187e-05, "loss": 1.4985, "step": 190 }, { "epoch": 0.05338078291814947, "grad_norm": 0.42708720087605334, "learning_rate": 1.0655737704918034e-05, "loss": 1.4229, "step": 195 }, { "epoch": 0.05474952094169176, "grad_norm": 0.4211311153329928, "learning_rate": 1.0928961748633882e-05, "loss": 1.4669, "step": 200 }, { "epoch": 0.05611825896523406, "grad_norm": 0.4553582543296074, "learning_rate": 1.1202185792349727e-05, "loss": 1.4444, "step": 205 }, { "epoch": 0.05748699698877635, "grad_norm": 0.43888054055275466, "learning_rate": 1.1475409836065575e-05, "loss": 1.4622, "step": 210 }, { "epoch": 0.058855735012318645, "grad_norm": 0.43839040680573754, "learning_rate": 1.1748633879781421e-05, "loss": 1.3921, "step": 215 }, { "epoch": 0.06022447303586093, "grad_norm": 0.442929256824184, "learning_rate": 1.202185792349727e-05, "loss": 1.4695, "step": 220 }, { "epoch": 0.06159321105940323, "grad_norm": 0.42979267539703714, "learning_rate": 1.2295081967213116e-05, "loss": 1.4444, "step": 225 }, { "epoch": 0.06296194908294553, "grad_norm": 0.43942113048062365, "learning_rate": 1.2568306010928964e-05, "loss": 1.4581, "step": 230 }, { "epoch": 0.06433068710648782, "grad_norm": 0.43105155511527676, "learning_rate": 1.284153005464481e-05, "loss": 1.4335, "step": 235 }, { "epoch": 0.06569942513003012, "grad_norm": 0.4347334195115935, "learning_rate": 1.3114754098360655e-05, "loss": 1.5437, "step": 240 }, { "epoch": 0.0670681631535724, "grad_norm": 0.41558733251657665, "learning_rate": 1.3387978142076503e-05, "loss": 1.4246, "step": 245 }, { "epoch": 0.0684369011771147, "grad_norm": 0.4166574660442, "learning_rate": 1.366120218579235e-05, "loss": 1.4672, "step": 250 }, { "epoch": 0.069805639200657, "grad_norm": 0.40343261677226544, "learning_rate": 1.3934426229508198e-05, "loss": 1.4695, "step": 255 }, { "epoch": 0.0711743772241993, "grad_norm": 0.41945935612278784, "learning_rate": 1.4207650273224044e-05, "loss": 1.4173, "step": 260 }, { "epoch": 0.07254311524774158, "grad_norm": 0.4091410585881606, "learning_rate": 1.4480874316939892e-05, "loss": 1.3669, "step": 265 }, { "epoch": 0.07391185327128387, "grad_norm": 0.40428264801464686, "learning_rate": 1.4754098360655739e-05, "loss": 1.4712, "step": 270 }, { "epoch": 0.07528059129482617, "grad_norm": 0.4074017153449847, "learning_rate": 1.5027322404371585e-05, "loss": 1.4477, "step": 275 }, { "epoch": 0.07664932931836846, "grad_norm": 0.41299926058270664, "learning_rate": 1.5300546448087432e-05, "loss": 1.4015, "step": 280 }, { "epoch": 0.07801806734191076, "grad_norm": 0.39622800283246956, "learning_rate": 1.5573770491803278e-05, "loss": 1.443, "step": 285 }, { "epoch": 0.07938680536545305, "grad_norm": 0.4050370810855124, "learning_rate": 1.5846994535519128e-05, "loss": 1.4454, "step": 290 }, { "epoch": 0.08075554338899535, "grad_norm": 0.405591352468174, "learning_rate": 1.6120218579234975e-05, "loss": 1.4077, "step": 295 }, { "epoch": 0.08212428141253764, "grad_norm": 0.4188674346008248, "learning_rate": 1.639344262295082e-05, "loss": 1.4503, "step": 300 }, { "epoch": 0.08349301943607994, "grad_norm": 0.4015518431467613, "learning_rate": 1.6666666666666667e-05, "loss": 1.4027, "step": 305 }, { "epoch": 0.08486175745962223, "grad_norm": 0.40825550305371383, "learning_rate": 1.6939890710382517e-05, "loss": 1.4021, "step": 310 }, { "epoch": 0.08623049548316453, "grad_norm": 0.4103687720629476, "learning_rate": 1.721311475409836e-05, "loss": 1.4304, "step": 315 }, { "epoch": 0.08759923350670681, "grad_norm": 0.4048205442390437, "learning_rate": 1.7486338797814207e-05, "loss": 1.4, "step": 320 }, { "epoch": 0.08896797153024912, "grad_norm": 0.41661892344943224, "learning_rate": 1.7759562841530057e-05, "loss": 1.4035, "step": 325 }, { "epoch": 0.0903367095537914, "grad_norm": 0.4308332237873975, "learning_rate": 1.8032786885245903e-05, "loss": 1.4302, "step": 330 }, { "epoch": 0.0917054475773337, "grad_norm": 0.41012677210314796, "learning_rate": 1.830601092896175e-05, "loss": 1.3969, "step": 335 }, { "epoch": 0.09307418560087599, "grad_norm": 0.40743633096989823, "learning_rate": 1.85792349726776e-05, "loss": 1.4425, "step": 340 }, { "epoch": 0.09444292362441829, "grad_norm": 0.39899819661715447, "learning_rate": 1.8852459016393446e-05, "loss": 1.4303, "step": 345 }, { "epoch": 0.09581166164796058, "grad_norm": 0.41870768132807307, "learning_rate": 1.912568306010929e-05, "loss": 1.4002, "step": 350 }, { "epoch": 0.09718039967150287, "grad_norm": 0.40412869941796903, "learning_rate": 1.939890710382514e-05, "loss": 1.403, "step": 355 }, { "epoch": 0.09854913769504517, "grad_norm": 0.40478167238457397, "learning_rate": 1.9672131147540985e-05, "loss": 1.4227, "step": 360 }, { "epoch": 0.09991787571858746, "grad_norm": 0.41341003482817107, "learning_rate": 1.994535519125683e-05, "loss": 1.4302, "step": 365 }, { "epoch": 0.10128661374212976, "grad_norm": 0.41658269655258945, "learning_rate": 1.999992692147127e-05, "loss": 1.445, "step": 370 }, { "epoch": 0.10265535176567205, "grad_norm": 0.40746899077621546, "learning_rate": 1.999963004177886e-05, "loss": 1.458, "step": 375 }, { "epoch": 0.10402408978921435, "grad_norm": 0.3960893049170255, "learning_rate": 1.9999104800289367e-05, "loss": 1.4382, "step": 380 }, { "epoch": 0.10539282781275663, "grad_norm": 0.3950983233516754, "learning_rate": 1.9998351208997734e-05, "loss": 1.4365, "step": 385 }, { "epoch": 0.10676156583629894, "grad_norm": 0.38845890924227394, "learning_rate": 1.9997369285113754e-05, "loss": 1.3731, "step": 390 }, { "epoch": 0.10813030385984122, "grad_norm": 0.39526066204949717, "learning_rate": 1.9996159051061638e-05, "loss": 1.4111, "step": 395 }, { "epoch": 0.10949904188338352, "grad_norm": 0.4133763360642962, "learning_rate": 1.9994720534479543e-05, "loss": 1.4294, "step": 400 }, { "epoch": 0.11086777990692581, "grad_norm": 0.39671784657884285, "learning_rate": 1.999305376821889e-05, "loss": 1.3708, "step": 405 }, { "epoch": 0.11223651793046811, "grad_norm": 0.3847825062400942, "learning_rate": 1.999115879034368e-05, "loss": 1.4638, "step": 410 }, { "epoch": 0.1136052559540104, "grad_norm": 0.41459555705408424, "learning_rate": 1.9989035644129553e-05, "loss": 1.4105, "step": 415 }, { "epoch": 0.1149739939775527, "grad_norm": 0.4270268157501032, "learning_rate": 1.998668437806286e-05, "loss": 1.4243, "step": 420 }, { "epoch": 0.11634273200109499, "grad_norm": 0.42083549100480927, "learning_rate": 1.998410504583952e-05, "loss": 1.4285, "step": 425 }, { "epoch": 0.11771147002463729, "grad_norm": 0.3929543512101535, "learning_rate": 1.998129770636381e-05, "loss": 1.3692, "step": 430 }, { "epoch": 0.11908020804817958, "grad_norm": 0.40163841769109054, "learning_rate": 1.9978262423747003e-05, "loss": 1.3842, "step": 435 }, { "epoch": 0.12044894607172187, "grad_norm": 0.40750648380479343, "learning_rate": 1.997499926730593e-05, "loss": 1.4412, "step": 440 }, { "epoch": 0.12181768409526417, "grad_norm": 0.4029670313800167, "learning_rate": 1.9971508311561373e-05, "loss": 1.4481, "step": 445 }, { "epoch": 0.12318642211880645, "grad_norm": 0.3934576991168861, "learning_rate": 1.996778963623637e-05, "loss": 1.3706, "step": 450 }, { "epoch": 0.12455516014234876, "grad_norm": 0.4126991514447529, "learning_rate": 1.9963843326254406e-05, "loss": 1.4096, "step": 455 }, { "epoch": 0.12592389816589106, "grad_norm": 0.400896785065692, "learning_rate": 1.9959669471737456e-05, "loss": 1.3284, "step": 460 }, { "epoch": 0.12729263618943334, "grad_norm": 0.4084884002554967, "learning_rate": 1.9955268168003938e-05, "loss": 1.4346, "step": 465 }, { "epoch": 0.12866137421297563, "grad_norm": 0.4031311283565323, "learning_rate": 1.9950639515566537e-05, "loss": 1.3632, "step": 470 }, { "epoch": 0.13003011223651792, "grad_norm": 0.4017916786924823, "learning_rate": 1.99457836201299e-05, "loss": 1.4083, "step": 475 }, { "epoch": 0.13139885026006023, "grad_norm": 0.38801834482390074, "learning_rate": 1.9940700592588228e-05, "loss": 1.3823, "step": 480 }, { "epoch": 0.13276758828360252, "grad_norm": 0.40725954819087723, "learning_rate": 1.993539054902275e-05, "loss": 1.369, "step": 485 }, { "epoch": 0.1341363263071448, "grad_norm": 0.42191731863948084, "learning_rate": 1.992985361069906e-05, "loss": 1.3907, "step": 490 }, { "epoch": 0.1355050643306871, "grad_norm": 0.3981980013765845, "learning_rate": 1.9924089904064354e-05, "loss": 1.4253, "step": 495 }, { "epoch": 0.1368738023542294, "grad_norm": 0.3968358847514907, "learning_rate": 1.9918099560744545e-05, "loss": 1.4478, "step": 500 }, { "epoch": 0.1382425403777717, "grad_norm": 0.4020407861005804, "learning_rate": 1.991188271754125e-05, "loss": 1.4063, "step": 505 }, { "epoch": 0.139611278401314, "grad_norm": 0.3983008409093381, "learning_rate": 1.990543951642866e-05, "loss": 1.3838, "step": 510 }, { "epoch": 0.14098001642485627, "grad_norm": 0.38772247745474425, "learning_rate": 1.9898770104550335e-05, "loss": 1.4193, "step": 515 }, { "epoch": 0.1423487544483986, "grad_norm": 0.4172192092889838, "learning_rate": 1.9891874634215784e-05, "loss": 1.3831, "step": 520 }, { "epoch": 0.14371749247194088, "grad_norm": 0.39316534732966774, "learning_rate": 1.9884753262897042e-05, "loss": 1.3813, "step": 525 }, { "epoch": 0.14508623049548316, "grad_norm": 0.40825756476362335, "learning_rate": 1.9877406153225028e-05, "loss": 1.4233, "step": 530 }, { "epoch": 0.14645496851902545, "grad_norm": 0.4303738354233706, "learning_rate": 1.9869833472985882e-05, "loss": 1.4267, "step": 535 }, { "epoch": 0.14782370654256774, "grad_norm": 0.39536263390575654, "learning_rate": 1.9862035395117075e-05, "loss": 1.3688, "step": 540 }, { "epoch": 0.14919244456611006, "grad_norm": 0.39303998157704906, "learning_rate": 1.9854012097703515e-05, "loss": 1.4259, "step": 545 }, { "epoch": 0.15056118258965234, "grad_norm": 0.3939411519174778, "learning_rate": 1.9845763763973433e-05, "loss": 1.354, "step": 550 }, { "epoch": 0.15192992061319463, "grad_norm": 0.39277763892363776, "learning_rate": 1.9837290582294233e-05, "loss": 1.3468, "step": 555 }, { "epoch": 0.15329865863673692, "grad_norm": 0.39121990838351217, "learning_rate": 1.9828592746168172e-05, "loss": 1.3508, "step": 560 }, { "epoch": 0.15466739666027923, "grad_norm": 0.4035101254769364, "learning_rate": 1.981967045422795e-05, "loss": 1.369, "step": 565 }, { "epoch": 0.15603613468382152, "grad_norm": 0.38407623227911974, "learning_rate": 1.9810523910232165e-05, "loss": 1.4039, "step": 570 }, { "epoch": 0.1574048727073638, "grad_norm": 0.3915340445821982, "learning_rate": 1.9801153323060667e-05, "loss": 1.3634, "step": 575 }, { "epoch": 0.1587736107309061, "grad_norm": 0.39241210575593644, "learning_rate": 1.9791558906709787e-05, "loss": 1.4257, "step": 580 }, { "epoch": 0.1601423487544484, "grad_norm": 0.3946963973614525, "learning_rate": 1.9781740880287444e-05, "loss": 1.3136, "step": 585 }, { "epoch": 0.1615110867779907, "grad_norm": 0.39148042578465725, "learning_rate": 1.9771699468008156e-05, "loss": 1.3654, "step": 590 }, { "epoch": 0.16287982480153299, "grad_norm": 0.3895106088217061, "learning_rate": 1.9761434899187893e-05, "loss": 1.3762, "step": 595 }, { "epoch": 0.16424856282507527, "grad_norm": 0.3997937267692177, "learning_rate": 1.9750947408238872e-05, "loss": 1.4064, "step": 600 }, { "epoch": 0.1656173008486176, "grad_norm": 0.4002294154735745, "learning_rate": 1.974023723466418e-05, "loss": 1.3771, "step": 605 }, { "epoch": 0.16698603887215988, "grad_norm": 0.40550620529076886, "learning_rate": 1.9729304623052315e-05, "loss": 1.4551, "step": 610 }, { "epoch": 0.16835477689570216, "grad_norm": 0.40410760524836825, "learning_rate": 1.9718149823071592e-05, "loss": 1.3394, "step": 615 }, { "epoch": 0.16972351491924445, "grad_norm": 0.40142195594131513, "learning_rate": 1.970677308946446e-05, "loss": 1.3872, "step": 620 }, { "epoch": 0.17109225294278674, "grad_norm": 0.4041136169142662, "learning_rate": 1.9695174682041652e-05, "loss": 1.3644, "step": 625 }, { "epoch": 0.17246099096632905, "grad_norm": 0.39436310577761385, "learning_rate": 1.9683354865676298e-05, "loss": 1.434, "step": 630 }, { "epoch": 0.17382972898987134, "grad_norm": 0.400517293758074, "learning_rate": 1.9671313910297826e-05, "loss": 1.3941, "step": 635 }, { "epoch": 0.17519846701341363, "grad_norm": 0.4059408812570113, "learning_rate": 1.9659052090885834e-05, "loss": 1.3907, "step": 640 }, { "epoch": 0.17656720503695592, "grad_norm": 0.3929001302562383, "learning_rate": 1.9646569687463796e-05, "loss": 1.351, "step": 645 }, { "epoch": 0.17793594306049823, "grad_norm": 0.4090718300758474, "learning_rate": 1.9633866985092655e-05, "loss": 1.4102, "step": 650 }, { "epoch": 0.17930468108404052, "grad_norm": 0.40995200812054255, "learning_rate": 1.9620944273864343e-05, "loss": 1.3575, "step": 655 }, { "epoch": 0.1806734191075828, "grad_norm": 0.40276642674265933, "learning_rate": 1.960780184889514e-05, "loss": 1.3941, "step": 660 }, { "epoch": 0.1820421571311251, "grad_norm": 0.400136397383414, "learning_rate": 1.9594440010318924e-05, "loss": 1.4171, "step": 665 }, { "epoch": 0.1834108951546674, "grad_norm": 0.40872717576906, "learning_rate": 1.9580859063280326e-05, "loss": 1.391, "step": 670 }, { "epoch": 0.1847796331782097, "grad_norm": 0.39405384932557985, "learning_rate": 1.956705931792777e-05, "loss": 1.4042, "step": 675 }, { "epoch": 0.18614837120175198, "grad_norm": 0.396632571946212, "learning_rate": 1.9553041089406387e-05, "loss": 1.3598, "step": 680 }, { "epoch": 0.18751710922529427, "grad_norm": 0.39535225715229844, "learning_rate": 1.95388046978508e-05, "loss": 1.3652, "step": 685 }, { "epoch": 0.18888584724883659, "grad_norm": 0.39747240968211434, "learning_rate": 1.9524350468377828e-05, "loss": 1.3857, "step": 690 }, { "epoch": 0.19025458527237887, "grad_norm": 0.4045442697980633, "learning_rate": 1.9509678731079074e-05, "loss": 1.3724, "step": 695 }, { "epoch": 0.19162332329592116, "grad_norm": 0.39648112930634666, "learning_rate": 1.949478982101336e-05, "loss": 1.3642, "step": 700 }, { "epoch": 0.19299206131946345, "grad_norm": 0.3966559822195587, "learning_rate": 1.947968407819909e-05, "loss": 1.3704, "step": 705 }, { "epoch": 0.19436079934300574, "grad_norm": 0.3803799928500003, "learning_rate": 1.9464361847606486e-05, "loss": 1.3718, "step": 710 }, { "epoch": 0.19572953736654805, "grad_norm": 0.3920580621341347, "learning_rate": 1.9448823479149705e-05, "loss": 1.3994, "step": 715 }, { "epoch": 0.19709827539009034, "grad_norm": 0.38159551727148144, "learning_rate": 1.9433069327678847e-05, "loss": 1.3539, "step": 720 }, { "epoch": 0.19846701341363263, "grad_norm": 0.3979848267470623, "learning_rate": 1.9417099752971858e-05, "loss": 1.3824, "step": 725 }, { "epoch": 0.1998357514371749, "grad_norm": 0.36228811525683013, "learning_rate": 1.9400915119726305e-05, "loss": 1.2942, "step": 730 }, { "epoch": 0.20120448946071723, "grad_norm": 0.40336839547536973, "learning_rate": 1.938451579755106e-05, "loss": 1.3305, "step": 735 }, { "epoch": 0.20257322748425952, "grad_norm": 0.4093706132368206, "learning_rate": 1.9367902160957843e-05, "loss": 1.4047, "step": 740 }, { "epoch": 0.2039419655078018, "grad_norm": 0.4012678664539993, "learning_rate": 1.9351074589352684e-05, "loss": 1.34, "step": 745 }, { "epoch": 0.2053107035313441, "grad_norm": 0.3982341789573293, "learning_rate": 1.933403346702725e-05, "loss": 1.3518, "step": 750 }, { "epoch": 0.2066794415548864, "grad_norm": 0.3895851509088621, "learning_rate": 1.931677918315007e-05, "loss": 1.3698, "step": 755 }, { "epoch": 0.2080481795784287, "grad_norm": 0.40873783451123424, "learning_rate": 1.9299312131757645e-05, "loss": 1.3768, "step": 760 }, { "epoch": 0.20941691760197098, "grad_norm": 0.38863227885235524, "learning_rate": 1.928163271174546e-05, "loss": 1.368, "step": 765 }, { "epoch": 0.21078565562551327, "grad_norm": 0.40218624042286666, "learning_rate": 1.9263741326858866e-05, "loss": 1.4002, "step": 770 }, { "epoch": 0.21215439364905558, "grad_norm": 0.38750285278794777, "learning_rate": 1.9245638385683857e-05, "loss": 1.3808, "step": 775 }, { "epoch": 0.21352313167259787, "grad_norm": 0.39332958062090156, "learning_rate": 1.9227324301637747e-05, "loss": 1.3991, "step": 780 }, { "epoch": 0.21489186969614016, "grad_norm": 0.39759945858946194, "learning_rate": 1.9208799492959723e-05, "loss": 1.3765, "step": 785 }, { "epoch": 0.21626060771968245, "grad_norm": 0.3779639266645249, "learning_rate": 1.9190064382701296e-05, "loss": 1.378, "step": 790 }, { "epoch": 0.21762934574322473, "grad_norm": 0.3980685187213535, "learning_rate": 1.917111939871664e-05, "loss": 1.3559, "step": 795 }, { "epoch": 0.21899808376676705, "grad_norm": 0.3955990129038042, "learning_rate": 1.915196497365282e-05, "loss": 1.3688, "step": 800 }, { "epoch": 0.22036682179030934, "grad_norm": 0.3814772041353902, "learning_rate": 1.9132601544939914e-05, "loss": 1.3106, "step": 805 }, { "epoch": 0.22173555981385162, "grad_norm": 0.4033095650227253, "learning_rate": 1.9113029554781014e-05, "loss": 1.3389, "step": 810 }, { "epoch": 0.2231042978373939, "grad_norm": 0.3934860150762494, "learning_rate": 1.9093249450142144e-05, "loss": 1.3343, "step": 815 }, { "epoch": 0.22447303586093623, "grad_norm": 0.4008166941226203, "learning_rate": 1.907326168274204e-05, "loss": 1.3627, "step": 820 }, { "epoch": 0.2258417738844785, "grad_norm": 0.38863913805386574, "learning_rate": 1.905306670904184e-05, "loss": 1.3978, "step": 825 }, { "epoch": 0.2272105119080208, "grad_norm": 0.39887884072448215, "learning_rate": 1.9032664990234648e-05, "loss": 1.3548, "step": 830 }, { "epoch": 0.2285792499315631, "grad_norm": 0.3942528539065832, "learning_rate": 1.9012056992235025e-05, "loss": 1.3589, "step": 835 }, { "epoch": 0.2299479879551054, "grad_norm": 0.4003093485695292, "learning_rate": 1.899124318566832e-05, "loss": 1.4008, "step": 840 }, { "epoch": 0.2313167259786477, "grad_norm": 0.38851656838431264, "learning_rate": 1.897022404585996e-05, "loss": 1.3396, "step": 845 }, { "epoch": 0.23268546400218998, "grad_norm": 0.39282411404685663, "learning_rate": 1.894900005282454e-05, "loss": 1.3118, "step": 850 }, { "epoch": 0.23405420202573227, "grad_norm": 0.3995600743628729, "learning_rate": 1.892757169125492e-05, "loss": 1.3261, "step": 855 }, { "epoch": 0.23542294004927458, "grad_norm": 0.40585943733175783, "learning_rate": 1.8905939450511117e-05, "loss": 1.2896, "step": 860 }, { "epoch": 0.23679167807281687, "grad_norm": 0.38676778157652897, "learning_rate": 1.888410382460915e-05, "loss": 1.3769, "step": 865 }, { "epoch": 0.23816041609635916, "grad_norm": 0.4006135132967053, "learning_rate": 1.8862065312209735e-05, "loss": 1.3744, "step": 870 }, { "epoch": 0.23952915411990144, "grad_norm": 0.38183538010624907, "learning_rate": 1.8839824416606932e-05, "loss": 1.3994, "step": 875 }, { "epoch": 0.24089789214344373, "grad_norm": 0.40012172515979694, "learning_rate": 1.8817381645716613e-05, "loss": 1.3736, "step": 880 }, { "epoch": 0.24226663016698605, "grad_norm": 0.4036320450863606, "learning_rate": 1.879473751206489e-05, "loss": 1.3679, "step": 885 }, { "epoch": 0.24363536819052833, "grad_norm": 0.39250548017399456, "learning_rate": 1.8771892532776406e-05, "loss": 1.352, "step": 890 }, { "epoch": 0.24500410621407062, "grad_norm": 0.384618237208868, "learning_rate": 1.8748847229562504e-05, "loss": 1.3415, "step": 895 }, { "epoch": 0.2463728442376129, "grad_norm": 0.3847231065307316, "learning_rate": 1.8725602128709348e-05, "loss": 1.436, "step": 900 }, { "epoch": 0.24774158226115522, "grad_norm": 0.39329176880661687, "learning_rate": 1.8702157761065877e-05, "loss": 1.344, "step": 905 }, { "epoch": 0.2491103202846975, "grad_norm": 0.4202514383981607, "learning_rate": 1.8678514662031688e-05, "loss": 1.3384, "step": 910 }, { "epoch": 0.2504790583082398, "grad_norm": 0.3866724615350219, "learning_rate": 1.8654673371544815e-05, "loss": 1.3374, "step": 915 }, { "epoch": 0.2518477963317821, "grad_norm": 0.4109333525553728, "learning_rate": 1.8630634434069397e-05, "loss": 1.3949, "step": 920 }, { "epoch": 0.2532165343553244, "grad_norm": 0.3782421450027903, "learning_rate": 1.860639839858324e-05, "loss": 1.3162, "step": 925 }, { "epoch": 0.2545852723788667, "grad_norm": 0.3909532902784873, "learning_rate": 1.8581965818565278e-05, "loss": 1.3829, "step": 930 }, { "epoch": 0.255954010402409, "grad_norm": 0.4051170609748319, "learning_rate": 1.855733725198295e-05, "loss": 1.3462, "step": 935 }, { "epoch": 0.25732274842595126, "grad_norm": 0.39349552836368845, "learning_rate": 1.8532513261279433e-05, "loss": 1.4015, "step": 940 }, { "epoch": 0.2586914864494936, "grad_norm": 0.39347934180028793, "learning_rate": 1.8507494413360808e-05, "loss": 1.3367, "step": 945 }, { "epoch": 0.26006022447303584, "grad_norm": 0.4158855876544813, "learning_rate": 1.848228127958312e-05, "loss": 1.353, "step": 950 }, { "epoch": 0.26142896249657815, "grad_norm": 0.3793682443730692, "learning_rate": 1.8456874435739337e-05, "loss": 1.3398, "step": 955 }, { "epoch": 0.26279770052012047, "grad_norm": 0.38047885709566037, "learning_rate": 1.843127446204616e-05, "loss": 1.329, "step": 960 }, { "epoch": 0.26416643854366273, "grad_norm": 0.3991275596683987, "learning_rate": 1.8405481943130827e-05, "loss": 1.3115, "step": 965 }, { "epoch": 0.26553517656720504, "grad_norm": 0.389649193690539, "learning_rate": 1.8379497468017726e-05, "loss": 1.3746, "step": 970 }, { "epoch": 0.2669039145907473, "grad_norm": 0.40470694231468296, "learning_rate": 1.8353321630114952e-05, "loss": 1.3354, "step": 975 }, { "epoch": 0.2682726526142896, "grad_norm": 0.40734933283960667, "learning_rate": 1.832695502720076e-05, "loss": 1.3295, "step": 980 }, { "epoch": 0.26964139063783193, "grad_norm": 0.39323434740751423, "learning_rate": 1.8300398261409912e-05, "loss": 1.3069, "step": 985 }, { "epoch": 0.2710101286613742, "grad_norm": 0.3837841492160018, "learning_rate": 1.8273651939219914e-05, "loss": 1.3543, "step": 990 }, { "epoch": 0.2723788666849165, "grad_norm": 0.38795485386053835, "learning_rate": 1.8246716671437186e-05, "loss": 1.3798, "step": 995 }, { "epoch": 0.2737476047084588, "grad_norm": 0.38509661706339016, "learning_rate": 1.8219593073183106e-05, "loss": 1.3604, "step": 1000 }, { "epoch": 0.2751163427320011, "grad_norm": 0.3882331224756399, "learning_rate": 1.8192281763879946e-05, "loss": 1.3417, "step": 1005 }, { "epoch": 0.2764850807555434, "grad_norm": 0.3842715356979672, "learning_rate": 1.816478336723675e-05, "loss": 1.4061, "step": 1010 }, { "epoch": 0.27785381877908566, "grad_norm": 0.4013303335538485, "learning_rate": 1.8137098511235084e-05, "loss": 1.3509, "step": 1015 }, { "epoch": 0.279222556802628, "grad_norm": 0.389250409772092, "learning_rate": 1.810922782811468e-05, "loss": 1.3371, "step": 1020 }, { "epoch": 0.2805912948261703, "grad_norm": 0.3891214912082097, "learning_rate": 1.808117195435901e-05, "loss": 1.3756, "step": 1025 }, { "epoch": 0.28196003284971255, "grad_norm": 0.39295322317665116, "learning_rate": 1.805293153068076e-05, "loss": 1.3525, "step": 1030 }, { "epoch": 0.28332877087325486, "grad_norm": 0.38744159952014173, "learning_rate": 1.802450720200718e-05, "loss": 1.3508, "step": 1035 }, { "epoch": 0.2846975088967972, "grad_norm": 0.3971694811346686, "learning_rate": 1.7995899617465357e-05, "loss": 1.3242, "step": 1040 }, { "epoch": 0.28606624692033944, "grad_norm": 0.3892979964335795, "learning_rate": 1.7967109430367406e-05, "loss": 1.2919, "step": 1045 }, { "epoch": 0.28743498494388176, "grad_norm": 0.40651116778504365, "learning_rate": 1.793813729819553e-05, "loss": 1.4047, "step": 1050 }, { "epoch": 0.288803722967424, "grad_norm": 0.39365601735181455, "learning_rate": 1.7908983882587038e-05, "loss": 1.3622, "step": 1055 }, { "epoch": 0.29017246099096633, "grad_norm": 0.38324138023204, "learning_rate": 1.787964984931919e-05, "loss": 1.3773, "step": 1060 }, { "epoch": 0.29154119901450865, "grad_norm": 0.3887104304366927, "learning_rate": 1.7850135868294023e-05, "loss": 1.3973, "step": 1065 }, { "epoch": 0.2929099370380509, "grad_norm": 0.4087532681641359, "learning_rate": 1.782044261352305e-05, "loss": 1.3246, "step": 1070 }, { "epoch": 0.2942786750615932, "grad_norm": 0.39381567774906023, "learning_rate": 1.7790570763111864e-05, "loss": 1.3683, "step": 1075 }, { "epoch": 0.2956474130851355, "grad_norm": 0.38720936830053604, "learning_rate": 1.7760520999244638e-05, "loss": 1.375, "step": 1080 }, { "epoch": 0.2970161511086778, "grad_norm": 0.38699812271583683, "learning_rate": 1.7730294008168578e-05, "loss": 1.403, "step": 1085 }, { "epoch": 0.2983848891322201, "grad_norm": 0.3938155567988771, "learning_rate": 1.7699890480178216e-05, "loss": 1.3567, "step": 1090 }, { "epoch": 0.29975362715576237, "grad_norm": 0.4051470823072413, "learning_rate": 1.766931110959967e-05, "loss": 1.4228, "step": 1095 }, { "epoch": 0.3011223651793047, "grad_norm": 0.4013404711196811, "learning_rate": 1.763855659477478e-05, "loss": 1.3689, "step": 1100 }, { "epoch": 0.302491103202847, "grad_norm": 0.4239789885244797, "learning_rate": 1.7607627638045156e-05, "loss": 1.3988, "step": 1105 }, { "epoch": 0.30385984122638926, "grad_norm": 0.4005302271519552, "learning_rate": 1.7576524945736137e-05, "loss": 1.3368, "step": 1110 }, { "epoch": 0.3052285792499316, "grad_norm": 0.3900827670562639, "learning_rate": 1.754524922814068e-05, "loss": 1.3633, "step": 1115 }, { "epoch": 0.30659731727347384, "grad_norm": 0.39004513777273, "learning_rate": 1.751380119950311e-05, "loss": 1.4024, "step": 1120 }, { "epoch": 0.30796605529701615, "grad_norm": 0.3955440868581025, "learning_rate": 1.7482181578002837e-05, "loss": 1.3667, "step": 1125 }, { "epoch": 0.30933479332055847, "grad_norm": 0.41261816602720663, "learning_rate": 1.745039108573793e-05, "loss": 1.357, "step": 1130 }, { "epoch": 0.3107035313441007, "grad_norm": 0.37732770567847945, "learning_rate": 1.7418430448708644e-05, "loss": 1.3337, "step": 1135 }, { "epoch": 0.31207226936764304, "grad_norm": 0.3764972884033891, "learning_rate": 1.738630039680083e-05, "loss": 1.3556, "step": 1140 }, { "epoch": 0.3134410073911853, "grad_norm": 0.40764805847336166, "learning_rate": 1.7354001663769278e-05, "loss": 1.3679, "step": 1145 }, { "epoch": 0.3148097454147276, "grad_norm": 0.3794101029629702, "learning_rate": 1.7321534987220942e-05, "loss": 1.3176, "step": 1150 }, { "epoch": 0.31617848343826993, "grad_norm": 0.40248164169870704, "learning_rate": 1.728890110859812e-05, "loss": 1.3458, "step": 1155 }, { "epoch": 0.3175472214618122, "grad_norm": 0.39012890957455026, "learning_rate": 1.7256100773161492e-05, "loss": 1.296, "step": 1160 }, { "epoch": 0.3189159594853545, "grad_norm": 0.39856381680118136, "learning_rate": 1.7223134729973134e-05, "loss": 1.3614, "step": 1165 }, { "epoch": 0.3202846975088968, "grad_norm": 0.40493526767721577, "learning_rate": 1.7190003731879375e-05, "loss": 1.3533, "step": 1170 }, { "epoch": 0.3216534355324391, "grad_norm": 0.38868285184435786, "learning_rate": 1.715670853549364e-05, "loss": 1.3377, "step": 1175 }, { "epoch": 0.3230221735559814, "grad_norm": 0.3840977803535061, "learning_rate": 1.7123249901179142e-05, "loss": 1.3753, "step": 1180 }, { "epoch": 0.32439091157952366, "grad_norm": 0.37510550805682613, "learning_rate": 1.708962859303154e-05, "loss": 1.3557, "step": 1185 }, { "epoch": 0.32575964960306597, "grad_norm": 0.39070855119805253, "learning_rate": 1.7055845378861476e-05, "loss": 1.3584, "step": 1190 }, { "epoch": 0.3271283876266083, "grad_norm": 0.39021160718672115, "learning_rate": 1.7021901030177036e-05, "loss": 1.3399, "step": 1195 }, { "epoch": 0.32849712565015055, "grad_norm": 0.3983342506003189, "learning_rate": 1.698779632216615e-05, "loss": 1.3965, "step": 1200 }, { "epoch": 0.32986586367369286, "grad_norm": 0.38690562360273995, "learning_rate": 1.6953532033678874e-05, "loss": 1.4221, "step": 1205 }, { "epoch": 0.3312346016972352, "grad_norm": 0.39029540911089444, "learning_rate": 1.69191089472096e-05, "loss": 1.3201, "step": 1210 }, { "epoch": 0.33260333972077744, "grad_norm": 0.40512131600219015, "learning_rate": 1.688452784887921e-05, "loss": 1.3495, "step": 1215 }, { "epoch": 0.33397207774431975, "grad_norm": 0.3804586676740124, "learning_rate": 1.684978952841709e-05, "loss": 1.3471, "step": 1220 }, { "epoch": 0.335340815767862, "grad_norm": 0.38428741373037245, "learning_rate": 1.681489477914312e-05, "loss": 1.3196, "step": 1225 }, { "epoch": 0.3367095537914043, "grad_norm": 0.384047691017285, "learning_rate": 1.677984439794954e-05, "loss": 1.3303, "step": 1230 }, { "epoch": 0.33807829181494664, "grad_norm": 0.3839900646299871, "learning_rate": 1.6744639185282784e-05, "loss": 1.2792, "step": 1235 }, { "epoch": 0.3394470298384889, "grad_norm": 0.3884042208952324, "learning_rate": 1.670927994512514e-05, "loss": 1.3223, "step": 1240 }, { "epoch": 0.3408157678620312, "grad_norm": 0.3911385295369844, "learning_rate": 1.667376748497646e-05, "loss": 1.3546, "step": 1245 }, { "epoch": 0.3421845058855735, "grad_norm": 0.3807410219808285, "learning_rate": 1.6638102615835658e-05, "loss": 1.3148, "step": 1250 }, { "epoch": 0.3435532439091158, "grad_norm": 0.39423887853753, "learning_rate": 1.6602286152182236e-05, "loss": 1.361, "step": 1255 }, { "epoch": 0.3449219819326581, "grad_norm": 0.39187255483162625, "learning_rate": 1.6566318911957647e-05, "loss": 1.339, "step": 1260 }, { "epoch": 0.34629071995620037, "grad_norm": 0.38859511955741677, "learning_rate": 1.6530201716546647e-05, "loss": 1.3556, "step": 1265 }, { "epoch": 0.3476594579797427, "grad_norm": 0.3993643770992966, "learning_rate": 1.649393539075851e-05, "loss": 1.3476, "step": 1270 }, { "epoch": 0.349028196003285, "grad_norm": 0.39035176617866785, "learning_rate": 1.6457520762808217e-05, "loss": 1.3228, "step": 1275 }, { "epoch": 0.35039693402682726, "grad_norm": 0.3830707286890397, "learning_rate": 1.6420958664297514e-05, "loss": 1.3094, "step": 1280 }, { "epoch": 0.35176567205036957, "grad_norm": 0.36321444690348953, "learning_rate": 1.638424993019595e-05, "loss": 1.3853, "step": 1285 }, { "epoch": 0.35313441007391183, "grad_norm": 0.371835147945527, "learning_rate": 1.634739539882178e-05, "loss": 1.3477, "step": 1290 }, { "epoch": 0.35450314809745415, "grad_norm": 0.3906744323143432, "learning_rate": 1.6310395911822848e-05, "loss": 1.3149, "step": 1295 }, { "epoch": 0.35587188612099646, "grad_norm": 0.37408867437468385, "learning_rate": 1.6273252314157352e-05, "loss": 1.3135, "step": 1300 }, { "epoch": 0.3572406241445387, "grad_norm": 0.37403440281560435, "learning_rate": 1.6235965454074535e-05, "loss": 1.2301, "step": 1305 }, { "epoch": 0.35860936216808104, "grad_norm": 0.3867708656040568, "learning_rate": 1.619853618309535e-05, "loss": 1.3716, "step": 1310 }, { "epoch": 0.3599781001916233, "grad_norm": 0.3885328796144267, "learning_rate": 1.6160965355992966e-05, "loss": 1.366, "step": 1315 }, { "epoch": 0.3613468382151656, "grad_norm": 0.403719948973195, "learning_rate": 1.6123253830773293e-05, "loss": 1.3661, "step": 1320 }, { "epoch": 0.3627155762387079, "grad_norm": 0.3827196192323514, "learning_rate": 1.6085402468655356e-05, "loss": 1.3567, "step": 1325 }, { "epoch": 0.3640843142622502, "grad_norm": 0.3968500270407259, "learning_rate": 1.6047412134051645e-05, "loss": 1.4044, "step": 1330 }, { "epoch": 0.3654530522857925, "grad_norm": 0.38847711656237954, "learning_rate": 1.6009283694548365e-05, "loss": 1.3591, "step": 1335 }, { "epoch": 0.3668217903093348, "grad_norm": 0.4036707276972636, "learning_rate": 1.5971018020885623e-05, "loss": 1.3916, "step": 1340 }, { "epoch": 0.3681905283328771, "grad_norm": 0.39526777473833846, "learning_rate": 1.593261598693755e-05, "loss": 1.3478, "step": 1345 }, { "epoch": 0.3695592663564194, "grad_norm": 0.3837685918142903, "learning_rate": 1.5894078469692343e-05, "loss": 1.3604, "step": 1350 }, { "epoch": 0.37092800437996165, "grad_norm": 0.4184389090754904, "learning_rate": 1.585540634923223e-05, "loss": 1.4042, "step": 1355 }, { "epoch": 0.37229674240350397, "grad_norm": 0.4132734637825316, "learning_rate": 1.5816600508713372e-05, "loss": 1.3901, "step": 1360 }, { "epoch": 0.3736654804270463, "grad_norm": 0.3876929955451876, "learning_rate": 1.5777661834345708e-05, "loss": 1.3296, "step": 1365 }, { "epoch": 0.37503421845058854, "grad_norm": 0.39294627819783745, "learning_rate": 1.57385912153727e-05, "loss": 1.3581, "step": 1370 }, { "epoch": 0.37640295647413086, "grad_norm": 0.3875865415567979, "learning_rate": 1.5699389544051028e-05, "loss": 1.3167, "step": 1375 }, { "epoch": 0.37777169449767317, "grad_norm": 0.3863898154888909, "learning_rate": 1.566005771563023e-05, "loss": 1.3694, "step": 1380 }, { "epoch": 0.37914043252121543, "grad_norm": 0.38577432820095847, "learning_rate": 1.5620596628332242e-05, "loss": 1.323, "step": 1385 }, { "epoch": 0.38050917054475775, "grad_norm": 0.3971392209603939, "learning_rate": 1.5581007183330877e-05, "loss": 1.3432, "step": 1390 }, { "epoch": 0.3818779085683, "grad_norm": 0.40311344444707337, "learning_rate": 1.554129028473127e-05, "loss": 1.3802, "step": 1395 }, { "epoch": 0.3832466465918423, "grad_norm": 0.40527858159368374, "learning_rate": 1.5501446839549207e-05, "loss": 1.3445, "step": 1400 }, { "epoch": 0.38461538461538464, "grad_norm": 0.4025758901452214, "learning_rate": 1.5461477757690424e-05, "loss": 1.3321, "step": 1405 }, { "epoch": 0.3859841226389269, "grad_norm": 0.3855393398790419, "learning_rate": 1.542138395192983e-05, "loss": 1.3405, "step": 1410 }, { "epoch": 0.3873528606624692, "grad_norm": 0.38944762738859545, "learning_rate": 1.538116633789065e-05, "loss": 1.3289, "step": 1415 }, { "epoch": 0.38872159868601147, "grad_norm": 0.4028684971128393, "learning_rate": 1.5340825834023526e-05, "loss": 1.3798, "step": 1420 }, { "epoch": 0.3900903367095538, "grad_norm": 0.37654892707807947, "learning_rate": 1.530036336158553e-05, "loss": 1.3432, "step": 1425 }, { "epoch": 0.3914590747330961, "grad_norm": 0.3986727790830818, "learning_rate": 1.5259779844619152e-05, "loss": 1.3422, "step": 1430 }, { "epoch": 0.39282781275663836, "grad_norm": 0.4029735444716236, "learning_rate": 1.5219076209931159e-05, "loss": 1.3136, "step": 1435 }, { "epoch": 0.3941965507801807, "grad_norm": 0.41012078556001347, "learning_rate": 1.5178253387071458e-05, "loss": 1.4002, "step": 1440 }, { "epoch": 0.395565288803723, "grad_norm": 0.40771281882367433, "learning_rate": 1.5137312308311857e-05, "loss": 1.3684, "step": 1445 }, { "epoch": 0.39693402682726525, "grad_norm": 0.38445172076552614, "learning_rate": 1.5096253908624778e-05, "loss": 1.3137, "step": 1450 }, { "epoch": 0.39830276485080757, "grad_norm": 0.3869608798791883, "learning_rate": 1.5055079125661908e-05, "loss": 1.2812, "step": 1455 }, { "epoch": 0.3996715028743498, "grad_norm": 0.3987871366347409, "learning_rate": 1.5013788899732775e-05, "loss": 1.3394, "step": 1460 }, { "epoch": 0.40104024089789214, "grad_norm": 0.3970490183809672, "learning_rate": 1.4972384173783284e-05, "loss": 1.3544, "step": 1465 }, { "epoch": 0.40240897892143446, "grad_norm": 0.3925074770480503, "learning_rate": 1.493086589337418e-05, "loss": 1.3294, "step": 1470 }, { "epoch": 0.4037777169449767, "grad_norm": 0.40973551864547075, "learning_rate": 1.4889235006659448e-05, "loss": 1.3675, "step": 1475 }, { "epoch": 0.40514645496851903, "grad_norm": 0.3812235133361238, "learning_rate": 1.484749246436468e-05, "loss": 1.3288, "step": 1480 }, { "epoch": 0.40651519299206135, "grad_norm": 0.3827389411340829, "learning_rate": 1.4805639219765337e-05, "loss": 1.4128, "step": 1485 }, { "epoch": 0.4078839310156036, "grad_norm": 0.37343264715636293, "learning_rate": 1.476367622866499e-05, "loss": 1.3419, "step": 1490 }, { "epoch": 0.4092526690391459, "grad_norm": 0.4116326320287375, "learning_rate": 1.4721604449373505e-05, "loss": 1.297, "step": 1495 }, { "epoch": 0.4106214070626882, "grad_norm": 0.38955441577160277, "learning_rate": 1.4679424842685137e-05, "loss": 1.3138, "step": 1500 }, { "epoch": 0.4119901450862305, "grad_norm": 0.38444568811838686, "learning_rate": 1.4637138371856601e-05, "loss": 1.3284, "step": 1505 }, { "epoch": 0.4133588831097728, "grad_norm": 0.3938144538178601, "learning_rate": 1.4594746002585072e-05, "loss": 1.3498, "step": 1510 }, { "epoch": 0.41472762113331507, "grad_norm": 0.39696555787686244, "learning_rate": 1.4552248702986127e-05, "loss": 1.3524, "step": 1515 }, { "epoch": 0.4160963591568574, "grad_norm": 0.3802714110333375, "learning_rate": 1.4509647443571643e-05, "loss": 1.288, "step": 1520 }, { "epoch": 0.41746509718039965, "grad_norm": 0.38532238774246486, "learning_rate": 1.446694319722763e-05, "loss": 1.3656, "step": 1525 }, { "epoch": 0.41883383520394196, "grad_norm": 0.3866193343216763, "learning_rate": 1.4424136939192009e-05, "loss": 1.3696, "step": 1530 }, { "epoch": 0.4202025732274843, "grad_norm": 0.37501111401637777, "learning_rate": 1.4381229647032346e-05, "loss": 1.3512, "step": 1535 }, { "epoch": 0.42157131125102654, "grad_norm": 0.3700012749063157, "learning_rate": 1.4338222300623533e-05, "loss": 1.3092, "step": 1540 }, { "epoch": 0.42294004927456885, "grad_norm": 0.38806617391675347, "learning_rate": 1.4295115882125393e-05, "loss": 1.3471, "step": 1545 }, { "epoch": 0.42430878729811117, "grad_norm": 0.39116631927262135, "learning_rate": 1.4251911375960261e-05, "loss": 1.4043, "step": 1550 }, { "epoch": 0.4256775253216534, "grad_norm": 0.3907489868883888, "learning_rate": 1.4208609768790513e-05, "loss": 1.3476, "step": 1555 }, { "epoch": 0.42704626334519574, "grad_norm": 0.40188011896979126, "learning_rate": 1.4165212049496013e-05, "loss": 1.3398, "step": 1560 }, { "epoch": 0.428415001368738, "grad_norm": 0.39583117671628015, "learning_rate": 1.4121719209151545e-05, "loss": 1.3827, "step": 1565 }, { "epoch": 0.4297837393922803, "grad_norm": 0.3948036088686505, "learning_rate": 1.4078132241004174e-05, "loss": 1.3509, "step": 1570 }, { "epoch": 0.43115247741582263, "grad_norm": 0.3844018510427473, "learning_rate": 1.4034452140450561e-05, "loss": 1.3619, "step": 1575 }, { "epoch": 0.4325212154393649, "grad_norm": 0.40654356409937986, "learning_rate": 1.3990679905014235e-05, "loss": 1.2983, "step": 1580 }, { "epoch": 0.4338899534629072, "grad_norm": 0.401731234750035, "learning_rate": 1.3946816534322815e-05, "loss": 1.3573, "step": 1585 }, { "epoch": 0.43525869148644947, "grad_norm": 0.40604771069832035, "learning_rate": 1.3902863030085176e-05, "loss": 1.333, "step": 1590 }, { "epoch": 0.4366274295099918, "grad_norm": 0.3821888564061559, "learning_rate": 1.3858820396068572e-05, "loss": 1.3062, "step": 1595 }, { "epoch": 0.4379961675335341, "grad_norm": 0.3992586017031061, "learning_rate": 1.3814689638075725e-05, "loss": 1.3671, "step": 1600 }, { "epoch": 0.43936490555707636, "grad_norm": 0.4018110425948897, "learning_rate": 1.3770471763921833e-05, "loss": 1.3709, "step": 1605 }, { "epoch": 0.4407336435806187, "grad_norm": 0.387994622633552, "learning_rate": 1.372616778341158e-05, "loss": 1.3138, "step": 1610 }, { "epoch": 0.442102381604161, "grad_norm": 0.3974604574879118, "learning_rate": 1.3681778708316054e-05, "loss": 1.4087, "step": 1615 }, { "epoch": 0.44347111962770325, "grad_norm": 0.39844593969419934, "learning_rate": 1.3637305552349656e-05, "loss": 1.3599, "step": 1620 }, { "epoch": 0.44483985765124556, "grad_norm": 0.3857194594528982, "learning_rate": 1.3592749331146941e-05, "loss": 1.2897, "step": 1625 }, { "epoch": 0.4462085956747878, "grad_norm": 0.3750305160936592, "learning_rate": 1.3548111062239432e-05, "loss": 1.2819, "step": 1630 }, { "epoch": 0.44757733369833014, "grad_norm": 0.4007647993356723, "learning_rate": 1.350339176503237e-05, "loss": 1.3194, "step": 1635 }, { "epoch": 0.44894607172187245, "grad_norm": 0.4013887994126227, "learning_rate": 1.3458592460781446e-05, "loss": 1.3199, "step": 1640 }, { "epoch": 0.4503148097454147, "grad_norm": 0.3981164766269796, "learning_rate": 1.341371417256947e-05, "loss": 1.3464, "step": 1645 }, { "epoch": 0.451683547768957, "grad_norm": 0.39477092552678616, "learning_rate": 1.3368757925283015e-05, "loss": 1.3529, "step": 1650 }, { "epoch": 0.45305228579249934, "grad_norm": 0.3896314211042654, "learning_rate": 1.3323724745589007e-05, "loss": 1.3881, "step": 1655 }, { "epoch": 0.4544210238160416, "grad_norm": 0.3954835469300208, "learning_rate": 1.3278615661911274e-05, "loss": 1.3124, "step": 1660 }, { "epoch": 0.4557897618395839, "grad_norm": 0.3708379387210574, "learning_rate": 1.3233431704407072e-05, "loss": 1.2866, "step": 1665 }, { "epoch": 0.4571584998631262, "grad_norm": 0.38510327105983166, "learning_rate": 1.318817390494355e-05, "loss": 1.2968, "step": 1670 }, { "epoch": 0.4585272378866685, "grad_norm": 0.38362958034859684, "learning_rate": 1.3142843297074182e-05, "loss": 1.3155, "step": 1675 }, { "epoch": 0.4598959759102108, "grad_norm": 0.4082500339960413, "learning_rate": 1.3097440916015179e-05, "loss": 1.3646, "step": 1680 }, { "epoch": 0.46126471393375307, "grad_norm": 0.3927808793669975, "learning_rate": 1.3051967798621834e-05, "loss": 1.3165, "step": 1685 }, { "epoch": 0.4626334519572954, "grad_norm": 0.4002328011893471, "learning_rate": 1.300642498336484e-05, "loss": 1.3287, "step": 1690 }, { "epoch": 0.46400218998083764, "grad_norm": 0.38096569725907414, "learning_rate": 1.2960813510306599e-05, "loss": 1.319, "step": 1695 }, { "epoch": 0.46537092800437996, "grad_norm": 0.38869356480860534, "learning_rate": 1.2915134421077433e-05, "loss": 1.3763, "step": 1700 }, { "epoch": 0.4667396660279223, "grad_norm": 0.38290589898614796, "learning_rate": 1.2869388758851828e-05, "loss": 1.3204, "step": 1705 }, { "epoch": 0.46810840405146453, "grad_norm": 0.41880465304512116, "learning_rate": 1.2823577568324604e-05, "loss": 1.3858, "step": 1710 }, { "epoch": 0.46947714207500685, "grad_norm": 0.4031243358335159, "learning_rate": 1.2777701895687034e-05, "loss": 1.3346, "step": 1715 }, { "epoch": 0.47084588009854916, "grad_norm": 0.3919138760842289, "learning_rate": 1.2731762788602988e-05, "loss": 1.4046, "step": 1720 }, { "epoch": 0.4722146181220914, "grad_norm": 0.40045145219664335, "learning_rate": 1.2685761296184987e-05, "loss": 1.3808, "step": 1725 }, { "epoch": 0.47358335614563374, "grad_norm": 0.38631426280799075, "learning_rate": 1.2639698468970237e-05, "loss": 1.3563, "step": 1730 }, { "epoch": 0.474952094169176, "grad_norm": 0.3891224496732836, "learning_rate": 1.259357535889666e-05, "loss": 1.368, "step": 1735 }, { "epoch": 0.4763208321927183, "grad_norm": 0.40044455029632475, "learning_rate": 1.2547393019278853e-05, "loss": 1.3548, "step": 1740 }, { "epoch": 0.47768957021626063, "grad_norm": 0.3899915581292829, "learning_rate": 1.2501152504784044e-05, "loss": 1.3291, "step": 1745 }, { "epoch": 0.4790583082398029, "grad_norm": 0.3869143385442828, "learning_rate": 1.2454854871407993e-05, "loss": 1.3473, "step": 1750 }, { "epoch": 0.4804270462633452, "grad_norm": 0.3824047072867315, "learning_rate": 1.2408501176450898e-05, "loss": 1.3196, "step": 1755 }, { "epoch": 0.48179578428688746, "grad_norm": 0.4057108906200233, "learning_rate": 1.2362092478493226e-05, "loss": 1.3554, "step": 1760 }, { "epoch": 0.4831645223104298, "grad_norm": 0.3776742288321122, "learning_rate": 1.2315629837371556e-05, "loss": 1.3014, "step": 1765 }, { "epoch": 0.4845332603339721, "grad_norm": 0.3977510874849243, "learning_rate": 1.2269114314154365e-05, "loss": 1.3967, "step": 1770 }, { "epoch": 0.48590199835751435, "grad_norm": 0.3697505648941807, "learning_rate": 1.2222546971117797e-05, "loss": 1.3171, "step": 1775 }, { "epoch": 0.48727073638105667, "grad_norm": 0.40788370625821474, "learning_rate": 1.2175928871721411e-05, "loss": 1.3222, "step": 1780 }, { "epoch": 0.488639474404599, "grad_norm": 0.39090111193092664, "learning_rate": 1.2129261080583897e-05, "loss": 1.3676, "step": 1785 }, { "epoch": 0.49000821242814124, "grad_norm": 0.3983751803246069, "learning_rate": 1.2082544663458736e-05, "loss": 1.3364, "step": 1790 }, { "epoch": 0.49137695045168356, "grad_norm": 0.3848833412781912, "learning_rate": 1.20357806872099e-05, "loss": 1.2978, "step": 1795 }, { "epoch": 0.4927456884752258, "grad_norm": 0.37897812252926655, "learning_rate": 1.1988970219787467e-05, "loss": 1.3333, "step": 1800 }, { "epoch": 0.49411442649876813, "grad_norm": 0.3948100327011319, "learning_rate": 1.1942114330203227e-05, "loss": 1.3665, "step": 1805 }, { "epoch": 0.49548316452231045, "grad_norm": 0.3844309877664426, "learning_rate": 1.1895214088506284e-05, "loss": 1.3383, "step": 1810 }, { "epoch": 0.4968519025458527, "grad_norm": 0.3785571762413702, "learning_rate": 1.1848270565758616e-05, "loss": 1.2997, "step": 1815 }, { "epoch": 0.498220640569395, "grad_norm": 0.3873700740980167, "learning_rate": 1.1801284834010596e-05, "loss": 1.3332, "step": 1820 }, { "epoch": 0.49958937859293734, "grad_norm": 0.3971195600650439, "learning_rate": 1.1754257966276544e-05, "loss": 1.3468, "step": 1825 }, { "epoch": 0.5009581166164796, "grad_norm": 0.37924897770225724, "learning_rate": 1.1707191036510189e-05, "loss": 1.3152, "step": 1830 }, { "epoch": 0.5023268546400219, "grad_norm": 0.37889309262468557, "learning_rate": 1.1660085119580165e-05, "loss": 1.3074, "step": 1835 }, { "epoch": 0.5036955926635642, "grad_norm": 0.38526267448234497, "learning_rate": 1.1612941291245456e-05, "loss": 1.2863, "step": 1840 }, { "epoch": 0.5050643306871064, "grad_norm": 0.4007327343298376, "learning_rate": 1.1565760628130824e-05, "loss": 1.348, "step": 1845 }, { "epoch": 0.5064330687106487, "grad_norm": 0.3917232580410847, "learning_rate": 1.1518544207702238e-05, "loss": 1.3193, "step": 1850 }, { "epoch": 0.5078018067341911, "grad_norm": 0.41154960813532804, "learning_rate": 1.1471293108242251e-05, "loss": 1.3921, "step": 1855 }, { "epoch": 0.5091705447577334, "grad_norm": 0.3808027664016428, "learning_rate": 1.1424008408825383e-05, "loss": 1.337, "step": 1860 }, { "epoch": 0.5105392827812757, "grad_norm": 0.38767901381709907, "learning_rate": 1.1376691189293474e-05, "loss": 1.3142, "step": 1865 }, { "epoch": 0.511908020804818, "grad_norm": 0.39402951617560983, "learning_rate": 1.1329342530231036e-05, "loss": 1.358, "step": 1870 }, { "epoch": 0.5132767588283602, "grad_norm": 0.38538496638703307, "learning_rate": 1.128196351294055e-05, "loss": 1.2994, "step": 1875 }, { "epoch": 0.5146454968519025, "grad_norm": 0.38082093170114384, "learning_rate": 1.1234555219417804e-05, "loss": 1.3208, "step": 1880 }, { "epoch": 0.5160142348754448, "grad_norm": 0.3748307863634792, "learning_rate": 1.1187118732327167e-05, "loss": 1.3506, "step": 1885 }, { "epoch": 0.5173829728989872, "grad_norm": 0.38705593276141625, "learning_rate": 1.1139655134976855e-05, "loss": 1.2906, "step": 1890 }, { "epoch": 0.5187517109225295, "grad_norm": 0.39888481423655353, "learning_rate": 1.1092165511294206e-05, "loss": 1.367, "step": 1895 }, { "epoch": 0.5201204489460717, "grad_norm": 0.3971141048112466, "learning_rate": 1.104465094580093e-05, "loss": 1.3405, "step": 1900 }, { "epoch": 0.521489186969614, "grad_norm": 0.3837313790185174, "learning_rate": 1.0997112523588322e-05, "loss": 1.3351, "step": 1905 }, { "epoch": 0.5228579249931563, "grad_norm": 0.39019433157562, "learning_rate": 1.0949551330292502e-05, "loss": 1.2482, "step": 1910 }, { "epoch": 0.5242266630166986, "grad_norm": 0.39062012523465395, "learning_rate": 1.090196845206961e-05, "loss": 1.343, "step": 1915 }, { "epoch": 0.5255954010402409, "grad_norm": 0.3948259084399907, "learning_rate": 1.0854364975571004e-05, "loss": 1.3001, "step": 1920 }, { "epoch": 0.5269641390637831, "grad_norm": 0.3845620013681871, "learning_rate": 1.0806741987918448e-05, "loss": 1.3449, "step": 1925 }, { "epoch": 0.5283328770873255, "grad_norm": 0.3976520791295208, "learning_rate": 1.075910057667928e-05, "loss": 1.3007, "step": 1930 }, { "epoch": 0.5297016151108678, "grad_norm": 0.3817709747936168, "learning_rate": 1.071144182984158e-05, "loss": 1.3453, "step": 1935 }, { "epoch": 0.5310703531344101, "grad_norm": 0.38542609340341144, "learning_rate": 1.0663766835789327e-05, "loss": 1.3187, "step": 1940 }, { "epoch": 0.5324390911579524, "grad_norm": 0.3775171570445251, "learning_rate": 1.0616076683277524e-05, "loss": 1.3315, "step": 1945 }, { "epoch": 0.5338078291814946, "grad_norm": 0.40024298975169065, "learning_rate": 1.056837246140736e-05, "loss": 1.3772, "step": 1950 }, { "epoch": 0.5351765672050369, "grad_norm": 0.3882397104456436, "learning_rate": 1.0520655259601325e-05, "loss": 1.3284, "step": 1955 }, { "epoch": 0.5365453052285792, "grad_norm": 0.3892143818918897, "learning_rate": 1.0472926167578323e-05, "loss": 1.302, "step": 1960 }, { "epoch": 0.5379140432521216, "grad_norm": 0.41282669532669847, "learning_rate": 1.042518627532881e-05, "loss": 1.3261, "step": 1965 }, { "epoch": 0.5392827812756639, "grad_norm": 0.37154889529358315, "learning_rate": 1.0377436673089873e-05, "loss": 1.3142, "step": 1970 }, { "epoch": 0.5406515192992062, "grad_norm": 0.4001478843028637, "learning_rate": 1.0329678451320352e-05, "loss": 1.3894, "step": 1975 }, { "epoch": 0.5420202573227484, "grad_norm": 0.4082737912884333, "learning_rate": 1.0281912700675937e-05, "loss": 1.38, "step": 1980 }, { "epoch": 0.5433889953462907, "grad_norm": 0.3976694377900842, "learning_rate": 1.0234140511984246e-05, "loss": 1.337, "step": 1985 }, { "epoch": 0.544757733369833, "grad_norm": 0.4090490835246688, "learning_rate": 1.0186362976219926e-05, "loss": 1.3389, "step": 1990 }, { "epoch": 0.5461264713933753, "grad_norm": 0.39123091240448754, "learning_rate": 1.0138581184479737e-05, "loss": 1.2726, "step": 1995 }, { "epoch": 0.5474952094169176, "grad_norm": 0.383808685203924, "learning_rate": 1.0090796227957633e-05, "loss": 1.3344, "step": 2000 }, { "epoch": 0.5488639474404599, "grad_norm": 0.39282920042398406, "learning_rate": 1.0043009197919836e-05, "loss": 1.3188, "step": 2005 }, { "epoch": 0.5502326854640022, "grad_norm": 0.3877735838132022, "learning_rate": 9.99522118567993e-06, "loss": 1.3412, "step": 2010 }, { "epoch": 0.5516014234875445, "grad_norm": 0.3855977732261345, "learning_rate": 9.947433282573926e-06, "loss": 1.31, "step": 2015 }, { "epoch": 0.5529701615110868, "grad_norm": 0.39182359203627265, "learning_rate": 9.899646579935336e-06, "loss": 1.3288, "step": 2020 }, { "epoch": 0.5543388995346291, "grad_norm": 0.37896193597955086, "learning_rate": 9.85186216907027e-06, "loss": 1.3222, "step": 2025 }, { "epoch": 0.5557076375581713, "grad_norm": 0.39627324276578396, "learning_rate": 9.80408114123249e-06, "loss": 1.3659, "step": 2030 }, { "epoch": 0.5570763755817136, "grad_norm": 0.39331176053433153, "learning_rate": 9.756304587598503e-06, "loss": 1.4234, "step": 2035 }, { "epoch": 0.558445113605256, "grad_norm": 0.38413569943245757, "learning_rate": 9.708533599242643e-06, "loss": 1.2809, "step": 2040 }, { "epoch": 0.5598138516287983, "grad_norm": 0.378050582923272, "learning_rate": 9.660769267112137e-06, "loss": 1.3, "step": 2045 }, { "epoch": 0.5611825896523406, "grad_norm": 0.3810287903511996, "learning_rate": 9.61301268200222e-06, "loss": 1.328, "step": 2050 }, { "epoch": 0.5625513276758828, "grad_norm": 0.3968883040797777, "learning_rate": 9.5652649345312e-06, "loss": 1.3861, "step": 2055 }, { "epoch": 0.5639200656994251, "grad_norm": 0.381754630880566, "learning_rate": 9.517527115115554e-06, "loss": 1.2931, "step": 2060 }, { "epoch": 0.5652888037229674, "grad_norm": 0.3934475062551264, "learning_rate": 9.46980031394504e-06, "loss": 1.357, "step": 2065 }, { "epoch": 0.5666575417465097, "grad_norm": 0.3764960791780743, "learning_rate": 9.422085620957795e-06, "loss": 1.3112, "step": 2070 }, { "epoch": 0.568026279770052, "grad_norm": 0.3819475902039454, "learning_rate": 9.374384125815427e-06, "loss": 1.317, "step": 2075 }, { "epoch": 0.5693950177935944, "grad_norm": 0.39075935188345634, "learning_rate": 9.326696917878159e-06, "loss": 1.329, "step": 2080 }, { "epoch": 0.5707637558171366, "grad_norm": 0.4115410325839856, "learning_rate": 9.27902508617993e-06, "loss": 1.3405, "step": 2085 }, { "epoch": 0.5721324938406789, "grad_norm": 0.3929327529237431, "learning_rate": 9.23136971940353e-06, "loss": 1.2811, "step": 2090 }, { "epoch": 0.5735012318642212, "grad_norm": 0.38257057379328274, "learning_rate": 9.183731905855746e-06, "loss": 1.3234, "step": 2095 }, { "epoch": 0.5748699698877635, "grad_norm": 0.38674820150451145, "learning_rate": 9.136112733442493e-06, "loss": 1.3466, "step": 2100 }, { "epoch": 0.5762387079113058, "grad_norm": 0.39368922489348585, "learning_rate": 9.088513289643982e-06, "loss": 1.3149, "step": 2105 }, { "epoch": 0.577607445934848, "grad_norm": 0.38199335563674713, "learning_rate": 9.040934661489885e-06, "loss": 1.3576, "step": 2110 }, { "epoch": 0.5789761839583903, "grad_norm": 0.39323361965983067, "learning_rate": 8.9933779355345e-06, "loss": 1.3159, "step": 2115 }, { "epoch": 0.5803449219819327, "grad_norm": 0.39256654497554794, "learning_rate": 8.945844197831945e-06, "loss": 1.3623, "step": 2120 }, { "epoch": 0.581713660005475, "grad_norm": 0.3905410660434604, "learning_rate": 8.898334533911362e-06, "loss": 1.3074, "step": 2125 }, { "epoch": 0.5830823980290173, "grad_norm": 0.3700729231212675, "learning_rate": 8.850850028752108e-06, "loss": 1.3436, "step": 2130 }, { "epoch": 0.5844511360525595, "grad_norm": 0.3701852614302805, "learning_rate": 8.803391766758998e-06, "loss": 1.2672, "step": 2135 }, { "epoch": 0.5858198740761018, "grad_norm": 0.3809658344783075, "learning_rate": 8.755960831737529e-06, "loss": 1.3638, "step": 2140 }, { "epoch": 0.5871886120996441, "grad_norm": 0.3852501406589481, "learning_rate": 8.708558306869125e-06, "loss": 1.3389, "step": 2145 }, { "epoch": 0.5885573501231864, "grad_norm": 0.4003501899309122, "learning_rate": 8.661185274686418e-06, "loss": 1.3228, "step": 2150 }, { "epoch": 0.5899260881467288, "grad_norm": 0.38933438340786924, "learning_rate": 8.613842817048503e-06, "loss": 1.3627, "step": 2155 }, { "epoch": 0.591294826170271, "grad_norm": 0.39608781513600605, "learning_rate": 8.566532015116251e-06, "loss": 1.3626, "step": 2160 }, { "epoch": 0.5926635641938133, "grad_norm": 0.37582004076749886, "learning_rate": 8.51925394932761e-06, "loss": 1.2766, "step": 2165 }, { "epoch": 0.5940323022173556, "grad_norm": 0.39046439420533485, "learning_rate": 8.47200969937293e-06, "loss": 1.3342, "step": 2170 }, { "epoch": 0.5954010402408979, "grad_norm": 0.3846853640345239, "learning_rate": 8.424800344170308e-06, "loss": 1.2973, "step": 2175 }, { "epoch": 0.5967697782644402, "grad_norm": 0.3794031254673053, "learning_rate": 8.37762696184095e-06, "loss": 1.3288, "step": 2180 }, { "epoch": 0.5981385162879824, "grad_norm": 0.38487683969290326, "learning_rate": 8.330490629684547e-06, "loss": 1.3374, "step": 2185 }, { "epoch": 0.5995072543115247, "grad_norm": 0.38666671699682115, "learning_rate": 8.28339242415468e-06, "loss": 1.3202, "step": 2190 }, { "epoch": 0.600875992335067, "grad_norm": 0.3788541652969095, "learning_rate": 8.236333420834216e-06, "loss": 1.2778, "step": 2195 }, { "epoch": 0.6022447303586094, "grad_norm": 0.4003559394681289, "learning_rate": 8.189314694410781e-06, "loss": 1.3349, "step": 2200 }, { "epoch": 0.6036134683821517, "grad_norm": 0.40672129713209243, "learning_rate": 8.14233731865218e-06, "loss": 1.3343, "step": 2205 }, { "epoch": 0.604982206405694, "grad_norm": 0.392344465653996, "learning_rate": 8.0954023663819e-06, "loss": 1.2804, "step": 2210 }, { "epoch": 0.6063509444292362, "grad_norm": 0.396931681160349, "learning_rate": 8.048510909454601e-06, "loss": 1.3712, "step": 2215 }, { "epoch": 0.6077196824527785, "grad_norm": 0.3882127247657791, "learning_rate": 8.001664018731637e-06, "loss": 1.3313, "step": 2220 }, { "epoch": 0.6090884204763208, "grad_norm": 0.3727086035483801, "learning_rate": 7.954862764056602e-06, "loss": 1.2715, "step": 2225 }, { "epoch": 0.6104571584998632, "grad_norm": 0.39519191088539024, "learning_rate": 7.908108214230902e-06, "loss": 1.3018, "step": 2230 }, { "epoch": 0.6118258965234055, "grad_norm": 0.39954515244878724, "learning_rate": 7.861401436989337e-06, "loss": 1.3697, "step": 2235 }, { "epoch": 0.6131946345469477, "grad_norm": 0.3961862899698687, "learning_rate": 7.814743498975731e-06, "loss": 1.3651, "step": 2240 }, { "epoch": 0.61456337257049, "grad_norm": 0.38565731038310297, "learning_rate": 7.768135465718559e-06, "loss": 1.3462, "step": 2245 }, { "epoch": 0.6159321105940323, "grad_norm": 0.38796255525719986, "learning_rate": 7.72157840160662e-06, "loss": 1.3087, "step": 2250 }, { "epoch": 0.6173008486175746, "grad_norm": 0.3947254708668997, "learning_rate": 7.67507336986474e-06, "loss": 1.262, "step": 2255 }, { "epoch": 0.6186695866411169, "grad_norm": 0.39167259958488637, "learning_rate": 7.628621432529461e-06, "loss": 1.3117, "step": 2260 }, { "epoch": 0.6200383246646591, "grad_norm": 0.39458361491505556, "learning_rate": 7.582223650424825e-06, "loss": 1.3067, "step": 2265 }, { "epoch": 0.6214070626882015, "grad_norm": 0.3948240515553033, "learning_rate": 7.5358810831381225e-06, "loss": 1.3177, "step": 2270 }, { "epoch": 0.6227758007117438, "grad_norm": 0.3814626352546649, "learning_rate": 7.489594788995698e-06, "loss": 1.3259, "step": 2275 }, { "epoch": 0.6241445387352861, "grad_norm": 0.39070667312634827, "learning_rate": 7.443365825038793e-06, "loss": 1.3212, "step": 2280 }, { "epoch": 0.6255132767588284, "grad_norm": 0.38913877038236866, "learning_rate": 7.397195246999391e-06, "loss": 1.346, "step": 2285 }, { "epoch": 0.6268820147823706, "grad_norm": 0.4070153012541291, "learning_rate": 7.351084109276119e-06, "loss": 1.3826, "step": 2290 }, { "epoch": 0.6282507528059129, "grad_norm": 0.36428744429147014, "learning_rate": 7.305033464910164e-06, "loss": 1.2659, "step": 2295 }, { "epoch": 0.6296194908294552, "grad_norm": 0.3877290507395246, "learning_rate": 7.25904436556122e-06, "loss": 1.304, "step": 2300 }, { "epoch": 0.6309882288529975, "grad_norm": 0.3803920075270922, "learning_rate": 7.21311786148348e-06, "loss": 1.3333, "step": 2305 }, { "epoch": 0.6323569668765399, "grad_norm": 0.3822541348646928, "learning_rate": 7.167255001501651e-06, "loss": 1.3307, "step": 2310 }, { "epoch": 0.6337257049000822, "grad_norm": 0.3977624877197784, "learning_rate": 7.121456832986988e-06, "loss": 1.3329, "step": 2315 }, { "epoch": 0.6350944429236244, "grad_norm": 0.38990040940646387, "learning_rate": 7.075724401833395e-06, "loss": 1.3647, "step": 2320 }, { "epoch": 0.6364631809471667, "grad_norm": 0.39550975070062716, "learning_rate": 7.030058752433526e-06, "loss": 1.2715, "step": 2325 }, { "epoch": 0.637831918970709, "grad_norm": 0.3857148917491234, "learning_rate": 6.984460927654937e-06, "loss": 1.3191, "step": 2330 }, { "epoch": 0.6392006569942513, "grad_norm": 0.38855941306596253, "learning_rate": 6.938931968816275e-06, "loss": 1.2979, "step": 2335 }, { "epoch": 0.6405693950177936, "grad_norm": 0.3892088605550479, "learning_rate": 6.893472915663493e-06, "loss": 1.3826, "step": 2340 }, { "epoch": 0.6419381330413358, "grad_norm": 0.4024108620394257, "learning_rate": 6.8480848063461035e-06, "loss": 1.2943, "step": 2345 }, { "epoch": 0.6433068710648782, "grad_norm": 0.3976600247063458, "learning_rate": 6.8027686773934765e-06, "loss": 1.3241, "step": 2350 }, { "epoch": 0.6446756090884205, "grad_norm": 0.38538373822690386, "learning_rate": 6.7575255636911626e-06, "loss": 1.3316, "step": 2355 }, { "epoch": 0.6460443471119628, "grad_norm": 0.3961847974725629, "learning_rate": 6.7123564984572596e-06, "loss": 1.3481, "step": 2360 }, { "epoch": 0.6474130851355051, "grad_norm": 0.3739018408786073, "learning_rate": 6.667262513218824e-06, "loss": 1.2857, "step": 2365 }, { "epoch": 0.6487818231590473, "grad_norm": 0.4058027362822848, "learning_rate": 6.622244637788302e-06, "loss": 1.3623, "step": 2370 }, { "epoch": 0.6501505611825896, "grad_norm": 0.3887614780397367, "learning_rate": 6.577303900240023e-06, "loss": 1.3466, "step": 2375 }, { "epoch": 0.6515192992061319, "grad_norm": 0.39583219449776136, "learning_rate": 6.532441326886716e-06, "loss": 1.3826, "step": 2380 }, { "epoch": 0.6528880372296743, "grad_norm": 0.38600694323078566, "learning_rate": 6.487657942256069e-06, "loss": 1.3124, "step": 2385 }, { "epoch": 0.6542567752532166, "grad_norm": 0.35581732986644143, "learning_rate": 6.442954769067341e-06, "loss": 1.3352, "step": 2390 }, { "epoch": 0.6556255132767588, "grad_norm": 0.3977530272775962, "learning_rate": 6.398332828207996e-06, "loss": 1.3226, "step": 2395 }, { "epoch": 0.6569942513003011, "grad_norm": 0.3928175408666658, "learning_rate": 6.3537931387103925e-06, "loss": 1.3262, "step": 2400 }, { "epoch": 0.6583629893238434, "grad_norm": 0.3795840583774513, "learning_rate": 6.309336717728516e-06, "loss": 1.3203, "step": 2405 }, { "epoch": 0.6597317273473857, "grad_norm": 0.396679060826067, "learning_rate": 6.264964580514745e-06, "loss": 1.2842, "step": 2410 }, { "epoch": 0.661100465370928, "grad_norm": 0.3699325537006795, "learning_rate": 6.220677740396668e-06, "loss": 1.2936, "step": 2415 }, { "epoch": 0.6624692033944704, "grad_norm": 0.38098356916961845, "learning_rate": 6.176477208753944e-06, "loss": 1.276, "step": 2420 }, { "epoch": 0.6638379414180126, "grad_norm": 0.38873579028440075, "learning_rate": 6.132363994995194e-06, "loss": 1.2348, "step": 2425 }, { "epoch": 0.6652066794415549, "grad_norm": 0.39405073870487267, "learning_rate": 6.088339106534971e-06, "loss": 1.2659, "step": 2430 }, { "epoch": 0.6665754174650972, "grad_norm": 0.3833620202872981, "learning_rate": 6.044403548770735e-06, "loss": 1.3103, "step": 2435 }, { "epoch": 0.6679441554886395, "grad_norm": 0.39527725366864974, "learning_rate": 6.000558325059894e-06, "loss": 1.3423, "step": 2440 }, { "epoch": 0.6693128935121818, "grad_norm": 0.38996615233504806, "learning_rate": 5.956804436696904e-06, "loss": 1.303, "step": 2445 }, { "epoch": 0.670681631535724, "grad_norm": 0.3943489977366666, "learning_rate": 5.9131428828903905e-06, "loss": 1.3547, "step": 2450 }, { "epoch": 0.6720503695592663, "grad_norm": 0.3899208460332646, "learning_rate": 5.8695746607403285e-06, "loss": 1.2969, "step": 2455 }, { "epoch": 0.6734191075828087, "grad_norm": 0.3940611248671062, "learning_rate": 5.826100765215273e-06, "loss": 1.318, "step": 2460 }, { "epoch": 0.674787845606351, "grad_norm": 0.3930083338553073, "learning_rate": 5.782722189129655e-06, "loss": 1.2906, "step": 2465 }, { "epoch": 0.6761565836298933, "grad_norm": 0.39311675749157776, "learning_rate": 5.739439923121077e-06, "loss": 1.3569, "step": 2470 }, { "epoch": 0.6775253216534355, "grad_norm": 0.4114954929599597, "learning_rate": 5.6962549556277134e-06, "loss": 1.3773, "step": 2475 }, { "epoch": 0.6788940596769778, "grad_norm": 0.38652001787654966, "learning_rate": 5.653168272865724e-06, "loss": 1.3008, "step": 2480 }, { "epoch": 0.6802627977005201, "grad_norm": 0.3878760723325068, "learning_rate": 5.6101808588067505e-06, "loss": 1.2992, "step": 2485 }, { "epoch": 0.6816315357240624, "grad_norm": 0.39884745029373364, "learning_rate": 5.56729369515542e-06, "loss": 1.3088, "step": 2490 }, { "epoch": 0.6830002737476047, "grad_norm": 0.38961302860625757, "learning_rate": 5.52450776132694e-06, "loss": 1.2904, "step": 2495 }, { "epoch": 0.684369011771147, "grad_norm": 0.3970778873479824, "learning_rate": 5.481824034424741e-06, "loss": 1.3464, "step": 2500 }, { "epoch": 0.6857377497946893, "grad_norm": 0.4014161218820373, "learning_rate": 5.439243489218138e-06, "loss": 1.3376, "step": 2505 }, { "epoch": 0.6871064878182316, "grad_norm": 0.3874962119825426, "learning_rate": 5.396767098120087e-06, "loss": 1.3078, "step": 2510 }, { "epoch": 0.6884752258417739, "grad_norm": 0.37396226223807555, "learning_rate": 5.354395831164982e-06, "loss": 1.3097, "step": 2515 }, { "epoch": 0.6898439638653162, "grad_norm": 0.3792883560955228, "learning_rate": 5.312130655986485e-06, "loss": 1.3158, "step": 2520 }, { "epoch": 0.6912127018888585, "grad_norm": 0.3798203691318429, "learning_rate": 5.269972537795434e-06, "loss": 1.3087, "step": 2525 }, { "epoch": 0.6925814399124007, "grad_norm": 0.38253235315284295, "learning_rate": 5.227922439357823e-06, "loss": 1.3198, "step": 2530 }, { "epoch": 0.693950177935943, "grad_norm": 0.3817014501814054, "learning_rate": 5.1859813209727775e-06, "loss": 1.321, "step": 2535 }, { "epoch": 0.6953189159594854, "grad_norm": 0.3950376035714224, "learning_rate": 5.144150140450643e-06, "loss": 1.3024, "step": 2540 }, { "epoch": 0.6966876539830277, "grad_norm": 0.4084629435066235, "learning_rate": 5.102429853091128e-06, "loss": 1.3437, "step": 2545 }, { "epoch": 0.69805639200657, "grad_norm": 0.3808185831587325, "learning_rate": 5.060821411661459e-06, "loss": 1.3264, "step": 2550 }, { "epoch": 0.6994251300301122, "grad_norm": 0.3937638284493571, "learning_rate": 5.019325766374625e-06, "loss": 1.314, "step": 2555 }, { "epoch": 0.7007938680536545, "grad_norm": 0.3920386340433558, "learning_rate": 4.977943864867712e-06, "loss": 1.3271, "step": 2560 }, { "epoch": 0.7021626060771968, "grad_norm": 0.3898729624487372, "learning_rate": 4.936676652180215e-06, "loss": 1.3467, "step": 2565 }, { "epoch": 0.7035313441007391, "grad_norm": 0.39902316286587364, "learning_rate": 4.89552507073248e-06, "loss": 1.3102, "step": 2570 }, { "epoch": 0.7049000821242815, "grad_norm": 0.3673934948431135, "learning_rate": 4.854490060304192e-06, "loss": 1.2612, "step": 2575 }, { "epoch": 0.7062688201478237, "grad_norm": 0.3922497888421968, "learning_rate": 4.813572558012892e-06, "loss": 1.329, "step": 2580 }, { "epoch": 0.707637558171366, "grad_norm": 0.3860080742782407, "learning_rate": 4.772773498292579e-06, "loss": 1.3332, "step": 2585 }, { "epoch": 0.7090062961949083, "grad_norm": 0.3695087849805836, "learning_rate": 4.732093812872391e-06, "loss": 1.2697, "step": 2590 }, { "epoch": 0.7103750342184506, "grad_norm": 0.4012769440751328, "learning_rate": 4.691534430755302e-06, "loss": 1.3213, "step": 2595 }, { "epoch": 0.7117437722419929, "grad_norm": 0.38657730746893026, "learning_rate": 4.651096278196916e-06, "loss": 1.3665, "step": 2600 }, { "epoch": 0.7131125102655351, "grad_norm": 0.38560469814966725, "learning_rate": 4.610780278684315e-06, "loss": 1.3369, "step": 2605 }, { "epoch": 0.7144812482890774, "grad_norm": 0.39542221048440374, "learning_rate": 4.570587352914977e-06, "loss": 1.3048, "step": 2610 }, { "epoch": 0.7158499863126198, "grad_norm": 0.3910002548641959, "learning_rate": 4.530518418775734e-06, "loss": 1.3247, "step": 2615 }, { "epoch": 0.7172187243361621, "grad_norm": 0.3825608407076146, "learning_rate": 4.490574391321814e-06, "loss": 1.2967, "step": 2620 }, { "epoch": 0.7185874623597044, "grad_norm": 0.4028460389272689, "learning_rate": 4.450756182755963e-06, "loss": 1.3671, "step": 2625 }, { "epoch": 0.7199562003832466, "grad_norm": 0.40485919090431155, "learning_rate": 4.411064702407585e-06, "loss": 1.3556, "step": 2630 }, { "epoch": 0.7213249384067889, "grad_norm": 0.39021669719654245, "learning_rate": 4.371500856711988e-06, "loss": 1.3774, "step": 2635 }, { "epoch": 0.7226936764303312, "grad_norm": 0.3958279714364836, "learning_rate": 4.332065549189697e-06, "loss": 1.2982, "step": 2640 }, { "epoch": 0.7240624144538735, "grad_norm": 0.4060690524084766, "learning_rate": 4.292759680425794e-06, "loss": 1.3767, "step": 2645 }, { "epoch": 0.7254311524774159, "grad_norm": 0.40926421108177974, "learning_rate": 4.253584148049369e-06, "loss": 1.339, "step": 2650 }, { "epoch": 0.7267998905009582, "grad_norm": 0.37375698755072967, "learning_rate": 4.214539846713024e-06, "loss": 1.2961, "step": 2655 }, { "epoch": 0.7281686285245004, "grad_norm": 0.3968852233360893, "learning_rate": 4.175627668072425e-06, "loss": 1.2962, "step": 2660 }, { "epoch": 0.7295373665480427, "grad_norm": 0.4073445114415598, "learning_rate": 4.136848500765948e-06, "loss": 1.4056, "step": 2665 }, { "epoch": 0.730906104571585, "grad_norm": 0.3950383171619063, "learning_rate": 4.098203230394399e-06, "loss": 1.3192, "step": 2670 }, { "epoch": 0.7322748425951273, "grad_norm": 0.3986322017273241, "learning_rate": 4.059692739500761e-06, "loss": 1.3485, "step": 2675 }, { "epoch": 0.7336435806186696, "grad_norm": 0.3848930065792978, "learning_rate": 4.02131790755006e-06, "loss": 1.378, "step": 2680 }, { "epoch": 0.7350123186422118, "grad_norm": 0.39139666434515563, "learning_rate": 3.983079610909283e-06, "loss": 1.3334, "step": 2685 }, { "epoch": 0.7363810566657542, "grad_norm": 0.38858454704969564, "learning_rate": 3.944978722827347e-06, "loss": 1.3235, "step": 2690 }, { "epoch": 0.7377497946892965, "grad_norm": 0.390122324459421, "learning_rate": 3.907016113415166e-06, "loss": 1.3583, "step": 2695 }, { "epoch": 0.7391185327128388, "grad_norm": 0.3919502066927978, "learning_rate": 3.869192649625792e-06, "loss": 1.3092, "step": 2700 }, { "epoch": 0.7404872707363811, "grad_norm": 0.39777490757009926, "learning_rate": 3.831509195234598e-06, "loss": 1.3358, "step": 2705 }, { "epoch": 0.7418560087599233, "grad_norm": 0.39353868348346743, "learning_rate": 3.793966610819545e-06, "loss": 1.3515, "step": 2710 }, { "epoch": 0.7432247467834656, "grad_norm": 0.39342447998988617, "learning_rate": 3.756565753741569e-06, "loss": 1.2885, "step": 2715 }, { "epoch": 0.7445934848070079, "grad_norm": 0.37781062124467246, "learning_rate": 3.7193074781249585e-06, "loss": 1.276, "step": 2720 }, { "epoch": 0.7459622228305502, "grad_norm": 0.38397833073923665, "learning_rate": 3.6821926348378666e-06, "loss": 1.3031, "step": 2725 }, { "epoch": 0.7473309608540926, "grad_norm": 0.3993776914034358, "learning_rate": 3.6452220714728883e-06, "loss": 1.3719, "step": 2730 }, { "epoch": 0.7486996988776348, "grad_norm": 0.387591612397714, "learning_rate": 3.608396632327684e-06, "loss": 1.3006, "step": 2735 }, { "epoch": 0.7500684369011771, "grad_norm": 0.4059102325008591, "learning_rate": 3.5717171583857115e-06, "loss": 1.3689, "step": 2740 }, { "epoch": 0.7514371749247194, "grad_norm": 0.3772375625162727, "learning_rate": 3.5351844872970233e-06, "loss": 1.3091, "step": 2745 }, { "epoch": 0.7528059129482617, "grad_norm": 0.3853940006094772, "learning_rate": 3.498799453359124e-06, "loss": 1.3454, "step": 2750 }, { "epoch": 0.754174650971804, "grad_norm": 0.3771144405512011, "learning_rate": 3.462562887497927e-06, "loss": 1.2998, "step": 2755 }, { "epoch": 0.7555433889953463, "grad_norm": 0.39604150906533137, "learning_rate": 3.4264756172487813e-06, "loss": 1.3129, "step": 2760 }, { "epoch": 0.7569121270188885, "grad_norm": 0.3843657920576998, "learning_rate": 3.390538466737564e-06, "loss": 1.2981, "step": 2765 }, { "epoch": 0.7582808650424309, "grad_norm": 0.39238834291334906, "learning_rate": 3.3547522566618593e-06, "loss": 1.2697, "step": 2770 }, { "epoch": 0.7596496030659732, "grad_norm": 0.4030855832137369, "learning_rate": 3.319117804272236e-06, "loss": 1.3152, "step": 2775 }, { "epoch": 0.7610183410895155, "grad_norm": 0.38853914068753576, "learning_rate": 3.283635923353553e-06, "loss": 1.3125, "step": 2780 }, { "epoch": 0.7623870791130578, "grad_norm": 0.39575307616938343, "learning_rate": 3.248307424206395e-06, "loss": 1.4215, "step": 2785 }, { "epoch": 0.7637558171366, "grad_norm": 0.3903008557955737, "learning_rate": 3.2131331136285717e-06, "loss": 1.3034, "step": 2790 }, { "epoch": 0.7651245551601423, "grad_norm": 0.38760587663102863, "learning_rate": 3.1781137948966754e-06, "loss": 1.291, "step": 2795 }, { "epoch": 0.7664932931836846, "grad_norm": 0.3917779069091076, "learning_rate": 3.1432502677477494e-06, "loss": 1.3318, "step": 2800 }, { "epoch": 0.767862031207227, "grad_norm": 0.3978532288818969, "learning_rate": 3.108543328361017e-06, "loss": 1.3074, "step": 2805 }, { "epoch": 0.7692307692307693, "grad_norm": 0.3931552217963087, "learning_rate": 3.0739937693397113e-06, "loss": 1.3181, "step": 2810 }, { "epoch": 0.7705995072543115, "grad_norm": 0.38132685696899987, "learning_rate": 3.0396023796929597e-06, "loss": 1.3202, "step": 2815 }, { "epoch": 0.7719682452778538, "grad_norm": 0.3895674318092009, "learning_rate": 3.0053699448177687e-06, "loss": 1.2912, "step": 2820 }, { "epoch": 0.7733369833013961, "grad_norm": 0.3923145835781509, "learning_rate": 2.971297246481101e-06, "loss": 1.2874, "step": 2825 }, { "epoch": 0.7747057213249384, "grad_norm": 0.3964424700594031, "learning_rate": 2.937385062802004e-06, "loss": 1.3723, "step": 2830 }, { "epoch": 0.7760744593484807, "grad_norm": 0.39752354179970156, "learning_rate": 2.9036341682338466e-06, "loss": 1.3329, "step": 2835 }, { "epoch": 0.7774431973720229, "grad_norm": 0.39323110376722525, "learning_rate": 2.870045333546644e-06, "loss": 1.2905, "step": 2840 }, { "epoch": 0.7788119353955653, "grad_norm": 0.3924911670687984, "learning_rate": 2.8366193258094355e-06, "loss": 1.292, "step": 2845 }, { "epoch": 0.7801806734191076, "grad_norm": 0.372337059578975, "learning_rate": 2.8033569083727797e-06, "loss": 1.3368, "step": 2850 }, { "epoch": 0.7815494114426499, "grad_norm": 0.386256983183401, "learning_rate": 2.7702588408513276e-06, "loss": 1.3236, "step": 2855 }, { "epoch": 0.7829181494661922, "grad_norm": 0.4151085195929508, "learning_rate": 2.7373258791064572e-06, "loss": 1.3342, "step": 2860 }, { "epoch": 0.7842868874897345, "grad_norm": 0.39410387314037304, "learning_rate": 2.7045587752290224e-06, "loss": 1.28, "step": 2865 }, { "epoch": 0.7856556255132767, "grad_norm": 0.40070665770653285, "learning_rate": 2.6719582775221862e-06, "loss": 1.3336, "step": 2870 }, { "epoch": 0.787024363536819, "grad_norm": 0.3842777541641571, "learning_rate": 2.6395251304843137e-06, "loss": 1.2757, "step": 2875 }, { "epoch": 0.7883931015603614, "grad_norm": 0.40580643578053277, "learning_rate": 2.6072600747919773e-06, "loss": 1.3196, "step": 2880 }, { "epoch": 0.7897618395839037, "grad_norm": 0.388179506322075, "learning_rate": 2.575163847283053e-06, "loss": 1.293, "step": 2885 }, { "epoch": 0.791130577607446, "grad_norm": 0.3732450173203748, "learning_rate": 2.543237180939875e-06, "loss": 1.2751, "step": 2890 }, { "epoch": 0.7924993156309882, "grad_norm": 0.38506878802774175, "learning_rate": 2.5114808048725035e-06, "loss": 1.3083, "step": 2895 }, { "epoch": 0.7938680536545305, "grad_norm": 0.4075860384902042, "learning_rate": 2.479895444302086e-06, "loss": 1.307, "step": 2900 }, { "epoch": 0.7952367916780728, "grad_norm": 0.39287683770161946, "learning_rate": 2.4484818205442763e-06, "loss": 1.3645, "step": 2905 }, { "epoch": 0.7966055297016151, "grad_norm": 0.3882005380231924, "learning_rate": 2.417240650992767e-06, "loss": 1.3512, "step": 2910 }, { "epoch": 0.7979742677251574, "grad_norm": 0.3866431713970654, "learning_rate": 2.3861726491029237e-06, "loss": 1.2793, "step": 2915 }, { "epoch": 0.7993430057486997, "grad_norm": 0.3798968085022952, "learning_rate": 2.355278524375465e-06, "loss": 1.2865, "step": 2920 }, { "epoch": 0.800711743772242, "grad_norm": 0.39453328832547074, "learning_rate": 2.324558982340275e-06, "loss": 1.2576, "step": 2925 }, { "epoch": 0.8020804817957843, "grad_norm": 0.3764906658011197, "learning_rate": 2.2940147245402944e-06, "loss": 1.3167, "step": 2930 }, { "epoch": 0.8034492198193266, "grad_norm": 0.3823354264503374, "learning_rate": 2.2636464485154875e-06, "loss": 1.334, "step": 2935 }, { "epoch": 0.8048179578428689, "grad_norm": 0.3911098766218853, "learning_rate": 2.23345484778692e-06, "loss": 1.3856, "step": 2940 }, { "epoch": 0.8061866958664111, "grad_norm": 0.4024307217930752, "learning_rate": 2.2034406118409178e-06, "loss": 1.3704, "step": 2945 }, { "epoch": 0.8075554338899534, "grad_norm": 0.38312668558103635, "learning_rate": 2.1736044261133305e-06, "loss": 1.32, "step": 2950 }, { "epoch": 0.8089241719134957, "grad_norm": 0.39944235808095113, "learning_rate": 2.1439469719738615e-06, "loss": 1.3348, "step": 2955 }, { "epoch": 0.8102929099370381, "grad_norm": 0.37362368240209753, "learning_rate": 2.1144689267105213e-06, "loss": 1.3058, "step": 2960 }, { "epoch": 0.8116616479605804, "grad_norm": 0.3977426993128578, "learning_rate": 2.0851709635141526e-06, "loss": 1.329, "step": 2965 }, { "epoch": 0.8130303859841227, "grad_norm": 0.390250388707541, "learning_rate": 2.0560537514630595e-06, "loss": 1.3435, "step": 2970 }, { "epoch": 0.8143991240076649, "grad_norm": 0.41070469591969644, "learning_rate": 2.0271179555077357e-06, "loss": 1.3172, "step": 2975 }, { "epoch": 0.8157678620312072, "grad_norm": 0.378348117809942, "learning_rate": 1.998364236455661e-06, "loss": 1.3027, "step": 2980 }, { "epoch": 0.8171366000547495, "grad_norm": 0.380190609171277, "learning_rate": 1.969793250956221e-06, "loss": 1.2577, "step": 2985 }, { "epoch": 0.8185053380782918, "grad_norm": 0.38655656053220244, "learning_rate": 1.9414056514857205e-06, "loss": 1.3137, "step": 2990 }, { "epoch": 0.8198740761018342, "grad_norm": 0.38867016212920535, "learning_rate": 1.913202086332463e-06, "loss": 1.3597, "step": 2995 }, { "epoch": 0.8212428141253764, "grad_norm": 0.38641157633799067, "learning_rate": 1.8851831995819569e-06, "loss": 1.3184, "step": 3000 }, { "epoch": 0.8226115521489187, "grad_norm": 0.39081114681314566, "learning_rate": 1.8573496311022133e-06, "loss": 1.3219, "step": 3005 }, { "epoch": 0.823980290172461, "grad_norm": 0.37733544335013924, "learning_rate": 1.8297020165291158e-06, "loss": 1.2927, "step": 3010 }, { "epoch": 0.8253490281960033, "grad_norm": 0.38748075262103254, "learning_rate": 1.8022409872519197e-06, "loss": 1.3184, "step": 3015 }, { "epoch": 0.8267177662195456, "grad_norm": 0.40138106139790847, "learning_rate": 1.7749671703988226e-06, "loss": 1.3312, "step": 3020 }, { "epoch": 0.8280865042430878, "grad_norm": 0.3876362883781664, "learning_rate": 1.7478811888226555e-06, "loss": 1.3101, "step": 3025 }, { "epoch": 0.8294552422666301, "grad_norm": 0.3936459314036851, "learning_rate": 1.7209836610866426e-06, "loss": 1.3431, "step": 3030 }, { "epoch": 0.8308239802901725, "grad_norm": 0.38764243587502206, "learning_rate": 1.694275201450284e-06, "loss": 1.3386, "step": 3035 }, { "epoch": 0.8321927183137148, "grad_norm": 0.3908568598595201, "learning_rate": 1.6677564198553332e-06, "loss": 1.3342, "step": 3040 }, { "epoch": 0.8335614563372571, "grad_norm": 0.38588972696708185, "learning_rate": 1.6414279219118568e-06, "loss": 1.3527, "step": 3045 }, { "epoch": 0.8349301943607993, "grad_norm": 0.3907806499233958, "learning_rate": 1.6152903088844051e-06, "loss": 1.3104, "step": 3050 }, { "epoch": 0.8362989323843416, "grad_norm": 0.3825824287507951, "learning_rate": 1.5893441776782947e-06, "loss": 1.3062, "step": 3055 }, { "epoch": 0.8376676704078839, "grad_norm": 0.3842728055099776, "learning_rate": 1.5635901208259608e-06, "loss": 1.3581, "step": 3060 }, { "epoch": 0.8390364084314262, "grad_norm": 0.3905623616859072, "learning_rate": 1.5380287264734285e-06, "loss": 1.3148, "step": 3065 }, { "epoch": 0.8404051464549686, "grad_norm": 0.39844460827488454, "learning_rate": 1.5126605783668945e-06, "loss": 1.3074, "step": 3070 }, { "epoch": 0.8417738844785108, "grad_norm": 0.39217288885390583, "learning_rate": 1.4874862558393787e-06, "loss": 1.3171, "step": 3075 }, { "epoch": 0.8431426225020531, "grad_norm": 0.38303308206260517, "learning_rate": 1.462506333797501e-06, "loss": 1.2985, "step": 3080 }, { "epoch": 0.8445113605255954, "grad_norm": 0.37178133903556354, "learning_rate": 1.4377213827083602e-06, "loss": 1.278, "step": 3085 }, { "epoch": 0.8458800985491377, "grad_norm": 0.383873936282195, "learning_rate": 1.413131968586491e-06, "loss": 1.2989, "step": 3090 }, { "epoch": 0.84724883657268, "grad_norm": 0.3877611566994199, "learning_rate": 1.3887386529809454e-06, "loss": 1.3543, "step": 3095 }, { "epoch": 0.8486175745962223, "grad_norm": 0.3844140889489012, "learning_rate": 1.364541992962476e-06, "loss": 1.345, "step": 3100 }, { "epoch": 0.8499863126197645, "grad_norm": 0.3791779816963577, "learning_rate": 1.3405425411108008e-06, "loss": 1.3202, "step": 3105 }, { "epoch": 0.8513550506433069, "grad_norm": 0.37802806800903216, "learning_rate": 1.3167408455019903e-06, "loss": 1.3, "step": 3110 }, { "epoch": 0.8527237886668492, "grad_norm": 0.38421255363489804, "learning_rate": 1.2931374496959548e-06, "loss": 1.3033, "step": 3115 }, { "epoch": 0.8540925266903915, "grad_norm": 0.3890245703911996, "learning_rate": 1.2697328927240238e-06, "loss": 1.3155, "step": 3120 }, { "epoch": 0.8554612647139338, "grad_norm": 0.38965532355017113, "learning_rate": 1.2465277090766381e-06, "loss": 1.3408, "step": 3125 }, { "epoch": 0.856830002737476, "grad_norm": 0.38847581936098935, "learning_rate": 1.2235224286911495e-06, "loss": 1.3619, "step": 3130 }, { "epoch": 0.8581987407610183, "grad_norm": 0.3901578768967818, "learning_rate": 1.2007175769397117e-06, "loss": 1.3714, "step": 3135 }, { "epoch": 0.8595674787845606, "grad_norm": 0.38170228292380465, "learning_rate": 1.178113674617285e-06, "loss": 1.3144, "step": 3140 }, { "epoch": 0.860936216808103, "grad_norm": 0.3852269233134472, "learning_rate": 1.1557112379297385e-06, "loss": 1.3542, "step": 3145 }, { "epoch": 0.8623049548316453, "grad_norm": 0.385512160462019, "learning_rate": 1.1335107784820741e-06, "loss": 1.3556, "step": 3150 }, { "epoch": 0.8636736928551875, "grad_norm": 0.39758307605071247, "learning_rate": 1.1115128032667288e-06, "loss": 1.2992, "step": 3155 }, { "epoch": 0.8650424308787298, "grad_norm": 0.3702170793773961, "learning_rate": 1.0897178146520014e-06, "loss": 1.3861, "step": 3160 }, { "epoch": 0.8664111689022721, "grad_norm": 0.3862887179246134, "learning_rate": 1.0681263103705853e-06, "loss": 1.3317, "step": 3165 }, { "epoch": 0.8677799069258144, "grad_norm": 0.3979382501364706, "learning_rate": 1.0467387835081944e-06, "loss": 1.351, "step": 3170 }, { "epoch": 0.8691486449493567, "grad_norm": 0.3853783562962658, "learning_rate": 1.0255557224923018e-06, "loss": 1.3474, "step": 3175 }, { "epoch": 0.8705173829728989, "grad_norm": 0.3921947417156507, "learning_rate": 1.004577611080998e-06, "loss": 1.3162, "step": 3180 }, { "epoch": 0.8718861209964412, "grad_norm": 0.38229199053031476, "learning_rate": 9.838049283519258e-07, "loss": 1.3265, "step": 3185 }, { "epoch": 0.8732548590199836, "grad_norm": 0.38057187484097027, "learning_rate": 9.63238148691351e-07, "loss": 1.3087, "step": 3190 }, { "epoch": 0.8746235970435259, "grad_norm": 0.3942222628186198, "learning_rate": 9.42877741783328e-07, "loss": 1.366, "step": 3195 }, { "epoch": 0.8759923350670682, "grad_norm": 0.38514993886291565, "learning_rate": 9.227241725989699e-07, "loss": 1.3212, "step": 3200 }, { "epoch": 0.8773610730906105, "grad_norm": 0.3836381492014198, "learning_rate": 9.027779013858284e-07, "loss": 1.2787, "step": 3205 }, { "epoch": 0.8787298111141527, "grad_norm": 0.38734970197899504, "learning_rate": 8.830393836573947e-07, "loss": 1.3387, "step": 3210 }, { "epoch": 0.880098549137695, "grad_norm": 0.3817865137800505, "learning_rate": 8.635090701826799e-07, "loss": 1.3753, "step": 3215 }, { "epoch": 0.8814672871612373, "grad_norm": 0.38200788300985267, "learning_rate": 8.441874069759337e-07, "loss": 1.2776, "step": 3220 }, { "epoch": 0.8828360251847797, "grad_norm": 0.37300363369723033, "learning_rate": 8.250748352864546e-07, "loss": 1.317, "step": 3225 }, { "epoch": 0.884204763208322, "grad_norm": 0.39737199946658025, "learning_rate": 8.061717915885103e-07, "loss": 1.3048, "step": 3230 }, { "epoch": 0.8855735012318642, "grad_norm": 0.3771135467998678, "learning_rate": 7.874787075713742e-07, "loss": 1.2507, "step": 3235 }, { "epoch": 0.8869422392554065, "grad_norm": 0.37723109683472905, "learning_rate": 7.689960101294691e-07, "loss": 1.3081, "step": 3240 }, { "epoch": 0.8883109772789488, "grad_norm": 0.3871833961992763, "learning_rate": 7.507241213526073e-07, "loss": 1.3122, "step": 3245 }, { "epoch": 0.8896797153024911, "grad_norm": 0.3774090621861952, "learning_rate": 7.326634585163617e-07, "loss": 1.3243, "step": 3250 }, { "epoch": 0.8910484533260334, "grad_norm": 0.39061075194420053, "learning_rate": 7.148144340725371e-07, "loss": 1.3123, "step": 3255 }, { "epoch": 0.8924171913495756, "grad_norm": 0.3844913895203542, "learning_rate": 6.971774556397415e-07, "loss": 1.3238, "step": 3260 }, { "epoch": 0.893785929373118, "grad_norm": 0.3954658749141559, "learning_rate": 6.797529259940827e-07, "loss": 1.3421, "step": 3265 }, { "epoch": 0.8951546673966603, "grad_norm": 0.38848733821104914, "learning_rate": 6.625412430599765e-07, "loss": 1.3485, "step": 3270 }, { "epoch": 0.8965234054202026, "grad_norm": 0.38275187348582623, "learning_rate": 6.455427999010466e-07, "loss": 1.3343, "step": 3275 }, { "epoch": 0.8978921434437449, "grad_norm": 0.3946581641011151, "learning_rate": 6.287579847111569e-07, "loss": 1.3362, "step": 3280 }, { "epoch": 0.8992608814672871, "grad_norm": 0.3931302904815353, "learning_rate": 6.121871808055479e-07, "loss": 1.3095, "step": 3285 }, { "epoch": 0.9006296194908294, "grad_norm": 0.39874206377284055, "learning_rate": 5.958307666120733e-07, "loss": 1.2925, "step": 3290 }, { "epoch": 0.9019983575143717, "grad_norm": 0.38856468686205425, "learning_rate": 5.796891156625639e-07, "loss": 1.2878, "step": 3295 }, { "epoch": 0.903367095537914, "grad_norm": 0.3830311422743617, "learning_rate": 5.637625965843041e-07, "loss": 1.3247, "step": 3300 }, { "epoch": 0.9047358335614564, "grad_norm": 0.39286283244773124, "learning_rate": 5.480515730915992e-07, "loss": 1.2902, "step": 3305 }, { "epoch": 0.9061045715849987, "grad_norm": 0.37864973007608094, "learning_rate": 5.325564039774777e-07, "loss": 1.308, "step": 3310 }, { "epoch": 0.9074733096085409, "grad_norm": 0.3932055048918823, "learning_rate": 5.172774431054995e-07, "loss": 1.3245, "step": 3315 }, { "epoch": 0.9088420476320832, "grad_norm": 0.3861464572512248, "learning_rate": 5.022150394016701e-07, "loss": 1.3345, "step": 3320 }, { "epoch": 0.9102107856556255, "grad_norm": 0.3849211658260205, "learning_rate": 4.873695368464693e-07, "loss": 1.3522, "step": 3325 }, { "epoch": 0.9115795236791678, "grad_norm": 0.3850036124623499, "learning_rate": 4.72741274467009e-07, "loss": 1.3205, "step": 3330 }, { "epoch": 0.9129482617027102, "grad_norm": 0.3858389082786252, "learning_rate": 4.5833058632927417e-07, "loss": 1.2984, "step": 3335 }, { "epoch": 0.9143169997262524, "grad_norm": 0.3974200110164499, "learning_rate": 4.441378015305031e-07, "loss": 1.2907, "step": 3340 }, { "epoch": 0.9156857377497947, "grad_norm": 0.3885026941845055, "learning_rate": 4.3016324419167365e-07, "loss": 1.3571, "step": 3345 }, { "epoch": 0.917054475773337, "grad_norm": 0.391701334822345, "learning_rate": 4.164072334500935e-07, "loss": 1.2946, "step": 3350 }, { "epoch": 0.9184232137968793, "grad_norm": 0.371298260058105, "learning_rate": 4.028700834521193e-07, "loss": 1.2734, "step": 3355 }, { "epoch": 0.9197919518204216, "grad_norm": 0.39257123679148415, "learning_rate": 3.8955210334597595e-07, "loss": 1.3792, "step": 3360 }, { "epoch": 0.9211606898439638, "grad_norm": 0.37739382436504915, "learning_rate": 3.764535972747052e-07, "loss": 1.3182, "step": 3365 }, { "epoch": 0.9225294278675061, "grad_norm": 0.4040354445639437, "learning_rate": 3.6357486436921164e-07, "loss": 1.3149, "step": 3370 }, { "epoch": 0.9238981658910485, "grad_norm": 0.3885892463097669, "learning_rate": 3.5091619874143446e-07, "loss": 1.3612, "step": 3375 }, { "epoch": 0.9252669039145908, "grad_norm": 0.39478973344601664, "learning_rate": 3.3847788947763194e-07, "loss": 1.3338, "step": 3380 }, { "epoch": 0.9266356419381331, "grad_norm": 0.4002415191432085, "learning_rate": 3.2626022063177997e-07, "loss": 1.3854, "step": 3385 }, { "epoch": 0.9280043799616753, "grad_norm": 0.3871640152525417, "learning_rate": 3.142634712190795e-07, "loss": 1.2663, "step": 3390 }, { "epoch": 0.9293731179852176, "grad_norm": 0.3844217189319342, "learning_rate": 3.0248791520959387e-07, "loss": 1.3304, "step": 3395 }, { "epoch": 0.9307418560087599, "grad_norm": 0.40010784379733844, "learning_rate": 2.909338215219859e-07, "loss": 1.3458, "step": 3400 }, { "epoch": 0.9321105940323022, "grad_norm": 0.38274676057072776, "learning_rate": 2.7960145401737415e-07, "loss": 1.2606, "step": 3405 }, { "epoch": 0.9334793320558445, "grad_norm": 0.39275920250834023, "learning_rate": 2.6849107149331756e-07, "loss": 1.2825, "step": 3410 }, { "epoch": 0.9348480700793868, "grad_norm": 0.3947633255379109, "learning_rate": 2.576029276778924e-07, "loss": 1.3441, "step": 3415 }, { "epoch": 0.9362168081029291, "grad_norm": 0.41407416506322803, "learning_rate": 2.4693727122390597e-07, "loss": 1.371, "step": 3420 }, { "epoch": 0.9375855461264714, "grad_norm": 0.37625661449174036, "learning_rate": 2.3649434570321984e-07, "loss": 1.2862, "step": 3425 }, { "epoch": 0.9389542841500137, "grad_norm": 0.37104631834710733, "learning_rate": 2.2627438960117876e-07, "loss": 1.2833, "step": 3430 }, { "epoch": 0.940323022173556, "grad_norm": 0.3868409112001441, "learning_rate": 2.1627763631117182e-07, "loss": 1.3551, "step": 3435 }, { "epoch": 0.9416917601970983, "grad_norm": 0.4061569745095073, "learning_rate": 2.0650431412930104e-07, "loss": 1.3273, "step": 3440 }, { "epoch": 0.9430604982206405, "grad_norm": 0.3923199404039606, "learning_rate": 1.969546462491634e-07, "loss": 1.3093, "step": 3445 }, { "epoch": 0.9444292362441828, "grad_norm": 0.40035298460473323, "learning_rate": 1.876288507567592e-07, "loss": 1.2859, "step": 3450 }, { "epoch": 0.9457979742677252, "grad_norm": 0.39270377743419116, "learning_rate": 1.785271406255107e-07, "loss": 1.3086, "step": 3455 }, { "epoch": 0.9471667122912675, "grad_norm": 0.391272234134139, "learning_rate": 1.6964972371139588e-07, "loss": 1.3324, "step": 3460 }, { "epoch": 0.9485354503148098, "grad_norm": 0.3971408419550116, "learning_rate": 1.609968027482012e-07, "loss": 1.3241, "step": 3465 }, { "epoch": 0.949904188338352, "grad_norm": 0.397310674294107, "learning_rate": 1.5256857534289626e-07, "loss": 1.344, "step": 3470 }, { "epoch": 0.9512729263618943, "grad_norm": 0.37737586107823284, "learning_rate": 1.443652339711199e-07, "loss": 1.3227, "step": 3475 }, { "epoch": 0.9526416643854366, "grad_norm": 0.3953334287452581, "learning_rate": 1.3638696597277678e-07, "loss": 1.3323, "step": 3480 }, { "epoch": 0.9540104024089789, "grad_norm": 0.39727863290285664, "learning_rate": 1.2863395354777097e-07, "loss": 1.2965, "step": 3485 }, { "epoch": 0.9553791404325213, "grad_norm": 0.3915026545665889, "learning_rate": 1.211063737518392e-07, "loss": 1.2945, "step": 3490 }, { "epoch": 0.9567478784560635, "grad_norm": 0.3963237143375255, "learning_rate": 1.1380439849250414e-07, "loss": 1.3079, "step": 3495 }, { "epoch": 0.9581166164796058, "grad_norm": 0.38628531323230814, "learning_rate": 1.0672819452515526e-07, "loss": 1.347, "step": 3500 }, { "epoch": 0.9594853545031481, "grad_norm": 0.39250735459020125, "learning_rate": 9.987792344923753e-08, "loss": 1.3292, "step": 3505 }, { "epoch": 0.9608540925266904, "grad_norm": 0.39214485753203543, "learning_rate": 9.32537417045576e-08, "loss": 1.2703, "step": 3510 }, { "epoch": 0.9622228305502327, "grad_norm": 0.38156832143297204, "learning_rate": 8.685580056771781e-08, "loss": 1.3404, "step": 3515 }, { "epoch": 0.9635915685737749, "grad_norm": 0.3907257773875925, "learning_rate": 8.0684246148659e-08, "loss": 1.2681, "step": 3520 }, { "epoch": 0.9649603065973172, "grad_norm": 0.4194193748219591, "learning_rate": 7.473921938731865e-08, "loss": 1.382, "step": 3525 }, { "epoch": 0.9663290446208596, "grad_norm": 0.3858101635490962, "learning_rate": 6.902085605042019e-08, "loss": 1.3671, "step": 3530 }, { "epoch": 0.9676977826444019, "grad_norm": 0.3813102160301827, "learning_rate": 6.352928672836767e-08, "loss": 1.3013, "step": 3535 }, { "epoch": 0.9690665206679442, "grad_norm": 0.38117598807083153, "learning_rate": 5.82646368322648e-08, "loss": 1.3406, "step": 3540 }, { "epoch": 0.9704352586914865, "grad_norm": 0.3869856138402919, "learning_rate": 5.3227026591049505e-08, "loss": 1.3311, "step": 3545 }, { "epoch": 0.9718039967150287, "grad_norm": 0.3866858707608126, "learning_rate": 4.841657104875275e-08, "loss": 1.3593, "step": 3550 }, { "epoch": 0.973172734738571, "grad_norm": 0.3866568363141188, "learning_rate": 4.3833380061865104e-08, "loss": 1.3318, "step": 3555 }, { "epoch": 0.9745414727621133, "grad_norm": 0.39553603732869785, "learning_rate": 3.947755829683097e-08, "loss": 1.3403, "step": 3560 }, { "epoch": 0.9759102107856557, "grad_norm": 0.3990993940103786, "learning_rate": 3.5349205227660496e-08, "loss": 1.3812, "step": 3565 }, { "epoch": 0.977278948809198, "grad_norm": 0.3925614431869845, "learning_rate": 3.144841513365249e-08, "loss": 1.3025, "step": 3570 }, { "epoch": 0.9786476868327402, "grad_norm": 0.3883373657835178, "learning_rate": 2.7775277097247255e-08, "loss": 1.313, "step": 3575 }, { "epoch": 0.9800164248562825, "grad_norm": 0.38572501548529475, "learning_rate": 2.4329875001989356e-08, "loss": 1.3058, "step": 3580 }, { "epoch": 0.9813851628798248, "grad_norm": 0.3838975342711645, "learning_rate": 2.1112287530609122e-08, "loss": 1.3165, "step": 3585 }, { "epoch": 0.9827539009033671, "grad_norm": 0.3918735272830008, "learning_rate": 1.812258816323187e-08, "loss": 1.3388, "step": 3590 }, { "epoch": 0.9841226389269094, "grad_norm": 0.38644635827545804, "learning_rate": 1.5360845175695916e-08, "loss": 1.3378, "step": 3595 }, { "epoch": 0.9854913769504516, "grad_norm": 0.4046028492421605, "learning_rate": 1.2827121637992712e-08, "loss": 1.3104, "step": 3600 }, { "epoch": 0.986860114973994, "grad_norm": 0.37588594630122674, "learning_rate": 1.0521475412830218e-08, "loss": 1.3345, "step": 3605 }, { "epoch": 0.9882288529975363, "grad_norm": 0.39556552792879707, "learning_rate": 8.44395915430729e-09, "loss": 1.3184, "step": 3610 }, { "epoch": 0.9895975910210786, "grad_norm": 0.3745912876489376, "learning_rate": 6.5946203067135395e-09, "loss": 1.3036, "step": 3615 }, { "epoch": 0.9909663290446209, "grad_norm": 0.39112472329549625, "learning_rate": 4.9735011034457434e-09, "loss": 1.2682, "step": 3620 }, { "epoch": 0.9923350670681631, "grad_norm": 0.380307545731869, "learning_rate": 3.580638566043071e-09, "loss": 1.2837, "step": 3625 }, { "epoch": 0.9937038050917054, "grad_norm": 0.3955743424810463, "learning_rate": 2.416064503342197e-09, "loss": 1.3092, "step": 3630 }, { "epoch": 0.9950725431152477, "grad_norm": 0.3853245568354318, "learning_rate": 1.4798055107489996e-09, "loss": 1.3025, "step": 3635 }, { "epoch": 0.99644128113879, "grad_norm": 0.39071516823719027, "learning_rate": 7.718829696334862e-10, "loss": 1.3332, "step": 3640 }, { "epoch": 0.9978100191623324, "grad_norm": 0.38346889130554324, "learning_rate": 2.9231304683907667e-10, "loss": 1.3369, "step": 3645 }, { "epoch": 0.9991787571858747, "grad_norm": 0.4139215786504167, "learning_rate": 4.1106694317338826e-11, "loss": 1.3518, "step": 3650 }, { "epoch": 1.0, "eval_loss": 1.320330262184143, "eval_runtime": 951.5767, "eval_samples_per_second": 92.065, "eval_steps_per_second": 5.755, "step": 3653 }, { "epoch": 1.0, "step": 3653, "total_flos": 66190143651840.0, "train_loss": 1.3573657579367173, "train_runtime": 8062.8425, "train_samples_per_second": 7.249, "train_steps_per_second": 0.453 } ], "logging_steps": 5, "max_steps": 3653, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 66190143651840.0, "train_batch_size": 4, "trial_name": null, "trial_params": null }