{ "best_metric": null, "best_model_checkpoint": null, "epoch": 99.2, "eval_steps": 50, "global_step": 6200, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.16, "grad_norm": 0.730820894241333, "learning_rate": 0.00019967741935483872, "loss": 2.5389, "step": 10 }, { "epoch": 0.32, "grad_norm": 0.7758501172065735, "learning_rate": 0.00019935483870967745, "loss": 1.8047, "step": 20 }, { "epoch": 0.48, "grad_norm": 4.077883720397949, "learning_rate": 0.00019906451612903227, "loss": 1.5731, "step": 30 }, { "epoch": 0.64, "grad_norm": 0.27714163064956665, "learning_rate": 0.00019874193548387098, "loss": 1.508, "step": 40 }, { "epoch": 0.8, "grad_norm": 0.1314336508512497, "learning_rate": 0.00019841935483870968, "loss": 1.4705, "step": 50 }, { "epoch": 0.96, "grad_norm": 0.1689644455909729, "learning_rate": 0.0001980967741935484, "loss": 1.4734, "step": 60 }, { "epoch": 1.12, "grad_norm": 0.14989492297172546, "learning_rate": 0.00019777419354838712, "loss": 1.4617, "step": 70 }, { "epoch": 1.28, "grad_norm": 0.6179828643798828, "learning_rate": 0.00019745161290322583, "loss": 1.4413, "step": 80 }, { "epoch": 1.44, "grad_norm": 2.594003677368164, "learning_rate": 0.0001971290322580645, "loss": 1.353, "step": 90 }, { "epoch": 1.6, "grad_norm": 0.08125967532396317, "learning_rate": 0.00019680645161290324, "loss": 1.3227, "step": 100 }, { "epoch": 1.76, "grad_norm": 0.09471185505390167, "learning_rate": 0.00019648387096774195, "loss": 1.3214, "step": 110 }, { "epoch": 1.92, "grad_norm": 0.10813447833061218, "learning_rate": 0.00019616129032258065, "loss": 1.3236, "step": 120 }, { "epoch": 2.08, "grad_norm": 0.07840854674577713, "learning_rate": 0.0001958387096774194, "loss": 1.3222, "step": 130 }, { "epoch": 2.24, "grad_norm": 0.08169138431549072, "learning_rate": 0.00019551612903225807, "loss": 1.3196, "step": 140 }, { "epoch": 2.4, "grad_norm": 0.15425510704517365, "learning_rate": 0.00019519354838709677, "loss": 1.3191, "step": 150 }, { "epoch": 2.56, "grad_norm": 0.08724704384803772, "learning_rate": 0.0001948709677419355, "loss": 1.3223, "step": 160 }, { "epoch": 2.7199999999999998, "grad_norm": 0.1369275003671646, "learning_rate": 0.0001945483870967742, "loss": 1.3209, "step": 170 }, { "epoch": 2.88, "grad_norm": 0.12385846674442291, "learning_rate": 0.0001942258064516129, "loss": 1.3211, "step": 180 }, { "epoch": 3.04, "grad_norm": 0.1191711276769638, "learning_rate": 0.00019390322580645162, "loss": 1.3225, "step": 190 }, { "epoch": 3.2, "grad_norm": 0.0959768146276474, "learning_rate": 0.00019358064516129033, "loss": 1.3192, "step": 200 }, { "epoch": 3.36, "grad_norm": 0.08907590806484222, "learning_rate": 0.00019325806451612904, "loss": 1.319, "step": 210 }, { "epoch": 3.52, "grad_norm": 0.1829214245080948, "learning_rate": 0.00019293548387096777, "loss": 1.3198, "step": 220 }, { "epoch": 3.68, "grad_norm": 0.10891153663396835, "learning_rate": 0.00019261290322580645, "loss": 1.32, "step": 230 }, { "epoch": 3.84, "grad_norm": 0.10339193791151047, "learning_rate": 0.00019229032258064516, "loss": 1.3191, "step": 240 }, { "epoch": 4.0, "grad_norm": 0.0916794016957283, "learning_rate": 0.0001919677419354839, "loss": 1.3223, "step": 250 }, { "epoch": 4.16, "grad_norm": 0.15707476437091827, "learning_rate": 0.0001916451612903226, "loss": 1.3196, "step": 260 }, { "epoch": 4.32, "grad_norm": 0.11287837475538254, "learning_rate": 0.0001913225806451613, "loss": 1.3168, "step": 270 }, { "epoch": 4.48, "grad_norm": 0.19220077991485596, "learning_rate": 0.000191, "loss": 1.3164, "step": 280 }, { "epoch": 4.64, "grad_norm": 0.08514147996902466, "learning_rate": 0.00019067741935483871, "loss": 1.3193, "step": 290 }, { "epoch": 4.8, "grad_norm": 0.12788636982440948, "learning_rate": 0.00019035483870967742, "loss": 1.3194, "step": 300 }, { "epoch": 4.96, "grad_norm": 0.14810492098331451, "learning_rate": 0.00019003225806451615, "loss": 1.3193, "step": 310 }, { "epoch": 5.12, "grad_norm": 0.16166192293167114, "learning_rate": 0.00018970967741935486, "loss": 1.3183, "step": 320 }, { "epoch": 5.28, "grad_norm": 0.23055078089237213, "learning_rate": 0.00018938709677419354, "loss": 1.3156, "step": 330 }, { "epoch": 5.44, "grad_norm": 0.12718935310840607, "learning_rate": 0.00018906451612903227, "loss": 1.3163, "step": 340 }, { "epoch": 5.6, "grad_norm": 0.20426005125045776, "learning_rate": 0.00018874193548387098, "loss": 1.3157, "step": 350 }, { "epoch": 5.76, "grad_norm": 0.11526554077863693, "learning_rate": 0.00018841935483870968, "loss": 1.3161, "step": 360 }, { "epoch": 5.92, "grad_norm": 0.1523907333612442, "learning_rate": 0.0001880967741935484, "loss": 1.3176, "step": 370 }, { "epoch": 6.08, "grad_norm": 0.2060098499059677, "learning_rate": 0.0001877741935483871, "loss": 1.3142, "step": 380 }, { "epoch": 6.24, "grad_norm": 0.24104587733745575, "learning_rate": 0.0001874516129032258, "loss": 1.3123, "step": 390 }, { "epoch": 6.4, "grad_norm": 0.17447544634342194, "learning_rate": 0.00018712903225806454, "loss": 1.31, "step": 400 }, { "epoch": 6.5600000000000005, "grad_norm": 0.1756139099597931, "learning_rate": 0.00018680645161290324, "loss": 1.3142, "step": 410 }, { "epoch": 6.72, "grad_norm": 0.17624247074127197, "learning_rate": 0.00018648387096774195, "loss": 1.3157, "step": 420 }, { "epoch": 6.88, "grad_norm": 0.1455477774143219, "learning_rate": 0.00018616129032258065, "loss": 1.3155, "step": 430 }, { "epoch": 7.04, "grad_norm": 0.21999657154083252, "learning_rate": 0.00018583870967741936, "loss": 1.3143, "step": 440 }, { "epoch": 7.2, "grad_norm": 0.24742542207241058, "learning_rate": 0.00018551612903225807, "loss": 1.3003, "step": 450 }, { "epoch": 7.36, "grad_norm": 0.2892099618911743, "learning_rate": 0.0001851935483870968, "loss": 1.3064, "step": 460 }, { "epoch": 7.52, "grad_norm": 0.2647784650325775, "learning_rate": 0.00018487096774193548, "loss": 1.3058, "step": 470 }, { "epoch": 7.68, "grad_norm": 0.2782827913761139, "learning_rate": 0.0001845483870967742, "loss": 1.3097, "step": 480 }, { "epoch": 7.84, "grad_norm": 0.17432355880737305, "learning_rate": 0.00018422580645161292, "loss": 1.3141, "step": 490 }, { "epoch": 8.0, "grad_norm": 0.20223839581012726, "learning_rate": 0.00018390322580645163, "loss": 1.3125, "step": 500 }, { "epoch": 8.16, "grad_norm": 0.39165350794792175, "learning_rate": 0.00018358064516129033, "loss": 1.2921, "step": 510 }, { "epoch": 8.32, "grad_norm": 0.35312631726264954, "learning_rate": 0.00018325806451612904, "loss": 1.2952, "step": 520 }, { "epoch": 8.48, "grad_norm": 0.34015020728111267, "learning_rate": 0.00018293548387096774, "loss": 1.297, "step": 530 }, { "epoch": 8.64, "grad_norm": 0.3401159346103668, "learning_rate": 0.00018261290322580648, "loss": 1.298, "step": 540 }, { "epoch": 8.8, "grad_norm": 0.3697248697280884, "learning_rate": 0.00018229032258064518, "loss": 1.3025, "step": 550 }, { "epoch": 8.96, "grad_norm": 0.31431111693382263, "learning_rate": 0.00018196774193548386, "loss": 1.3042, "step": 560 }, { "epoch": 9.12, "grad_norm": 0.5856003761291504, "learning_rate": 0.0001816451612903226, "loss": 1.2838, "step": 570 }, { "epoch": 9.28, "grad_norm": 0.5546128749847412, "learning_rate": 0.0001813225806451613, "loss": 1.2766, "step": 580 }, { "epoch": 9.44, "grad_norm": 0.5272495150566101, "learning_rate": 0.000181, "loss": 1.2825, "step": 590 }, { "epoch": 9.6, "grad_norm": 0.37607982754707336, "learning_rate": 0.00018067741935483874, "loss": 1.2855, "step": 600 }, { "epoch": 9.76, "grad_norm": 0.4203632175922394, "learning_rate": 0.00018035483870967742, "loss": 1.2905, "step": 610 }, { "epoch": 9.92, "grad_norm": 0.5385815501213074, "learning_rate": 0.00018003225806451613, "loss": 1.2862, "step": 620 }, { "epoch": 10.08, "grad_norm": 0.6651461124420166, "learning_rate": 0.00017970967741935486, "loss": 1.2753, "step": 630 }, { "epoch": 10.24, "grad_norm": 0.6366870999336243, "learning_rate": 0.00017938709677419357, "loss": 1.2522, "step": 640 }, { "epoch": 10.4, "grad_norm": 0.662529468536377, "learning_rate": 0.00017906451612903225, "loss": 1.2568, "step": 650 }, { "epoch": 10.56, "grad_norm": 0.569629967212677, "learning_rate": 0.00017874193548387098, "loss": 1.2596, "step": 660 }, { "epoch": 10.72, "grad_norm": 0.5770915150642395, "learning_rate": 0.00017841935483870969, "loss": 1.2699, "step": 670 }, { "epoch": 10.88, "grad_norm": 0.5522739887237549, "learning_rate": 0.0001780967741935484, "loss": 1.2682, "step": 680 }, { "epoch": 11.04, "grad_norm": 0.7272565960884094, "learning_rate": 0.00017777419354838712, "loss": 1.2536, "step": 690 }, { "epoch": 11.2, "grad_norm": 0.9644076824188232, "learning_rate": 0.0001774516129032258, "loss": 1.22, "step": 700 }, { "epoch": 11.36, "grad_norm": 0.7902132272720337, "learning_rate": 0.0001771290322580645, "loss": 1.2189, "step": 710 }, { "epoch": 11.52, "grad_norm": 0.7859879732131958, "learning_rate": 0.00017680645161290324, "loss": 1.231, "step": 720 }, { "epoch": 11.68, "grad_norm": 0.7887389659881592, "learning_rate": 0.00017648387096774195, "loss": 1.2358, "step": 730 }, { "epoch": 11.84, "grad_norm": 0.7551841139793396, "learning_rate": 0.00017616129032258066, "loss": 1.2444, "step": 740 }, { "epoch": 12.0, "grad_norm": 0.6560466885566711, "learning_rate": 0.00017583870967741936, "loss": 1.2326, "step": 750 }, { "epoch": 12.16, "grad_norm": 1.0130267143249512, "learning_rate": 0.00017551612903225807, "loss": 1.167, "step": 760 }, { "epoch": 12.32, "grad_norm": 0.8946852087974548, "learning_rate": 0.00017519354838709677, "loss": 1.1776, "step": 770 }, { "epoch": 12.48, "grad_norm": 1.0538991689682007, "learning_rate": 0.0001748709677419355, "loss": 1.1792, "step": 780 }, { "epoch": 12.64, "grad_norm": 1.0721157789230347, "learning_rate": 0.00017454838709677421, "loss": 1.1915, "step": 790 }, { "epoch": 12.8, "grad_norm": 0.8850572109222412, "learning_rate": 0.0001742258064516129, "loss": 1.193, "step": 800 }, { "epoch": 12.96, "grad_norm": 1.0722297430038452, "learning_rate": 0.00017390322580645163, "loss": 1.1925, "step": 810 }, { "epoch": 13.12, "grad_norm": 1.2234816551208496, "learning_rate": 0.00017358064516129033, "loss": 1.1253, "step": 820 }, { "epoch": 13.28, "grad_norm": 1.289270043373108, "learning_rate": 0.00017325806451612904, "loss": 1.1006, "step": 830 }, { "epoch": 13.44, "grad_norm": 1.1833810806274414, "learning_rate": 0.00017293548387096775, "loss": 1.102, "step": 840 }, { "epoch": 13.6, "grad_norm": 1.2299394607543945, "learning_rate": 0.00017261290322580645, "loss": 1.1233, "step": 850 }, { "epoch": 13.76, "grad_norm": 1.1658111810684204, "learning_rate": 0.00017229032258064516, "loss": 1.1253, "step": 860 }, { "epoch": 13.92, "grad_norm": 1.1958309412002563, "learning_rate": 0.0001719677419354839, "loss": 1.1253, "step": 870 }, { "epoch": 14.08, "grad_norm": 1.418826699256897, "learning_rate": 0.0001716451612903226, "loss": 1.0708, "step": 880 }, { "epoch": 14.24, "grad_norm": 1.5377317667007446, "learning_rate": 0.0001713225806451613, "loss": 1.0089, "step": 890 }, { "epoch": 14.4, "grad_norm": 1.5082684755325317, "learning_rate": 0.000171, "loss": 1.0263, "step": 900 }, { "epoch": 14.56, "grad_norm": 1.5111769437789917, "learning_rate": 0.00017067741935483872, "loss": 1.0347, "step": 910 }, { "epoch": 14.72, "grad_norm": 1.4334949254989624, "learning_rate": 0.00017035483870967742, "loss": 1.037, "step": 920 }, { "epoch": 14.88, "grad_norm": 2.889636278152466, "learning_rate": 0.00017003225806451616, "loss": 1.0348, "step": 930 }, { "epoch": 15.04, "grad_norm": 1.4357563257217407, "learning_rate": 0.00016970967741935483, "loss": 1.0162, "step": 940 }, { "epoch": 15.2, "grad_norm": 1.679137110710144, "learning_rate": 0.00016938709677419357, "loss": 0.9209, "step": 950 }, { "epoch": 15.36, "grad_norm": 1.6809172630310059, "learning_rate": 0.00016906451612903227, "loss": 0.9224, "step": 960 }, { "epoch": 15.52, "grad_norm": 2.0711913108825684, "learning_rate": 0.00016874193548387098, "loss": 0.9374, "step": 970 }, { "epoch": 15.68, "grad_norm": 1.6750158071517944, "learning_rate": 0.00016841935483870969, "loss": 0.9386, "step": 980 }, { "epoch": 15.84, "grad_norm": 1.7292613983154297, "learning_rate": 0.0001680967741935484, "loss": 0.949, "step": 990 }, { "epoch": 16.0, "grad_norm": 1.7566314935684204, "learning_rate": 0.0001677741935483871, "loss": 0.9549, "step": 1000 }, { "epoch": 16.16, "grad_norm": 1.9804630279541016, "learning_rate": 0.00016745161290322583, "loss": 0.8478, "step": 1010 }, { "epoch": 16.32, "grad_norm": 1.8297134637832642, "learning_rate": 0.00016712903225806454, "loss": 0.813, "step": 1020 }, { "epoch": 16.48, "grad_norm": 2.0066633224487305, "learning_rate": 0.00016680645161290322, "loss": 0.835, "step": 1030 }, { "epoch": 16.64, "grad_norm": 1.9635342359542847, "learning_rate": 0.00016648387096774195, "loss": 0.8383, "step": 1040 }, { "epoch": 16.8, "grad_norm": 1.8945978879928589, "learning_rate": 0.00016616129032258066, "loss": 0.8675, "step": 1050 }, { "epoch": 16.96, "grad_norm": 1.917394757270813, "learning_rate": 0.00016583870967741936, "loss": 0.8529, "step": 1060 }, { "epoch": 17.12, "grad_norm": 1.9029055833816528, "learning_rate": 0.00016551612903225807, "loss": 0.7593, "step": 1070 }, { "epoch": 17.28, "grad_norm": 2.5159807205200195, "learning_rate": 0.00016519354838709678, "loss": 0.7418, "step": 1080 }, { "epoch": 17.44, "grad_norm": 1.9958096742630005, "learning_rate": 0.00016487096774193548, "loss": 0.7489, "step": 1090 }, { "epoch": 17.6, "grad_norm": 2.0814130306243896, "learning_rate": 0.00016454838709677421, "loss": 0.7502, "step": 1100 }, { "epoch": 17.76, "grad_norm": 2.4516828060150146, "learning_rate": 0.00016422580645161292, "loss": 0.7607, "step": 1110 }, { "epoch": 17.92, "grad_norm": 2.1293084621429443, "learning_rate": 0.0001639032258064516, "loss": 0.7671, "step": 1120 }, { "epoch": 18.08, "grad_norm": 2.431506872177124, "learning_rate": 0.00016358064516129033, "loss": 0.7117, "step": 1130 }, { "epoch": 18.24, "grad_norm": 2.156888723373413, "learning_rate": 0.00016325806451612904, "loss": 0.6516, "step": 1140 }, { "epoch": 18.4, "grad_norm": 2.404205322265625, "learning_rate": 0.00016293548387096775, "loss": 0.6595, "step": 1150 }, { "epoch": 18.56, "grad_norm": 2.5965211391448975, "learning_rate": 0.00016261290322580648, "loss": 0.6738, "step": 1160 }, { "epoch": 18.72, "grad_norm": 2.4526731967926025, "learning_rate": 0.00016229032258064516, "loss": 0.6581, "step": 1170 }, { "epoch": 18.88, "grad_norm": 2.5058436393737793, "learning_rate": 0.00016196774193548386, "loss": 0.6821, "step": 1180 }, { "epoch": 19.04, "grad_norm": 2.1841073036193848, "learning_rate": 0.0001616451612903226, "loss": 0.6644, "step": 1190 }, { "epoch": 19.2, "grad_norm": 2.2386417388916016, "learning_rate": 0.0001613225806451613, "loss": 0.572, "step": 1200 }, { "epoch": 19.36, "grad_norm": 2.338043689727783, "learning_rate": 0.000161, "loss": 0.5822, "step": 1210 }, { "epoch": 19.52, "grad_norm": 2.37507700920105, "learning_rate": 0.00016067741935483872, "loss": 0.5925, "step": 1220 }, { "epoch": 19.68, "grad_norm": 2.571474552154541, "learning_rate": 0.00016035483870967742, "loss": 0.6041, "step": 1230 }, { "epoch": 19.84, "grad_norm": 2.4885830879211426, "learning_rate": 0.00016003225806451613, "loss": 0.6237, "step": 1240 }, { "epoch": 20.0, "grad_norm": 2.9572579860687256, "learning_rate": 0.00015970967741935486, "loss": 0.6022, "step": 1250 }, { "epoch": 20.16, "grad_norm": 2.3328864574432373, "learning_rate": 0.00015938709677419354, "loss": 0.5042, "step": 1260 }, { "epoch": 20.32, "grad_norm": 2.6092636585235596, "learning_rate": 0.00015906451612903225, "loss": 0.5112, "step": 1270 }, { "epoch": 20.48, "grad_norm": 2.577423572540283, "learning_rate": 0.00015874193548387098, "loss": 0.5242, "step": 1280 }, { "epoch": 20.64, "grad_norm": 2.496828556060791, "learning_rate": 0.0001584193548387097, "loss": 0.5326, "step": 1290 }, { "epoch": 20.8, "grad_norm": 2.479008197784424, "learning_rate": 0.0001580967741935484, "loss": 0.5427, "step": 1300 }, { "epoch": 20.96, "grad_norm": 3.1138076782226562, "learning_rate": 0.0001577741935483871, "loss": 0.5466, "step": 1310 }, { "epoch": 21.12, "grad_norm": 2.6781411170959473, "learning_rate": 0.0001574516129032258, "loss": 0.4764, "step": 1320 }, { "epoch": 21.28, "grad_norm": 3.1711771488189697, "learning_rate": 0.0001571290322580645, "loss": 0.4623, "step": 1330 }, { "epoch": 21.44, "grad_norm": 2.4297163486480713, "learning_rate": 0.00015680645161290325, "loss": 0.4542, "step": 1340 }, { "epoch": 21.6, "grad_norm": 2.8714449405670166, "learning_rate": 0.00015648387096774195, "loss": 0.4752, "step": 1350 }, { "epoch": 21.76, "grad_norm": 2.6193573474884033, "learning_rate": 0.00015616129032258066, "loss": 0.484, "step": 1360 }, { "epoch": 21.92, "grad_norm": 3.0310218334198, "learning_rate": 0.00015583870967741936, "loss": 0.4792, "step": 1370 }, { "epoch": 22.08, "grad_norm": 2.4924027919769287, "learning_rate": 0.00015551612903225807, "loss": 0.4529, "step": 1380 }, { "epoch": 22.24, "grad_norm": 2.0856494903564453, "learning_rate": 0.00015519354838709678, "loss": 0.4077, "step": 1390 }, { "epoch": 22.4, "grad_norm": 2.794166088104248, "learning_rate": 0.0001548709677419355, "loss": 0.4142, "step": 1400 }, { "epoch": 22.56, "grad_norm": 2.5052692890167236, "learning_rate": 0.0001545483870967742, "loss": 0.4209, "step": 1410 }, { "epoch": 22.72, "grad_norm": 2.6941845417022705, "learning_rate": 0.00015422580645161292, "loss": 0.4299, "step": 1420 }, { "epoch": 22.88, "grad_norm": 3.0589232444763184, "learning_rate": 0.00015390322580645163, "loss": 0.4398, "step": 1430 }, { "epoch": 23.04, "grad_norm": 2.647005558013916, "learning_rate": 0.00015358064516129033, "loss": 0.4283, "step": 1440 }, { "epoch": 23.2, "grad_norm": 2.0591695308685303, "learning_rate": 0.00015325806451612904, "loss": 0.3667, "step": 1450 }, { "epoch": 23.36, "grad_norm": 2.557981014251709, "learning_rate": 0.00015293548387096775, "loss": 0.3748, "step": 1460 }, { "epoch": 23.52, "grad_norm": 2.7171032428741455, "learning_rate": 0.00015261290322580645, "loss": 0.3771, "step": 1470 }, { "epoch": 23.68, "grad_norm": 2.8117640018463135, "learning_rate": 0.00015229032258064516, "loss": 0.3857, "step": 1480 }, { "epoch": 23.84, "grad_norm": 2.635938882827759, "learning_rate": 0.0001519677419354839, "loss": 0.3955, "step": 1490 }, { "epoch": 24.0, "grad_norm": 2.530909299850464, "learning_rate": 0.00015164516129032257, "loss": 0.4003, "step": 1500 }, { "epoch": 24.16, "grad_norm": 2.716811418533325, "learning_rate": 0.0001513225806451613, "loss": 0.3257, "step": 1510 }, { "epoch": 24.32, "grad_norm": 2.6447346210479736, "learning_rate": 0.000151, "loss": 0.3342, "step": 1520 }, { "epoch": 24.48, "grad_norm": 2.419239044189453, "learning_rate": 0.00015067741935483872, "loss": 0.3443, "step": 1530 }, { "epoch": 24.64, "grad_norm": 2.9881439208984375, "learning_rate": 0.00015035483870967742, "loss": 0.351, "step": 1540 }, { "epoch": 24.8, "grad_norm": 2.9499566555023193, "learning_rate": 0.00015003225806451613, "loss": 0.3626, "step": 1550 }, { "epoch": 24.96, "grad_norm": 3.1362252235412598, "learning_rate": 0.00014970967741935484, "loss": 0.3572, "step": 1560 }, { "epoch": 25.12, "grad_norm": 2.9108505249023438, "learning_rate": 0.00014938709677419357, "loss": 0.3141, "step": 1570 }, { "epoch": 25.28, "grad_norm": 2.4450314044952393, "learning_rate": 0.00014906451612903228, "loss": 0.3062, "step": 1580 }, { "epoch": 25.44, "grad_norm": 3.3903231620788574, "learning_rate": 0.00014874193548387095, "loss": 0.3137, "step": 1590 }, { "epoch": 25.6, "grad_norm": 2.7393417358398438, "learning_rate": 0.0001484193548387097, "loss": 0.3276, "step": 1600 }, { "epoch": 25.76, "grad_norm": 2.811481475830078, "learning_rate": 0.0001480967741935484, "loss": 0.3337, "step": 1610 }, { "epoch": 25.92, "grad_norm": 2.320997953414917, "learning_rate": 0.0001477741935483871, "loss": 0.3257, "step": 1620 }, { "epoch": 26.08, "grad_norm": 2.6050562858581543, "learning_rate": 0.00014745161290322583, "loss": 0.3058, "step": 1630 }, { "epoch": 26.24, "grad_norm": 2.153953790664673, "learning_rate": 0.0001471290322580645, "loss": 0.2795, "step": 1640 }, { "epoch": 26.4, "grad_norm": 2.5632269382476807, "learning_rate": 0.00014680645161290322, "loss": 0.284, "step": 1650 }, { "epoch": 26.56, "grad_norm": 2.355788230895996, "learning_rate": 0.00014648387096774195, "loss": 0.2955, "step": 1660 }, { "epoch": 26.72, "grad_norm": 2.563262462615967, "learning_rate": 0.00014616129032258066, "loss": 0.2965, "step": 1670 }, { "epoch": 26.88, "grad_norm": 2.3319389820098877, "learning_rate": 0.00014583870967741936, "loss": 0.3023, "step": 1680 }, { "epoch": 27.04, "grad_norm": 1.6385563611984253, "learning_rate": 0.00014551612903225807, "loss": 0.295, "step": 1690 }, { "epoch": 27.2, "grad_norm": 2.8859646320343018, "learning_rate": 0.00014519354838709678, "loss": 0.2576, "step": 1700 }, { "epoch": 27.36, "grad_norm": 2.1979823112487793, "learning_rate": 0.00014487096774193548, "loss": 0.2686, "step": 1710 }, { "epoch": 27.52, "grad_norm": 2.4276225566864014, "learning_rate": 0.00014454838709677422, "loss": 0.2667, "step": 1720 }, { "epoch": 27.68, "grad_norm": 2.81669282913208, "learning_rate": 0.0001442258064516129, "loss": 0.2726, "step": 1730 }, { "epoch": 27.84, "grad_norm": 3.1404976844787598, "learning_rate": 0.0001439032258064516, "loss": 0.2855, "step": 1740 }, { "epoch": 28.0, "grad_norm": 3.198063373565674, "learning_rate": 0.00014358064516129034, "loss": 0.2911, "step": 1750 }, { "epoch": 28.16, "grad_norm": 2.2350854873657227, "learning_rate": 0.00014325806451612904, "loss": 0.2378, "step": 1760 }, { "epoch": 28.32, "grad_norm": 2.4164915084838867, "learning_rate": 0.00014293548387096775, "loss": 0.2397, "step": 1770 }, { "epoch": 28.48, "grad_norm": 2.8860878944396973, "learning_rate": 0.00014261290322580645, "loss": 0.2505, "step": 1780 }, { "epoch": 28.64, "grad_norm": 2.7349557876586914, "learning_rate": 0.00014229032258064516, "loss": 0.262, "step": 1790 }, { "epoch": 28.8, "grad_norm": 2.604602813720703, "learning_rate": 0.00014196774193548387, "loss": 0.2639, "step": 1800 }, { "epoch": 28.96, "grad_norm": 2.5355823040008545, "learning_rate": 0.0001416451612903226, "loss": 0.2666, "step": 1810 }, { "epoch": 29.12, "grad_norm": 2.2154204845428467, "learning_rate": 0.0001413225806451613, "loss": 0.233, "step": 1820 }, { "epoch": 29.28, "grad_norm": 2.8737149238586426, "learning_rate": 0.000141, "loss": 0.2245, "step": 1830 }, { "epoch": 29.44, "grad_norm": 2.359257221221924, "learning_rate": 0.00014067741935483872, "loss": 0.2314, "step": 1840 }, { "epoch": 29.6, "grad_norm": 2.4809696674346924, "learning_rate": 0.00014035483870967742, "loss": 0.24, "step": 1850 }, { "epoch": 29.76, "grad_norm": 2.770080804824829, "learning_rate": 0.00014003225806451613, "loss": 0.2471, "step": 1860 }, { "epoch": 29.92, "grad_norm": 2.3980712890625, "learning_rate": 0.00013970967741935486, "loss": 0.2459, "step": 1870 }, { "epoch": 30.08, "grad_norm": 2.22361421585083, "learning_rate": 0.00013938709677419354, "loss": 0.2293, "step": 1880 }, { "epoch": 30.24, "grad_norm": 2.23351788520813, "learning_rate": 0.00013906451612903228, "loss": 0.2161, "step": 1890 }, { "epoch": 30.4, "grad_norm": 2.324384927749634, "learning_rate": 0.00013874193548387098, "loss": 0.2181, "step": 1900 }, { "epoch": 30.56, "grad_norm": 2.4072265625, "learning_rate": 0.0001384193548387097, "loss": 0.2252, "step": 1910 }, { "epoch": 30.72, "grad_norm": 2.428682565689087, "learning_rate": 0.0001380967741935484, "loss": 0.2332, "step": 1920 }, { "epoch": 30.88, "grad_norm": 2.5889081954956055, "learning_rate": 0.0001377741935483871, "loss": 0.234, "step": 1930 }, { "epoch": 31.04, "grad_norm": 2.178375005722046, "learning_rate": 0.0001374516129032258, "loss": 0.2221, "step": 1940 }, { "epoch": 31.2, "grad_norm": 2.2690091133117676, "learning_rate": 0.00013712903225806451, "loss": 0.1978, "step": 1950 }, { "epoch": 31.36, "grad_norm": 2.445535659790039, "learning_rate": 0.00013680645161290325, "loss": 0.2075, "step": 1960 }, { "epoch": 31.52, "grad_norm": 2.297614097595215, "learning_rate": 0.00013648387096774193, "loss": 0.2105, "step": 1970 }, { "epoch": 31.68, "grad_norm": 2.456094980239868, "learning_rate": 0.00013616129032258066, "loss": 0.2183, "step": 1980 }, { "epoch": 31.84, "grad_norm": 2.9073026180267334, "learning_rate": 0.00013583870967741937, "loss": 0.2219, "step": 1990 }, { "epoch": 32.0, "grad_norm": 2.5645761489868164, "learning_rate": 0.00013551612903225807, "loss": 0.2244, "step": 2000 }, { "epoch": 32.16, "grad_norm": 1.9080466032028198, "learning_rate": 0.00013519354838709678, "loss": 0.1852, "step": 2010 }, { "epoch": 32.32, "grad_norm": 2.4245429039001465, "learning_rate": 0.00013487096774193548, "loss": 0.1941, "step": 2020 }, { "epoch": 32.48, "grad_norm": 2.873189687728882, "learning_rate": 0.0001345483870967742, "loss": 0.198, "step": 2030 }, { "epoch": 32.64, "grad_norm": 3.123337984085083, "learning_rate": 0.00013422580645161292, "loss": 0.203, "step": 2040 }, { "epoch": 32.8, "grad_norm": 2.592850923538208, "learning_rate": 0.00013390322580645163, "loss": 0.2095, "step": 2050 }, { "epoch": 32.96, "grad_norm": 2.4933552742004395, "learning_rate": 0.0001335806451612903, "loss": 0.2053, "step": 2060 }, { "epoch": 33.12, "grad_norm": 1.862947940826416, "learning_rate": 0.00013325806451612904, "loss": 0.1886, "step": 2070 }, { "epoch": 33.28, "grad_norm": 2.4500813484191895, "learning_rate": 0.00013293548387096775, "loss": 0.1834, "step": 2080 }, { "epoch": 33.44, "grad_norm": 2.4213106632232666, "learning_rate": 0.00013261290322580645, "loss": 0.1916, "step": 2090 }, { "epoch": 33.6, "grad_norm": 2.4137024879455566, "learning_rate": 0.0001322903225806452, "loss": 0.1889, "step": 2100 }, { "epoch": 33.76, "grad_norm": 1.9225903749465942, "learning_rate": 0.00013196774193548387, "loss": 0.1932, "step": 2110 }, { "epoch": 33.92, "grad_norm": 2.627751350402832, "learning_rate": 0.00013164516129032257, "loss": 0.1994, "step": 2120 }, { "epoch": 34.08, "grad_norm": 2.1216776371002197, "learning_rate": 0.0001313225806451613, "loss": 0.1854, "step": 2130 }, { "epoch": 34.24, "grad_norm": 1.79988431930542, "learning_rate": 0.000131, "loss": 0.1764, "step": 2140 }, { "epoch": 34.4, "grad_norm": 2.725584030151367, "learning_rate": 0.00013067741935483872, "loss": 0.1724, "step": 2150 }, { "epoch": 34.56, "grad_norm": 3.0872607231140137, "learning_rate": 0.00013035483870967743, "loss": 0.1864, "step": 2160 }, { "epoch": 34.72, "grad_norm": 2.670351982116699, "learning_rate": 0.00013003225806451613, "loss": 0.189, "step": 2170 }, { "epoch": 34.88, "grad_norm": 2.467918872833252, "learning_rate": 0.00012970967741935484, "loss": 0.1892, "step": 2180 }, { "epoch": 35.04, "grad_norm": 1.591225504875183, "learning_rate": 0.00012938709677419357, "loss": 0.1862, "step": 2190 }, { "epoch": 35.2, "grad_norm": 2.1949949264526367, "learning_rate": 0.00012906451612903225, "loss": 0.1594, "step": 2200 }, { "epoch": 35.36, "grad_norm": 2.2106893062591553, "learning_rate": 0.00012874193548387096, "loss": 0.1699, "step": 2210 }, { "epoch": 35.52, "grad_norm": 1.835793375968933, "learning_rate": 0.0001284193548387097, "loss": 0.1699, "step": 2220 }, { "epoch": 35.68, "grad_norm": 1.770975947380066, "learning_rate": 0.0001280967741935484, "loss": 0.1786, "step": 2230 }, { "epoch": 35.84, "grad_norm": 2.7737338542938232, "learning_rate": 0.0001277741935483871, "loss": 0.1799, "step": 2240 }, { "epoch": 36.0, "grad_norm": 2.317680835723877, "learning_rate": 0.0001274516129032258, "loss": 0.1807, "step": 2250 }, { "epoch": 36.16, "grad_norm": 2.137967586517334, "learning_rate": 0.00012712903225806451, "loss": 0.155, "step": 2260 }, { "epoch": 36.32, "grad_norm": 1.764809012413025, "learning_rate": 0.00012680645161290322, "loss": 0.1602, "step": 2270 }, { "epoch": 36.48, "grad_norm": 2.4306418895721436, "learning_rate": 0.00012648387096774195, "loss": 0.1644, "step": 2280 }, { "epoch": 36.64, "grad_norm": 1.7215269804000854, "learning_rate": 0.00012616129032258066, "loss": 0.1675, "step": 2290 }, { "epoch": 36.8, "grad_norm": 2.1169686317443848, "learning_rate": 0.00012583870967741937, "loss": 0.169, "step": 2300 }, { "epoch": 36.96, "grad_norm": 2.576965093612671, "learning_rate": 0.00012551612903225807, "loss": 0.1759, "step": 2310 }, { "epoch": 37.12, "grad_norm": 2.349083185195923, "learning_rate": 0.00012519354838709678, "loss": 0.1569, "step": 2320 }, { "epoch": 37.28, "grad_norm": 1.698199987411499, "learning_rate": 0.00012487096774193549, "loss": 0.1523, "step": 2330 }, { "epoch": 37.44, "grad_norm": 2.2408599853515625, "learning_rate": 0.00012454838709677422, "loss": 0.1579, "step": 2340 }, { "epoch": 37.6, "grad_norm": 1.7537211179733276, "learning_rate": 0.0001242258064516129, "loss": 0.1613, "step": 2350 }, { "epoch": 37.76, "grad_norm": 2.2816600799560547, "learning_rate": 0.0001239032258064516, "loss": 0.1604, "step": 2360 }, { "epoch": 37.92, "grad_norm": 2.135903835296631, "learning_rate": 0.00012358064516129034, "loss": 0.1666, "step": 2370 }, { "epoch": 38.08, "grad_norm": 1.4444613456726074, "learning_rate": 0.00012325806451612904, "loss": 0.1574, "step": 2380 }, { "epoch": 38.24, "grad_norm": 2.1274943351745605, "learning_rate": 0.00012293548387096775, "loss": 0.1497, "step": 2390 }, { "epoch": 38.4, "grad_norm": 2.0582194328308105, "learning_rate": 0.00012261290322580646, "loss": 0.1502, "step": 2400 }, { "epoch": 38.56, "grad_norm": 1.99543035030365, "learning_rate": 0.00012229032258064516, "loss": 0.1527, "step": 2410 }, { "epoch": 38.72, "grad_norm": 2.4016642570495605, "learning_rate": 0.00012196774193548388, "loss": 0.1579, "step": 2420 }, { "epoch": 38.88, "grad_norm": 1.889532446861267, "learning_rate": 0.0001216451612903226, "loss": 0.1581, "step": 2430 }, { "epoch": 39.04, "grad_norm": 1.5487650632858276, "learning_rate": 0.0001213225806451613, "loss": 0.1557, "step": 2440 }, { "epoch": 39.2, "grad_norm": 1.7775896787643433, "learning_rate": 0.000121, "loss": 0.1433, "step": 2450 }, { "epoch": 39.36, "grad_norm": 1.629921555519104, "learning_rate": 0.00012067741935483872, "loss": 0.1477, "step": 2460 }, { "epoch": 39.52, "grad_norm": 1.9216187000274658, "learning_rate": 0.00012035483870967743, "loss": 0.1481, "step": 2470 }, { "epoch": 39.68, "grad_norm": 2.4521870613098145, "learning_rate": 0.00012003225806451615, "loss": 0.1548, "step": 2480 }, { "epoch": 39.84, "grad_norm": 2.893139600753784, "learning_rate": 0.00011970967741935484, "loss": 0.1542, "step": 2490 }, { "epoch": 40.0, "grad_norm": 1.7710884809494019, "learning_rate": 0.00011938709677419356, "loss": 0.1539, "step": 2500 }, { "epoch": 40.16, "grad_norm": 2.467780590057373, "learning_rate": 0.00011906451612903226, "loss": 0.1384, "step": 2510 }, { "epoch": 40.32, "grad_norm": 1.692098617553711, "learning_rate": 0.00011874193548387098, "loss": 0.1398, "step": 2520 }, { "epoch": 40.48, "grad_norm": 2.4998574256896973, "learning_rate": 0.00011841935483870968, "loss": 0.141, "step": 2530 }, { "epoch": 40.64, "grad_norm": 2.455094575881958, "learning_rate": 0.00011809677419354838, "loss": 0.1472, "step": 2540 }, { "epoch": 40.8, "grad_norm": 2.2501373291015625, "learning_rate": 0.0001177741935483871, "loss": 0.1478, "step": 2550 }, { "epoch": 40.96, "grad_norm": 1.7352842092514038, "learning_rate": 0.00011745161290322582, "loss": 0.1523, "step": 2560 }, { "epoch": 41.12, "grad_norm": 1.6160428524017334, "learning_rate": 0.00011712903225806453, "loss": 0.1383, "step": 2570 }, { "epoch": 41.28, "grad_norm": 1.7578778266906738, "learning_rate": 0.00011680645161290322, "loss": 0.1344, "step": 2580 }, { "epoch": 41.44, "grad_norm": 1.668540120124817, "learning_rate": 0.00011648387096774194, "loss": 0.137, "step": 2590 }, { "epoch": 41.6, "grad_norm": 1.9453370571136475, "learning_rate": 0.00011616129032258065, "loss": 0.1445, "step": 2600 }, { "epoch": 41.76, "grad_norm": 2.1329894065856934, "learning_rate": 0.00011583870967741937, "loss": 0.1401, "step": 2610 }, { "epoch": 41.92, "grad_norm": 2.9534623622894287, "learning_rate": 0.00011551612903225807, "loss": 0.1451, "step": 2620 }, { "epoch": 42.08, "grad_norm": 1.3531379699707031, "learning_rate": 0.00011519354838709677, "loss": 0.1375, "step": 2630 }, { "epoch": 42.24, "grad_norm": 2.283677816390991, "learning_rate": 0.00011487096774193549, "loss": 0.1307, "step": 2640 }, { "epoch": 42.4, "grad_norm": 1.6229647397994995, "learning_rate": 0.0001145483870967742, "loss": 0.1313, "step": 2650 }, { "epoch": 42.56, "grad_norm": 1.5721402168273926, "learning_rate": 0.00011422580645161291, "loss": 0.1322, "step": 2660 }, { "epoch": 42.72, "grad_norm": 1.6071618795394897, "learning_rate": 0.0001139032258064516, "loss": 0.1376, "step": 2670 }, { "epoch": 42.88, "grad_norm": 1.4502424001693726, "learning_rate": 0.00011358064516129032, "loss": 0.1368, "step": 2680 }, { "epoch": 43.04, "grad_norm": 1.2102515697479248, "learning_rate": 0.00011325806451612903, "loss": 0.1363, "step": 2690 }, { "epoch": 43.2, "grad_norm": 1.4724687337875366, "learning_rate": 0.00011293548387096775, "loss": 0.1241, "step": 2700 }, { "epoch": 43.36, "grad_norm": 2.126573085784912, "learning_rate": 0.00011261290322580647, "loss": 0.1281, "step": 2710 }, { "epoch": 43.52, "grad_norm": 2.0378658771514893, "learning_rate": 0.00011229032258064516, "loss": 0.1352, "step": 2720 }, { "epoch": 43.68, "grad_norm": 1.978238821029663, "learning_rate": 0.00011196774193548387, "loss": 0.1357, "step": 2730 }, { "epoch": 43.84, "grad_norm": 1.5167338848114014, "learning_rate": 0.00011164516129032259, "loss": 0.1378, "step": 2740 }, { "epoch": 44.0, "grad_norm": 2.6229848861694336, "learning_rate": 0.0001113225806451613, "loss": 0.1368, "step": 2750 }, { "epoch": 44.16, "grad_norm": 2.586956024169922, "learning_rate": 0.00011100000000000001, "loss": 0.1215, "step": 2760 }, { "epoch": 44.32, "grad_norm": 1.5434893369674683, "learning_rate": 0.00011067741935483871, "loss": 0.122, "step": 2770 }, { "epoch": 44.48, "grad_norm": 1.6991748809814453, "learning_rate": 0.00011035483870967743, "loss": 0.1288, "step": 2780 }, { "epoch": 44.64, "grad_norm": 1.6747218370437622, "learning_rate": 0.00011003225806451613, "loss": 0.1298, "step": 2790 }, { "epoch": 44.8, "grad_norm": 1.8977768421173096, "learning_rate": 0.00010970967741935485, "loss": 0.1274, "step": 2800 }, { "epoch": 44.96, "grad_norm": 2.1921539306640625, "learning_rate": 0.00010938709677419355, "loss": 0.1328, "step": 2810 }, { "epoch": 45.12, "grad_norm": 2.201087474822998, "learning_rate": 0.00010906451612903225, "loss": 0.1199, "step": 2820 }, { "epoch": 45.28, "grad_norm": 1.6816660165786743, "learning_rate": 0.00010874193548387097, "loss": 0.1214, "step": 2830 }, { "epoch": 45.44, "grad_norm": 1.0567231178283691, "learning_rate": 0.00010841935483870969, "loss": 0.1234, "step": 2840 }, { "epoch": 45.6, "grad_norm": 1.3555560111999512, "learning_rate": 0.0001080967741935484, "loss": 0.1236, "step": 2850 }, { "epoch": 45.76, "grad_norm": 2.097130537033081, "learning_rate": 0.00010777419354838709, "loss": 0.1275, "step": 2860 }, { "epoch": 45.92, "grad_norm": 2.127514123916626, "learning_rate": 0.00010745161290322581, "loss": 0.1286, "step": 2870 }, { "epoch": 46.08, "grad_norm": 1.4086848497390747, "learning_rate": 0.00010712903225806452, "loss": 0.1203, "step": 2880 }, { "epoch": 46.24, "grad_norm": 1.6010750532150269, "learning_rate": 0.00010680645161290324, "loss": 0.1175, "step": 2890 }, { "epoch": 46.4, "grad_norm": 1.5432184934616089, "learning_rate": 0.00010648387096774196, "loss": 0.1163, "step": 2900 }, { "epoch": 46.56, "grad_norm": 3.0002214908599854, "learning_rate": 0.00010616129032258065, "loss": 0.1257, "step": 2910 }, { "epoch": 46.72, "grad_norm": 1.6969698667526245, "learning_rate": 0.00010583870967741935, "loss": 0.1207, "step": 2920 }, { "epoch": 46.88, "grad_norm": 1.35691237449646, "learning_rate": 0.00010551612903225807, "loss": 0.1257, "step": 2930 }, { "epoch": 47.04, "grad_norm": 1.2951841354370117, "learning_rate": 0.00010519354838709678, "loss": 0.1226, "step": 2940 }, { "epoch": 47.2, "grad_norm": 1.980775237083435, "learning_rate": 0.0001048709677419355, "loss": 0.1121, "step": 2950 }, { "epoch": 47.36, "grad_norm": 1.7462553977966309, "learning_rate": 0.00010454838709677419, "loss": 0.1193, "step": 2960 }, { "epoch": 47.52, "grad_norm": 1.3529144525527954, "learning_rate": 0.00010422580645161291, "loss": 0.115, "step": 2970 }, { "epoch": 47.68, "grad_norm": 1.7577213048934937, "learning_rate": 0.00010390322580645162, "loss": 0.1201, "step": 2980 }, { "epoch": 47.84, "grad_norm": 1.2776310443878174, "learning_rate": 0.00010358064516129034, "loss": 0.1198, "step": 2990 }, { "epoch": 48.0, "grad_norm": 1.3981118202209473, "learning_rate": 0.00010325806451612903, "loss": 0.1204, "step": 3000 }, { "epoch": 48.16, "grad_norm": 1.208207130432129, "learning_rate": 0.00010293548387096774, "loss": 0.1063, "step": 3010 }, { "epoch": 48.32, "grad_norm": 1.3536474704742432, "learning_rate": 0.00010261290322580646, "loss": 0.1123, "step": 3020 }, { "epoch": 48.48, "grad_norm": 1.3631086349487305, "learning_rate": 0.00010229032258064516, "loss": 0.115, "step": 3030 }, { "epoch": 48.64, "grad_norm": 1.2178473472595215, "learning_rate": 0.00010196774193548388, "loss": 0.1143, "step": 3040 }, { "epoch": 48.8, "grad_norm": 2.1066434383392334, "learning_rate": 0.00010164516129032258, "loss": 0.1234, "step": 3050 }, { "epoch": 48.96, "grad_norm": 1.818852186203003, "learning_rate": 0.0001013225806451613, "loss": 0.1242, "step": 3060 }, { "epoch": 49.12, "grad_norm": 1.2515507936477661, "learning_rate": 0.000101, "loss": 0.1118, "step": 3070 }, { "epoch": 49.28, "grad_norm": 1.3625049591064453, "learning_rate": 0.00010067741935483872, "loss": 0.1097, "step": 3080 }, { "epoch": 49.44, "grad_norm": 1.7623980045318604, "learning_rate": 0.00010035483870967743, "loss": 0.1126, "step": 3090 }, { "epoch": 49.6, "grad_norm": 1.3493503332138062, "learning_rate": 0.00010003225806451612, "loss": 0.1139, "step": 3100 }, { "epoch": 49.76, "grad_norm": 1.5169587135314941, "learning_rate": 9.970967741935484e-05, "loss": 0.1154, "step": 3110 }, { "epoch": 49.92, "grad_norm": 2.1966283321380615, "learning_rate": 9.938709677419356e-05, "loss": 0.1171, "step": 3120 }, { "epoch": 50.08, "grad_norm": 1.6833268404006958, "learning_rate": 9.906451612903225e-05, "loss": 0.113, "step": 3130 }, { "epoch": 50.24, "grad_norm": 1.4803876876831055, "learning_rate": 9.874193548387097e-05, "loss": 0.1054, "step": 3140 }, { "epoch": 50.4, "grad_norm": 1.3878694772720337, "learning_rate": 9.841935483870969e-05, "loss": 0.1079, "step": 3150 }, { "epoch": 50.56, "grad_norm": 1.64521062374115, "learning_rate": 9.809677419354838e-05, "loss": 0.1113, "step": 3160 }, { "epoch": 50.72, "grad_norm": 1.4510949850082397, "learning_rate": 9.77741935483871e-05, "loss": 0.1147, "step": 3170 }, { "epoch": 50.88, "grad_norm": 1.4455626010894775, "learning_rate": 9.745161290322581e-05, "loss": 0.1137, "step": 3180 }, { "epoch": 51.04, "grad_norm": 1.1447566747665405, "learning_rate": 9.712903225806452e-05, "loss": 0.1111, "step": 3190 }, { "epoch": 51.2, "grad_norm": 1.1405003070831299, "learning_rate": 9.680645161290322e-05, "loss": 0.1049, "step": 3200 }, { "epoch": 51.36, "grad_norm": 1.3528752326965332, "learning_rate": 9.648387096774194e-05, "loss": 0.1081, "step": 3210 }, { "epoch": 51.52, "grad_norm": 1.6244195699691772, "learning_rate": 9.616129032258065e-05, "loss": 0.1083, "step": 3220 }, { "epoch": 51.68, "grad_norm": 1.4993714094161987, "learning_rate": 9.583870967741936e-05, "loss": 0.1082, "step": 3230 }, { "epoch": 51.84, "grad_norm": 1.1913836002349854, "learning_rate": 9.551612903225808e-05, "loss": 0.1101, "step": 3240 }, { "epoch": 52.0, "grad_norm": 1.444038987159729, "learning_rate": 9.519354838709678e-05, "loss": 0.1136, "step": 3250 }, { "epoch": 52.16, "grad_norm": 1.5950225591659546, "learning_rate": 9.487096774193549e-05, "loss": 0.1015, "step": 3260 }, { "epoch": 52.32, "grad_norm": 1.5138227939605713, "learning_rate": 9.45483870967742e-05, "loss": 0.1044, "step": 3270 }, { "epoch": 52.48, "grad_norm": 1.7190989255905151, "learning_rate": 9.422580645161291e-05, "loss": 0.1049, "step": 3280 }, { "epoch": 52.64, "grad_norm": 2.0404067039489746, "learning_rate": 9.390322580645162e-05, "loss": 0.1084, "step": 3290 }, { "epoch": 52.8, "grad_norm": 0.8778730630874634, "learning_rate": 9.358064516129033e-05, "loss": 0.1089, "step": 3300 }, { "epoch": 52.96, "grad_norm": 1.302156686782837, "learning_rate": 9.325806451612905e-05, "loss": 0.1064, "step": 3310 }, { "epoch": 53.12, "grad_norm": 1.0487626791000366, "learning_rate": 9.293548387096774e-05, "loss": 0.1041, "step": 3320 }, { "epoch": 53.28, "grad_norm": 1.5156755447387695, "learning_rate": 9.261290322580646e-05, "loss": 0.1025, "step": 3330 }, { "epoch": 53.44, "grad_norm": 1.1233017444610596, "learning_rate": 9.229032258064516e-05, "loss": 0.1061, "step": 3340 }, { "epoch": 53.6, "grad_norm": 1.3437849283218384, "learning_rate": 9.196774193548387e-05, "loss": 0.1072, "step": 3350 }, { "epoch": 53.76, "grad_norm": 1.4854118824005127, "learning_rate": 9.164516129032259e-05, "loss": 0.1047, "step": 3360 }, { "epoch": 53.92, "grad_norm": 1.4864035844802856, "learning_rate": 9.13225806451613e-05, "loss": 0.1068, "step": 3370 }, { "epoch": 54.08, "grad_norm": 1.0401661396026611, "learning_rate": 9.1e-05, "loss": 0.1034, "step": 3380 }, { "epoch": 54.24, "grad_norm": 0.8670012354850769, "learning_rate": 9.067741935483871e-05, "loss": 0.0974, "step": 3390 }, { "epoch": 54.4, "grad_norm": 1.1909117698669434, "learning_rate": 9.035483870967743e-05, "loss": 0.1007, "step": 3400 }, { "epoch": 54.56, "grad_norm": 1.3757662773132324, "learning_rate": 9.003225806451614e-05, "loss": 0.1038, "step": 3410 }, { "epoch": 54.72, "grad_norm": 0.9391764402389526, "learning_rate": 8.970967741935484e-05, "loss": 0.1056, "step": 3420 }, { "epoch": 54.88, "grad_norm": 1.0170888900756836, "learning_rate": 8.938709677419356e-05, "loss": 0.1023, "step": 3430 }, { "epoch": 55.04, "grad_norm": 1.0093938112258911, "learning_rate": 8.906451612903227e-05, "loss": 0.1058, "step": 3440 }, { "epoch": 55.2, "grad_norm": 1.8801138401031494, "learning_rate": 8.874193548387097e-05, "loss": 0.099, "step": 3450 }, { "epoch": 55.36, "grad_norm": 1.5623430013656616, "learning_rate": 8.841935483870968e-05, "loss": 0.1046, "step": 3460 }, { "epoch": 55.52, "grad_norm": 1.7789967060089111, "learning_rate": 8.809677419354839e-05, "loss": 0.1038, "step": 3470 }, { "epoch": 55.68, "grad_norm": 1.797473669052124, "learning_rate": 8.777419354838709e-05, "loss": 0.1062, "step": 3480 }, { "epoch": 55.84, "grad_norm": 2.0963029861450195, "learning_rate": 8.745161290322581e-05, "loss": 0.104, "step": 3490 }, { "epoch": 56.0, "grad_norm": 1.711259126663208, "learning_rate": 8.712903225806452e-05, "loss": 0.107, "step": 3500 }, { "epoch": 56.16, "grad_norm": 1.2828420400619507, "learning_rate": 8.680645161290322e-05, "loss": 0.0967, "step": 3510 }, { "epoch": 56.32, "grad_norm": 1.2137057781219482, "learning_rate": 8.648387096774194e-05, "loss": 0.0971, "step": 3520 }, { "epoch": 56.48, "grad_norm": 0.9034223556518555, "learning_rate": 8.616129032258065e-05, "loss": 0.1006, "step": 3530 }, { "epoch": 56.64, "grad_norm": 1.3555456399917603, "learning_rate": 8.583870967741936e-05, "loss": 0.1018, "step": 3540 }, { "epoch": 56.8, "grad_norm": 1.7107219696044922, "learning_rate": 8.551612903225806e-05, "loss": 0.1041, "step": 3550 }, { "epoch": 56.96, "grad_norm": 1.4730465412139893, "learning_rate": 8.519354838709678e-05, "loss": 0.1051, "step": 3560 }, { "epoch": 57.12, "grad_norm": 1.1063456535339355, "learning_rate": 8.487096774193549e-05, "loss": 0.0968, "step": 3570 }, { "epoch": 57.28, "grad_norm": 0.9461372494697571, "learning_rate": 8.45483870967742e-05, "loss": 0.0958, "step": 3580 }, { "epoch": 57.44, "grad_norm": 0.8849135041236877, "learning_rate": 8.422580645161291e-05, "loss": 0.0984, "step": 3590 }, { "epoch": 57.6, "grad_norm": 0.8585776090621948, "learning_rate": 8.390322580645161e-05, "loss": 0.1, "step": 3600 }, { "epoch": 57.76, "grad_norm": 1.3201136589050293, "learning_rate": 8.358064516129033e-05, "loss": 0.1013, "step": 3610 }, { "epoch": 57.92, "grad_norm": 0.9458446502685547, "learning_rate": 8.325806451612905e-05, "loss": 0.1042, "step": 3620 }, { "epoch": 58.08, "grad_norm": 1.1642829179763794, "learning_rate": 8.293548387096774e-05, "loss": 0.0974, "step": 3630 }, { "epoch": 58.24, "grad_norm": 0.7290544509887695, "learning_rate": 8.261290322580646e-05, "loss": 0.0947, "step": 3640 }, { "epoch": 58.4, "grad_norm": 1.4136425256729126, "learning_rate": 8.229032258064517e-05, "loss": 0.0966, "step": 3650 }, { "epoch": 58.56, "grad_norm": 1.6335221529006958, "learning_rate": 8.196774193548387e-05, "loss": 0.0976, "step": 3660 }, { "epoch": 58.72, "grad_norm": 0.7871268391609192, "learning_rate": 8.164516129032258e-05, "loss": 0.0996, "step": 3670 }, { "epoch": 58.88, "grad_norm": 1.2082417011260986, "learning_rate": 8.13225806451613e-05, "loss": 0.1004, "step": 3680 }, { "epoch": 59.04, "grad_norm": 0.9343645572662354, "learning_rate": 8.1e-05, "loss": 0.0991, "step": 3690 }, { "epoch": 59.2, "grad_norm": 0.9338416457176208, "learning_rate": 8.067741935483871e-05, "loss": 0.0937, "step": 3700 }, { "epoch": 59.36, "grad_norm": 0.8510011434555054, "learning_rate": 8.035483870967743e-05, "loss": 0.0956, "step": 3710 }, { "epoch": 59.52, "grad_norm": 0.7901081442832947, "learning_rate": 8.003225806451614e-05, "loss": 0.0961, "step": 3720 }, { "epoch": 59.68, "grad_norm": 0.8202269673347473, "learning_rate": 7.970967741935484e-05, "loss": 0.0971, "step": 3730 }, { "epoch": 59.84, "grad_norm": 0.5913488268852234, "learning_rate": 7.938709677419355e-05, "loss": 0.0986, "step": 3740 }, { "epoch": 60.0, "grad_norm": 0.8020423650741577, "learning_rate": 7.906451612903227e-05, "loss": 0.1004, "step": 3750 }, { "epoch": 60.16, "grad_norm": 1.2948077917099, "learning_rate": 7.874193548387097e-05, "loss": 0.092, "step": 3760 }, { "epoch": 60.32, "grad_norm": 0.8374224305152893, "learning_rate": 7.841935483870968e-05, "loss": 0.0954, "step": 3770 }, { "epoch": 60.48, "grad_norm": 1.1318821907043457, "learning_rate": 7.80967741935484e-05, "loss": 0.0961, "step": 3780 }, { "epoch": 60.64, "grad_norm": 1.1860573291778564, "learning_rate": 7.777419354838709e-05, "loss": 0.0958, "step": 3790 }, { "epoch": 60.8, "grad_norm": 1.1659208536148071, "learning_rate": 7.745161290322581e-05, "loss": 0.0975, "step": 3800 }, { "epoch": 60.96, "grad_norm": 1.3710063695907593, "learning_rate": 7.712903225806452e-05, "loss": 0.0997, "step": 3810 }, { "epoch": 61.12, "grad_norm": 0.7192108035087585, "learning_rate": 7.680645161290323e-05, "loss": 0.0931, "step": 3820 }, { "epoch": 61.28, "grad_norm": 0.9875419735908508, "learning_rate": 7.648387096774194e-05, "loss": 0.0933, "step": 3830 }, { "epoch": 61.44, "grad_norm": 1.3378856182098389, "learning_rate": 7.616129032258065e-05, "loss": 0.0939, "step": 3840 }, { "epoch": 61.6, "grad_norm": 0.9775927662849426, "learning_rate": 7.583870967741936e-05, "loss": 0.0958, "step": 3850 }, { "epoch": 61.76, "grad_norm": 0.815967857837677, "learning_rate": 7.551612903225806e-05, "loss": 0.0971, "step": 3860 }, { "epoch": 61.92, "grad_norm": 0.711142897605896, "learning_rate": 7.519354838709678e-05, "loss": 0.098, "step": 3870 }, { "epoch": 62.08, "grad_norm": 0.757862389087677, "learning_rate": 7.487096774193548e-05, "loss": 0.0948, "step": 3880 }, { "epoch": 62.24, "grad_norm": 0.9134785532951355, "learning_rate": 7.45483870967742e-05, "loss": 0.0913, "step": 3890 }, { "epoch": 62.4, "grad_norm": 1.1586359739303589, "learning_rate": 7.422580645161292e-05, "loss": 0.0933, "step": 3900 }, { "epoch": 62.56, "grad_norm": 0.93504399061203, "learning_rate": 7.390322580645161e-05, "loss": 0.0946, "step": 3910 }, { "epoch": 62.72, "grad_norm": 1.2178099155426025, "learning_rate": 7.358064516129033e-05, "loss": 0.0972, "step": 3920 }, { "epoch": 62.88, "grad_norm": 1.491281270980835, "learning_rate": 7.325806451612903e-05, "loss": 0.095, "step": 3930 }, { "epoch": 63.04, "grad_norm": 0.789310097694397, "learning_rate": 7.293548387096774e-05, "loss": 0.0949, "step": 3940 }, { "epoch": 63.2, "grad_norm": 0.8426668643951416, "learning_rate": 7.261290322580645e-05, "loss": 0.0925, "step": 3950 }, { "epoch": 63.36, "grad_norm": 0.6679518818855286, "learning_rate": 7.229032258064517e-05, "loss": 0.0935, "step": 3960 }, { "epoch": 63.52, "grad_norm": 1.1688344478607178, "learning_rate": 7.196774193548387e-05, "loss": 0.095, "step": 3970 }, { "epoch": 63.68, "grad_norm": 0.9475545883178711, "learning_rate": 7.164516129032258e-05, "loss": 0.0954, "step": 3980 }, { "epoch": 63.84, "grad_norm": 0.7896549701690674, "learning_rate": 7.13225806451613e-05, "loss": 0.0973, "step": 3990 }, { "epoch": 64.0, "grad_norm": 0.9919887185096741, "learning_rate": 7.1e-05, "loss": 0.0984, "step": 4000 }, { "epoch": 64.16, "grad_norm": 0.8883840441703796, "learning_rate": 7.067741935483871e-05, "loss": 0.0909, "step": 4010 }, { "epoch": 64.32, "grad_norm": 1.3879891633987427, "learning_rate": 7.035483870967742e-05, "loss": 0.0924, "step": 4020 }, { "epoch": 64.48, "grad_norm": 0.6254050135612488, "learning_rate": 7.003225806451614e-05, "loss": 0.0937, "step": 4030 }, { "epoch": 64.64, "grad_norm": 0.8522732257843018, "learning_rate": 6.970967741935484e-05, "loss": 0.0944, "step": 4040 }, { "epoch": 64.8, "grad_norm": 0.8993662595748901, "learning_rate": 6.938709677419355e-05, "loss": 0.0946, "step": 4050 }, { "epoch": 64.96, "grad_norm": 0.6210134029388428, "learning_rate": 6.906451612903227e-05, "loss": 0.0961, "step": 4060 }, { "epoch": 65.12, "grad_norm": 1.4237314462661743, "learning_rate": 6.874193548387096e-05, "loss": 0.0915, "step": 4070 }, { "epoch": 65.28, "grad_norm": 0.590337336063385, "learning_rate": 6.841935483870968e-05, "loss": 0.0903, "step": 4080 }, { "epoch": 65.44, "grad_norm": 0.8954355120658875, "learning_rate": 6.809677419354839e-05, "loss": 0.0915, "step": 4090 }, { "epoch": 65.6, "grad_norm": 1.272484302520752, "learning_rate": 6.77741935483871e-05, "loss": 0.0919, "step": 4100 }, { "epoch": 65.76, "grad_norm": 0.5799235105514526, "learning_rate": 6.745161290322581e-05, "loss": 0.0953, "step": 4110 }, { "epoch": 65.92, "grad_norm": 0.5489311814308167, "learning_rate": 6.712903225806452e-05, "loss": 0.0944, "step": 4120 }, { "epoch": 66.08, "grad_norm": 0.454113632440567, "learning_rate": 6.680645161290323e-05, "loss": 0.0925, "step": 4130 }, { "epoch": 66.24, "grad_norm": 0.46685802936553955, "learning_rate": 6.648387096774193e-05, "loss": 0.09, "step": 4140 }, { "epoch": 66.4, "grad_norm": 0.5726438164710999, "learning_rate": 6.616129032258065e-05, "loss": 0.0921, "step": 4150 }, { "epoch": 66.56, "grad_norm": 0.9260037541389465, "learning_rate": 6.583870967741936e-05, "loss": 0.0937, "step": 4160 }, { "epoch": 66.72, "grad_norm": 0.6579046249389648, "learning_rate": 6.551612903225806e-05, "loss": 0.0927, "step": 4170 }, { "epoch": 66.88, "grad_norm": 1.529139518737793, "learning_rate": 6.519354838709678e-05, "loss": 0.0941, "step": 4180 }, { "epoch": 67.04, "grad_norm": 0.49119099974632263, "learning_rate": 6.487096774193549e-05, "loss": 0.0922, "step": 4190 }, { "epoch": 67.2, "grad_norm": 0.5991088151931763, "learning_rate": 6.45483870967742e-05, "loss": 0.0891, "step": 4200 }, { "epoch": 67.36, "grad_norm": 0.5796908140182495, "learning_rate": 6.42258064516129e-05, "loss": 0.0897, "step": 4210 }, { "epoch": 67.52, "grad_norm": 0.4511209726333618, "learning_rate": 6.390322580645162e-05, "loss": 0.0907, "step": 4220 }, { "epoch": 67.68, "grad_norm": 0.35779204964637756, "learning_rate": 6.358064516129033e-05, "loss": 0.0924, "step": 4230 }, { "epoch": 67.84, "grad_norm": 0.5029581189155579, "learning_rate": 6.325806451612903e-05, "loss": 0.093, "step": 4240 }, { "epoch": 68.0, "grad_norm": 0.5110875964164734, "learning_rate": 6.293548387096775e-05, "loss": 0.0937, "step": 4250 }, { "epoch": 68.16, "grad_norm": 0.390669584274292, "learning_rate": 6.261290322580645e-05, "loss": 0.0886, "step": 4260 }, { "epoch": 68.32, "grad_norm": 0.6021801233291626, "learning_rate": 6.229032258064517e-05, "loss": 0.0876, "step": 4270 }, { "epoch": 68.48, "grad_norm": 0.41805922985076904, "learning_rate": 6.196774193548387e-05, "loss": 0.0898, "step": 4280 }, { "epoch": 68.64, "grad_norm": 0.40193480253219604, "learning_rate": 6.164516129032258e-05, "loss": 0.0911, "step": 4290 }, { "epoch": 68.8, "grad_norm": 1.7792600393295288, "learning_rate": 6.13225806451613e-05, "loss": 0.0931, "step": 4300 }, { "epoch": 68.96, "grad_norm": 0.4636439383029938, "learning_rate": 6.1e-05, "loss": 0.0924, "step": 4310 }, { "epoch": 69.12, "grad_norm": 0.8066326975822449, "learning_rate": 6.067741935483872e-05, "loss": 0.0896, "step": 4320 }, { "epoch": 69.28, "grad_norm": 0.5064697265625, "learning_rate": 6.035483870967742e-05, "loss": 0.088, "step": 4330 }, { "epoch": 69.44, "grad_norm": 0.40276482701301575, "learning_rate": 6.003225806451613e-05, "loss": 0.0896, "step": 4340 }, { "epoch": 69.6, "grad_norm": 0.39066290855407715, "learning_rate": 5.970967741935484e-05, "loss": 0.0912, "step": 4350 }, { "epoch": 69.76, "grad_norm": 0.5445528030395508, "learning_rate": 5.938709677419355e-05, "loss": 0.0932, "step": 4360 }, { "epoch": 69.92, "grad_norm": 0.44867417216300964, "learning_rate": 5.906451612903226e-05, "loss": 0.0933, "step": 4370 }, { "epoch": 70.08, "grad_norm": 1.0833595991134644, "learning_rate": 5.874193548387097e-05, "loss": 0.0907, "step": 4380 }, { "epoch": 70.24, "grad_norm": 0.5424122214317322, "learning_rate": 5.841935483870968e-05, "loss": 0.0883, "step": 4390 }, { "epoch": 70.4, "grad_norm": 0.46907031536102295, "learning_rate": 5.809677419354839e-05, "loss": 0.0887, "step": 4400 }, { "epoch": 70.56, "grad_norm": 0.40855908393859863, "learning_rate": 5.77741935483871e-05, "loss": 0.089, "step": 4410 }, { "epoch": 70.72, "grad_norm": 0.37600624561309814, "learning_rate": 5.745161290322581e-05, "loss": 0.0915, "step": 4420 }, { "epoch": 70.88, "grad_norm": 0.6693145036697388, "learning_rate": 5.712903225806452e-05, "loss": 0.0919, "step": 4430 }, { "epoch": 71.04, "grad_norm": 0.4039818346500397, "learning_rate": 5.6806451612903234e-05, "loss": 0.0909, "step": 4440 }, { "epoch": 71.2, "grad_norm": 0.3611462414264679, "learning_rate": 5.648387096774193e-05, "loss": 0.0874, "step": 4450 }, { "epoch": 71.36, "grad_norm": 0.402322381734848, "learning_rate": 5.616129032258065e-05, "loss": 0.0877, "step": 4460 }, { "epoch": 71.52, "grad_norm": 0.48968443274497986, "learning_rate": 5.583870967741935e-05, "loss": 0.0893, "step": 4470 }, { "epoch": 71.68, "grad_norm": 0.4324292242527008, "learning_rate": 5.5516129032258065e-05, "loss": 0.0898, "step": 4480 }, { "epoch": 71.84, "grad_norm": 0.44097113609313965, "learning_rate": 5.519354838709677e-05, "loss": 0.0903, "step": 4490 }, { "epoch": 72.0, "grad_norm": 0.38512125611305237, "learning_rate": 5.4870967741935484e-05, "loss": 0.0921, "step": 4500 }, { "epoch": 72.16, "grad_norm": 0.37481260299682617, "learning_rate": 5.45483870967742e-05, "loss": 0.0867, "step": 4510 }, { "epoch": 72.32, "grad_norm": 0.4293956160545349, "learning_rate": 5.4225806451612904e-05, "loss": 0.0876, "step": 4520 }, { "epoch": 72.48, "grad_norm": 0.4083200991153717, "learning_rate": 5.390322580645162e-05, "loss": 0.0886, "step": 4530 }, { "epoch": 72.64, "grad_norm": 0.3753615617752075, "learning_rate": 5.358064516129032e-05, "loss": 0.09, "step": 4540 }, { "epoch": 72.8, "grad_norm": 0.3720114529132843, "learning_rate": 5.3258064516129036e-05, "loss": 0.0892, "step": 4550 }, { "epoch": 72.96, "grad_norm": 0.3209419250488281, "learning_rate": 5.293548387096774e-05, "loss": 0.0915, "step": 4560 }, { "epoch": 73.12, "grad_norm": 0.35520559549331665, "learning_rate": 5.2612903225806455e-05, "loss": 0.0871, "step": 4570 }, { "epoch": 73.28, "grad_norm": 0.3513979911804199, "learning_rate": 5.229032258064517e-05, "loss": 0.087, "step": 4580 }, { "epoch": 73.44, "grad_norm": 0.5205033421516418, "learning_rate": 5.1967741935483874e-05, "loss": 0.088, "step": 4590 }, { "epoch": 73.6, "grad_norm": 0.3683389127254486, "learning_rate": 5.164516129032259e-05, "loss": 0.0886, "step": 4600 }, { "epoch": 73.76, "grad_norm": 0.45570969581604004, "learning_rate": 5.132258064516129e-05, "loss": 0.0902, "step": 4610 }, { "epoch": 73.92, "grad_norm": 0.34694406390190125, "learning_rate": 5.1000000000000006e-05, "loss": 0.0909, "step": 4620 }, { "epoch": 74.08, "grad_norm": 0.3391146957874298, "learning_rate": 5.0677419354838706e-05, "loss": 0.0888, "step": 4630 }, { "epoch": 74.24, "grad_norm": 0.45434221625328064, "learning_rate": 5.035483870967742e-05, "loss": 0.0866, "step": 4640 }, { "epoch": 74.4, "grad_norm": 0.4046492576599121, "learning_rate": 5.003225806451614e-05, "loss": 0.0881, "step": 4650 }, { "epoch": 74.56, "grad_norm": 0.36426255106925964, "learning_rate": 4.970967741935484e-05, "loss": 0.0887, "step": 4660 }, { "epoch": 74.72, "grad_norm": 0.38272616267204285, "learning_rate": 4.938709677419355e-05, "loss": 0.0884, "step": 4670 }, { "epoch": 74.88, "grad_norm": 0.4772709012031555, "learning_rate": 4.9064516129032264e-05, "loss": 0.0902, "step": 4680 }, { "epoch": 75.04, "grad_norm": 0.32595178484916687, "learning_rate": 4.874193548387097e-05, "loss": 0.0893, "step": 4690 }, { "epoch": 75.2, "grad_norm": 0.4148579239845276, "learning_rate": 4.8419354838709676e-05, "loss": 0.0856, "step": 4700 }, { "epoch": 75.36, "grad_norm": 0.43145865201950073, "learning_rate": 4.809677419354839e-05, "loss": 0.0866, "step": 4710 }, { "epoch": 75.52, "grad_norm": 0.4075751006603241, "learning_rate": 4.7774193548387096e-05, "loss": 0.0881, "step": 4720 }, { "epoch": 75.68, "grad_norm": 0.33043548464775085, "learning_rate": 4.745161290322581e-05, "loss": 0.0898, "step": 4730 }, { "epoch": 75.84, "grad_norm": 0.36957836151123047, "learning_rate": 4.712903225806452e-05, "loss": 0.0904, "step": 4740 }, { "epoch": 76.0, "grad_norm": 0.3838096857070923, "learning_rate": 4.680645161290323e-05, "loss": 0.0896, "step": 4750 }, { "epoch": 76.16, "grad_norm": 0.42275360226631165, "learning_rate": 4.648387096774194e-05, "loss": 0.0867, "step": 4760 }, { "epoch": 76.32, "grad_norm": 0.5027480721473694, "learning_rate": 4.616129032258065e-05, "loss": 0.0864, "step": 4770 }, { "epoch": 76.48, "grad_norm": 0.36403143405914307, "learning_rate": 4.583870967741935e-05, "loss": 0.088, "step": 4780 }, { "epoch": 76.64, "grad_norm": 0.3124987781047821, "learning_rate": 4.5516129032258066e-05, "loss": 0.0881, "step": 4790 }, { "epoch": 76.8, "grad_norm": 0.38175061345100403, "learning_rate": 4.519354838709678e-05, "loss": 0.0891, "step": 4800 }, { "epoch": 76.96, "grad_norm": 0.41596823930740356, "learning_rate": 4.4870967741935485e-05, "loss": 0.0898, "step": 4810 }, { "epoch": 77.12, "grad_norm": 0.31852710247039795, "learning_rate": 4.45483870967742e-05, "loss": 0.0864, "step": 4820 }, { "epoch": 77.28, "grad_norm": 0.4198415279388428, "learning_rate": 4.4225806451612905e-05, "loss": 0.0857, "step": 4830 }, { "epoch": 77.44, "grad_norm": 0.392374724149704, "learning_rate": 4.390322580645162e-05, "loss": 0.0864, "step": 4840 }, { "epoch": 77.6, "grad_norm": 0.4310876429080963, "learning_rate": 4.3580645161290324e-05, "loss": 0.0893, "step": 4850 }, { "epoch": 77.76, "grad_norm": 0.3877263367176056, "learning_rate": 4.325806451612903e-05, "loss": 0.0888, "step": 4860 }, { "epoch": 77.92, "grad_norm": 0.31759241223335266, "learning_rate": 4.293548387096775e-05, "loss": 0.0902, "step": 4870 }, { "epoch": 78.08, "grad_norm": 0.3832705616950989, "learning_rate": 4.2612903225806456e-05, "loss": 0.0883, "step": 4880 }, { "epoch": 78.24, "grad_norm": 0.3763550817966461, "learning_rate": 4.229032258064516e-05, "loss": 0.086, "step": 4890 }, { "epoch": 78.4, "grad_norm": 0.45972681045532227, "learning_rate": 4.1967741935483875e-05, "loss": 0.087, "step": 4900 }, { "epoch": 78.56, "grad_norm": 0.4792405664920807, "learning_rate": 4.164516129032258e-05, "loss": 0.0886, "step": 4910 }, { "epoch": 78.72, "grad_norm": 0.34303322434425354, "learning_rate": 4.132258064516129e-05, "loss": 0.0884, "step": 4920 }, { "epoch": 78.88, "grad_norm": 0.33531343936920166, "learning_rate": 4.1e-05, "loss": 0.0892, "step": 4930 }, { "epoch": 79.04, "grad_norm": 0.3543599545955658, "learning_rate": 4.0677419354838713e-05, "loss": 0.0891, "step": 4940 }, { "epoch": 79.2, "grad_norm": 0.3615137040615082, "learning_rate": 4.035483870967742e-05, "loss": 0.0853, "step": 4950 }, { "epoch": 79.36, "grad_norm": 0.40525078773498535, "learning_rate": 4.003225806451613e-05, "loss": 0.0864, "step": 4960 }, { "epoch": 79.52, "grad_norm": 0.3445266783237457, "learning_rate": 3.970967741935484e-05, "loss": 0.088, "step": 4970 }, { "epoch": 79.68, "grad_norm": 0.4922199547290802, "learning_rate": 3.938709677419355e-05, "loss": 0.0885, "step": 4980 }, { "epoch": 79.84, "grad_norm": 0.3156628906726837, "learning_rate": 3.906451612903226e-05, "loss": 0.0888, "step": 4990 }, { "epoch": 80.0, "grad_norm": 0.32096874713897705, "learning_rate": 3.8741935483870964e-05, "loss": 0.0889, "step": 5000 }, { "epoch": 80.16, "grad_norm": 0.3683740794658661, "learning_rate": 3.8419354838709684e-05, "loss": 0.0853, "step": 5010 }, { "epoch": 80.32, "grad_norm": 0.43887290358543396, "learning_rate": 3.809677419354839e-05, "loss": 0.0855, "step": 5020 }, { "epoch": 80.48, "grad_norm": 0.4039384424686432, "learning_rate": 3.7774193548387096e-05, "loss": 0.0865, "step": 5030 }, { "epoch": 80.64, "grad_norm": 0.3594481647014618, "learning_rate": 3.745161290322581e-05, "loss": 0.0871, "step": 5040 }, { "epoch": 80.8, "grad_norm": 0.3050839304924011, "learning_rate": 3.7129032258064516e-05, "loss": 0.0892, "step": 5050 }, { "epoch": 80.96, "grad_norm": 0.3346671164035797, "learning_rate": 3.680645161290323e-05, "loss": 0.0898, "step": 5060 }, { "epoch": 81.12, "grad_norm": 0.32323309779167175, "learning_rate": 3.648387096774194e-05, "loss": 0.0871, "step": 5070 }, { "epoch": 81.28, "grad_norm": 0.5844283699989319, "learning_rate": 3.616129032258065e-05, "loss": 0.085, "step": 5080 }, { "epoch": 81.44, "grad_norm": 0.39134910702705383, "learning_rate": 3.583870967741936e-05, "loss": 0.0868, "step": 5090 }, { "epoch": 81.6, "grad_norm": 0.3730677366256714, "learning_rate": 3.551612903225807e-05, "loss": 0.0879, "step": 5100 }, { "epoch": 81.76, "grad_norm": 0.40242692828178406, "learning_rate": 3.519354838709677e-05, "loss": 0.0878, "step": 5110 }, { "epoch": 81.92, "grad_norm": 0.40863659977912903, "learning_rate": 3.4870967741935486e-05, "loss": 0.0887, "step": 5120 }, { "epoch": 82.08, "grad_norm": 0.3297955095767975, "learning_rate": 3.454838709677419e-05, "loss": 0.0871, "step": 5130 }, { "epoch": 82.24, "grad_norm": 0.39270907640457153, "learning_rate": 3.4225806451612905e-05, "loss": 0.0851, "step": 5140 }, { "epoch": 82.4, "grad_norm": 0.35612475872039795, "learning_rate": 3.390322580645162e-05, "loss": 0.0864, "step": 5150 }, { "epoch": 82.56, "grad_norm": 0.419009804725647, "learning_rate": 3.3580645161290325e-05, "loss": 0.0872, "step": 5160 }, { "epoch": 82.72, "grad_norm": 0.44817715883255005, "learning_rate": 3.325806451612903e-05, "loss": 0.0882, "step": 5170 }, { "epoch": 82.88, "grad_norm": 0.36784857511520386, "learning_rate": 3.2935483870967744e-05, "loss": 0.0878, "step": 5180 }, { "epoch": 83.04, "grad_norm": 0.42855191230773926, "learning_rate": 3.261290322580645e-05, "loss": 0.088, "step": 5190 }, { "epoch": 83.2, "grad_norm": 0.34956803917884827, "learning_rate": 3.229032258064516e-05, "loss": 0.0848, "step": 5200 }, { "epoch": 83.36, "grad_norm": 0.3654519021511078, "learning_rate": 3.1967741935483876e-05, "loss": 0.0862, "step": 5210 }, { "epoch": 83.52, "grad_norm": 0.3838617205619812, "learning_rate": 3.164516129032258e-05, "loss": 0.0862, "step": 5220 }, { "epoch": 83.68, "grad_norm": 0.36475858092308044, "learning_rate": 3.1322580645161295e-05, "loss": 0.0876, "step": 5230 }, { "epoch": 83.84, "grad_norm": 0.39168354868888855, "learning_rate": 3.1e-05, "loss": 0.0882, "step": 5240 }, { "epoch": 84.0, "grad_norm": 0.38403555750846863, "learning_rate": 3.067741935483871e-05, "loss": 0.0885, "step": 5250 }, { "epoch": 84.16, "grad_norm": 0.3362351059913635, "learning_rate": 3.035483870967742e-05, "loss": 0.0852, "step": 5260 }, { "epoch": 84.32, "grad_norm": 0.37289589643478394, "learning_rate": 3.0032258064516127e-05, "loss": 0.0852, "step": 5270 }, { "epoch": 84.48, "grad_norm": 0.3794906437397003, "learning_rate": 2.9709677419354843e-05, "loss": 0.0866, "step": 5280 }, { "epoch": 84.64, "grad_norm": 0.46871262788772583, "learning_rate": 2.938709677419355e-05, "loss": 0.0871, "step": 5290 }, { "epoch": 84.8, "grad_norm": 0.3682585060596466, "learning_rate": 2.906451612903226e-05, "loss": 0.0879, "step": 5300 }, { "epoch": 84.96, "grad_norm": 0.36091622710227966, "learning_rate": 2.874193548387097e-05, "loss": 0.0884, "step": 5310 }, { "epoch": 85.12, "grad_norm": 0.3225698173046112, "learning_rate": 2.8419354838709678e-05, "loss": 0.0858, "step": 5320 }, { "epoch": 85.28, "grad_norm": 0.36931437253952026, "learning_rate": 2.8096774193548388e-05, "loss": 0.0853, "step": 5330 }, { "epoch": 85.44, "grad_norm": 0.3807561993598938, "learning_rate": 2.77741935483871e-05, "loss": 0.0859, "step": 5340 }, { "epoch": 85.6, "grad_norm": 0.4048006534576416, "learning_rate": 2.745161290322581e-05, "loss": 0.0865, "step": 5350 }, { "epoch": 85.76, "grad_norm": 0.3060702979564667, "learning_rate": 2.712903225806452e-05, "loss": 0.0874, "step": 5360 }, { "epoch": 85.92, "grad_norm": 0.3515985608100891, "learning_rate": 2.6806451612903226e-05, "loss": 0.088, "step": 5370 }, { "epoch": 86.08, "grad_norm": 0.3089566230773926, "learning_rate": 2.6483870967741936e-05, "loss": 0.0871, "step": 5380 }, { "epoch": 86.24, "grad_norm": 0.42711663246154785, "learning_rate": 2.6161290322580645e-05, "loss": 0.0844, "step": 5390 }, { "epoch": 86.4, "grad_norm": 0.37982869148254395, "learning_rate": 2.5838709677419355e-05, "loss": 0.0861, "step": 5400 }, { "epoch": 86.56, "grad_norm": 0.38406598567962646, "learning_rate": 2.5516129032258068e-05, "loss": 0.0866, "step": 5410 }, { "epoch": 86.72, "grad_norm": 0.37800315022468567, "learning_rate": 2.5193548387096777e-05, "loss": 0.0868, "step": 5420 }, { "epoch": 86.88, "grad_norm": 0.39667201042175293, "learning_rate": 2.4870967741935487e-05, "loss": 0.0874, "step": 5430 }, { "epoch": 87.04, "grad_norm": 0.30831629037857056, "learning_rate": 2.4548387096774193e-05, "loss": 0.087, "step": 5440 }, { "epoch": 87.2, "grad_norm": 0.3226090669631958, "learning_rate": 2.4225806451612903e-05, "loss": 0.0851, "step": 5450 }, { "epoch": 87.36, "grad_norm": 0.3767179250717163, "learning_rate": 2.3903225806451616e-05, "loss": 0.0847, "step": 5460 }, { "epoch": 87.52, "grad_norm": 0.4611583650112152, "learning_rate": 2.3580645161290325e-05, "loss": 0.086, "step": 5470 }, { "epoch": 87.68, "grad_norm": 0.4075451195240021, "learning_rate": 2.325806451612903e-05, "loss": 0.087, "step": 5480 }, { "epoch": 87.84, "grad_norm": 0.3248244822025299, "learning_rate": 2.293548387096774e-05, "loss": 0.0874, "step": 5490 }, { "epoch": 88.0, "grad_norm": 0.3889995813369751, "learning_rate": 2.2612903225806454e-05, "loss": 0.0876, "step": 5500 }, { "epoch": 88.16, "grad_norm": 0.3838357925415039, "learning_rate": 2.229032258064516e-05, "loss": 0.0847, "step": 5510 }, { "epoch": 88.32, "grad_norm": 0.35209789872169495, "learning_rate": 2.196774193548387e-05, "loss": 0.0843, "step": 5520 }, { "epoch": 88.48, "grad_norm": 0.3327479958534241, "learning_rate": 2.1645161290322583e-05, "loss": 0.086, "step": 5530 }, { "epoch": 88.64, "grad_norm": 0.36551928520202637, "learning_rate": 2.1322580645161293e-05, "loss": 0.0863, "step": 5540 }, { "epoch": 88.8, "grad_norm": 0.38019606471061707, "learning_rate": 2.1e-05, "loss": 0.0866, "step": 5550 }, { "epoch": 88.96, "grad_norm": 0.3546133041381836, "learning_rate": 2.0677419354838712e-05, "loss": 0.0877, "step": 5560 }, { "epoch": 89.12, "grad_norm": 0.276177316904068, "learning_rate": 2.035483870967742e-05, "loss": 0.0856, "step": 5570 }, { "epoch": 89.28, "grad_norm": 0.36877840757369995, "learning_rate": 2.003225806451613e-05, "loss": 0.0845, "step": 5580 }, { "epoch": 89.44, "grad_norm": 0.3937029540538788, "learning_rate": 1.9709677419354837e-05, "loss": 0.0853, "step": 5590 }, { "epoch": 89.6, "grad_norm": 0.3682461977005005, "learning_rate": 1.938709677419355e-05, "loss": 0.0861, "step": 5600 }, { "epoch": 89.76, "grad_norm": 0.3939973711967468, "learning_rate": 1.906451612903226e-05, "loss": 0.0863, "step": 5610 }, { "epoch": 89.92, "grad_norm": 0.35440462827682495, "learning_rate": 1.8741935483870966e-05, "loss": 0.0871, "step": 5620 }, { "epoch": 90.08, "grad_norm": 0.3386973738670349, "learning_rate": 1.841935483870968e-05, "loss": 0.0852, "step": 5630 }, { "epoch": 90.24, "grad_norm": 0.359678715467453, "learning_rate": 1.809677419354839e-05, "loss": 0.084, "step": 5640 }, { "epoch": 90.4, "grad_norm": 0.38088712096214294, "learning_rate": 1.7774193548387098e-05, "loss": 0.0851, "step": 5650 }, { "epoch": 90.56, "grad_norm": 0.4327407479286194, "learning_rate": 1.7451612903225808e-05, "loss": 0.0861, "step": 5660 }, { "epoch": 90.72, "grad_norm": 0.3494204580783844, "learning_rate": 1.7129032258064517e-05, "loss": 0.086, "step": 5670 }, { "epoch": 90.88, "grad_norm": 0.3655516803264618, "learning_rate": 1.6806451612903227e-05, "loss": 0.0867, "step": 5680 }, { "epoch": 91.04, "grad_norm": 0.3738172948360443, "learning_rate": 1.6483870967741937e-05, "loss": 0.0867, "step": 5690 }, { "epoch": 91.2, "grad_norm": 0.33062541484832764, "learning_rate": 1.6161290322580646e-05, "loss": 0.0841, "step": 5700 }, { "epoch": 91.36, "grad_norm": 0.3357619345188141, "learning_rate": 1.5838709677419356e-05, "loss": 0.0845, "step": 5710 }, { "epoch": 91.52, "grad_norm": 0.3250825107097626, "learning_rate": 1.5516129032258065e-05, "loss": 0.0858, "step": 5720 }, { "epoch": 91.68, "grad_norm": 0.4149022102355957, "learning_rate": 1.5193548387096777e-05, "loss": 0.0857, "step": 5730 }, { "epoch": 91.84, "grad_norm": 0.3375944495201111, "learning_rate": 1.4870967741935485e-05, "loss": 0.0863, "step": 5740 }, { "epoch": 92.0, "grad_norm": 0.39512765407562256, "learning_rate": 1.4548387096774194e-05, "loss": 0.0867, "step": 5750 }, { "epoch": 92.16, "grad_norm": 0.36112552881240845, "learning_rate": 1.4225806451612905e-05, "loss": 0.0838, "step": 5760 }, { "epoch": 92.32, "grad_norm": 0.3789423704147339, "learning_rate": 1.3903225806451613e-05, "loss": 0.0839, "step": 5770 }, { "epoch": 92.48, "grad_norm": 0.4131874740123749, "learning_rate": 1.3580645161290323e-05, "loss": 0.0845, "step": 5780 }, { "epoch": 92.64, "grad_norm": 0.4238837659358978, "learning_rate": 1.3258064516129033e-05, "loss": 0.0859, "step": 5790 }, { "epoch": 92.8, "grad_norm": 0.4758787751197815, "learning_rate": 1.2935483870967744e-05, "loss": 0.0862, "step": 5800 }, { "epoch": 92.96, "grad_norm": 0.4589889943599701, "learning_rate": 1.2612903225806452e-05, "loss": 0.0867, "step": 5810 }, { "epoch": 93.12, "grad_norm": 0.3451363146305084, "learning_rate": 1.2290322580645163e-05, "loss": 0.0844, "step": 5820 }, { "epoch": 93.28, "grad_norm": 0.3493219316005707, "learning_rate": 1.1967741935483871e-05, "loss": 0.0847, "step": 5830 }, { "epoch": 93.44, "grad_norm": 0.3858239948749542, "learning_rate": 1.1645161290322582e-05, "loss": 0.085, "step": 5840 }, { "epoch": 93.6, "grad_norm": 0.4442474842071533, "learning_rate": 1.132258064516129e-05, "loss": 0.0851, "step": 5850 }, { "epoch": 93.76, "grad_norm": 0.41025981307029724, "learning_rate": 1.1000000000000001e-05, "loss": 0.0858, "step": 5860 }, { "epoch": 93.92, "grad_norm": 0.39134228229522705, "learning_rate": 1.0677419354838711e-05, "loss": 0.0855, "step": 5870 }, { "epoch": 94.08, "grad_norm": 0.39811140298843384, "learning_rate": 1.0354838709677419e-05, "loss": 0.0848, "step": 5880 }, { "epoch": 94.24, "grad_norm": 0.37056440114974976, "learning_rate": 1.003225806451613e-05, "loss": 0.0834, "step": 5890 }, { "epoch": 94.4, "grad_norm": 0.41117194294929504, "learning_rate": 9.709677419354838e-06, "loss": 0.0846, "step": 5900 }, { "epoch": 94.56, "grad_norm": 0.3775732219219208, "learning_rate": 9.38709677419355e-06, "loss": 0.0846, "step": 5910 }, { "epoch": 94.72, "grad_norm": 0.4182074964046478, "learning_rate": 9.064516129032259e-06, "loss": 0.085, "step": 5920 }, { "epoch": 94.88, "grad_norm": 0.3670865595340729, "learning_rate": 8.741935483870969e-06, "loss": 0.0854, "step": 5930 }, { "epoch": 95.04, "grad_norm": 0.3670247793197632, "learning_rate": 8.419354838709678e-06, "loss": 0.0853, "step": 5940 }, { "epoch": 95.2, "grad_norm": 0.3669385313987732, "learning_rate": 8.096774193548388e-06, "loss": 0.0839, "step": 5950 }, { "epoch": 95.36, "grad_norm": 0.41123315691947937, "learning_rate": 7.774193548387097e-06, "loss": 0.0838, "step": 5960 }, { "epoch": 95.52, "grad_norm": 0.4040362238883972, "learning_rate": 7.451612903225806e-06, "loss": 0.0847, "step": 5970 }, { "epoch": 95.68, "grad_norm": 0.40712663531303406, "learning_rate": 7.1290322580645166e-06, "loss": 0.0851, "step": 5980 }, { "epoch": 95.84, "grad_norm": 0.46202871203422546, "learning_rate": 6.806451612903226e-06, "loss": 0.0853, "step": 5990 }, { "epoch": 96.0, "grad_norm": 0.36839255690574646, "learning_rate": 6.483870967741936e-06, "loss": 0.0856, "step": 6000 }, { "epoch": 96.16, "grad_norm": 0.34025129675865173, "learning_rate": 6.161290322580645e-06, "loss": 0.084, "step": 6010 }, { "epoch": 96.32, "grad_norm": 0.37305212020874023, "learning_rate": 5.838709677419355e-06, "loss": 0.0838, "step": 6020 }, { "epoch": 96.48, "grad_norm": 0.3265480101108551, "learning_rate": 5.5161290322580645e-06, "loss": 0.0843, "step": 6030 }, { "epoch": 96.64, "grad_norm": 0.42518019676208496, "learning_rate": 5.193548387096774e-06, "loss": 0.0843, "step": 6040 }, { "epoch": 96.8, "grad_norm": 0.40051665902137756, "learning_rate": 4.870967741935484e-06, "loss": 0.0847, "step": 6050 }, { "epoch": 96.96, "grad_norm": 0.4043062627315521, "learning_rate": 4.548387096774194e-06, "loss": 0.0849, "step": 6060 }, { "epoch": 97.12, "grad_norm": 0.350195050239563, "learning_rate": 4.225806451612904e-06, "loss": 0.0845, "step": 6070 }, { "epoch": 97.28, "grad_norm": 0.4239977300167084, "learning_rate": 3.903225806451613e-06, "loss": 0.084, "step": 6080 }, { "epoch": 97.44, "grad_norm": 0.3817554712295532, "learning_rate": 3.5806451612903225e-06, "loss": 0.0842, "step": 6090 }, { "epoch": 97.6, "grad_norm": 0.3956522047519684, "learning_rate": 3.258064516129032e-06, "loss": 0.0837, "step": 6100 }, { "epoch": 97.76, "grad_norm": 0.4088551104068756, "learning_rate": 2.935483870967742e-06, "loss": 0.0842, "step": 6110 }, { "epoch": 97.92, "grad_norm": 0.4062351882457733, "learning_rate": 2.6129032258064518e-06, "loss": 0.084, "step": 6120 }, { "epoch": 98.08, "grad_norm": 0.3595307767391205, "learning_rate": 2.2903225806451614e-06, "loss": 0.0836, "step": 6130 }, { "epoch": 98.24, "grad_norm": 0.4038846790790558, "learning_rate": 1.967741935483871e-06, "loss": 0.0831, "step": 6140 }, { "epoch": 98.4, "grad_norm": 0.411885529756546, "learning_rate": 1.6451612903225808e-06, "loss": 0.0839, "step": 6150 }, { "epoch": 98.56, "grad_norm": 0.3277604579925537, "learning_rate": 1.3225806451612904e-06, "loss": 0.084, "step": 6160 }, { "epoch": 98.72, "grad_norm": 0.3658997118473053, "learning_rate": 1.0000000000000002e-06, "loss": 0.0843, "step": 6170 }, { "epoch": 98.88, "grad_norm": 0.4060486853122711, "learning_rate": 6.774193548387097e-07, "loss": 0.0838, "step": 6180 }, { "epoch": 99.04, "grad_norm": 0.3686084449291229, "learning_rate": 3.548387096774194e-07, "loss": 0.0835, "step": 6190 }, { "epoch": 99.2, "grad_norm": 0.36468222737312317, "learning_rate": 3.2258064516129035e-08, "loss": 0.0837, "step": 6200 }, { "epoch": 99.2, "step": 6200, "total_flos": 5.286498354069504e+17, "train_loss": 0.3455863463590222, "train_runtime": 16034.7635, "train_samples_per_second": 6.236, "train_steps_per_second": 0.387 } ], "logging_steps": 10, "max_steps": 6200, "num_input_tokens_seen": 0, "num_train_epochs": 100, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5.286498354069504e+17, "train_batch_size": 2, "trial_name": null, "trial_params": null }