{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0004173042147726, "eval_steps": 899, "global_step": 1798, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0005564056196967589, "grad_norm": 1.3203125, "learning_rate": 2e-05, "loss": 1.7806, "step": 1 }, { "epoch": 0.0005564056196967589, "eval_loss": 1.7791756391525269, "eval_runtime": 1750.0366, "eval_samples_per_second": 2.681, "eval_steps_per_second": 1.341, "step": 1 }, { "epoch": 0.0011128112393935178, "grad_norm": 1.1875, "learning_rate": 4e-05, "loss": 1.7467, "step": 2 }, { "epoch": 0.0016692168590902768, "grad_norm": 1.1484375, "learning_rate": 6e-05, "loss": 1.6933, "step": 3 }, { "epoch": 0.0022256224787870356, "grad_norm": 1.0390625, "learning_rate": 8e-05, "loss": 1.6916, "step": 4 }, { "epoch": 0.0027820280984837947, "grad_norm": 1.078125, "learning_rate": 0.0001, "loss": 1.6973, "step": 5 }, { "epoch": 0.0033384337181805537, "grad_norm": 0.9140625, "learning_rate": 0.00012, "loss": 1.6518, "step": 6 }, { "epoch": 0.0038948393378773127, "grad_norm": 0.796875, "learning_rate": 0.00014, "loss": 1.6383, "step": 7 }, { "epoch": 0.004451244957574071, "grad_norm": 0.75, "learning_rate": 0.00016, "loss": 1.5714, "step": 8 }, { "epoch": 0.005007650577270831, "grad_norm": 0.70703125, "learning_rate": 0.00018, "loss": 1.553, "step": 9 }, { "epoch": 0.005564056196967589, "grad_norm": 0.734375, "learning_rate": 0.0002, "loss": 1.5321, "step": 10 }, { "epoch": 0.006120461816664348, "grad_norm": 0.84765625, "learning_rate": 0.0001999999829570899, "loss": 1.5589, "step": 11 }, { "epoch": 0.006676867436361107, "grad_norm": 0.72265625, "learning_rate": 0.0001999999318283654, "loss": 1.4519, "step": 12 }, { "epoch": 0.007233273056057866, "grad_norm": 0.6640625, "learning_rate": 0.0001999998466138439, "loss": 1.4035, "step": 13 }, { "epoch": 0.007789678675754625, "grad_norm": 0.68359375, "learning_rate": 0.0001999997273135545, "loss": 1.4071, "step": 14 }, { "epoch": 0.008346084295451384, "grad_norm": 0.6796875, "learning_rate": 0.0001999995739275378, "loss": 1.2862, "step": 15 }, { "epoch": 0.008902489915148143, "grad_norm": 0.609375, "learning_rate": 0.00019999938645584616, "loss": 1.2801, "step": 16 }, { "epoch": 0.009458895534844901, "grad_norm": 0.5390625, "learning_rate": 0.0001999991648985434, "loss": 1.2358, "step": 17 }, { "epoch": 0.010015301154541661, "grad_norm": 0.56640625, "learning_rate": 0.0001999989092557051, "loss": 1.2122, "step": 18 }, { "epoch": 0.01057170677423842, "grad_norm": 0.56640625, "learning_rate": 0.00019999861952741839, "loss": 1.2294, "step": 19 }, { "epoch": 0.011128112393935179, "grad_norm": 0.53125, "learning_rate": 0.00019999829571378197, "loss": 1.2641, "step": 20 }, { "epoch": 0.011684518013631937, "grad_norm": 0.482421875, "learning_rate": 0.0001999979378149063, "loss": 1.2373, "step": 21 }, { "epoch": 0.012240923633328696, "grad_norm": 0.474609375, "learning_rate": 0.0001999975458309133, "loss": 1.2389, "step": 22 }, { "epoch": 0.012797329253025456, "grad_norm": 0.47265625, "learning_rate": 0.00019999711976193663, "loss": 1.2716, "step": 23 }, { "epoch": 0.013353734872722215, "grad_norm": 0.455078125, "learning_rate": 0.0001999966596081215, "loss": 1.2038, "step": 24 }, { "epoch": 0.013910140492418973, "grad_norm": 0.462890625, "learning_rate": 0.00019999616536962473, "loss": 1.2588, "step": 25 }, { "epoch": 0.014466546112115732, "grad_norm": 0.451171875, "learning_rate": 0.00019999563704661483, "loss": 1.198, "step": 26 }, { "epoch": 0.01502295173181249, "grad_norm": 0.4296875, "learning_rate": 0.00019999507463927188, "loss": 1.2047, "step": 27 }, { "epoch": 0.01557935735150925, "grad_norm": 0.431640625, "learning_rate": 0.00019999447814778755, "loss": 1.1884, "step": 28 }, { "epoch": 0.01613576297120601, "grad_norm": 0.435546875, "learning_rate": 0.0001999938475723652, "loss": 1.2156, "step": 29 }, { "epoch": 0.016692168590902768, "grad_norm": 0.4375, "learning_rate": 0.00019999318291321968, "loss": 1.1636, "step": 30 }, { "epoch": 0.017248574210599527, "grad_norm": 0.4140625, "learning_rate": 0.00019999248417057767, "loss": 1.1857, "step": 31 }, { "epoch": 0.017804979830296285, "grad_norm": 0.439453125, "learning_rate": 0.00019999175134467726, "loss": 1.1862, "step": 32 }, { "epoch": 0.018361385449993044, "grad_norm": 0.41015625, "learning_rate": 0.00019999098443576827, "loss": 1.2155, "step": 33 }, { "epoch": 0.018917791069689802, "grad_norm": 0.427734375, "learning_rate": 0.00019999018344411207, "loss": 1.1009, "step": 34 }, { "epoch": 0.019474196689386564, "grad_norm": 0.421875, "learning_rate": 0.00019998934836998173, "loss": 1.1714, "step": 35 }, { "epoch": 0.020030602309083323, "grad_norm": 0.42578125, "learning_rate": 0.00019998847921366187, "loss": 1.0824, "step": 36 }, { "epoch": 0.02058700792878008, "grad_norm": 0.390625, "learning_rate": 0.00019998757597544875, "loss": 1.0703, "step": 37 }, { "epoch": 0.02114341354847684, "grad_norm": 0.392578125, "learning_rate": 0.00019998663865565023, "loss": 1.1183, "step": 38 }, { "epoch": 0.0216998191681736, "grad_norm": 0.400390625, "learning_rate": 0.00019998566725458586, "loss": 1.1328, "step": 39 }, { "epoch": 0.022256224787870357, "grad_norm": 0.396484375, "learning_rate": 0.0001999846617725867, "loss": 1.1076, "step": 40 }, { "epoch": 0.022812630407567116, "grad_norm": 0.39453125, "learning_rate": 0.00019998362220999548, "loss": 1.1143, "step": 41 }, { "epoch": 0.023369036027263874, "grad_norm": 0.419921875, "learning_rate": 0.00019998254856716657, "loss": 1.1768, "step": 42 }, { "epoch": 0.023925441646960633, "grad_norm": 0.400390625, "learning_rate": 0.0001999814408444659, "loss": 1.0872, "step": 43 }, { "epoch": 0.02448184726665739, "grad_norm": 0.396484375, "learning_rate": 0.00019998029904227103, "loss": 1.122, "step": 44 }, { "epoch": 0.025038252886354154, "grad_norm": 0.39453125, "learning_rate": 0.0001999791231609712, "loss": 1.0025, "step": 45 }, { "epoch": 0.025594658506050912, "grad_norm": 0.4140625, "learning_rate": 0.00019997791320096718, "loss": 1.1672, "step": 46 }, { "epoch": 0.02615106412574767, "grad_norm": 0.40234375, "learning_rate": 0.00019997666916267144, "loss": 1.1096, "step": 47 }, { "epoch": 0.02670746974544443, "grad_norm": 0.39453125, "learning_rate": 0.000199975391046508, "loss": 1.1446, "step": 48 }, { "epoch": 0.027263875365141188, "grad_norm": 0.400390625, "learning_rate": 0.0001999740788529125, "loss": 1.1254, "step": 49 }, { "epoch": 0.027820280984837947, "grad_norm": 0.396484375, "learning_rate": 0.00019997273258233223, "loss": 1.1143, "step": 50 }, { "epoch": 0.028376686604534705, "grad_norm": 0.404296875, "learning_rate": 0.00019997135223522605, "loss": 1.1123, "step": 51 }, { "epoch": 0.028933092224231464, "grad_norm": 0.384765625, "learning_rate": 0.00019996993781206448, "loss": 1.0469, "step": 52 }, { "epoch": 0.029489497843928222, "grad_norm": 0.388671875, "learning_rate": 0.00019996848931332965, "loss": 1.1145, "step": 53 }, { "epoch": 0.03004590346362498, "grad_norm": 0.38671875, "learning_rate": 0.0001999670067395153, "loss": 1.0489, "step": 54 }, { "epoch": 0.030602309083321743, "grad_norm": 0.390625, "learning_rate": 0.00019996549009112675, "loss": 1.0671, "step": 55 }, { "epoch": 0.0311587147030185, "grad_norm": 0.3828125, "learning_rate": 0.00019996393936868098, "loss": 1.0341, "step": 56 }, { "epoch": 0.03171512032271526, "grad_norm": 0.3828125, "learning_rate": 0.00019996235457270654, "loss": 1.0725, "step": 57 }, { "epoch": 0.03227152594241202, "grad_norm": 0.388671875, "learning_rate": 0.00019996073570374367, "loss": 1.1031, "step": 58 }, { "epoch": 0.032827931562108774, "grad_norm": 0.380859375, "learning_rate": 0.0001999590827623441, "loss": 0.9974, "step": 59 }, { "epoch": 0.033384337181805536, "grad_norm": 0.392578125, "learning_rate": 0.00019995739574907133, "loss": 1.1398, "step": 60 }, { "epoch": 0.0339407428015023, "grad_norm": 0.40234375, "learning_rate": 0.00019995567466450035, "loss": 1.0728, "step": 61 }, { "epoch": 0.03449714842119905, "grad_norm": 0.380859375, "learning_rate": 0.00019995391950921782, "loss": 0.9649, "step": 62 }, { "epoch": 0.035053554040895815, "grad_norm": 0.384765625, "learning_rate": 0.000199952130283822, "loss": 1.1536, "step": 63 }, { "epoch": 0.03560995966059257, "grad_norm": 0.376953125, "learning_rate": 0.0001999503069889227, "loss": 1.0135, "step": 64 }, { "epoch": 0.03616636528028933, "grad_norm": 0.39453125, "learning_rate": 0.0001999484496251415, "loss": 1.0599, "step": 65 }, { "epoch": 0.03672277089998609, "grad_norm": 0.384765625, "learning_rate": 0.00019994655819311144, "loss": 1.0521, "step": 66 }, { "epoch": 0.03727917651968285, "grad_norm": 0.388671875, "learning_rate": 0.00019994463269347725, "loss": 1.0737, "step": 67 }, { "epoch": 0.037835582139379605, "grad_norm": 0.388671875, "learning_rate": 0.00019994267312689525, "loss": 1.0856, "step": 68 }, { "epoch": 0.03839198775907637, "grad_norm": 0.365234375, "learning_rate": 0.00019994067949403337, "loss": 0.9674, "step": 69 }, { "epoch": 0.03894839337877313, "grad_norm": 0.390625, "learning_rate": 0.00019993865179557117, "loss": 1.1069, "step": 70 }, { "epoch": 0.039504798998469884, "grad_norm": 0.3828125, "learning_rate": 0.00019993659003219978, "loss": 1.0198, "step": 71 }, { "epoch": 0.040061204618166646, "grad_norm": 0.384765625, "learning_rate": 0.00019993449420462197, "loss": 1.0864, "step": 72 }, { "epoch": 0.0406176102378634, "grad_norm": 0.38671875, "learning_rate": 0.00019993236431355217, "loss": 1.037, "step": 73 }, { "epoch": 0.04117401585756016, "grad_norm": 0.39453125, "learning_rate": 0.00019993020035971633, "loss": 1.0724, "step": 74 }, { "epoch": 0.04173042147725692, "grad_norm": 0.369140625, "learning_rate": 0.00019992800234385201, "loss": 0.9854, "step": 75 }, { "epoch": 0.04228682709695368, "grad_norm": 0.37890625, "learning_rate": 0.00019992577026670854, "loss": 1.0655, "step": 76 }, { "epoch": 0.042843232716650435, "grad_norm": 0.392578125, "learning_rate": 0.00019992350412904663, "loss": 1.0487, "step": 77 }, { "epoch": 0.0433996383363472, "grad_norm": 0.369140625, "learning_rate": 0.00019992120393163876, "loss": 0.9593, "step": 78 }, { "epoch": 0.04395604395604396, "grad_norm": 0.40234375, "learning_rate": 0.00019991886967526897, "loss": 1.1132, "step": 79 }, { "epoch": 0.044512449575740715, "grad_norm": 0.36328125, "learning_rate": 0.00019991650136073287, "loss": 0.9343, "step": 80 }, { "epoch": 0.04506885519543748, "grad_norm": 0.369140625, "learning_rate": 0.0001999140989888378, "loss": 1.0228, "step": 81 }, { "epoch": 0.04562526081513423, "grad_norm": 0.384765625, "learning_rate": 0.00019991166256040256, "loss": 1.0239, "step": 82 }, { "epoch": 0.046181666434830994, "grad_norm": 0.38671875, "learning_rate": 0.00019990919207625764, "loss": 1.0706, "step": 83 }, { "epoch": 0.04673807205452775, "grad_norm": 0.38671875, "learning_rate": 0.00019990668753724515, "loss": 1.0069, "step": 84 }, { "epoch": 0.04729447767422451, "grad_norm": 0.38671875, "learning_rate": 0.00019990414894421876, "loss": 1.0511, "step": 85 }, { "epoch": 0.047850883293921266, "grad_norm": 0.37890625, "learning_rate": 0.0001999015762980438, "loss": 0.9398, "step": 86 }, { "epoch": 0.04840728891361803, "grad_norm": 0.396484375, "learning_rate": 0.00019989896959959712, "loss": 1.1496, "step": 87 }, { "epoch": 0.04896369453331478, "grad_norm": 0.369140625, "learning_rate": 0.00019989632884976727, "loss": 1.0488, "step": 88 }, { "epoch": 0.049520100153011545, "grad_norm": 0.39453125, "learning_rate": 0.00019989365404945436, "loss": 1.0843, "step": 89 }, { "epoch": 0.05007650577270831, "grad_norm": 0.375, "learning_rate": 0.00019989094519957015, "loss": 1.0822, "step": 90 }, { "epoch": 0.05063291139240506, "grad_norm": 0.373046875, "learning_rate": 0.00019988820230103796, "loss": 1.0301, "step": 91 }, { "epoch": 0.051189317012101825, "grad_norm": 0.380859375, "learning_rate": 0.0001998854253547927, "loss": 1.0098, "step": 92 }, { "epoch": 0.05174572263179858, "grad_norm": 0.361328125, "learning_rate": 0.00019988261436178093, "loss": 0.9847, "step": 93 }, { "epoch": 0.05230212825149534, "grad_norm": 0.38671875, "learning_rate": 0.0001998797693229608, "loss": 1.0233, "step": 94 }, { "epoch": 0.0528585338711921, "grad_norm": 0.388671875, "learning_rate": 0.0001998768902393021, "loss": 1.0747, "step": 95 }, { "epoch": 0.05341493949088886, "grad_norm": 0.3828125, "learning_rate": 0.00019987397711178614, "loss": 0.9823, "step": 96 }, { "epoch": 0.053971345110585614, "grad_norm": 0.390625, "learning_rate": 0.0001998710299414059, "loss": 1.0238, "step": 97 }, { "epoch": 0.054527750730282376, "grad_norm": 0.38671875, "learning_rate": 0.00019986804872916593, "loss": 0.9832, "step": 98 }, { "epoch": 0.05508415634997914, "grad_norm": 0.373046875, "learning_rate": 0.00019986503347608245, "loss": 0.9833, "step": 99 }, { "epoch": 0.05564056196967589, "grad_norm": 0.392578125, "learning_rate": 0.0001998619841831832, "loss": 1.0971, "step": 100 }, { "epoch": 0.056196967589372655, "grad_norm": 0.376953125, "learning_rate": 0.0001998589008515076, "loss": 1.0077, "step": 101 }, { "epoch": 0.05675337320906941, "grad_norm": 0.390625, "learning_rate": 0.0001998557834821065, "loss": 1.0066, "step": 102 }, { "epoch": 0.05730977882876617, "grad_norm": 0.380859375, "learning_rate": 0.00019985263207604264, "loss": 1.0267, "step": 103 }, { "epoch": 0.05786618444846293, "grad_norm": 0.373046875, "learning_rate": 0.0001998494466343901, "loss": 0.975, "step": 104 }, { "epoch": 0.05842259006815969, "grad_norm": 0.3828125, "learning_rate": 0.00019984622715823474, "loss": 1.0134, "step": 105 }, { "epoch": 0.058978995687856445, "grad_norm": 0.392578125, "learning_rate": 0.00019984297364867385, "loss": 1.1247, "step": 106 }, { "epoch": 0.05953540130755321, "grad_norm": 0.376953125, "learning_rate": 0.00019983968610681648, "loss": 0.974, "step": 107 }, { "epoch": 0.06009180692724996, "grad_norm": 0.380859375, "learning_rate": 0.0001998363645337832, "loss": 1.0099, "step": 108 }, { "epoch": 0.060648212546946724, "grad_norm": 0.3515625, "learning_rate": 0.00019983300893070623, "loss": 0.8479, "step": 109 }, { "epoch": 0.061204618166643486, "grad_norm": 0.375, "learning_rate": 0.00019982961929872927, "loss": 1.0034, "step": 110 }, { "epoch": 0.06176102378634024, "grad_norm": 0.380859375, "learning_rate": 0.00019982619563900782, "loss": 1.0165, "step": 111 }, { "epoch": 0.062317429406037, "grad_norm": 0.390625, "learning_rate": 0.00019982273795270876, "loss": 1.0338, "step": 112 }, { "epoch": 0.06287383502573377, "grad_norm": 0.392578125, "learning_rate": 0.00019981924624101071, "loss": 1.0465, "step": 113 }, { "epoch": 0.06343024064543051, "grad_norm": 0.369140625, "learning_rate": 0.00019981572050510387, "loss": 0.9284, "step": 114 }, { "epoch": 0.06398664626512728, "grad_norm": 0.40234375, "learning_rate": 0.00019981216074619, "loss": 1.074, "step": 115 }, { "epoch": 0.06454305188482404, "grad_norm": 0.375, "learning_rate": 0.00019980856696548242, "loss": 0.8981, "step": 116 }, { "epoch": 0.0650994575045208, "grad_norm": 0.375, "learning_rate": 0.00019980493916420616, "loss": 1.0605, "step": 117 }, { "epoch": 0.06565586312421755, "grad_norm": 0.37890625, "learning_rate": 0.00019980127734359782, "loss": 1.0433, "step": 118 }, { "epoch": 0.06621226874391431, "grad_norm": 0.38671875, "learning_rate": 0.00019979758150490546, "loss": 1.0483, "step": 119 }, { "epoch": 0.06676867436361107, "grad_norm": 0.392578125, "learning_rate": 0.0001997938516493889, "loss": 1.0543, "step": 120 }, { "epoch": 0.06732507998330783, "grad_norm": 0.376953125, "learning_rate": 0.0001997900877783195, "loss": 1.0162, "step": 121 }, { "epoch": 0.0678814856030046, "grad_norm": 0.369140625, "learning_rate": 0.0001997862898929802, "loss": 0.8468, "step": 122 }, { "epoch": 0.06843789122270134, "grad_norm": 0.37890625, "learning_rate": 0.00019978245799466554, "loss": 1.0417, "step": 123 }, { "epoch": 0.0689942968423981, "grad_norm": 0.3828125, "learning_rate": 0.00019977859208468161, "loss": 0.9901, "step": 124 }, { "epoch": 0.06955070246209487, "grad_norm": 0.404296875, "learning_rate": 0.00019977469216434619, "loss": 1.0117, "step": 125 }, { "epoch": 0.07010710808179163, "grad_norm": 0.375, "learning_rate": 0.0001997707582349886, "loss": 0.9719, "step": 126 }, { "epoch": 0.07066351370148838, "grad_norm": 0.369140625, "learning_rate": 0.00019976679029794968, "loss": 0.9176, "step": 127 }, { "epoch": 0.07121991932118514, "grad_norm": 0.400390625, "learning_rate": 0.00019976278835458203, "loss": 1.0633, "step": 128 }, { "epoch": 0.0717763249408819, "grad_norm": 0.380859375, "learning_rate": 0.00019975875240624968, "loss": 0.9845, "step": 129 }, { "epoch": 0.07233273056057866, "grad_norm": 0.376953125, "learning_rate": 0.00019975468245432833, "loss": 0.9352, "step": 130 }, { "epoch": 0.07288913618027543, "grad_norm": 0.3828125, "learning_rate": 0.0001997505785002053, "loss": 1.0966, "step": 131 }, { "epoch": 0.07344554179997217, "grad_norm": 0.380859375, "learning_rate": 0.0001997464405452794, "loss": 0.9278, "step": 132 }, { "epoch": 0.07400194741966894, "grad_norm": 0.359375, "learning_rate": 0.0001997422685909611, "loss": 0.9244, "step": 133 }, { "epoch": 0.0745583530393657, "grad_norm": 0.3671875, "learning_rate": 0.00019973806263867247, "loss": 0.8941, "step": 134 }, { "epoch": 0.07511475865906246, "grad_norm": 0.380859375, "learning_rate": 0.00019973382268984713, "loss": 0.9702, "step": 135 }, { "epoch": 0.07567116427875921, "grad_norm": 0.376953125, "learning_rate": 0.00019972954874593026, "loss": 0.9809, "step": 136 }, { "epoch": 0.07622756989845597, "grad_norm": 0.37890625, "learning_rate": 0.00019972524080837873, "loss": 1.0604, "step": 137 }, { "epoch": 0.07678397551815273, "grad_norm": 0.361328125, "learning_rate": 0.00019972089887866092, "loss": 0.9702, "step": 138 }, { "epoch": 0.0773403811378495, "grad_norm": 0.37890625, "learning_rate": 0.00019971652295825681, "loss": 0.9903, "step": 139 }, { "epoch": 0.07789678675754626, "grad_norm": 0.36328125, "learning_rate": 0.00019971211304865795, "loss": 0.979, "step": 140 }, { "epoch": 0.078453192377243, "grad_norm": 0.380859375, "learning_rate": 0.00019970766915136746, "loss": 1.1067, "step": 141 }, { "epoch": 0.07900959799693977, "grad_norm": 0.357421875, "learning_rate": 0.0001997031912679002, "loss": 0.966, "step": 142 }, { "epoch": 0.07956600361663653, "grad_norm": 0.375, "learning_rate": 0.00019969867939978236, "loss": 1.0207, "step": 143 }, { "epoch": 0.08012240923633329, "grad_norm": 0.37890625, "learning_rate": 0.00019969413354855195, "loss": 0.9127, "step": 144 }, { "epoch": 0.08067881485603004, "grad_norm": 0.400390625, "learning_rate": 0.0001996895537157584, "loss": 1.0944, "step": 145 }, { "epoch": 0.0812352204757268, "grad_norm": 0.35546875, "learning_rate": 0.00019968493990296277, "loss": 0.948, "step": 146 }, { "epoch": 0.08179162609542356, "grad_norm": 0.3828125, "learning_rate": 0.00019968029211173778, "loss": 0.9824, "step": 147 }, { "epoch": 0.08234803171512033, "grad_norm": 0.37890625, "learning_rate": 0.00019967561034366763, "loss": 0.9328, "step": 148 }, { "epoch": 0.08290443733481709, "grad_norm": 0.37109375, "learning_rate": 0.00019967089460034814, "loss": 1.0162, "step": 149 }, { "epoch": 0.08346084295451384, "grad_norm": 0.33984375, "learning_rate": 0.0001996661448833867, "loss": 0.7696, "step": 150 }, { "epoch": 0.0840172485742106, "grad_norm": 0.361328125, "learning_rate": 0.00019966136119440233, "loss": 0.9133, "step": 151 }, { "epoch": 0.08457365419390736, "grad_norm": 0.384765625, "learning_rate": 0.00019965654353502554, "loss": 1.0071, "step": 152 }, { "epoch": 0.08513005981360412, "grad_norm": 0.392578125, "learning_rate": 0.0001996516919068985, "loss": 0.9797, "step": 153 }, { "epoch": 0.08568646543330087, "grad_norm": 0.388671875, "learning_rate": 0.00019964680631167492, "loss": 0.9521, "step": 154 }, { "epoch": 0.08624287105299763, "grad_norm": 0.392578125, "learning_rate": 0.0001996418867510201, "loss": 1.034, "step": 155 }, { "epoch": 0.0867992766726944, "grad_norm": 0.373046875, "learning_rate": 0.00019963693322661087, "loss": 0.9971, "step": 156 }, { "epoch": 0.08735568229239116, "grad_norm": 0.37890625, "learning_rate": 0.00019963194574013573, "loss": 1.0262, "step": 157 }, { "epoch": 0.08791208791208792, "grad_norm": 0.380859375, "learning_rate": 0.0001996269242932947, "loss": 0.9913, "step": 158 }, { "epoch": 0.08846849353178467, "grad_norm": 0.408203125, "learning_rate": 0.00019962186888779936, "loss": 1.0822, "step": 159 }, { "epoch": 0.08902489915148143, "grad_norm": 0.36328125, "learning_rate": 0.00019961677952537292, "loss": 0.8287, "step": 160 }, { "epoch": 0.08958130477117819, "grad_norm": 0.37890625, "learning_rate": 0.0001996116562077501, "loss": 0.9663, "step": 161 }, { "epoch": 0.09013771039087495, "grad_norm": 0.333984375, "learning_rate": 0.00019960649893667722, "loss": 0.7657, "step": 162 }, { "epoch": 0.0906941160105717, "grad_norm": 0.35546875, "learning_rate": 0.0001996013077139122, "loss": 0.8837, "step": 163 }, { "epoch": 0.09125052163026846, "grad_norm": 0.376953125, "learning_rate": 0.0001995960825412245, "loss": 0.9727, "step": 164 }, { "epoch": 0.09180692724996523, "grad_norm": 0.3671875, "learning_rate": 0.0001995908234203952, "loss": 0.9438, "step": 165 }, { "epoch": 0.09236333286966199, "grad_norm": 0.369140625, "learning_rate": 0.00019958553035321686, "loss": 0.9364, "step": 166 }, { "epoch": 0.09291973848935874, "grad_norm": 0.353515625, "learning_rate": 0.00019958020334149366, "loss": 0.9435, "step": 167 }, { "epoch": 0.0934761441090555, "grad_norm": 0.384765625, "learning_rate": 0.00019957484238704139, "loss": 1.034, "step": 168 }, { "epoch": 0.09403254972875226, "grad_norm": 0.37109375, "learning_rate": 0.00019956944749168737, "loss": 0.9662, "step": 169 }, { "epoch": 0.09458895534844902, "grad_norm": 0.38671875, "learning_rate": 0.00019956401865727053, "loss": 1.0057, "step": 170 }, { "epoch": 0.09514536096814578, "grad_norm": 0.361328125, "learning_rate": 0.00019955855588564125, "loss": 0.9569, "step": 171 }, { "epoch": 0.09570176658784253, "grad_norm": 0.37109375, "learning_rate": 0.00019955305917866164, "loss": 0.9651, "step": 172 }, { "epoch": 0.0962581722075393, "grad_norm": 0.3515625, "learning_rate": 0.00019954752853820523, "loss": 0.8911, "step": 173 }, { "epoch": 0.09681457782723606, "grad_norm": 0.3515625, "learning_rate": 0.00019954196396615721, "loss": 0.9656, "step": 174 }, { "epoch": 0.09737098344693282, "grad_norm": 0.373046875, "learning_rate": 0.00019953636546441431, "loss": 0.8851, "step": 175 }, { "epoch": 0.09792738906662957, "grad_norm": 0.357421875, "learning_rate": 0.00019953073303488486, "loss": 0.91, "step": 176 }, { "epoch": 0.09848379468632633, "grad_norm": 0.3828125, "learning_rate": 0.00019952506667948868, "loss": 1.0532, "step": 177 }, { "epoch": 0.09904020030602309, "grad_norm": 0.384765625, "learning_rate": 0.00019951936640015723, "loss": 1.052, "step": 178 }, { "epoch": 0.09959660592571985, "grad_norm": 0.36328125, "learning_rate": 0.00019951363219883344, "loss": 0.9436, "step": 179 }, { "epoch": 0.10015301154541661, "grad_norm": 0.3828125, "learning_rate": 0.0001995078640774719, "loss": 1.0067, "step": 180 }, { "epoch": 0.10070941716511336, "grad_norm": 0.353515625, "learning_rate": 0.00019950206203803874, "loss": 0.8923, "step": 181 }, { "epoch": 0.10126582278481013, "grad_norm": 0.3671875, "learning_rate": 0.00019949622608251156, "loss": 0.9681, "step": 182 }, { "epoch": 0.10182222840450689, "grad_norm": 0.3828125, "learning_rate": 0.00019949035621287967, "loss": 0.9994, "step": 183 }, { "epoch": 0.10237863402420365, "grad_norm": 0.3671875, "learning_rate": 0.00019948445243114384, "loss": 1.0139, "step": 184 }, { "epoch": 0.1029350396439004, "grad_norm": 0.365234375, "learning_rate": 0.0001994785147393164, "loss": 0.9698, "step": 185 }, { "epoch": 0.10349144526359716, "grad_norm": 0.373046875, "learning_rate": 0.0001994725431394213, "loss": 0.9727, "step": 186 }, { "epoch": 0.10404785088329392, "grad_norm": 0.34765625, "learning_rate": 0.00019946653763349394, "loss": 0.9109, "step": 187 }, { "epoch": 0.10460425650299068, "grad_norm": 0.380859375, "learning_rate": 0.0001994604982235814, "loss": 0.9881, "step": 188 }, { "epoch": 0.10516066212268745, "grad_norm": 0.3828125, "learning_rate": 0.00019945442491174227, "loss": 1.0062, "step": 189 }, { "epoch": 0.1057170677423842, "grad_norm": 0.375, "learning_rate": 0.00019944831770004665, "loss": 0.9374, "step": 190 }, { "epoch": 0.10627347336208096, "grad_norm": 0.3671875, "learning_rate": 0.00019944217659057626, "loss": 0.907, "step": 191 }, { "epoch": 0.10682987898177772, "grad_norm": 0.365234375, "learning_rate": 0.00019943600158542436, "loss": 0.9287, "step": 192 }, { "epoch": 0.10738628460147448, "grad_norm": 0.376953125, "learning_rate": 0.00019942979268669573, "loss": 0.9948, "step": 193 }, { "epoch": 0.10794269022117123, "grad_norm": 0.373046875, "learning_rate": 0.0001994235498965067, "loss": 0.9651, "step": 194 }, { "epoch": 0.10849909584086799, "grad_norm": 0.3671875, "learning_rate": 0.00019941727321698521, "loss": 0.9267, "step": 195 }, { "epoch": 0.10905550146056475, "grad_norm": 0.349609375, "learning_rate": 0.00019941096265027074, "loss": 0.8551, "step": 196 }, { "epoch": 0.10961190708026151, "grad_norm": 0.369140625, "learning_rate": 0.00019940461819851425, "loss": 0.9683, "step": 197 }, { "epoch": 0.11016831269995828, "grad_norm": 0.357421875, "learning_rate": 0.00019939823986387834, "loss": 0.9127, "step": 198 }, { "epoch": 0.11072471831965502, "grad_norm": 0.373046875, "learning_rate": 0.00019939182764853707, "loss": 1.0557, "step": 199 }, { "epoch": 0.11128112393935179, "grad_norm": 0.37890625, "learning_rate": 0.00019938538155467615, "loss": 0.9967, "step": 200 }, { "epoch": 0.11183752955904855, "grad_norm": 0.376953125, "learning_rate": 0.00019937890158449272, "loss": 0.9391, "step": 201 }, { "epoch": 0.11239393517874531, "grad_norm": 2.109375, "learning_rate": 0.00019937238774019557, "loss": 1.9551, "step": 202 }, { "epoch": 0.11295034079844206, "grad_norm": 0.37890625, "learning_rate": 0.000199365840024005, "loss": 0.9682, "step": 203 }, { "epoch": 0.11350674641813882, "grad_norm": 0.373046875, "learning_rate": 0.00019935925843815285, "loss": 0.961, "step": 204 }, { "epoch": 0.11406315203783558, "grad_norm": 0.42578125, "learning_rate": 0.0001993526429848825, "loss": 1.0718, "step": 205 }, { "epoch": 0.11461955765753234, "grad_norm": 0.37890625, "learning_rate": 0.0001993459936664489, "loss": 0.9232, "step": 206 }, { "epoch": 0.1151759632772291, "grad_norm": 0.396484375, "learning_rate": 0.0001993393104851185, "loss": 1.0022, "step": 207 }, { "epoch": 0.11573236889692586, "grad_norm": 0.37890625, "learning_rate": 0.00019933259344316934, "loss": 0.9549, "step": 208 }, { "epoch": 0.11628877451662262, "grad_norm": 0.357421875, "learning_rate": 0.00019932584254289096, "loss": 0.9472, "step": 209 }, { "epoch": 0.11684518013631938, "grad_norm": 2.859375, "learning_rate": 0.00019931905778658443, "loss": 1.9077, "step": 210 }, { "epoch": 0.11740158575601614, "grad_norm": 0.388671875, "learning_rate": 0.00019931223917656247, "loss": 1.0227, "step": 211 }, { "epoch": 0.11795799137571289, "grad_norm": 0.404296875, "learning_rate": 0.00019930538671514923, "loss": 1.0234, "step": 212 }, { "epoch": 0.11851439699540965, "grad_norm": 0.373046875, "learning_rate": 0.00019929850040468035, "loss": 0.9573, "step": 213 }, { "epoch": 0.11907080261510641, "grad_norm": 0.365234375, "learning_rate": 0.0001992915802475032, "loss": 0.9576, "step": 214 }, { "epoch": 0.11962720823480318, "grad_norm": 0.365234375, "learning_rate": 0.0001992846262459765, "loss": 0.9715, "step": 215 }, { "epoch": 0.12018361385449992, "grad_norm": 0.376953125, "learning_rate": 0.0001992776384024706, "loss": 0.8743, "step": 216 }, { "epoch": 0.12074001947419669, "grad_norm": 0.365234375, "learning_rate": 0.00019927061671936737, "loss": 0.8957, "step": 217 }, { "epoch": 0.12129642509389345, "grad_norm": 0.37890625, "learning_rate": 0.0001992635611990602, "loss": 0.9718, "step": 218 }, { "epoch": 0.12185283071359021, "grad_norm": 0.369140625, "learning_rate": 0.00019925647184395402, "loss": 0.8916, "step": 219 }, { "epoch": 0.12240923633328697, "grad_norm": 0.361328125, "learning_rate": 0.0001992493486564653, "loss": 0.9546, "step": 220 }, { "epoch": 0.12296564195298372, "grad_norm": 0.369140625, "learning_rate": 0.00019924219163902204, "loss": 0.9285, "step": 221 }, { "epoch": 0.12352204757268048, "grad_norm": 0.36328125, "learning_rate": 0.00019923500079406373, "loss": 0.9095, "step": 222 }, { "epoch": 0.12407845319237724, "grad_norm": 0.390625, "learning_rate": 0.00019922777612404148, "loss": 0.9921, "step": 223 }, { "epoch": 0.124634858812074, "grad_norm": 0.376953125, "learning_rate": 0.00019922051763141788, "loss": 0.9756, "step": 224 }, { "epoch": 0.12519126443177075, "grad_norm": 0.345703125, "learning_rate": 0.00019921322531866702, "loss": 0.832, "step": 225 }, { "epoch": 0.12574767005146753, "grad_norm": 0.322265625, "learning_rate": 0.00019920589918827455, "loss": 0.741, "step": 226 }, { "epoch": 0.12630407567116428, "grad_norm": 0.3671875, "learning_rate": 0.0001991985392427376, "loss": 0.8839, "step": 227 }, { "epoch": 0.12686048129086103, "grad_norm": 0.37890625, "learning_rate": 0.00019919114548456497, "loss": 1.0329, "step": 228 }, { "epoch": 0.1274168869105578, "grad_norm": 0.369140625, "learning_rate": 0.0001991837179162768, "loss": 0.9777, "step": 229 }, { "epoch": 0.12797329253025455, "grad_norm": 0.357421875, "learning_rate": 0.00019917625654040485, "loss": 0.9311, "step": 230 }, { "epoch": 0.12852969814995133, "grad_norm": 0.38671875, "learning_rate": 0.0001991687613594924, "loss": 0.9973, "step": 231 }, { "epoch": 0.12908610376964808, "grad_norm": 0.375, "learning_rate": 0.00019916123237609428, "loss": 1.054, "step": 232 }, { "epoch": 0.12964250938934482, "grad_norm": 0.36328125, "learning_rate": 0.00019915366959277674, "loss": 0.886, "step": 233 }, { "epoch": 0.1301989150090416, "grad_norm": 0.37109375, "learning_rate": 0.00019914607301211765, "loss": 0.9879, "step": 234 }, { "epoch": 0.13075532062873835, "grad_norm": 0.333984375, "learning_rate": 0.00019913844263670638, "loss": 0.8272, "step": 235 }, { "epoch": 0.1313117262484351, "grad_norm": 0.37890625, "learning_rate": 0.00019913077846914377, "loss": 0.8881, "step": 236 }, { "epoch": 0.13186813186813187, "grad_norm": 0.37109375, "learning_rate": 0.00019912308051204226, "loss": 0.9737, "step": 237 }, { "epoch": 0.13242453748782862, "grad_norm": 0.40625, "learning_rate": 0.00019911534876802572, "loss": 1.1547, "step": 238 }, { "epoch": 0.1329809431075254, "grad_norm": 0.37109375, "learning_rate": 0.00019910758323972961, "loss": 0.9975, "step": 239 }, { "epoch": 0.13353734872722214, "grad_norm": 0.36328125, "learning_rate": 0.00019909978392980087, "loss": 0.9454, "step": 240 }, { "epoch": 0.1340937543469189, "grad_norm": 0.37890625, "learning_rate": 0.00019909195084089792, "loss": 0.969, "step": 241 }, { "epoch": 0.13465015996661567, "grad_norm": 0.357421875, "learning_rate": 0.00019908408397569075, "loss": 1.0012, "step": 242 }, { "epoch": 0.13520656558631242, "grad_norm": 0.365234375, "learning_rate": 0.0001990761833368609, "loss": 0.9292, "step": 243 }, { "epoch": 0.1357629712060092, "grad_norm": 0.35546875, "learning_rate": 0.0001990682489271013, "loss": 0.8401, "step": 244 }, { "epoch": 0.13631937682570594, "grad_norm": 0.36328125, "learning_rate": 0.00019906028074911645, "loss": 0.9614, "step": 245 }, { "epoch": 0.1368757824454027, "grad_norm": 0.388671875, "learning_rate": 0.00019905227880562246, "loss": 0.9454, "step": 246 }, { "epoch": 0.13743218806509946, "grad_norm": 0.349609375, "learning_rate": 0.0001990442430993468, "loss": 0.8745, "step": 247 }, { "epoch": 0.1379885936847962, "grad_norm": 0.361328125, "learning_rate": 0.00019903617363302846, "loss": 0.9185, "step": 248 }, { "epoch": 0.138544999304493, "grad_norm": 0.37109375, "learning_rate": 0.00019902807040941806, "loss": 0.9842, "step": 249 }, { "epoch": 0.13910140492418974, "grad_norm": 0.37890625, "learning_rate": 0.00019901993343127763, "loss": 1.0057, "step": 250 }, { "epoch": 0.13965781054388648, "grad_norm": 0.357421875, "learning_rate": 0.0001990117627013807, "loss": 0.8891, "step": 251 }, { "epoch": 0.14021421616358326, "grad_norm": 0.361328125, "learning_rate": 0.00019900355822251233, "loss": 0.9715, "step": 252 }, { "epoch": 0.14077062178328, "grad_norm": 0.34375, "learning_rate": 0.0001989953199974691, "loss": 0.8475, "step": 253 }, { "epoch": 0.14132702740297676, "grad_norm": 0.3515625, "learning_rate": 0.0001989870480290591, "loss": 0.9162, "step": 254 }, { "epoch": 0.14188343302267353, "grad_norm": 0.349609375, "learning_rate": 0.00019897874232010186, "loss": 0.8788, "step": 255 }, { "epoch": 0.14243983864237028, "grad_norm": 0.7734375, "learning_rate": 0.0001989704028734285, "loss": 1.8645, "step": 256 }, { "epoch": 0.14299624426206706, "grad_norm": 0.53125, "learning_rate": 0.00019896202969188148, "loss": 1.0641, "step": 257 }, { "epoch": 0.1435526498817638, "grad_norm": 0.37109375, "learning_rate": 0.000198953622778315, "loss": 1.0014, "step": 258 }, { "epoch": 0.14410905550146055, "grad_norm": 0.39453125, "learning_rate": 0.00019894518213559457, "loss": 1.0273, "step": 259 }, { "epoch": 0.14466546112115733, "grad_norm": 0.361328125, "learning_rate": 0.0001989367077665972, "loss": 0.9224, "step": 260 }, { "epoch": 0.14522186674085408, "grad_norm": 0.353515625, "learning_rate": 0.00019892819967421154, "loss": 0.9585, "step": 261 }, { "epoch": 0.14577827236055085, "grad_norm": 0.3828125, "learning_rate": 0.0001989196578613376, "loss": 0.9612, "step": 262 }, { "epoch": 0.1463346779802476, "grad_norm": 0.3515625, "learning_rate": 0.0001989110823308869, "loss": 0.8635, "step": 263 }, { "epoch": 0.14689108359994435, "grad_norm": 0.37109375, "learning_rate": 0.0001989024730857825, "loss": 0.9478, "step": 264 }, { "epoch": 0.14744748921964113, "grad_norm": 0.376953125, "learning_rate": 0.00019889383012895896, "loss": 1.0103, "step": 265 }, { "epoch": 0.14800389483933787, "grad_norm": 0.353515625, "learning_rate": 0.00019888515346336226, "loss": 0.8712, "step": 266 }, { "epoch": 0.14856030045903465, "grad_norm": 0.3671875, "learning_rate": 0.00019887644309195, "loss": 0.9338, "step": 267 }, { "epoch": 0.1491167060787314, "grad_norm": 0.345703125, "learning_rate": 0.00019886769901769104, "loss": 0.9435, "step": 268 }, { "epoch": 0.14967311169842815, "grad_norm": 0.369140625, "learning_rate": 0.00019885892124356597, "loss": 1.0054, "step": 269 }, { "epoch": 0.15022951731812492, "grad_norm": 0.375, "learning_rate": 0.00019885010977256675, "loss": 0.9563, "step": 270 }, { "epoch": 0.15078592293782167, "grad_norm": 0.36328125, "learning_rate": 0.0001988412646076968, "loss": 0.889, "step": 271 }, { "epoch": 0.15134232855751842, "grad_norm": 0.34375, "learning_rate": 0.00019883238575197113, "loss": 0.8797, "step": 272 }, { "epoch": 0.1518987341772152, "grad_norm": 0.357421875, "learning_rate": 0.00019882347320841615, "loss": 0.9045, "step": 273 }, { "epoch": 0.15245513979691194, "grad_norm": 0.34765625, "learning_rate": 0.00019881452698006973, "loss": 0.8376, "step": 274 }, { "epoch": 0.15301154541660872, "grad_norm": 0.361328125, "learning_rate": 0.0001988055470699813, "loss": 0.9139, "step": 275 }, { "epoch": 0.15356795103630547, "grad_norm": 0.361328125, "learning_rate": 0.00019879653348121176, "loss": 0.9675, "step": 276 }, { "epoch": 0.15412435665600221, "grad_norm": 0.345703125, "learning_rate": 0.00019878748621683344, "loss": 0.8359, "step": 277 }, { "epoch": 0.154680762275699, "grad_norm": 0.353515625, "learning_rate": 0.00019877840527993018, "loss": 0.9296, "step": 278 }, { "epoch": 0.15523716789539574, "grad_norm": 0.357421875, "learning_rate": 0.00019876929067359726, "loss": 0.9716, "step": 279 }, { "epoch": 0.15579357351509251, "grad_norm": 0.37109375, "learning_rate": 0.00019876014240094154, "loss": 0.9313, "step": 280 }, { "epoch": 0.15634997913478926, "grad_norm": 0.357421875, "learning_rate": 0.0001987509604650812, "loss": 0.9064, "step": 281 }, { "epoch": 0.156906384754486, "grad_norm": 0.369140625, "learning_rate": 0.00019874174486914602, "loss": 1.0199, "step": 282 }, { "epoch": 0.1574627903741828, "grad_norm": 0.365234375, "learning_rate": 0.00019873249561627723, "loss": 0.9184, "step": 283 }, { "epoch": 0.15801919599387954, "grad_norm": 0.337890625, "learning_rate": 0.0001987232127096275, "loss": 0.8171, "step": 284 }, { "epoch": 0.15857560161357628, "grad_norm": 0.376953125, "learning_rate": 0.00019871389615236094, "loss": 0.9977, "step": 285 }, { "epoch": 0.15913200723327306, "grad_norm": 0.369140625, "learning_rate": 0.00019870454594765323, "loss": 0.9063, "step": 286 }, { "epoch": 0.1596884128529698, "grad_norm": 0.3671875, "learning_rate": 0.00019869516209869144, "loss": 1.0768, "step": 287 }, { "epoch": 0.16024481847266658, "grad_norm": 0.3671875, "learning_rate": 0.00019868574460867416, "loss": 0.9082, "step": 288 }, { "epoch": 0.16080122409236333, "grad_norm": 0.36328125, "learning_rate": 0.00019867629348081138, "loss": 0.9355, "step": 289 }, { "epoch": 0.16135762971206008, "grad_norm": 0.3515625, "learning_rate": 0.0001986668087183246, "loss": 0.8966, "step": 290 }, { "epoch": 0.16191403533175686, "grad_norm": 0.3515625, "learning_rate": 0.0001986572903244468, "loss": 0.8964, "step": 291 }, { "epoch": 0.1624704409514536, "grad_norm": 0.3671875, "learning_rate": 0.00019864773830242244, "loss": 0.9721, "step": 292 }, { "epoch": 0.16302684657115038, "grad_norm": 0.36328125, "learning_rate": 0.0001986381526555073, "loss": 0.9809, "step": 293 }, { "epoch": 0.16358325219084713, "grad_norm": 0.376953125, "learning_rate": 0.0001986285333869688, "loss": 0.9851, "step": 294 }, { "epoch": 0.16413965781054388, "grad_norm": 0.349609375, "learning_rate": 0.00019861888050008575, "loss": 0.8559, "step": 295 }, { "epoch": 0.16469606343024065, "grad_norm": 0.35546875, "learning_rate": 0.00019860919399814837, "loss": 0.9342, "step": 296 }, { "epoch": 0.1652524690499374, "grad_norm": 0.375, "learning_rate": 0.0001985994738844584, "loss": 0.9079, "step": 297 }, { "epoch": 0.16580887466963418, "grad_norm": 0.33984375, "learning_rate": 0.00019858972016232907, "loss": 0.8145, "step": 298 }, { "epoch": 0.16636528028933092, "grad_norm": 0.361328125, "learning_rate": 0.00019857993283508497, "loss": 0.9047, "step": 299 }, { "epoch": 0.16692168590902767, "grad_norm": 0.373046875, "learning_rate": 0.00019857011190606215, "loss": 0.9684, "step": 300 }, { "epoch": 0.16747809152872445, "grad_norm": 0.380859375, "learning_rate": 0.00019856025737860828, "loss": 1.0272, "step": 301 }, { "epoch": 0.1680344971484212, "grad_norm": 0.359375, "learning_rate": 0.0001985503692560822, "loss": 0.9, "step": 302 }, { "epoch": 0.16859090276811795, "grad_norm": 0.376953125, "learning_rate": 0.0001985404475418545, "loss": 1.0117, "step": 303 }, { "epoch": 0.16914730838781472, "grad_norm": 0.3515625, "learning_rate": 0.00019853049223930697, "loss": 0.8626, "step": 304 }, { "epoch": 0.16970371400751147, "grad_norm": 0.359375, "learning_rate": 0.00019852050335183305, "loss": 0.9383, "step": 305 }, { "epoch": 0.17026011962720825, "grad_norm": 0.357421875, "learning_rate": 0.00019851048088283742, "loss": 0.9119, "step": 306 }, { "epoch": 0.170816525246905, "grad_norm": 0.373046875, "learning_rate": 0.00019850042483573644, "loss": 0.9732, "step": 307 }, { "epoch": 0.17137293086660174, "grad_norm": 0.373046875, "learning_rate": 0.00019849033521395773, "loss": 0.9586, "step": 308 }, { "epoch": 0.17192933648629852, "grad_norm": 0.37109375, "learning_rate": 0.00019848021202094042, "loss": 0.943, "step": 309 }, { "epoch": 0.17248574210599527, "grad_norm": 0.361328125, "learning_rate": 0.0001984700552601351, "loss": 0.9244, "step": 310 }, { "epoch": 0.17304214772569204, "grad_norm": 0.369140625, "learning_rate": 0.00019845986493500378, "loss": 1.0124, "step": 311 }, { "epoch": 0.1735985533453888, "grad_norm": 0.3671875, "learning_rate": 0.00019844964104901992, "loss": 1.0024, "step": 312 }, { "epoch": 0.17415495896508554, "grad_norm": 0.3671875, "learning_rate": 0.00019843938360566843, "loss": 0.856, "step": 313 }, { "epoch": 0.1747113645847823, "grad_norm": 0.369140625, "learning_rate": 0.0001984290926084456, "loss": 0.9451, "step": 314 }, { "epoch": 0.17526777020447906, "grad_norm": 0.3671875, "learning_rate": 0.00019841876806085923, "loss": 0.9995, "step": 315 }, { "epoch": 0.17582417582417584, "grad_norm": 0.376953125, "learning_rate": 0.00019840840996642852, "loss": 0.9539, "step": 316 }, { "epoch": 0.17638058144387259, "grad_norm": 0.357421875, "learning_rate": 0.00019839801832868413, "loss": 0.9055, "step": 317 }, { "epoch": 0.17693698706356933, "grad_norm": 0.376953125, "learning_rate": 0.0001983875931511681, "loss": 0.8727, "step": 318 }, { "epoch": 0.1774933926832661, "grad_norm": 0.36328125, "learning_rate": 0.00019837713443743397, "loss": 0.899, "step": 319 }, { "epoch": 0.17804979830296286, "grad_norm": 0.3515625, "learning_rate": 0.00019836664219104666, "loss": 0.8607, "step": 320 }, { "epoch": 0.1786062039226596, "grad_norm": 0.359375, "learning_rate": 0.00019835611641558253, "loss": 1.0124, "step": 321 }, { "epoch": 0.17916260954235638, "grad_norm": 0.369140625, "learning_rate": 0.0001983455571146294, "loss": 1.0013, "step": 322 }, { "epoch": 0.17971901516205313, "grad_norm": 0.34765625, "learning_rate": 0.00019833496429178652, "loss": 0.8477, "step": 323 }, { "epoch": 0.1802754207817499, "grad_norm": 0.357421875, "learning_rate": 0.00019832433795066446, "loss": 0.889, "step": 324 }, { "epoch": 0.18083182640144665, "grad_norm": 0.3359375, "learning_rate": 0.00019831367809488537, "loss": 0.789, "step": 325 }, { "epoch": 0.1813882320211434, "grad_norm": 0.34765625, "learning_rate": 0.0001983029847280827, "loss": 0.8853, "step": 326 }, { "epoch": 0.18194463764084018, "grad_norm": 0.3671875, "learning_rate": 0.0001982922578539014, "loss": 0.9427, "step": 327 }, { "epoch": 0.18250104326053693, "grad_norm": 0.361328125, "learning_rate": 0.0001982814974759978, "loss": 0.8221, "step": 328 }, { "epoch": 0.1830574488802337, "grad_norm": 0.384765625, "learning_rate": 0.00019827070359803968, "loss": 0.9266, "step": 329 }, { "epoch": 0.18361385449993045, "grad_norm": 0.38671875, "learning_rate": 0.0001982598762237062, "loss": 1.0117, "step": 330 }, { "epoch": 0.1841702601196272, "grad_norm": 0.416015625, "learning_rate": 0.00019824901535668796, "loss": 0.9188, "step": 331 }, { "epoch": 0.18472666573932398, "grad_norm": 0.3828125, "learning_rate": 0.000198238121000687, "loss": 0.952, "step": 332 }, { "epoch": 0.18528307135902072, "grad_norm": 0.353515625, "learning_rate": 0.00019822719315941672, "loss": 0.9938, "step": 333 }, { "epoch": 0.18583947697871747, "grad_norm": 0.3515625, "learning_rate": 0.000198216231836602, "loss": 0.8886, "step": 334 }, { "epoch": 0.18639588259841425, "grad_norm": 0.361328125, "learning_rate": 0.00019820523703597908, "loss": 0.8996, "step": 335 }, { "epoch": 0.186952288218111, "grad_norm": 0.359375, "learning_rate": 0.0001981942087612956, "loss": 0.8635, "step": 336 }, { "epoch": 0.18750869383780777, "grad_norm": 0.37890625, "learning_rate": 0.0001981831470163107, "loss": 1.029, "step": 337 }, { "epoch": 0.18806509945750452, "grad_norm": 0.359375, "learning_rate": 0.0001981720518047948, "loss": 0.9771, "step": 338 }, { "epoch": 0.18862150507720127, "grad_norm": 0.36328125, "learning_rate": 0.0001981609231305298, "loss": 0.919, "step": 339 }, { "epoch": 0.18917791069689804, "grad_norm": 0.359375, "learning_rate": 0.00019814976099730904, "loss": 0.9555, "step": 340 }, { "epoch": 0.1897343163165948, "grad_norm": 0.3828125, "learning_rate": 0.00019813856540893722, "loss": 1.0713, "step": 341 }, { "epoch": 0.19029072193629157, "grad_norm": 0.36328125, "learning_rate": 0.00019812733636923044, "loss": 0.9679, "step": 342 }, { "epoch": 0.19084712755598832, "grad_norm": 0.3515625, "learning_rate": 0.0001981160738820162, "loss": 0.9542, "step": 343 }, { "epoch": 0.19140353317568506, "grad_norm": 0.37109375, "learning_rate": 0.0001981047779511334, "loss": 0.991, "step": 344 }, { "epoch": 0.19195993879538184, "grad_norm": 0.498046875, "learning_rate": 0.0001980934485804324, "loss": 0.9589, "step": 345 }, { "epoch": 0.1925163444150786, "grad_norm": 0.3515625, "learning_rate": 0.00019808208577377486, "loss": 0.9242, "step": 346 }, { "epoch": 0.19307275003477536, "grad_norm": 0.375, "learning_rate": 0.0001980706895350339, "loss": 0.9109, "step": 347 }, { "epoch": 0.1936291556544721, "grad_norm": 0.37890625, "learning_rate": 0.00019805925986809403, "loss": 1.0326, "step": 348 }, { "epoch": 0.19418556127416886, "grad_norm": 0.369140625, "learning_rate": 0.00019804779677685112, "loss": 0.9841, "step": 349 }, { "epoch": 0.19474196689386564, "grad_norm": 0.359375, "learning_rate": 0.0001980363002652125, "loss": 0.9203, "step": 350 }, { "epoch": 0.19529837251356238, "grad_norm": 0.33984375, "learning_rate": 0.0001980247703370968, "loss": 0.8824, "step": 351 }, { "epoch": 0.19585477813325913, "grad_norm": 0.326171875, "learning_rate": 0.00019801320699643416, "loss": 0.7956, "step": 352 }, { "epoch": 0.1964111837529559, "grad_norm": 0.369140625, "learning_rate": 0.00019800161024716598, "loss": 0.922, "step": 353 }, { "epoch": 0.19696758937265266, "grad_norm": 0.369140625, "learning_rate": 0.00019798998009324512, "loss": 0.9232, "step": 354 }, { "epoch": 0.19752399499234943, "grad_norm": 0.34765625, "learning_rate": 0.00019797831653863582, "loss": 0.8868, "step": 355 }, { "epoch": 0.19808040061204618, "grad_norm": 0.3515625, "learning_rate": 0.0001979666195873137, "loss": 0.9103, "step": 356 }, { "epoch": 0.19863680623174293, "grad_norm": 0.353515625, "learning_rate": 0.00019795488924326574, "loss": 0.8998, "step": 357 }, { "epoch": 0.1991932118514397, "grad_norm": 0.361328125, "learning_rate": 0.00019794312551049039, "loss": 0.8973, "step": 358 }, { "epoch": 0.19974961747113645, "grad_norm": 0.369140625, "learning_rate": 0.00019793132839299732, "loss": 0.9458, "step": 359 }, { "epoch": 0.20030602309083323, "grad_norm": 0.3671875, "learning_rate": 0.00019791949789480775, "loss": 1.0083, "step": 360 }, { "epoch": 0.20086242871052998, "grad_norm": 0.35546875, "learning_rate": 0.00019790763401995415, "loss": 0.9661, "step": 361 }, { "epoch": 0.20141883433022673, "grad_norm": 0.365234375, "learning_rate": 0.00019789573677248047, "loss": 0.8475, "step": 362 }, { "epoch": 0.2019752399499235, "grad_norm": 0.345703125, "learning_rate": 0.00019788380615644196, "loss": 0.8632, "step": 363 }, { "epoch": 0.20253164556962025, "grad_norm": 0.359375, "learning_rate": 0.00019787184217590523, "loss": 0.9522, "step": 364 }, { "epoch": 0.20308805118931703, "grad_norm": 0.3671875, "learning_rate": 0.00019785984483494838, "loss": 0.8749, "step": 365 }, { "epoch": 0.20364445680901377, "grad_norm": 0.337890625, "learning_rate": 0.00019784781413766073, "loss": 0.8714, "step": 366 }, { "epoch": 0.20420086242871052, "grad_norm": 0.373046875, "learning_rate": 0.0001978357500881431, "loss": 0.9581, "step": 367 }, { "epoch": 0.2047572680484073, "grad_norm": 0.37109375, "learning_rate": 0.00019782365269050756, "loss": 1.0496, "step": 368 }, { "epoch": 0.20531367366810405, "grad_norm": 0.34765625, "learning_rate": 0.00019781152194887764, "loss": 0.8804, "step": 369 }, { "epoch": 0.2058700792878008, "grad_norm": 0.3359375, "learning_rate": 0.00019779935786738822, "loss": 0.8474, "step": 370 }, { "epoch": 0.20642648490749757, "grad_norm": 0.349609375, "learning_rate": 0.0001977871604501855, "loss": 0.8413, "step": 371 }, { "epoch": 0.20698289052719432, "grad_norm": 0.3515625, "learning_rate": 0.00019777492970142707, "loss": 0.8636, "step": 372 }, { "epoch": 0.2075392961468911, "grad_norm": 0.345703125, "learning_rate": 0.00019776266562528192, "loss": 0.8351, "step": 373 }, { "epoch": 0.20809570176658784, "grad_norm": 0.330078125, "learning_rate": 0.0001977503682259303, "loss": 0.8751, "step": 374 }, { "epoch": 0.2086521073862846, "grad_norm": 0.3515625, "learning_rate": 0.00019773803750756393, "loss": 0.9113, "step": 375 }, { "epoch": 0.20920851300598137, "grad_norm": 0.3671875, "learning_rate": 0.0001977256734743858, "loss": 0.9433, "step": 376 }, { "epoch": 0.20976491862567812, "grad_norm": 0.34375, "learning_rate": 0.0001977132761306103, "loss": 0.8863, "step": 377 }, { "epoch": 0.2103213242453749, "grad_norm": 0.34765625, "learning_rate": 0.00019770084548046318, "loss": 0.8073, "step": 378 }, { "epoch": 0.21087772986507164, "grad_norm": 0.345703125, "learning_rate": 0.00019768838152818153, "loss": 0.9198, "step": 379 }, { "epoch": 0.2114341354847684, "grad_norm": 0.357421875, "learning_rate": 0.0001976758842780138, "loss": 0.9204, "step": 380 }, { "epoch": 0.21199054110446516, "grad_norm": 0.359375, "learning_rate": 0.00019766335373421973, "loss": 0.9714, "step": 381 }, { "epoch": 0.2125469467241619, "grad_norm": 0.3671875, "learning_rate": 0.00019765078990107048, "loss": 1.0368, "step": 382 }, { "epoch": 0.21310335234385866, "grad_norm": 0.353515625, "learning_rate": 0.0001976381927828486, "loss": 0.8881, "step": 383 }, { "epoch": 0.21365975796355544, "grad_norm": 0.359375, "learning_rate": 0.00019762556238384784, "loss": 1.0236, "step": 384 }, { "epoch": 0.21421616358325218, "grad_norm": 0.37109375, "learning_rate": 0.0001976128987083734, "loss": 0.9628, "step": 385 }, { "epoch": 0.21477256920294896, "grad_norm": 0.349609375, "learning_rate": 0.0001976002017607418, "loss": 0.8861, "step": 386 }, { "epoch": 0.2153289748226457, "grad_norm": 0.380859375, "learning_rate": 0.00019758747154528092, "loss": 1.0946, "step": 387 }, { "epoch": 0.21588538044234246, "grad_norm": 0.36328125, "learning_rate": 0.00019757470806632994, "loss": 1.0047, "step": 388 }, { "epoch": 0.21644178606203923, "grad_norm": 0.35546875, "learning_rate": 0.00019756191132823937, "loss": 0.8831, "step": 389 }, { "epoch": 0.21699819168173598, "grad_norm": 0.373046875, "learning_rate": 0.00019754908133537113, "loss": 0.8999, "step": 390 }, { "epoch": 0.21755459730143276, "grad_norm": 0.361328125, "learning_rate": 0.00019753621809209842, "loss": 0.8368, "step": 391 }, { "epoch": 0.2181110029211295, "grad_norm": 0.345703125, "learning_rate": 0.00019752332160280576, "loss": 0.9262, "step": 392 }, { "epoch": 0.21866740854082625, "grad_norm": 0.365234375, "learning_rate": 0.000197510391871889, "loss": 0.9496, "step": 393 }, { "epoch": 0.21922381416052303, "grad_norm": 0.353515625, "learning_rate": 0.0001974974289037554, "loss": 0.9303, "step": 394 }, { "epoch": 0.21978021978021978, "grad_norm": 0.375, "learning_rate": 0.0001974844327028235, "loss": 0.924, "step": 395 }, { "epoch": 0.22033662539991655, "grad_norm": 0.3671875, "learning_rate": 0.00019747140327352306, "loss": 0.9763, "step": 396 }, { "epoch": 0.2208930310196133, "grad_norm": 0.3359375, "learning_rate": 0.0001974583406202954, "loss": 0.7589, "step": 397 }, { "epoch": 0.22144943663931005, "grad_norm": 0.35546875, "learning_rate": 0.00019744524474759296, "loss": 0.9391, "step": 398 }, { "epoch": 0.22200584225900682, "grad_norm": 0.376953125, "learning_rate": 0.0001974321156598796, "loss": 0.9961, "step": 399 }, { "epoch": 0.22256224787870357, "grad_norm": 0.35546875, "learning_rate": 0.00019741895336163046, "loss": 0.9214, "step": 400 }, { "epoch": 0.22311865349840032, "grad_norm": 0.359375, "learning_rate": 0.000197405757857332, "loss": 0.8915, "step": 401 }, { "epoch": 0.2236750591180971, "grad_norm": 0.337890625, "learning_rate": 0.00019739252915148208, "loss": 0.8334, "step": 402 }, { "epoch": 0.22423146473779385, "grad_norm": 0.359375, "learning_rate": 0.00019737926724858976, "loss": 0.9007, "step": 403 }, { "epoch": 0.22478787035749062, "grad_norm": 0.353515625, "learning_rate": 0.0001973659721531755, "loss": 0.8694, "step": 404 }, { "epoch": 0.22534427597718737, "grad_norm": 0.34375, "learning_rate": 0.00019735264386977098, "loss": 0.9033, "step": 405 }, { "epoch": 0.22590068159688412, "grad_norm": 0.353515625, "learning_rate": 0.00019733928240291932, "loss": 0.8603, "step": 406 }, { "epoch": 0.2264570872165809, "grad_norm": 0.365234375, "learning_rate": 0.00019732588775717486, "loss": 1.024, "step": 407 }, { "epoch": 0.22701349283627764, "grad_norm": 0.345703125, "learning_rate": 0.00019731245993710327, "loss": 0.8826, "step": 408 }, { "epoch": 0.22756989845597442, "grad_norm": 0.345703125, "learning_rate": 0.00019729899894728158, "loss": 0.8364, "step": 409 }, { "epoch": 0.22812630407567117, "grad_norm": 0.373046875, "learning_rate": 0.000197285504792298, "loss": 1.0239, "step": 410 }, { "epoch": 0.22868270969536791, "grad_norm": 0.3515625, "learning_rate": 0.00019727197747675218, "loss": 0.9286, "step": 411 }, { "epoch": 0.2292391153150647, "grad_norm": 0.3515625, "learning_rate": 0.000197258417005255, "loss": 0.8623, "step": 412 }, { "epoch": 0.22979552093476144, "grad_norm": 0.349609375, "learning_rate": 0.00019724482338242866, "loss": 0.8571, "step": 413 }, { "epoch": 0.2303519265544582, "grad_norm": 0.36328125, "learning_rate": 0.00019723119661290664, "loss": 1.0023, "step": 414 }, { "epoch": 0.23090833217415496, "grad_norm": 0.345703125, "learning_rate": 0.00019721753670133376, "loss": 0.8866, "step": 415 }, { "epoch": 0.2314647377938517, "grad_norm": 0.33203125, "learning_rate": 0.0001972038436523661, "loss": 0.7956, "step": 416 }, { "epoch": 0.2320211434135485, "grad_norm": 0.369140625, "learning_rate": 0.00019719011747067108, "loss": 0.9645, "step": 417 }, { "epoch": 0.23257754903324523, "grad_norm": 0.37890625, "learning_rate": 0.00019717635816092732, "loss": 0.9324, "step": 418 }, { "epoch": 0.23313395465294198, "grad_norm": 0.361328125, "learning_rate": 0.00019716256572782482, "loss": 0.9626, "step": 419 }, { "epoch": 0.23369036027263876, "grad_norm": 0.34375, "learning_rate": 0.0001971487401760649, "loss": 0.8419, "step": 420 }, { "epoch": 0.2342467658923355, "grad_norm": 0.33984375, "learning_rate": 0.00019713488151036003, "loss": 0.8471, "step": 421 }, { "epoch": 0.23480317151203228, "grad_norm": 0.361328125, "learning_rate": 0.00019712098973543408, "loss": 1.0177, "step": 422 }, { "epoch": 0.23535957713172903, "grad_norm": 0.35546875, "learning_rate": 0.0001971070648560222, "loss": 0.944, "step": 423 }, { "epoch": 0.23591598275142578, "grad_norm": 0.3359375, "learning_rate": 0.0001970931068768708, "loss": 0.8007, "step": 424 }, { "epoch": 0.23647238837112255, "grad_norm": 0.3671875, "learning_rate": 0.0001970791158027375, "loss": 0.9658, "step": 425 }, { "epoch": 0.2370287939908193, "grad_norm": 0.35546875, "learning_rate": 0.00019706509163839136, "loss": 0.8966, "step": 426 }, { "epoch": 0.23758519961051608, "grad_norm": 0.365234375, "learning_rate": 0.0001970510343886126, "loss": 1.0448, "step": 427 }, { "epoch": 0.23814160523021283, "grad_norm": 0.37109375, "learning_rate": 0.0001970369440581927, "loss": 0.937, "step": 428 }, { "epoch": 0.23869801084990958, "grad_norm": 0.3515625, "learning_rate": 0.00019702282065193455, "loss": 0.9181, "step": 429 }, { "epoch": 0.23925441646960635, "grad_norm": 0.34375, "learning_rate": 0.00019700866417465216, "loss": 0.9101, "step": 430 }, { "epoch": 0.2398108220893031, "grad_norm": 0.359375, "learning_rate": 0.00019699447463117093, "loss": 0.9438, "step": 431 }, { "epoch": 0.24036722770899985, "grad_norm": 0.34765625, "learning_rate": 0.00019698025202632743, "loss": 0.8315, "step": 432 }, { "epoch": 0.24092363332869662, "grad_norm": 0.353515625, "learning_rate": 0.00019696599636496962, "loss": 0.8385, "step": 433 }, { "epoch": 0.24148003894839337, "grad_norm": 0.33984375, "learning_rate": 0.0001969517076519566, "loss": 0.8667, "step": 434 }, { "epoch": 0.24203644456809015, "grad_norm": 0.349609375, "learning_rate": 0.00019693738589215883, "loss": 0.881, "step": 435 }, { "epoch": 0.2425928501877869, "grad_norm": 0.36328125, "learning_rate": 0.00019692303109045795, "loss": 0.9509, "step": 436 }, { "epoch": 0.24314925580748364, "grad_norm": 0.357421875, "learning_rate": 0.00019690864325174699, "loss": 0.9065, "step": 437 }, { "epoch": 0.24370566142718042, "grad_norm": 0.380859375, "learning_rate": 0.0001968942223809301, "loss": 1.072, "step": 438 }, { "epoch": 0.24426206704687717, "grad_norm": 0.333984375, "learning_rate": 0.0001968797684829228, "loss": 0.9068, "step": 439 }, { "epoch": 0.24481847266657394, "grad_norm": 0.33203125, "learning_rate": 0.00019686528156265175, "loss": 0.8168, "step": 440 }, { "epoch": 0.2453748782862707, "grad_norm": 0.3515625, "learning_rate": 0.000196850761625055, "loss": 0.8748, "step": 441 }, { "epoch": 0.24593128390596744, "grad_norm": 0.330078125, "learning_rate": 0.00019683620867508176, "loss": 0.7686, "step": 442 }, { "epoch": 0.24648768952566422, "grad_norm": 0.365234375, "learning_rate": 0.00019682162271769256, "loss": 0.9036, "step": 443 }, { "epoch": 0.24704409514536096, "grad_norm": 0.357421875, "learning_rate": 0.0001968070037578591, "loss": 0.9472, "step": 444 }, { "epoch": 0.24760050076505774, "grad_norm": 0.33984375, "learning_rate": 0.00019679235180056437, "loss": 0.8882, "step": 445 }, { "epoch": 0.2481569063847545, "grad_norm": 0.36328125, "learning_rate": 0.00019677766685080264, "loss": 0.9693, "step": 446 }, { "epoch": 0.24871331200445124, "grad_norm": 0.357421875, "learning_rate": 0.00019676294891357937, "loss": 0.8764, "step": 447 }, { "epoch": 0.249269717624148, "grad_norm": 0.38671875, "learning_rate": 0.00019674819799391129, "loss": 0.9671, "step": 448 }, { "epoch": 0.24982612324384476, "grad_norm": 0.345703125, "learning_rate": 0.0001967334140968264, "loss": 0.9398, "step": 449 }, { "epoch": 0.2503825288635415, "grad_norm": 0.345703125, "learning_rate": 0.00019671859722736389, "loss": 0.9572, "step": 450 }, { "epoch": 0.2509389344832383, "grad_norm": 0.357421875, "learning_rate": 0.0001967037473905742, "loss": 0.958, "step": 451 }, { "epoch": 0.25149534010293506, "grad_norm": 0.35546875, "learning_rate": 0.00019668886459151908, "loss": 0.8959, "step": 452 }, { "epoch": 0.2520517457226318, "grad_norm": 0.35546875, "learning_rate": 0.00019667394883527139, "loss": 0.9529, "step": 453 }, { "epoch": 0.25260815134232856, "grad_norm": 0.361328125, "learning_rate": 0.00019665900012691529, "loss": 0.9213, "step": 454 }, { "epoch": 0.25316455696202533, "grad_norm": 0.359375, "learning_rate": 0.0001966440184715462, "loss": 0.952, "step": 455 }, { "epoch": 0.25372096258172205, "grad_norm": 0.365234375, "learning_rate": 0.00019662900387427075, "loss": 0.9548, "step": 456 }, { "epoch": 0.25427736820141883, "grad_norm": 0.3359375, "learning_rate": 0.00019661395634020673, "loss": 0.8359, "step": 457 }, { "epoch": 0.2548337738211156, "grad_norm": 0.345703125, "learning_rate": 0.00019659887587448327, "loss": 0.9204, "step": 458 }, { "epoch": 0.2553901794408123, "grad_norm": 0.359375, "learning_rate": 0.00019658376248224063, "loss": 0.9216, "step": 459 }, { "epoch": 0.2559465850605091, "grad_norm": 0.365234375, "learning_rate": 0.00019656861616863038, "loss": 0.9383, "step": 460 }, { "epoch": 0.2565029906802059, "grad_norm": 0.35546875, "learning_rate": 0.00019655343693881526, "loss": 0.9428, "step": 461 }, { "epoch": 0.25705939629990265, "grad_norm": 0.3515625, "learning_rate": 0.0001965382247979692, "loss": 0.9357, "step": 462 }, { "epoch": 0.2576158019195994, "grad_norm": 0.35546875, "learning_rate": 0.00019652297975127737, "loss": 0.967, "step": 463 }, { "epoch": 0.25817220753929615, "grad_norm": 0.349609375, "learning_rate": 0.0001965077018039362, "loss": 0.8993, "step": 464 }, { "epoch": 0.2587286131589929, "grad_norm": 0.359375, "learning_rate": 0.00019649239096115336, "loss": 1.0113, "step": 465 }, { "epoch": 0.25928501877868965, "grad_norm": 0.33984375, "learning_rate": 0.00019647704722814755, "loss": 0.9117, "step": 466 }, { "epoch": 0.2598414243983864, "grad_norm": 0.345703125, "learning_rate": 0.00019646167061014892, "loss": 0.9318, "step": 467 }, { "epoch": 0.2603978300180832, "grad_norm": 0.34375, "learning_rate": 0.00019644626111239863, "loss": 0.9237, "step": 468 }, { "epoch": 0.2609542356377799, "grad_norm": 0.37109375, "learning_rate": 0.0001964308187401492, "loss": 0.9926, "step": 469 }, { "epoch": 0.2615106412574767, "grad_norm": 0.337890625, "learning_rate": 0.0001964153434986643, "loss": 0.8647, "step": 470 }, { "epoch": 0.26206704687717347, "grad_norm": 0.330078125, "learning_rate": 0.0001963998353932187, "loss": 0.7973, "step": 471 }, { "epoch": 0.2626234524968702, "grad_norm": 0.35546875, "learning_rate": 0.00019638429442909853, "loss": 0.8922, "step": 472 }, { "epoch": 0.26317985811656697, "grad_norm": 0.3671875, "learning_rate": 0.00019636872061160109, "loss": 0.9267, "step": 473 }, { "epoch": 0.26373626373626374, "grad_norm": 0.34765625, "learning_rate": 0.00019635311394603476, "loss": 0.8743, "step": 474 }, { "epoch": 0.2642926693559605, "grad_norm": 0.33203125, "learning_rate": 0.00019633747443771924, "loss": 0.7997, "step": 475 }, { "epoch": 0.26484907497565724, "grad_norm": 0.36328125, "learning_rate": 0.00019632180209198542, "loss": 0.9695, "step": 476 }, { "epoch": 0.265405480595354, "grad_norm": 0.3359375, "learning_rate": 0.00019630609691417527, "loss": 0.7992, "step": 477 }, { "epoch": 0.2659618862150508, "grad_norm": 0.3125, "learning_rate": 0.00019629035890964208, "loss": 0.7188, "step": 478 }, { "epoch": 0.2665182918347475, "grad_norm": 0.384765625, "learning_rate": 0.0001962745880837503, "loss": 0.93, "step": 479 }, { "epoch": 0.2670746974544443, "grad_norm": 0.341796875, "learning_rate": 0.00019625878444187551, "loss": 0.8121, "step": 480 }, { "epoch": 0.26763110307414106, "grad_norm": 0.357421875, "learning_rate": 0.00019624294798940451, "loss": 0.9366, "step": 481 }, { "epoch": 0.2681875086938378, "grad_norm": 0.359375, "learning_rate": 0.0001962270787317353, "loss": 0.94, "step": 482 }, { "epoch": 0.26874391431353456, "grad_norm": 0.357421875, "learning_rate": 0.000196211176674277, "loss": 0.9317, "step": 483 }, { "epoch": 0.26930031993323134, "grad_norm": 0.34375, "learning_rate": 0.00019619524182245, "loss": 0.8751, "step": 484 }, { "epoch": 0.2698567255529281, "grad_norm": 0.345703125, "learning_rate": 0.00019617927418168587, "loss": 0.9093, "step": 485 }, { "epoch": 0.27041313117262483, "grad_norm": 0.33984375, "learning_rate": 0.00019616327375742722, "loss": 0.8399, "step": 486 }, { "epoch": 0.2709695367923216, "grad_norm": 0.353515625, "learning_rate": 0.00019614724055512799, "loss": 0.9025, "step": 487 }, { "epoch": 0.2715259424120184, "grad_norm": 0.353515625, "learning_rate": 0.0001961311745802532, "loss": 0.8661, "step": 488 }, { "epoch": 0.2720823480317151, "grad_norm": 0.34375, "learning_rate": 0.00019611507583827903, "loss": 0.8429, "step": 489 }, { "epoch": 0.2726387536514119, "grad_norm": 0.3515625, "learning_rate": 0.00019609894433469295, "loss": 0.9272, "step": 490 }, { "epoch": 0.27319515927110866, "grad_norm": 0.353515625, "learning_rate": 0.0001960827800749935, "loss": 0.87, "step": 491 }, { "epoch": 0.2737515648908054, "grad_norm": 0.373046875, "learning_rate": 0.00019606658306469034, "loss": 0.865, "step": 492 }, { "epoch": 0.27430797051050215, "grad_norm": 0.328125, "learning_rate": 0.00019605035330930442, "loss": 0.8939, "step": 493 }, { "epoch": 0.27486437613019893, "grad_norm": 0.3515625, "learning_rate": 0.00019603409081436773, "loss": 0.9009, "step": 494 }, { "epoch": 0.27542078174989565, "grad_norm": 0.349609375, "learning_rate": 0.0001960177955854235, "loss": 0.8828, "step": 495 }, { "epoch": 0.2759771873695924, "grad_norm": 0.349609375, "learning_rate": 0.00019600146762802613, "loss": 0.9509, "step": 496 }, { "epoch": 0.2765335929892892, "grad_norm": 0.333984375, "learning_rate": 0.00019598510694774108, "loss": 0.7998, "step": 497 }, { "epoch": 0.277089998608986, "grad_norm": 0.349609375, "learning_rate": 0.00019596871355014506, "loss": 0.8909, "step": 498 }, { "epoch": 0.2776464042286827, "grad_norm": 0.36328125, "learning_rate": 0.00019595228744082586, "loss": 0.9643, "step": 499 }, { "epoch": 0.2782028098483795, "grad_norm": 0.3515625, "learning_rate": 0.00019593582862538248, "loss": 0.9311, "step": 500 }, { "epoch": 0.27875921546807625, "grad_norm": 0.33984375, "learning_rate": 0.00019591933710942503, "loss": 0.8344, "step": 501 }, { "epoch": 0.27931562108777297, "grad_norm": 0.34765625, "learning_rate": 0.00019590281289857478, "loss": 0.9109, "step": 502 }, { "epoch": 0.27987202670746975, "grad_norm": 0.357421875, "learning_rate": 0.00019588625599846417, "loss": 0.994, "step": 503 }, { "epoch": 0.2804284323271665, "grad_norm": 0.35546875, "learning_rate": 0.00019586966641473673, "loss": 0.8917, "step": 504 }, { "epoch": 0.28098483794686324, "grad_norm": 0.33984375, "learning_rate": 0.00019585304415304711, "loss": 0.8417, "step": 505 }, { "epoch": 0.28154124356656, "grad_norm": 0.341796875, "learning_rate": 0.00019583638921906124, "loss": 0.9425, "step": 506 }, { "epoch": 0.2820976491862568, "grad_norm": 0.333984375, "learning_rate": 0.00019581970161845603, "loss": 0.8792, "step": 507 }, { "epoch": 0.2826540548059535, "grad_norm": 0.349609375, "learning_rate": 0.0001958029813569196, "loss": 0.9791, "step": 508 }, { "epoch": 0.2832104604256503, "grad_norm": 0.330078125, "learning_rate": 0.00019578622844015117, "loss": 0.8546, "step": 509 }, { "epoch": 0.28376686604534707, "grad_norm": 0.359375, "learning_rate": 0.00019576944287386113, "loss": 0.8836, "step": 510 }, { "epoch": 0.28432327166504384, "grad_norm": 0.341796875, "learning_rate": 0.000195752624663771, "loss": 0.8764, "step": 511 }, { "epoch": 0.28487967728474056, "grad_norm": 0.369140625, "learning_rate": 0.00019573577381561336, "loss": 0.9411, "step": 512 }, { "epoch": 0.28543608290443734, "grad_norm": 0.345703125, "learning_rate": 0.00019571889033513198, "loss": 0.9222, "step": 513 }, { "epoch": 0.2859924885241341, "grad_norm": 0.365234375, "learning_rate": 0.00019570197422808173, "loss": 0.9914, "step": 514 }, { "epoch": 0.28654889414383083, "grad_norm": 0.34765625, "learning_rate": 0.00019568502550022858, "loss": 0.8788, "step": 515 }, { "epoch": 0.2871052997635276, "grad_norm": 0.3359375, "learning_rate": 0.0001956680441573497, "loss": 0.811, "step": 516 }, { "epoch": 0.2876617053832244, "grad_norm": 0.3515625, "learning_rate": 0.00019565103020523328, "loss": 0.9191, "step": 517 }, { "epoch": 0.2882181110029211, "grad_norm": 0.345703125, "learning_rate": 0.00019563398364967868, "loss": 0.9129, "step": 518 }, { "epoch": 0.2887745166226179, "grad_norm": 0.357421875, "learning_rate": 0.00019561690449649636, "loss": 0.9591, "step": 519 }, { "epoch": 0.28933092224231466, "grad_norm": 0.326171875, "learning_rate": 0.00019559979275150783, "loss": 0.8129, "step": 520 }, { "epoch": 0.2898873278620114, "grad_norm": 0.349609375, "learning_rate": 0.00019558264842054585, "loss": 0.8881, "step": 521 }, { "epoch": 0.29044373348170816, "grad_norm": 0.37890625, "learning_rate": 0.00019556547150945418, "loss": 0.8941, "step": 522 }, { "epoch": 0.29100013910140493, "grad_norm": 0.33984375, "learning_rate": 0.0001955482620240877, "loss": 0.8872, "step": 523 }, { "epoch": 0.2915565447211017, "grad_norm": 0.3359375, "learning_rate": 0.0001955310199703124, "loss": 0.7887, "step": 524 }, { "epoch": 0.2921129503407984, "grad_norm": 0.34375, "learning_rate": 0.00019551374535400543, "loss": 0.8684, "step": 525 }, { "epoch": 0.2926693559604952, "grad_norm": 0.3203125, "learning_rate": 0.0001954964381810549, "loss": 0.7739, "step": 526 }, { "epoch": 0.293225761580192, "grad_norm": 0.33203125, "learning_rate": 0.00019547909845736013, "loss": 0.8568, "step": 527 }, { "epoch": 0.2937821671998887, "grad_norm": 0.353515625, "learning_rate": 0.00019546172618883155, "loss": 0.9188, "step": 528 }, { "epoch": 0.2943385728195855, "grad_norm": 0.330078125, "learning_rate": 0.00019544432138139062, "loss": 0.8119, "step": 529 }, { "epoch": 0.29489497843928225, "grad_norm": 0.34765625, "learning_rate": 0.00019542688404096986, "loss": 0.9192, "step": 530 }, { "epoch": 0.29545138405897897, "grad_norm": 0.33984375, "learning_rate": 0.000195409414173513, "loss": 0.8616, "step": 531 }, { "epoch": 0.29600778967867575, "grad_norm": 0.34375, "learning_rate": 0.00019539191178497472, "loss": 0.9191, "step": 532 }, { "epoch": 0.2965641952983725, "grad_norm": 0.32421875, "learning_rate": 0.00019537437688132094, "loss": 0.7765, "step": 533 }, { "epoch": 0.2971206009180693, "grad_norm": 0.34765625, "learning_rate": 0.00019535680946852852, "loss": 0.9185, "step": 534 }, { "epoch": 0.297677006537766, "grad_norm": 0.353515625, "learning_rate": 0.00019533920955258546, "loss": 0.8514, "step": 535 }, { "epoch": 0.2982334121574628, "grad_norm": 0.80078125, "learning_rate": 0.00019532157713949083, "loss": 1.8071, "step": 536 }, { "epoch": 0.29878981777715957, "grad_norm": 0.361328125, "learning_rate": 0.00019530391223525478, "loss": 0.9006, "step": 537 }, { "epoch": 0.2993462233968563, "grad_norm": 0.357421875, "learning_rate": 0.00019528621484589859, "loss": 0.9186, "step": 538 }, { "epoch": 0.29990262901655307, "grad_norm": 0.333984375, "learning_rate": 0.00019526848497745448, "loss": 0.8292, "step": 539 }, { "epoch": 0.30045903463624984, "grad_norm": 0.3359375, "learning_rate": 0.00019525072263596585, "loss": 0.7303, "step": 540 }, { "epoch": 0.30101544025594656, "grad_norm": 0.357421875, "learning_rate": 0.0001952329278274872, "loss": 0.9214, "step": 541 }, { "epoch": 0.30157184587564334, "grad_norm": 0.3515625, "learning_rate": 0.00019521510055808396, "loss": 0.9173, "step": 542 }, { "epoch": 0.3021282514953401, "grad_norm": 0.345703125, "learning_rate": 0.00019519724083383273, "loss": 0.8349, "step": 543 }, { "epoch": 0.30268465711503684, "grad_norm": 0.353515625, "learning_rate": 0.00019517934866082116, "loss": 0.9247, "step": 544 }, { "epoch": 0.3032410627347336, "grad_norm": 0.3671875, "learning_rate": 0.0001951614240451479, "loss": 0.8788, "step": 545 }, { "epoch": 0.3037974683544304, "grad_norm": 0.33203125, "learning_rate": 0.00019514346699292274, "loss": 0.8521, "step": 546 }, { "epoch": 0.30435387397412716, "grad_norm": 0.34375, "learning_rate": 0.00019512547751026648, "loss": 0.8488, "step": 547 }, { "epoch": 0.3049102795938239, "grad_norm": 0.3671875, "learning_rate": 0.00019510745560331097, "loss": 0.9655, "step": 548 }, { "epoch": 0.30546668521352066, "grad_norm": 0.375, "learning_rate": 0.00019508940127819913, "loss": 1.0282, "step": 549 }, { "epoch": 0.30602309083321744, "grad_norm": 0.37109375, "learning_rate": 0.00019507131454108497, "loss": 1.0152, "step": 550 }, { "epoch": 0.30657949645291416, "grad_norm": 0.330078125, "learning_rate": 0.0001950531953981334, "loss": 0.8472, "step": 551 }, { "epoch": 0.30713590207261093, "grad_norm": 0.345703125, "learning_rate": 0.0001950350438555206, "loss": 0.8322, "step": 552 }, { "epoch": 0.3076923076923077, "grad_norm": 0.357421875, "learning_rate": 0.00019501685991943357, "loss": 0.9437, "step": 553 }, { "epoch": 0.30824871331200443, "grad_norm": 0.330078125, "learning_rate": 0.0001949986435960705, "loss": 0.817, "step": 554 }, { "epoch": 0.3088051189317012, "grad_norm": 0.53125, "learning_rate": 0.00019498039489164058, "loss": 0.9954, "step": 555 }, { "epoch": 0.309361524551398, "grad_norm": 0.34765625, "learning_rate": 0.00019496211381236401, "loss": 0.8284, "step": 556 }, { "epoch": 0.3099179301710947, "grad_norm": 0.357421875, "learning_rate": 0.0001949438003644721, "loss": 0.8699, "step": 557 }, { "epoch": 0.3104743357907915, "grad_norm": 0.3515625, "learning_rate": 0.00019492545455420703, "loss": 0.8738, "step": 558 }, { "epoch": 0.31103074141048825, "grad_norm": 0.34765625, "learning_rate": 0.0001949070763878222, "loss": 0.894, "step": 559 }, { "epoch": 0.31158714703018503, "grad_norm": 0.33203125, "learning_rate": 0.00019488866587158196, "loss": 0.8258, "step": 560 }, { "epoch": 0.31214355264988175, "grad_norm": 0.3671875, "learning_rate": 0.00019487022301176168, "loss": 0.9347, "step": 561 }, { "epoch": 0.3126999582695785, "grad_norm": 0.34765625, "learning_rate": 0.00019485174781464773, "loss": 0.9083, "step": 562 }, { "epoch": 0.3132563638892753, "grad_norm": 0.330078125, "learning_rate": 0.00019483324028653754, "loss": 0.784, "step": 563 }, { "epoch": 0.313812769508972, "grad_norm": 0.345703125, "learning_rate": 0.0001948147004337396, "loss": 0.8954, "step": 564 }, { "epoch": 0.3143691751286688, "grad_norm": 0.345703125, "learning_rate": 0.0001947961282625733, "loss": 0.8873, "step": 565 }, { "epoch": 0.3149255807483656, "grad_norm": 0.337890625, "learning_rate": 0.00019477752377936917, "loss": 0.9064, "step": 566 }, { "epoch": 0.3154819863680623, "grad_norm": 0.333984375, "learning_rate": 0.00019475888699046866, "loss": 0.8424, "step": 567 }, { "epoch": 0.31603839198775907, "grad_norm": 0.384765625, "learning_rate": 0.0001947402179022243, "loss": 0.9908, "step": 568 }, { "epoch": 0.31659479760745585, "grad_norm": 0.353515625, "learning_rate": 0.00019472151652099963, "loss": 0.9116, "step": 569 }, { "epoch": 0.31715120322715257, "grad_norm": 0.3359375, "learning_rate": 0.0001947027828531691, "loss": 0.9074, "step": 570 }, { "epoch": 0.31770760884684934, "grad_norm": 0.357421875, "learning_rate": 0.00019468401690511832, "loss": 0.9643, "step": 571 }, { "epoch": 0.3182640144665461, "grad_norm": 0.353515625, "learning_rate": 0.00019466521868324375, "loss": 0.9002, "step": 572 }, { "epoch": 0.3188204200862429, "grad_norm": 0.357421875, "learning_rate": 0.0001946463881939529, "loss": 0.9181, "step": 573 }, { "epoch": 0.3193768257059396, "grad_norm": 0.34765625, "learning_rate": 0.00019462752544366436, "loss": 0.9275, "step": 574 }, { "epoch": 0.3199332313256364, "grad_norm": 0.341796875, "learning_rate": 0.0001946086304388076, "loss": 0.8565, "step": 575 }, { "epoch": 0.32048963694533317, "grad_norm": 0.3671875, "learning_rate": 0.0001945897031858232, "loss": 0.9334, "step": 576 }, { "epoch": 0.3210460425650299, "grad_norm": 0.35546875, "learning_rate": 0.0001945707436911626, "loss": 0.9068, "step": 577 }, { "epoch": 0.32160244818472666, "grad_norm": 0.34765625, "learning_rate": 0.00019455175196128838, "loss": 0.8463, "step": 578 }, { "epoch": 0.32215885380442344, "grad_norm": 0.349609375, "learning_rate": 0.00019453272800267393, "loss": 0.8119, "step": 579 }, { "epoch": 0.32271525942412016, "grad_norm": 0.349609375, "learning_rate": 0.00019451367182180378, "loss": 0.9386, "step": 580 }, { "epoch": 0.32327166504381694, "grad_norm": 0.349609375, "learning_rate": 0.00019449458342517338, "loss": 0.9527, "step": 581 }, { "epoch": 0.3238280706635137, "grad_norm": 0.359375, "learning_rate": 0.00019447546281928916, "loss": 0.9448, "step": 582 }, { "epoch": 0.3243844762832105, "grad_norm": 0.330078125, "learning_rate": 0.00019445631001066853, "loss": 0.8528, "step": 583 }, { "epoch": 0.3249408819029072, "grad_norm": 0.353515625, "learning_rate": 0.00019443712500583993, "loss": 0.9443, "step": 584 }, { "epoch": 0.325497287522604, "grad_norm": 0.361328125, "learning_rate": 0.00019441790781134265, "loss": 0.9369, "step": 585 }, { "epoch": 0.32605369314230076, "grad_norm": 0.34765625, "learning_rate": 0.0001943986584337271, "loss": 0.8631, "step": 586 }, { "epoch": 0.3266100987619975, "grad_norm": 0.359375, "learning_rate": 0.0001943793768795545, "loss": 0.921, "step": 587 }, { "epoch": 0.32716650438169426, "grad_norm": 0.345703125, "learning_rate": 0.0001943600631553972, "loss": 0.8827, "step": 588 }, { "epoch": 0.32772291000139103, "grad_norm": 0.36328125, "learning_rate": 0.00019434071726783843, "loss": 0.9897, "step": 589 }, { "epoch": 0.32827931562108775, "grad_norm": 0.333984375, "learning_rate": 0.0001943213392234724, "loss": 0.7868, "step": 590 }, { "epoch": 0.32883572124078453, "grad_norm": 0.34765625, "learning_rate": 0.00019430192902890423, "loss": 0.8787, "step": 591 }, { "epoch": 0.3293921268604813, "grad_norm": 0.341796875, "learning_rate": 0.0001942824866907501, "loss": 0.861, "step": 592 }, { "epoch": 0.329948532480178, "grad_norm": 0.337890625, "learning_rate": 0.00019426301221563702, "loss": 0.8415, "step": 593 }, { "epoch": 0.3305049380998748, "grad_norm": 0.361328125, "learning_rate": 0.00019424350561020312, "loss": 0.9665, "step": 594 }, { "epoch": 0.3310613437195716, "grad_norm": 0.353515625, "learning_rate": 0.0001942239668810973, "loss": 0.934, "step": 595 }, { "epoch": 0.33161774933926835, "grad_norm": 0.353515625, "learning_rate": 0.00019420439603497956, "loss": 0.9518, "step": 596 }, { "epoch": 0.3321741549589651, "grad_norm": 0.330078125, "learning_rate": 0.00019418479307852074, "loss": 0.8112, "step": 597 }, { "epoch": 0.33273056057866185, "grad_norm": 0.33203125, "learning_rate": 0.0001941651580184027, "loss": 0.7649, "step": 598 }, { "epoch": 0.3332869661983586, "grad_norm": 0.3515625, "learning_rate": 0.00019414549086131815, "loss": 0.9321, "step": 599 }, { "epoch": 0.33384337181805535, "grad_norm": 0.3671875, "learning_rate": 0.0001941257916139709, "loss": 0.9537, "step": 600 }, { "epoch": 0.3343997774377521, "grad_norm": 0.3515625, "learning_rate": 0.0001941060602830755, "loss": 0.961, "step": 601 }, { "epoch": 0.3349561830574489, "grad_norm": 0.33984375, "learning_rate": 0.00019408629687535763, "loss": 0.9001, "step": 602 }, { "epoch": 0.3355125886771456, "grad_norm": 0.326171875, "learning_rate": 0.00019406650139755375, "loss": 0.8772, "step": 603 }, { "epoch": 0.3360689942968424, "grad_norm": 0.357421875, "learning_rate": 0.00019404667385641128, "loss": 0.9444, "step": 604 }, { "epoch": 0.33662539991653917, "grad_norm": 0.3359375, "learning_rate": 0.0001940268142586887, "loss": 0.8729, "step": 605 }, { "epoch": 0.3371818055362359, "grad_norm": 0.333984375, "learning_rate": 0.0001940069226111552, "loss": 0.8825, "step": 606 }, { "epoch": 0.33773821115593267, "grad_norm": 0.361328125, "learning_rate": 0.00019398699892059112, "loss": 0.9166, "step": 607 }, { "epoch": 0.33829461677562944, "grad_norm": 0.35546875, "learning_rate": 0.00019396704319378754, "loss": 0.9447, "step": 608 }, { "epoch": 0.3388510223953262, "grad_norm": 0.3515625, "learning_rate": 0.0001939470554375466, "loss": 0.932, "step": 609 }, { "epoch": 0.33940742801502294, "grad_norm": 0.345703125, "learning_rate": 0.0001939270356586812, "loss": 0.8981, "step": 610 }, { "epoch": 0.3399638336347197, "grad_norm": 0.337890625, "learning_rate": 0.00019390698386401533, "loss": 0.8561, "step": 611 }, { "epoch": 0.3405202392544165, "grad_norm": 0.359375, "learning_rate": 0.00019388690006038375, "loss": 0.8992, "step": 612 }, { "epoch": 0.3410766448741132, "grad_norm": 0.3515625, "learning_rate": 0.00019386678425463221, "loss": 0.952, "step": 613 }, { "epoch": 0.34163305049381, "grad_norm": 0.34375, "learning_rate": 0.00019384663645361736, "loss": 0.8244, "step": 614 }, { "epoch": 0.34218945611350676, "grad_norm": 0.357421875, "learning_rate": 0.00019382645666420674, "loss": 0.9347, "step": 615 }, { "epoch": 0.3427458617332035, "grad_norm": 0.357421875, "learning_rate": 0.00019380624489327876, "loss": 0.9098, "step": 616 }, { "epoch": 0.34330226735290026, "grad_norm": 1.4375, "learning_rate": 0.00019378600114772283, "loss": 1.9243, "step": 617 }, { "epoch": 0.34385867297259703, "grad_norm": 0.326171875, "learning_rate": 0.00019376572543443915, "loss": 0.8078, "step": 618 }, { "epoch": 0.34441507859229376, "grad_norm": 0.333984375, "learning_rate": 0.00019374541776033888, "loss": 0.8494, "step": 619 }, { "epoch": 0.34497148421199053, "grad_norm": 0.3515625, "learning_rate": 0.00019372507813234405, "loss": 0.8841, "step": 620 }, { "epoch": 0.3455278898316873, "grad_norm": 0.33984375, "learning_rate": 0.0001937047065573876, "loss": 0.8424, "step": 621 }, { "epoch": 0.3460842954513841, "grad_norm": 0.333984375, "learning_rate": 0.0001936843030424133, "loss": 0.8028, "step": 622 }, { "epoch": 0.3466407010710808, "grad_norm": 0.337890625, "learning_rate": 0.00019366386759437596, "loss": 0.8883, "step": 623 }, { "epoch": 0.3471971066907776, "grad_norm": 0.328125, "learning_rate": 0.0001936434002202411, "loss": 0.7923, "step": 624 }, { "epoch": 0.34775351231047436, "grad_norm": 0.357421875, "learning_rate": 0.00019362290092698515, "loss": 0.8398, "step": 625 }, { "epoch": 0.3483099179301711, "grad_norm": 0.330078125, "learning_rate": 0.0001936023697215956, "loss": 0.866, "step": 626 }, { "epoch": 0.34886632354986785, "grad_norm": 0.328125, "learning_rate": 0.00019358180661107051, "loss": 0.8925, "step": 627 }, { "epoch": 0.3494227291695646, "grad_norm": 0.33984375, "learning_rate": 0.0001935612116024191, "loss": 0.8689, "step": 628 }, { "epoch": 0.34997913478926135, "grad_norm": 0.345703125, "learning_rate": 0.0001935405847026613, "loss": 0.8397, "step": 629 }, { "epoch": 0.3505355404089581, "grad_norm": 0.34375, "learning_rate": 0.00019351992591882798, "loss": 0.8053, "step": 630 }, { "epoch": 0.3510919460286549, "grad_norm": 0.3515625, "learning_rate": 0.00019349923525796084, "loss": 0.8547, "step": 631 }, { "epoch": 0.3516483516483517, "grad_norm": 0.3515625, "learning_rate": 0.00019347851272711247, "loss": 0.944, "step": 632 }, { "epoch": 0.3522047572680484, "grad_norm": 0.35546875, "learning_rate": 0.00019345775833334633, "loss": 0.8962, "step": 633 }, { "epoch": 0.35276116288774517, "grad_norm": 0.33203125, "learning_rate": 0.0001934369720837367, "loss": 0.8379, "step": 634 }, { "epoch": 0.35331756850744195, "grad_norm": 0.345703125, "learning_rate": 0.00019341615398536877, "loss": 0.9129, "step": 635 }, { "epoch": 0.35387397412713867, "grad_norm": 0.32421875, "learning_rate": 0.00019339530404533854, "loss": 0.8244, "step": 636 }, { "epoch": 0.35443037974683544, "grad_norm": 0.33203125, "learning_rate": 0.00019337442227075285, "loss": 0.8405, "step": 637 }, { "epoch": 0.3549867853665322, "grad_norm": 0.33984375, "learning_rate": 0.0001933535086687295, "loss": 0.906, "step": 638 }, { "epoch": 0.35554319098622894, "grad_norm": 0.34375, "learning_rate": 0.000193332563246397, "loss": 0.9176, "step": 639 }, { "epoch": 0.3560995966059257, "grad_norm": 0.33203125, "learning_rate": 0.0001933115860108948, "loss": 0.8164, "step": 640 }, { "epoch": 0.3566560022256225, "grad_norm": 0.34765625, "learning_rate": 0.00019329057696937317, "loss": 0.8414, "step": 641 }, { "epoch": 0.3572124078453192, "grad_norm": 0.357421875, "learning_rate": 0.00019326953612899318, "loss": 0.9027, "step": 642 }, { "epoch": 0.357768813465016, "grad_norm": 0.3671875, "learning_rate": 0.0001932484634969268, "loss": 0.8986, "step": 643 }, { "epoch": 0.35832521908471276, "grad_norm": 0.3515625, "learning_rate": 0.00019322735908035678, "loss": 0.891, "step": 644 }, { "epoch": 0.35888162470440954, "grad_norm": 0.353515625, "learning_rate": 0.0001932062228864768, "loss": 0.9787, "step": 645 }, { "epoch": 0.35943803032410626, "grad_norm": 0.33984375, "learning_rate": 0.0001931850549224912, "loss": 0.8145, "step": 646 }, { "epoch": 0.35999443594380304, "grad_norm": 0.34375, "learning_rate": 0.00019316385519561538, "loss": 0.8913, "step": 647 }, { "epoch": 0.3605508415634998, "grad_norm": 0.328125, "learning_rate": 0.00019314262371307532, "loss": 0.7981, "step": 648 }, { "epoch": 0.36110724718319653, "grad_norm": 0.337890625, "learning_rate": 0.00019312136048210801, "loss": 0.8772, "step": 649 }, { "epoch": 0.3616636528028933, "grad_norm": 0.326171875, "learning_rate": 0.0001931000655099612, "loss": 0.8837, "step": 650 }, { "epoch": 0.3622200584225901, "grad_norm": 0.345703125, "learning_rate": 0.00019307873880389345, "loss": 0.923, "step": 651 }, { "epoch": 0.3627764640422868, "grad_norm": 0.337890625, "learning_rate": 0.00019305738037117412, "loss": 0.8609, "step": 652 }, { "epoch": 0.3633328696619836, "grad_norm": 0.34765625, "learning_rate": 0.00019303599021908342, "loss": 0.8668, "step": 653 }, { "epoch": 0.36388927528168036, "grad_norm": 0.3515625, "learning_rate": 0.00019301456835491237, "loss": 0.9394, "step": 654 }, { "epoch": 0.3644456809013771, "grad_norm": 0.328125, "learning_rate": 0.0001929931147859628, "loss": 0.7789, "step": 655 }, { "epoch": 0.36500208652107385, "grad_norm": 0.359375, "learning_rate": 0.00019297162951954728, "loss": 0.9009, "step": 656 }, { "epoch": 0.36555849214077063, "grad_norm": 0.333984375, "learning_rate": 0.00019295011256298929, "loss": 0.7917, "step": 657 }, { "epoch": 0.3661148977604674, "grad_norm": 0.353515625, "learning_rate": 0.00019292856392362303, "loss": 0.9266, "step": 658 }, { "epoch": 0.3666713033801641, "grad_norm": 0.3359375, "learning_rate": 0.00019290698360879355, "loss": 0.849, "step": 659 }, { "epoch": 0.3672277089998609, "grad_norm": 0.3359375, "learning_rate": 0.0001928853716258567, "loss": 0.8509, "step": 660 }, { "epoch": 0.3677841146195577, "grad_norm": 0.326171875, "learning_rate": 0.00019286372798217905, "loss": 0.8732, "step": 661 }, { "epoch": 0.3683405202392544, "grad_norm": 0.345703125, "learning_rate": 0.00019284205268513803, "loss": 0.8759, "step": 662 }, { "epoch": 0.3688969258589512, "grad_norm": 1.8046875, "learning_rate": 0.00019282034574212185, "loss": 1.8908, "step": 663 }, { "epoch": 0.36945333147864795, "grad_norm": 0.33203125, "learning_rate": 0.00019279860716052947, "loss": 0.8435, "step": 664 }, { "epoch": 0.37000973709834467, "grad_norm": 0.353515625, "learning_rate": 0.00019277683694777074, "loss": 0.8151, "step": 665 }, { "epoch": 0.37056614271804145, "grad_norm": 0.36328125, "learning_rate": 0.00019275503511126618, "loss": 0.9638, "step": 666 }, { "epoch": 0.3711225483377382, "grad_norm": 0.337890625, "learning_rate": 0.00019273320165844707, "loss": 0.9098, "step": 667 }, { "epoch": 0.37167895395743494, "grad_norm": 0.359375, "learning_rate": 0.00019271133659675563, "loss": 0.9032, "step": 668 }, { "epoch": 0.3722353595771317, "grad_norm": 0.33984375, "learning_rate": 0.00019268943993364462, "loss": 0.7979, "step": 669 }, { "epoch": 0.3727917651968285, "grad_norm": 0.353515625, "learning_rate": 0.00019266751167657778, "loss": 0.9499, "step": 670 }, { "epoch": 0.37334817081652527, "grad_norm": 0.3359375, "learning_rate": 0.0001926455518330295, "loss": 0.8741, "step": 671 }, { "epoch": 0.373904576436222, "grad_norm": 0.349609375, "learning_rate": 0.000192623560410485, "loss": 0.9162, "step": 672 }, { "epoch": 0.37446098205591877, "grad_norm": 0.33203125, "learning_rate": 0.00019260153741644021, "loss": 0.8135, "step": 673 }, { "epoch": 0.37501738767561554, "grad_norm": 0.3125, "learning_rate": 0.00019257948285840188, "loss": 0.6664, "step": 674 }, { "epoch": 0.37557379329531226, "grad_norm": 0.359375, "learning_rate": 0.00019255739674388746, "loss": 0.9409, "step": 675 }, { "epoch": 0.37613019891500904, "grad_norm": 0.34765625, "learning_rate": 0.00019253527908042522, "loss": 0.9325, "step": 676 }, { "epoch": 0.3766866045347058, "grad_norm": 0.3359375, "learning_rate": 0.00019251312987555408, "loss": 0.8681, "step": 677 }, { "epoch": 0.37724301015440254, "grad_norm": 0.3515625, "learning_rate": 0.00019249094913682382, "loss": 0.9045, "step": 678 }, { "epoch": 0.3777994157740993, "grad_norm": 0.345703125, "learning_rate": 0.00019246873687179495, "loss": 0.8807, "step": 679 }, { "epoch": 0.3783558213937961, "grad_norm": 0.35546875, "learning_rate": 0.00019244649308803867, "loss": 0.9102, "step": 680 }, { "epoch": 0.37891222701349286, "grad_norm": 0.333984375, "learning_rate": 0.00019242421779313692, "loss": 0.8124, "step": 681 }, { "epoch": 0.3794686326331896, "grad_norm": 0.37109375, "learning_rate": 0.00019240191099468253, "loss": 0.9337, "step": 682 }, { "epoch": 0.38002503825288636, "grad_norm": 0.90625, "learning_rate": 0.00019237957270027884, "loss": 1.8244, "step": 683 }, { "epoch": 0.38058144387258314, "grad_norm": 0.349609375, "learning_rate": 0.0001923572029175401, "loss": 0.8975, "step": 684 }, { "epoch": 0.38113784949227986, "grad_norm": 0.34375, "learning_rate": 0.0001923348016540912, "loss": 0.8755, "step": 685 }, { "epoch": 0.38169425511197663, "grad_norm": 0.328125, "learning_rate": 0.00019231236891756787, "loss": 0.8125, "step": 686 }, { "epoch": 0.3822506607316734, "grad_norm": 0.337890625, "learning_rate": 0.00019228990471561636, "loss": 0.8673, "step": 687 }, { "epoch": 0.38280706635137013, "grad_norm": 0.33203125, "learning_rate": 0.0001922674090558939, "loss": 0.8462, "step": 688 }, { "epoch": 0.3833634719710669, "grad_norm": 0.3671875, "learning_rate": 0.00019224488194606826, "loss": 1.0229, "step": 689 }, { "epoch": 0.3839198775907637, "grad_norm": 0.333984375, "learning_rate": 0.00019222232339381802, "loss": 0.8654, "step": 690 }, { "epoch": 0.3844762832104604, "grad_norm": 0.361328125, "learning_rate": 0.00019219973340683244, "loss": 0.877, "step": 691 }, { "epoch": 0.3850326888301572, "grad_norm": 0.357421875, "learning_rate": 0.00019217711199281148, "loss": 0.9084, "step": 692 }, { "epoch": 0.38558909444985395, "grad_norm": 0.34765625, "learning_rate": 0.00019215445915946584, "loss": 0.9667, "step": 693 }, { "epoch": 0.38614550006955073, "grad_norm": 0.32421875, "learning_rate": 0.00019213177491451692, "loss": 0.8251, "step": 694 }, { "epoch": 0.38670190568924745, "grad_norm": 0.341796875, "learning_rate": 0.00019210905926569688, "loss": 0.8764, "step": 695 }, { "epoch": 0.3872583113089442, "grad_norm": 0.36328125, "learning_rate": 0.0001920863122207485, "loss": 0.8484, "step": 696 }, { "epoch": 0.387814716928641, "grad_norm": 0.373046875, "learning_rate": 0.00019206353378742528, "loss": 1.0047, "step": 697 }, { "epoch": 0.3883711225483377, "grad_norm": 0.337890625, "learning_rate": 0.00019204072397349144, "loss": 0.8996, "step": 698 }, { "epoch": 0.3889275281680345, "grad_norm": 0.34375, "learning_rate": 0.00019201788278672193, "loss": 0.9355, "step": 699 }, { "epoch": 0.3894839337877313, "grad_norm": 0.345703125, "learning_rate": 0.00019199501023490232, "loss": 0.8971, "step": 700 }, { "epoch": 0.390040339407428, "grad_norm": 0.357421875, "learning_rate": 0.0001919721063258289, "loss": 0.8437, "step": 701 }, { "epoch": 0.39059674502712477, "grad_norm": 0.349609375, "learning_rate": 0.0001919491710673087, "loss": 0.9162, "step": 702 }, { "epoch": 0.39115315064682155, "grad_norm": 0.349609375, "learning_rate": 0.00019192620446715933, "loss": 0.9021, "step": 703 }, { "epoch": 0.39170955626651827, "grad_norm": 0.326171875, "learning_rate": 0.00019190320653320918, "loss": 0.8099, "step": 704 }, { "epoch": 0.39226596188621504, "grad_norm": 0.359375, "learning_rate": 0.0001918801772732973, "loss": 0.9569, "step": 705 }, { "epoch": 0.3928223675059118, "grad_norm": 0.349609375, "learning_rate": 0.0001918571166952734, "loss": 0.9673, "step": 706 }, { "epoch": 0.3933787731256086, "grad_norm": 0.337890625, "learning_rate": 0.0001918340248069978, "loss": 0.8443, "step": 707 }, { "epoch": 0.3939351787453053, "grad_norm": 0.359375, "learning_rate": 0.00019181090161634167, "loss": 0.9339, "step": 708 }, { "epoch": 0.3944915843650021, "grad_norm": 0.3359375, "learning_rate": 0.00019178774713118663, "loss": 0.834, "step": 709 }, { "epoch": 0.39504798998469887, "grad_norm": 0.357421875, "learning_rate": 0.00019176456135942515, "loss": 0.9933, "step": 710 }, { "epoch": 0.3956043956043956, "grad_norm": 0.34765625, "learning_rate": 0.00019174134430896027, "loss": 0.924, "step": 711 }, { "epoch": 0.39616080122409236, "grad_norm": 0.341796875, "learning_rate": 0.0001917180959877057, "loss": 0.8976, "step": 712 }, { "epoch": 0.39671720684378914, "grad_norm": 0.33984375, "learning_rate": 0.00019169481640358583, "loss": 0.8488, "step": 713 }, { "epoch": 0.39727361246348586, "grad_norm": 0.337890625, "learning_rate": 0.00019167150556453569, "loss": 0.8426, "step": 714 }, { "epoch": 0.39783001808318263, "grad_norm": 0.357421875, "learning_rate": 0.00019164816347850096, "loss": 1.0289, "step": 715 }, { "epoch": 0.3983864237028794, "grad_norm": 0.341796875, "learning_rate": 0.00019162479015343803, "loss": 0.8953, "step": 716 }, { "epoch": 0.39894282932257613, "grad_norm": 0.34765625, "learning_rate": 0.00019160138559731386, "loss": 0.9065, "step": 717 }, { "epoch": 0.3994992349422729, "grad_norm": 0.349609375, "learning_rate": 0.00019157794981810607, "loss": 0.908, "step": 718 }, { "epoch": 0.4000556405619697, "grad_norm": 0.337890625, "learning_rate": 0.00019155448282380294, "loss": 0.796, "step": 719 }, { "epoch": 0.40061204618166646, "grad_norm": 0.37109375, "learning_rate": 0.00019153098462240342, "loss": 0.9603, "step": 720 }, { "epoch": 0.4011684518013632, "grad_norm": 0.341796875, "learning_rate": 0.00019150745522191701, "loss": 0.9096, "step": 721 }, { "epoch": 0.40172485742105996, "grad_norm": 0.357421875, "learning_rate": 0.00019148389463036395, "loss": 1.0184, "step": 722 }, { "epoch": 0.40228126304075673, "grad_norm": 0.33203125, "learning_rate": 0.00019146030285577506, "loss": 0.8274, "step": 723 }, { "epoch": 0.40283766866045345, "grad_norm": 0.353515625, "learning_rate": 0.00019143667990619177, "loss": 0.9454, "step": 724 }, { "epoch": 0.4033940742801502, "grad_norm": 0.330078125, "learning_rate": 0.00019141302578966614, "loss": 0.8516, "step": 725 }, { "epoch": 0.403950479899847, "grad_norm": 0.349609375, "learning_rate": 0.00019138934051426092, "loss": 0.9234, "step": 726 }, { "epoch": 0.4045068855195437, "grad_norm": 0.337890625, "learning_rate": 0.00019136562408804934, "loss": 0.8836, "step": 727 }, { "epoch": 0.4050632911392405, "grad_norm": 0.349609375, "learning_rate": 0.00019134187651911546, "loss": 0.9283, "step": 728 }, { "epoch": 0.4056196967589373, "grad_norm": 0.353515625, "learning_rate": 0.00019131809781555373, "loss": 0.9731, "step": 729 }, { "epoch": 0.40617610237863405, "grad_norm": 0.326171875, "learning_rate": 0.00019129428798546938, "loss": 0.7996, "step": 730 }, { "epoch": 0.4067325079983308, "grad_norm": 0.33984375, "learning_rate": 0.00019127044703697816, "loss": 0.8759, "step": 731 }, { "epoch": 0.40728891361802755, "grad_norm": 0.345703125, "learning_rate": 0.00019124657497820644, "loss": 0.9137, "step": 732 }, { "epoch": 0.4078453192377243, "grad_norm": 0.3515625, "learning_rate": 0.00019122267181729123, "loss": 0.8381, "step": 733 }, { "epoch": 0.40840172485742104, "grad_norm": 0.365234375, "learning_rate": 0.0001911987375623801, "loss": 0.9209, "step": 734 }, { "epoch": 0.4089581304771178, "grad_norm": 0.345703125, "learning_rate": 0.00019117477222163127, "loss": 0.8937, "step": 735 }, { "epoch": 0.4095145360968146, "grad_norm": 0.35546875, "learning_rate": 0.0001911507758032135, "loss": 0.9494, "step": 736 }, { "epoch": 0.4100709417165113, "grad_norm": 0.353515625, "learning_rate": 0.00019112674831530616, "loss": 0.9375, "step": 737 }, { "epoch": 0.4106273473362081, "grad_norm": 0.369140625, "learning_rate": 0.00019110268976609923, "loss": 0.8973, "step": 738 }, { "epoch": 0.41118375295590487, "grad_norm": 0.36328125, "learning_rate": 0.00019107860016379324, "loss": 0.9579, "step": 739 }, { "epoch": 0.4117401585756016, "grad_norm": 0.328125, "learning_rate": 0.00019105447951659935, "loss": 0.7406, "step": 740 }, { "epoch": 0.41229656419529837, "grad_norm": 0.365234375, "learning_rate": 0.0001910303278327393, "loss": 0.9315, "step": 741 }, { "epoch": 0.41285296981499514, "grad_norm": 0.35546875, "learning_rate": 0.00019100614512044536, "loss": 0.8731, "step": 742 }, { "epoch": 0.4134093754346919, "grad_norm": 0.349609375, "learning_rate": 0.0001909819313879604, "loss": 0.8679, "step": 743 }, { "epoch": 0.41396578105438864, "grad_norm": 0.345703125, "learning_rate": 0.00019095768664353785, "loss": 0.8881, "step": 744 }, { "epoch": 0.4145221866740854, "grad_norm": 0.3359375, "learning_rate": 0.0001909334108954418, "loss": 0.8616, "step": 745 }, { "epoch": 0.4150785922937822, "grad_norm": 0.3359375, "learning_rate": 0.00019090910415194682, "loss": 0.8466, "step": 746 }, { "epoch": 0.4156349979134789, "grad_norm": 0.359375, "learning_rate": 0.000190884766421338, "loss": 0.963, "step": 747 }, { "epoch": 0.4161914035331757, "grad_norm": 0.3359375, "learning_rate": 0.00019086039771191112, "loss": 0.8879, "step": 748 }, { "epoch": 0.41674780915287246, "grad_norm": 0.357421875, "learning_rate": 0.0001908359980319724, "loss": 0.9345, "step": 749 }, { "epoch": 0.4173042147725692, "grad_norm": 0.3671875, "learning_rate": 0.00019081156738983874, "loss": 0.9356, "step": 750 }, { "epoch": 0.41786062039226596, "grad_norm": 0.33203125, "learning_rate": 0.00019078710579383746, "loss": 0.8295, "step": 751 }, { "epoch": 0.41841702601196273, "grad_norm": 0.345703125, "learning_rate": 0.00019076261325230655, "loss": 0.9092, "step": 752 }, { "epoch": 0.41897343163165945, "grad_norm": 0.32421875, "learning_rate": 0.00019073808977359446, "loss": 0.83, "step": 753 }, { "epoch": 0.41952983725135623, "grad_norm": 0.318359375, "learning_rate": 0.0001907135353660602, "loss": 0.7945, "step": 754 }, { "epoch": 0.420086242871053, "grad_norm": 0.353515625, "learning_rate": 0.00019068895003807339, "loss": 0.959, "step": 755 }, { "epoch": 0.4206426484907498, "grad_norm": 0.361328125, "learning_rate": 0.0001906643337980141, "loss": 0.8404, "step": 756 }, { "epoch": 0.4211990541104465, "grad_norm": 0.32421875, "learning_rate": 0.000190639686654273, "loss": 0.811, "step": 757 }, { "epoch": 0.4217554597301433, "grad_norm": 0.34375, "learning_rate": 0.00019061500861525127, "loss": 0.9065, "step": 758 }, { "epoch": 0.42231186534984005, "grad_norm": 0.357421875, "learning_rate": 0.00019059029968936062, "loss": 0.8662, "step": 759 }, { "epoch": 0.4228682709695368, "grad_norm": 0.33203125, "learning_rate": 0.0001905655598850233, "loss": 0.7609, "step": 760 }, { "epoch": 0.42342467658923355, "grad_norm": 0.33984375, "learning_rate": 0.00019054078921067205, "loss": 0.8826, "step": 761 }, { "epoch": 0.4239810822089303, "grad_norm": 0.328125, "learning_rate": 0.00019051598767475016, "loss": 0.8253, "step": 762 }, { "epoch": 0.42453748782862705, "grad_norm": 0.3203125, "learning_rate": 0.00019049115528571147, "loss": 0.7537, "step": 763 }, { "epoch": 0.4250938934483238, "grad_norm": 0.3515625, "learning_rate": 0.00019046629205202027, "loss": 0.9363, "step": 764 }, { "epoch": 0.4256502990680206, "grad_norm": 0.3359375, "learning_rate": 0.00019044139798215143, "loss": 0.8248, "step": 765 }, { "epoch": 0.4262067046877173, "grad_norm": 0.322265625, "learning_rate": 0.00019041647308459022, "loss": 0.7364, "step": 766 }, { "epoch": 0.4267631103074141, "grad_norm": 0.345703125, "learning_rate": 0.00019039151736783262, "loss": 0.9075, "step": 767 }, { "epoch": 0.42731951592711087, "grad_norm": 0.341796875, "learning_rate": 0.00019036653084038486, "loss": 0.966, "step": 768 }, { "epoch": 0.42787592154680765, "grad_norm": 0.3515625, "learning_rate": 0.00019034151351076387, "loss": 0.8645, "step": 769 }, { "epoch": 0.42843232716650437, "grad_norm": 0.3359375, "learning_rate": 0.00019031646538749707, "loss": 0.824, "step": 770 }, { "epoch": 0.42898873278620114, "grad_norm": 0.33984375, "learning_rate": 0.00019029138647912218, "loss": 0.8858, "step": 771 }, { "epoch": 0.4295451384058979, "grad_norm": 0.328125, "learning_rate": 0.00019026627679418767, "loss": 0.8066, "step": 772 }, { "epoch": 0.43010154402559464, "grad_norm": 0.3203125, "learning_rate": 0.00019024113634125234, "loss": 0.8256, "step": 773 }, { "epoch": 0.4306579496452914, "grad_norm": 0.33984375, "learning_rate": 0.0001902159651288855, "loss": 0.8785, "step": 774 }, { "epoch": 0.4312143552649882, "grad_norm": 0.337890625, "learning_rate": 0.000190190763165667, "loss": 0.8643, "step": 775 }, { "epoch": 0.4317707608846849, "grad_norm": 0.34375, "learning_rate": 0.00019016553046018712, "loss": 0.9032, "step": 776 }, { "epoch": 0.4323271665043817, "grad_norm": 0.34375, "learning_rate": 0.00019014026702104662, "loss": 0.8785, "step": 777 }, { "epoch": 0.43288357212407846, "grad_norm": 0.33984375, "learning_rate": 0.00019011497285685678, "loss": 0.8607, "step": 778 }, { "epoch": 0.4334399777437752, "grad_norm": 0.3671875, "learning_rate": 0.0001900896479762393, "loss": 0.8355, "step": 779 }, { "epoch": 0.43399638336347196, "grad_norm": 0.376953125, "learning_rate": 0.00019006429238782637, "loss": 0.9727, "step": 780 }, { "epoch": 0.43455278898316874, "grad_norm": 0.34375, "learning_rate": 0.00019003890610026067, "loss": 0.8807, "step": 781 }, { "epoch": 0.4351091946028655, "grad_norm": 0.33984375, "learning_rate": 0.00019001348912219533, "loss": 0.9325, "step": 782 }, { "epoch": 0.43566560022256223, "grad_norm": 0.3359375, "learning_rate": 0.00018998804146229387, "loss": 0.8559, "step": 783 }, { "epoch": 0.436222005842259, "grad_norm": 0.337890625, "learning_rate": 0.00018996256312923043, "loss": 0.8304, "step": 784 }, { "epoch": 0.4367784114619558, "grad_norm": 0.353515625, "learning_rate": 0.00018993705413168944, "loss": 0.9767, "step": 785 }, { "epoch": 0.4373348170816525, "grad_norm": 0.33984375, "learning_rate": 0.0001899115144783659, "loss": 0.8711, "step": 786 }, { "epoch": 0.4378912227013493, "grad_norm": 0.345703125, "learning_rate": 0.00018988594417796516, "loss": 0.8525, "step": 787 }, { "epoch": 0.43844762832104606, "grad_norm": 0.341796875, "learning_rate": 0.00018986034323920312, "loss": 0.83, "step": 788 }, { "epoch": 0.4390040339407428, "grad_norm": 0.333984375, "learning_rate": 0.000189834711670806, "loss": 0.8088, "step": 789 }, { "epoch": 0.43956043956043955, "grad_norm": 0.318359375, "learning_rate": 0.00018980904948151062, "loss": 0.784, "step": 790 }, { "epoch": 0.44011684518013633, "grad_norm": 0.33984375, "learning_rate": 0.00018978335668006407, "loss": 0.9129, "step": 791 }, { "epoch": 0.4406732507998331, "grad_norm": 0.36328125, "learning_rate": 0.000189757633275224, "loss": 1.0234, "step": 792 }, { "epoch": 0.4412296564195298, "grad_norm": 0.349609375, "learning_rate": 0.0001897318792757584, "loss": 0.8711, "step": 793 }, { "epoch": 0.4417860620392266, "grad_norm": 0.34375, "learning_rate": 0.00018970609469044577, "loss": 0.9124, "step": 794 }, { "epoch": 0.4423424676589234, "grad_norm": 0.310546875, "learning_rate": 0.00018968027952807494, "loss": 0.7081, "step": 795 }, { "epoch": 0.4428988732786201, "grad_norm": 0.31640625, "learning_rate": 0.00018965443379744535, "loss": 0.7776, "step": 796 }, { "epoch": 0.4434552788983169, "grad_norm": 0.32421875, "learning_rate": 0.00018962855750736656, "loss": 0.8216, "step": 797 }, { "epoch": 0.44401168451801365, "grad_norm": 0.345703125, "learning_rate": 0.00018960265066665882, "loss": 0.8406, "step": 798 }, { "epoch": 0.44456809013771037, "grad_norm": 0.33984375, "learning_rate": 0.0001895767132841527, "loss": 0.9074, "step": 799 }, { "epoch": 0.44512449575740715, "grad_norm": 0.37109375, "learning_rate": 0.00018955074536868913, "loss": 0.88, "step": 800 }, { "epoch": 0.4456809013771039, "grad_norm": 0.361328125, "learning_rate": 0.0001895247469291195, "loss": 0.9454, "step": 801 }, { "epoch": 0.44623730699680064, "grad_norm": 0.330078125, "learning_rate": 0.00018949871797430557, "loss": 0.7475, "step": 802 }, { "epoch": 0.4467937126164974, "grad_norm": 0.353515625, "learning_rate": 0.00018947265851311954, "loss": 0.9466, "step": 803 }, { "epoch": 0.4473501182361942, "grad_norm": 0.328125, "learning_rate": 0.000189446568554444, "loss": 0.791, "step": 804 }, { "epoch": 0.44790652385589097, "grad_norm": 0.345703125, "learning_rate": 0.00018942044810717188, "loss": 0.898, "step": 805 }, { "epoch": 0.4484629294755877, "grad_norm": 0.318359375, "learning_rate": 0.00018939429718020661, "loss": 0.7666, "step": 806 }, { "epoch": 0.44901933509528447, "grad_norm": 0.34375, "learning_rate": 0.0001893681157824619, "loss": 0.9343, "step": 807 }, { "epoch": 0.44957574071498124, "grad_norm": 0.337890625, "learning_rate": 0.00018934190392286198, "loss": 0.8074, "step": 808 }, { "epoch": 0.45013214633467796, "grad_norm": 0.333984375, "learning_rate": 0.00018931566161034126, "loss": 0.8394, "step": 809 }, { "epoch": 0.45068855195437474, "grad_norm": 0.3359375, "learning_rate": 0.00018928938885384472, "loss": 0.8655, "step": 810 }, { "epoch": 0.4512449575740715, "grad_norm": 0.322265625, "learning_rate": 0.00018926308566232763, "loss": 0.792, "step": 811 }, { "epoch": 0.45180136319376824, "grad_norm": 0.34765625, "learning_rate": 0.00018923675204475564, "loss": 0.8638, "step": 812 }, { "epoch": 0.452357768813465, "grad_norm": 0.33984375, "learning_rate": 0.00018921038801010477, "loss": 0.8151, "step": 813 }, { "epoch": 0.4529141744331618, "grad_norm": 0.3359375, "learning_rate": 0.00018918399356736146, "loss": 0.8198, "step": 814 }, { "epoch": 0.4534705800528585, "grad_norm": 0.30078125, "learning_rate": 0.00018915756872552242, "loss": 0.6904, "step": 815 }, { "epoch": 0.4540269856725553, "grad_norm": 0.333984375, "learning_rate": 0.0001891311134935948, "loss": 0.8762, "step": 816 }, { "epoch": 0.45458339129225206, "grad_norm": 0.3125, "learning_rate": 0.0001891046278805961, "loss": 0.7527, "step": 817 }, { "epoch": 0.45513979691194884, "grad_norm": 0.33984375, "learning_rate": 0.00018907811189555412, "loss": 0.7777, "step": 818 }, { "epoch": 0.45569620253164556, "grad_norm": 0.3359375, "learning_rate": 0.00018905156554750708, "loss": 0.8556, "step": 819 }, { "epoch": 0.45625260815134233, "grad_norm": 0.328125, "learning_rate": 0.0001890249888455035, "loss": 0.8273, "step": 820 }, { "epoch": 0.4568090137710391, "grad_norm": 0.330078125, "learning_rate": 0.0001889983817986023, "loss": 0.8405, "step": 821 }, { "epoch": 0.45736541939073583, "grad_norm": 0.341796875, "learning_rate": 0.00018897174441587266, "loss": 0.9006, "step": 822 }, { "epoch": 0.4579218250104326, "grad_norm": 0.349609375, "learning_rate": 0.00018894507670639418, "loss": 0.9191, "step": 823 }, { "epoch": 0.4584782306301294, "grad_norm": 0.357421875, "learning_rate": 0.00018891837867925678, "loss": 0.9777, "step": 824 }, { "epoch": 0.4590346362498261, "grad_norm": 0.34375, "learning_rate": 0.00018889165034356072, "loss": 0.892, "step": 825 }, { "epoch": 0.4595910418695229, "grad_norm": 0.353515625, "learning_rate": 0.00018886489170841648, "loss": 0.8615, "step": 826 }, { "epoch": 0.46014744748921965, "grad_norm": 0.341796875, "learning_rate": 0.00018883810278294503, "loss": 0.8765, "step": 827 }, { "epoch": 0.4607038531089164, "grad_norm": 0.34375, "learning_rate": 0.00018881128357627763, "loss": 0.8803, "step": 828 }, { "epoch": 0.46126025872861315, "grad_norm": 0.349609375, "learning_rate": 0.00018878443409755575, "loss": 0.9017, "step": 829 }, { "epoch": 0.4618166643483099, "grad_norm": 0.330078125, "learning_rate": 0.00018875755435593129, "loss": 0.8139, "step": 830 }, { "epoch": 0.4623730699680067, "grad_norm": 0.34375, "learning_rate": 0.0001887306443605664, "loss": 0.9611, "step": 831 }, { "epoch": 0.4629294755877034, "grad_norm": 0.34765625, "learning_rate": 0.00018870370412063365, "loss": 0.9311, "step": 832 }, { "epoch": 0.4634858812074002, "grad_norm": 0.357421875, "learning_rate": 0.00018867673364531578, "loss": 0.9108, "step": 833 }, { "epoch": 0.464042286827097, "grad_norm": 0.36328125, "learning_rate": 0.00018864973294380587, "loss": 1.0142, "step": 834 }, { "epoch": 0.4645986924467937, "grad_norm": 0.328125, "learning_rate": 0.00018862270202530738, "loss": 0.744, "step": 835 }, { "epoch": 0.46515509806649047, "grad_norm": 0.341796875, "learning_rate": 0.00018859564089903404, "loss": 0.8419, "step": 836 }, { "epoch": 0.46571150368618724, "grad_norm": 0.322265625, "learning_rate": 0.00018856854957420983, "loss": 0.7806, "step": 837 }, { "epoch": 0.46626790930588397, "grad_norm": 0.3515625, "learning_rate": 0.00018854142806006902, "loss": 0.9135, "step": 838 }, { "epoch": 0.46682431492558074, "grad_norm": 0.330078125, "learning_rate": 0.00018851427636585623, "loss": 0.836, "step": 839 }, { "epoch": 0.4673807205452775, "grad_norm": 0.337890625, "learning_rate": 0.00018848709450082637, "loss": 0.8926, "step": 840 }, { "epoch": 0.4679371261649743, "grad_norm": 0.35546875, "learning_rate": 0.0001884598824742445, "loss": 0.8962, "step": 841 }, { "epoch": 0.468493531784671, "grad_norm": 0.330078125, "learning_rate": 0.0001884326402953862, "loss": 0.8185, "step": 842 }, { "epoch": 0.4690499374043678, "grad_norm": 0.35546875, "learning_rate": 0.0001884053679735371, "loss": 0.948, "step": 843 }, { "epoch": 0.46960634302406457, "grad_norm": 0.365234375, "learning_rate": 0.0001883780655179932, "loss": 1.0652, "step": 844 }, { "epoch": 0.4701627486437613, "grad_norm": 0.35546875, "learning_rate": 0.00018835073293806075, "loss": 0.9291, "step": 845 }, { "epoch": 0.47071915426345806, "grad_norm": 0.34765625, "learning_rate": 0.00018832337024305638, "loss": 0.9503, "step": 846 }, { "epoch": 0.47127555988315484, "grad_norm": 0.337890625, "learning_rate": 0.00018829597744230678, "loss": 0.8685, "step": 847 }, { "epoch": 0.47183196550285156, "grad_norm": 0.3359375, "learning_rate": 0.00018826855454514907, "loss": 0.8796, "step": 848 }, { "epoch": 0.47238837112254833, "grad_norm": 0.32421875, "learning_rate": 0.00018824110156093053, "loss": 0.7906, "step": 849 }, { "epoch": 0.4729447767422451, "grad_norm": 0.34375, "learning_rate": 0.00018821361849900877, "loss": 0.9085, "step": 850 }, { "epoch": 0.47350118236194183, "grad_norm": 0.330078125, "learning_rate": 0.00018818610536875164, "loss": 0.9018, "step": 851 }, { "epoch": 0.4740575879816386, "grad_norm": 0.33203125, "learning_rate": 0.00018815856217953714, "loss": 0.8632, "step": 852 }, { "epoch": 0.4746139936013354, "grad_norm": 0.33984375, "learning_rate": 0.00018813098894075365, "loss": 0.9047, "step": 853 }, { "epoch": 0.47517039922103216, "grad_norm": 0.34375, "learning_rate": 0.00018810338566179973, "loss": 0.9046, "step": 854 }, { "epoch": 0.4757268048407289, "grad_norm": 0.328125, "learning_rate": 0.00018807575235208415, "loss": 0.8683, "step": 855 }, { "epoch": 0.47628321046042565, "grad_norm": 0.330078125, "learning_rate": 0.00018804808902102597, "loss": 0.8541, "step": 856 }, { "epoch": 0.47683961608012243, "grad_norm": 0.33203125, "learning_rate": 0.00018802039567805448, "loss": 0.8484, "step": 857 }, { "epoch": 0.47739602169981915, "grad_norm": 0.345703125, "learning_rate": 0.00018799267233260918, "loss": 0.8559, "step": 858 }, { "epoch": 0.4779524273195159, "grad_norm": 0.3359375, "learning_rate": 0.00018796491899413976, "loss": 0.8646, "step": 859 }, { "epoch": 0.4785088329392127, "grad_norm": 0.349609375, "learning_rate": 0.0001879371356721062, "loss": 0.9625, "step": 860 }, { "epoch": 0.4790652385589094, "grad_norm": 0.3203125, "learning_rate": 0.0001879093223759787, "loss": 0.7927, "step": 861 }, { "epoch": 0.4796216441786062, "grad_norm": 0.32421875, "learning_rate": 0.00018788147911523762, "loss": 0.8716, "step": 862 }, { "epoch": 0.480178049798303, "grad_norm": 0.3515625, "learning_rate": 0.00018785360589937354, "loss": 0.7913, "step": 863 }, { "epoch": 0.4807344554179997, "grad_norm": 0.34375, "learning_rate": 0.00018782570273788736, "loss": 0.8636, "step": 864 }, { "epoch": 0.48129086103769647, "grad_norm": 0.330078125, "learning_rate": 0.00018779776964029, "loss": 0.8082, "step": 865 }, { "epoch": 0.48184726665739325, "grad_norm": 0.33984375, "learning_rate": 0.00018776980661610273, "loss": 0.8549, "step": 866 }, { "epoch": 0.48240367227709, "grad_norm": 0.33203125, "learning_rate": 0.000187741813674857, "loss": 0.8543, "step": 867 }, { "epoch": 0.48296007789678674, "grad_norm": 0.337890625, "learning_rate": 0.00018771379082609436, "loss": 0.8735, "step": 868 }, { "epoch": 0.4835164835164835, "grad_norm": 0.33984375, "learning_rate": 0.0001876857380793667, "loss": 0.8338, "step": 869 }, { "epoch": 0.4840728891361803, "grad_norm": 0.333984375, "learning_rate": 0.00018765765544423606, "loss": 0.8904, "step": 870 }, { "epoch": 0.484629294755877, "grad_norm": 0.326171875, "learning_rate": 0.00018762954293027453, "loss": 0.7981, "step": 871 }, { "epoch": 0.4851857003755738, "grad_norm": 0.341796875, "learning_rate": 0.00018760140054706453, "loss": 0.9093, "step": 872 }, { "epoch": 0.48574210599527057, "grad_norm": 0.341796875, "learning_rate": 0.00018757322830419867, "loss": 0.9216, "step": 873 }, { "epoch": 0.4862985116149673, "grad_norm": 0.31640625, "learning_rate": 0.00018754502621127966, "loss": 0.7949, "step": 874 }, { "epoch": 0.48685491723466406, "grad_norm": 0.345703125, "learning_rate": 0.00018751679427792038, "loss": 0.9682, "step": 875 }, { "epoch": 0.48741132285436084, "grad_norm": 0.328125, "learning_rate": 0.00018748853251374396, "loss": 0.792, "step": 876 }, { "epoch": 0.48796772847405756, "grad_norm": 0.349609375, "learning_rate": 0.00018746024092838366, "loss": 0.8901, "step": 877 }, { "epoch": 0.48852413409375434, "grad_norm": 0.349609375, "learning_rate": 0.00018743191953148286, "loss": 0.9955, "step": 878 }, { "epoch": 0.4890805397134511, "grad_norm": 0.34375, "learning_rate": 0.00018740356833269516, "loss": 0.9246, "step": 879 }, { "epoch": 0.4896369453331479, "grad_norm": 0.34375, "learning_rate": 0.00018737518734168428, "loss": 0.9122, "step": 880 }, { "epoch": 0.4901933509528446, "grad_norm": 0.3359375, "learning_rate": 0.0001873467765681242, "loss": 0.9119, "step": 881 }, { "epoch": 0.4907497565725414, "grad_norm": 0.322265625, "learning_rate": 0.00018731833602169884, "loss": 0.7738, "step": 882 }, { "epoch": 0.49130616219223816, "grad_norm": 0.33203125, "learning_rate": 0.00018728986571210244, "loss": 0.8866, "step": 883 }, { "epoch": 0.4918625678119349, "grad_norm": 0.333984375, "learning_rate": 0.00018726136564903938, "loss": 0.9322, "step": 884 }, { "epoch": 0.49241897343163166, "grad_norm": 0.318359375, "learning_rate": 0.0001872328358422241, "loss": 0.7841, "step": 885 }, { "epoch": 0.49297537905132843, "grad_norm": 0.349609375, "learning_rate": 0.00018720427630138122, "loss": 0.9459, "step": 886 }, { "epoch": 0.49353178467102515, "grad_norm": 0.330078125, "learning_rate": 0.0001871756870362455, "loss": 0.8372, "step": 887 }, { "epoch": 0.49408819029072193, "grad_norm": 0.3359375, "learning_rate": 0.00018714706805656185, "loss": 0.8801, "step": 888 }, { "epoch": 0.4946445959104187, "grad_norm": 0.31640625, "learning_rate": 0.00018711841937208527, "loss": 0.7583, "step": 889 }, { "epoch": 0.4952010015301155, "grad_norm": 0.333984375, "learning_rate": 0.00018708974099258086, "loss": 0.8007, "step": 890 }, { "epoch": 0.4957574071498122, "grad_norm": 0.349609375, "learning_rate": 0.0001870610329278239, "loss": 0.9087, "step": 891 }, { "epoch": 0.496313812769509, "grad_norm": 0.333984375, "learning_rate": 0.00018703229518759985, "loss": 0.8809, "step": 892 }, { "epoch": 0.49687021838920575, "grad_norm": 0.33203125, "learning_rate": 0.00018700352778170407, "loss": 0.8517, "step": 893 }, { "epoch": 0.4974266240089025, "grad_norm": 0.345703125, "learning_rate": 0.00018697473071994225, "loss": 0.9004, "step": 894 }, { "epoch": 0.49798302962859925, "grad_norm": 0.33203125, "learning_rate": 0.0001869459040121301, "loss": 0.8786, "step": 895 }, { "epoch": 0.498539435248296, "grad_norm": 0.34765625, "learning_rate": 0.0001869170476680934, "loss": 0.911, "step": 896 }, { "epoch": 0.49909584086799275, "grad_norm": 0.34765625, "learning_rate": 0.0001868881616976681, "loss": 0.8994, "step": 897 }, { "epoch": 0.4996522464876895, "grad_norm": 0.328125, "learning_rate": 0.00018685924611070025, "loss": 0.8162, "step": 898 }, { "epoch": 0.5002086521073863, "grad_norm": 0.33984375, "learning_rate": 0.00018683030091704592, "loss": 0.891, "step": 899 }, { "epoch": 0.5002086521073863, "eval_loss": 0.8054929971694946, "eval_runtime": 1766.0961, "eval_samples_per_second": 2.656, "eval_steps_per_second": 1.328, "step": 899 }, { "epoch": 0.500765057727083, "grad_norm": 0.326171875, "learning_rate": 0.0001868013261265713, "loss": 0.8493, "step": 900 }, { "epoch": 0.5013214633467798, "grad_norm": 0.3515625, "learning_rate": 0.00018677232174915275, "loss": 0.8253, "step": 901 }, { "epoch": 0.5018778689664766, "grad_norm": 0.341796875, "learning_rate": 0.0001867432877946766, "loss": 0.8701, "step": 902 }, { "epoch": 0.5024342745861733, "grad_norm": 0.359375, "learning_rate": 0.0001867142242730393, "loss": 0.9443, "step": 903 }, { "epoch": 0.5029906802058701, "grad_norm": 0.3515625, "learning_rate": 0.00018668513119414744, "loss": 0.9508, "step": 904 }, { "epoch": 0.5035470858255668, "grad_norm": 0.330078125, "learning_rate": 0.00018665600856791764, "loss": 0.7881, "step": 905 }, { "epoch": 0.5041034914452636, "grad_norm": 0.345703125, "learning_rate": 0.00018662685640427653, "loss": 0.8958, "step": 906 }, { "epoch": 0.5046598970649604, "grad_norm": 0.345703125, "learning_rate": 0.00018659767471316087, "loss": 0.8597, "step": 907 }, { "epoch": 0.5052163026846571, "grad_norm": 0.33984375, "learning_rate": 0.0001865684635045175, "loss": 0.8495, "step": 908 }, { "epoch": 0.5057727083043538, "grad_norm": 0.33984375, "learning_rate": 0.00018653922278830335, "loss": 0.8809, "step": 909 }, { "epoch": 0.5063291139240507, "grad_norm": 0.330078125, "learning_rate": 0.00018650995257448526, "loss": 0.8809, "step": 910 }, { "epoch": 0.5068855195437474, "grad_norm": 0.337890625, "learning_rate": 0.0001864806528730403, "loss": 0.894, "step": 911 }, { "epoch": 0.5074419251634441, "grad_norm": 0.333984375, "learning_rate": 0.00018645132369395545, "loss": 0.8067, "step": 912 }, { "epoch": 0.5079983307831409, "grad_norm": 0.34375, "learning_rate": 0.00018642196504722785, "loss": 0.8951, "step": 913 }, { "epoch": 0.5085547364028377, "grad_norm": 0.33984375, "learning_rate": 0.00018639257694286462, "loss": 0.846, "step": 914 }, { "epoch": 0.5091111420225344, "grad_norm": 0.326171875, "learning_rate": 0.00018636315939088292, "loss": 0.8076, "step": 915 }, { "epoch": 0.5096675476422312, "grad_norm": 0.322265625, "learning_rate": 0.00018633371240131, "loss": 0.7998, "step": 916 }, { "epoch": 0.5102239532619279, "grad_norm": 0.330078125, "learning_rate": 0.00018630423598418308, "loss": 0.7803, "step": 917 }, { "epoch": 0.5107803588816247, "grad_norm": 0.353515625, "learning_rate": 0.00018627473014954946, "loss": 0.9279, "step": 918 }, { "epoch": 0.5113367645013215, "grad_norm": 0.341796875, "learning_rate": 0.0001862451949074664, "loss": 0.9039, "step": 919 }, { "epoch": 0.5118931701210182, "grad_norm": 0.326171875, "learning_rate": 0.00018621563026800127, "loss": 0.7871, "step": 920 }, { "epoch": 0.512449575740715, "grad_norm": 0.3359375, "learning_rate": 0.00018618603624123145, "loss": 0.8431, "step": 921 }, { "epoch": 0.5130059813604118, "grad_norm": 0.33203125, "learning_rate": 0.00018615641283724425, "loss": 0.8864, "step": 922 }, { "epoch": 0.5135623869801085, "grad_norm": 0.3359375, "learning_rate": 0.00018612676006613706, "loss": 0.8569, "step": 923 }, { "epoch": 0.5141187925998053, "grad_norm": 0.345703125, "learning_rate": 0.00018609707793801726, "loss": 0.9234, "step": 924 }, { "epoch": 0.514675198219502, "grad_norm": 0.34765625, "learning_rate": 0.0001860673664630023, "loss": 0.8782, "step": 925 }, { "epoch": 0.5152316038391987, "grad_norm": 0.333984375, "learning_rate": 0.00018603762565121953, "loss": 0.8889, "step": 926 }, { "epoch": 0.5157880094588956, "grad_norm": 0.341796875, "learning_rate": 0.00018600785551280637, "loss": 0.9133, "step": 927 }, { "epoch": 0.5163444150785923, "grad_norm": 0.337890625, "learning_rate": 0.0001859780560579102, "loss": 0.8214, "step": 928 }, { "epoch": 0.516900820698289, "grad_norm": 0.310546875, "learning_rate": 0.00018594822729668843, "loss": 0.7542, "step": 929 }, { "epoch": 0.5174572263179859, "grad_norm": 0.326171875, "learning_rate": 0.00018591836923930843, "loss": 0.8442, "step": 930 }, { "epoch": 0.5180136319376826, "grad_norm": 0.359375, "learning_rate": 0.00018588848189594756, "loss": 0.9669, "step": 931 }, { "epoch": 0.5185700375573793, "grad_norm": 0.33203125, "learning_rate": 0.00018585856527679316, "loss": 0.8215, "step": 932 }, { "epoch": 0.5191264431770761, "grad_norm": 0.34765625, "learning_rate": 0.00018582861939204256, "loss": 0.8499, "step": 933 }, { "epoch": 0.5196828487967728, "grad_norm": 0.345703125, "learning_rate": 0.00018579864425190309, "loss": 0.8397, "step": 934 }, { "epoch": 0.5202392544164696, "grad_norm": 0.326171875, "learning_rate": 0.00018576863986659198, "loss": 0.8204, "step": 935 }, { "epoch": 0.5207956600361664, "grad_norm": 0.314453125, "learning_rate": 0.0001857386062463365, "loss": 0.765, "step": 936 }, { "epoch": 0.5213520656558631, "grad_norm": 0.341796875, "learning_rate": 0.0001857085434013738, "loss": 0.8809, "step": 937 }, { "epoch": 0.5219084712755598, "grad_norm": 0.326171875, "learning_rate": 0.0001856784513419511, "loss": 0.8837, "step": 938 }, { "epoch": 0.5224648768952567, "grad_norm": 0.3359375, "learning_rate": 0.00018564833007832555, "loss": 0.8835, "step": 939 }, { "epoch": 0.5230212825149534, "grad_norm": 0.345703125, "learning_rate": 0.00018561817962076414, "loss": 0.8856, "step": 940 }, { "epoch": 0.5235776881346501, "grad_norm": 0.33203125, "learning_rate": 0.00018558799997954402, "loss": 0.8138, "step": 941 }, { "epoch": 0.5241340937543469, "grad_norm": 0.34375, "learning_rate": 0.00018555779116495206, "loss": 0.8903, "step": 942 }, { "epoch": 0.5246904993740437, "grad_norm": 0.32421875, "learning_rate": 0.00018552755318728523, "loss": 0.7882, "step": 943 }, { "epoch": 0.5252469049937404, "grad_norm": 0.341796875, "learning_rate": 0.00018549728605685042, "loss": 0.8532, "step": 944 }, { "epoch": 0.5258033106134372, "grad_norm": 0.318359375, "learning_rate": 0.00018546698978396436, "loss": 0.7801, "step": 945 }, { "epoch": 0.5263597162331339, "grad_norm": 0.328125, "learning_rate": 0.00018543666437895385, "loss": 0.8937, "step": 946 }, { "epoch": 0.5269161218528308, "grad_norm": 0.341796875, "learning_rate": 0.00018540630985215548, "loss": 0.9237, "step": 947 }, { "epoch": 0.5274725274725275, "grad_norm": 0.349609375, "learning_rate": 0.00018537592621391591, "loss": 0.8728, "step": 948 }, { "epoch": 0.5280289330922242, "grad_norm": 0.33203125, "learning_rate": 0.00018534551347459163, "loss": 0.8142, "step": 949 }, { "epoch": 0.528585338711921, "grad_norm": 0.337890625, "learning_rate": 0.0001853150716445491, "loss": 0.9059, "step": 950 }, { "epoch": 0.5291417443316178, "grad_norm": 0.337890625, "learning_rate": 0.00018528460073416456, "loss": 0.8826, "step": 951 }, { "epoch": 0.5296981499513145, "grad_norm": 0.359375, "learning_rate": 0.00018525410075382438, "loss": 0.9263, "step": 952 }, { "epoch": 0.5302545555710113, "grad_norm": 0.341796875, "learning_rate": 0.0001852235717139247, "loss": 0.8751, "step": 953 }, { "epoch": 0.530810961190708, "grad_norm": 0.330078125, "learning_rate": 0.00018519301362487158, "loss": 0.833, "step": 954 }, { "epoch": 0.5313673668104048, "grad_norm": 0.34765625, "learning_rate": 0.00018516242649708103, "loss": 0.9195, "step": 955 }, { "epoch": 0.5319237724301016, "grad_norm": 0.359375, "learning_rate": 0.00018513181034097886, "loss": 0.961, "step": 956 }, { "epoch": 0.5324801780497983, "grad_norm": 0.326171875, "learning_rate": 0.0001851011651670009, "loss": 0.8375, "step": 957 }, { "epoch": 0.533036583669495, "grad_norm": 0.3359375, "learning_rate": 0.00018507049098559275, "loss": 0.8556, "step": 958 }, { "epoch": 0.5335929892891919, "grad_norm": 0.35546875, "learning_rate": 0.00018503978780721, "loss": 0.9757, "step": 959 }, { "epoch": 0.5341493949088886, "grad_norm": 0.349609375, "learning_rate": 0.00018500905564231807, "loss": 0.9229, "step": 960 }, { "epoch": 0.5347058005285853, "grad_norm": 0.330078125, "learning_rate": 0.0001849782945013923, "loss": 0.8092, "step": 961 }, { "epoch": 0.5352622061482821, "grad_norm": 0.359375, "learning_rate": 0.00018494750439491782, "loss": 0.8918, "step": 962 }, { "epoch": 0.5358186117679788, "grad_norm": 0.328125, "learning_rate": 0.00018491668533338973, "loss": 0.8153, "step": 963 }, { "epoch": 0.5363750173876756, "grad_norm": 0.333984375, "learning_rate": 0.00018488583732731293, "loss": 0.8872, "step": 964 }, { "epoch": 0.5369314230073724, "grad_norm": 0.35546875, "learning_rate": 0.00018485496038720222, "loss": 0.9895, "step": 965 }, { "epoch": 0.5374878286270691, "grad_norm": 0.34375, "learning_rate": 0.00018482405452358228, "loss": 0.8771, "step": 966 }, { "epoch": 0.5380442342467658, "grad_norm": 0.353515625, "learning_rate": 0.00018479311974698765, "loss": 0.9115, "step": 967 }, { "epoch": 0.5386006398664627, "grad_norm": 0.341796875, "learning_rate": 0.00018476215606796263, "loss": 0.8777, "step": 968 }, { "epoch": 0.5391570454861594, "grad_norm": 0.34375, "learning_rate": 0.00018473116349706153, "loss": 0.9589, "step": 969 }, { "epoch": 0.5397134511058562, "grad_norm": 0.306640625, "learning_rate": 0.00018470014204484834, "loss": 0.7294, "step": 970 }, { "epoch": 0.5402698567255529, "grad_norm": 0.333984375, "learning_rate": 0.000184669091721897, "loss": 0.8296, "step": 971 }, { "epoch": 0.5408262623452497, "grad_norm": 0.3515625, "learning_rate": 0.00018463801253879128, "loss": 0.9549, "step": 972 }, { "epoch": 0.5413826679649465, "grad_norm": 0.33203125, "learning_rate": 0.0001846069045061248, "loss": 0.8513, "step": 973 }, { "epoch": 0.5419390735846432, "grad_norm": 0.333984375, "learning_rate": 0.00018457576763450097, "loss": 0.879, "step": 974 }, { "epoch": 0.5424954792043399, "grad_norm": 0.337890625, "learning_rate": 0.00018454460193453303, "loss": 0.8198, "step": 975 }, { "epoch": 0.5430518848240368, "grad_norm": 0.333984375, "learning_rate": 0.00018451340741684408, "loss": 0.8208, "step": 976 }, { "epoch": 0.5436082904437335, "grad_norm": 0.345703125, "learning_rate": 0.000184482184092067, "loss": 0.9505, "step": 977 }, { "epoch": 0.5441646960634302, "grad_norm": 0.34375, "learning_rate": 0.00018445093197084456, "loss": 0.8958, "step": 978 }, { "epoch": 0.544721101683127, "grad_norm": 0.33203125, "learning_rate": 0.00018441965106382927, "loss": 0.8845, "step": 979 }, { "epoch": 0.5452775073028238, "grad_norm": 0.31640625, "learning_rate": 0.0001843883413816835, "loss": 0.7969, "step": 980 }, { "epoch": 0.5458339129225205, "grad_norm": 0.333984375, "learning_rate": 0.0001843570029350794, "loss": 0.8301, "step": 981 }, { "epoch": 0.5463903185422173, "grad_norm": 0.306640625, "learning_rate": 0.00018432563573469895, "loss": 0.7873, "step": 982 }, { "epoch": 0.546946724161914, "grad_norm": 0.34765625, "learning_rate": 0.0001842942397912339, "loss": 0.8894, "step": 983 }, { "epoch": 0.5475031297816108, "grad_norm": 0.33984375, "learning_rate": 0.00018426281511538584, "loss": 0.91, "step": 984 }, { "epoch": 0.5480595354013076, "grad_norm": 0.328125, "learning_rate": 0.00018423136171786611, "loss": 0.823, "step": 985 }, { "epoch": 0.5486159410210043, "grad_norm": 0.318359375, "learning_rate": 0.00018419987960939586, "loss": 0.8033, "step": 986 }, { "epoch": 0.549172346640701, "grad_norm": 0.34375, "learning_rate": 0.00018416836880070603, "loss": 0.8735, "step": 987 }, { "epoch": 0.5497287522603979, "grad_norm": 0.3203125, "learning_rate": 0.00018413682930253734, "loss": 0.8009, "step": 988 }, { "epoch": 0.5502851578800946, "grad_norm": 0.34765625, "learning_rate": 0.00018410526112564026, "loss": 0.9164, "step": 989 }, { "epoch": 0.5508415634997913, "grad_norm": 0.328125, "learning_rate": 0.0001840736642807751, "loss": 0.869, "step": 990 }, { "epoch": 0.5513979691194881, "grad_norm": 0.3359375, "learning_rate": 0.00018404203877871186, "loss": 0.7933, "step": 991 }, { "epoch": 0.5519543747391848, "grad_norm": 0.330078125, "learning_rate": 0.0001840103846302304, "loss": 0.8246, "step": 992 }, { "epoch": 0.5525107803588816, "grad_norm": 0.337890625, "learning_rate": 0.0001839787018461203, "loss": 0.8753, "step": 993 }, { "epoch": 0.5530671859785784, "grad_norm": 0.330078125, "learning_rate": 0.00018394699043718085, "loss": 0.8591, "step": 994 }, { "epoch": 0.5536235915982751, "grad_norm": 0.333984375, "learning_rate": 0.00018391525041422114, "loss": 0.8757, "step": 995 }, { "epoch": 0.554179997217972, "grad_norm": 0.333984375, "learning_rate": 0.00018388348178806008, "loss": 0.8981, "step": 996 }, { "epoch": 0.5547364028376687, "grad_norm": 0.326171875, "learning_rate": 0.00018385168456952621, "loss": 0.7889, "step": 997 }, { "epoch": 0.5552928084573654, "grad_norm": 0.333984375, "learning_rate": 0.0001838198587694579, "loss": 0.8176, "step": 998 }, { "epoch": 0.5558492140770622, "grad_norm": 0.34765625, "learning_rate": 0.0001837880043987032, "loss": 0.9225, "step": 999 }, { "epoch": 0.556405619696759, "grad_norm": 0.328125, "learning_rate": 0.00018375612146812, "loss": 0.8243, "step": 1000 }, { "epoch": 0.5569620253164557, "grad_norm": 0.32421875, "learning_rate": 0.00018372420998857584, "loss": 0.7683, "step": 1001 }, { "epoch": 0.5575184309361525, "grad_norm": 0.33203125, "learning_rate": 0.00018369226997094792, "loss": 0.906, "step": 1002 }, { "epoch": 0.5580748365558492, "grad_norm": 0.341796875, "learning_rate": 0.0001836603014261234, "loss": 0.8942, "step": 1003 }, { "epoch": 0.5586312421755459, "grad_norm": 0.34375, "learning_rate": 0.00018362830436499886, "loss": 0.8148, "step": 1004 }, { "epoch": 0.5591876477952428, "grad_norm": 0.33984375, "learning_rate": 0.0001835962787984809, "loss": 0.9291, "step": 1005 }, { "epoch": 0.5597440534149395, "grad_norm": 0.318359375, "learning_rate": 0.00018356422473748564, "loss": 0.7764, "step": 1006 }, { "epoch": 0.5603004590346362, "grad_norm": 0.3359375, "learning_rate": 0.00018353214219293896, "loss": 0.8792, "step": 1007 }, { "epoch": 0.560856864654333, "grad_norm": 0.34375, "learning_rate": 0.0001835000311757765, "loss": 0.8608, "step": 1008 }, { "epoch": 0.5614132702740298, "grad_norm": 0.33984375, "learning_rate": 0.0001834678916969435, "loss": 0.927, "step": 1009 }, { "epoch": 0.5619696758937265, "grad_norm": 0.314453125, "learning_rate": 0.00018343572376739505, "loss": 0.7639, "step": 1010 }, { "epoch": 0.5625260815134233, "grad_norm": 0.345703125, "learning_rate": 0.00018340352739809577, "loss": 0.8458, "step": 1011 }, { "epoch": 0.56308248713312, "grad_norm": 0.341796875, "learning_rate": 0.0001833713026000201, "loss": 0.9023, "step": 1012 }, { "epoch": 0.5636388927528168, "grad_norm": 0.322265625, "learning_rate": 0.0001833390493841521, "loss": 0.8104, "step": 1013 }, { "epoch": 0.5641952983725136, "grad_norm": 0.349609375, "learning_rate": 0.00018330676776148557, "loss": 0.8363, "step": 1014 }, { "epoch": 0.5647517039922103, "grad_norm": 0.33203125, "learning_rate": 0.00018327445774302397, "loss": 0.8123, "step": 1015 }, { "epoch": 0.565308109611907, "grad_norm": 0.357421875, "learning_rate": 0.00018324211933978043, "loss": 0.8886, "step": 1016 }, { "epoch": 0.5658645152316039, "grad_norm": 0.31640625, "learning_rate": 0.0001832097525627777, "loss": 0.7553, "step": 1017 }, { "epoch": 0.5664209208513006, "grad_norm": 0.33984375, "learning_rate": 0.00018317735742304832, "loss": 0.8144, "step": 1018 }, { "epoch": 0.5669773264709974, "grad_norm": 0.3359375, "learning_rate": 0.00018314493393163444, "loss": 0.8803, "step": 1019 }, { "epoch": 0.5675337320906941, "grad_norm": 0.330078125, "learning_rate": 0.0001831124820995879, "loss": 0.8016, "step": 1020 }, { "epoch": 0.5680901377103909, "grad_norm": 0.3203125, "learning_rate": 0.00018308000193797012, "loss": 0.7583, "step": 1021 }, { "epoch": 0.5686465433300877, "grad_norm": 0.3359375, "learning_rate": 0.00018304749345785223, "loss": 0.8315, "step": 1022 }, { "epoch": 0.5692029489497844, "grad_norm": 0.322265625, "learning_rate": 0.00018301495667031505, "loss": 0.7795, "step": 1023 }, { "epoch": 0.5697593545694811, "grad_norm": 0.3203125, "learning_rate": 0.000182982391586449, "loss": 0.7642, "step": 1024 }, { "epoch": 0.570315760189178, "grad_norm": 0.357421875, "learning_rate": 0.00018294979821735413, "loss": 0.8609, "step": 1025 }, { "epoch": 0.5708721658088747, "grad_norm": 0.39453125, "learning_rate": 0.00018291717657414013, "loss": 0.9518, "step": 1026 }, { "epoch": 0.5714285714285714, "grad_norm": 0.33984375, "learning_rate": 0.00018288452666792644, "loss": 0.8299, "step": 1027 }, { "epoch": 0.5719849770482682, "grad_norm": 0.3515625, "learning_rate": 0.000182851848509842, "loss": 0.9225, "step": 1028 }, { "epoch": 0.572541382667965, "grad_norm": 0.3203125, "learning_rate": 0.00018281914211102542, "loss": 0.7879, "step": 1029 }, { "epoch": 0.5730977882876617, "grad_norm": 0.3046875, "learning_rate": 0.00018278640748262498, "loss": 0.6934, "step": 1030 }, { "epoch": 0.5736541939073585, "grad_norm": 0.33984375, "learning_rate": 0.00018275364463579853, "loss": 0.9117, "step": 1031 }, { "epoch": 0.5742105995270552, "grad_norm": 0.34375, "learning_rate": 0.00018272085358171354, "loss": 0.8398, "step": 1032 }, { "epoch": 0.5747670051467519, "grad_norm": 0.3203125, "learning_rate": 0.00018268803433154712, "loss": 0.7672, "step": 1033 }, { "epoch": 0.5753234107664488, "grad_norm": 0.3203125, "learning_rate": 0.00018265518689648596, "loss": 0.7822, "step": 1034 }, { "epoch": 0.5758798163861455, "grad_norm": 0.318359375, "learning_rate": 0.00018262231128772643, "loss": 0.7824, "step": 1035 }, { "epoch": 0.5764362220058422, "grad_norm": 0.357421875, "learning_rate": 0.00018258940751647443, "loss": 0.9375, "step": 1036 }, { "epoch": 0.576992627625539, "grad_norm": 0.32421875, "learning_rate": 0.00018255647559394542, "loss": 0.8067, "step": 1037 }, { "epoch": 0.5775490332452358, "grad_norm": 0.345703125, "learning_rate": 0.0001825235155313646, "loss": 0.912, "step": 1038 }, { "epoch": 0.5781054388649325, "grad_norm": 0.3515625, "learning_rate": 0.00018249052733996664, "loss": 0.9291, "step": 1039 }, { "epoch": 0.5786618444846293, "grad_norm": 0.341796875, "learning_rate": 0.0001824575110309958, "loss": 0.8022, "step": 1040 }, { "epoch": 0.579218250104326, "grad_norm": 0.333984375, "learning_rate": 0.00018242446661570605, "loss": 0.8948, "step": 1041 }, { "epoch": 0.5797746557240228, "grad_norm": 0.37109375, "learning_rate": 0.00018239139410536077, "loss": 0.9978, "step": 1042 }, { "epoch": 0.5803310613437196, "grad_norm": 0.32421875, "learning_rate": 0.00018235829351123304, "loss": 0.8297, "step": 1043 }, { "epoch": 0.5808874669634163, "grad_norm": 0.337890625, "learning_rate": 0.00018232516484460541, "loss": 0.8944, "step": 1044 }, { "epoch": 0.5814438725831131, "grad_norm": 0.3359375, "learning_rate": 0.00018229200811677016, "loss": 0.9035, "step": 1045 }, { "epoch": 0.5820002782028099, "grad_norm": 0.322265625, "learning_rate": 0.00018225882333902896, "loss": 0.7308, "step": 1046 }, { "epoch": 0.5825566838225066, "grad_norm": 0.34765625, "learning_rate": 0.0001822256105226931, "loss": 0.8482, "step": 1047 }, { "epoch": 0.5831130894422034, "grad_norm": 0.31640625, "learning_rate": 0.00018219236967908348, "loss": 0.7428, "step": 1048 }, { "epoch": 0.5836694950619001, "grad_norm": 0.33984375, "learning_rate": 0.00018215910081953053, "loss": 0.8924, "step": 1049 }, { "epoch": 0.5842259006815969, "grad_norm": 0.314453125, "learning_rate": 0.00018212580395537417, "loss": 0.7806, "step": 1050 }, { "epoch": 0.5847823063012937, "grad_norm": 0.333984375, "learning_rate": 0.0001820924790979639, "loss": 0.8435, "step": 1051 }, { "epoch": 0.5853387119209904, "grad_norm": 0.34375, "learning_rate": 0.00018205912625865881, "loss": 0.9288, "step": 1052 }, { "epoch": 0.5858951175406871, "grad_norm": 0.32421875, "learning_rate": 0.0001820257454488275, "loss": 0.7463, "step": 1053 }, { "epoch": 0.586451523160384, "grad_norm": 0.326171875, "learning_rate": 0.00018199233667984802, "loss": 0.7873, "step": 1054 }, { "epoch": 0.5870079287800807, "grad_norm": 0.341796875, "learning_rate": 0.00018195889996310813, "loss": 0.8771, "step": 1055 }, { "epoch": 0.5875643343997774, "grad_norm": 0.330078125, "learning_rate": 0.0001819254353100049, "loss": 0.8398, "step": 1056 }, { "epoch": 0.5881207400194742, "grad_norm": 0.345703125, "learning_rate": 0.00018189194273194513, "loss": 0.9242, "step": 1057 }, { "epoch": 0.588677145639171, "grad_norm": 0.341796875, "learning_rate": 0.00018185842224034495, "loss": 0.889, "step": 1058 }, { "epoch": 0.5892335512588677, "grad_norm": 0.33984375, "learning_rate": 0.00018182487384663017, "loss": 0.9042, "step": 1059 }, { "epoch": 0.5897899568785645, "grad_norm": 0.328125, "learning_rate": 0.000181791297562236, "loss": 0.8742, "step": 1060 }, { "epoch": 0.5903463624982612, "grad_norm": 0.328125, "learning_rate": 0.00018175769339860714, "loss": 0.8162, "step": 1061 }, { "epoch": 0.5909027681179579, "grad_norm": 0.345703125, "learning_rate": 0.00018172406136719794, "loss": 0.9006, "step": 1062 }, { "epoch": 0.5914591737376548, "grad_norm": 0.322265625, "learning_rate": 0.00018169040147947212, "loss": 0.7795, "step": 1063 }, { "epoch": 0.5920155793573515, "grad_norm": 0.345703125, "learning_rate": 0.00018165671374690293, "loss": 0.8475, "step": 1064 }, { "epoch": 0.5925719849770482, "grad_norm": 0.322265625, "learning_rate": 0.00018162299818097303, "loss": 0.803, "step": 1065 }, { "epoch": 0.593128390596745, "grad_norm": 0.3359375, "learning_rate": 0.0001815892547931748, "loss": 0.8489, "step": 1066 }, { "epoch": 0.5936847962164418, "grad_norm": 0.330078125, "learning_rate": 0.00018155548359500981, "loss": 0.8463, "step": 1067 }, { "epoch": 0.5942412018361386, "grad_norm": 0.328125, "learning_rate": 0.00018152168459798935, "loss": 0.8165, "step": 1068 }, { "epoch": 0.5947976074558353, "grad_norm": 0.330078125, "learning_rate": 0.00018148785781363401, "loss": 0.8304, "step": 1069 }, { "epoch": 0.595354013075532, "grad_norm": 0.306640625, "learning_rate": 0.000181454003253474, "loss": 0.7645, "step": 1070 }, { "epoch": 0.5959104186952289, "grad_norm": 0.31640625, "learning_rate": 0.00018142012092904885, "loss": 0.8276, "step": 1071 }, { "epoch": 0.5964668243149256, "grad_norm": 0.337890625, "learning_rate": 0.00018138621085190767, "loss": 0.889, "step": 1072 }, { "epoch": 0.5970232299346223, "grad_norm": 0.34765625, "learning_rate": 0.000181352273033609, "loss": 0.83, "step": 1073 }, { "epoch": 0.5975796355543191, "grad_norm": 0.3515625, "learning_rate": 0.00018131830748572078, "loss": 0.9765, "step": 1074 }, { "epoch": 0.5981360411740159, "grad_norm": 0.337890625, "learning_rate": 0.00018128431421982047, "loss": 0.8316, "step": 1075 }, { "epoch": 0.5986924467937126, "grad_norm": 0.33984375, "learning_rate": 0.00018125029324749499, "loss": 0.867, "step": 1076 }, { "epoch": 0.5992488524134094, "grad_norm": 0.314453125, "learning_rate": 0.00018121624458034062, "loss": 0.7405, "step": 1077 }, { "epoch": 0.5998052580331061, "grad_norm": 0.333984375, "learning_rate": 0.0001811821682299631, "loss": 0.9073, "step": 1078 }, { "epoch": 0.6003616636528029, "grad_norm": 0.34375, "learning_rate": 0.0001811480642079777, "loss": 0.8368, "step": 1079 }, { "epoch": 0.6009180692724997, "grad_norm": 0.3125, "learning_rate": 0.00018111393252600902, "loss": 0.7902, "step": 1080 }, { "epoch": 0.6014744748921964, "grad_norm": 0.330078125, "learning_rate": 0.00018107977319569115, "loss": 0.8296, "step": 1081 }, { "epoch": 0.6020308805118931, "grad_norm": 0.32421875, "learning_rate": 0.00018104558622866756, "loss": 0.8123, "step": 1082 }, { "epoch": 0.60258728613159, "grad_norm": 0.33984375, "learning_rate": 0.00018101137163659115, "loss": 0.8452, "step": 1083 }, { "epoch": 0.6031436917512867, "grad_norm": 0.353515625, "learning_rate": 0.0001809771294311242, "loss": 0.8303, "step": 1084 }, { "epoch": 0.6037000973709834, "grad_norm": 0.3203125, "learning_rate": 0.00018094285962393856, "loss": 0.7941, "step": 1085 }, { "epoch": 0.6042565029906802, "grad_norm": 0.345703125, "learning_rate": 0.0001809085622267153, "loss": 0.9412, "step": 1086 }, { "epoch": 0.604812908610377, "grad_norm": 0.333984375, "learning_rate": 0.00018087423725114498, "loss": 0.8406, "step": 1087 }, { "epoch": 0.6053693142300737, "grad_norm": 0.34765625, "learning_rate": 0.00018083988470892756, "loss": 0.9588, "step": 1088 }, { "epoch": 0.6059257198497705, "grad_norm": 0.349609375, "learning_rate": 0.0001808055046117724, "loss": 0.9362, "step": 1089 }, { "epoch": 0.6064821254694672, "grad_norm": 0.33203125, "learning_rate": 0.00018077109697139818, "loss": 0.8893, "step": 1090 }, { "epoch": 0.607038531089164, "grad_norm": 0.328125, "learning_rate": 0.00018073666179953306, "loss": 0.8135, "step": 1091 }, { "epoch": 0.6075949367088608, "grad_norm": 0.337890625, "learning_rate": 0.00018070219910791457, "loss": 0.8649, "step": 1092 }, { "epoch": 0.6081513423285575, "grad_norm": 0.322265625, "learning_rate": 0.0001806677089082896, "loss": 0.8343, "step": 1093 }, { "epoch": 0.6087077479482543, "grad_norm": 0.3359375, "learning_rate": 0.00018063319121241434, "loss": 0.8991, "step": 1094 }, { "epoch": 0.609264153567951, "grad_norm": 0.3359375, "learning_rate": 0.00018059864603205456, "loss": 0.8552, "step": 1095 }, { "epoch": 0.6098205591876478, "grad_norm": 0.322265625, "learning_rate": 0.00018056407337898515, "loss": 0.7787, "step": 1096 }, { "epoch": 0.6103769648073446, "grad_norm": 0.369140625, "learning_rate": 0.00018052947326499056, "loss": 0.8752, "step": 1097 }, { "epoch": 0.6109333704270413, "grad_norm": 0.337890625, "learning_rate": 0.0001804948457018645, "loss": 0.851, "step": 1098 }, { "epoch": 0.611489776046738, "grad_norm": 0.3203125, "learning_rate": 0.00018046019070141006, "loss": 0.8067, "step": 1099 }, { "epoch": 0.6120461816664349, "grad_norm": 0.333984375, "learning_rate": 0.00018042550827543962, "loss": 0.8336, "step": 1100 }, { "epoch": 0.6126025872861316, "grad_norm": 0.33984375, "learning_rate": 0.00018039079843577507, "loss": 0.8959, "step": 1101 }, { "epoch": 0.6131589929058283, "grad_norm": 0.328125, "learning_rate": 0.0001803560611942475, "loss": 0.8092, "step": 1102 }, { "epoch": 0.6137153985255251, "grad_norm": 0.33203125, "learning_rate": 0.00018032129656269738, "loss": 0.8899, "step": 1103 }, { "epoch": 0.6142718041452219, "grad_norm": 0.365234375, "learning_rate": 0.0001802865045529745, "loss": 0.9702, "step": 1104 }, { "epoch": 0.6148282097649186, "grad_norm": 0.31640625, "learning_rate": 0.00018025168517693803, "loss": 0.7923, "step": 1105 }, { "epoch": 0.6153846153846154, "grad_norm": 0.328125, "learning_rate": 0.00018021683844645644, "loss": 0.9082, "step": 1106 }, { "epoch": 0.6159410210043121, "grad_norm": 0.33203125, "learning_rate": 0.0001801819643734075, "loss": 0.8512, "step": 1107 }, { "epoch": 0.6164974266240089, "grad_norm": 0.333984375, "learning_rate": 0.00018014706296967838, "loss": 0.8654, "step": 1108 }, { "epoch": 0.6170538322437057, "grad_norm": 0.333984375, "learning_rate": 0.00018011213424716545, "loss": 0.9283, "step": 1109 }, { "epoch": 0.6176102378634024, "grad_norm": 0.31640625, "learning_rate": 0.00018007717821777445, "loss": 0.7825, "step": 1110 }, { "epoch": 0.6181666434830991, "grad_norm": 0.33203125, "learning_rate": 0.00018004219489342046, "loss": 0.8412, "step": 1111 }, { "epoch": 0.618723049102796, "grad_norm": 0.33203125, "learning_rate": 0.00018000718428602786, "loss": 0.8454, "step": 1112 }, { "epoch": 0.6192794547224927, "grad_norm": 0.337890625, "learning_rate": 0.00017997214640753023, "loss": 0.9071, "step": 1113 }, { "epoch": 0.6198358603421894, "grad_norm": 0.333984375, "learning_rate": 0.00017993708126987058, "loss": 0.9327, "step": 1114 }, { "epoch": 0.6203922659618862, "grad_norm": 0.314453125, "learning_rate": 0.0001799019888850011, "loss": 0.7861, "step": 1115 }, { "epoch": 0.620948671581583, "grad_norm": 0.3203125, "learning_rate": 0.00017986686926488337, "loss": 0.8135, "step": 1116 }, { "epoch": 0.6215050772012798, "grad_norm": 0.34375, "learning_rate": 0.00017983172242148814, "loss": 0.9299, "step": 1117 }, { "epoch": 0.6220614828209765, "grad_norm": 0.345703125, "learning_rate": 0.00017979654836679555, "loss": 0.8861, "step": 1118 }, { "epoch": 0.6226178884406732, "grad_norm": 0.314453125, "learning_rate": 0.00017976134711279496, "loss": 0.7301, "step": 1119 }, { "epoch": 0.6231742940603701, "grad_norm": 0.33203125, "learning_rate": 0.00017972611867148497, "loss": 0.8426, "step": 1120 }, { "epoch": 0.6237306996800668, "grad_norm": 0.345703125, "learning_rate": 0.00017969086305487354, "loss": 0.9108, "step": 1121 }, { "epoch": 0.6242871052997635, "grad_norm": 0.345703125, "learning_rate": 0.00017965558027497775, "loss": 0.9061, "step": 1122 }, { "epoch": 0.6248435109194603, "grad_norm": 0.326171875, "learning_rate": 0.00017962027034382411, "loss": 0.7835, "step": 1123 }, { "epoch": 0.625399916539157, "grad_norm": 0.353515625, "learning_rate": 0.00017958493327344826, "loss": 0.92, "step": 1124 }, { "epoch": 0.6259563221588538, "grad_norm": 0.345703125, "learning_rate": 0.00017954956907589517, "loss": 0.9074, "step": 1125 }, { "epoch": 0.6265127277785506, "grad_norm": 0.33203125, "learning_rate": 0.00017951417776321894, "loss": 0.8155, "step": 1126 }, { "epoch": 0.6270691333982473, "grad_norm": 0.337890625, "learning_rate": 0.00017947875934748306, "loss": 0.8642, "step": 1127 }, { "epoch": 0.627625539017944, "grad_norm": 0.35546875, "learning_rate": 0.00017944331384076015, "loss": 0.9221, "step": 1128 }, { "epoch": 0.6281819446376409, "grad_norm": 0.330078125, "learning_rate": 0.00017940784125513212, "loss": 0.8576, "step": 1129 }, { "epoch": 0.6287383502573376, "grad_norm": 0.34375, "learning_rate": 0.0001793723416026901, "loss": 0.8981, "step": 1130 }, { "epoch": 0.6292947558770343, "grad_norm": 0.359375, "learning_rate": 0.0001793368148955344, "loss": 0.8868, "step": 1131 }, { "epoch": 0.6298511614967311, "grad_norm": 0.359375, "learning_rate": 0.00017930126114577464, "loss": 0.9415, "step": 1132 }, { "epoch": 0.6304075671164279, "grad_norm": 0.328125, "learning_rate": 0.00017926568036552953, "loss": 0.7919, "step": 1133 }, { "epoch": 0.6309639727361246, "grad_norm": 0.357421875, "learning_rate": 0.00017923007256692713, "loss": 0.9498, "step": 1134 }, { "epoch": 0.6315203783558214, "grad_norm": 0.34375, "learning_rate": 0.0001791944377621047, "loss": 0.9427, "step": 1135 }, { "epoch": 0.6320767839755181, "grad_norm": 0.3359375, "learning_rate": 0.00017915877596320855, "loss": 0.8621, "step": 1136 }, { "epoch": 0.6326331895952149, "grad_norm": 0.302734375, "learning_rate": 0.00017912308718239433, "loss": 0.7246, "step": 1137 }, { "epoch": 0.6331895952149117, "grad_norm": 0.365234375, "learning_rate": 0.0001790873714318269, "loss": 0.9493, "step": 1138 }, { "epoch": 0.6337460008346084, "grad_norm": 0.328125, "learning_rate": 0.00017905162872368022, "loss": 0.8226, "step": 1139 }, { "epoch": 0.6343024064543051, "grad_norm": 0.34765625, "learning_rate": 0.0001790158590701375, "loss": 0.8976, "step": 1140 }, { "epoch": 0.634858812074002, "grad_norm": 0.345703125, "learning_rate": 0.0001789800624833911, "loss": 0.9134, "step": 1141 }, { "epoch": 0.6354152176936987, "grad_norm": 0.328125, "learning_rate": 0.0001789442389756426, "loss": 0.8021, "step": 1142 }, { "epoch": 0.6359716233133955, "grad_norm": 0.33984375, "learning_rate": 0.00017890838855910276, "loss": 0.8757, "step": 1143 }, { "epoch": 0.6365280289330922, "grad_norm": 0.328125, "learning_rate": 0.00017887251124599143, "loss": 0.8675, "step": 1144 }, { "epoch": 0.637084434552789, "grad_norm": 0.326171875, "learning_rate": 0.00017883660704853775, "loss": 0.811, "step": 1145 }, { "epoch": 0.6376408401724858, "grad_norm": 0.33984375, "learning_rate": 0.0001788006759789799, "loss": 0.8422, "step": 1146 }, { "epoch": 0.6381972457921825, "grad_norm": 0.318359375, "learning_rate": 0.00017876471804956535, "loss": 0.751, "step": 1147 }, { "epoch": 0.6387536514118792, "grad_norm": 0.326171875, "learning_rate": 0.0001787287332725506, "loss": 0.793, "step": 1148 }, { "epoch": 0.6393100570315761, "grad_norm": 0.326171875, "learning_rate": 0.00017869272166020134, "loss": 0.7835, "step": 1149 }, { "epoch": 0.6398664626512728, "grad_norm": 0.33203125, "learning_rate": 0.0001786566832247925, "loss": 0.8919, "step": 1150 }, { "epoch": 0.6404228682709695, "grad_norm": 0.337890625, "learning_rate": 0.00017862061797860798, "loss": 0.8714, "step": 1151 }, { "epoch": 0.6409792738906663, "grad_norm": 0.33984375, "learning_rate": 0.000178584525933941, "loss": 0.9351, "step": 1152 }, { "epoch": 0.641535679510363, "grad_norm": 0.33203125, "learning_rate": 0.0001785484071030938, "loss": 0.92, "step": 1153 }, { "epoch": 0.6420920851300598, "grad_norm": 0.333984375, "learning_rate": 0.00017851226149837778, "loss": 0.8925, "step": 1154 }, { "epoch": 0.6426484907497566, "grad_norm": 0.33984375, "learning_rate": 0.00017847608913211342, "loss": 0.9021, "step": 1155 }, { "epoch": 0.6432048963694533, "grad_norm": 0.337890625, "learning_rate": 0.00017843989001663043, "loss": 0.8723, "step": 1156 }, { "epoch": 0.64376130198915, "grad_norm": 0.33984375, "learning_rate": 0.00017840366416426755, "loss": 0.9426, "step": 1157 }, { "epoch": 0.6443177076088469, "grad_norm": 0.3515625, "learning_rate": 0.00017836741158737266, "loss": 0.8317, "step": 1158 }, { "epoch": 0.6448741132285436, "grad_norm": 0.3359375, "learning_rate": 0.0001783311322983028, "loss": 0.906, "step": 1159 }, { "epoch": 0.6454305188482403, "grad_norm": 0.33203125, "learning_rate": 0.00017829482630942394, "loss": 0.8644, "step": 1160 }, { "epoch": 0.6459869244679372, "grad_norm": 0.3125, "learning_rate": 0.0001782584936331114, "loss": 0.7903, "step": 1161 }, { "epoch": 0.6465433300876339, "grad_norm": 0.33203125, "learning_rate": 0.0001782221342817494, "loss": 0.8909, "step": 1162 }, { "epoch": 0.6470997357073306, "grad_norm": 0.333984375, "learning_rate": 0.00017818574826773133, "loss": 0.8217, "step": 1163 }, { "epoch": 0.6476561413270274, "grad_norm": 0.31640625, "learning_rate": 0.0001781493356034597, "loss": 0.7876, "step": 1164 }, { "epoch": 0.6482125469467241, "grad_norm": 0.337890625, "learning_rate": 0.00017811289630134601, "loss": 0.8962, "step": 1165 }, { "epoch": 0.648768952566421, "grad_norm": 0.32421875, "learning_rate": 0.00017807643037381094, "loss": 0.8133, "step": 1166 }, { "epoch": 0.6493253581861177, "grad_norm": 0.302734375, "learning_rate": 0.0001780399378332842, "loss": 0.6899, "step": 1167 }, { "epoch": 0.6498817638058144, "grad_norm": 0.33984375, "learning_rate": 0.00017800341869220456, "loss": 0.8604, "step": 1168 }, { "epoch": 0.6504381694255112, "grad_norm": 0.328125, "learning_rate": 0.00017796687296301985, "loss": 0.7746, "step": 1169 }, { "epoch": 0.650994575045208, "grad_norm": 0.330078125, "learning_rate": 0.00017793030065818699, "loss": 0.8599, "step": 1170 }, { "epoch": 0.6515509806649047, "grad_norm": 0.326171875, "learning_rate": 0.00017789370179017194, "loss": 0.8365, "step": 1171 }, { "epoch": 0.6521073862846015, "grad_norm": 0.32421875, "learning_rate": 0.00017785707637144976, "loss": 0.8358, "step": 1172 }, { "epoch": 0.6526637919042982, "grad_norm": 0.3359375, "learning_rate": 0.0001778204244145045, "loss": 0.8689, "step": 1173 }, { "epoch": 0.653220197523995, "grad_norm": 0.32421875, "learning_rate": 0.00017778374593182931, "loss": 0.8538, "step": 1174 }, { "epoch": 0.6537766031436918, "grad_norm": 0.337890625, "learning_rate": 0.00017774704093592632, "loss": 0.8759, "step": 1175 }, { "epoch": 0.6543330087633885, "grad_norm": 0.337890625, "learning_rate": 0.0001777103094393067, "loss": 0.902, "step": 1176 }, { "epoch": 0.6548894143830852, "grad_norm": 0.3203125, "learning_rate": 0.00017767355145449073, "loss": 0.8503, "step": 1177 }, { "epoch": 0.6554458200027821, "grad_norm": 0.322265625, "learning_rate": 0.0001776367669940077, "loss": 0.8096, "step": 1178 }, { "epoch": 0.6560022256224788, "grad_norm": 0.32421875, "learning_rate": 0.0001775999560703958, "loss": 0.7987, "step": 1179 }, { "epoch": 0.6565586312421755, "grad_norm": 0.333984375, "learning_rate": 0.0001775631186962024, "loss": 0.8789, "step": 1180 }, { "epoch": 0.6571150368618723, "grad_norm": 0.3359375, "learning_rate": 0.00017752625488398383, "loss": 0.8525, "step": 1181 }, { "epoch": 0.6576714424815691, "grad_norm": 0.328125, "learning_rate": 0.0001774893646463054, "loss": 0.8729, "step": 1182 }, { "epoch": 0.6582278481012658, "grad_norm": 0.337890625, "learning_rate": 0.00017745244799574145, "loss": 0.9459, "step": 1183 }, { "epoch": 0.6587842537209626, "grad_norm": 0.333984375, "learning_rate": 0.00017741550494487532, "loss": 0.8822, "step": 1184 }, { "epoch": 0.6593406593406593, "grad_norm": 0.35546875, "learning_rate": 0.00017737853550629936, "loss": 0.9891, "step": 1185 }, { "epoch": 0.659897064960356, "grad_norm": 0.31640625, "learning_rate": 0.0001773415396926149, "loss": 0.7885, "step": 1186 }, { "epoch": 0.6604534705800529, "grad_norm": 0.337890625, "learning_rate": 0.00017730451751643228, "loss": 0.8582, "step": 1187 }, { "epoch": 0.6610098761997496, "grad_norm": 0.328125, "learning_rate": 0.00017726746899037079, "loss": 0.9027, "step": 1188 }, { "epoch": 0.6615662818194463, "grad_norm": 0.32421875, "learning_rate": 0.00017723039412705877, "loss": 0.802, "step": 1189 }, { "epoch": 0.6621226874391432, "grad_norm": 0.33984375, "learning_rate": 0.00017719329293913342, "loss": 0.8545, "step": 1190 }, { "epoch": 0.6626790930588399, "grad_norm": 0.326171875, "learning_rate": 0.00017715616543924104, "loss": 0.8496, "step": 1191 }, { "epoch": 0.6632354986785367, "grad_norm": 0.31640625, "learning_rate": 0.0001771190116400368, "loss": 0.7292, "step": 1192 }, { "epoch": 0.6637919042982334, "grad_norm": 0.326171875, "learning_rate": 0.00017708183155418493, "loss": 0.7995, "step": 1193 }, { "epoch": 0.6643483099179301, "grad_norm": 0.328125, "learning_rate": 0.00017704462519435853, "loss": 0.8858, "step": 1194 }, { "epoch": 0.664904715537627, "grad_norm": 0.318359375, "learning_rate": 0.0001770073925732397, "loss": 0.8412, "step": 1195 }, { "epoch": 0.6654611211573237, "grad_norm": 0.31640625, "learning_rate": 0.0001769701337035195, "loss": 0.8375, "step": 1196 }, { "epoch": 0.6660175267770204, "grad_norm": 0.337890625, "learning_rate": 0.0001769328485978979, "loss": 0.8676, "step": 1197 }, { "epoch": 0.6665739323967173, "grad_norm": 0.361328125, "learning_rate": 0.00017689553726908387, "loss": 0.9652, "step": 1198 }, { "epoch": 0.667130338016414, "grad_norm": 0.32421875, "learning_rate": 0.00017685819972979524, "loss": 0.7994, "step": 1199 }, { "epoch": 0.6676867436361107, "grad_norm": 0.333984375, "learning_rate": 0.0001768208359927588, "loss": 0.7993, "step": 1200 }, { "epoch": 0.6682431492558075, "grad_norm": 0.328125, "learning_rate": 0.00017678344607071034, "loss": 0.7557, "step": 1201 }, { "epoch": 0.6687995548755042, "grad_norm": 0.322265625, "learning_rate": 0.0001767460299763945, "loss": 0.8166, "step": 1202 }, { "epoch": 0.669355960495201, "grad_norm": 0.33203125, "learning_rate": 0.00017670858772256487, "loss": 0.8687, "step": 1203 }, { "epoch": 0.6699123661148978, "grad_norm": 0.328125, "learning_rate": 0.00017667111932198392, "loss": 0.8069, "step": 1204 }, { "epoch": 0.6704687717345945, "grad_norm": 0.333984375, "learning_rate": 0.0001766336247874231, "loss": 0.8764, "step": 1205 }, { "epoch": 0.6710251773542912, "grad_norm": 0.310546875, "learning_rate": 0.00017659610413166269, "loss": 0.7519, "step": 1206 }, { "epoch": 0.6715815829739881, "grad_norm": 0.353515625, "learning_rate": 0.00017655855736749198, "loss": 0.945, "step": 1207 }, { "epoch": 0.6721379885936848, "grad_norm": 0.328125, "learning_rate": 0.00017652098450770902, "loss": 0.8002, "step": 1208 }, { "epoch": 0.6726943942133815, "grad_norm": 0.333984375, "learning_rate": 0.00017648338556512086, "loss": 0.8359, "step": 1209 }, { "epoch": 0.6732507998330783, "grad_norm": 0.337890625, "learning_rate": 0.00017644576055254337, "loss": 0.8427, "step": 1210 }, { "epoch": 0.6738072054527751, "grad_norm": 0.33984375, "learning_rate": 0.00017640810948280143, "loss": 0.8537, "step": 1211 }, { "epoch": 0.6743636110724718, "grad_norm": 0.34765625, "learning_rate": 0.00017637043236872865, "loss": 0.924, "step": 1212 }, { "epoch": 0.6749200166921686, "grad_norm": 0.34765625, "learning_rate": 0.00017633272922316757, "loss": 0.9067, "step": 1213 }, { "epoch": 0.6754764223118653, "grad_norm": 0.345703125, "learning_rate": 0.00017629500005896968, "loss": 0.9031, "step": 1214 }, { "epoch": 0.6760328279315622, "grad_norm": 0.328125, "learning_rate": 0.0001762572448889952, "loss": 0.7696, "step": 1215 }, { "epoch": 0.6765892335512589, "grad_norm": 0.30078125, "learning_rate": 0.00017621946372611336, "loss": 0.7062, "step": 1216 }, { "epoch": 0.6771456391709556, "grad_norm": 0.337890625, "learning_rate": 0.00017618165658320214, "loss": 0.8407, "step": 1217 }, { "epoch": 0.6777020447906524, "grad_norm": 0.33203125, "learning_rate": 0.00017614382347314842, "loss": 0.8517, "step": 1218 }, { "epoch": 0.6782584504103492, "grad_norm": 0.33203125, "learning_rate": 0.00017610596440884792, "loss": 0.8495, "step": 1219 }, { "epoch": 0.6788148560300459, "grad_norm": 0.33984375, "learning_rate": 0.00017606807940320524, "loss": 0.8023, "step": 1220 }, { "epoch": 0.6793712616497427, "grad_norm": 0.353515625, "learning_rate": 0.00017603016846913375, "loss": 0.8952, "step": 1221 }, { "epoch": 0.6799276672694394, "grad_norm": 0.337890625, "learning_rate": 0.00017599223161955573, "loss": 0.8634, "step": 1222 }, { "epoch": 0.6804840728891361, "grad_norm": 0.3515625, "learning_rate": 0.0001759542688674023, "loss": 0.9523, "step": 1223 }, { "epoch": 0.681040478508833, "grad_norm": 0.314453125, "learning_rate": 0.0001759162802256133, "loss": 0.7981, "step": 1224 }, { "epoch": 0.6815968841285297, "grad_norm": 0.33984375, "learning_rate": 0.00017587826570713752, "loss": 0.9015, "step": 1225 }, { "epoch": 0.6821532897482264, "grad_norm": 0.33203125, "learning_rate": 0.00017584022532493254, "loss": 0.8965, "step": 1226 }, { "epoch": 0.6827096953679233, "grad_norm": 0.330078125, "learning_rate": 0.00017580215909196467, "loss": 0.8798, "step": 1227 }, { "epoch": 0.68326610098762, "grad_norm": 0.32421875, "learning_rate": 0.00017576406702120916, "loss": 0.8244, "step": 1228 }, { "epoch": 0.6838225066073167, "grad_norm": 0.322265625, "learning_rate": 0.00017572594912564996, "loss": 0.8822, "step": 1229 }, { "epoch": 0.6843789122270135, "grad_norm": 0.328125, "learning_rate": 0.00017568780541827987, "loss": 0.8287, "step": 1230 }, { "epoch": 0.6849353178467102, "grad_norm": 0.33203125, "learning_rate": 0.00017564963591210054, "loss": 0.8288, "step": 1231 }, { "epoch": 0.685491723466407, "grad_norm": 0.3515625, "learning_rate": 0.0001756114406201223, "loss": 0.9075, "step": 1232 }, { "epoch": 0.6860481290861038, "grad_norm": 0.337890625, "learning_rate": 0.00017557321955536436, "loss": 0.9005, "step": 1233 }, { "epoch": 0.6866045347058005, "grad_norm": 0.322265625, "learning_rate": 0.00017553497273085466, "loss": 0.8201, "step": 1234 }, { "epoch": 0.6871609403254972, "grad_norm": 0.328125, "learning_rate": 0.00017549670015962998, "loss": 0.8144, "step": 1235 }, { "epoch": 0.6877173459451941, "grad_norm": 0.337890625, "learning_rate": 0.00017545840185473582, "loss": 0.8734, "step": 1236 }, { "epoch": 0.6882737515648908, "grad_norm": 0.33984375, "learning_rate": 0.00017542007782922646, "loss": 0.8758, "step": 1237 }, { "epoch": 0.6888301571845875, "grad_norm": 0.322265625, "learning_rate": 0.00017538172809616494, "loss": 0.7767, "step": 1238 }, { "epoch": 0.6893865628042843, "grad_norm": 0.314453125, "learning_rate": 0.00017534335266862315, "loss": 0.7504, "step": 1239 }, { "epoch": 0.6899429684239811, "grad_norm": 0.328125, "learning_rate": 0.0001753049515596816, "loss": 0.8237, "step": 1240 }, { "epoch": 0.6904993740436779, "grad_norm": 0.3359375, "learning_rate": 0.00017526652478242966, "loss": 0.7673, "step": 1241 }, { "epoch": 0.6910557796633746, "grad_norm": 0.3203125, "learning_rate": 0.00017522807234996538, "loss": 0.8724, "step": 1242 }, { "epoch": 0.6916121852830713, "grad_norm": 0.326171875, "learning_rate": 0.00017518959427539562, "loss": 0.8233, "step": 1243 }, { "epoch": 0.6921685909027682, "grad_norm": 0.306640625, "learning_rate": 0.00017515109057183596, "loss": 0.6871, "step": 1244 }, { "epoch": 0.6927249965224649, "grad_norm": 0.33203125, "learning_rate": 0.00017511256125241068, "loss": 0.7386, "step": 1245 }, { "epoch": 0.6932814021421616, "grad_norm": 0.337890625, "learning_rate": 0.00017507400633025278, "loss": 0.7724, "step": 1246 }, { "epoch": 0.6938378077618584, "grad_norm": 0.32421875, "learning_rate": 0.00017503542581850408, "loss": 0.81, "step": 1247 }, { "epoch": 0.6943942133815552, "grad_norm": 0.80859375, "learning_rate": 0.00017499681973031503, "loss": 1.8259, "step": 1248 }, { "epoch": 0.6949506190012519, "grad_norm": 0.326171875, "learning_rate": 0.00017495818807884486, "loss": 0.7351, "step": 1249 }, { "epoch": 0.6955070246209487, "grad_norm": 0.3515625, "learning_rate": 0.00017491953087726142, "loss": 0.9045, "step": 1250 }, { "epoch": 0.6960634302406454, "grad_norm": 0.31640625, "learning_rate": 0.00017488084813874143, "loss": 0.6979, "step": 1251 }, { "epoch": 0.6966198358603422, "grad_norm": 0.353515625, "learning_rate": 0.00017484213987647014, "loss": 0.8801, "step": 1252 }, { "epoch": 0.697176241480039, "grad_norm": 0.35546875, "learning_rate": 0.00017480340610364161, "loss": 0.8701, "step": 1253 }, { "epoch": 0.6977326470997357, "grad_norm": 0.341796875, "learning_rate": 0.0001747646468334586, "loss": 0.8198, "step": 1254 }, { "epoch": 0.6982890527194324, "grad_norm": 0.3359375, "learning_rate": 0.00017472586207913245, "loss": 0.8907, "step": 1255 }, { "epoch": 0.6988454583391293, "grad_norm": 0.345703125, "learning_rate": 0.00017468705185388333, "loss": 0.8712, "step": 1256 }, { "epoch": 0.699401863958826, "grad_norm": 0.365234375, "learning_rate": 0.00017464821617093997, "loss": 1.0164, "step": 1257 }, { "epoch": 0.6999582695785227, "grad_norm": 0.349609375, "learning_rate": 0.00017460935504353984, "loss": 0.9493, "step": 1258 }, { "epoch": 0.7005146751982195, "grad_norm": 0.322265625, "learning_rate": 0.0001745704684849291, "loss": 0.7055, "step": 1259 }, { "epoch": 0.7010710808179162, "grad_norm": 0.33203125, "learning_rate": 0.00017453155650836255, "loss": 0.8317, "step": 1260 }, { "epoch": 0.701627486437613, "grad_norm": 0.326171875, "learning_rate": 0.00017449261912710364, "loss": 0.8239, "step": 1261 }, { "epoch": 0.7021838920573098, "grad_norm": 0.318359375, "learning_rate": 0.0001744536563544245, "loss": 0.771, "step": 1262 }, { "epoch": 0.7027402976770065, "grad_norm": 0.345703125, "learning_rate": 0.00017441466820360596, "loss": 0.9036, "step": 1263 }, { "epoch": 0.7032967032967034, "grad_norm": 0.34765625, "learning_rate": 0.00017437565468793732, "loss": 0.7954, "step": 1264 }, { "epoch": 0.7038531089164001, "grad_norm": 0.333984375, "learning_rate": 0.00017433661582071677, "loss": 0.8681, "step": 1265 }, { "epoch": 0.7044095145360968, "grad_norm": 0.33984375, "learning_rate": 0.00017429755161525105, "loss": 0.8394, "step": 1266 }, { "epoch": 0.7049659201557936, "grad_norm": 0.337890625, "learning_rate": 0.00017425846208485543, "loss": 0.8418, "step": 1267 }, { "epoch": 0.7055223257754903, "grad_norm": 0.396484375, "learning_rate": 0.00017421934724285394, "loss": 0.9183, "step": 1268 }, { "epoch": 0.7060787313951871, "grad_norm": 0.337890625, "learning_rate": 0.00017418020710257916, "loss": 0.8866, "step": 1269 }, { "epoch": 0.7066351370148839, "grad_norm": 0.32421875, "learning_rate": 0.00017414104167737238, "loss": 0.8109, "step": 1270 }, { "epoch": 0.7071915426345806, "grad_norm": 0.33984375, "learning_rate": 0.0001741018509805834, "loss": 0.922, "step": 1271 }, { "epoch": 0.7077479482542773, "grad_norm": 0.34375, "learning_rate": 0.00017406263502557073, "loss": 0.9158, "step": 1272 }, { "epoch": 0.7083043538739742, "grad_norm": 0.337890625, "learning_rate": 0.00017402339382570147, "loss": 0.8446, "step": 1273 }, { "epoch": 0.7088607594936709, "grad_norm": 0.33984375, "learning_rate": 0.00017398412739435125, "loss": 0.881, "step": 1274 }, { "epoch": 0.7094171651133676, "grad_norm": 0.337890625, "learning_rate": 0.00017394483574490437, "loss": 0.9032, "step": 1275 }, { "epoch": 0.7099735707330644, "grad_norm": 0.333984375, "learning_rate": 0.00017390551889075374, "loss": 0.8782, "step": 1276 }, { "epoch": 0.7105299763527612, "grad_norm": 0.33203125, "learning_rate": 0.00017386617684530083, "loss": 0.8744, "step": 1277 }, { "epoch": 0.7110863819724579, "grad_norm": 0.330078125, "learning_rate": 0.00017382680962195566, "loss": 0.8058, "step": 1278 }, { "epoch": 0.7116427875921547, "grad_norm": 0.314453125, "learning_rate": 0.0001737874172341369, "loss": 0.6973, "step": 1279 }, { "epoch": 0.7121991932118514, "grad_norm": 0.34375, "learning_rate": 0.00017374799969527174, "loss": 0.8964, "step": 1280 }, { "epoch": 0.7127555988315482, "grad_norm": 0.3203125, "learning_rate": 0.000173708557018796, "loss": 0.8244, "step": 1281 }, { "epoch": 0.713312004451245, "grad_norm": 0.337890625, "learning_rate": 0.00017366908921815404, "loss": 0.8864, "step": 1282 }, { "epoch": 0.7138684100709417, "grad_norm": 0.328125, "learning_rate": 0.0001736295963067988, "loss": 0.8349, "step": 1283 }, { "epoch": 0.7144248156906384, "grad_norm": 0.34375, "learning_rate": 0.0001735900782981917, "loss": 0.8759, "step": 1284 }, { "epoch": 0.7149812213103353, "grad_norm": 0.306640625, "learning_rate": 0.00017355053520580284, "loss": 0.7247, "step": 1285 }, { "epoch": 0.715537626930032, "grad_norm": 0.3203125, "learning_rate": 0.0001735109670431108, "loss": 0.8187, "step": 1286 }, { "epoch": 0.7160940325497287, "grad_norm": 0.328125, "learning_rate": 0.00017347137382360268, "loss": 0.8287, "step": 1287 }, { "epoch": 0.7166504381694255, "grad_norm": 0.333984375, "learning_rate": 0.00017343175556077417, "loss": 0.8427, "step": 1288 }, { "epoch": 0.7172068437891223, "grad_norm": 0.318359375, "learning_rate": 0.00017339211226812947, "loss": 0.7638, "step": 1289 }, { "epoch": 0.7177632494088191, "grad_norm": 0.341796875, "learning_rate": 0.00017335244395918136, "loss": 0.8827, "step": 1290 }, { "epoch": 0.7183196550285158, "grad_norm": 0.337890625, "learning_rate": 0.00017331275064745105, "loss": 0.925, "step": 1291 }, { "epoch": 0.7188760606482125, "grad_norm": 0.310546875, "learning_rate": 0.00017327303234646838, "loss": 0.7816, "step": 1292 }, { "epoch": 0.7194324662679094, "grad_norm": 0.33984375, "learning_rate": 0.00017323328906977166, "loss": 0.9068, "step": 1293 }, { "epoch": 0.7199888718876061, "grad_norm": 0.322265625, "learning_rate": 0.00017319352083090765, "loss": 0.8168, "step": 1294 }, { "epoch": 0.7205452775073028, "grad_norm": 0.326171875, "learning_rate": 0.00017315372764343173, "loss": 0.8039, "step": 1295 }, { "epoch": 0.7211016831269996, "grad_norm": 0.318359375, "learning_rate": 0.00017311390952090775, "loss": 0.7918, "step": 1296 }, { "epoch": 0.7216580887466963, "grad_norm": 0.330078125, "learning_rate": 0.000173074066476908, "loss": 0.8398, "step": 1297 }, { "epoch": 0.7222144943663931, "grad_norm": 0.337890625, "learning_rate": 0.00017303419852501332, "loss": 0.7868, "step": 1298 }, { "epoch": 0.7227708999860899, "grad_norm": 0.318359375, "learning_rate": 0.00017299430567881306, "loss": 0.7343, "step": 1299 }, { "epoch": 0.7233273056057866, "grad_norm": 0.318359375, "learning_rate": 0.000172954387951905, "loss": 0.8459, "step": 1300 }, { "epoch": 0.7238837112254833, "grad_norm": 0.341796875, "learning_rate": 0.0001729144453578954, "loss": 0.8439, "step": 1301 }, { "epoch": 0.7244401168451802, "grad_norm": 0.33984375, "learning_rate": 0.00017287447791039905, "loss": 0.8974, "step": 1302 }, { "epoch": 0.7249965224648769, "grad_norm": 0.333984375, "learning_rate": 0.00017283448562303922, "loss": 0.8415, "step": 1303 }, { "epoch": 0.7255529280845736, "grad_norm": 0.310546875, "learning_rate": 0.00017279446850944753, "loss": 0.8327, "step": 1304 }, { "epoch": 0.7261093337042704, "grad_norm": 0.35546875, "learning_rate": 0.0001727544265832642, "loss": 0.9265, "step": 1305 }, { "epoch": 0.7266657393239672, "grad_norm": 0.33203125, "learning_rate": 0.0001727143598581378, "loss": 0.8761, "step": 1306 }, { "epoch": 0.7272221449436639, "grad_norm": 0.3359375, "learning_rate": 0.00017267426834772546, "loss": 0.8675, "step": 1307 }, { "epoch": 0.7277785505633607, "grad_norm": 0.33203125, "learning_rate": 0.00017263415206569268, "loss": 0.8685, "step": 1308 }, { "epoch": 0.7283349561830574, "grad_norm": 0.326171875, "learning_rate": 0.00017259401102571336, "loss": 0.882, "step": 1309 }, { "epoch": 0.7288913618027542, "grad_norm": 0.314453125, "learning_rate": 0.00017255384524146998, "loss": 0.7924, "step": 1310 }, { "epoch": 0.729447767422451, "grad_norm": 0.337890625, "learning_rate": 0.0001725136547266534, "loss": 0.8117, "step": 1311 }, { "epoch": 0.7300041730421477, "grad_norm": 0.326171875, "learning_rate": 0.0001724734394949628, "loss": 0.8178, "step": 1312 }, { "epoch": 0.7305605786618445, "grad_norm": 0.310546875, "learning_rate": 0.00017243319956010587, "loss": 0.7766, "step": 1313 }, { "epoch": 0.7311169842815413, "grad_norm": 0.328125, "learning_rate": 0.00017239293493579884, "loss": 0.8458, "step": 1314 }, { "epoch": 0.731673389901238, "grad_norm": 0.328125, "learning_rate": 0.00017235264563576612, "loss": 0.8981, "step": 1315 }, { "epoch": 0.7322297955209348, "grad_norm": 0.310546875, "learning_rate": 0.0001723123316737407, "loss": 0.7909, "step": 1316 }, { "epoch": 0.7327862011406315, "grad_norm": 0.33203125, "learning_rate": 0.00017227199306346392, "loss": 0.9362, "step": 1317 }, { "epoch": 0.7333426067603283, "grad_norm": 0.326171875, "learning_rate": 0.0001722316298186855, "loss": 0.8683, "step": 1318 }, { "epoch": 0.7338990123800251, "grad_norm": 0.333984375, "learning_rate": 0.00017219124195316362, "loss": 0.8612, "step": 1319 }, { "epoch": 0.7344554179997218, "grad_norm": 0.322265625, "learning_rate": 0.0001721508294806648, "loss": 0.862, "step": 1320 }, { "epoch": 0.7350118236194185, "grad_norm": 0.341796875, "learning_rate": 0.00017211039241496394, "loss": 0.8873, "step": 1321 }, { "epoch": 0.7355682292391154, "grad_norm": 0.328125, "learning_rate": 0.0001720699307698444, "loss": 0.9028, "step": 1322 }, { "epoch": 0.7361246348588121, "grad_norm": 0.3203125, "learning_rate": 0.0001720294445590978, "loss": 0.8096, "step": 1323 }, { "epoch": 0.7366810404785088, "grad_norm": 0.328125, "learning_rate": 0.00017198893379652424, "loss": 0.8278, "step": 1324 }, { "epoch": 0.7372374460982056, "grad_norm": 0.333984375, "learning_rate": 0.00017194839849593214, "loss": 0.8649, "step": 1325 }, { "epoch": 0.7377938517179023, "grad_norm": 0.326171875, "learning_rate": 0.00017190783867113827, "loss": 0.8515, "step": 1326 }, { "epoch": 0.7383502573375991, "grad_norm": 0.31640625, "learning_rate": 0.00017186725433596779, "loss": 0.7985, "step": 1327 }, { "epoch": 0.7389066629572959, "grad_norm": 0.328125, "learning_rate": 0.00017182664550425422, "loss": 0.7828, "step": 1328 }, { "epoch": 0.7394630685769926, "grad_norm": 0.326171875, "learning_rate": 0.00017178601218983941, "loss": 0.7932, "step": 1329 }, { "epoch": 0.7400194741966893, "grad_norm": 0.341796875, "learning_rate": 0.0001717453544065735, "loss": 0.8869, "step": 1330 }, { "epoch": 0.7405758798163862, "grad_norm": 0.33203125, "learning_rate": 0.00017170467216831513, "loss": 0.8317, "step": 1331 }, { "epoch": 0.7411322854360829, "grad_norm": 0.3125, "learning_rate": 0.00017166396548893106, "loss": 0.7774, "step": 1332 }, { "epoch": 0.7416886910557796, "grad_norm": 0.357421875, "learning_rate": 0.00017162323438229662, "loss": 0.8914, "step": 1333 }, { "epoch": 0.7422450966754764, "grad_norm": 0.322265625, "learning_rate": 0.00017158247886229523, "loss": 0.8156, "step": 1334 }, { "epoch": 0.7428015022951732, "grad_norm": 0.3359375, "learning_rate": 0.0001715416989428188, "loss": 0.829, "step": 1335 }, { "epoch": 0.7433579079148699, "grad_norm": 0.345703125, "learning_rate": 0.00017150089463776754, "loss": 0.8551, "step": 1336 }, { "epoch": 0.7439143135345667, "grad_norm": 0.328125, "learning_rate": 0.00017146006596104982, "loss": 0.8098, "step": 1337 }, { "epoch": 0.7444707191542634, "grad_norm": 0.306640625, "learning_rate": 0.00017141921292658253, "loss": 0.702, "step": 1338 }, { "epoch": 0.7450271247739603, "grad_norm": 0.314453125, "learning_rate": 0.0001713783355482907, "loss": 0.805, "step": 1339 }, { "epoch": 0.745583530393657, "grad_norm": 0.31640625, "learning_rate": 0.00017133743384010778, "loss": 0.7702, "step": 1340 }, { "epoch": 0.7461399360133537, "grad_norm": 0.330078125, "learning_rate": 0.0001712965078159754, "loss": 0.8625, "step": 1341 }, { "epoch": 0.7466963416330505, "grad_norm": 0.3125, "learning_rate": 0.00017125555748984353, "loss": 0.7743, "step": 1342 }, { "epoch": 0.7472527472527473, "grad_norm": 0.326171875, "learning_rate": 0.00017121458287567044, "loss": 0.8225, "step": 1343 }, { "epoch": 0.747809152872444, "grad_norm": 0.353515625, "learning_rate": 0.00017117358398742268, "loss": 0.8579, "step": 1344 }, { "epoch": 0.7483655584921408, "grad_norm": 0.328125, "learning_rate": 0.00017113256083907505, "loss": 0.8478, "step": 1345 }, { "epoch": 0.7489219641118375, "grad_norm": 0.33203125, "learning_rate": 0.00017109151344461062, "loss": 0.8209, "step": 1346 }, { "epoch": 0.7494783697315343, "grad_norm": 0.318359375, "learning_rate": 0.0001710504418180207, "loss": 0.7909, "step": 1347 }, { "epoch": 0.7500347753512311, "grad_norm": 0.33203125, "learning_rate": 0.00017100934597330495, "loss": 0.8893, "step": 1348 }, { "epoch": 0.7505911809709278, "grad_norm": 0.33203125, "learning_rate": 0.0001709682259244712, "loss": 0.7629, "step": 1349 }, { "epoch": 0.7511475865906245, "grad_norm": 0.328125, "learning_rate": 0.00017092708168553556, "loss": 0.8391, "step": 1350 }, { "epoch": 0.7517039922103214, "grad_norm": 0.333984375, "learning_rate": 0.00017088591327052237, "loss": 0.8319, "step": 1351 }, { "epoch": 0.7522603978300181, "grad_norm": 0.33203125, "learning_rate": 0.0001708447206934642, "loss": 0.8358, "step": 1352 }, { "epoch": 0.7528168034497148, "grad_norm": 0.326171875, "learning_rate": 0.00017080350396840192, "loss": 0.8513, "step": 1353 }, { "epoch": 0.7533732090694116, "grad_norm": 0.326171875, "learning_rate": 0.0001707622631093846, "loss": 0.8373, "step": 1354 }, { "epoch": 0.7539296146891084, "grad_norm": 0.3359375, "learning_rate": 0.00017072099813046944, "loss": 0.874, "step": 1355 }, { "epoch": 0.7544860203088051, "grad_norm": 0.31640625, "learning_rate": 0.00017067970904572206, "loss": 0.7982, "step": 1356 }, { "epoch": 0.7550424259285019, "grad_norm": 0.3359375, "learning_rate": 0.0001706383958692161, "loss": 0.8553, "step": 1357 }, { "epoch": 0.7555988315481986, "grad_norm": 0.33203125, "learning_rate": 0.00017059705861503351, "loss": 0.8443, "step": 1358 }, { "epoch": 0.7561552371678953, "grad_norm": 0.3203125, "learning_rate": 0.0001705556972972645, "loss": 0.8124, "step": 1359 }, { "epoch": 0.7567116427875922, "grad_norm": 0.369140625, "learning_rate": 0.0001705143119300073, "loss": 0.9886, "step": 1360 }, { "epoch": 0.7572680484072889, "grad_norm": 0.33203125, "learning_rate": 0.00017047290252736853, "loss": 0.8099, "step": 1361 }, { "epoch": 0.7578244540269857, "grad_norm": 0.322265625, "learning_rate": 0.00017043146910346294, "loss": 0.767, "step": 1362 }, { "epoch": 0.7583808596466824, "grad_norm": 0.33203125, "learning_rate": 0.00017039001167241336, "loss": 0.9191, "step": 1363 }, { "epoch": 0.7589372652663792, "grad_norm": 0.333984375, "learning_rate": 0.000170348530248351, "loss": 0.8611, "step": 1364 }, { "epoch": 0.759493670886076, "grad_norm": 0.306640625, "learning_rate": 0.00017030702484541505, "loss": 0.7678, "step": 1365 }, { "epoch": 0.7600500765057727, "grad_norm": 0.3359375, "learning_rate": 0.00017026549547775302, "loss": 0.8689, "step": 1366 }, { "epoch": 0.7606064821254694, "grad_norm": 0.333984375, "learning_rate": 0.0001702239421595206, "loss": 0.8887, "step": 1367 }, { "epoch": 0.7611628877451663, "grad_norm": 0.3203125, "learning_rate": 0.00017018236490488144, "loss": 0.8028, "step": 1368 }, { "epoch": 0.761719293364863, "grad_norm": 0.322265625, "learning_rate": 0.00017014076372800761, "loss": 0.8483, "step": 1369 }, { "epoch": 0.7622756989845597, "grad_norm": 0.310546875, "learning_rate": 0.0001700991386430791, "loss": 0.776, "step": 1370 }, { "epoch": 0.7628321046042565, "grad_norm": 0.330078125, "learning_rate": 0.0001700574896642843, "loss": 0.9116, "step": 1371 }, { "epoch": 0.7633885102239533, "grad_norm": 0.3125, "learning_rate": 0.00017001581680581944, "loss": 0.7911, "step": 1372 }, { "epoch": 0.76394491584365, "grad_norm": 0.328125, "learning_rate": 0.00016997412008188918, "loss": 0.828, "step": 1373 }, { "epoch": 0.7645013214633468, "grad_norm": 0.30859375, "learning_rate": 0.00016993239950670614, "loss": 0.8273, "step": 1374 }, { "epoch": 0.7650577270830435, "grad_norm": 0.3203125, "learning_rate": 0.00016989065509449115, "loss": 0.789, "step": 1375 }, { "epoch": 0.7656141327027403, "grad_norm": 0.333984375, "learning_rate": 0.00016984888685947312, "loss": 0.8335, "step": 1376 }, { "epoch": 0.7661705383224371, "grad_norm": 0.341796875, "learning_rate": 0.0001698070948158891, "loss": 0.9027, "step": 1377 }, { "epoch": 0.7667269439421338, "grad_norm": 0.921875, "learning_rate": 0.00016976527897798424, "loss": 1.8603, "step": 1378 }, { "epoch": 0.7672833495618305, "grad_norm": 0.330078125, "learning_rate": 0.00016972343936001182, "loss": 0.8506, "step": 1379 }, { "epoch": 0.7678397551815274, "grad_norm": 0.326171875, "learning_rate": 0.0001696815759762332, "loss": 0.8211, "step": 1380 }, { "epoch": 0.7683961608012241, "grad_norm": 0.326171875, "learning_rate": 0.0001696396888409179, "loss": 0.8601, "step": 1381 }, { "epoch": 0.7689525664209208, "grad_norm": 0.33203125, "learning_rate": 0.00016959777796834343, "loss": 0.8136, "step": 1382 }, { "epoch": 0.7695089720406176, "grad_norm": 0.322265625, "learning_rate": 0.0001695558433727955, "loss": 0.8032, "step": 1383 }, { "epoch": 0.7700653776603144, "grad_norm": 0.35546875, "learning_rate": 0.00016951388506856785, "loss": 0.9557, "step": 1384 }, { "epoch": 0.7706217832800111, "grad_norm": 0.359375, "learning_rate": 0.00016947190306996235, "loss": 0.9532, "step": 1385 }, { "epoch": 0.7711781888997079, "grad_norm": 0.333984375, "learning_rate": 0.00016942989739128885, "loss": 0.9016, "step": 1386 }, { "epoch": 0.7717345945194046, "grad_norm": 0.310546875, "learning_rate": 0.0001693878680468653, "loss": 0.7211, "step": 1387 }, { "epoch": 0.7722910001391015, "grad_norm": 0.330078125, "learning_rate": 0.00016934581505101786, "loss": 0.8251, "step": 1388 }, { "epoch": 0.7728474057587982, "grad_norm": 0.333984375, "learning_rate": 0.00016930373841808055, "loss": 0.9217, "step": 1389 }, { "epoch": 0.7734038113784949, "grad_norm": 0.314453125, "learning_rate": 0.00016926163816239557, "loss": 0.798, "step": 1390 }, { "epoch": 0.7739602169981917, "grad_norm": 0.361328125, "learning_rate": 0.0001692195142983131, "loss": 0.9148, "step": 1391 }, { "epoch": 0.7745166226178885, "grad_norm": 0.32421875, "learning_rate": 0.00016917736684019148, "loss": 0.7434, "step": 1392 }, { "epoch": 0.7750730282375852, "grad_norm": 0.318359375, "learning_rate": 0.00016913519580239697, "loss": 0.7526, "step": 1393 }, { "epoch": 0.775629433857282, "grad_norm": 0.32421875, "learning_rate": 0.00016909300119930388, "loss": 0.7844, "step": 1394 }, { "epoch": 0.7761858394769787, "grad_norm": 0.326171875, "learning_rate": 0.00016905078304529464, "loss": 0.8256, "step": 1395 }, { "epoch": 0.7767422450966754, "grad_norm": 0.337890625, "learning_rate": 0.00016900854135475966, "loss": 0.8872, "step": 1396 }, { "epoch": 0.7772986507163723, "grad_norm": 0.328125, "learning_rate": 0.0001689662761420973, "loss": 0.8541, "step": 1397 }, { "epoch": 0.777855056336069, "grad_norm": 0.330078125, "learning_rate": 0.00016892398742171408, "loss": 0.8455, "step": 1398 }, { "epoch": 0.7784114619557657, "grad_norm": 0.328125, "learning_rate": 0.0001688816752080244, "loss": 0.9086, "step": 1399 }, { "epoch": 0.7789678675754625, "grad_norm": 0.310546875, "learning_rate": 0.00016883933951545075, "loss": 0.7954, "step": 1400 }, { "epoch": 0.7795242731951593, "grad_norm": 0.353515625, "learning_rate": 0.0001687969803584236, "loss": 0.8604, "step": 1401 }, { "epoch": 0.780080678814856, "grad_norm": 0.30859375, "learning_rate": 0.00016875459775138142, "loss": 0.7879, "step": 1402 }, { "epoch": 0.7806370844345528, "grad_norm": 0.30859375, "learning_rate": 0.00016871219170877066, "loss": 0.6737, "step": 1403 }, { "epoch": 0.7811934900542495, "grad_norm": 0.326171875, "learning_rate": 0.00016866976224504576, "loss": 0.8641, "step": 1404 }, { "epoch": 0.7817498956739463, "grad_norm": 0.3125, "learning_rate": 0.00016862730937466916, "loss": 0.7594, "step": 1405 }, { "epoch": 0.7823063012936431, "grad_norm": 0.306640625, "learning_rate": 0.0001685848331121113, "loss": 0.7111, "step": 1406 }, { "epoch": 0.7828627069133398, "grad_norm": 0.326171875, "learning_rate": 0.00016854233347185052, "loss": 0.7913, "step": 1407 }, { "epoch": 0.7834191125330365, "grad_norm": 0.330078125, "learning_rate": 0.00016849981046837317, "loss": 0.8784, "step": 1408 }, { "epoch": 0.7839755181527334, "grad_norm": 0.31640625, "learning_rate": 0.0001684572641161736, "loss": 0.8471, "step": 1409 }, { "epoch": 0.7845319237724301, "grad_norm": 0.322265625, "learning_rate": 0.00016841469442975405, "loss": 0.8297, "step": 1410 }, { "epoch": 0.7850883293921269, "grad_norm": 0.3359375, "learning_rate": 0.00016837210142362473, "loss": 0.907, "step": 1411 }, { "epoch": 0.7856447350118236, "grad_norm": 0.31640625, "learning_rate": 0.00016832948511230386, "loss": 0.7822, "step": 1412 }, { "epoch": 0.7862011406315204, "grad_norm": 0.328125, "learning_rate": 0.00016828684551031757, "loss": 0.8785, "step": 1413 }, { "epoch": 0.7867575462512172, "grad_norm": 0.33984375, "learning_rate": 0.00016824418263219987, "loss": 0.9351, "step": 1414 }, { "epoch": 0.7873139518709139, "grad_norm": 0.33984375, "learning_rate": 0.00016820149649249277, "loss": 0.8879, "step": 1415 }, { "epoch": 0.7878703574906106, "grad_norm": 0.33203125, "learning_rate": 0.00016815878710574618, "loss": 0.8392, "step": 1416 }, { "epoch": 0.7884267631103075, "grad_norm": 0.328125, "learning_rate": 0.000168116054486518, "loss": 0.812, "step": 1417 }, { "epoch": 0.7889831687300042, "grad_norm": 0.345703125, "learning_rate": 0.00016807329864937392, "loss": 0.8525, "step": 1418 }, { "epoch": 0.7895395743497009, "grad_norm": 0.341796875, "learning_rate": 0.00016803051960888766, "loss": 0.9073, "step": 1419 }, { "epoch": 0.7900959799693977, "grad_norm": 0.3203125, "learning_rate": 0.00016798771737964077, "loss": 0.7898, "step": 1420 }, { "epoch": 0.7906523855890945, "grad_norm": 0.33203125, "learning_rate": 0.00016794489197622279, "loss": 0.8647, "step": 1421 }, { "epoch": 0.7912087912087912, "grad_norm": 0.322265625, "learning_rate": 0.0001679020434132311, "loss": 0.8098, "step": 1422 }, { "epoch": 0.791765196828488, "grad_norm": 0.33984375, "learning_rate": 0.00016785917170527095, "loss": 0.8732, "step": 1423 }, { "epoch": 0.7923216024481847, "grad_norm": 0.353515625, "learning_rate": 0.00016781627686695552, "loss": 0.8421, "step": 1424 }, { "epoch": 0.7928780080678814, "grad_norm": 0.3359375, "learning_rate": 0.0001677733589129059, "loss": 0.9202, "step": 1425 }, { "epoch": 0.7934344136875783, "grad_norm": 0.337890625, "learning_rate": 0.000167730417857751, "loss": 0.8072, "step": 1426 }, { "epoch": 0.793990819307275, "grad_norm": 0.33203125, "learning_rate": 0.00016768745371612765, "loss": 0.8307, "step": 1427 }, { "epoch": 0.7945472249269717, "grad_norm": 0.326171875, "learning_rate": 0.00016764446650268053, "loss": 0.8375, "step": 1428 }, { "epoch": 0.7951036305466685, "grad_norm": 0.9453125, "learning_rate": 0.0001676014562320622, "loss": 1.8024, "step": 1429 }, { "epoch": 0.7956600361663653, "grad_norm": 0.30859375, "learning_rate": 0.000167558422918933, "loss": 0.6899, "step": 1430 }, { "epoch": 0.796216441786062, "grad_norm": 0.328125, "learning_rate": 0.00016751536657796123, "loss": 0.8481, "step": 1431 }, { "epoch": 0.7967728474057588, "grad_norm": 0.333984375, "learning_rate": 0.00016747228722382298, "loss": 0.8259, "step": 1432 }, { "epoch": 0.7973292530254555, "grad_norm": 0.333984375, "learning_rate": 0.00016742918487120222, "loss": 0.8527, "step": 1433 }, { "epoch": 0.7978856586451523, "grad_norm": 0.306640625, "learning_rate": 0.00016738605953479075, "loss": 0.6968, "step": 1434 }, { "epoch": 0.7984420642648491, "grad_norm": 0.353515625, "learning_rate": 0.00016734291122928816, "loss": 0.9613, "step": 1435 }, { "epoch": 0.7989984698845458, "grad_norm": 0.333984375, "learning_rate": 0.00016729973996940193, "loss": 0.8458, "step": 1436 }, { "epoch": 0.7995548755042426, "grad_norm": 0.326171875, "learning_rate": 0.0001672565457698473, "loss": 0.8313, "step": 1437 }, { "epoch": 0.8001112811239394, "grad_norm": 0.318359375, "learning_rate": 0.00016721332864534744, "loss": 0.7054, "step": 1438 }, { "epoch": 0.8006676867436361, "grad_norm": 0.3359375, "learning_rate": 0.00016717008861063322, "loss": 0.9309, "step": 1439 }, { "epoch": 0.8012240923633329, "grad_norm": 0.3046875, "learning_rate": 0.00016712682568044334, "loss": 0.7426, "step": 1440 }, { "epoch": 0.8017804979830296, "grad_norm": 0.333984375, "learning_rate": 0.00016708353986952433, "loss": 0.8588, "step": 1441 }, { "epoch": 0.8023369036027264, "grad_norm": 0.34765625, "learning_rate": 0.00016704023119263058, "loss": 0.9493, "step": 1442 }, { "epoch": 0.8028933092224232, "grad_norm": 0.326171875, "learning_rate": 0.0001669968996645241, "loss": 0.838, "step": 1443 }, { "epoch": 0.8034497148421199, "grad_norm": 0.310546875, "learning_rate": 0.00016695354529997486, "loss": 0.7359, "step": 1444 }, { "epoch": 0.8040061204618166, "grad_norm": 0.353515625, "learning_rate": 0.00016691016811376058, "loss": 0.905, "step": 1445 }, { "epoch": 0.8045625260815135, "grad_norm": 0.322265625, "learning_rate": 0.00016686676812066668, "loss": 0.8376, "step": 1446 }, { "epoch": 0.8051189317012102, "grad_norm": 0.302734375, "learning_rate": 0.00016682334533548642, "loss": 0.7095, "step": 1447 }, { "epoch": 0.8056753373209069, "grad_norm": 0.337890625, "learning_rate": 0.00016677989977302078, "loss": 0.8286, "step": 1448 }, { "epoch": 0.8062317429406037, "grad_norm": 0.333984375, "learning_rate": 0.0001667364314480786, "loss": 0.8728, "step": 1449 }, { "epoch": 0.8067881485603005, "grad_norm": 0.337890625, "learning_rate": 0.00016669294037547635, "loss": 0.8136, "step": 1450 }, { "epoch": 0.8073445541799972, "grad_norm": 0.353515625, "learning_rate": 0.00016664942657003838, "loss": 0.8813, "step": 1451 }, { "epoch": 0.807900959799694, "grad_norm": 0.310546875, "learning_rate": 0.00016660589004659663, "loss": 0.7414, "step": 1452 }, { "epoch": 0.8084573654193907, "grad_norm": 0.3203125, "learning_rate": 0.000166562330819991, "loss": 0.8392, "step": 1453 }, { "epoch": 0.8090137710390874, "grad_norm": 0.337890625, "learning_rate": 0.00016651874890506897, "loss": 0.8908, "step": 1454 }, { "epoch": 0.8095701766587843, "grad_norm": 0.33984375, "learning_rate": 0.0001664751443166858, "loss": 0.8836, "step": 1455 }, { "epoch": 0.810126582278481, "grad_norm": 0.328125, "learning_rate": 0.0001664315170697044, "loss": 0.8345, "step": 1456 }, { "epoch": 0.8106829878981777, "grad_norm": 0.322265625, "learning_rate": 0.00016638786717899551, "loss": 0.7562, "step": 1457 }, { "epoch": 0.8112393935178746, "grad_norm": 0.333984375, "learning_rate": 0.0001663441946594376, "loss": 0.8809, "step": 1458 }, { "epoch": 0.8117957991375713, "grad_norm": 0.333984375, "learning_rate": 0.0001663004995259168, "loss": 0.8177, "step": 1459 }, { "epoch": 0.8123522047572681, "grad_norm": 0.3203125, "learning_rate": 0.00016625678179332692, "loss": 0.8005, "step": 1460 }, { "epoch": 0.8129086103769648, "grad_norm": 0.32421875, "learning_rate": 0.00016621304147656948, "loss": 0.8027, "step": 1461 }, { "epoch": 0.8134650159966615, "grad_norm": 0.3203125, "learning_rate": 0.0001661692785905538, "loss": 0.8325, "step": 1462 }, { "epoch": 0.8140214216163584, "grad_norm": 0.32421875, "learning_rate": 0.0001661254931501968, "loss": 0.7392, "step": 1463 }, { "epoch": 0.8145778272360551, "grad_norm": 0.326171875, "learning_rate": 0.00016608168517042305, "loss": 0.819, "step": 1464 }, { "epoch": 0.8151342328557518, "grad_norm": 0.32421875, "learning_rate": 0.00016603785466616492, "loss": 0.7979, "step": 1465 }, { "epoch": 0.8156906384754486, "grad_norm": 0.3125, "learning_rate": 0.0001659940016523624, "loss": 0.7334, "step": 1466 }, { "epoch": 0.8162470440951454, "grad_norm": 0.30859375, "learning_rate": 0.0001659501261439631, "loss": 0.6871, "step": 1467 }, { "epoch": 0.8168034497148421, "grad_norm": 0.3359375, "learning_rate": 0.00016590622815592236, "loss": 0.902, "step": 1468 }, { "epoch": 0.8173598553345389, "grad_norm": 0.337890625, "learning_rate": 0.0001658623077032032, "loss": 0.896, "step": 1469 }, { "epoch": 0.8179162609542356, "grad_norm": 0.328125, "learning_rate": 0.00016581836480077623, "loss": 0.8791, "step": 1470 }, { "epoch": 0.8184726665739324, "grad_norm": 0.431640625, "learning_rate": 0.00016577439946361978, "loss": 0.9967, "step": 1471 }, { "epoch": 0.8190290721936292, "grad_norm": 0.328125, "learning_rate": 0.0001657304117067198, "loss": 0.8007, "step": 1472 }, { "epoch": 0.8195854778133259, "grad_norm": 0.357421875, "learning_rate": 0.00016568640154506983, "loss": 0.9419, "step": 1473 }, { "epoch": 0.8201418834330226, "grad_norm": 0.328125, "learning_rate": 0.00016564236899367115, "loss": 0.8375, "step": 1474 }, { "epoch": 0.8206982890527195, "grad_norm": 0.322265625, "learning_rate": 0.00016559831406753257, "loss": 0.7735, "step": 1475 }, { "epoch": 0.8212546946724162, "grad_norm": 0.3203125, "learning_rate": 0.00016555423678167063, "loss": 0.7933, "step": 1476 }, { "epoch": 0.8218111002921129, "grad_norm": 0.318359375, "learning_rate": 0.00016551013715110932, "loss": 0.8085, "step": 1477 }, { "epoch": 0.8223675059118097, "grad_norm": 0.3125, "learning_rate": 0.00016546601519088053, "loss": 0.7552, "step": 1478 }, { "epoch": 0.8229239115315065, "grad_norm": 0.3203125, "learning_rate": 0.00016542187091602345, "loss": 0.7744, "step": 1479 }, { "epoch": 0.8234803171512032, "grad_norm": 0.322265625, "learning_rate": 0.0001653777043415851, "loss": 0.8328, "step": 1480 }, { "epoch": 0.8240367227709, "grad_norm": 0.349609375, "learning_rate": 0.00016533351548261998, "loss": 0.8444, "step": 1481 }, { "epoch": 0.8245931283905967, "grad_norm": 0.337890625, "learning_rate": 0.0001652893043541902, "loss": 0.8248, "step": 1482 }, { "epoch": 0.8251495340102935, "grad_norm": 0.330078125, "learning_rate": 0.00016524507097136554, "loss": 0.8688, "step": 1483 }, { "epoch": 0.8257059396299903, "grad_norm": 0.322265625, "learning_rate": 0.0001652008153492233, "loss": 0.8503, "step": 1484 }, { "epoch": 0.826262345249687, "grad_norm": 0.3125, "learning_rate": 0.00016515653750284837, "loss": 0.8137, "step": 1485 }, { "epoch": 0.8268187508693838, "grad_norm": 0.306640625, "learning_rate": 0.0001651122374473332, "loss": 0.7943, "step": 1486 }, { "epoch": 0.8273751564890806, "grad_norm": 0.318359375, "learning_rate": 0.00016506791519777786, "loss": 0.7787, "step": 1487 }, { "epoch": 0.8279315621087773, "grad_norm": 0.333984375, "learning_rate": 0.00016502357076928989, "loss": 0.856, "step": 1488 }, { "epoch": 0.8284879677284741, "grad_norm": 0.314453125, "learning_rate": 0.0001649792041769845, "loss": 0.7854, "step": 1489 }, { "epoch": 0.8290443733481708, "grad_norm": 0.3203125, "learning_rate": 0.00016493481543598443, "loss": 0.7728, "step": 1490 }, { "epoch": 0.8296007789678675, "grad_norm": 0.318359375, "learning_rate": 0.00016489040456141988, "loss": 0.7262, "step": 1491 }, { "epoch": 0.8301571845875644, "grad_norm": 0.3125, "learning_rate": 0.0001648459715684287, "loss": 0.721, "step": 1492 }, { "epoch": 0.8307135902072611, "grad_norm": 0.359375, "learning_rate": 0.00016480151647215626, "loss": 0.8713, "step": 1493 }, { "epoch": 0.8312699958269578, "grad_norm": 0.3359375, "learning_rate": 0.00016475703928775542, "loss": 0.8769, "step": 1494 }, { "epoch": 0.8318264014466547, "grad_norm": 0.326171875, "learning_rate": 0.00016471254003038656, "loss": 0.8117, "step": 1495 }, { "epoch": 0.8323828070663514, "grad_norm": 0.326171875, "learning_rate": 0.0001646680187152177, "loss": 0.8064, "step": 1496 }, { "epoch": 0.8329392126860481, "grad_norm": 0.3125, "learning_rate": 0.00016462347535742419, "loss": 0.783, "step": 1497 }, { "epoch": 0.8334956183057449, "grad_norm": 0.33203125, "learning_rate": 0.00016457890997218907, "loss": 0.8462, "step": 1498 }, { "epoch": 0.8340520239254416, "grad_norm": 0.32421875, "learning_rate": 0.00016453432257470282, "loss": 0.8125, "step": 1499 }, { "epoch": 0.8346084295451384, "grad_norm": 0.3203125, "learning_rate": 0.0001644897131801634, "loss": 0.812, "step": 1500 }, { "epoch": 0.8351648351648352, "grad_norm": 0.310546875, "learning_rate": 0.00016444508180377628, "loss": 0.722, "step": 1501 }, { "epoch": 0.8357212407845319, "grad_norm": 0.318359375, "learning_rate": 0.00016440042846075446, "loss": 0.8175, "step": 1502 }, { "epoch": 0.8362776464042286, "grad_norm": 0.33984375, "learning_rate": 0.00016435575316631833, "loss": 0.9485, "step": 1503 }, { "epoch": 0.8368340520239255, "grad_norm": 0.3125, "learning_rate": 0.0001643110559356959, "loss": 0.7406, "step": 1504 }, { "epoch": 0.8373904576436222, "grad_norm": 0.330078125, "learning_rate": 0.00016426633678412263, "loss": 0.7846, "step": 1505 }, { "epoch": 0.8379468632633189, "grad_norm": 0.32421875, "learning_rate": 0.00016422159572684127, "loss": 0.8702, "step": 1506 }, { "epoch": 0.8385032688830157, "grad_norm": 0.3359375, "learning_rate": 0.00016417683277910228, "loss": 0.8541, "step": 1507 }, { "epoch": 0.8390596745027125, "grad_norm": 0.32421875, "learning_rate": 0.00016413204795616345, "loss": 0.8826, "step": 1508 }, { "epoch": 0.8396160801224093, "grad_norm": 0.31640625, "learning_rate": 0.00016408724127329005, "loss": 0.7689, "step": 1509 }, { "epoch": 0.840172485742106, "grad_norm": 0.337890625, "learning_rate": 0.0001640424127457548, "loss": 0.9144, "step": 1510 }, { "epoch": 0.8407288913618027, "grad_norm": 0.33203125, "learning_rate": 0.0001639975623888379, "loss": 0.8451, "step": 1511 }, { "epoch": 0.8412852969814996, "grad_norm": 0.33203125, "learning_rate": 0.00016395269021782696, "loss": 0.8054, "step": 1512 }, { "epoch": 0.8418417026011963, "grad_norm": 0.33984375, "learning_rate": 0.00016390779624801699, "loss": 0.8637, "step": 1513 }, { "epoch": 0.842398108220893, "grad_norm": 0.333984375, "learning_rate": 0.00016386288049471048, "loss": 0.8536, "step": 1514 }, { "epoch": 0.8429545138405898, "grad_norm": 0.3359375, "learning_rate": 0.00016381794297321735, "loss": 0.8799, "step": 1515 }, { "epoch": 0.8435109194602866, "grad_norm": 0.330078125, "learning_rate": 0.0001637729836988549, "loss": 0.814, "step": 1516 }, { "epoch": 0.8440673250799833, "grad_norm": 0.31640625, "learning_rate": 0.00016372800268694792, "loss": 0.7759, "step": 1517 }, { "epoch": 0.8446237306996801, "grad_norm": 0.33984375, "learning_rate": 0.00016368299995282848, "loss": 0.851, "step": 1518 }, { "epoch": 0.8451801363193768, "grad_norm": 0.349609375, "learning_rate": 0.0001636379755118362, "loss": 0.8151, "step": 1519 }, { "epoch": 0.8457365419390735, "grad_norm": 0.318359375, "learning_rate": 0.00016359292937931797, "loss": 0.7879, "step": 1520 }, { "epoch": 0.8462929475587704, "grad_norm": 0.326171875, "learning_rate": 0.00016354786157062818, "loss": 0.825, "step": 1521 }, { "epoch": 0.8468493531784671, "grad_norm": 0.31640625, "learning_rate": 0.00016350277210112855, "loss": 0.7414, "step": 1522 }, { "epoch": 0.8474057587981638, "grad_norm": 0.31640625, "learning_rate": 0.0001634576609861882, "loss": 0.7679, "step": 1523 }, { "epoch": 0.8479621644178607, "grad_norm": 0.322265625, "learning_rate": 0.00016341252824118358, "loss": 0.8039, "step": 1524 }, { "epoch": 0.8485185700375574, "grad_norm": 0.306640625, "learning_rate": 0.00016336737388149863, "loss": 0.6974, "step": 1525 }, { "epoch": 0.8490749756572541, "grad_norm": 0.33203125, "learning_rate": 0.0001633221979225245, "loss": 0.852, "step": 1526 }, { "epoch": 0.8496313812769509, "grad_norm": 0.39453125, "learning_rate": 0.00016327700037965987, "loss": 0.9108, "step": 1527 }, { "epoch": 0.8501877868966476, "grad_norm": 0.35546875, "learning_rate": 0.00016323178126831066, "loss": 0.8568, "step": 1528 }, { "epoch": 0.8507441925163444, "grad_norm": 0.3359375, "learning_rate": 0.00016318654060389016, "loss": 0.8606, "step": 1529 }, { "epoch": 0.8513005981360412, "grad_norm": 0.314453125, "learning_rate": 0.00016314127840181902, "loss": 0.7187, "step": 1530 }, { "epoch": 0.8518570037557379, "grad_norm": 0.3203125, "learning_rate": 0.00016309599467752522, "loss": 0.8346, "step": 1531 }, { "epoch": 0.8524134093754346, "grad_norm": 0.326171875, "learning_rate": 0.00016305068944644417, "loss": 0.9214, "step": 1532 }, { "epoch": 0.8529698149951315, "grad_norm": 0.3203125, "learning_rate": 0.00016300536272401844, "loss": 0.7717, "step": 1533 }, { "epoch": 0.8535262206148282, "grad_norm": 0.330078125, "learning_rate": 0.00016296001452569808, "loss": 0.8721, "step": 1534 }, { "epoch": 0.854082626234525, "grad_norm": 0.3359375, "learning_rate": 0.0001629146448669403, "loss": 0.8538, "step": 1535 }, { "epoch": 0.8546390318542217, "grad_norm": 0.345703125, "learning_rate": 0.00016286925376320986, "loss": 0.9144, "step": 1536 }, { "epoch": 0.8551954374739185, "grad_norm": 0.3359375, "learning_rate": 0.00016282384122997856, "loss": 0.8563, "step": 1537 }, { "epoch": 0.8557518430936153, "grad_norm": 0.32421875, "learning_rate": 0.0001627784072827257, "loss": 0.8297, "step": 1538 }, { "epoch": 0.856308248713312, "grad_norm": 0.310546875, "learning_rate": 0.0001627329519369378, "loss": 0.7608, "step": 1539 }, { "epoch": 0.8568646543330087, "grad_norm": 0.328125, "learning_rate": 0.00016268747520810869, "loss": 0.8589, "step": 1540 }, { "epoch": 0.8574210599527056, "grad_norm": 0.318359375, "learning_rate": 0.00016264197711173946, "loss": 0.8031, "step": 1541 }, { "epoch": 0.8579774655724023, "grad_norm": 0.310546875, "learning_rate": 0.00016259645766333858, "loss": 0.7716, "step": 1542 }, { "epoch": 0.858533871192099, "grad_norm": 0.30859375, "learning_rate": 0.00016255091687842165, "loss": 0.7499, "step": 1543 }, { "epoch": 0.8590902768117958, "grad_norm": 0.337890625, "learning_rate": 0.00016250535477251168, "loss": 0.8882, "step": 1544 }, { "epoch": 0.8596466824314926, "grad_norm": 0.322265625, "learning_rate": 0.00016245977136113883, "loss": 0.7705, "step": 1545 }, { "epoch": 0.8602030880511893, "grad_norm": 0.333984375, "learning_rate": 0.00016241416665984063, "loss": 0.8502, "step": 1546 }, { "epoch": 0.8607594936708861, "grad_norm": 0.314453125, "learning_rate": 0.00016236854068416178, "loss": 0.774, "step": 1547 }, { "epoch": 0.8613158992905828, "grad_norm": 0.349609375, "learning_rate": 0.00016232289344965428, "loss": 0.9005, "step": 1548 }, { "epoch": 0.8618723049102796, "grad_norm": 0.333984375, "learning_rate": 0.0001622772249718774, "loss": 0.8641, "step": 1549 }, { "epoch": 0.8624287105299764, "grad_norm": 0.349609375, "learning_rate": 0.00016223153526639754, "loss": 0.9873, "step": 1550 }, { "epoch": 0.8629851161496731, "grad_norm": 0.333984375, "learning_rate": 0.00016218582434878847, "loss": 0.8727, "step": 1551 }, { "epoch": 0.8635415217693698, "grad_norm": 0.3203125, "learning_rate": 0.0001621400922346311, "loss": 0.802, "step": 1552 }, { "epoch": 0.8640979273890667, "grad_norm": 0.3203125, "learning_rate": 0.00016209433893951366, "loss": 0.8282, "step": 1553 }, { "epoch": 0.8646543330087634, "grad_norm": 0.333984375, "learning_rate": 0.00016204856447903145, "loss": 0.9093, "step": 1554 }, { "epoch": 0.8652107386284601, "grad_norm": 0.345703125, "learning_rate": 0.0001620027688687871, "loss": 0.9119, "step": 1555 }, { "epoch": 0.8657671442481569, "grad_norm": 0.326171875, "learning_rate": 0.00016195695212439044, "loss": 0.8263, "step": 1556 }, { "epoch": 0.8663235498678536, "grad_norm": 0.357421875, "learning_rate": 0.00016191111426145846, "loss": 0.9783, "step": 1557 }, { "epoch": 0.8668799554875504, "grad_norm": 0.330078125, "learning_rate": 0.00016186525529561537, "loss": 0.8366, "step": 1558 }, { "epoch": 0.8674363611072472, "grad_norm": 0.318359375, "learning_rate": 0.00016181937524249263, "loss": 0.7817, "step": 1559 }, { "epoch": 0.8679927667269439, "grad_norm": 0.328125, "learning_rate": 0.00016177347411772876, "loss": 0.7799, "step": 1560 }, { "epoch": 0.8685491723466408, "grad_norm": 0.33203125, "learning_rate": 0.00016172755193696956, "loss": 0.8473, "step": 1561 }, { "epoch": 0.8691055779663375, "grad_norm": 0.33203125, "learning_rate": 0.000161681608715868, "loss": 0.8397, "step": 1562 }, { "epoch": 0.8696619835860342, "grad_norm": 0.330078125, "learning_rate": 0.00016163564447008416, "loss": 0.8296, "step": 1563 }, { "epoch": 0.870218389205731, "grad_norm": 0.32421875, "learning_rate": 0.00016158965921528537, "loss": 0.7982, "step": 1564 }, { "epoch": 0.8707747948254277, "grad_norm": 0.310546875, "learning_rate": 0.0001615436529671461, "loss": 0.7385, "step": 1565 }, { "epoch": 0.8713312004451245, "grad_norm": 0.330078125, "learning_rate": 0.0001614976257413479, "loss": 0.8126, "step": 1566 }, { "epoch": 0.8718876060648213, "grad_norm": 0.306640625, "learning_rate": 0.00016145157755357957, "loss": 0.756, "step": 1567 }, { "epoch": 0.872444011684518, "grad_norm": 0.326171875, "learning_rate": 0.00016140550841953698, "loss": 0.8072, "step": 1568 }, { "epoch": 0.8730004173042147, "grad_norm": 0.3515625, "learning_rate": 0.00016135941835492318, "loss": 0.9268, "step": 1569 }, { "epoch": 0.8735568229239116, "grad_norm": 0.330078125, "learning_rate": 0.0001613133073754484, "loss": 0.8406, "step": 1570 }, { "epoch": 0.8741132285436083, "grad_norm": 0.341796875, "learning_rate": 0.0001612671754968299, "loss": 0.8901, "step": 1571 }, { "epoch": 0.874669634163305, "grad_norm": 0.33984375, "learning_rate": 0.00016122102273479207, "loss": 0.8701, "step": 1572 }, { "epoch": 0.8752260397830018, "grad_norm": 0.322265625, "learning_rate": 0.00016117484910506652, "loss": 0.7697, "step": 1573 }, { "epoch": 0.8757824454026986, "grad_norm": 0.33984375, "learning_rate": 0.0001611286546233919, "loss": 0.9444, "step": 1574 }, { "epoch": 0.8763388510223953, "grad_norm": 0.34375, "learning_rate": 0.00016108243930551392, "loss": 0.8555, "step": 1575 }, { "epoch": 0.8768952566420921, "grad_norm": 0.3203125, "learning_rate": 0.00016103620316718553, "loss": 0.7459, "step": 1576 }, { "epoch": 0.8774516622617888, "grad_norm": 0.322265625, "learning_rate": 0.00016098994622416665, "loss": 0.8395, "step": 1577 }, { "epoch": 0.8780080678814856, "grad_norm": 0.3125, "learning_rate": 0.00016094366849222437, "loss": 0.799, "step": 1578 }, { "epoch": 0.8785644735011824, "grad_norm": 0.337890625, "learning_rate": 0.0001608973699871328, "loss": 0.9022, "step": 1579 }, { "epoch": 0.8791208791208791, "grad_norm": 0.32421875, "learning_rate": 0.00016085105072467318, "loss": 0.7485, "step": 1580 }, { "epoch": 0.8796772847405758, "grad_norm": 0.330078125, "learning_rate": 0.00016080471072063386, "loss": 0.8096, "step": 1581 }, { "epoch": 0.8802336903602727, "grad_norm": 0.328125, "learning_rate": 0.00016075834999081012, "loss": 0.8698, "step": 1582 }, { "epoch": 0.8807900959799694, "grad_norm": 0.314453125, "learning_rate": 0.00016071196855100444, "loss": 0.7957, "step": 1583 }, { "epoch": 0.8813465015996662, "grad_norm": 0.328125, "learning_rate": 0.0001606655664170263, "loss": 0.8163, "step": 1584 }, { "epoch": 0.8819029072193629, "grad_norm": 0.328125, "learning_rate": 0.00016061914360469227, "loss": 0.8946, "step": 1585 }, { "epoch": 0.8824593128390597, "grad_norm": 0.328125, "learning_rate": 0.00016057270012982596, "loss": 0.8084, "step": 1586 }, { "epoch": 0.8830157184587565, "grad_norm": 0.341796875, "learning_rate": 0.00016052623600825797, "loss": 0.8569, "step": 1587 }, { "epoch": 0.8835721240784532, "grad_norm": 0.326171875, "learning_rate": 0.000160479751255826, "loss": 0.771, "step": 1588 }, { "epoch": 0.8841285296981499, "grad_norm": 0.3125, "learning_rate": 0.00016043324588837474, "loss": 0.7227, "step": 1589 }, { "epoch": 0.8846849353178468, "grad_norm": 0.322265625, "learning_rate": 0.00016038671992175594, "loss": 0.8204, "step": 1590 }, { "epoch": 0.8852413409375435, "grad_norm": 0.333984375, "learning_rate": 0.00016034017337182834, "loss": 0.9365, "step": 1591 }, { "epoch": 0.8857977465572402, "grad_norm": 0.328125, "learning_rate": 0.00016029360625445773, "loss": 0.7505, "step": 1592 }, { "epoch": 0.886354152176937, "grad_norm": 0.31640625, "learning_rate": 0.00016024701858551694, "loss": 0.8051, "step": 1593 }, { "epoch": 0.8869105577966337, "grad_norm": 0.333984375, "learning_rate": 0.00016020041038088568, "loss": 0.8519, "step": 1594 }, { "epoch": 0.8874669634163305, "grad_norm": 0.3359375, "learning_rate": 0.00016015378165645076, "loss": 0.9041, "step": 1595 }, { "epoch": 0.8880233690360273, "grad_norm": 0.32421875, "learning_rate": 0.000160107132428106, "loss": 0.7955, "step": 1596 }, { "epoch": 0.888579774655724, "grad_norm": 0.3125, "learning_rate": 0.0001600604627117521, "loss": 0.7547, "step": 1597 }, { "epoch": 0.8891361802754207, "grad_norm": 0.33203125, "learning_rate": 0.0001600137725232969, "loss": 0.8384, "step": 1598 }, { "epoch": 0.8896925858951176, "grad_norm": 0.322265625, "learning_rate": 0.0001599670618786551, "loss": 0.7619, "step": 1599 }, { "epoch": 0.8902489915148143, "grad_norm": 0.328125, "learning_rate": 0.00015992033079374838, "loss": 0.8609, "step": 1600 }, { "epoch": 0.890805397134511, "grad_norm": 0.34765625, "learning_rate": 0.00015987357928450542, "loss": 0.9125, "step": 1601 }, { "epoch": 0.8913618027542078, "grad_norm": 0.333984375, "learning_rate": 0.0001598268073668619, "loss": 0.8257, "step": 1602 }, { "epoch": 0.8919182083739046, "grad_norm": 0.32421875, "learning_rate": 0.00015978001505676035, "loss": 0.766, "step": 1603 }, { "epoch": 0.8924746139936013, "grad_norm": 0.33984375, "learning_rate": 0.0001597332023701504, "loss": 0.8824, "step": 1604 }, { "epoch": 0.8930310196132981, "grad_norm": 0.33203125, "learning_rate": 0.00015968636932298843, "loss": 0.8271, "step": 1605 }, { "epoch": 0.8935874252329948, "grad_norm": 0.32421875, "learning_rate": 0.00015963951593123797, "loss": 0.8136, "step": 1606 }, { "epoch": 0.8941438308526916, "grad_norm": 0.31640625, "learning_rate": 0.0001595926422108693, "loss": 0.7455, "step": 1607 }, { "epoch": 0.8947002364723884, "grad_norm": 0.328125, "learning_rate": 0.00015954574817785973, "loss": 0.8478, "step": 1608 }, { "epoch": 0.8952566420920851, "grad_norm": 0.3203125, "learning_rate": 0.00015949883384819353, "loss": 0.7687, "step": 1609 }, { "epoch": 0.8958130477117819, "grad_norm": 0.3359375, "learning_rate": 0.00015945189923786176, "loss": 0.8229, "step": 1610 }, { "epoch": 0.8963694533314787, "grad_norm": 0.3359375, "learning_rate": 0.00015940494436286252, "loss": 0.8361, "step": 1611 }, { "epoch": 0.8969258589511754, "grad_norm": 0.333984375, "learning_rate": 0.00015935796923920078, "loss": 0.7942, "step": 1612 }, { "epoch": 0.8974822645708722, "grad_norm": 0.337890625, "learning_rate": 0.0001593109738828883, "loss": 0.79, "step": 1613 }, { "epoch": 0.8980386701905689, "grad_norm": 0.33203125, "learning_rate": 0.00015926395830994394, "loss": 0.8377, "step": 1614 }, { "epoch": 0.8985950758102657, "grad_norm": 0.3203125, "learning_rate": 0.00015921692253639328, "loss": 0.796, "step": 1615 }, { "epoch": 0.8991514814299625, "grad_norm": 0.330078125, "learning_rate": 0.00015916986657826885, "loss": 0.7916, "step": 1616 }, { "epoch": 0.8997078870496592, "grad_norm": 0.3125, "learning_rate": 0.0001591227904516101, "loss": 0.7697, "step": 1617 }, { "epoch": 0.9002642926693559, "grad_norm": 0.32421875, "learning_rate": 0.00015907569417246333, "loss": 0.8299, "step": 1618 }, { "epoch": 0.9008206982890528, "grad_norm": 0.33984375, "learning_rate": 0.0001590285777568816, "loss": 0.7568, "step": 1619 }, { "epoch": 0.9013771039087495, "grad_norm": 0.333984375, "learning_rate": 0.00015898144122092505, "loss": 0.8744, "step": 1620 }, { "epoch": 0.9019335095284462, "grad_norm": 0.3125, "learning_rate": 0.00015893428458066042, "loss": 0.7819, "step": 1621 }, { "epoch": 0.902489915148143, "grad_norm": 0.328125, "learning_rate": 0.00015888710785216153, "loss": 0.8183, "step": 1622 }, { "epoch": 0.9030463207678397, "grad_norm": 0.322265625, "learning_rate": 0.0001588399110515089, "loss": 0.7904, "step": 1623 }, { "epoch": 0.9036027263875365, "grad_norm": 0.3359375, "learning_rate": 0.00015879269419479002, "loss": 0.8461, "step": 1624 }, { "epoch": 0.9041591320072333, "grad_norm": 0.33984375, "learning_rate": 0.00015874545729809906, "loss": 0.8309, "step": 1625 }, { "epoch": 0.90471553762693, "grad_norm": 0.310546875, "learning_rate": 0.00015869820037753716, "loss": 0.6568, "step": 1626 }, { "epoch": 0.9052719432466267, "grad_norm": 0.333984375, "learning_rate": 0.0001586509234492122, "loss": 0.8832, "step": 1627 }, { "epoch": 0.9058283488663236, "grad_norm": 0.328125, "learning_rate": 0.00015860362652923894, "loss": 0.7721, "step": 1628 }, { "epoch": 0.9063847544860203, "grad_norm": 0.328125, "learning_rate": 0.0001585563096337389, "loss": 0.8272, "step": 1629 }, { "epoch": 0.906941160105717, "grad_norm": 0.3359375, "learning_rate": 0.00015850897277884044, "loss": 0.9139, "step": 1630 }, { "epoch": 0.9074975657254138, "grad_norm": 0.322265625, "learning_rate": 0.0001584616159806787, "loss": 0.8265, "step": 1631 }, { "epoch": 0.9080539713451106, "grad_norm": 0.3125, "learning_rate": 0.00015841423925539564, "loss": 0.7456, "step": 1632 }, { "epoch": 0.9086103769648074, "grad_norm": 0.3125, "learning_rate": 0.00015836684261914, "loss": 0.7247, "step": 1633 }, { "epoch": 0.9091667825845041, "grad_norm": 0.3203125, "learning_rate": 0.00015831942608806734, "loss": 0.7684, "step": 1634 }, { "epoch": 0.9097231882042008, "grad_norm": 0.31640625, "learning_rate": 0.00015827198967833993, "loss": 0.7793, "step": 1635 }, { "epoch": 0.9102795938238977, "grad_norm": 0.33984375, "learning_rate": 0.00015822453340612693, "loss": 0.8197, "step": 1636 }, { "epoch": 0.9108359994435944, "grad_norm": 0.341796875, "learning_rate": 0.0001581770572876041, "loss": 0.8746, "step": 1637 }, { "epoch": 0.9113924050632911, "grad_norm": 0.333984375, "learning_rate": 0.00015812956133895412, "loss": 0.8732, "step": 1638 }, { "epoch": 0.9119488106829879, "grad_norm": 0.33203125, "learning_rate": 0.00015808204557636642, "loss": 0.8501, "step": 1639 }, { "epoch": 0.9125052163026847, "grad_norm": 0.322265625, "learning_rate": 0.00015803451001603706, "loss": 0.7846, "step": 1640 }, { "epoch": 0.9130616219223814, "grad_norm": 0.328125, "learning_rate": 0.00015798695467416897, "loss": 0.871, "step": 1641 }, { "epoch": 0.9136180275420782, "grad_norm": 0.318359375, "learning_rate": 0.00015793937956697175, "loss": 0.7881, "step": 1642 }, { "epoch": 0.9141744331617749, "grad_norm": 0.310546875, "learning_rate": 0.00015789178471066178, "loss": 0.7453, "step": 1643 }, { "epoch": 0.9147308387814717, "grad_norm": 0.32421875, "learning_rate": 0.00015784417012146217, "loss": 0.8814, "step": 1644 }, { "epoch": 0.9152872444011685, "grad_norm": 0.35546875, "learning_rate": 0.0001577965358156027, "loss": 0.8565, "step": 1645 }, { "epoch": 0.9158436500208652, "grad_norm": 0.330078125, "learning_rate": 0.00015774888180931993, "loss": 0.8185, "step": 1646 }, { "epoch": 0.9164000556405619, "grad_norm": 0.345703125, "learning_rate": 0.0001577012081188572, "loss": 0.8507, "step": 1647 }, { "epoch": 0.9169564612602588, "grad_norm": 0.32421875, "learning_rate": 0.00015765351476046435, "loss": 0.721, "step": 1648 }, { "epoch": 0.9175128668799555, "grad_norm": 0.333984375, "learning_rate": 0.00015760580175039812, "loss": 0.8975, "step": 1649 }, { "epoch": 0.9180692724996522, "grad_norm": 0.306640625, "learning_rate": 0.00015755806910492187, "loss": 0.7393, "step": 1650 }, { "epoch": 0.918625678119349, "grad_norm": 0.330078125, "learning_rate": 0.00015751031684030568, "loss": 0.8464, "step": 1651 }, { "epoch": 0.9191820837390458, "grad_norm": 0.32421875, "learning_rate": 0.0001574625449728263, "loss": 0.7609, "step": 1652 }, { "epoch": 0.9197384893587425, "grad_norm": 0.328125, "learning_rate": 0.00015741475351876717, "loss": 0.8405, "step": 1653 }, { "epoch": 0.9202948949784393, "grad_norm": 0.337890625, "learning_rate": 0.00015736694249441832, "loss": 0.8624, "step": 1654 }, { "epoch": 0.920851300598136, "grad_norm": 0.3203125, "learning_rate": 0.00015731911191607661, "loss": 0.7953, "step": 1655 }, { "epoch": 0.9214077062178327, "grad_norm": 0.318359375, "learning_rate": 0.0001572712618000455, "loss": 0.7722, "step": 1656 }, { "epoch": 0.9219641118375296, "grad_norm": 0.380859375, "learning_rate": 0.00015722339216263503, "loss": 0.8173, "step": 1657 }, { "epoch": 0.9225205174572263, "grad_norm": 0.314453125, "learning_rate": 0.00015717550302016197, "loss": 0.7558, "step": 1658 }, { "epoch": 0.9230769230769231, "grad_norm": 0.33203125, "learning_rate": 0.00015712759438894978, "loss": 0.83, "step": 1659 }, { "epoch": 0.9236333286966198, "grad_norm": 0.326171875, "learning_rate": 0.00015707966628532847, "loss": 0.8345, "step": 1660 }, { "epoch": 0.9241897343163166, "grad_norm": 0.322265625, "learning_rate": 0.00015703171872563472, "loss": 0.7731, "step": 1661 }, { "epoch": 0.9247461399360134, "grad_norm": 0.306640625, "learning_rate": 0.00015698375172621188, "loss": 0.7117, "step": 1662 }, { "epoch": 0.9253025455557101, "grad_norm": 0.318359375, "learning_rate": 0.00015693576530340987, "loss": 0.7655, "step": 1663 }, { "epoch": 0.9258589511754068, "grad_norm": 0.3203125, "learning_rate": 0.00015688775947358524, "loss": 0.8557, "step": 1664 }, { "epoch": 0.9264153567951037, "grad_norm": 0.30859375, "learning_rate": 0.0001568397342531012, "loss": 0.7155, "step": 1665 }, { "epoch": 0.9269717624148004, "grad_norm": 0.3203125, "learning_rate": 0.0001567916896583275, "loss": 0.7699, "step": 1666 }, { "epoch": 0.9275281680344971, "grad_norm": 0.34375, "learning_rate": 0.00015674362570564063, "loss": 0.9217, "step": 1667 }, { "epoch": 0.928084573654194, "grad_norm": 0.330078125, "learning_rate": 0.0001566955424114235, "loss": 0.8761, "step": 1668 }, { "epoch": 0.9286409792738907, "grad_norm": 0.330078125, "learning_rate": 0.00015664743979206574, "loss": 0.854, "step": 1669 }, { "epoch": 0.9291973848935874, "grad_norm": 0.361328125, "learning_rate": 0.00015659931786396345, "loss": 0.9302, "step": 1670 }, { "epoch": 0.9297537905132842, "grad_norm": 0.310546875, "learning_rate": 0.00015655117664351945, "loss": 0.6646, "step": 1671 }, { "epoch": 0.9303101961329809, "grad_norm": 0.30078125, "learning_rate": 0.00015650301614714303, "loss": 0.6723, "step": 1672 }, { "epoch": 0.9308666017526777, "grad_norm": 0.333984375, "learning_rate": 0.00015645483639125014, "loss": 0.8624, "step": 1673 }, { "epoch": 0.9314230073723745, "grad_norm": 0.33984375, "learning_rate": 0.0001564066373922632, "loss": 0.8976, "step": 1674 }, { "epoch": 0.9319794129920712, "grad_norm": 0.33203125, "learning_rate": 0.00015635841916661127, "loss": 0.8596, "step": 1675 }, { "epoch": 0.9325358186117679, "grad_norm": 0.341796875, "learning_rate": 0.00015631018173072988, "loss": 0.9422, "step": 1676 }, { "epoch": 0.9330922242314648, "grad_norm": 0.32421875, "learning_rate": 0.00015626192510106123, "loss": 0.7931, "step": 1677 }, { "epoch": 0.9336486298511615, "grad_norm": 0.337890625, "learning_rate": 0.0001562136492940539, "loss": 0.8655, "step": 1678 }, { "epoch": 0.9342050354708582, "grad_norm": 0.298828125, "learning_rate": 0.00015616535432616316, "loss": 0.7388, "step": 1679 }, { "epoch": 0.934761441090555, "grad_norm": 0.3359375, "learning_rate": 0.0001561170402138507, "loss": 0.8805, "step": 1680 }, { "epoch": 0.9353178467102518, "grad_norm": 0.318359375, "learning_rate": 0.00015606870697358483, "loss": 0.7576, "step": 1681 }, { "epoch": 0.9358742523299486, "grad_norm": 0.33984375, "learning_rate": 0.0001560203546218403, "loss": 0.8602, "step": 1682 }, { "epoch": 0.9364306579496453, "grad_norm": 0.330078125, "learning_rate": 0.00015597198317509838, "loss": 0.8555, "step": 1683 }, { "epoch": 0.936987063569342, "grad_norm": 0.328125, "learning_rate": 0.00015592359264984693, "loss": 0.8234, "step": 1684 }, { "epoch": 0.9375434691890389, "grad_norm": 0.326171875, "learning_rate": 0.00015587518306258024, "loss": 0.8587, "step": 1685 }, { "epoch": 0.9380998748087356, "grad_norm": 0.330078125, "learning_rate": 0.0001558267544297991, "loss": 0.7904, "step": 1686 }, { "epoch": 0.9386562804284323, "grad_norm": 0.33984375, "learning_rate": 0.00015577830676801079, "loss": 0.8887, "step": 1687 }, { "epoch": 0.9392126860481291, "grad_norm": 0.341796875, "learning_rate": 0.00015572984009372916, "loss": 0.8488, "step": 1688 }, { "epoch": 0.9397690916678259, "grad_norm": 0.41015625, "learning_rate": 0.0001556813544234744, "loss": 0.7988, "step": 1689 }, { "epoch": 0.9403254972875226, "grad_norm": 0.314453125, "learning_rate": 0.0001556328497737733, "loss": 0.8036, "step": 1690 }, { "epoch": 0.9408819029072194, "grad_norm": 0.330078125, "learning_rate": 0.00015558432616115901, "loss": 0.8355, "step": 1691 }, { "epoch": 0.9414383085269161, "grad_norm": 0.326171875, "learning_rate": 0.00015553578360217126, "loss": 0.8544, "step": 1692 }, { "epoch": 0.9419947141466128, "grad_norm": 0.341796875, "learning_rate": 0.0001554872221133562, "loss": 0.8712, "step": 1693 }, { "epoch": 0.9425511197663097, "grad_norm": 0.306640625, "learning_rate": 0.00015543864171126633, "loss": 0.696, "step": 1694 }, { "epoch": 0.9431075253860064, "grad_norm": 0.32421875, "learning_rate": 0.0001553900424124607, "loss": 0.8413, "step": 1695 }, { "epoch": 0.9436639310057031, "grad_norm": 0.328125, "learning_rate": 0.00015534142423350483, "loss": 0.8712, "step": 1696 }, { "epoch": 0.9442203366254, "grad_norm": 0.31640625, "learning_rate": 0.00015529278719097057, "loss": 0.8315, "step": 1697 }, { "epoch": 0.9447767422450967, "grad_norm": 0.3203125, "learning_rate": 0.00015524413130143632, "loss": 0.8684, "step": 1698 }, { "epoch": 0.9453331478647934, "grad_norm": 0.99609375, "learning_rate": 0.00015519545658148674, "loss": 1.8628, "step": 1699 }, { "epoch": 0.9458895534844902, "grad_norm": 0.322265625, "learning_rate": 0.0001551467630477131, "loss": 0.8537, "step": 1700 }, { "epoch": 0.9464459591041869, "grad_norm": 0.333984375, "learning_rate": 0.00015509805071671292, "loss": 0.8349, "step": 1701 }, { "epoch": 0.9470023647238837, "grad_norm": 0.3203125, "learning_rate": 0.0001550493196050903, "loss": 0.7872, "step": 1702 }, { "epoch": 0.9475587703435805, "grad_norm": 0.34375, "learning_rate": 0.0001550005697294555, "loss": 0.9317, "step": 1703 }, { "epoch": 0.9481151759632772, "grad_norm": 0.337890625, "learning_rate": 0.00015495180110642543, "loss": 0.909, "step": 1704 }, { "epoch": 0.9486715815829739, "grad_norm": 0.330078125, "learning_rate": 0.0001549030137526232, "loss": 0.8417, "step": 1705 }, { "epoch": 0.9492279872026708, "grad_norm": 0.31640625, "learning_rate": 0.00015485420768467843, "loss": 0.7273, "step": 1706 }, { "epoch": 0.9497843928223675, "grad_norm": 0.3203125, "learning_rate": 0.00015480538291922703, "loss": 0.8405, "step": 1707 }, { "epoch": 0.9503407984420643, "grad_norm": 0.3515625, "learning_rate": 0.00015475653947291136, "loss": 0.9213, "step": 1708 }, { "epoch": 0.950897204061761, "grad_norm": 0.326171875, "learning_rate": 0.00015470767736238007, "loss": 0.8166, "step": 1709 }, { "epoch": 0.9514536096814578, "grad_norm": 0.3203125, "learning_rate": 0.0001546587966042883, "loss": 0.8303, "step": 1710 }, { "epoch": 0.9520100153011546, "grad_norm": 0.3359375, "learning_rate": 0.0001546098972152973, "loss": 0.8412, "step": 1711 }, { "epoch": 0.9525664209208513, "grad_norm": 0.310546875, "learning_rate": 0.00015456097921207496, "loss": 0.7055, "step": 1712 }, { "epoch": 0.953122826540548, "grad_norm": 0.361328125, "learning_rate": 0.00015451204261129534, "loss": 0.925, "step": 1713 }, { "epoch": 0.9536792321602449, "grad_norm": 0.353515625, "learning_rate": 0.0001544630874296389, "loss": 0.8671, "step": 1714 }, { "epoch": 0.9542356377799416, "grad_norm": 0.326171875, "learning_rate": 0.00015441411368379235, "loss": 0.7963, "step": 1715 }, { "epoch": 0.9547920433996383, "grad_norm": 0.3125, "learning_rate": 0.00015436512139044886, "loss": 0.7544, "step": 1716 }, { "epoch": 0.9553484490193351, "grad_norm": 0.318359375, "learning_rate": 0.00015431611056630784, "loss": 0.7664, "step": 1717 }, { "epoch": 0.9559048546390319, "grad_norm": 0.328125, "learning_rate": 0.000154267081228075, "loss": 0.8951, "step": 1718 }, { "epoch": 0.9564612602587286, "grad_norm": 0.33203125, "learning_rate": 0.00015421803339246245, "loss": 0.8817, "step": 1719 }, { "epoch": 0.9570176658784254, "grad_norm": 0.326171875, "learning_rate": 0.00015416896707618852, "loss": 0.8008, "step": 1720 }, { "epoch": 0.9575740714981221, "grad_norm": 0.357421875, "learning_rate": 0.00015411988229597782, "loss": 0.9524, "step": 1721 }, { "epoch": 0.9581304771178188, "grad_norm": 0.330078125, "learning_rate": 0.00015407077906856137, "loss": 0.7774, "step": 1722 }, { "epoch": 0.9586868827375157, "grad_norm": 0.318359375, "learning_rate": 0.0001540216574106764, "loss": 0.7962, "step": 1723 }, { "epoch": 0.9592432883572124, "grad_norm": 0.33984375, "learning_rate": 0.00015397251733906636, "loss": 0.8501, "step": 1724 }, { "epoch": 0.9597996939769091, "grad_norm": 0.322265625, "learning_rate": 0.00015392335887048114, "loss": 0.7948, "step": 1725 }, { "epoch": 0.960356099596606, "grad_norm": 0.33984375, "learning_rate": 0.00015387418202167673, "loss": 0.8632, "step": 1726 }, { "epoch": 0.9609125052163027, "grad_norm": 0.35546875, "learning_rate": 0.0001538249868094155, "loss": 0.9704, "step": 1727 }, { "epoch": 0.9614689108359994, "grad_norm": 0.30859375, "learning_rate": 0.00015377577325046606, "loss": 0.7435, "step": 1728 }, { "epoch": 0.9620253164556962, "grad_norm": 0.330078125, "learning_rate": 0.00015372654136160322, "loss": 0.8626, "step": 1729 }, { "epoch": 0.9625817220753929, "grad_norm": 0.30859375, "learning_rate": 0.0001536772911596081, "loss": 0.7785, "step": 1730 }, { "epoch": 0.9631381276950898, "grad_norm": 0.318359375, "learning_rate": 0.000153628022661268, "loss": 0.7472, "step": 1731 }, { "epoch": 0.9636945333147865, "grad_norm": 0.326171875, "learning_rate": 0.0001535787358833765, "loss": 0.854, "step": 1732 }, { "epoch": 0.9642509389344832, "grad_norm": 0.330078125, "learning_rate": 0.00015352943084273344, "loss": 0.8813, "step": 1733 }, { "epoch": 0.96480734455418, "grad_norm": 0.35546875, "learning_rate": 0.0001534801075561448, "loss": 0.9251, "step": 1734 }, { "epoch": 0.9653637501738768, "grad_norm": 0.306640625, "learning_rate": 0.00015343076604042285, "loss": 0.7338, "step": 1735 }, { "epoch": 0.9659201557935735, "grad_norm": 0.32421875, "learning_rate": 0.00015338140631238604, "loss": 0.761, "step": 1736 }, { "epoch": 0.9664765614132703, "grad_norm": 0.33203125, "learning_rate": 0.00015333202838885908, "loss": 0.8448, "step": 1737 }, { "epoch": 0.967032967032967, "grad_norm": 0.32421875, "learning_rate": 0.0001532826322866728, "loss": 0.7478, "step": 1738 }, { "epoch": 0.9675893726526638, "grad_norm": 0.328125, "learning_rate": 0.00015323321802266428, "loss": 0.8116, "step": 1739 }, { "epoch": 0.9681457782723606, "grad_norm": 0.32421875, "learning_rate": 0.00015318378561367672, "loss": 0.8851, "step": 1740 }, { "epoch": 0.9687021838920573, "grad_norm": 0.333984375, "learning_rate": 0.00015313433507655968, "loss": 0.8524, "step": 1741 }, { "epoch": 0.969258589511754, "grad_norm": 0.314453125, "learning_rate": 0.00015308486642816869, "loss": 0.7722, "step": 1742 }, { "epoch": 0.9698149951314509, "grad_norm": 0.33984375, "learning_rate": 0.00015303537968536557, "loss": 0.9203, "step": 1743 }, { "epoch": 0.9703714007511476, "grad_norm": 0.333984375, "learning_rate": 0.00015298587486501828, "loss": 0.8607, "step": 1744 }, { "epoch": 0.9709278063708443, "grad_norm": 0.32421875, "learning_rate": 0.00015293635198400095, "loss": 0.7585, "step": 1745 }, { "epoch": 0.9714842119905411, "grad_norm": 0.318359375, "learning_rate": 0.00015288681105919385, "loss": 0.7475, "step": 1746 }, { "epoch": 0.9720406176102379, "grad_norm": 0.322265625, "learning_rate": 0.0001528372521074834, "loss": 0.7884, "step": 1747 }, { "epoch": 0.9725970232299346, "grad_norm": 0.328125, "learning_rate": 0.00015278767514576223, "loss": 0.7938, "step": 1748 }, { "epoch": 0.9731534288496314, "grad_norm": 0.326171875, "learning_rate": 0.00015273808019092896, "loss": 0.8201, "step": 1749 }, { "epoch": 0.9737098344693281, "grad_norm": 0.3125, "learning_rate": 0.00015268846725988855, "loss": 0.7848, "step": 1750 }, { "epoch": 0.9742662400890248, "grad_norm": 0.328125, "learning_rate": 0.0001526388363695519, "loss": 0.8647, "step": 1751 }, { "epoch": 0.9748226457087217, "grad_norm": 0.34765625, "learning_rate": 0.00015258918753683612, "loss": 0.8837, "step": 1752 }, { "epoch": 0.9753790513284184, "grad_norm": 0.302734375, "learning_rate": 0.00015253952077866437, "loss": 0.72, "step": 1753 }, { "epoch": 0.9759354569481151, "grad_norm": 0.33984375, "learning_rate": 0.0001524898361119661, "loss": 0.9329, "step": 1754 }, { "epoch": 0.976491862567812, "grad_norm": 0.30859375, "learning_rate": 0.00015244013355367663, "loss": 0.7853, "step": 1755 }, { "epoch": 0.9770482681875087, "grad_norm": 0.333984375, "learning_rate": 0.00015239041312073754, "loss": 0.7884, "step": 1756 }, { "epoch": 0.9776046738072055, "grad_norm": 0.328125, "learning_rate": 0.0001523406748300964, "loss": 0.8844, "step": 1757 }, { "epoch": 0.9781610794269022, "grad_norm": 0.34375, "learning_rate": 0.00015229091869870697, "loss": 0.8389, "step": 1758 }, { "epoch": 0.978717485046599, "grad_norm": 0.34375, "learning_rate": 0.00015224114474352897, "loss": 0.9325, "step": 1759 }, { "epoch": 0.9792738906662958, "grad_norm": 0.3359375, "learning_rate": 0.00015219135298152835, "loss": 0.8748, "step": 1760 }, { "epoch": 0.9798302962859925, "grad_norm": 0.3359375, "learning_rate": 0.00015214154342967691, "loss": 0.8733, "step": 1761 }, { "epoch": 0.9803867019056892, "grad_norm": 0.326171875, "learning_rate": 0.00015209171610495277, "loss": 0.8178, "step": 1762 }, { "epoch": 0.980943107525386, "grad_norm": 0.322265625, "learning_rate": 0.0001520418710243399, "loss": 0.7557, "step": 1763 }, { "epoch": 0.9814995131450828, "grad_norm": 0.337890625, "learning_rate": 0.0001519920082048285, "loss": 0.8423, "step": 1764 }, { "epoch": 0.9820559187647795, "grad_norm": 0.333984375, "learning_rate": 0.00015194212766341458, "loss": 0.8604, "step": 1765 }, { "epoch": 0.9826123243844763, "grad_norm": 0.326171875, "learning_rate": 0.00015189222941710043, "loss": 0.849, "step": 1766 }, { "epoch": 0.983168730004173, "grad_norm": 0.322265625, "learning_rate": 0.0001518423134828943, "loss": 0.8249, "step": 1767 }, { "epoch": 0.9837251356238698, "grad_norm": 0.32421875, "learning_rate": 0.00015179237987781037, "loss": 0.8117, "step": 1768 }, { "epoch": 0.9842815412435666, "grad_norm": 0.326171875, "learning_rate": 0.00015174242861886892, "loss": 0.7628, "step": 1769 }, { "epoch": 0.9848379468632633, "grad_norm": 0.298828125, "learning_rate": 0.00015169245972309633, "loss": 0.7333, "step": 1770 }, { "epoch": 0.98539435248296, "grad_norm": 0.333984375, "learning_rate": 0.00015164247320752488, "loss": 0.8816, "step": 1771 }, { "epoch": 0.9859507581026569, "grad_norm": 0.3359375, "learning_rate": 0.0001515924690891928, "loss": 0.814, "step": 1772 }, { "epoch": 0.9865071637223536, "grad_norm": 0.318359375, "learning_rate": 0.00015154244738514448, "loss": 0.7076, "step": 1773 }, { "epoch": 0.9870635693420503, "grad_norm": 0.326171875, "learning_rate": 0.00015149240811243022, "loss": 0.8046, "step": 1774 }, { "epoch": 0.9876199749617471, "grad_norm": 0.318359375, "learning_rate": 0.0001514423512881063, "loss": 0.7358, "step": 1775 }, { "epoch": 0.9881763805814439, "grad_norm": 0.314453125, "learning_rate": 0.000151392276929235, "loss": 0.7509, "step": 1776 }, { "epoch": 0.9887327862011406, "grad_norm": 0.3125, "learning_rate": 0.00015134218505288458, "loss": 0.719, "step": 1777 }, { "epoch": 0.9892891918208374, "grad_norm": 0.365234375, "learning_rate": 0.00015129207567612932, "loss": 0.945, "step": 1778 }, { "epoch": 0.9898455974405341, "grad_norm": 0.328125, "learning_rate": 0.00015124194881604929, "loss": 0.8144, "step": 1779 }, { "epoch": 0.990402003060231, "grad_norm": 0.330078125, "learning_rate": 0.00015119180448973073, "loss": 0.7926, "step": 1780 }, { "epoch": 0.9909584086799277, "grad_norm": 0.337890625, "learning_rate": 0.0001511416427142658, "loss": 0.9045, "step": 1781 }, { "epoch": 0.9915148142996244, "grad_norm": 0.318359375, "learning_rate": 0.00015109146350675236, "loss": 0.7689, "step": 1782 }, { "epoch": 0.9920712199193212, "grad_norm": 0.33203125, "learning_rate": 0.00015104126688429462, "loss": 0.8457, "step": 1783 }, { "epoch": 0.992627625539018, "grad_norm": 0.341796875, "learning_rate": 0.00015099105286400238, "loss": 0.8154, "step": 1784 }, { "epoch": 0.9931840311587147, "grad_norm": 0.326171875, "learning_rate": 0.00015094082146299153, "loss": 0.7393, "step": 1785 }, { "epoch": 0.9937404367784115, "grad_norm": 0.306640625, "learning_rate": 0.0001508905726983839, "loss": 0.7191, "step": 1786 }, { "epoch": 0.9942968423981082, "grad_norm": 0.314453125, "learning_rate": 0.0001508403065873071, "loss": 0.7611, "step": 1787 }, { "epoch": 0.994853248017805, "grad_norm": 0.330078125, "learning_rate": 0.00015079002314689484, "loss": 0.7972, "step": 1788 }, { "epoch": 0.9954096536375018, "grad_norm": 0.3203125, "learning_rate": 0.0001507397223942866, "loss": 0.7863, "step": 1789 }, { "epoch": 0.9959660592571985, "grad_norm": 0.318359375, "learning_rate": 0.00015068940434662776, "loss": 0.801, "step": 1790 }, { "epoch": 0.9965224648768952, "grad_norm": 0.3046875, "learning_rate": 0.00015063906902106975, "loss": 0.7198, "step": 1791 }, { "epoch": 0.997078870496592, "grad_norm": 0.302734375, "learning_rate": 0.0001505887164347697, "loss": 0.6595, "step": 1792 }, { "epoch": 0.9976352761162888, "grad_norm": 0.318359375, "learning_rate": 0.00015053834660489073, "loss": 0.8524, "step": 1793 }, { "epoch": 0.9981916817359855, "grad_norm": 0.310546875, "learning_rate": 0.00015048795954860172, "loss": 0.8229, "step": 1794 }, { "epoch": 0.9987480873556823, "grad_norm": 0.310546875, "learning_rate": 0.00015043755528307767, "loss": 0.8114, "step": 1795 }, { "epoch": 0.999304492975379, "grad_norm": 0.328125, "learning_rate": 0.0001503871338254992, "loss": 0.868, "step": 1796 }, { "epoch": 0.9998608985950758, "grad_norm": 0.3203125, "learning_rate": 0.00015033669519305286, "loss": 0.771, "step": 1797 }, { "epoch": 1.0004173042147726, "grad_norm": 0.318359375, "learning_rate": 0.0001502862394029311, "loss": 0.7104, "step": 1798 }, { "epoch": 1.0004173042147726, "eval_loss": 0.7584885954856873, "eval_runtime": 1763.8079, "eval_samples_per_second": 2.66, "eval_steps_per_second": 1.33, "step": 1798 } ], "logging_steps": 1, "max_steps": 5391, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 899, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 7.49249300155859e+17, "train_batch_size": 2, "trial_name": null, "trial_params": null }