{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.9995724668661823, "eval_steps": 500, "global_step": 877, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.003420265070542967, "grad_norm": 10.156329759741077, "learning_rate": 0.0, "loss": 2.1575, "step": 1 }, { "epoch": 0.006840530141085934, "grad_norm": 184.3295302770049, "learning_rate": 1.41e-05, "loss": 4.88, "step": 2 }, { "epoch": 0.010260795211628902, "grad_norm": 182.65124398096702, "learning_rate": 1.41e-05, "loss": 4.8994, "step": 3 }, { "epoch": 0.013681060282171868, "grad_norm": 30.32918049437796, "learning_rate": 1.41e-05, "loss": 3.5419, "step": 4 }, { "epoch": 0.017101325352714837, "grad_norm": 16.444831424361816, "learning_rate": 1.41e-05, "loss": 2.7311, "step": 5 }, { "epoch": 0.020521590423257803, "grad_norm": 6.633419850116959, "learning_rate": 1.41e-05, "loss": 2.2268, "step": 6 }, { "epoch": 0.02394185549380077, "grad_norm": 9.401679436345786, "learning_rate": 1.41e-05, "loss": 2.1353, "step": 7 }, { "epoch": 0.027362120564343735, "grad_norm": 7.95097945213086, "learning_rate": 1.41e-05, "loss": 2.0867, "step": 8 }, { "epoch": 0.030782385634886705, "grad_norm": 3.9617869212350962, "learning_rate": 1.41e-05, "loss": 1.9823, "step": 9 }, { "epoch": 0.034202650705429674, "grad_norm": 9.402637091305385, "learning_rate": 1.41e-05, "loss": 1.9361, "step": 10 }, { "epoch": 0.03762291577597264, "grad_norm": 3.7983366071849862, "learning_rate": 1.41e-05, "loss": 1.9295, "step": 11 }, { "epoch": 0.041043180846515606, "grad_norm": 4.558710902554969, "learning_rate": 1.41e-05, "loss": 1.8787, "step": 12 }, { "epoch": 0.04446344591705857, "grad_norm": 3.9168701968301134, "learning_rate": 1.41e-05, "loss": 1.7936, "step": 13 }, { "epoch": 0.04788371098760154, "grad_norm": 3.763989925966716, "learning_rate": 1.41e-05, "loss": 1.7617, "step": 14 }, { "epoch": 0.051303976058144504, "grad_norm": 3.4821708929604007, "learning_rate": 1.41e-05, "loss": 1.6881, "step": 15 }, { "epoch": 0.05472424112868747, "grad_norm": 2.1426267148325997, "learning_rate": 1.41e-05, "loss": 1.6905, "step": 16 }, { "epoch": 0.05814450619923044, "grad_norm": 3.6919608929938255, "learning_rate": 1.41e-05, "loss": 1.6473, "step": 17 }, { "epoch": 0.06156477126977341, "grad_norm": 2.725651684433171, "learning_rate": 1.41e-05, "loss": 1.7064, "step": 18 }, { "epoch": 0.06498503634031637, "grad_norm": 2.9127756737882824, "learning_rate": 1.41e-05, "loss": 1.5901, "step": 19 }, { "epoch": 0.06840530141085935, "grad_norm": 2.664503176902388, "learning_rate": 1.41e-05, "loss": 1.6196, "step": 20 }, { "epoch": 0.07182556648140231, "grad_norm": 2.5256162285977077, "learning_rate": 1.41e-05, "loss": 1.4734, "step": 21 }, { "epoch": 0.07524583155194528, "grad_norm": 2.232926984374301, "learning_rate": 1.41e-05, "loss": 1.4862, "step": 22 }, { "epoch": 0.07866609662248825, "grad_norm": 1.8467296080556872, "learning_rate": 1.41e-05, "loss": 1.4919, "step": 23 }, { "epoch": 0.08208636169303121, "grad_norm": 2.133405884576976, "learning_rate": 1.41e-05, "loss": 1.4962, "step": 24 }, { "epoch": 0.08550662676357418, "grad_norm": 2.0723150961914256, "learning_rate": 1.41e-05, "loss": 1.3745, "step": 25 }, { "epoch": 0.08892689183411714, "grad_norm": 2.2473775199610166, "learning_rate": 1.41e-05, "loss": 1.4678, "step": 26 }, { "epoch": 0.09234715690466011, "grad_norm": 3.1366913970429366, "learning_rate": 1.41e-05, "loss": 1.3231, "step": 27 }, { "epoch": 0.09576742197520308, "grad_norm": 2.2703890784758585, "learning_rate": 1.41e-05, "loss": 1.3328, "step": 28 }, { "epoch": 0.09918768704574604, "grad_norm": 2.7735799657439943, "learning_rate": 1.41e-05, "loss": 1.1703, "step": 29 }, { "epoch": 0.10260795211628901, "grad_norm": 2.5181744230002203, "learning_rate": 1.41e-05, "loss": 1.257, "step": 30 }, { "epoch": 0.10602821718683197, "grad_norm": 2.5187681530121546, "learning_rate": 1.41e-05, "loss": 1.1321, "step": 31 }, { "epoch": 0.10944848225737494, "grad_norm": 2.5085106975094678, "learning_rate": 1.41e-05, "loss": 1.0929, "step": 32 }, { "epoch": 0.11286874732791792, "grad_norm": 2.1171154989402887, "learning_rate": 1.41e-05, "loss": 1.1053, "step": 33 }, { "epoch": 0.11628901239846089, "grad_norm": 2.4016503905187725, "learning_rate": 1.41e-05, "loss": 0.9514, "step": 34 }, { "epoch": 0.11970927746900385, "grad_norm": 4.505185827517653, "learning_rate": 1.41e-05, "loss": 1.1443, "step": 35 }, { "epoch": 0.12312954253954682, "grad_norm": 2.283031446044169, "learning_rate": 1.41e-05, "loss": 1.1263, "step": 36 }, { "epoch": 0.12654980761008977, "grad_norm": 2.474396323802316, "learning_rate": 1.41e-05, "loss": 0.9587, "step": 37 }, { "epoch": 0.12997007268063274, "grad_norm": 3.3515325695241667, "learning_rate": 1.41e-05, "loss": 0.8727, "step": 38 }, { "epoch": 0.1333903377511757, "grad_norm": 2.3607499413214694, "learning_rate": 1.41e-05, "loss": 0.929, "step": 39 }, { "epoch": 0.1368106028217187, "grad_norm": 2.4723425187830537, "learning_rate": 1.41e-05, "loss": 1.0473, "step": 40 }, { "epoch": 0.14023086789226166, "grad_norm": 2.183114877668465, "learning_rate": 1.41e-05, "loss": 0.854, "step": 41 }, { "epoch": 0.14365113296280463, "grad_norm": 2.3660790712775555, "learning_rate": 1.41e-05, "loss": 0.8244, "step": 42 }, { "epoch": 0.1470713980333476, "grad_norm": 1.9551361934244411, "learning_rate": 1.41e-05, "loss": 0.8276, "step": 43 }, { "epoch": 0.15049166310389056, "grad_norm": 3.163303383692252, "learning_rate": 1.41e-05, "loss": 0.8468, "step": 44 }, { "epoch": 0.15391192817443353, "grad_norm": 3.163426556149129, "learning_rate": 1.41e-05, "loss": 0.7213, "step": 45 }, { "epoch": 0.1573321932449765, "grad_norm": 2.7934537634938783, "learning_rate": 1.41e-05, "loss": 0.7035, "step": 46 }, { "epoch": 0.16075245831551946, "grad_norm": 1.8747838269439079, "learning_rate": 1.41e-05, "loss": 0.7371, "step": 47 }, { "epoch": 0.16417272338606242, "grad_norm": 2.7812346413247484, "learning_rate": 1.41e-05, "loss": 0.7602, "step": 48 }, { "epoch": 0.1675929884566054, "grad_norm": 2.378687443172595, "learning_rate": 1.41e-05, "loss": 0.8561, "step": 49 }, { "epoch": 0.17101325352714836, "grad_norm": 2.3489013022006056, "learning_rate": 1.41e-05, "loss": 0.5867, "step": 50 }, { "epoch": 0.17443351859769132, "grad_norm": 1.7892416152344746, "learning_rate": 1.41e-05, "loss": 0.4853, "step": 51 }, { "epoch": 0.1778537836682343, "grad_norm": 2.077063902741533, "learning_rate": 1.41e-05, "loss": 0.6375, "step": 52 }, { "epoch": 0.18127404873877725, "grad_norm": 2.2312628788967412, "learning_rate": 1.41e-05, "loss": 0.6439, "step": 53 }, { "epoch": 0.18469431380932022, "grad_norm": 2.996527586876452, "learning_rate": 1.41e-05, "loss": 0.5925, "step": 54 }, { "epoch": 0.1881145788798632, "grad_norm": 1.8130852934136357, "learning_rate": 1.41e-05, "loss": 0.5575, "step": 55 }, { "epoch": 0.19153484395040615, "grad_norm": 1.5587429488330231, "learning_rate": 1.41e-05, "loss": 0.5954, "step": 56 }, { "epoch": 0.19495510902094912, "grad_norm": 1.7839250019187394, "learning_rate": 1.41e-05, "loss": 0.6196, "step": 57 }, { "epoch": 0.19837537409149208, "grad_norm": 1.6933241820211493, "learning_rate": 1.41e-05, "loss": 0.4969, "step": 58 }, { "epoch": 0.20179563916203505, "grad_norm": 1.731887729628178, "learning_rate": 1.41e-05, "loss": 0.6548, "step": 59 }, { "epoch": 0.20521590423257802, "grad_norm": 2.3987513930499995, "learning_rate": 1.41e-05, "loss": 0.4595, "step": 60 }, { "epoch": 0.20863616930312098, "grad_norm": 2.639090838839165, "learning_rate": 1.41e-05, "loss": 0.5341, "step": 61 }, { "epoch": 0.21205643437366395, "grad_norm": 1.7474798765435617, "learning_rate": 1.41e-05, "loss": 0.6015, "step": 62 }, { "epoch": 0.21547669944420692, "grad_norm": 3.571763528033457, "learning_rate": 1.41e-05, "loss": 0.4851, "step": 63 }, { "epoch": 0.21889696451474988, "grad_norm": 1.9011170860485487, "learning_rate": 1.41e-05, "loss": 0.4911, "step": 64 }, { "epoch": 0.22231722958529285, "grad_norm": 2.625704311611419, "learning_rate": 1.41e-05, "loss": 0.5247, "step": 65 }, { "epoch": 0.22573749465583584, "grad_norm": 1.8272075489091404, "learning_rate": 1.41e-05, "loss": 0.4123, "step": 66 }, { "epoch": 0.2291577597263788, "grad_norm": 2.005801690308512, "learning_rate": 1.41e-05, "loss": 0.437, "step": 67 }, { "epoch": 0.23257802479692177, "grad_norm": 2.009191411617363, "learning_rate": 1.41e-05, "loss": 0.4553, "step": 68 }, { "epoch": 0.23599828986746474, "grad_norm": 1.9856248091054856, "learning_rate": 1.41e-05, "loss": 0.4911, "step": 69 }, { "epoch": 0.2394185549380077, "grad_norm": 1.9582222549850024, "learning_rate": 1.41e-05, "loss": 0.4345, "step": 70 }, { "epoch": 0.24283882000855067, "grad_norm": 1.6336088110301379, "learning_rate": 1.41e-05, "loss": 0.4649, "step": 71 }, { "epoch": 0.24625908507909364, "grad_norm": 1.946765368736228, "learning_rate": 1.41e-05, "loss": 0.3968, "step": 72 }, { "epoch": 0.2496793501496366, "grad_norm": 1.9515164464490413, "learning_rate": 1.41e-05, "loss": 0.405, "step": 73 }, { "epoch": 0.25309961522017954, "grad_norm": 1.5507200892542412, "learning_rate": 1.41e-05, "loss": 0.2883, "step": 74 }, { "epoch": 0.25651988029072254, "grad_norm": 1.8953946078798778, "learning_rate": 1.41e-05, "loss": 0.3697, "step": 75 }, { "epoch": 0.2599401453612655, "grad_norm": 1.609567252811076, "learning_rate": 1.41e-05, "loss": 0.3912, "step": 76 }, { "epoch": 0.26336041043180847, "grad_norm": 1.841628814862342, "learning_rate": 1.41e-05, "loss": 0.3753, "step": 77 }, { "epoch": 0.2667806755023514, "grad_norm": 1.3866520663429722, "learning_rate": 1.41e-05, "loss": 0.2978, "step": 78 }, { "epoch": 0.2702009405728944, "grad_norm": 1.8662166292908873, "learning_rate": 1.41e-05, "loss": 0.325, "step": 79 }, { "epoch": 0.2736212056434374, "grad_norm": 1.86990469712952, "learning_rate": 1.41e-05, "loss": 0.3608, "step": 80 }, { "epoch": 0.27704147071398033, "grad_norm": 1.7776490861424512, "learning_rate": 1.41e-05, "loss": 0.3164, "step": 81 }, { "epoch": 0.2804617357845233, "grad_norm": 1.5257018693275182, "learning_rate": 1.41e-05, "loss": 0.2736, "step": 82 }, { "epoch": 0.28388200085506626, "grad_norm": 1.2107786177311144, "learning_rate": 1.41e-05, "loss": 0.2293, "step": 83 }, { "epoch": 0.28730226592560926, "grad_norm": 1.7075082762966491, "learning_rate": 1.41e-05, "loss": 0.3312, "step": 84 }, { "epoch": 0.2907225309961522, "grad_norm": 1.4271608707593932, "learning_rate": 1.41e-05, "loss": 0.2731, "step": 85 }, { "epoch": 0.2941427960666952, "grad_norm": 2.0611639591955466, "learning_rate": 1.41e-05, "loss": 0.2708, "step": 86 }, { "epoch": 0.2975630611372381, "grad_norm": 1.2754227728205574, "learning_rate": 1.41e-05, "loss": 0.2301, "step": 87 }, { "epoch": 0.3009833262077811, "grad_norm": 1.312428216280633, "learning_rate": 1.41e-05, "loss": 0.2219, "step": 88 }, { "epoch": 0.30440359127832406, "grad_norm": 1.36584246397569, "learning_rate": 1.41e-05, "loss": 0.2761, "step": 89 }, { "epoch": 0.30782385634886705, "grad_norm": 1.1768002468324985, "learning_rate": 1.41e-05, "loss": 0.1884, "step": 90 }, { "epoch": 0.31124412141941, "grad_norm": 1.402779753210287, "learning_rate": 1.41e-05, "loss": 0.24, "step": 91 }, { "epoch": 0.314664386489953, "grad_norm": 1.2917943074724656, "learning_rate": 1.41e-05, "loss": 0.2487, "step": 92 }, { "epoch": 0.3180846515604959, "grad_norm": 1.4168579440960474, "learning_rate": 1.41e-05, "loss": 0.2429, "step": 93 }, { "epoch": 0.3215049166310389, "grad_norm": 1.1040710078361045, "learning_rate": 1.41e-05, "loss": 0.1676, "step": 94 }, { "epoch": 0.32492518170158186, "grad_norm": 1.6692824066389471, "learning_rate": 1.41e-05, "loss": 0.2456, "step": 95 }, { "epoch": 0.32834544677212485, "grad_norm": 1.4794418835245668, "learning_rate": 1.41e-05, "loss": 0.2346, "step": 96 }, { "epoch": 0.3317657118426678, "grad_norm": 1.2426804107195764, "learning_rate": 1.41e-05, "loss": 0.2286, "step": 97 }, { "epoch": 0.3351859769132108, "grad_norm": 1.5976518274152767, "learning_rate": 1.41e-05, "loss": 0.2356, "step": 98 }, { "epoch": 0.3386062419837537, "grad_norm": 1.3380817380576575, "learning_rate": 1.41e-05, "loss": 0.2361, "step": 99 }, { "epoch": 0.3420265070542967, "grad_norm": 1.8499907335313168, "learning_rate": 1.41e-05, "loss": 0.2144, "step": 100 }, { "epoch": 0.34544677212483965, "grad_norm": 1.5519650604852626, "learning_rate": 1.41e-05, "loss": 0.2211, "step": 101 }, { "epoch": 0.34886703719538265, "grad_norm": 1.4110088914262258, "learning_rate": 1.41e-05, "loss": 0.2422, "step": 102 }, { "epoch": 0.3522873022659256, "grad_norm": 1.4052513404508866, "learning_rate": 1.41e-05, "loss": 0.1939, "step": 103 }, { "epoch": 0.3557075673364686, "grad_norm": 1.377026410169766, "learning_rate": 1.41e-05, "loss": 0.193, "step": 104 }, { "epoch": 0.3591278324070115, "grad_norm": 1.0901968649772857, "learning_rate": 1.41e-05, "loss": 0.1705, "step": 105 }, { "epoch": 0.3625480974775545, "grad_norm": 1.2716891700474584, "learning_rate": 1.41e-05, "loss": 0.1998, "step": 106 }, { "epoch": 0.3659683625480975, "grad_norm": 1.269092339983086, "learning_rate": 1.41e-05, "loss": 0.1845, "step": 107 }, { "epoch": 0.36938862761864044, "grad_norm": 1.0134148691503204, "learning_rate": 1.41e-05, "loss": 0.1457, "step": 108 }, { "epoch": 0.37280889268918344, "grad_norm": 1.0973618617348355, "learning_rate": 1.41e-05, "loss": 0.1829, "step": 109 }, { "epoch": 0.3762291577597264, "grad_norm": 1.3476960032405298, "learning_rate": 1.41e-05, "loss": 0.1936, "step": 110 }, { "epoch": 0.37964942283026937, "grad_norm": 1.1243588113049168, "learning_rate": 1.41e-05, "loss": 0.1763, "step": 111 }, { "epoch": 0.3830696879008123, "grad_norm": 1.756776902549336, "learning_rate": 1.41e-05, "loss": 0.1772, "step": 112 }, { "epoch": 0.3864899529713553, "grad_norm": 1.0653616123456815, "learning_rate": 1.41e-05, "loss": 0.1485, "step": 113 }, { "epoch": 0.38991021804189824, "grad_norm": 1.322506116172751, "learning_rate": 1.41e-05, "loss": 0.1511, "step": 114 }, { "epoch": 0.39333048311244123, "grad_norm": 1.134939568700316, "learning_rate": 1.41e-05, "loss": 0.1595, "step": 115 }, { "epoch": 0.39675074818298417, "grad_norm": 1.2351108355732012, "learning_rate": 1.41e-05, "loss": 0.1653, "step": 116 }, { "epoch": 0.40017101325352716, "grad_norm": 1.291174811726909, "learning_rate": 1.41e-05, "loss": 0.1414, "step": 117 }, { "epoch": 0.4035912783240701, "grad_norm": 1.3110441911359814, "learning_rate": 1.41e-05, "loss": 0.1602, "step": 118 }, { "epoch": 0.4070115433946131, "grad_norm": 1.0324677945393839, "learning_rate": 1.41e-05, "loss": 0.1589, "step": 119 }, { "epoch": 0.41043180846515603, "grad_norm": 1.1528407686685203, "learning_rate": 1.41e-05, "loss": 0.142, "step": 120 }, { "epoch": 0.413852073535699, "grad_norm": 1.259359755456077, "learning_rate": 1.41e-05, "loss": 0.1671, "step": 121 }, { "epoch": 0.41727233860624197, "grad_norm": 0.9342347863650876, "learning_rate": 1.41e-05, "loss": 0.123, "step": 122 }, { "epoch": 0.42069260367678496, "grad_norm": 1.0262297732751853, "learning_rate": 1.41e-05, "loss": 0.1379, "step": 123 }, { "epoch": 0.4241128687473279, "grad_norm": 0.8528122922890459, "learning_rate": 1.41e-05, "loss": 0.1115, "step": 124 }, { "epoch": 0.4275331338178709, "grad_norm": 1.0773688633850005, "learning_rate": 1.41e-05, "loss": 0.1368, "step": 125 }, { "epoch": 0.43095339888841383, "grad_norm": 1.02543648603311, "learning_rate": 1.41e-05, "loss": 0.1116, "step": 126 }, { "epoch": 0.4343736639589568, "grad_norm": 0.8824964139082571, "learning_rate": 1.41e-05, "loss": 0.1149, "step": 127 }, { "epoch": 0.43779392902949976, "grad_norm": 1.1649358884512249, "learning_rate": 1.41e-05, "loss": 0.127, "step": 128 }, { "epoch": 0.44121419410004276, "grad_norm": 1.0614948335318424, "learning_rate": 1.41e-05, "loss": 0.1248, "step": 129 }, { "epoch": 0.4446344591705857, "grad_norm": 0.7776250289623791, "learning_rate": 1.41e-05, "loss": 0.1145, "step": 130 }, { "epoch": 0.4480547242411287, "grad_norm": 0.8384984456348001, "learning_rate": 1.41e-05, "loss": 0.0921, "step": 131 }, { "epoch": 0.4514749893116717, "grad_norm": 1.0151085870429353, "learning_rate": 1.41e-05, "loss": 0.1226, "step": 132 }, { "epoch": 0.4548952543822146, "grad_norm": 0.7615407647681977, "learning_rate": 1.41e-05, "loss": 0.0951, "step": 133 }, { "epoch": 0.4583155194527576, "grad_norm": 0.9555465270405811, "learning_rate": 1.41e-05, "loss": 0.1186, "step": 134 }, { "epoch": 0.46173578452330055, "grad_norm": 0.9186643738206431, "learning_rate": 1.41e-05, "loss": 0.1207, "step": 135 }, { "epoch": 0.46515604959384355, "grad_norm": 0.8860210174403064, "learning_rate": 1.41e-05, "loss": 0.1195, "step": 136 }, { "epoch": 0.4685763146643865, "grad_norm": 0.8817179419098865, "learning_rate": 1.41e-05, "loss": 0.0897, "step": 137 }, { "epoch": 0.4719965797349295, "grad_norm": 0.8289163487116727, "learning_rate": 1.41e-05, "loss": 0.1003, "step": 138 }, { "epoch": 0.4754168448054724, "grad_norm": 0.904736501387183, "learning_rate": 1.41e-05, "loss": 0.0924, "step": 139 }, { "epoch": 0.4788371098760154, "grad_norm": 0.943604807746642, "learning_rate": 1.41e-05, "loss": 0.1109, "step": 140 }, { "epoch": 0.48225737494655835, "grad_norm": 0.9402997510757618, "learning_rate": 1.41e-05, "loss": 0.1145, "step": 141 }, { "epoch": 0.48567764001710134, "grad_norm": 0.7366551780058475, "learning_rate": 1.41e-05, "loss": 0.0899, "step": 142 }, { "epoch": 0.4890979050876443, "grad_norm": 0.9208800933128714, "learning_rate": 1.41e-05, "loss": 0.1167, "step": 143 }, { "epoch": 0.4925181701581873, "grad_norm": 0.750448679907497, "learning_rate": 1.41e-05, "loss": 0.0809, "step": 144 }, { "epoch": 0.4959384352287302, "grad_norm": 0.8092636444718945, "learning_rate": 1.41e-05, "loss": 0.0908, "step": 145 }, { "epoch": 0.4993587002992732, "grad_norm": 0.8625942305572842, "learning_rate": 1.41e-05, "loss": 0.1031, "step": 146 }, { "epoch": 0.5027789653698161, "grad_norm": 0.9198105373299891, "learning_rate": 1.41e-05, "loss": 0.1113, "step": 147 }, { "epoch": 0.5061992304403591, "grad_norm": 0.7848012242788545, "learning_rate": 1.41e-05, "loss": 0.0865, "step": 148 }, { "epoch": 0.5096194955109021, "grad_norm": 0.9324979893254519, "learning_rate": 1.41e-05, "loss": 0.0971, "step": 149 }, { "epoch": 0.5130397605814451, "grad_norm": 0.7143510040718761, "learning_rate": 1.41e-05, "loss": 0.0795, "step": 150 }, { "epoch": 0.516460025651988, "grad_norm": 1.0676584465757966, "learning_rate": 1.41e-05, "loss": 0.1135, "step": 151 }, { "epoch": 0.519880290722531, "grad_norm": 0.6822736052229258, "learning_rate": 1.41e-05, "loss": 0.0733, "step": 152 }, { "epoch": 0.523300555793074, "grad_norm": 0.8374258007686984, "learning_rate": 1.41e-05, "loss": 0.0907, "step": 153 }, { "epoch": 0.5267208208636169, "grad_norm": 0.8341896688419776, "learning_rate": 1.41e-05, "loss": 0.0939, "step": 154 }, { "epoch": 0.5301410859341599, "grad_norm": 0.7611257220630606, "learning_rate": 1.41e-05, "loss": 0.0869, "step": 155 }, { "epoch": 0.5335613510047028, "grad_norm": 0.740325417856638, "learning_rate": 1.41e-05, "loss": 0.0832, "step": 156 }, { "epoch": 0.5369816160752459, "grad_norm": 0.7042263268553958, "learning_rate": 1.41e-05, "loss": 0.0716, "step": 157 }, { "epoch": 0.5404018811457888, "grad_norm": 0.7335797206103793, "learning_rate": 1.41e-05, "loss": 0.0807, "step": 158 }, { "epoch": 0.5438221462163317, "grad_norm": 0.8019722685081757, "learning_rate": 1.41e-05, "loss": 0.0826, "step": 159 }, { "epoch": 0.5472424112868748, "grad_norm": 0.6929468279193534, "learning_rate": 1.41e-05, "loss": 0.0821, "step": 160 }, { "epoch": 0.5506626763574177, "grad_norm": 0.6854252563729888, "learning_rate": 1.41e-05, "loss": 0.0675, "step": 161 }, { "epoch": 0.5540829414279607, "grad_norm": 0.6741146081895844, "learning_rate": 1.41e-05, "loss": 0.0749, "step": 162 }, { "epoch": 0.5575032064985036, "grad_norm": 0.6902694268516201, "learning_rate": 1.41e-05, "loss": 0.0707, "step": 163 }, { "epoch": 0.5609234715690466, "grad_norm": 0.6948144741970704, "learning_rate": 1.41e-05, "loss": 0.0795, "step": 164 }, { "epoch": 0.5643437366395896, "grad_norm": 0.7169974641783955, "learning_rate": 1.41e-05, "loss": 0.0818, "step": 165 }, { "epoch": 0.5677640017101325, "grad_norm": 0.6384211122986987, "learning_rate": 1.41e-05, "loss": 0.0735, "step": 166 }, { "epoch": 0.5711842667806755, "grad_norm": 0.7417356609403256, "learning_rate": 1.41e-05, "loss": 0.0842, "step": 167 }, { "epoch": 0.5746045318512185, "grad_norm": 0.66647458408933, "learning_rate": 1.41e-05, "loss": 0.0814, "step": 168 }, { "epoch": 0.5780247969217615, "grad_norm": 0.7378707726234803, "learning_rate": 1.41e-05, "loss": 0.0746, "step": 169 }, { "epoch": 0.5814450619923044, "grad_norm": 0.6812399388437269, "learning_rate": 1.41e-05, "loss": 0.0828, "step": 170 }, { "epoch": 0.5848653270628473, "grad_norm": 0.6793042032294047, "learning_rate": 1.41e-05, "loss": 0.0724, "step": 171 }, { "epoch": 0.5882855921333904, "grad_norm": 0.6062435025746024, "learning_rate": 1.41e-05, "loss": 0.0649, "step": 172 }, { "epoch": 0.5917058572039333, "grad_norm": 0.6813501327595134, "learning_rate": 1.41e-05, "loss": 0.0711, "step": 173 }, { "epoch": 0.5951261222744763, "grad_norm": 0.672907340612349, "learning_rate": 1.41e-05, "loss": 0.0769, "step": 174 }, { "epoch": 0.5985463873450192, "grad_norm": 0.5410302293555107, "learning_rate": 1.41e-05, "loss": 0.0588, "step": 175 }, { "epoch": 0.6019666524155622, "grad_norm": 0.6636988853462809, "learning_rate": 1.41e-05, "loss": 0.0702, "step": 176 }, { "epoch": 0.6053869174861052, "grad_norm": 0.5918082791413835, "learning_rate": 1.41e-05, "loss": 0.0626, "step": 177 }, { "epoch": 0.6088071825566481, "grad_norm": 0.5816411742116699, "learning_rate": 1.41e-05, "loss": 0.059, "step": 178 }, { "epoch": 0.6122274476271911, "grad_norm": 0.6819230663398479, "learning_rate": 1.41e-05, "loss": 0.074, "step": 179 }, { "epoch": 0.6156477126977341, "grad_norm": 0.6727834641069249, "learning_rate": 1.41e-05, "loss": 0.0739, "step": 180 }, { "epoch": 0.619067977768277, "grad_norm": 0.5791689945437091, "learning_rate": 1.41e-05, "loss": 0.0553, "step": 181 }, { "epoch": 0.62248824283882, "grad_norm": 0.6060693467771211, "learning_rate": 1.41e-05, "loss": 0.0572, "step": 182 }, { "epoch": 0.6259085079093629, "grad_norm": 0.6611378853926027, "learning_rate": 1.41e-05, "loss": 0.0697, "step": 183 }, { "epoch": 0.629328772979906, "grad_norm": 0.5887572335679387, "learning_rate": 1.41e-05, "loss": 0.0603, "step": 184 }, { "epoch": 0.6327490380504489, "grad_norm": 0.538930391171475, "learning_rate": 1.41e-05, "loss": 0.0462, "step": 185 }, { "epoch": 0.6361693031209918, "grad_norm": 0.5617056384496442, "learning_rate": 1.41e-05, "loss": 0.0549, "step": 186 }, { "epoch": 0.6395895681915349, "grad_norm": 0.5912693387471951, "learning_rate": 1.41e-05, "loss": 0.0574, "step": 187 }, { "epoch": 0.6430098332620778, "grad_norm": 0.5373216387973052, "learning_rate": 1.41e-05, "loss": 0.0574, "step": 188 }, { "epoch": 0.6464300983326208, "grad_norm": 0.6151671164129469, "learning_rate": 1.41e-05, "loss": 0.0533, "step": 189 }, { "epoch": 0.6498503634031637, "grad_norm": 0.5394742707247884, "learning_rate": 1.41e-05, "loss": 0.0575, "step": 190 }, { "epoch": 0.6532706284737068, "grad_norm": 0.5752514447611141, "learning_rate": 1.41e-05, "loss": 0.0574, "step": 191 }, { "epoch": 0.6566908935442497, "grad_norm": 0.5136422669182581, "learning_rate": 1.41e-05, "loss": 0.054, "step": 192 }, { "epoch": 0.6601111586147926, "grad_norm": 0.6261951776293332, "learning_rate": 1.41e-05, "loss": 0.0612, "step": 193 }, { "epoch": 0.6635314236853356, "grad_norm": 0.5067466055288193, "learning_rate": 1.41e-05, "loss": 0.054, "step": 194 }, { "epoch": 0.6669516887558786, "grad_norm": 0.5892942817895197, "learning_rate": 1.41e-05, "loss": 0.0572, "step": 195 }, { "epoch": 0.6703719538264216, "grad_norm": 0.5652390088377635, "learning_rate": 1.41e-05, "loss": 0.0655, "step": 196 }, { "epoch": 0.6737922188969645, "grad_norm": 0.5285822929717092, "learning_rate": 1.41e-05, "loss": 0.0494, "step": 197 }, { "epoch": 0.6772124839675074, "grad_norm": 0.5246191872665474, "learning_rate": 1.41e-05, "loss": 0.0614, "step": 198 }, { "epoch": 0.6806327490380505, "grad_norm": 0.5387443642901717, "learning_rate": 1.41e-05, "loss": 0.0531, "step": 199 }, { "epoch": 0.6840530141085934, "grad_norm": 0.5238421425157995, "learning_rate": 1.41e-05, "loss": 0.0512, "step": 200 }, { "epoch": 0.6874732791791364, "grad_norm": 0.4882221825014573, "learning_rate": 1.41e-05, "loss": 0.0489, "step": 201 }, { "epoch": 0.6908935442496793, "grad_norm": 0.5646516308611281, "learning_rate": 1.41e-05, "loss": 0.0512, "step": 202 }, { "epoch": 0.6943138093202224, "grad_norm": 0.5778234440206057, "learning_rate": 1.41e-05, "loss": 0.0634, "step": 203 }, { "epoch": 0.6977340743907653, "grad_norm": 0.5293741220834561, "learning_rate": 1.41e-05, "loss": 0.056, "step": 204 }, { "epoch": 0.7011543394613082, "grad_norm": 0.527641578215631, "learning_rate": 1.41e-05, "loss": 0.0561, "step": 205 }, { "epoch": 0.7045746045318512, "grad_norm": 0.4000174523279179, "learning_rate": 1.41e-05, "loss": 0.0401, "step": 206 }, { "epoch": 0.7079948696023942, "grad_norm": 0.5298298097123045, "learning_rate": 1.41e-05, "loss": 0.0609, "step": 207 }, { "epoch": 0.7114151346729372, "grad_norm": 0.5349126180633413, "learning_rate": 1.41e-05, "loss": 0.052, "step": 208 }, { "epoch": 0.7148353997434801, "grad_norm": 0.5748237411918808, "learning_rate": 1.41e-05, "loss": 0.0651, "step": 209 }, { "epoch": 0.718255664814023, "grad_norm": 0.4989630175919984, "learning_rate": 1.41e-05, "loss": 0.0502, "step": 210 }, { "epoch": 0.7216759298845661, "grad_norm": 0.5604803895822472, "learning_rate": 1.41e-05, "loss": 0.054, "step": 211 }, { "epoch": 0.725096194955109, "grad_norm": 0.6062121706000653, "learning_rate": 1.41e-05, "loss": 0.0644, "step": 212 }, { "epoch": 0.728516460025652, "grad_norm": 0.49966896896386376, "learning_rate": 1.41e-05, "loss": 0.0469, "step": 213 }, { "epoch": 0.731936725096195, "grad_norm": 0.44454274508391683, "learning_rate": 1.41e-05, "loss": 0.0428, "step": 214 }, { "epoch": 0.735356990166738, "grad_norm": 0.563352438281472, "learning_rate": 1.41e-05, "loss": 0.055, "step": 215 }, { "epoch": 0.7387772552372809, "grad_norm": 0.6191641486840816, "learning_rate": 1.41e-05, "loss": 0.063, "step": 216 }, { "epoch": 0.7421975203078238, "grad_norm": 0.4827104401750106, "learning_rate": 1.41e-05, "loss": 0.0496, "step": 217 }, { "epoch": 0.7456177853783669, "grad_norm": 0.4656663916077845, "learning_rate": 1.41e-05, "loss": 0.0528, "step": 218 }, { "epoch": 0.7490380504489098, "grad_norm": 0.4785719764753492, "learning_rate": 1.41e-05, "loss": 0.0438, "step": 219 }, { "epoch": 0.7524583155194527, "grad_norm": 0.5448096052982832, "learning_rate": 1.41e-05, "loss": 0.0422, "step": 220 }, { "epoch": 0.7558785805899957, "grad_norm": 0.5065090698982245, "learning_rate": 1.41e-05, "loss": 0.0489, "step": 221 }, { "epoch": 0.7592988456605387, "grad_norm": 0.5059637979656728, "learning_rate": 1.41e-05, "loss": 0.0485, "step": 222 }, { "epoch": 0.7627191107310817, "grad_norm": 0.49583806553345944, "learning_rate": 1.41e-05, "loss": 0.0525, "step": 223 }, { "epoch": 0.7661393758016246, "grad_norm": 0.4012288566084756, "learning_rate": 1.41e-05, "loss": 0.0384, "step": 224 }, { "epoch": 0.7695596408721675, "grad_norm": 0.49449469823574593, "learning_rate": 1.41e-05, "loss": 0.05, "step": 225 }, { "epoch": 0.7729799059427106, "grad_norm": 0.45672465215152086, "learning_rate": 1.41e-05, "loss": 0.0479, "step": 226 }, { "epoch": 0.7764001710132535, "grad_norm": 0.4909318818146214, "learning_rate": 1.41e-05, "loss": 0.0499, "step": 227 }, { "epoch": 0.7798204360837965, "grad_norm": 0.40392311014383353, "learning_rate": 1.41e-05, "loss": 0.0406, "step": 228 }, { "epoch": 0.7832407011543394, "grad_norm": 0.4131849222227493, "learning_rate": 1.41e-05, "loss": 0.0388, "step": 229 }, { "epoch": 0.7866609662248825, "grad_norm": 0.5611423817088044, "learning_rate": 1.41e-05, "loss": 0.0594, "step": 230 }, { "epoch": 0.7900812312954254, "grad_norm": 0.47364645404517464, "learning_rate": 1.41e-05, "loss": 0.0522, "step": 231 }, { "epoch": 0.7935014963659683, "grad_norm": 0.4545208834696141, "learning_rate": 1.41e-05, "loss": 0.048, "step": 232 }, { "epoch": 0.7969217614365113, "grad_norm": 0.4113439560012879, "learning_rate": 1.41e-05, "loss": 0.0393, "step": 233 }, { "epoch": 0.8003420265070543, "grad_norm": 0.42334212760511825, "learning_rate": 1.41e-05, "loss": 0.0404, "step": 234 }, { "epoch": 0.8037622915775973, "grad_norm": 0.4768224768125407, "learning_rate": 1.41e-05, "loss": 0.0479, "step": 235 }, { "epoch": 0.8071825566481402, "grad_norm": 0.43678389875922824, "learning_rate": 1.41e-05, "loss": 0.0442, "step": 236 }, { "epoch": 0.8106028217186833, "grad_norm": 0.43277663891476426, "learning_rate": 1.41e-05, "loss": 0.0409, "step": 237 }, { "epoch": 0.8140230867892262, "grad_norm": 0.4443462520817696, "learning_rate": 1.41e-05, "loss": 0.0465, "step": 238 }, { "epoch": 0.8174433518597691, "grad_norm": 0.43839684525150946, "learning_rate": 1.41e-05, "loss": 0.0415, "step": 239 }, { "epoch": 0.8208636169303121, "grad_norm": 0.43297499486580887, "learning_rate": 1.41e-05, "loss": 0.0445, "step": 240 }, { "epoch": 0.8242838820008551, "grad_norm": 0.38283138235459124, "learning_rate": 1.41e-05, "loss": 0.0404, "step": 241 }, { "epoch": 0.827704147071398, "grad_norm": 0.40333504070789256, "learning_rate": 1.41e-05, "loss": 0.0373, "step": 242 }, { "epoch": 0.831124412141941, "grad_norm": 0.40950403615291003, "learning_rate": 1.41e-05, "loss": 0.0423, "step": 243 }, { "epoch": 0.8345446772124839, "grad_norm": 0.4314317792707956, "learning_rate": 1.41e-05, "loss": 0.0424, "step": 244 }, { "epoch": 0.837964942283027, "grad_norm": 0.4189847423801003, "learning_rate": 1.41e-05, "loss": 0.0434, "step": 245 }, { "epoch": 0.8413852073535699, "grad_norm": 0.4593355477826361, "learning_rate": 1.41e-05, "loss": 0.0456, "step": 246 }, { "epoch": 0.8448054724241129, "grad_norm": 0.43148788855113257, "learning_rate": 1.41e-05, "loss": 0.0429, "step": 247 }, { "epoch": 0.8482257374946558, "grad_norm": 0.41015663281431336, "learning_rate": 1.41e-05, "loss": 0.0411, "step": 248 }, { "epoch": 0.8516460025651988, "grad_norm": 0.39479744422344626, "learning_rate": 1.41e-05, "loss": 0.0408, "step": 249 }, { "epoch": 0.8550662676357418, "grad_norm": 0.49034951176740266, "learning_rate": 1.41e-05, "loss": 0.056, "step": 250 }, { "epoch": 0.8584865327062847, "grad_norm": 0.4241786968876316, "learning_rate": 1.41e-05, "loss": 0.0414, "step": 251 }, { "epoch": 0.8619067977768277, "grad_norm": 0.40856481199229994, "learning_rate": 1.41e-05, "loss": 0.0384, "step": 252 }, { "epoch": 0.8653270628473707, "grad_norm": 0.44863803613683295, "learning_rate": 1.41e-05, "loss": 0.0454, "step": 253 }, { "epoch": 0.8687473279179136, "grad_norm": 0.4315245075802853, "learning_rate": 1.41e-05, "loss": 0.0425, "step": 254 }, { "epoch": 0.8721675929884566, "grad_norm": 0.4300931739828693, "learning_rate": 1.41e-05, "loss": 0.0386, "step": 255 }, { "epoch": 0.8755878580589995, "grad_norm": 0.41516523943726863, "learning_rate": 1.41e-05, "loss": 0.044, "step": 256 }, { "epoch": 0.8790081231295426, "grad_norm": 0.41978138385172276, "learning_rate": 1.41e-05, "loss": 0.0376, "step": 257 }, { "epoch": 0.8824283882000855, "grad_norm": 0.38286432398443565, "learning_rate": 1.41e-05, "loss": 0.0357, "step": 258 }, { "epoch": 0.8858486532706284, "grad_norm": 1.0190380922512265, "learning_rate": 1.41e-05, "loss": 0.0413, "step": 259 }, { "epoch": 0.8892689183411714, "grad_norm": 0.46857802366186146, "learning_rate": 1.41e-05, "loss": 0.0468, "step": 260 }, { "epoch": 0.8926891834117144, "grad_norm": 0.38236119116632356, "learning_rate": 1.41e-05, "loss": 0.0383, "step": 261 }, { "epoch": 0.8961094484822574, "grad_norm": 0.4551658885616909, "learning_rate": 1.41e-05, "loss": 0.0453, "step": 262 }, { "epoch": 0.8995297135528003, "grad_norm": 0.4286695716016865, "learning_rate": 1.41e-05, "loss": 0.0408, "step": 263 }, { "epoch": 0.9029499786233434, "grad_norm": 0.43951997404486826, "learning_rate": 1.41e-05, "loss": 0.0395, "step": 264 }, { "epoch": 0.9063702436938863, "grad_norm": 0.4395704115846384, "learning_rate": 1.41e-05, "loss": 0.0419, "step": 265 }, { "epoch": 0.9097905087644292, "grad_norm": 0.3839941648623069, "learning_rate": 1.41e-05, "loss": 0.0374, "step": 266 }, { "epoch": 0.9132107738349722, "grad_norm": 0.922460042226267, "learning_rate": 1.41e-05, "loss": 0.0372, "step": 267 }, { "epoch": 0.9166310389055152, "grad_norm": 0.44811463225682846, "learning_rate": 1.41e-05, "loss": 0.043, "step": 268 }, { "epoch": 0.9200513039760582, "grad_norm": 0.4172986063615246, "learning_rate": 1.41e-05, "loss": 0.0365, "step": 269 }, { "epoch": 0.9234715690466011, "grad_norm": 0.37219225103541853, "learning_rate": 1.41e-05, "loss": 0.0377, "step": 270 }, { "epoch": 0.926891834117144, "grad_norm": 0.3447899677572012, "learning_rate": 1.41e-05, "loss": 0.0317, "step": 271 }, { "epoch": 0.9303120991876871, "grad_norm": 0.3964175357323583, "learning_rate": 1.41e-05, "loss": 0.034, "step": 272 }, { "epoch": 0.93373236425823, "grad_norm": 0.36028858741706155, "learning_rate": 1.41e-05, "loss": 0.0351, "step": 273 }, { "epoch": 0.937152629328773, "grad_norm": 0.46256283545817356, "learning_rate": 1.41e-05, "loss": 0.041, "step": 274 }, { "epoch": 0.9405728943993159, "grad_norm": 0.3781578237183448, "learning_rate": 1.41e-05, "loss": 0.0366, "step": 275 }, { "epoch": 0.943993159469859, "grad_norm": 0.3818392999731623, "learning_rate": 1.41e-05, "loss": 0.0321, "step": 276 }, { "epoch": 0.9474134245404019, "grad_norm": 0.38568204387768645, "learning_rate": 1.41e-05, "loss": 0.0336, "step": 277 }, { "epoch": 0.9508336896109448, "grad_norm": 0.3905057560704898, "learning_rate": 1.41e-05, "loss": 0.0349, "step": 278 }, { "epoch": 0.9542539546814878, "grad_norm": 0.36210340527903007, "learning_rate": 1.41e-05, "loss": 0.0351, "step": 279 }, { "epoch": 0.9576742197520308, "grad_norm": 0.3784578196301756, "learning_rate": 1.41e-05, "loss": 0.0382, "step": 280 }, { "epoch": 0.9610944848225738, "grad_norm": 0.4415630974468222, "learning_rate": 1.41e-05, "loss": 0.0417, "step": 281 }, { "epoch": 0.9645147498931167, "grad_norm": 0.36220744727441767, "learning_rate": 1.41e-05, "loss": 0.0336, "step": 282 }, { "epoch": 0.9679350149636596, "grad_norm": 0.4266084587936558, "learning_rate": 1.41e-05, "loss": 0.0384, "step": 283 }, { "epoch": 0.9713552800342027, "grad_norm": 0.3599843427563046, "learning_rate": 1.41e-05, "loss": 0.035, "step": 284 }, { "epoch": 0.9747755451047456, "grad_norm": 0.38722401771389997, "learning_rate": 1.41e-05, "loss": 0.0348, "step": 285 }, { "epoch": 0.9781958101752886, "grad_norm": 0.40946001056055625, "learning_rate": 1.41e-05, "loss": 0.0373, "step": 286 }, { "epoch": 0.9816160752458315, "grad_norm": 0.41550323303820474, "learning_rate": 1.41e-05, "loss": 0.04, "step": 287 }, { "epoch": 0.9850363403163745, "grad_norm": 0.37048272944847027, "learning_rate": 1.41e-05, "loss": 0.0309, "step": 288 }, { "epoch": 0.9884566053869175, "grad_norm": 0.3620011070515116, "learning_rate": 1.41e-05, "loss": 0.0341, "step": 289 }, { "epoch": 0.9918768704574604, "grad_norm": 0.4132102658215653, "learning_rate": 1.41e-05, "loss": 0.0383, "step": 290 }, { "epoch": 0.9952971355280035, "grad_norm": 0.3852351778306183, "learning_rate": 1.41e-05, "loss": 0.0357, "step": 291 }, { "epoch": 0.9987174005985464, "grad_norm": 0.37414254856973184, "learning_rate": 1.41e-05, "loss": 0.0348, "step": 292 }, { "epoch": 1.0021376656690895, "grad_norm": 0.31841258787737364, "learning_rate": 1.41e-05, "loss": 0.0308, "step": 293 }, { "epoch": 1.0055579307396323, "grad_norm": 0.30923263652050564, "learning_rate": 1.41e-05, "loss": 0.0263, "step": 294 }, { "epoch": 1.0089781958101753, "grad_norm": 0.3118052788978316, "learning_rate": 1.41e-05, "loss": 0.0243, "step": 295 }, { "epoch": 1.0123984608807182, "grad_norm": 0.2908942931886208, "learning_rate": 1.41e-05, "loss": 0.0239, "step": 296 }, { "epoch": 1.0158187259512612, "grad_norm": 0.2904620406307295, "learning_rate": 1.41e-05, "loss": 0.0222, "step": 297 }, { "epoch": 1.0192389910218043, "grad_norm": 0.28692807204183246, "learning_rate": 1.41e-05, "loss": 0.0225, "step": 298 }, { "epoch": 1.022659256092347, "grad_norm": 0.32077856448530445, "learning_rate": 1.41e-05, "loss": 0.025, "step": 299 }, { "epoch": 1.0260795211628901, "grad_norm": 0.28877834524497115, "learning_rate": 1.41e-05, "loss": 0.0214, "step": 300 }, { "epoch": 1.0294997862334332, "grad_norm": 0.27974144729840145, "learning_rate": 1.41e-05, "loss": 0.0226, "step": 301 }, { "epoch": 1.032920051303976, "grad_norm": 0.2976755981973845, "learning_rate": 1.41e-05, "loss": 0.025, "step": 302 }, { "epoch": 1.036340316374519, "grad_norm": 0.2909288296995041, "learning_rate": 1.41e-05, "loss": 0.0241, "step": 303 }, { "epoch": 1.039760581445062, "grad_norm": 0.26580362507255084, "learning_rate": 1.41e-05, "loss": 0.0227, "step": 304 }, { "epoch": 1.043180846515605, "grad_norm": 0.3375634201266733, "learning_rate": 1.41e-05, "loss": 0.0261, "step": 305 }, { "epoch": 1.046601111586148, "grad_norm": 0.2797681439324777, "learning_rate": 1.41e-05, "loss": 0.023, "step": 306 }, { "epoch": 1.0500213766566908, "grad_norm": 0.3197672217160271, "learning_rate": 1.41e-05, "loss": 0.0258, "step": 307 }, { "epoch": 1.0534416417272339, "grad_norm": 0.2834421342617412, "learning_rate": 1.41e-05, "loss": 0.0231, "step": 308 }, { "epoch": 1.056861906797777, "grad_norm": 0.30638464271397325, "learning_rate": 1.41e-05, "loss": 0.0237, "step": 309 }, { "epoch": 1.0602821718683197, "grad_norm": 0.2772969145815675, "learning_rate": 1.41e-05, "loss": 0.0227, "step": 310 }, { "epoch": 1.0637024369388628, "grad_norm": 0.27940000335528603, "learning_rate": 1.41e-05, "loss": 0.0232, "step": 311 }, { "epoch": 1.0671227020094056, "grad_norm": 0.2793001905583242, "learning_rate": 1.41e-05, "loss": 0.024, "step": 312 }, { "epoch": 1.0705429670799487, "grad_norm": 0.2615333153159237, "learning_rate": 1.41e-05, "loss": 0.0219, "step": 313 }, { "epoch": 1.0739632321504917, "grad_norm": 0.2876480036224866, "learning_rate": 1.41e-05, "loss": 0.0253, "step": 314 }, { "epoch": 1.0773834972210345, "grad_norm": 0.30478384773242695, "learning_rate": 1.41e-05, "loss": 0.0252, "step": 315 }, { "epoch": 1.0808037622915776, "grad_norm": 0.2528846190302502, "learning_rate": 1.41e-05, "loss": 0.0206, "step": 316 }, { "epoch": 1.0842240273621206, "grad_norm": 0.2435449168695854, "learning_rate": 1.41e-05, "loss": 0.0224, "step": 317 }, { "epoch": 1.0876442924326635, "grad_norm": 0.3079236522189953, "learning_rate": 1.41e-05, "loss": 0.0253, "step": 318 }, { "epoch": 1.0910645575032065, "grad_norm": 0.2675432502865711, "learning_rate": 1.41e-05, "loss": 0.0239, "step": 319 }, { "epoch": 1.0944848225737496, "grad_norm": 0.3004587131742902, "learning_rate": 1.41e-05, "loss": 0.0239, "step": 320 }, { "epoch": 1.0979050876442924, "grad_norm": 0.29332378399603554, "learning_rate": 1.41e-05, "loss": 0.0247, "step": 321 }, { "epoch": 1.1013253527148354, "grad_norm": 0.23762802516720216, "learning_rate": 1.41e-05, "loss": 0.0242, "step": 322 }, { "epoch": 1.1047456177853783, "grad_norm": 0.26342688417785715, "learning_rate": 1.41e-05, "loss": 0.0244, "step": 323 }, { "epoch": 1.1081658828559213, "grad_norm": 0.3055498089807232, "learning_rate": 1.41e-05, "loss": 0.0265, "step": 324 }, { "epoch": 1.1115861479264644, "grad_norm": 0.23787182811525248, "learning_rate": 1.41e-05, "loss": 0.0226, "step": 325 }, { "epoch": 1.1150064129970072, "grad_norm": 0.2647940637873088, "learning_rate": 1.41e-05, "loss": 0.0242, "step": 326 }, { "epoch": 1.1184266780675503, "grad_norm": 0.2913294201237873, "learning_rate": 1.41e-05, "loss": 0.0261, "step": 327 }, { "epoch": 1.1218469431380933, "grad_norm": 0.2581915579915153, "learning_rate": 1.41e-05, "loss": 0.0228, "step": 328 }, { "epoch": 1.1252672082086361, "grad_norm": 0.2500171384316944, "learning_rate": 1.41e-05, "loss": 0.0206, "step": 329 }, { "epoch": 1.1286874732791792, "grad_norm": 0.2689440388389032, "learning_rate": 1.41e-05, "loss": 0.0237, "step": 330 }, { "epoch": 1.1321077383497222, "grad_norm": 0.2551497009606492, "learning_rate": 1.41e-05, "loss": 0.0216, "step": 331 }, { "epoch": 1.135528003420265, "grad_norm": 0.25757413049163996, "learning_rate": 1.41e-05, "loss": 0.0224, "step": 332 }, { "epoch": 1.138948268490808, "grad_norm": 0.2699506485343775, "learning_rate": 1.41e-05, "loss": 0.0218, "step": 333 }, { "epoch": 1.142368533561351, "grad_norm": 0.25767905743807257, "learning_rate": 1.41e-05, "loss": 0.0227, "step": 334 }, { "epoch": 1.145788798631894, "grad_norm": 0.2273460686067317, "learning_rate": 1.41e-05, "loss": 0.0226, "step": 335 }, { "epoch": 1.149209063702437, "grad_norm": 0.31200680170359746, "learning_rate": 1.41e-05, "loss": 0.0196, "step": 336 }, { "epoch": 1.1526293287729799, "grad_norm": 0.23994204361556526, "learning_rate": 1.41e-05, "loss": 0.022, "step": 337 }, { "epoch": 1.156049593843523, "grad_norm": 0.2867059420441838, "learning_rate": 1.41e-05, "loss": 0.0256, "step": 338 }, { "epoch": 1.1594698589140657, "grad_norm": 0.2523475692407159, "learning_rate": 1.41e-05, "loss": 0.0212, "step": 339 }, { "epoch": 1.1628901239846088, "grad_norm": 0.24499307458764866, "learning_rate": 1.41e-05, "loss": 0.0198, "step": 340 }, { "epoch": 1.1663103890551518, "grad_norm": 0.2436283905820348, "learning_rate": 1.41e-05, "loss": 0.0216, "step": 341 }, { "epoch": 1.1697306541256947, "grad_norm": 0.25602523324809817, "learning_rate": 1.41e-05, "loss": 0.0216, "step": 342 }, { "epoch": 1.1731509191962377, "grad_norm": 0.25021177697750296, "learning_rate": 1.41e-05, "loss": 0.0233, "step": 343 }, { "epoch": 1.1765711842667808, "grad_norm": 0.28379855080073285, "learning_rate": 1.41e-05, "loss": 0.0259, "step": 344 }, { "epoch": 1.1799914493373236, "grad_norm": 0.25642024168179856, "learning_rate": 1.41e-05, "loss": 0.0208, "step": 345 }, { "epoch": 1.1834117144078666, "grad_norm": 0.24053576849839317, "learning_rate": 1.41e-05, "loss": 0.0206, "step": 346 }, { "epoch": 1.1868319794784097, "grad_norm": 0.26458302891901536, "learning_rate": 1.41e-05, "loss": 0.025, "step": 347 }, { "epoch": 1.1902522445489525, "grad_norm": 0.25340954518832737, "learning_rate": 1.41e-05, "loss": 0.0232, "step": 348 }, { "epoch": 1.1936725096194956, "grad_norm": 0.2822910057628071, "learning_rate": 1.41e-05, "loss": 0.0258, "step": 349 }, { "epoch": 1.1970927746900384, "grad_norm": 0.24401825117116366, "learning_rate": 1.41e-05, "loss": 0.0207, "step": 350 }, { "epoch": 1.2005130397605814, "grad_norm": 0.25351547147013626, "learning_rate": 1.41e-05, "loss": 0.0232, "step": 351 }, { "epoch": 1.2039333048311245, "grad_norm": 0.2514277079230116, "learning_rate": 1.41e-05, "loss": 0.0236, "step": 352 }, { "epoch": 1.2073535699016673, "grad_norm": 0.2499624840027022, "learning_rate": 1.41e-05, "loss": 0.0218, "step": 353 }, { "epoch": 1.2107738349722104, "grad_norm": 0.25541034452698597, "learning_rate": 1.41e-05, "loss": 0.0226, "step": 354 }, { "epoch": 1.2141941000427534, "grad_norm": 0.26116617349172355, "learning_rate": 1.41e-05, "loss": 0.0233, "step": 355 }, { "epoch": 1.2176143651132962, "grad_norm": 0.3144220810007225, "learning_rate": 1.41e-05, "loss": 0.0238, "step": 356 }, { "epoch": 1.2210346301838393, "grad_norm": 0.2397648420075889, "learning_rate": 1.41e-05, "loss": 0.0218, "step": 357 }, { "epoch": 1.2244548952543823, "grad_norm": 0.23056631938102404, "learning_rate": 1.41e-05, "loss": 0.0211, "step": 358 }, { "epoch": 1.2278751603249252, "grad_norm": 0.6359133071495683, "learning_rate": 1.41e-05, "loss": 0.0221, "step": 359 }, { "epoch": 1.2312954253954682, "grad_norm": 0.2494306454740628, "learning_rate": 1.41e-05, "loss": 0.0212, "step": 360 }, { "epoch": 1.234715690466011, "grad_norm": 0.2661512294789303, "learning_rate": 1.41e-05, "loss": 0.0254, "step": 361 }, { "epoch": 1.238135955536554, "grad_norm": 0.7710920857502603, "learning_rate": 1.41e-05, "loss": 0.025, "step": 362 }, { "epoch": 1.2415562206070971, "grad_norm": 0.3942637887185519, "learning_rate": 1.41e-05, "loss": 0.0236, "step": 363 }, { "epoch": 1.24497648567764, "grad_norm": 0.27087010860427124, "learning_rate": 1.41e-05, "loss": 0.0252, "step": 364 }, { "epoch": 1.248396750748183, "grad_norm": 0.24919842057781497, "learning_rate": 1.41e-05, "loss": 0.022, "step": 365 }, { "epoch": 1.2518170158187258, "grad_norm": 0.247824008548709, "learning_rate": 1.41e-05, "loss": 0.0235, "step": 366 }, { "epoch": 1.255237280889269, "grad_norm": 0.2113113650016047, "learning_rate": 1.41e-05, "loss": 0.0202, "step": 367 }, { "epoch": 1.258657545959812, "grad_norm": 0.2259169472785117, "learning_rate": 1.41e-05, "loss": 0.0192, "step": 368 }, { "epoch": 1.262077811030355, "grad_norm": 0.2752050707801685, "learning_rate": 1.41e-05, "loss": 0.0254, "step": 369 }, { "epoch": 1.2654980761008978, "grad_norm": 0.23830392409535447, "learning_rate": 1.41e-05, "loss": 0.0229, "step": 370 }, { "epoch": 1.2689183411714409, "grad_norm": 0.24302497993054667, "learning_rate": 1.41e-05, "loss": 0.0219, "step": 371 }, { "epoch": 1.2723386062419837, "grad_norm": 0.24504347526605771, "learning_rate": 1.41e-05, "loss": 0.0224, "step": 372 }, { "epoch": 1.2757588713125267, "grad_norm": 0.2346626865644856, "learning_rate": 1.41e-05, "loss": 0.0209, "step": 373 }, { "epoch": 1.2791791363830698, "grad_norm": 0.24245589828220748, "learning_rate": 1.41e-05, "loss": 0.0218, "step": 374 }, { "epoch": 1.2825994014536126, "grad_norm": 0.282140360657227, "learning_rate": 1.41e-05, "loss": 0.0234, "step": 375 }, { "epoch": 1.2860196665241557, "grad_norm": 0.26692130820889887, "learning_rate": 1.41e-05, "loss": 0.02, "step": 376 }, { "epoch": 1.2894399315946985, "grad_norm": 0.2691231899902035, "learning_rate": 1.41e-05, "loss": 0.0221, "step": 377 }, { "epoch": 1.2928601966652415, "grad_norm": 0.2281687474950334, "learning_rate": 1.41e-05, "loss": 0.0222, "step": 378 }, { "epoch": 1.2962804617357846, "grad_norm": 0.373488505328009, "learning_rate": 1.41e-05, "loss": 0.0217, "step": 379 }, { "epoch": 1.2997007268063274, "grad_norm": 0.2345889821941022, "learning_rate": 1.41e-05, "loss": 0.021, "step": 380 }, { "epoch": 1.3031209918768705, "grad_norm": 0.2105687923684266, "learning_rate": 1.41e-05, "loss": 0.0204, "step": 381 }, { "epoch": 1.3065412569474133, "grad_norm": 0.23464935916200427, "learning_rate": 1.41e-05, "loss": 0.0213, "step": 382 }, { "epoch": 1.3099615220179563, "grad_norm": 0.23646699592209017, "learning_rate": 1.41e-05, "loss": 0.0211, "step": 383 }, { "epoch": 1.3133817870884994, "grad_norm": 0.2631733185003215, "learning_rate": 1.41e-05, "loss": 0.0219, "step": 384 }, { "epoch": 1.3168020521590424, "grad_norm": 0.25931322588284866, "learning_rate": 1.41e-05, "loss": 0.0229, "step": 385 }, { "epoch": 1.3202223172295853, "grad_norm": 0.28688530125524797, "learning_rate": 1.41e-05, "loss": 0.0226, "step": 386 }, { "epoch": 1.3236425823001283, "grad_norm": 0.25729658498109925, "learning_rate": 1.41e-05, "loss": 0.0229, "step": 387 }, { "epoch": 1.3270628473706712, "grad_norm": 0.25760420815172724, "learning_rate": 1.41e-05, "loss": 0.0207, "step": 388 }, { "epoch": 1.3304831124412142, "grad_norm": 0.32657940169239236, "learning_rate": 1.41e-05, "loss": 0.0217, "step": 389 }, { "epoch": 1.3339033775117572, "grad_norm": 0.22557643602886415, "learning_rate": 1.41e-05, "loss": 0.0204, "step": 390 }, { "epoch": 1.3373236425823, "grad_norm": 0.2661843657709089, "learning_rate": 1.41e-05, "loss": 0.0243, "step": 391 }, { "epoch": 1.3407439076528431, "grad_norm": 0.2974504589558172, "learning_rate": 1.41e-05, "loss": 0.0241, "step": 392 }, { "epoch": 1.344164172723386, "grad_norm": 0.26843823846639436, "learning_rate": 1.41e-05, "loss": 0.0234, "step": 393 }, { "epoch": 1.347584437793929, "grad_norm": 0.3218845205293079, "learning_rate": 1.41e-05, "loss": 0.0221, "step": 394 }, { "epoch": 1.351004702864472, "grad_norm": 0.2428094674581459, "learning_rate": 1.41e-05, "loss": 0.0222, "step": 395 }, { "epoch": 1.354424967935015, "grad_norm": 0.2455927243682846, "learning_rate": 1.41e-05, "loss": 0.0211, "step": 396 }, { "epoch": 1.357845233005558, "grad_norm": 0.2733123552071639, "learning_rate": 1.41e-05, "loss": 0.0212, "step": 397 }, { "epoch": 1.361265498076101, "grad_norm": 0.2659612332359629, "learning_rate": 1.41e-05, "loss": 0.0203, "step": 398 }, { "epoch": 1.3646857631466438, "grad_norm": 0.25254417828478626, "learning_rate": 1.41e-05, "loss": 0.0201, "step": 399 }, { "epoch": 1.3681060282171869, "grad_norm": 0.2598602588125554, "learning_rate": 1.41e-05, "loss": 0.0217, "step": 400 }, { "epoch": 1.37152629328773, "grad_norm": 0.25849620548333013, "learning_rate": 1.41e-05, "loss": 0.0209, "step": 401 }, { "epoch": 1.3749465583582727, "grad_norm": 0.3137023430125288, "learning_rate": 1.41e-05, "loss": 0.0211, "step": 402 }, { "epoch": 1.3783668234288158, "grad_norm": 0.2113162287336589, "learning_rate": 1.41e-05, "loss": 0.0195, "step": 403 }, { "epoch": 1.3817870884993586, "grad_norm": 0.2503171178420045, "learning_rate": 1.41e-05, "loss": 0.0234, "step": 404 }, { "epoch": 1.3852073535699017, "grad_norm": 0.23119898543579737, "learning_rate": 1.41e-05, "loss": 0.0209, "step": 405 }, { "epoch": 1.3886276186404447, "grad_norm": 0.26039590115583117, "learning_rate": 1.41e-05, "loss": 0.0244, "step": 406 }, { "epoch": 1.3920478837109875, "grad_norm": 0.23886269607593336, "learning_rate": 1.41e-05, "loss": 0.0223, "step": 407 }, { "epoch": 1.3954681487815306, "grad_norm": 0.2714710828662534, "learning_rate": 1.41e-05, "loss": 0.0247, "step": 408 }, { "epoch": 1.3988884138520734, "grad_norm": 0.27319555229782644, "learning_rate": 1.41e-05, "loss": 0.0244, "step": 409 }, { "epoch": 1.4023086789226165, "grad_norm": 0.24019773624911636, "learning_rate": 1.41e-05, "loss": 0.0218, "step": 410 }, { "epoch": 1.4057289439931595, "grad_norm": 0.2708671308101268, "learning_rate": 1.41e-05, "loss": 0.0219, "step": 411 }, { "epoch": 1.4091492090637026, "grad_norm": 0.22702757974948617, "learning_rate": 1.41e-05, "loss": 0.0198, "step": 412 }, { "epoch": 1.4125694741342454, "grad_norm": 0.25555770375998416, "learning_rate": 1.41e-05, "loss": 0.0224, "step": 413 }, { "epoch": 1.4159897392047884, "grad_norm": 0.23780595691689027, "learning_rate": 1.41e-05, "loss": 0.0213, "step": 414 }, { "epoch": 1.4194100042753313, "grad_norm": 0.26451555701259444, "learning_rate": 1.41e-05, "loss": 0.0211, "step": 415 }, { "epoch": 1.4228302693458743, "grad_norm": 0.24436230077463678, "learning_rate": 1.41e-05, "loss": 0.0226, "step": 416 }, { "epoch": 1.4262505344164174, "grad_norm": 0.2216762882841438, "learning_rate": 1.41e-05, "loss": 0.021, "step": 417 }, { "epoch": 1.4296707994869602, "grad_norm": 0.26405628674467824, "learning_rate": 1.41e-05, "loss": 0.0219, "step": 418 }, { "epoch": 1.4330910645575032, "grad_norm": 0.23921814427898216, "learning_rate": 1.41e-05, "loss": 0.021, "step": 419 }, { "epoch": 1.436511329628046, "grad_norm": 0.265453432302215, "learning_rate": 1.41e-05, "loss": 0.0226, "step": 420 }, { "epoch": 1.4399315946985891, "grad_norm": 0.24492268359315303, "learning_rate": 1.41e-05, "loss": 0.0227, "step": 421 }, { "epoch": 1.4433518597691322, "grad_norm": 0.25191935351384814, "learning_rate": 1.41e-05, "loss": 0.0215, "step": 422 }, { "epoch": 1.4467721248396752, "grad_norm": 0.27197993846345997, "learning_rate": 1.41e-05, "loss": 0.0229, "step": 423 }, { "epoch": 1.450192389910218, "grad_norm": 0.2075718815542033, "learning_rate": 1.41e-05, "loss": 0.0183, "step": 424 }, { "epoch": 1.453612654980761, "grad_norm": 0.26376818444878, "learning_rate": 1.41e-05, "loss": 0.0221, "step": 425 }, { "epoch": 1.457032920051304, "grad_norm": 0.26080564628665626, "learning_rate": 1.41e-05, "loss": 0.0229, "step": 426 }, { "epoch": 1.460453185121847, "grad_norm": 0.2434588148667514, "learning_rate": 1.41e-05, "loss": 0.022, "step": 427 }, { "epoch": 1.46387345019239, "grad_norm": 0.20943782903685337, "learning_rate": 1.41e-05, "loss": 0.0185, "step": 428 }, { "epoch": 1.4672937152629328, "grad_norm": 0.24301387125104526, "learning_rate": 1.41e-05, "loss": 0.0208, "step": 429 }, { "epoch": 1.470713980333476, "grad_norm": 0.25766203422027834, "learning_rate": 1.41e-05, "loss": 0.0237, "step": 430 }, { "epoch": 1.4741342454040187, "grad_norm": 0.21556702000374744, "learning_rate": 1.41e-05, "loss": 0.0213, "step": 431 }, { "epoch": 1.4775545104745618, "grad_norm": 0.23784909822124217, "learning_rate": 1.41e-05, "loss": 0.0207, "step": 432 }, { "epoch": 1.4809747755451048, "grad_norm": 0.22414128247781562, "learning_rate": 1.41e-05, "loss": 0.0198, "step": 433 }, { "epoch": 1.4843950406156476, "grad_norm": 0.23386987857579358, "learning_rate": 1.41e-05, "loss": 0.0201, "step": 434 }, { "epoch": 1.4878153056861907, "grad_norm": 0.2853961947266083, "learning_rate": 1.41e-05, "loss": 0.023, "step": 435 }, { "epoch": 1.4912355707567335, "grad_norm": 0.24974092416116495, "learning_rate": 1.41e-05, "loss": 0.0209, "step": 436 }, { "epoch": 1.4946558358272766, "grad_norm": 0.23615654630506394, "learning_rate": 1.41e-05, "loss": 0.022, "step": 437 }, { "epoch": 1.4980761008978196, "grad_norm": 0.2284456001474283, "learning_rate": 1.41e-05, "loss": 0.0195, "step": 438 }, { "epoch": 1.5014963659683627, "grad_norm": 0.26063959990204033, "learning_rate": 1.41e-05, "loss": 0.0258, "step": 439 }, { "epoch": 1.5049166310389055, "grad_norm": 0.24423575575966894, "learning_rate": 1.41e-05, "loss": 0.0215, "step": 440 }, { "epoch": 1.5083368961094483, "grad_norm": 0.24270403980054853, "learning_rate": 1.41e-05, "loss": 0.0198, "step": 441 }, { "epoch": 1.5117571611799914, "grad_norm": 0.24660594248828135, "learning_rate": 1.41e-05, "loss": 0.0208, "step": 442 }, { "epoch": 1.5151774262505344, "grad_norm": 0.23764056229245814, "learning_rate": 1.41e-05, "loss": 0.0185, "step": 443 }, { "epoch": 1.5185976913210775, "grad_norm": 0.2327212499624357, "learning_rate": 1.41e-05, "loss": 0.0205, "step": 444 }, { "epoch": 1.5220179563916203, "grad_norm": 0.2644477025039793, "learning_rate": 1.41e-05, "loss": 0.0222, "step": 445 }, { "epoch": 1.5254382214621633, "grad_norm": 0.2385202884230959, "learning_rate": 1.41e-05, "loss": 0.023, "step": 446 }, { "epoch": 1.5288584865327062, "grad_norm": 0.24308327876991292, "learning_rate": 1.41e-05, "loss": 0.0198, "step": 447 }, { "epoch": 1.5322787516032492, "grad_norm": 0.30634205660699, "learning_rate": 1.41e-05, "loss": 0.026, "step": 448 }, { "epoch": 1.5356990166737923, "grad_norm": 0.21217473130594544, "learning_rate": 1.41e-05, "loss": 0.0199, "step": 449 }, { "epoch": 1.5391192817443353, "grad_norm": 0.2658497297922587, "learning_rate": 1.41e-05, "loss": 0.0221, "step": 450 }, { "epoch": 1.5425395468148781, "grad_norm": 0.21481338352800414, "learning_rate": 1.41e-05, "loss": 0.0197, "step": 451 }, { "epoch": 1.545959811885421, "grad_norm": 0.27821761611964035, "learning_rate": 1.41e-05, "loss": 0.0246, "step": 452 }, { "epoch": 1.549380076955964, "grad_norm": 0.2416104328737188, "learning_rate": 1.41e-05, "loss": 0.0216, "step": 453 }, { "epoch": 1.552800342026507, "grad_norm": 0.2322000952546936, "learning_rate": 1.41e-05, "loss": 0.0219, "step": 454 }, { "epoch": 1.5562206070970501, "grad_norm": 0.23752597721176905, "learning_rate": 1.41e-05, "loss": 0.0218, "step": 455 }, { "epoch": 1.559640872167593, "grad_norm": 0.24308738119455534, "learning_rate": 1.41e-05, "loss": 0.0205, "step": 456 }, { "epoch": 1.563061137238136, "grad_norm": 0.26073824102297316, "learning_rate": 1.41e-05, "loss": 0.0247, "step": 457 }, { "epoch": 1.5664814023086788, "grad_norm": 0.24050011455971732, "learning_rate": 1.41e-05, "loss": 0.0223, "step": 458 }, { "epoch": 1.5699016673792219, "grad_norm": 0.21610207139710053, "learning_rate": 1.41e-05, "loss": 0.0192, "step": 459 }, { "epoch": 1.573321932449765, "grad_norm": 0.25801793315676924, "learning_rate": 1.41e-05, "loss": 0.0236, "step": 460 }, { "epoch": 1.576742197520308, "grad_norm": 0.21228131497792052, "learning_rate": 1.41e-05, "loss": 0.0191, "step": 461 }, { "epoch": 1.5801624625908508, "grad_norm": 0.25197778439538576, "learning_rate": 1.41e-05, "loss": 0.022, "step": 462 }, { "epoch": 1.5835827276613936, "grad_norm": 0.2565258961786049, "learning_rate": 1.41e-05, "loss": 0.0219, "step": 463 }, { "epoch": 1.5870029927319367, "grad_norm": 0.2559923791329211, "learning_rate": 1.41e-05, "loss": 0.0217, "step": 464 }, { "epoch": 1.5904232578024797, "grad_norm": 0.2535527705987336, "learning_rate": 1.41e-05, "loss": 0.0221, "step": 465 }, { "epoch": 1.5938435228730228, "grad_norm": 0.28495146888910583, "learning_rate": 1.41e-05, "loss": 0.0272, "step": 466 }, { "epoch": 1.5972637879435656, "grad_norm": 0.26151357261060504, "learning_rate": 1.41e-05, "loss": 0.022, "step": 467 }, { "epoch": 1.6006840530141084, "grad_norm": 0.24889528453574525, "learning_rate": 1.41e-05, "loss": 0.0218, "step": 468 }, { "epoch": 1.6041043180846515, "grad_norm": 0.21219419933729458, "learning_rate": 1.41e-05, "loss": 0.0176, "step": 469 }, { "epoch": 1.6075245831551945, "grad_norm": 0.24995940249087767, "learning_rate": 1.41e-05, "loss": 0.0207, "step": 470 }, { "epoch": 1.6109448482257376, "grad_norm": 0.23890689663178338, "learning_rate": 1.41e-05, "loss": 0.0204, "step": 471 }, { "epoch": 1.6143651132962804, "grad_norm": 0.23090495982296502, "learning_rate": 1.41e-05, "loss": 0.0209, "step": 472 }, { "epoch": 1.6177853783668235, "grad_norm": 0.24984846656221682, "learning_rate": 1.41e-05, "loss": 0.0191, "step": 473 }, { "epoch": 1.6212056434373663, "grad_norm": 0.2864149291763989, "learning_rate": 1.41e-05, "loss": 0.0243, "step": 474 }, { "epoch": 1.6246259085079093, "grad_norm": 0.23371681508272213, "learning_rate": 1.41e-05, "loss": 0.0197, "step": 475 }, { "epoch": 1.6280461735784524, "grad_norm": 0.21807560914421245, "learning_rate": 1.41e-05, "loss": 0.019, "step": 476 }, { "epoch": 1.6314664386489954, "grad_norm": 0.24325192491721911, "learning_rate": 1.41e-05, "loss": 0.0212, "step": 477 }, { "epoch": 1.6348867037195383, "grad_norm": 0.21802834032775958, "learning_rate": 1.41e-05, "loss": 0.0183, "step": 478 }, { "epoch": 1.638306968790081, "grad_norm": 0.25015961868430003, "learning_rate": 1.41e-05, "loss": 0.0206, "step": 479 }, { "epoch": 1.6417272338606241, "grad_norm": 0.2047993695964016, "learning_rate": 1.41e-05, "loss": 0.0166, "step": 480 }, { "epoch": 1.6451474989311672, "grad_norm": 0.24407449809464615, "learning_rate": 1.41e-05, "loss": 0.0209, "step": 481 }, { "epoch": 1.6485677640017102, "grad_norm": 0.22905836500660803, "learning_rate": 1.41e-05, "loss": 0.0197, "step": 482 }, { "epoch": 1.651988029072253, "grad_norm": 0.2302659533712275, "learning_rate": 1.41e-05, "loss": 0.0184, "step": 483 }, { "epoch": 1.655408294142796, "grad_norm": 0.22469778783055896, "learning_rate": 1.41e-05, "loss": 0.022, "step": 484 }, { "epoch": 1.658828559213339, "grad_norm": 0.2403880470710616, "learning_rate": 1.41e-05, "loss": 0.0184, "step": 485 }, { "epoch": 1.662248824283882, "grad_norm": 0.21080166415704205, "learning_rate": 1.41e-05, "loss": 0.0194, "step": 486 }, { "epoch": 1.665669089354425, "grad_norm": 0.24366180394977457, "learning_rate": 1.41e-05, "loss": 0.0212, "step": 487 }, { "epoch": 1.669089354424968, "grad_norm": 0.2445725370169137, "learning_rate": 1.41e-05, "loss": 0.0201, "step": 488 }, { "epoch": 1.672509619495511, "grad_norm": 0.2400419596996237, "learning_rate": 1.41e-05, "loss": 0.0217, "step": 489 }, { "epoch": 1.6759298845660537, "grad_norm": 0.22845732992118456, "learning_rate": 1.41e-05, "loss": 0.0193, "step": 490 }, { "epoch": 1.6793501496365968, "grad_norm": 0.21963101467880122, "learning_rate": 1.41e-05, "loss": 0.0205, "step": 491 }, { "epoch": 1.6827704147071398, "grad_norm": 0.23832272298065416, "learning_rate": 1.41e-05, "loss": 0.0208, "step": 492 }, { "epoch": 1.6861906797776829, "grad_norm": 0.2701770424720032, "learning_rate": 1.41e-05, "loss": 0.0245, "step": 493 }, { "epoch": 1.6896109448482257, "grad_norm": 0.21999352850800136, "learning_rate": 1.41e-05, "loss": 0.0191, "step": 494 }, { "epoch": 1.6930312099187685, "grad_norm": 0.2303202011396172, "learning_rate": 1.41e-05, "loss": 0.0206, "step": 495 }, { "epoch": 1.6964514749893116, "grad_norm": 0.23207638496444852, "learning_rate": 1.41e-05, "loss": 0.0196, "step": 496 }, { "epoch": 1.6998717400598546, "grad_norm": 0.27476502715987833, "learning_rate": 1.41e-05, "loss": 0.0243, "step": 497 }, { "epoch": 1.7032920051303977, "grad_norm": 0.25943127784425596, "learning_rate": 1.41e-05, "loss": 0.0209, "step": 498 }, { "epoch": 1.7067122702009405, "grad_norm": 0.2582673258172837, "learning_rate": 1.41e-05, "loss": 0.0213, "step": 499 }, { "epoch": 1.7101325352714836, "grad_norm": 0.2270336558731971, "learning_rate": 1.41e-05, "loss": 0.0207, "step": 500 }, { "epoch": 1.7135528003420264, "grad_norm": 0.21716332073489333, "learning_rate": 1.41e-05, "loss": 0.017, "step": 501 }, { "epoch": 1.7169730654125694, "grad_norm": 0.22361076006822656, "learning_rate": 1.41e-05, "loss": 0.0198, "step": 502 }, { "epoch": 1.7203933304831125, "grad_norm": 0.25453906115757163, "learning_rate": 1.41e-05, "loss": 0.0215, "step": 503 }, { "epoch": 1.7238135955536555, "grad_norm": 0.2639274732447762, "learning_rate": 1.41e-05, "loss": 0.0211, "step": 504 }, { "epoch": 1.7272338606241984, "grad_norm": 0.25322752853555974, "learning_rate": 1.41e-05, "loss": 0.0228, "step": 505 }, { "epoch": 1.7306541256947412, "grad_norm": 0.23285923741403275, "learning_rate": 1.41e-05, "loss": 0.0205, "step": 506 }, { "epoch": 1.7340743907652842, "grad_norm": 0.2323064412599781, "learning_rate": 1.41e-05, "loss": 0.0195, "step": 507 }, { "epoch": 1.7374946558358273, "grad_norm": 0.23653121619208753, "learning_rate": 1.41e-05, "loss": 0.0211, "step": 508 }, { "epoch": 1.7409149209063703, "grad_norm": 0.2576532470063006, "learning_rate": 1.41e-05, "loss": 0.0216, "step": 509 }, { "epoch": 1.7443351859769132, "grad_norm": 0.3031982152602418, "learning_rate": 1.41e-05, "loss": 0.0212, "step": 510 }, { "epoch": 1.7477554510474562, "grad_norm": 0.21353603367638405, "learning_rate": 1.41e-05, "loss": 0.0187, "step": 511 }, { "epoch": 1.751175716117999, "grad_norm": 0.23897983103928835, "learning_rate": 1.41e-05, "loss": 0.0211, "step": 512 }, { "epoch": 1.754595981188542, "grad_norm": 0.26759279857241974, "learning_rate": 1.41e-05, "loss": 0.0225, "step": 513 }, { "epoch": 1.7580162462590851, "grad_norm": 0.21204297476027956, "learning_rate": 1.41e-05, "loss": 0.0182, "step": 514 }, { "epoch": 1.7614365113296282, "grad_norm": 0.2522156014013123, "learning_rate": 1.41e-05, "loss": 0.0224, "step": 515 }, { "epoch": 1.764856776400171, "grad_norm": 0.2483681502861828, "learning_rate": 1.41e-05, "loss": 0.0225, "step": 516 }, { "epoch": 1.7682770414707139, "grad_norm": 0.2757743080951287, "learning_rate": 1.41e-05, "loss": 0.0252, "step": 517 }, { "epoch": 1.771697306541257, "grad_norm": 0.20783407363518988, "learning_rate": 1.41e-05, "loss": 0.0189, "step": 518 }, { "epoch": 1.7751175716118, "grad_norm": 0.25743244499425244, "learning_rate": 1.41e-05, "loss": 0.022, "step": 519 }, { "epoch": 1.778537836682343, "grad_norm": 0.23017428673046214, "learning_rate": 1.41e-05, "loss": 0.0195, "step": 520 }, { "epoch": 1.7819581017528858, "grad_norm": 0.2315895066049246, "learning_rate": 1.41e-05, "loss": 0.0178, "step": 521 }, { "epoch": 1.7853783668234287, "grad_norm": 0.22782043412403868, "learning_rate": 1.41e-05, "loss": 0.0215, "step": 522 }, { "epoch": 1.7887986318939717, "grad_norm": 0.2724892881429743, "learning_rate": 1.41e-05, "loss": 0.0224, "step": 523 }, { "epoch": 1.7922188969645148, "grad_norm": 0.28330092514107247, "learning_rate": 1.41e-05, "loss": 0.02, "step": 524 }, { "epoch": 1.7956391620350578, "grad_norm": 0.23561517254595046, "learning_rate": 1.41e-05, "loss": 0.0199, "step": 525 }, { "epoch": 1.7990594271056009, "grad_norm": 0.27080028676550905, "learning_rate": 1.41e-05, "loss": 0.0213, "step": 526 }, { "epoch": 1.8024796921761437, "grad_norm": 0.20908394348220563, "learning_rate": 1.41e-05, "loss": 0.0181, "step": 527 }, { "epoch": 1.8058999572466865, "grad_norm": 0.24764828335378633, "learning_rate": 1.41e-05, "loss": 0.0208, "step": 528 }, { "epoch": 1.8093202223172296, "grad_norm": 0.29733366903532993, "learning_rate": 1.41e-05, "loss": 0.0218, "step": 529 }, { "epoch": 1.8127404873877726, "grad_norm": 0.2708827452564214, "learning_rate": 1.41e-05, "loss": 0.0226, "step": 530 }, { "epoch": 1.8161607524583157, "grad_norm": 0.2887461777468973, "learning_rate": 1.41e-05, "loss": 0.0226, "step": 531 }, { "epoch": 1.8195810175288585, "grad_norm": 0.2774042332926017, "learning_rate": 1.41e-05, "loss": 0.0246, "step": 532 }, { "epoch": 1.8230012825994013, "grad_norm": 0.23877829686783267, "learning_rate": 1.41e-05, "loss": 0.0202, "step": 533 }, { "epoch": 1.8264215476699444, "grad_norm": 0.2544233995482896, "learning_rate": 1.41e-05, "loss": 0.0204, "step": 534 }, { "epoch": 1.8298418127404874, "grad_norm": 0.26365092898440745, "learning_rate": 1.41e-05, "loss": 0.0223, "step": 535 }, { "epoch": 1.8332620778110305, "grad_norm": 0.21178009134118125, "learning_rate": 1.41e-05, "loss": 0.018, "step": 536 }, { "epoch": 1.8366823428815733, "grad_norm": 0.27115664081831004, "learning_rate": 1.41e-05, "loss": 0.0211, "step": 537 }, { "epoch": 1.8401026079521163, "grad_norm": 0.2653048117342068, "learning_rate": 1.41e-05, "loss": 0.0212, "step": 538 }, { "epoch": 1.8435228730226592, "grad_norm": 0.2757728122775812, "learning_rate": 1.41e-05, "loss": 0.0204, "step": 539 }, { "epoch": 1.8469431380932022, "grad_norm": 0.2664986190346045, "learning_rate": 1.41e-05, "loss": 0.0207, "step": 540 }, { "epoch": 1.8503634031637453, "grad_norm": 0.265518302380727, "learning_rate": 1.41e-05, "loss": 0.0179, "step": 541 }, { "epoch": 1.8537836682342883, "grad_norm": 0.2895623799588168, "learning_rate": 1.41e-05, "loss": 0.0228, "step": 542 }, { "epoch": 1.8572039333048311, "grad_norm": 0.23284728474850006, "learning_rate": 1.41e-05, "loss": 0.0192, "step": 543 }, { "epoch": 1.860624198375374, "grad_norm": 0.21864913411739378, "learning_rate": 1.41e-05, "loss": 0.0178, "step": 544 }, { "epoch": 1.864044463445917, "grad_norm": 0.25369925309969443, "learning_rate": 1.41e-05, "loss": 0.0204, "step": 545 }, { "epoch": 1.86746472851646, "grad_norm": 0.22162789215208448, "learning_rate": 1.41e-05, "loss": 0.0214, "step": 546 }, { "epoch": 1.870884993587003, "grad_norm": 0.228575157771765, "learning_rate": 1.41e-05, "loss": 0.0215, "step": 547 }, { "epoch": 1.874305258657546, "grad_norm": 0.29177202889082, "learning_rate": 1.41e-05, "loss": 0.0231, "step": 548 }, { "epoch": 1.8777255237280888, "grad_norm": 0.2526426700464145, "learning_rate": 1.41e-05, "loss": 0.022, "step": 549 }, { "epoch": 1.8811457887986318, "grad_norm": 0.25824463173696005, "learning_rate": 1.41e-05, "loss": 0.0238, "step": 550 }, { "epoch": 1.8845660538691749, "grad_norm": 0.22931141277969364, "learning_rate": 1.41e-05, "loss": 0.0195, "step": 551 }, { "epoch": 1.887986318939718, "grad_norm": 0.20128540781283585, "learning_rate": 1.41e-05, "loss": 0.018, "step": 552 }, { "epoch": 1.891406584010261, "grad_norm": 0.2726453682164885, "learning_rate": 1.41e-05, "loss": 0.0215, "step": 553 }, { "epoch": 1.8948268490808038, "grad_norm": 0.2379175413754306, "learning_rate": 1.41e-05, "loss": 0.02, "step": 554 }, { "epoch": 1.8982471141513466, "grad_norm": 0.2657340317097815, "learning_rate": 1.41e-05, "loss": 0.0215, "step": 555 }, { "epoch": 1.9016673792218897, "grad_norm": 0.2469655254797632, "learning_rate": 1.41e-05, "loss": 0.0218, "step": 556 }, { "epoch": 1.9050876442924327, "grad_norm": 0.2306898046638619, "learning_rate": 1.41e-05, "loss": 0.0206, "step": 557 }, { "epoch": 1.9085079093629758, "grad_norm": 0.2501966986368517, "learning_rate": 1.41e-05, "loss": 0.0208, "step": 558 }, { "epoch": 1.9119281744335186, "grad_norm": 0.26337448659750273, "learning_rate": 1.41e-05, "loss": 0.0224, "step": 559 }, { "epoch": 1.9153484395040614, "grad_norm": 0.23788126235745702, "learning_rate": 1.41e-05, "loss": 0.0207, "step": 560 }, { "epoch": 1.9187687045746045, "grad_norm": 0.24400460810027635, "learning_rate": 1.41e-05, "loss": 0.0199, "step": 561 }, { "epoch": 1.9221889696451475, "grad_norm": 0.2524032823359455, "learning_rate": 1.41e-05, "loss": 0.02, "step": 562 }, { "epoch": 1.9256092347156906, "grad_norm": 0.2288025108079512, "learning_rate": 1.41e-05, "loss": 0.0201, "step": 563 }, { "epoch": 1.9290294997862334, "grad_norm": 0.269452001004884, "learning_rate": 1.41e-05, "loss": 0.0214, "step": 564 }, { "epoch": 1.9324497648567764, "grad_norm": 0.2316242696085615, "learning_rate": 1.41e-05, "loss": 0.0196, "step": 565 }, { "epoch": 1.9358700299273193, "grad_norm": 0.221180291707861, "learning_rate": 1.41e-05, "loss": 0.0177, "step": 566 }, { "epoch": 1.9392902949978623, "grad_norm": 0.2607363855888593, "learning_rate": 1.41e-05, "loss": 0.0192, "step": 567 }, { "epoch": 1.9427105600684054, "grad_norm": 0.24514765767262625, "learning_rate": 1.41e-05, "loss": 0.0182, "step": 568 }, { "epoch": 1.9461308251389484, "grad_norm": 0.28916534222959817, "learning_rate": 1.41e-05, "loss": 0.0207, "step": 569 }, { "epoch": 1.9495510902094912, "grad_norm": 0.26414654167917073, "learning_rate": 1.41e-05, "loss": 0.0204, "step": 570 }, { "epoch": 1.952971355280034, "grad_norm": 0.2482098982197082, "learning_rate": 1.41e-05, "loss": 0.0206, "step": 571 }, { "epoch": 1.9563916203505771, "grad_norm": 0.2608562249261069, "learning_rate": 1.41e-05, "loss": 0.0208, "step": 572 }, { "epoch": 1.9598118854211202, "grad_norm": 0.25525358687825855, "learning_rate": 1.41e-05, "loss": 0.0213, "step": 573 }, { "epoch": 1.9632321504916632, "grad_norm": 0.23125518036524792, "learning_rate": 1.41e-05, "loss": 0.0212, "step": 574 }, { "epoch": 1.966652415562206, "grad_norm": 0.2432676298521834, "learning_rate": 1.41e-05, "loss": 0.0187, "step": 575 }, { "epoch": 1.9700726806327489, "grad_norm": 0.24502134522738972, "learning_rate": 1.41e-05, "loss": 0.0194, "step": 576 }, { "epoch": 1.973492945703292, "grad_norm": 0.26553777081307495, "learning_rate": 1.41e-05, "loss": 0.0212, "step": 577 }, { "epoch": 1.976913210773835, "grad_norm": 0.24919535132544937, "learning_rate": 1.41e-05, "loss": 0.0208, "step": 578 }, { "epoch": 1.980333475844378, "grad_norm": 0.2828767110434045, "learning_rate": 1.41e-05, "loss": 0.0238, "step": 579 }, { "epoch": 1.983753740914921, "grad_norm": 0.26780635861278174, "learning_rate": 1.41e-05, "loss": 0.0225, "step": 580 }, { "epoch": 1.987174005985464, "grad_norm": 0.26755832407143076, "learning_rate": 1.41e-05, "loss": 0.0197, "step": 581 }, { "epoch": 1.9905942710560067, "grad_norm": 0.23150201894075045, "learning_rate": 1.41e-05, "loss": 0.0225, "step": 582 }, { "epoch": 1.9940145361265498, "grad_norm": 0.2556938828383787, "learning_rate": 1.41e-05, "loss": 0.0217, "step": 583 }, { "epoch": 1.9974348011970928, "grad_norm": 0.24779362245147543, "learning_rate": 1.41e-05, "loss": 0.0209, "step": 584 }, { "epoch": 2.000855066267636, "grad_norm": 0.23220294742842815, "learning_rate": 1.41e-05, "loss": 0.0205, "step": 585 }, { "epoch": 2.004275331338179, "grad_norm": 0.19611669457431613, "learning_rate": 1.41e-05, "loss": 0.0136, "step": 586 }, { "epoch": 2.0076955964087215, "grad_norm": 0.18201675420540842, "learning_rate": 1.41e-05, "loss": 0.0146, "step": 587 }, { "epoch": 2.0111158614792646, "grad_norm": 0.2111608412027396, "learning_rate": 1.41e-05, "loss": 0.0165, "step": 588 }, { "epoch": 2.0145361265498076, "grad_norm": 0.19725473414903413, "learning_rate": 1.41e-05, "loss": 0.015, "step": 589 }, { "epoch": 2.0179563916203507, "grad_norm": 0.21354984112177863, "learning_rate": 1.41e-05, "loss": 0.0161, "step": 590 }, { "epoch": 2.0213766566908937, "grad_norm": 0.22940862450456587, "learning_rate": 1.41e-05, "loss": 0.0157, "step": 591 }, { "epoch": 2.0247969217614363, "grad_norm": 0.18767782352642373, "learning_rate": 1.41e-05, "loss": 0.0152, "step": 592 }, { "epoch": 2.0282171868319794, "grad_norm": 0.20643159409018694, "learning_rate": 1.41e-05, "loss": 0.0148, "step": 593 }, { "epoch": 2.0316374519025224, "grad_norm": 0.2037046870595368, "learning_rate": 1.41e-05, "loss": 0.0138, "step": 594 }, { "epoch": 2.0350577169730655, "grad_norm": 0.19705696567608205, "learning_rate": 1.41e-05, "loss": 0.0162, "step": 595 }, { "epoch": 2.0384779820436085, "grad_norm": 0.21000255959562664, "learning_rate": 1.41e-05, "loss": 0.0156, "step": 596 }, { "epoch": 2.041898247114151, "grad_norm": 0.1909421240987611, "learning_rate": 1.41e-05, "loss": 0.0136, "step": 597 }, { "epoch": 2.045318512184694, "grad_norm": 0.22254593762119865, "learning_rate": 1.41e-05, "loss": 0.0155, "step": 598 }, { "epoch": 2.0487387772552372, "grad_norm": 0.18039596470753322, "learning_rate": 1.41e-05, "loss": 0.0147, "step": 599 }, { "epoch": 2.0521590423257803, "grad_norm": 0.1836910759989102, "learning_rate": 1.41e-05, "loss": 0.0139, "step": 600 }, { "epoch": 2.0555793073963233, "grad_norm": 0.2007119695404956, "learning_rate": 1.41e-05, "loss": 0.015, "step": 601 }, { "epoch": 2.0589995724668664, "grad_norm": 0.17881220811912174, "learning_rate": 1.41e-05, "loss": 0.016, "step": 602 }, { "epoch": 2.062419837537409, "grad_norm": 0.18131752461639056, "learning_rate": 1.41e-05, "loss": 0.0152, "step": 603 }, { "epoch": 2.065840102607952, "grad_norm": 0.2091101968035948, "learning_rate": 1.41e-05, "loss": 0.0147, "step": 604 }, { "epoch": 2.069260367678495, "grad_norm": 0.1785580403127973, "learning_rate": 1.41e-05, "loss": 0.0142, "step": 605 }, { "epoch": 2.072680632749038, "grad_norm": 0.1952402053161801, "learning_rate": 1.41e-05, "loss": 0.0143, "step": 606 }, { "epoch": 2.076100897819581, "grad_norm": 0.16849209697986273, "learning_rate": 1.41e-05, "loss": 0.0135, "step": 607 }, { "epoch": 2.079521162890124, "grad_norm": 0.18960749528358953, "learning_rate": 1.41e-05, "loss": 0.0158, "step": 608 }, { "epoch": 2.082941427960667, "grad_norm": 0.1822523628091137, "learning_rate": 1.41e-05, "loss": 0.0157, "step": 609 }, { "epoch": 2.08636169303121, "grad_norm": 0.22244344533378702, "learning_rate": 1.41e-05, "loss": 0.0145, "step": 610 }, { "epoch": 2.089781958101753, "grad_norm": 0.16677137506686548, "learning_rate": 1.41e-05, "loss": 0.0144, "step": 611 }, { "epoch": 2.093202223172296, "grad_norm": 0.19008011213315038, "learning_rate": 1.41e-05, "loss": 0.0149, "step": 612 }, { "epoch": 2.096622488242839, "grad_norm": 0.1965146825157404, "learning_rate": 1.41e-05, "loss": 0.0158, "step": 613 }, { "epoch": 2.1000427533133816, "grad_norm": 0.17095730271612053, "learning_rate": 1.41e-05, "loss": 0.0148, "step": 614 }, { "epoch": 2.1034630183839247, "grad_norm": 0.18376098918489842, "learning_rate": 1.41e-05, "loss": 0.0133, "step": 615 }, { "epoch": 2.1068832834544677, "grad_norm": 0.18859892344363186, "learning_rate": 1.41e-05, "loss": 0.0155, "step": 616 }, { "epoch": 2.110303548525011, "grad_norm": 0.17678403587042746, "learning_rate": 1.41e-05, "loss": 0.013, "step": 617 }, { "epoch": 2.113723813595554, "grad_norm": 0.18933984169673357, "learning_rate": 1.41e-05, "loss": 0.0152, "step": 618 }, { "epoch": 2.1171440786660964, "grad_norm": 0.20060664740874196, "learning_rate": 1.41e-05, "loss": 0.0157, "step": 619 }, { "epoch": 2.1205643437366395, "grad_norm": 0.1803038264349401, "learning_rate": 1.41e-05, "loss": 0.0141, "step": 620 }, { "epoch": 2.1239846088071825, "grad_norm": 0.1910210164605926, "learning_rate": 1.41e-05, "loss": 0.015, "step": 621 }, { "epoch": 2.1274048738777256, "grad_norm": 0.1794365441256405, "learning_rate": 1.41e-05, "loss": 0.0137, "step": 622 }, { "epoch": 2.1308251389482686, "grad_norm": 0.15893205685478506, "learning_rate": 1.41e-05, "loss": 0.0136, "step": 623 }, { "epoch": 2.1342454040188112, "grad_norm": 0.19724991519919266, "learning_rate": 1.41e-05, "loss": 0.0143, "step": 624 }, { "epoch": 2.1376656690893543, "grad_norm": 0.17458605756275755, "learning_rate": 1.41e-05, "loss": 0.016, "step": 625 }, { "epoch": 2.1410859341598973, "grad_norm": 0.172521593573168, "learning_rate": 1.41e-05, "loss": 0.0141, "step": 626 }, { "epoch": 2.1445061992304404, "grad_norm": 0.14557636073307398, "learning_rate": 1.41e-05, "loss": 0.0135, "step": 627 }, { "epoch": 2.1479264643009834, "grad_norm": 0.22843890765431576, "learning_rate": 1.41e-05, "loss": 0.016, "step": 628 }, { "epoch": 2.1513467293715265, "grad_norm": 0.18196169286713992, "learning_rate": 1.41e-05, "loss": 0.0146, "step": 629 }, { "epoch": 2.154766994442069, "grad_norm": 0.17272140560874683, "learning_rate": 1.41e-05, "loss": 0.0135, "step": 630 }, { "epoch": 2.158187259512612, "grad_norm": 0.15595539359263064, "learning_rate": 1.41e-05, "loss": 0.014, "step": 631 }, { "epoch": 2.161607524583155, "grad_norm": 0.16465865001486402, "learning_rate": 1.41e-05, "loss": 0.0136, "step": 632 }, { "epoch": 2.1650277896536982, "grad_norm": 0.17986632980817158, "learning_rate": 1.41e-05, "loss": 0.015, "step": 633 }, { "epoch": 2.1684480547242413, "grad_norm": 0.18929877257592365, "learning_rate": 1.41e-05, "loss": 0.0163, "step": 634 }, { "epoch": 2.171868319794784, "grad_norm": 0.19542505698465773, "learning_rate": 1.41e-05, "loss": 0.0159, "step": 635 }, { "epoch": 2.175288584865327, "grad_norm": 0.17114565781513597, "learning_rate": 1.41e-05, "loss": 0.0145, "step": 636 }, { "epoch": 2.17870884993587, "grad_norm": 0.18331913158118285, "learning_rate": 1.41e-05, "loss": 0.0146, "step": 637 }, { "epoch": 2.182129115006413, "grad_norm": 0.16055196305467795, "learning_rate": 1.41e-05, "loss": 0.015, "step": 638 }, { "epoch": 2.185549380076956, "grad_norm": 0.15985440322290387, "learning_rate": 1.41e-05, "loss": 0.0141, "step": 639 }, { "epoch": 2.188969645147499, "grad_norm": 0.16417247457666104, "learning_rate": 1.41e-05, "loss": 0.0145, "step": 640 }, { "epoch": 2.1923899102180417, "grad_norm": 0.16210688538374043, "learning_rate": 1.41e-05, "loss": 0.0141, "step": 641 }, { "epoch": 2.195810175288585, "grad_norm": 0.16634144533077172, "learning_rate": 1.41e-05, "loss": 0.014, "step": 642 }, { "epoch": 2.199230440359128, "grad_norm": 0.18582606043218244, "learning_rate": 1.41e-05, "loss": 0.0156, "step": 643 }, { "epoch": 2.202650705429671, "grad_norm": 0.1747184961523454, "learning_rate": 1.41e-05, "loss": 0.0164, "step": 644 }, { "epoch": 2.206070970500214, "grad_norm": 0.15610336933105856, "learning_rate": 1.41e-05, "loss": 0.013, "step": 645 }, { "epoch": 2.2094912355707566, "grad_norm": 0.16920938302300378, "learning_rate": 1.41e-05, "loss": 0.0139, "step": 646 }, { "epoch": 2.2129115006412996, "grad_norm": 0.17648344233280136, "learning_rate": 1.41e-05, "loss": 0.0147, "step": 647 }, { "epoch": 2.2163317657118426, "grad_norm": 0.17020885614859071, "learning_rate": 1.41e-05, "loss": 0.0155, "step": 648 }, { "epoch": 2.2197520307823857, "grad_norm": 0.17716196526554184, "learning_rate": 1.41e-05, "loss": 0.0142, "step": 649 }, { "epoch": 2.2231722958529287, "grad_norm": 0.17375550168851311, "learning_rate": 1.41e-05, "loss": 0.0144, "step": 650 }, { "epoch": 2.2265925609234714, "grad_norm": 0.15660012931864775, "learning_rate": 1.41e-05, "loss": 0.0138, "step": 651 }, { "epoch": 2.2300128259940144, "grad_norm": 0.192388171113886, "learning_rate": 1.41e-05, "loss": 0.0151, "step": 652 }, { "epoch": 2.2334330910645575, "grad_norm": 0.17112223174304642, "learning_rate": 1.41e-05, "loss": 0.0137, "step": 653 }, { "epoch": 2.2368533561351005, "grad_norm": 0.17046414663473125, "learning_rate": 1.41e-05, "loss": 0.0163, "step": 654 }, { "epoch": 2.2402736212056436, "grad_norm": 0.14723103129157328, "learning_rate": 1.41e-05, "loss": 0.0141, "step": 655 }, { "epoch": 2.2436938862761866, "grad_norm": 0.16260020026720062, "learning_rate": 1.41e-05, "loss": 0.0134, "step": 656 }, { "epoch": 2.247114151346729, "grad_norm": 0.1605529186115587, "learning_rate": 1.41e-05, "loss": 0.0148, "step": 657 }, { "epoch": 2.2505344164172723, "grad_norm": 0.1649326783596916, "learning_rate": 1.41e-05, "loss": 0.0146, "step": 658 }, { "epoch": 2.2539546814878153, "grad_norm": 0.1664480794095166, "learning_rate": 1.41e-05, "loss": 0.0144, "step": 659 }, { "epoch": 2.2573749465583584, "grad_norm": 0.16946179512795181, "learning_rate": 1.41e-05, "loss": 0.0146, "step": 660 }, { "epoch": 2.2607952116289014, "grad_norm": 0.166434202624617, "learning_rate": 1.41e-05, "loss": 0.0135, "step": 661 }, { "epoch": 2.2642154766994445, "grad_norm": 0.1739998754952991, "learning_rate": 1.41e-05, "loss": 0.0142, "step": 662 }, { "epoch": 2.267635741769987, "grad_norm": 0.1809093145862042, "learning_rate": 1.41e-05, "loss": 0.014, "step": 663 }, { "epoch": 2.27105600684053, "grad_norm": 0.15727492334875873, "learning_rate": 1.41e-05, "loss": 0.0137, "step": 664 }, { "epoch": 2.274476271911073, "grad_norm": 0.18448939109027693, "learning_rate": 1.41e-05, "loss": 0.0157, "step": 665 }, { "epoch": 2.277896536981616, "grad_norm": 0.18236721526184443, "learning_rate": 1.41e-05, "loss": 0.0166, "step": 666 }, { "epoch": 2.2813168020521593, "grad_norm": 0.1695210468590141, "learning_rate": 1.41e-05, "loss": 0.0144, "step": 667 }, { "epoch": 2.284737067122702, "grad_norm": 0.17157602170689792, "learning_rate": 1.41e-05, "loss": 0.0153, "step": 668 }, { "epoch": 2.288157332193245, "grad_norm": 0.16583571772207567, "learning_rate": 1.41e-05, "loss": 0.0133, "step": 669 }, { "epoch": 2.291577597263788, "grad_norm": 0.15309231682278962, "learning_rate": 1.41e-05, "loss": 0.0134, "step": 670 }, { "epoch": 2.294997862334331, "grad_norm": 0.18010491103515566, "learning_rate": 1.41e-05, "loss": 0.0155, "step": 671 }, { "epoch": 2.298418127404874, "grad_norm": 0.165226793210629, "learning_rate": 1.41e-05, "loss": 0.0134, "step": 672 }, { "epoch": 2.3018383924754167, "grad_norm": 0.19017257841407276, "learning_rate": 1.41e-05, "loss": 0.016, "step": 673 }, { "epoch": 2.3052586575459597, "grad_norm": 0.16105989698952458, "learning_rate": 1.41e-05, "loss": 0.0154, "step": 674 }, { "epoch": 2.3086789226165028, "grad_norm": 0.17435397438208278, "learning_rate": 1.41e-05, "loss": 0.0136, "step": 675 }, { "epoch": 2.312099187687046, "grad_norm": 0.15489352805926157, "learning_rate": 1.41e-05, "loss": 0.0149, "step": 676 }, { "epoch": 2.315519452757589, "grad_norm": 0.15749574161575872, "learning_rate": 1.41e-05, "loss": 0.0136, "step": 677 }, { "epoch": 2.3189397178281315, "grad_norm": 0.17861788971942064, "learning_rate": 1.41e-05, "loss": 0.015, "step": 678 }, { "epoch": 2.3223599828986745, "grad_norm": 0.16852344878390402, "learning_rate": 1.41e-05, "loss": 0.0152, "step": 679 }, { "epoch": 2.3257802479692176, "grad_norm": 0.1903544466095535, "learning_rate": 1.41e-05, "loss": 0.0146, "step": 680 }, { "epoch": 2.3292005130397606, "grad_norm": 0.1938655036641079, "learning_rate": 1.41e-05, "loss": 0.0158, "step": 681 }, { "epoch": 2.3326207781103037, "grad_norm": 0.20203642790445486, "learning_rate": 1.41e-05, "loss": 0.0143, "step": 682 }, { "epoch": 2.3360410431808463, "grad_norm": 0.16328748761388734, "learning_rate": 1.41e-05, "loss": 0.0154, "step": 683 }, { "epoch": 2.3394613082513893, "grad_norm": 0.17437303421082254, "learning_rate": 1.41e-05, "loss": 0.0149, "step": 684 }, { "epoch": 2.3428815733219324, "grad_norm": 0.17762818156916235, "learning_rate": 1.41e-05, "loss": 0.0149, "step": 685 }, { "epoch": 2.3463018383924754, "grad_norm": 0.18594818733099802, "learning_rate": 1.41e-05, "loss": 0.0161, "step": 686 }, { "epoch": 2.3497221034630185, "grad_norm": 0.16139308140544603, "learning_rate": 1.41e-05, "loss": 0.0148, "step": 687 }, { "epoch": 2.3531423685335615, "grad_norm": 0.16108681383332696, "learning_rate": 1.41e-05, "loss": 0.0139, "step": 688 }, { "epoch": 2.3565626336041046, "grad_norm": 0.15992689822195585, "learning_rate": 1.41e-05, "loss": 0.0146, "step": 689 }, { "epoch": 2.359982898674647, "grad_norm": 0.19759610045563245, "learning_rate": 1.41e-05, "loss": 0.0158, "step": 690 }, { "epoch": 2.36340316374519, "grad_norm": 0.16149200857684542, "learning_rate": 1.41e-05, "loss": 0.014, "step": 691 }, { "epoch": 2.3668234288157333, "grad_norm": 0.15072989325955685, "learning_rate": 1.41e-05, "loss": 0.0144, "step": 692 }, { "epoch": 2.3702436938862763, "grad_norm": 0.14913939759012662, "learning_rate": 1.41e-05, "loss": 0.014, "step": 693 }, { "epoch": 2.3736639589568194, "grad_norm": 0.15636976839856181, "learning_rate": 1.41e-05, "loss": 0.0153, "step": 694 }, { "epoch": 2.377084224027362, "grad_norm": 0.18599274943363073, "learning_rate": 1.41e-05, "loss": 0.0141, "step": 695 }, { "epoch": 2.380504489097905, "grad_norm": 0.16568136382204926, "learning_rate": 1.41e-05, "loss": 0.013, "step": 696 }, { "epoch": 2.383924754168448, "grad_norm": 0.17332302727716312, "learning_rate": 1.41e-05, "loss": 0.0128, "step": 697 }, { "epoch": 2.387345019238991, "grad_norm": 0.1879178833471297, "learning_rate": 1.41e-05, "loss": 0.0146, "step": 698 }, { "epoch": 2.390765284309534, "grad_norm": 0.1726315547096704, "learning_rate": 1.41e-05, "loss": 0.0136, "step": 699 }, { "epoch": 2.3941855493800768, "grad_norm": 0.1720038600446686, "learning_rate": 1.41e-05, "loss": 0.0158, "step": 700 }, { "epoch": 2.39760581445062, "grad_norm": 0.17262771034044544, "learning_rate": 1.41e-05, "loss": 0.0146, "step": 701 }, { "epoch": 2.401026079521163, "grad_norm": 0.18381965008540402, "learning_rate": 1.41e-05, "loss": 0.0151, "step": 702 }, { "epoch": 2.404446344591706, "grad_norm": 0.1943786246553059, "learning_rate": 1.41e-05, "loss": 0.0148, "step": 703 }, { "epoch": 2.407866609662249, "grad_norm": 0.15285642974808178, "learning_rate": 1.41e-05, "loss": 0.0133, "step": 704 }, { "epoch": 2.4112868747327916, "grad_norm": 0.19330131469876405, "learning_rate": 1.41e-05, "loss": 0.0146, "step": 705 }, { "epoch": 2.4147071398033346, "grad_norm": 0.16935279234922626, "learning_rate": 1.41e-05, "loss": 0.0143, "step": 706 }, { "epoch": 2.4181274048738777, "grad_norm": 0.17130297774748693, "learning_rate": 1.41e-05, "loss": 0.016, "step": 707 }, { "epoch": 2.4215476699444207, "grad_norm": 0.16649757139294383, "learning_rate": 1.41e-05, "loss": 0.0135, "step": 708 }, { "epoch": 2.4249679350149638, "grad_norm": 0.1509762981692361, "learning_rate": 1.41e-05, "loss": 0.0131, "step": 709 }, { "epoch": 2.428388200085507, "grad_norm": 0.16903612165580836, "learning_rate": 1.41e-05, "loss": 0.015, "step": 710 }, { "epoch": 2.4318084651560494, "grad_norm": 0.16965158831335697, "learning_rate": 1.41e-05, "loss": 0.0139, "step": 711 }, { "epoch": 2.4352287302265925, "grad_norm": 0.17489335625033847, "learning_rate": 1.41e-05, "loss": 0.0127, "step": 712 }, { "epoch": 2.4386489952971355, "grad_norm": 0.17619733289193135, "learning_rate": 1.41e-05, "loss": 0.0144, "step": 713 }, { "epoch": 2.4420692603676786, "grad_norm": 0.17617881507074173, "learning_rate": 1.41e-05, "loss": 0.0152, "step": 714 }, { "epoch": 2.4454895254382216, "grad_norm": 0.18048518098716906, "learning_rate": 1.41e-05, "loss": 0.0139, "step": 715 }, { "epoch": 2.4489097905087647, "grad_norm": 0.2051952297924243, "learning_rate": 1.41e-05, "loss": 0.0146, "step": 716 }, { "epoch": 2.4523300555793073, "grad_norm": 0.15664970563386238, "learning_rate": 1.41e-05, "loss": 0.0143, "step": 717 }, { "epoch": 2.4557503206498503, "grad_norm": 0.16743723595706328, "learning_rate": 1.41e-05, "loss": 0.0152, "step": 718 }, { "epoch": 2.4591705857203934, "grad_norm": 0.17613442783841668, "learning_rate": 1.41e-05, "loss": 0.0162, "step": 719 }, { "epoch": 2.4625908507909364, "grad_norm": 0.1818706957568455, "learning_rate": 1.41e-05, "loss": 0.0135, "step": 720 }, { "epoch": 2.4660111158614795, "grad_norm": 0.17047836187664528, "learning_rate": 1.41e-05, "loss": 0.0141, "step": 721 }, { "epoch": 2.469431380932022, "grad_norm": 0.17741467330052185, "learning_rate": 1.41e-05, "loss": 0.0158, "step": 722 }, { "epoch": 2.472851646002565, "grad_norm": 0.1800167905144302, "learning_rate": 1.41e-05, "loss": 0.0156, "step": 723 }, { "epoch": 2.476271911073108, "grad_norm": 0.1520569062216661, "learning_rate": 1.41e-05, "loss": 0.014, "step": 724 }, { "epoch": 2.4796921761436512, "grad_norm": 0.19504143457216305, "learning_rate": 1.41e-05, "loss": 0.015, "step": 725 }, { "epoch": 2.4831124412141943, "grad_norm": 0.16406995475864966, "learning_rate": 1.41e-05, "loss": 0.0152, "step": 726 }, { "epoch": 2.486532706284737, "grad_norm": 0.1640747075139151, "learning_rate": 1.41e-05, "loss": 0.0138, "step": 727 }, { "epoch": 2.48995297135528, "grad_norm": 0.17515606314180554, "learning_rate": 1.41e-05, "loss": 0.0156, "step": 728 }, { "epoch": 2.493373236425823, "grad_norm": 0.17908072157399882, "learning_rate": 1.41e-05, "loss": 0.0155, "step": 729 }, { "epoch": 2.496793501496366, "grad_norm": 0.15664812551985538, "learning_rate": 1.41e-05, "loss": 0.0147, "step": 730 }, { "epoch": 2.500213766566909, "grad_norm": 0.165379248202876, "learning_rate": 1.41e-05, "loss": 0.0151, "step": 731 }, { "epoch": 2.5036340316374517, "grad_norm": 0.15423115964706519, "learning_rate": 1.41e-05, "loss": 0.0148, "step": 732 }, { "epoch": 2.5070542967079947, "grad_norm": 0.16485219288254147, "learning_rate": 1.41e-05, "loss": 0.0149, "step": 733 }, { "epoch": 2.510474561778538, "grad_norm": 0.176637422447442, "learning_rate": 1.41e-05, "loss": 0.0154, "step": 734 }, { "epoch": 2.513894826849081, "grad_norm": 0.17814754105281233, "learning_rate": 1.41e-05, "loss": 0.0154, "step": 735 }, { "epoch": 2.517315091919624, "grad_norm": 0.17118656720070607, "learning_rate": 1.41e-05, "loss": 0.0155, "step": 736 }, { "epoch": 2.5207353569901665, "grad_norm": 0.18260703508475823, "learning_rate": 1.41e-05, "loss": 0.0154, "step": 737 }, { "epoch": 2.52415562206071, "grad_norm": 0.17080402514684412, "learning_rate": 1.41e-05, "loss": 0.0136, "step": 738 }, { "epoch": 2.5275758871312526, "grad_norm": 0.15709135265175292, "learning_rate": 1.41e-05, "loss": 0.0134, "step": 739 }, { "epoch": 2.5309961522017956, "grad_norm": 0.18246981305252055, "learning_rate": 1.41e-05, "loss": 0.0142, "step": 740 }, { "epoch": 2.5344164172723387, "grad_norm": 0.1808585737198185, "learning_rate": 1.41e-05, "loss": 0.0139, "step": 741 }, { "epoch": 2.5378366823428817, "grad_norm": 0.16454795740570335, "learning_rate": 1.41e-05, "loss": 0.0153, "step": 742 }, { "epoch": 2.541256947413425, "grad_norm": 0.21032421935647805, "learning_rate": 1.41e-05, "loss": 0.0165, "step": 743 }, { "epoch": 2.5446772124839674, "grad_norm": 0.15249763976570654, "learning_rate": 1.41e-05, "loss": 0.0139, "step": 744 }, { "epoch": 2.5480974775545104, "grad_norm": 0.1613635941460105, "learning_rate": 1.41e-05, "loss": 0.0131, "step": 745 }, { "epoch": 2.5515177426250535, "grad_norm": 0.1700060737892137, "learning_rate": 1.41e-05, "loss": 0.0145, "step": 746 }, { "epoch": 2.5549380076955965, "grad_norm": 0.20547449901345968, "learning_rate": 1.41e-05, "loss": 0.0168, "step": 747 }, { "epoch": 2.5583582727661396, "grad_norm": 0.16484966537277382, "learning_rate": 1.41e-05, "loss": 0.0142, "step": 748 }, { "epoch": 2.561778537836682, "grad_norm": 0.17889219393433903, "learning_rate": 1.41e-05, "loss": 0.014, "step": 749 }, { "epoch": 2.5651988029072252, "grad_norm": 0.17679721987917787, "learning_rate": 1.41e-05, "loss": 0.0138, "step": 750 }, { "epoch": 2.5686190679777683, "grad_norm": 0.16210612559541202, "learning_rate": 1.41e-05, "loss": 0.0145, "step": 751 }, { "epoch": 2.5720393330483113, "grad_norm": 0.17192503019428163, "learning_rate": 1.41e-05, "loss": 0.0143, "step": 752 }, { "epoch": 2.5754595981188544, "grad_norm": 0.16114230533462182, "learning_rate": 1.41e-05, "loss": 0.0151, "step": 753 }, { "epoch": 2.578879863189397, "grad_norm": 0.16683326324078276, "learning_rate": 1.41e-05, "loss": 0.014, "step": 754 }, { "epoch": 2.58230012825994, "grad_norm": 0.15929030309974343, "learning_rate": 1.41e-05, "loss": 0.0135, "step": 755 }, { "epoch": 2.585720393330483, "grad_norm": 0.17728404681160545, "learning_rate": 1.41e-05, "loss": 0.0134, "step": 756 }, { "epoch": 2.589140658401026, "grad_norm": 0.16727441441267749, "learning_rate": 1.41e-05, "loss": 0.0143, "step": 757 }, { "epoch": 2.592560923471569, "grad_norm": 0.21841845784314023, "learning_rate": 1.41e-05, "loss": 0.0149, "step": 758 }, { "epoch": 2.595981188542112, "grad_norm": 0.1770465088709539, "learning_rate": 1.41e-05, "loss": 0.015, "step": 759 }, { "epoch": 2.599401453612655, "grad_norm": 0.1737371304814951, "learning_rate": 1.41e-05, "loss": 0.0156, "step": 760 }, { "epoch": 2.602821718683198, "grad_norm": 0.17069864594223258, "learning_rate": 1.41e-05, "loss": 0.0134, "step": 761 }, { "epoch": 2.606241983753741, "grad_norm": 0.16609359338004503, "learning_rate": 1.41e-05, "loss": 0.0158, "step": 762 }, { "epoch": 2.609662248824284, "grad_norm": 0.17378348121220705, "learning_rate": 1.41e-05, "loss": 0.0141, "step": 763 }, { "epoch": 2.6130825138948266, "grad_norm": 0.17355912661147865, "learning_rate": 1.41e-05, "loss": 0.0142, "step": 764 }, { "epoch": 2.61650277896537, "grad_norm": 0.20077889987954006, "learning_rate": 1.41e-05, "loss": 0.0154, "step": 765 }, { "epoch": 2.6199230440359127, "grad_norm": 0.16419074893195182, "learning_rate": 1.41e-05, "loss": 0.014, "step": 766 }, { "epoch": 2.6233433091064557, "grad_norm": 0.19476203022751112, "learning_rate": 1.41e-05, "loss": 0.0154, "step": 767 }, { "epoch": 2.626763574176999, "grad_norm": 0.16568838103054007, "learning_rate": 1.41e-05, "loss": 0.0148, "step": 768 }, { "epoch": 2.630183839247542, "grad_norm": 0.21099152906068663, "learning_rate": 1.41e-05, "loss": 0.017, "step": 769 }, { "epoch": 2.633604104318085, "grad_norm": 0.17959731227982292, "learning_rate": 1.41e-05, "loss": 0.014, "step": 770 }, { "epoch": 2.6370243693886275, "grad_norm": 0.16972780919186436, "learning_rate": 1.41e-05, "loss": 0.0139, "step": 771 }, { "epoch": 2.6404446344591705, "grad_norm": 0.1842651161264449, "learning_rate": 1.41e-05, "loss": 0.0157, "step": 772 }, { "epoch": 2.6438648995297136, "grad_norm": 0.17834439152890905, "learning_rate": 1.41e-05, "loss": 0.0153, "step": 773 }, { "epoch": 2.6472851646002566, "grad_norm": 0.18047152749420897, "learning_rate": 1.41e-05, "loss": 0.0156, "step": 774 }, { "epoch": 2.6507054296707997, "grad_norm": 0.17154124492966955, "learning_rate": 1.41e-05, "loss": 0.0147, "step": 775 }, { "epoch": 2.6541256947413423, "grad_norm": 0.17549885521671557, "learning_rate": 1.41e-05, "loss": 0.0161, "step": 776 }, { "epoch": 2.6575459598118853, "grad_norm": 0.17861344190439962, "learning_rate": 1.41e-05, "loss": 0.0151, "step": 777 }, { "epoch": 2.6609662248824284, "grad_norm": 0.16730938712809765, "learning_rate": 1.41e-05, "loss": 0.0145, "step": 778 }, { "epoch": 2.6643864899529714, "grad_norm": 0.1775818865748014, "learning_rate": 1.41e-05, "loss": 0.0159, "step": 779 }, { "epoch": 2.6678067550235145, "grad_norm": 0.191333871334205, "learning_rate": 1.41e-05, "loss": 0.013, "step": 780 }, { "epoch": 2.671227020094057, "grad_norm": 0.16907634204579375, "learning_rate": 1.41e-05, "loss": 0.0145, "step": 781 }, { "epoch": 2.6746472851646, "grad_norm": 0.18178434604030544, "learning_rate": 1.41e-05, "loss": 0.0135, "step": 782 }, { "epoch": 2.678067550235143, "grad_norm": 0.20503354023831713, "learning_rate": 1.41e-05, "loss": 0.0146, "step": 783 }, { "epoch": 2.6814878153056863, "grad_norm": 0.16391814776977276, "learning_rate": 1.41e-05, "loss": 0.0156, "step": 784 }, { "epoch": 2.6849080803762293, "grad_norm": 0.1839989787303562, "learning_rate": 1.41e-05, "loss": 0.0148, "step": 785 }, { "epoch": 2.688328345446772, "grad_norm": 0.19236560848567033, "learning_rate": 1.41e-05, "loss": 0.0154, "step": 786 }, { "epoch": 2.691748610517315, "grad_norm": 0.16811298067940522, "learning_rate": 1.41e-05, "loss": 0.0141, "step": 787 }, { "epoch": 2.695168875587858, "grad_norm": 0.19975209346986805, "learning_rate": 1.41e-05, "loss": 0.0159, "step": 788 }, { "epoch": 2.698589140658401, "grad_norm": 0.16338526349657181, "learning_rate": 1.41e-05, "loss": 0.0133, "step": 789 }, { "epoch": 2.702009405728944, "grad_norm": 0.17351362318047445, "learning_rate": 1.41e-05, "loss": 0.0159, "step": 790 }, { "epoch": 2.7054296707994867, "grad_norm": 0.17661074926464723, "learning_rate": 1.41e-05, "loss": 0.0154, "step": 791 }, { "epoch": 2.70884993587003, "grad_norm": 0.1762002497021315, "learning_rate": 1.41e-05, "loss": 0.016, "step": 792 }, { "epoch": 2.712270200940573, "grad_norm": 0.19710675588819915, "learning_rate": 1.41e-05, "loss": 0.0161, "step": 793 }, { "epoch": 2.715690466011116, "grad_norm": 0.1692117872805568, "learning_rate": 1.41e-05, "loss": 0.014, "step": 794 }, { "epoch": 2.719110731081659, "grad_norm": 0.20731978857327257, "learning_rate": 1.41e-05, "loss": 0.0156, "step": 795 }, { "epoch": 2.722530996152202, "grad_norm": 0.17806201725432017, "learning_rate": 1.41e-05, "loss": 0.0147, "step": 796 }, { "epoch": 2.725951261222745, "grad_norm": 0.16617443428776074, "learning_rate": 1.41e-05, "loss": 0.0143, "step": 797 }, { "epoch": 2.7293715262932876, "grad_norm": 0.20059647446620907, "learning_rate": 1.41e-05, "loss": 0.014, "step": 798 }, { "epoch": 2.7327917913638307, "grad_norm": 0.1773142221832559, "learning_rate": 1.41e-05, "loss": 0.0171, "step": 799 }, { "epoch": 2.7362120564343737, "grad_norm": 0.1979846553461278, "learning_rate": 1.41e-05, "loss": 0.0139, "step": 800 }, { "epoch": 2.7396323215049168, "grad_norm": 0.18511763263524342, "learning_rate": 1.41e-05, "loss": 0.0155, "step": 801 }, { "epoch": 2.74305258657546, "grad_norm": 0.17699099585761474, "learning_rate": 1.41e-05, "loss": 0.0118, "step": 802 }, { "epoch": 2.7464728516460024, "grad_norm": 0.19352657405112964, "learning_rate": 1.41e-05, "loss": 0.0163, "step": 803 }, { "epoch": 2.7498931167165455, "grad_norm": 0.19135035210758095, "learning_rate": 1.41e-05, "loss": 0.0156, "step": 804 }, { "epoch": 2.7533133817870885, "grad_norm": 0.1841148474869883, "learning_rate": 1.41e-05, "loss": 0.0155, "step": 805 }, { "epoch": 2.7567336468576316, "grad_norm": 0.19008317628622337, "learning_rate": 1.41e-05, "loss": 0.0152, "step": 806 }, { "epoch": 2.7601539119281746, "grad_norm": 0.1949140281276627, "learning_rate": 1.41e-05, "loss": 0.0143, "step": 807 }, { "epoch": 2.763574176998717, "grad_norm": 0.1944281661049677, "learning_rate": 1.41e-05, "loss": 0.0145, "step": 808 }, { "epoch": 2.7669944420692603, "grad_norm": 0.1674859606182937, "learning_rate": 1.41e-05, "loss": 0.0162, "step": 809 }, { "epoch": 2.7704147071398033, "grad_norm": 0.18850185666818006, "learning_rate": 1.41e-05, "loss": 0.0148, "step": 810 }, { "epoch": 2.7738349722103464, "grad_norm": 0.18475272615726215, "learning_rate": 1.41e-05, "loss": 0.0156, "step": 811 }, { "epoch": 2.7772552372808894, "grad_norm": 0.15365265630124134, "learning_rate": 1.41e-05, "loss": 0.0134, "step": 812 }, { "epoch": 2.780675502351432, "grad_norm": 0.16024686744719197, "learning_rate": 1.41e-05, "loss": 0.0143, "step": 813 }, { "epoch": 2.784095767421975, "grad_norm": 0.18228181650312342, "learning_rate": 1.41e-05, "loss": 0.0142, "step": 814 }, { "epoch": 2.787516032492518, "grad_norm": 0.15232575479299343, "learning_rate": 1.41e-05, "loss": 0.0141, "step": 815 }, { "epoch": 2.790936297563061, "grad_norm": 0.19473439330433306, "learning_rate": 1.41e-05, "loss": 0.0141, "step": 816 }, { "epoch": 2.794356562633604, "grad_norm": 0.20524717935488615, "learning_rate": 1.41e-05, "loss": 0.017, "step": 817 }, { "epoch": 2.797776827704147, "grad_norm": 0.1908817371704934, "learning_rate": 1.41e-05, "loss": 0.0147, "step": 818 }, { "epoch": 2.8011970927746903, "grad_norm": 0.1817449147111606, "learning_rate": 1.41e-05, "loss": 0.0146, "step": 819 }, { "epoch": 2.804617357845233, "grad_norm": 0.20849626412897376, "learning_rate": 1.41e-05, "loss": 0.0144, "step": 820 }, { "epoch": 2.808037622915776, "grad_norm": 0.17439115442992892, "learning_rate": 1.41e-05, "loss": 0.0165, "step": 821 }, { "epoch": 2.811457887986319, "grad_norm": 0.184212269622675, "learning_rate": 1.41e-05, "loss": 0.0139, "step": 822 }, { "epoch": 2.814878153056862, "grad_norm": 0.1742874984626096, "learning_rate": 1.41e-05, "loss": 0.0146, "step": 823 }, { "epoch": 2.818298418127405, "grad_norm": 0.16854467961459083, "learning_rate": 1.41e-05, "loss": 0.0145, "step": 824 }, { "epoch": 2.8217186831979477, "grad_norm": 0.17468169666793745, "learning_rate": 1.41e-05, "loss": 0.0149, "step": 825 }, { "epoch": 2.8251389482684908, "grad_norm": 0.16936997795252084, "learning_rate": 1.41e-05, "loss": 0.0139, "step": 826 }, { "epoch": 2.828559213339034, "grad_norm": 0.16572236927644424, "learning_rate": 1.41e-05, "loss": 0.0154, "step": 827 }, { "epoch": 2.831979478409577, "grad_norm": 0.1651962024013233, "learning_rate": 1.41e-05, "loss": 0.0127, "step": 828 }, { "epoch": 2.83539974348012, "grad_norm": 0.1657376085150308, "learning_rate": 1.41e-05, "loss": 0.0139, "step": 829 }, { "epoch": 2.8388200085506625, "grad_norm": 0.17582133731054114, "learning_rate": 1.41e-05, "loss": 0.0131, "step": 830 }, { "epoch": 2.8422402736212056, "grad_norm": 0.20723879555852162, "learning_rate": 1.41e-05, "loss": 0.0162, "step": 831 }, { "epoch": 2.8456605386917486, "grad_norm": 0.6025905984134068, "learning_rate": 1.41e-05, "loss": 0.0142, "step": 832 }, { "epoch": 2.8490808037622917, "grad_norm": 0.1678623970594971, "learning_rate": 1.41e-05, "loss": 0.0143, "step": 833 }, { "epoch": 2.8525010688328347, "grad_norm": 0.21086844483510658, "learning_rate": 1.41e-05, "loss": 0.0158, "step": 834 }, { "epoch": 2.8559213339033773, "grad_norm": 0.18493480224421863, "learning_rate": 1.41e-05, "loss": 0.0126, "step": 835 }, { "epoch": 2.8593415989739204, "grad_norm": 0.1904246277238966, "learning_rate": 1.41e-05, "loss": 0.0155, "step": 836 }, { "epoch": 2.8627618640444634, "grad_norm": 0.18774356330645445, "learning_rate": 1.41e-05, "loss": 0.0174, "step": 837 }, { "epoch": 2.8661821291150065, "grad_norm": 0.20481371594917794, "learning_rate": 1.41e-05, "loss": 0.0132, "step": 838 }, { "epoch": 2.8696023941855495, "grad_norm": 0.19259669092688833, "learning_rate": 1.41e-05, "loss": 0.0144, "step": 839 }, { "epoch": 2.873022659256092, "grad_norm": 0.20161241494039436, "learning_rate": 1.41e-05, "loss": 0.0159, "step": 840 }, { "epoch": 2.876442924326635, "grad_norm": 0.19885243582338097, "learning_rate": 1.41e-05, "loss": 0.0155, "step": 841 }, { "epoch": 2.8798631893971782, "grad_norm": 0.18089297372993837, "learning_rate": 1.41e-05, "loss": 0.015, "step": 842 }, { "epoch": 2.8832834544677213, "grad_norm": 0.19064879861955514, "learning_rate": 1.41e-05, "loss": 0.0132, "step": 843 }, { "epoch": 2.8867037195382643, "grad_norm": 0.18907692558323141, "learning_rate": 1.41e-05, "loss": 0.0146, "step": 844 }, { "epoch": 2.890123984608807, "grad_norm": 0.19585899032623924, "learning_rate": 1.41e-05, "loss": 0.0153, "step": 845 }, { "epoch": 2.8935442496793504, "grad_norm": 0.22806745672139891, "learning_rate": 1.41e-05, "loss": 0.0169, "step": 846 }, { "epoch": 2.896964514749893, "grad_norm": 0.19336051551290398, "learning_rate": 1.41e-05, "loss": 0.0155, "step": 847 }, { "epoch": 2.900384779820436, "grad_norm": 0.17663309403974298, "learning_rate": 1.41e-05, "loss": 0.013, "step": 848 }, { "epoch": 2.903805044890979, "grad_norm": 0.1884422733573296, "learning_rate": 1.41e-05, "loss": 0.0153, "step": 849 }, { "epoch": 2.907225309961522, "grad_norm": 0.19574362737358483, "learning_rate": 1.41e-05, "loss": 0.016, "step": 850 }, { "epoch": 2.9106455750320652, "grad_norm": 0.16218777133767662, "learning_rate": 1.41e-05, "loss": 0.0135, "step": 851 }, { "epoch": 2.914065840102608, "grad_norm": 0.1844053227394077, "learning_rate": 1.41e-05, "loss": 0.0147, "step": 852 }, { "epoch": 2.917486105173151, "grad_norm": 0.19292851510616071, "learning_rate": 1.41e-05, "loss": 0.0149, "step": 853 }, { "epoch": 2.920906370243694, "grad_norm": 0.2089035835617463, "learning_rate": 1.41e-05, "loss": 0.017, "step": 854 }, { "epoch": 2.924326635314237, "grad_norm": 0.7055863227979324, "learning_rate": 1.41e-05, "loss": 0.016, "step": 855 }, { "epoch": 2.92774690038478, "grad_norm": 0.18518793131577801, "learning_rate": 1.41e-05, "loss": 0.0141, "step": 856 }, { "epoch": 2.9311671654553226, "grad_norm": 0.17688744692401304, "learning_rate": 1.41e-05, "loss": 0.0144, "step": 857 }, { "epoch": 2.9345874305258657, "grad_norm": 0.20202970506605689, "learning_rate": 1.41e-05, "loss": 0.0144, "step": 858 }, { "epoch": 2.9380076955964087, "grad_norm": 0.19190788298961772, "learning_rate": 1.41e-05, "loss": 0.0154, "step": 859 }, { "epoch": 2.941427960666952, "grad_norm": 0.19843345575867563, "learning_rate": 1.41e-05, "loss": 0.0167, "step": 860 }, { "epoch": 2.944848225737495, "grad_norm": 0.18043514187233164, "learning_rate": 1.41e-05, "loss": 0.0124, "step": 861 }, { "epoch": 2.9482684908080374, "grad_norm": 0.1852741160641503, "learning_rate": 1.41e-05, "loss": 0.0153, "step": 862 }, { "epoch": 2.9516887558785805, "grad_norm": 0.19552493492009015, "learning_rate": 1.41e-05, "loss": 0.0161, "step": 863 }, { "epoch": 2.9551090209491235, "grad_norm": 0.1972068517674688, "learning_rate": 1.41e-05, "loss": 0.0151, "step": 864 }, { "epoch": 2.9585292860196666, "grad_norm": 0.2016593632101422, "learning_rate": 1.41e-05, "loss": 0.0161, "step": 865 }, { "epoch": 2.9619495510902096, "grad_norm": 0.18720575761526204, "learning_rate": 1.41e-05, "loss": 0.014, "step": 866 }, { "epoch": 2.9653698161607522, "grad_norm": 0.1948235931544798, "learning_rate": 1.41e-05, "loss": 0.0142, "step": 867 }, { "epoch": 2.9687900812312953, "grad_norm": 0.18580729685744493, "learning_rate": 1.41e-05, "loss": 0.0148, "step": 868 }, { "epoch": 2.9722103463018383, "grad_norm": 0.1747215712068137, "learning_rate": 1.41e-05, "loss": 0.0148, "step": 869 }, { "epoch": 2.9756306113723814, "grad_norm": 0.21255162912353637, "learning_rate": 1.41e-05, "loss": 0.0149, "step": 870 }, { "epoch": 2.9790508764429244, "grad_norm": 0.18848472150934822, "learning_rate": 1.41e-05, "loss": 0.0141, "step": 871 }, { "epoch": 2.982471141513467, "grad_norm": 0.18016163548131123, "learning_rate": 1.41e-05, "loss": 0.0147, "step": 872 }, { "epoch": 2.9858914065840105, "grad_norm": 0.1795705522164676, "learning_rate": 1.41e-05, "loss": 0.014, "step": 873 }, { "epoch": 2.989311671654553, "grad_norm": 0.17401025878895968, "learning_rate": 1.41e-05, "loss": 0.0138, "step": 874 }, { "epoch": 2.992731936725096, "grad_norm": 0.1783730795841426, "learning_rate": 1.41e-05, "loss": 0.0132, "step": 875 }, { "epoch": 2.9961522017956392, "grad_norm": 0.20051631261532152, "learning_rate": 1.41e-05, "loss": 0.0166, "step": 876 }, { "epoch": 2.9995724668661823, "grad_norm": 0.2056542584608787, "learning_rate": 1.41e-05, "loss": 0.0155, "step": 877 } ], "logging_steps": 1.0, "max_steps": 1460, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 367251916062720.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }