{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.4036326942482341, "eval_steps": 20, "global_step": 400, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0010090817356205853, "grad_norm": 48.721710205078125, "learning_rate": 8e-08, "loss": 0.8681, "step": 1 }, { "epoch": 0.0020181634712411706, "grad_norm": 44.42015075683594, "learning_rate": 1.6e-07, "loss": 0.6615, "step": 2 }, { "epoch": 0.0030272452068617556, "grad_norm": 42.49191665649414, "learning_rate": 2.4e-07, "loss": 0.6742, "step": 3 }, { "epoch": 0.004036326942482341, "grad_norm": 40.36923599243164, "learning_rate": 3.2e-07, "loss": 0.7824, "step": 4 }, { "epoch": 0.005045408678102927, "grad_norm": 57.19468688964844, "learning_rate": 4e-07, "loss": 1.0009, "step": 5 }, { "epoch": 0.006054490413723511, "grad_norm": 48.78745651245117, "learning_rate": 4.8e-07, "loss": 0.7613, "step": 6 }, { "epoch": 0.007063572149344097, "grad_norm": 37.92594528198242, "learning_rate": 5.6e-07, "loss": 0.6615, "step": 7 }, { "epoch": 0.008072653884964682, "grad_norm": 37.87478256225586, "learning_rate": 6.4e-07, "loss": 0.6779, "step": 8 }, { "epoch": 0.009081735620585268, "grad_norm": 43.225589752197266, "learning_rate": 7.2e-07, "loss": 0.796, "step": 9 }, { "epoch": 0.010090817356205853, "grad_norm": 43.1926383972168, "learning_rate": 8e-07, "loss": 0.8526, "step": 10 }, { "epoch": 0.011099899091826439, "grad_norm": 42.08029556274414, "learning_rate": 8.799999999999999e-07, "loss": 0.6605, "step": 11 }, { "epoch": 0.012108980827447022, "grad_norm": 49.14653396606445, "learning_rate": 9.6e-07, "loss": 0.6774, "step": 12 }, { "epoch": 0.013118062563067608, "grad_norm": 47.238338470458984, "learning_rate": 1.04e-06, "loss": 0.7313, "step": 13 }, { "epoch": 0.014127144298688193, "grad_norm": 37.998931884765625, "learning_rate": 1.12e-06, "loss": 0.6244, "step": 14 }, { "epoch": 0.015136226034308779, "grad_norm": 37.945064544677734, "learning_rate": 1.2e-06, "loss": 0.5214, "step": 15 }, { "epoch": 0.016145307769929364, "grad_norm": 33.03001403808594, "learning_rate": 1.28e-06, "loss": 0.5042, "step": 16 }, { "epoch": 0.017154389505549948, "grad_norm": 32.11518096923828, "learning_rate": 1.3600000000000001e-06, "loss": 0.3594, "step": 17 }, { "epoch": 0.018163471241170535, "grad_norm": 30.64830207824707, "learning_rate": 1.44e-06, "loss": 0.3533, "step": 18 }, { "epoch": 0.01917255297679112, "grad_norm": 41.34456253051758, "learning_rate": 1.5199999999999998e-06, "loss": 0.6093, "step": 19 }, { "epoch": 0.020181634712411706, "grad_norm": 35.51686477661133, "learning_rate": 1.6e-06, "loss": 0.3241, "step": 20 }, { "epoch": 0.020181634712411706, "eval_accuracy": 0.7758229284903518, "eval_loss": 0.3856516480445862, "eval_runtime": 63.0132, "eval_samples_per_second": 27.962, "eval_steps_per_second": 3.507, "step": 20 }, { "epoch": 0.02119071644803229, "grad_norm": 36.574764251708984, "learning_rate": 1.6799999999999998e-06, "loss": 0.3611, "step": 21 }, { "epoch": 0.022199798183652877, "grad_norm": 42.98014831542969, "learning_rate": 1.7599999999999999e-06, "loss": 0.5187, "step": 22 }, { "epoch": 0.02320887991927346, "grad_norm": 32.25895690917969, "learning_rate": 1.84e-06, "loss": 0.3201, "step": 23 }, { "epoch": 0.024217961654894045, "grad_norm": 45.763553619384766, "learning_rate": 1.92e-06, "loss": 0.3977, "step": 24 }, { "epoch": 0.025227043390514632, "grad_norm": 41.3906135559082, "learning_rate": 2e-06, "loss": 0.3863, "step": 25 }, { "epoch": 0.026236125126135216, "grad_norm": 55.49371337890625, "learning_rate": 2.08e-06, "loss": 0.4371, "step": 26 }, { "epoch": 0.027245206861755803, "grad_norm": 49.18912124633789, "learning_rate": 2.16e-06, "loss": 0.4072, "step": 27 }, { "epoch": 0.028254288597376387, "grad_norm": 53.16490173339844, "learning_rate": 2.24e-06, "loss": 0.5366, "step": 28 }, { "epoch": 0.029263370332996974, "grad_norm": 58.15668869018555, "learning_rate": 2.32e-06, "loss": 0.5429, "step": 29 }, { "epoch": 0.030272452068617558, "grad_norm": 32.979671478271484, "learning_rate": 2.4e-06, "loss": 0.2946, "step": 30 }, { "epoch": 0.03128153380423814, "grad_norm": 40.131385803222656, "learning_rate": 2.48e-06, "loss": 0.298, "step": 31 }, { "epoch": 0.03229061553985873, "grad_norm": 45.999542236328125, "learning_rate": 2.56e-06, "loss": 0.27, "step": 32 }, { "epoch": 0.033299697275479316, "grad_norm": 32.2443962097168, "learning_rate": 2.64e-06, "loss": 0.2337, "step": 33 }, { "epoch": 0.034308779011099896, "grad_norm": 28.031023025512695, "learning_rate": 2.7200000000000002e-06, "loss": 0.1835, "step": 34 }, { "epoch": 0.035317860746720484, "grad_norm": 22.68515396118164, "learning_rate": 2.8e-06, "loss": 0.1522, "step": 35 }, { "epoch": 0.03632694248234107, "grad_norm": 38.576019287109375, "learning_rate": 2.88e-06, "loss": 0.3371, "step": 36 }, { "epoch": 0.03733602421796166, "grad_norm": 23.268003463745117, "learning_rate": 2.96e-06, "loss": 0.161, "step": 37 }, { "epoch": 0.03834510595358224, "grad_norm": 43.69157409667969, "learning_rate": 3.0399999999999997e-06, "loss": 0.4692, "step": 38 }, { "epoch": 0.039354187689202826, "grad_norm": 33.641780853271484, "learning_rate": 3.1199999999999998e-06, "loss": 0.3815, "step": 39 }, { "epoch": 0.04036326942482341, "grad_norm": 32.81157302856445, "learning_rate": 3.2e-06, "loss": 0.3342, "step": 40 }, { "epoch": 0.04036326942482341, "eval_accuracy": 0.8359818388195233, "eval_loss": 0.30243048071861267, "eval_runtime": 62.0758, "eval_samples_per_second": 28.385, "eval_steps_per_second": 3.56, "step": 40 }, { "epoch": 0.04137235116044399, "grad_norm": 38.70158004760742, "learning_rate": 3.2799999999999995e-06, "loss": 0.3523, "step": 41 }, { "epoch": 0.04238143289606458, "grad_norm": 36.827938079833984, "learning_rate": 3.3599999999999996e-06, "loss": 0.3339, "step": 42 }, { "epoch": 0.04339051463168517, "grad_norm": 40.072757720947266, "learning_rate": 3.4399999999999997e-06, "loss": 0.3718, "step": 43 }, { "epoch": 0.044399596367305755, "grad_norm": 17.283546447753906, "learning_rate": 3.5199999999999998e-06, "loss": 0.1344, "step": 44 }, { "epoch": 0.045408678102926335, "grad_norm": 38.25215148925781, "learning_rate": 3.6e-06, "loss": 0.3276, "step": 45 }, { "epoch": 0.04641775983854692, "grad_norm": 31.07087516784668, "learning_rate": 3.68e-06, "loss": 0.3091, "step": 46 }, { "epoch": 0.04742684157416751, "grad_norm": 29.617158889770508, "learning_rate": 3.7599999999999996e-06, "loss": 0.2113, "step": 47 }, { "epoch": 0.04843592330978809, "grad_norm": 37.85402297973633, "learning_rate": 3.84e-06, "loss": 0.3061, "step": 48 }, { "epoch": 0.04944500504540868, "grad_norm": 92.3378677368164, "learning_rate": 3.92e-06, "loss": 0.8413, "step": 49 }, { "epoch": 0.050454086781029264, "grad_norm": 25.68213653564453, "learning_rate": 4e-06, "loss": 0.2466, "step": 50 }, { "epoch": 0.05146316851664985, "grad_norm": 22.958641052246094, "learning_rate": 4.08e-06, "loss": 0.2127, "step": 51 }, { "epoch": 0.05247225025227043, "grad_norm": 26.674156188964844, "learning_rate": 4.16e-06, "loss": 0.5098, "step": 52 }, { "epoch": 0.05348133198789102, "grad_norm": 18.660629272460938, "learning_rate": 4.24e-06, "loss": 0.2416, "step": 53 }, { "epoch": 0.054490413723511606, "grad_norm": 20.310880661010742, "learning_rate": 4.32e-06, "loss": 0.3086, "step": 54 }, { "epoch": 0.055499495459132187, "grad_norm": 20.308359146118164, "learning_rate": 4.4e-06, "loss": 0.3001, "step": 55 }, { "epoch": 0.056508577194752774, "grad_norm": 13.419134140014648, "learning_rate": 4.48e-06, "loss": 0.1779, "step": 56 }, { "epoch": 0.05751765893037336, "grad_norm": 14.848164558410645, "learning_rate": 4.5599999999999995e-06, "loss": 0.1883, "step": 57 }, { "epoch": 0.05852674066599395, "grad_norm": 23.216663360595703, "learning_rate": 4.64e-06, "loss": 0.3184, "step": 58 }, { "epoch": 0.05953582240161453, "grad_norm": 65.48796081542969, "learning_rate": 4.72e-06, "loss": 0.3397, "step": 59 }, { "epoch": 0.060544904137235116, "grad_norm": 97.63284301757812, "learning_rate": 4.8e-06, "loss": 0.3936, "step": 60 }, { "epoch": 0.060544904137235116, "eval_accuracy": 0.7474460839954598, "eval_loss": 0.4690244793891907, "eval_runtime": 62.2233, "eval_samples_per_second": 28.317, "eval_steps_per_second": 3.552, "step": 60 }, { "epoch": 0.0615539858728557, "grad_norm": 210.90390014648438, "learning_rate": 4.88e-06, "loss": 0.6895, "step": 61 }, { "epoch": 0.06256306760847628, "grad_norm": 46.58690643310547, "learning_rate": 4.96e-06, "loss": 0.29, "step": 62 }, { "epoch": 0.06357214934409687, "grad_norm": 31.453752517700195, "learning_rate": 5.04e-06, "loss": 0.3071, "step": 63 }, { "epoch": 0.06458123107971746, "grad_norm": 20.283885955810547, "learning_rate": 5.12e-06, "loss": 0.3571, "step": 64 }, { "epoch": 0.06559031281533804, "grad_norm": 42.37957763671875, "learning_rate": 5.2e-06, "loss": 0.5532, "step": 65 }, { "epoch": 0.06659939455095863, "grad_norm": 21.637910842895508, "learning_rate": 5.28e-06, "loss": 0.262, "step": 66 }, { "epoch": 0.06760847628657922, "grad_norm": 18.54781723022461, "learning_rate": 5.36e-06, "loss": 0.2257, "step": 67 }, { "epoch": 0.06861755802219979, "grad_norm": 20.521879196166992, "learning_rate": 5.4400000000000004e-06, "loss": 0.3857, "step": 68 }, { "epoch": 0.06962663975782038, "grad_norm": 29.46757698059082, "learning_rate": 5.52e-06, "loss": 0.1029, "step": 69 }, { "epoch": 0.07063572149344097, "grad_norm": 29.944223403930664, "learning_rate": 5.6e-06, "loss": 0.2712, "step": 70 }, { "epoch": 0.07164480322906155, "grad_norm": 24.391925811767578, "learning_rate": 5.68e-06, "loss": 0.3976, "step": 71 }, { "epoch": 0.07265388496468214, "grad_norm": 18.12702751159668, "learning_rate": 5.76e-06, "loss": 0.3472, "step": 72 }, { "epoch": 0.07366296670030273, "grad_norm": 10.216832160949707, "learning_rate": 5.84e-06, "loss": 0.2769, "step": 73 }, { "epoch": 0.07467204843592332, "grad_norm": 11.782180786132812, "learning_rate": 5.92e-06, "loss": 0.3737, "step": 74 }, { "epoch": 0.07568113017154389, "grad_norm": 33.01165771484375, "learning_rate": 6e-06, "loss": 0.2062, "step": 75 }, { "epoch": 0.07669021190716448, "grad_norm": 16.44459342956543, "learning_rate": 6.079999999999999e-06, "loss": 0.4451, "step": 76 }, { "epoch": 0.07769929364278506, "grad_norm": 8.65665340423584, "learning_rate": 6.1599999999999995e-06, "loss": 0.2799, "step": 77 }, { "epoch": 0.07870837537840565, "grad_norm": 9.738832473754883, "learning_rate": 6.2399999999999995e-06, "loss": 0.2986, "step": 78 }, { "epoch": 0.07971745711402624, "grad_norm": 5.850776195526123, "learning_rate": 6.32e-06, "loss": 0.2548, "step": 79 }, { "epoch": 0.08072653884964683, "grad_norm": 11.245881080627441, "learning_rate": 6.4e-06, "loss": 0.3249, "step": 80 }, { "epoch": 0.08072653884964683, "eval_accuracy": 0.8569807037457434, "eval_loss": 0.2707608640193939, "eval_runtime": 62.6893, "eval_samples_per_second": 28.107, "eval_steps_per_second": 3.525, "step": 80 }, { "epoch": 0.08173562058526741, "grad_norm": 9.138853073120117, "learning_rate": 6.48e-06, "loss": 0.2931, "step": 81 }, { "epoch": 0.08274470232088799, "grad_norm": 7.408812999725342, "learning_rate": 6.559999999999999e-06, "loss": 0.276, "step": 82 }, { "epoch": 0.08375378405650857, "grad_norm": 5.13024377822876, "learning_rate": 6.639999999999999e-06, "loss": 0.1842, "step": 83 }, { "epoch": 0.08476286579212916, "grad_norm": 5.102560043334961, "learning_rate": 6.719999999999999e-06, "loss": 0.2041, "step": 84 }, { "epoch": 0.08577194752774975, "grad_norm": 8.902843475341797, "learning_rate": 6.799999999999999e-06, "loss": 0.3015, "step": 85 }, { "epoch": 0.08678102926337034, "grad_norm": 8.0729398727417, "learning_rate": 6.879999999999999e-06, "loss": 0.3637, "step": 86 }, { "epoch": 0.08779011099899092, "grad_norm": 7.424192428588867, "learning_rate": 6.9599999999999994e-06, "loss": 0.1691, "step": 87 }, { "epoch": 0.08879919273461151, "grad_norm": 19.509435653686523, "learning_rate": 7.0399999999999995e-06, "loss": 0.3888, "step": 88 }, { "epoch": 0.08980827447023208, "grad_norm": 13.97599983215332, "learning_rate": 7.12e-06, "loss": 0.2415, "step": 89 }, { "epoch": 0.09081735620585267, "grad_norm": 14.023358345031738, "learning_rate": 7.2e-06, "loss": 0.2889, "step": 90 }, { "epoch": 0.09182643794147326, "grad_norm": 15.243273735046387, "learning_rate": 7.28e-06, "loss": 0.4061, "step": 91 }, { "epoch": 0.09283551967709384, "grad_norm": 9.57010555267334, "learning_rate": 7.36e-06, "loss": 0.2501, "step": 92 }, { "epoch": 0.09384460141271443, "grad_norm": 20.639694213867188, "learning_rate": 7.44e-06, "loss": 0.2877, "step": 93 }, { "epoch": 0.09485368314833502, "grad_norm": 75.29905700683594, "learning_rate": 7.519999999999999e-06, "loss": 0.4536, "step": 94 }, { "epoch": 0.0958627648839556, "grad_norm": 340.7889709472656, "learning_rate": 7.599999999999999e-06, "loss": 0.4443, "step": 95 }, { "epoch": 0.09687184661957618, "grad_norm": 16.866931915283203, "learning_rate": 7.68e-06, "loss": 0.292, "step": 96 }, { "epoch": 0.09788092835519677, "grad_norm": 12.300322532653809, "learning_rate": 7.76e-06, "loss": 0.2191, "step": 97 }, { "epoch": 0.09889001009081735, "grad_norm": 18.15094566345215, "learning_rate": 7.84e-06, "loss": 0.1821, "step": 98 }, { "epoch": 0.09989909182643794, "grad_norm": 31.187374114990234, "learning_rate": 7.92e-06, "loss": 0.1823, "step": 99 }, { "epoch": 0.10090817356205853, "grad_norm": 64.1741943359375, "learning_rate": 8e-06, "loss": 0.5276, "step": 100 }, { "epoch": 0.10090817356205853, "eval_accuracy": 0.8467650397275823, "eval_loss": 0.28612375259399414, "eval_runtime": 62.505, "eval_samples_per_second": 28.19, "eval_steps_per_second": 3.536, "step": 100 }, { "epoch": 0.10191725529767912, "grad_norm": 12.569557189941406, "learning_rate": 7.999975135834775e-06, "loss": 0.3077, "step": 101 }, { "epoch": 0.1029263370332997, "grad_norm": 7.0805277824401855, "learning_rate": 7.999900543648217e-06, "loss": 0.1759, "step": 102 }, { "epoch": 0.10393541876892028, "grad_norm": 9.746481895446777, "learning_rate": 7.999776224367659e-06, "loss": 0.2064, "step": 103 }, { "epoch": 0.10494450050454086, "grad_norm": 6.783945560455322, "learning_rate": 7.999602179538651e-06, "loss": 0.3028, "step": 104 }, { "epoch": 0.10595358224016145, "grad_norm": 4.683195114135742, "learning_rate": 7.999378411324933e-06, "loss": 0.1767, "step": 105 }, { "epoch": 0.10696266397578204, "grad_norm": 6.255279064178467, "learning_rate": 7.999104922508408e-06, "loss": 0.2984, "step": 106 }, { "epoch": 0.10797174571140263, "grad_norm": 5.866703987121582, "learning_rate": 7.99878171648911e-06, "loss": 0.2472, "step": 107 }, { "epoch": 0.10898082744702321, "grad_norm": 5.773608207702637, "learning_rate": 7.998408797285167e-06, "loss": 0.225, "step": 108 }, { "epoch": 0.1099899091826438, "grad_norm": 10.58280086517334, "learning_rate": 7.99798616953274e-06, "loss": 0.3962, "step": 109 }, { "epoch": 0.11099899091826437, "grad_norm": 5.545925140380859, "learning_rate": 7.997513838485971e-06, "loss": 0.1597, "step": 110 }, { "epoch": 0.11200807265388496, "grad_norm": 7.84637975692749, "learning_rate": 7.99699181001692e-06, "loss": 0.346, "step": 111 }, { "epoch": 0.11301715438950555, "grad_norm": 5.827946662902832, "learning_rate": 7.996420090615486e-06, "loss": 0.2507, "step": 112 }, { "epoch": 0.11402623612512613, "grad_norm": 8.125126838684082, "learning_rate": 7.995798687389334e-06, "loss": 0.2496, "step": 113 }, { "epoch": 0.11503531786074672, "grad_norm": 6.233455181121826, "learning_rate": 7.9951276080638e-06, "loss": 0.2129, "step": 114 }, { "epoch": 0.11604439959636731, "grad_norm": 8.677657127380371, "learning_rate": 7.994406860981797e-06, "loss": 0.3926, "step": 115 }, { "epoch": 0.1170534813319879, "grad_norm": 6.164268493652344, "learning_rate": 7.99363645510371e-06, "loss": 0.2335, "step": 116 }, { "epoch": 0.11806256306760847, "grad_norm": 5.605584144592285, "learning_rate": 7.992816400007294e-06, "loss": 0.2357, "step": 117 }, { "epoch": 0.11907164480322906, "grad_norm": 8.481727600097656, "learning_rate": 7.991946705887537e-06, "loss": 0.4145, "step": 118 }, { "epoch": 0.12008072653884964, "grad_norm": 7.141845226287842, "learning_rate": 7.99102738355655e-06, "loss": 0.3807, "step": 119 }, { "epoch": 0.12108980827447023, "grad_norm": 22.80964469909668, "learning_rate": 7.990058444443424e-06, "loss": 0.259, "step": 120 }, { "epoch": 0.12108980827447023, "eval_accuracy": 0.7729852440408627, "eval_loss": 0.44569578766822815, "eval_runtime": 62.4809, "eval_samples_per_second": 28.201, "eval_steps_per_second": 3.537, "step": 120 }, { "epoch": 0.12209889001009082, "grad_norm": 56.96179962158203, "learning_rate": 7.989039900594089e-06, "loss": 0.4721, "step": 121 }, { "epoch": 0.1231079717457114, "grad_norm": 9.661979675292969, "learning_rate": 7.987971764671168e-06, "loss": 0.2308, "step": 122 }, { "epoch": 0.124117053481332, "grad_norm": 20.324970245361328, "learning_rate": 7.986854049953814e-06, "loss": 0.4019, "step": 123 }, { "epoch": 0.12512613521695257, "grad_norm": 7.678678512573242, "learning_rate": 7.98568677033755e-06, "loss": 0.1287, "step": 124 }, { "epoch": 0.12613521695257315, "grad_norm": 5.97120475769043, "learning_rate": 7.984469940334089e-06, "loss": 0.3056, "step": 125 }, { "epoch": 0.12714429868819374, "grad_norm": 4.4789509773254395, "learning_rate": 7.983203575071166e-06, "loss": 0.156, "step": 126 }, { "epoch": 0.12815338042381433, "grad_norm": 7.833422660827637, "learning_rate": 7.981887690292338e-06, "loss": 0.3012, "step": 127 }, { "epoch": 0.12916246215943492, "grad_norm": 4.627663612365723, "learning_rate": 7.980522302356792e-06, "loss": 0.1701, "step": 128 }, { "epoch": 0.1301715438950555, "grad_norm": 23.75330924987793, "learning_rate": 7.979107428239143e-06, "loss": 0.5042, "step": 129 }, { "epoch": 0.1311806256306761, "grad_norm": 8.093461036682129, "learning_rate": 7.977643085529227e-06, "loss": 0.3362, "step": 130 }, { "epoch": 0.13218970736629668, "grad_norm": 8.479804992675781, "learning_rate": 7.97612929243187e-06, "loss": 0.2005, "step": 131 }, { "epoch": 0.13319878910191726, "grad_norm": 8.889948844909668, "learning_rate": 7.974566067766671e-06, "loss": 0.4108, "step": 132 }, { "epoch": 0.13420787083753785, "grad_norm": 6.473036289215088, "learning_rate": 7.972953430967771e-06, "loss": 0.3489, "step": 133 }, { "epoch": 0.13521695257315844, "grad_norm": 5.9677863121032715, "learning_rate": 7.971291402083606e-06, "loss": 0.2707, "step": 134 }, { "epoch": 0.136226034308779, "grad_norm": 9.331769943237305, "learning_rate": 7.969580001776653e-06, "loss": 0.4528, "step": 135 }, { "epoch": 0.13723511604439959, "grad_norm": 6.988079071044922, "learning_rate": 7.96781925132318e-06, "loss": 0.3895, "step": 136 }, { "epoch": 0.13824419778002017, "grad_norm": 5.6570281982421875, "learning_rate": 7.966009172612988e-06, "loss": 0.2324, "step": 137 }, { "epoch": 0.13925327951564076, "grad_norm": 7.507450103759766, "learning_rate": 7.964149788149122e-06, "loss": 0.3709, "step": 138 }, { "epoch": 0.14026236125126135, "grad_norm": 6.912793159484863, "learning_rate": 7.962241121047602e-06, "loss": 0.2004, "step": 139 }, { "epoch": 0.14127144298688193, "grad_norm": 7.456505298614502, "learning_rate": 7.960283195037138e-06, "loss": 0.1852, "step": 140 }, { "epoch": 0.14127144298688193, "eval_accuracy": 0.8382519863791147, "eval_loss": 0.30151110887527466, "eval_runtime": 59.0753, "eval_samples_per_second": 29.826, "eval_steps_per_second": 3.741, "step": 140 }, { "epoch": 0.14228052472250252, "grad_norm": 5.073808670043945, "learning_rate": 7.958276034458826e-06, "loss": 0.2379, "step": 141 }, { "epoch": 0.1432896064581231, "grad_norm": 10.782764434814453, "learning_rate": 7.956219664265852e-06, "loss": 0.1299, "step": 142 }, { "epoch": 0.1442986881937437, "grad_norm": 5.916919708251953, "learning_rate": 7.95411411002318e-06, "loss": 0.2939, "step": 143 }, { "epoch": 0.14530776992936428, "grad_norm": 4.290775775909424, "learning_rate": 7.951959397907236e-06, "loss": 0.2175, "step": 144 }, { "epoch": 0.14631685166498487, "grad_norm": 6.01127290725708, "learning_rate": 7.949755554705577e-06, "loss": 0.2962, "step": 145 }, { "epoch": 0.14732593340060546, "grad_norm": 5.430018901824951, "learning_rate": 7.947502607816566e-06, "loss": 0.3358, "step": 146 }, { "epoch": 0.14833501513622604, "grad_norm": 2.893122911453247, "learning_rate": 7.945200585249022e-06, "loss": 0.0949, "step": 147 }, { "epoch": 0.14934409687184663, "grad_norm": 5.385425567626953, "learning_rate": 7.942849515621881e-06, "loss": 0.314, "step": 148 }, { "epoch": 0.1503531786074672, "grad_norm": 4.384002208709717, "learning_rate": 7.940449428163837e-06, "loss": 0.2312, "step": 149 }, { "epoch": 0.15136226034308778, "grad_norm": 5.955286979675293, "learning_rate": 7.938000352712972e-06, "loss": 0.1949, "step": 150 }, { "epoch": 0.15237134207870837, "grad_norm": 6.1054487228393555, "learning_rate": 7.935502319716397e-06, "loss": 0.3697, "step": 151 }, { "epoch": 0.15338042381432895, "grad_norm": 3.2285423278808594, "learning_rate": 7.932955360229862e-06, "loss": 0.1118, "step": 152 }, { "epoch": 0.15438950554994954, "grad_norm": 7.409282684326172, "learning_rate": 7.930359505917381e-06, "loss": 0.4898, "step": 153 }, { "epoch": 0.15539858728557013, "grad_norm": 4.591885566711426, "learning_rate": 7.927714789050827e-06, "loss": 0.2288, "step": 154 }, { "epoch": 0.15640766902119072, "grad_norm": 5.779257297515869, "learning_rate": 7.925021242509538e-06, "loss": 0.2563, "step": 155 }, { "epoch": 0.1574167507568113, "grad_norm": 4.496645927429199, "learning_rate": 7.92227889977991e-06, "loss": 0.2096, "step": 156 }, { "epoch": 0.1584258324924319, "grad_norm": 6.980607509613037, "learning_rate": 7.919487794954972e-06, "loss": 0.3658, "step": 157 }, { "epoch": 0.15943491422805248, "grad_norm": 5.738483905792236, "learning_rate": 7.91664796273397e-06, "loss": 0.3123, "step": 158 }, { "epoch": 0.16044399596367306, "grad_norm": 6.03507137298584, "learning_rate": 7.913759438421932e-06, "loss": 0.3345, "step": 159 }, { "epoch": 0.16145307769929365, "grad_norm": 31.2845458984375, "learning_rate": 7.910822257929234e-06, "loss": 0.1122, "step": 160 }, { "epoch": 0.16145307769929365, "eval_accuracy": 0.8518728717366629, "eval_loss": 0.38947635889053345, "eval_runtime": 62.3612, "eval_samples_per_second": 28.255, "eval_steps_per_second": 3.544, "step": 160 }, { "epoch": 0.16246215943491424, "grad_norm": 6.148342609405518, "learning_rate": 7.907836457771143e-06, "loss": 0.2418, "step": 161 }, { "epoch": 0.16347124117053483, "grad_norm": 38.332550048828125, "learning_rate": 7.904802075067377e-06, "loss": 0.6998, "step": 162 }, { "epoch": 0.16448032290615539, "grad_norm": 8.149801254272461, "learning_rate": 7.901719147541628e-06, "loss": 0.1255, "step": 163 }, { "epoch": 0.16548940464177597, "grad_norm": 19.77474021911621, "learning_rate": 7.898587713521109e-06, "loss": 0.6065, "step": 164 }, { "epoch": 0.16649848637739656, "grad_norm": 5.291502952575684, "learning_rate": 7.895407811936064e-06, "loss": 0.1659, "step": 165 }, { "epoch": 0.16750756811301715, "grad_norm": 5.293869495391846, "learning_rate": 7.892179482319294e-06, "loss": 0.2073, "step": 166 }, { "epoch": 0.16851664984863773, "grad_norm": 6.291193962097168, "learning_rate": 7.88890276480566e-06, "loss": 0.3597, "step": 167 }, { "epoch": 0.16952573158425832, "grad_norm": 5.216683864593506, "learning_rate": 7.885577700131584e-06, "loss": 0.2888, "step": 168 }, { "epoch": 0.1705348133198789, "grad_norm": 3.573047637939453, "learning_rate": 7.882204329634543e-06, "loss": 0.1343, "step": 169 }, { "epoch": 0.1715438950554995, "grad_norm": 5.618640899658203, "learning_rate": 7.878782695252562e-06, "loss": 0.2477, "step": 170 }, { "epoch": 0.17255297679112008, "grad_norm": 6.469369411468506, "learning_rate": 7.875312839523677e-06, "loss": 0.3563, "step": 171 }, { "epoch": 0.17356205852674067, "grad_norm": 7.982197284698486, "learning_rate": 7.871794805585425e-06, "loss": 0.2748, "step": 172 }, { "epoch": 0.17457114026236126, "grad_norm": 8.727740287780762, "learning_rate": 7.868228637174292e-06, "loss": 0.2539, "step": 173 }, { "epoch": 0.17558022199798184, "grad_norm": 9.52997875213623, "learning_rate": 7.86461437862518e-06, "loss": 0.2352, "step": 174 }, { "epoch": 0.17658930373360243, "grad_norm": 8.447443008422852, "learning_rate": 7.86095207487085e-06, "loss": 0.3129, "step": 175 }, { "epoch": 0.17759838546922302, "grad_norm": 7.929843425750732, "learning_rate": 7.857241771441364e-06, "loss": 0.3138, "step": 176 }, { "epoch": 0.17860746720484358, "grad_norm": 7.068841934204102, "learning_rate": 7.853483514463521e-06, "loss": 0.1881, "step": 177 }, { "epoch": 0.17961654894046417, "grad_norm": 179.09567260742188, "learning_rate": 7.849677350660282e-06, "loss": 1.6187, "step": 178 }, { "epoch": 0.18062563067608475, "grad_norm": 14.378893852233887, "learning_rate": 7.84582332735019e-06, "loss": 0.44, "step": 179 }, { "epoch": 0.18163471241170534, "grad_norm": 10.581100463867188, "learning_rate": 7.841921492446781e-06, "loss": 0.1962, "step": 180 }, { "epoch": 0.18163471241170534, "eval_accuracy": 0.8229284903518729, "eval_loss": 0.2842705547809601, "eval_runtime": 63.0695, "eval_samples_per_second": 27.937, "eval_steps_per_second": 3.504, "step": 180 }, { "epoch": 0.18264379414732593, "grad_norm": 13.833659172058105, "learning_rate": 7.837971894457989e-06, "loss": 0.4059, "step": 181 }, { "epoch": 0.18365287588294651, "grad_norm": 6.910711288452148, "learning_rate": 7.833974582485544e-06, "loss": 0.3115, "step": 182 }, { "epoch": 0.1846619576185671, "grad_norm": 4.3872599601745605, "learning_rate": 7.829929606224356e-06, "loss": 0.1654, "step": 183 }, { "epoch": 0.1856710393541877, "grad_norm": 8.301304817199707, "learning_rate": 7.825837015961904e-06, "loss": 0.3223, "step": 184 }, { "epoch": 0.18668012108980828, "grad_norm": 6.772916316986084, "learning_rate": 7.82169686257761e-06, "loss": 0.207, "step": 185 }, { "epoch": 0.18768920282542886, "grad_norm": 10.749194145202637, "learning_rate": 7.817509197542204e-06, "loss": 0.3322, "step": 186 }, { "epoch": 0.18869828456104945, "grad_norm": 18.030078887939453, "learning_rate": 7.813274072917081e-06, "loss": 0.3675, "step": 187 }, { "epoch": 0.18970736629667004, "grad_norm": 15.416492462158203, "learning_rate": 7.80899154135366e-06, "loss": 0.2049, "step": 188 }, { "epoch": 0.19071644803229063, "grad_norm": 24.75461769104004, "learning_rate": 7.80466165609273e-06, "loss": 0.36, "step": 189 }, { "epoch": 0.1917255297679112, "grad_norm": 11.484724044799805, "learning_rate": 7.800284470963781e-06, "loss": 0.2014, "step": 190 }, { "epoch": 0.1927346115035318, "grad_norm": 19.571962356567383, "learning_rate": 7.795860040384339e-06, "loss": 0.5687, "step": 191 }, { "epoch": 0.19374369323915236, "grad_norm": 6.390493392944336, "learning_rate": 7.791388419359292e-06, "loss": 0.2563, "step": 192 }, { "epoch": 0.19475277497477295, "grad_norm": 9.572249412536621, "learning_rate": 7.786869663480201e-06, "loss": 0.3565, "step": 193 }, { "epoch": 0.19576185671039353, "grad_norm": 8.426647186279297, "learning_rate": 7.782303828924613e-06, "loss": 0.5456, "step": 194 }, { "epoch": 0.19677093844601412, "grad_norm": 14.882564544677734, "learning_rate": 7.77769097245536e-06, "loss": 0.2976, "step": 195 }, { "epoch": 0.1977800201816347, "grad_norm": 10.191393852233887, "learning_rate": 7.773031151419853e-06, "loss": 0.4641, "step": 196 }, { "epoch": 0.1987891019172553, "grad_norm": 7.604254245758057, "learning_rate": 7.768324423749376e-06, "loss": 0.2364, "step": 197 }, { "epoch": 0.19979818365287588, "grad_norm": 9.080754280090332, "learning_rate": 7.763570847958354e-06, "loss": 0.1754, "step": 198 }, { "epoch": 0.20080726538849647, "grad_norm": 10.90381908416748, "learning_rate": 7.758770483143633e-06, "loss": 0.1799, "step": 199 }, { "epoch": 0.20181634712411706, "grad_norm": 12.508880615234375, "learning_rate": 7.753923388983747e-06, "loss": 0.2855, "step": 200 }, { "epoch": 0.20181634712411706, "eval_accuracy": 0.840522133938706, "eval_loss": 0.29473429918289185, "eval_runtime": 62.4643, "eval_samples_per_second": 28.208, "eval_steps_per_second": 3.538, "step": 200 }, { "epoch": 0.20282542885973764, "grad_norm": 20.691980361938477, "learning_rate": 7.749029625738169e-06, "loss": 0.1942, "step": 201 }, { "epoch": 0.20383451059535823, "grad_norm": 10.143714904785156, "learning_rate": 7.744089254246569e-06, "loss": 0.2001, "step": 202 }, { "epoch": 0.20484359233097882, "grad_norm": 17.92678451538086, "learning_rate": 7.739102335928053e-06, "loss": 0.3823, "step": 203 }, { "epoch": 0.2058526740665994, "grad_norm": 12.694001197814941, "learning_rate": 7.734068932780405e-06, "loss": 0.274, "step": 204 }, { "epoch": 0.20686175580222, "grad_norm": 12.30357837677002, "learning_rate": 7.728989107379303e-06, "loss": 0.3017, "step": 205 }, { "epoch": 0.20787083753784055, "grad_norm": 7.952809810638428, "learning_rate": 7.72386292287756e-06, "loss": 0.14, "step": 206 }, { "epoch": 0.20887991927346114, "grad_norm": 69.02944946289062, "learning_rate": 7.718690443004324e-06, "loss": 0.4633, "step": 207 }, { "epoch": 0.20988900100908173, "grad_norm": 92.96024322509766, "learning_rate": 7.71347173206429e-06, "loss": 1.1202, "step": 208 }, { "epoch": 0.21089808274470231, "grad_norm": 83.07585906982422, "learning_rate": 7.708206854936908e-06, "loss": 0.3752, "step": 209 }, { "epoch": 0.2119071644803229, "grad_norm": 10.343903541564941, "learning_rate": 7.702895877075563e-06, "loss": 0.3543, "step": 210 }, { "epoch": 0.2129162462159435, "grad_norm": 34.449188232421875, "learning_rate": 7.697538864506767e-06, "loss": 0.5488, "step": 211 }, { "epoch": 0.21392532795156408, "grad_norm": 5.609644412994385, "learning_rate": 7.692135883829349e-06, "loss": 0.342, "step": 212 }, { "epoch": 0.21493440968718466, "grad_norm": 6.129890441894531, "learning_rate": 7.686687002213609e-06, "loss": 0.1932, "step": 213 }, { "epoch": 0.21594349142280525, "grad_norm": 17.425050735473633, "learning_rate": 7.681192287400491e-06, "loss": 0.4084, "step": 214 }, { "epoch": 0.21695257315842584, "grad_norm": 24.733633041381836, "learning_rate": 7.675651807700748e-06, "loss": 0.2141, "step": 215 }, { "epoch": 0.21796165489404642, "grad_norm": 33.499794006347656, "learning_rate": 7.670065631994078e-06, "loss": 0.5658, "step": 216 }, { "epoch": 0.218970736629667, "grad_norm": 9.948688507080078, "learning_rate": 7.664433829728277e-06, "loss": 0.2564, "step": 217 }, { "epoch": 0.2199798183652876, "grad_norm": 9.255337715148926, "learning_rate": 7.658756470918382e-06, "loss": 0.2294, "step": 218 }, { "epoch": 0.2209889001009082, "grad_norm": 30.673681259155273, "learning_rate": 7.65303362614578e-06, "loss": 0.4334, "step": 219 }, { "epoch": 0.22199798183652875, "grad_norm": 5.19248628616333, "learning_rate": 7.647265366557355e-06, "loss": 0.1593, "step": 220 }, { "epoch": 0.22199798183652875, "eval_accuracy": 0.8382519863791147, "eval_loss": 0.29325050115585327, "eval_runtime": 62.9477, "eval_samples_per_second": 27.991, "eval_steps_per_second": 3.511, "step": 220 }, { "epoch": 0.22300706357214933, "grad_norm": 5.615384101867676, "learning_rate": 7.641451763864587e-06, "loss": 0.225, "step": 221 }, { "epoch": 0.22401614530776992, "grad_norm": 6.7938032150268555, "learning_rate": 7.63559289034266e-06, "loss": 0.2151, "step": 222 }, { "epoch": 0.2250252270433905, "grad_norm": 2.8786470890045166, "learning_rate": 7.629688818829577e-06, "loss": 0.1089, "step": 223 }, { "epoch": 0.2260343087790111, "grad_norm": 7.671812057495117, "learning_rate": 7.623739622725244e-06, "loss": 0.289, "step": 224 }, { "epoch": 0.22704339051463168, "grad_norm": 5.496854782104492, "learning_rate": 7.617745375990556e-06, "loss": 0.24, "step": 225 }, { "epoch": 0.22805247225025227, "grad_norm": 6.458908557891846, "learning_rate": 7.611706153146485e-06, "loss": 0.3702, "step": 226 }, { "epoch": 0.22906155398587286, "grad_norm": 10.48363971710205, "learning_rate": 7.605622029273148e-06, "loss": 0.1634, "step": 227 }, { "epoch": 0.23007063572149344, "grad_norm": 32.09809875488281, "learning_rate": 7.599493080008873e-06, "loss": 0.7119, "step": 228 }, { "epoch": 0.23107971745711403, "grad_norm": 43.372257232666016, "learning_rate": 7.5933193815492675e-06, "loss": 0.4775, "step": 229 }, { "epoch": 0.23208879919273462, "grad_norm": 21.043750762939453, "learning_rate": 7.587101010646259e-06, "loss": 0.3401, "step": 230 }, { "epoch": 0.2330978809283552, "grad_norm": 13.805304527282715, "learning_rate": 7.58083804460715e-06, "loss": 0.1592, "step": 231 }, { "epoch": 0.2341069626639758, "grad_norm": 6.074002265930176, "learning_rate": 7.574530561293649e-06, "loss": 0.3836, "step": 232 }, { "epoch": 0.23511604439959638, "grad_norm": 4.921712875366211, "learning_rate": 7.5681786391209105e-06, "loss": 0.2796, "step": 233 }, { "epoch": 0.23612512613521694, "grad_norm": 5.469928741455078, "learning_rate": 7.561782357056557e-06, "loss": 0.2877, "step": 234 }, { "epoch": 0.23713420787083753, "grad_norm": 5.309055328369141, "learning_rate": 7.555341794619694e-06, "loss": 0.2664, "step": 235 }, { "epoch": 0.23814328960645811, "grad_norm": 55.794979095458984, "learning_rate": 7.548857031879926e-06, "loss": 0.1148, "step": 236 }, { "epoch": 0.2391523713420787, "grad_norm": 207.49688720703125, "learning_rate": 7.5423281494563595e-06, "loss": 0.3024, "step": 237 }, { "epoch": 0.2401614530776993, "grad_norm": 13.331132888793945, "learning_rate": 7.535755228516601e-06, "loss": 0.353, "step": 238 }, { "epoch": 0.24117053481331988, "grad_norm": 9.264041900634766, "learning_rate": 7.529138350775745e-06, "loss": 0.3563, "step": 239 }, { "epoch": 0.24217961654894046, "grad_norm": 4.584170818328857, "learning_rate": 7.522477598495363e-06, "loss": 0.2153, "step": 240 }, { "epoch": 0.24217961654894046, "eval_accuracy": 0.8376844494892168, "eval_loss": 0.2747989892959595, "eval_runtime": 62.479, "eval_samples_per_second": 28.201, "eval_steps_per_second": 3.537, "step": 240 }, { "epoch": 0.24318869828456105, "grad_norm": 4.465677738189697, "learning_rate": 7.515773054482478e-06, "loss": 0.2503, "step": 241 }, { "epoch": 0.24419778002018164, "grad_norm": 4.346131324768066, "learning_rate": 7.509024802088534e-06, "loss": 0.2141, "step": 242 }, { "epoch": 0.24520686175580222, "grad_norm": 7.381392002105713, "learning_rate": 7.502232925208363e-06, "loss": 0.2984, "step": 243 }, { "epoch": 0.2462159434914228, "grad_norm": 4.441753387451172, "learning_rate": 7.49539750827914e-06, "loss": 0.2058, "step": 244 }, { "epoch": 0.2472250252270434, "grad_norm": 7.497750282287598, "learning_rate": 7.488518636279331e-06, "loss": 0.4537, "step": 245 }, { "epoch": 0.248234106962664, "grad_norm": 9.178194046020508, "learning_rate": 7.4815963947276436e-06, "loss": 0.2963, "step": 246 }, { "epoch": 0.24924318869828457, "grad_norm": 7.65831184387207, "learning_rate": 7.474630869681954e-06, "loss": 0.3374, "step": 247 }, { "epoch": 0.25025227043390513, "grad_norm": 4.4074482917785645, "learning_rate": 7.467622147738246e-06, "loss": 0.176, "step": 248 }, { "epoch": 0.2512613521695257, "grad_norm": 3.8703436851501465, "learning_rate": 7.4605703160295315e-06, "loss": 0.1294, "step": 249 }, { "epoch": 0.2522704339051463, "grad_norm": 14.600114822387695, "learning_rate": 7.453475462224763e-06, "loss": 0.2959, "step": 250 }, { "epoch": 0.2532795156407669, "grad_norm": 7.458588600158691, "learning_rate": 7.44633767452775e-06, "loss": 0.3291, "step": 251 }, { "epoch": 0.2542885973763875, "grad_norm": 13.63289737701416, "learning_rate": 7.439157041676058e-06, "loss": 0.4166, "step": 252 }, { "epoch": 0.25529767911200807, "grad_norm": 8.358410835266113, "learning_rate": 7.431933652939908e-06, "loss": 0.257, "step": 253 }, { "epoch": 0.25630676084762866, "grad_norm": 5.803023338317871, "learning_rate": 7.424667598121065e-06, "loss": 0.2334, "step": 254 }, { "epoch": 0.25731584258324924, "grad_norm": 5.000588893890381, "learning_rate": 7.4173589675517245e-06, "loss": 0.2193, "step": 255 }, { "epoch": 0.25832492431886983, "grad_norm": 5.990160942077637, "learning_rate": 7.410007852093384e-06, "loss": 0.31, "step": 256 }, { "epoch": 0.2593340060544904, "grad_norm": 6.175328731536865, "learning_rate": 7.40261434313572e-06, "loss": 0.307, "step": 257 }, { "epoch": 0.260343087790111, "grad_norm": 6.047603130340576, "learning_rate": 7.395178532595444e-06, "loss": 0.223, "step": 258 }, { "epoch": 0.2613521695257316, "grad_norm": 5.447187423706055, "learning_rate": 7.387700512915168e-06, "loss": 0.0597, "step": 259 }, { "epoch": 0.2623612512613522, "grad_norm": 6.375972747802734, "learning_rate": 7.380180377062251e-06, "loss": 0.398, "step": 260 }, { "epoch": 0.2623612512613522, "eval_accuracy": 0.8320090805902384, "eval_loss": 0.2793748378753662, "eval_runtime": 62.5384, "eval_samples_per_second": 28.175, "eval_steps_per_second": 3.534, "step": 260 }, { "epoch": 0.26337033299697277, "grad_norm": 5.560431003570557, "learning_rate": 7.372618218527644e-06, "loss": 0.2861, "step": 261 }, { "epoch": 0.26437941473259335, "grad_norm": 5.246089935302734, "learning_rate": 7.365014131324725e-06, "loss": 0.1859, "step": 262 }, { "epoch": 0.26538849646821394, "grad_norm": 7.6310625076293945, "learning_rate": 7.3573682099881345e-06, "loss": 0.3164, "step": 263 }, { "epoch": 0.26639757820383453, "grad_norm": 6.964735507965088, "learning_rate": 7.349680549572598e-06, "loss": 0.3655, "step": 264 }, { "epoch": 0.2674066599394551, "grad_norm": 4.680076599121094, "learning_rate": 7.3419512456517455e-06, "loss": 0.1872, "step": 265 }, { "epoch": 0.2684157416750757, "grad_norm": 3.2980797290802, "learning_rate": 7.3341803943169214e-06, "loss": 0.0983, "step": 266 }, { "epoch": 0.2694248234106963, "grad_norm": 9.682284355163574, "learning_rate": 7.326368092175993e-06, "loss": 0.1637, "step": 267 }, { "epoch": 0.2704339051463169, "grad_norm": 58.10804748535156, "learning_rate": 7.3185144363521435e-06, "loss": 1.2893, "step": 268 }, { "epoch": 0.2714429868819374, "grad_norm": 3.9088146686553955, "learning_rate": 7.310619524482673e-06, "loss": 0.1103, "step": 269 }, { "epoch": 0.272452068617558, "grad_norm": 8.596792221069336, "learning_rate": 7.302683454717778e-06, "loss": 0.4524, "step": 270 }, { "epoch": 0.2734611503531786, "grad_norm": 18.94976234436035, "learning_rate": 7.294706325719331e-06, "loss": 0.3868, "step": 271 }, { "epoch": 0.27447023208879917, "grad_norm": 28.618101119995117, "learning_rate": 7.28668823665966e-06, "loss": 1.2269, "step": 272 }, { "epoch": 0.27547931382441976, "grad_norm": 7.027772903442383, "learning_rate": 7.2786292872203125e-06, "loss": 0.2546, "step": 273 }, { "epoch": 0.27648839556004035, "grad_norm": 4.971069812774658, "learning_rate": 7.270529577590812e-06, "loss": 0.2229, "step": 274 }, { "epoch": 0.27749747729566093, "grad_norm": 7.468736171722412, "learning_rate": 7.262389208467417e-06, "loss": 0.3655, "step": 275 }, { "epoch": 0.2785065590312815, "grad_norm": 7.4164509773254395, "learning_rate": 7.2542082810518696e-06, "loss": 0.3047, "step": 276 }, { "epoch": 0.2795156407669021, "grad_norm": 6.213995933532715, "learning_rate": 7.245986897050137e-06, "loss": 0.2879, "step": 277 }, { "epoch": 0.2805247225025227, "grad_norm": 6.1668548583984375, "learning_rate": 7.237725158671141e-06, "loss": 0.3405, "step": 278 }, { "epoch": 0.2815338042381433, "grad_norm": 11.742606163024902, "learning_rate": 7.229423168625498e-06, "loss": 0.3894, "step": 279 }, { "epoch": 0.28254288597376387, "grad_norm": 5.1043477058410645, "learning_rate": 7.2210810301242345e-06, "loss": 0.199, "step": 280 }, { "epoch": 0.28254288597376387, "eval_accuracy": 0.8382519863791147, "eval_loss": 0.28534775972366333, "eval_runtime": 62.6946, "eval_samples_per_second": 28.104, "eval_steps_per_second": 3.525, "step": 280 }, { "epoch": 0.28355196770938446, "grad_norm": 5.588554859161377, "learning_rate": 7.212698846877503e-06, "loss": 0.347, "step": 281 }, { "epoch": 0.28456104944500504, "grad_norm": 4.220617294311523, "learning_rate": 7.204276723093301e-06, "loss": 0.2267, "step": 282 }, { "epoch": 0.28557013118062563, "grad_norm": 36.09522247314453, "learning_rate": 7.195814763476164e-06, "loss": 0.6504, "step": 283 }, { "epoch": 0.2865792129162462, "grad_norm": 5.161518096923828, "learning_rate": 7.187313073225876e-06, "loss": 0.2736, "step": 284 }, { "epoch": 0.2875882946518668, "grad_norm": 5.858831882476807, "learning_rate": 7.178771758036152e-06, "loss": 0.3758, "step": 285 }, { "epoch": 0.2885973763874874, "grad_norm": 14.579549789428711, "learning_rate": 7.170190924093326e-06, "loss": 0.3338, "step": 286 }, { "epoch": 0.289606458123108, "grad_norm": 6.289809703826904, "learning_rate": 7.161570678075037e-06, "loss": 0.4051, "step": 287 }, { "epoch": 0.29061553985872857, "grad_norm": 4.26041841506958, "learning_rate": 7.152911127148893e-06, "loss": 0.2365, "step": 288 }, { "epoch": 0.29162462159434915, "grad_norm": 3.567761182785034, "learning_rate": 7.1442123789711495e-06, "loss": 0.1612, "step": 289 }, { "epoch": 0.29263370332996974, "grad_norm": 4.908886432647705, "learning_rate": 7.135474541685359e-06, "loss": 0.2345, "step": 290 }, { "epoch": 0.29364278506559033, "grad_norm": 5.828519821166992, "learning_rate": 7.126697723921041e-06, "loss": 0.3646, "step": 291 }, { "epoch": 0.2946518668012109, "grad_norm": 5.6283674240112305, "learning_rate": 7.117882034792315e-06, "loss": 0.3539, "step": 292 }, { "epoch": 0.2956609485368315, "grad_norm": 2.5755410194396973, "learning_rate": 7.109027583896559e-06, "loss": 0.072, "step": 293 }, { "epoch": 0.2966700302724521, "grad_norm": 5.203614234924316, "learning_rate": 7.1001344813130355e-06, "loss": 0.2994, "step": 294 }, { "epoch": 0.2976791120080727, "grad_norm": 4.4349822998046875, "learning_rate": 7.0912028376015315e-06, "loss": 0.1816, "step": 295 }, { "epoch": 0.29868819374369326, "grad_norm": 3.131836175918579, "learning_rate": 7.082232763800982e-06, "loss": 0.1107, "step": 296 }, { "epoch": 0.29969727547931385, "grad_norm": 7.76261568069458, "learning_rate": 7.073224371428083e-06, "loss": 0.3865, "step": 297 }, { "epoch": 0.3007063572149344, "grad_norm": 8.260970115661621, "learning_rate": 7.064177772475912e-06, "loss": 0.3415, "step": 298 }, { "epoch": 0.30171543895055497, "grad_norm": 7.274144172668457, "learning_rate": 7.055093079412536e-06, "loss": 0.406, "step": 299 }, { "epoch": 0.30272452068617556, "grad_norm": 6.949483394622803, "learning_rate": 7.04597040517961e-06, "loss": 0.2363, "step": 300 }, { "epoch": 0.30272452068617556, "eval_accuracy": 0.8422247446083996, "eval_loss": 0.29199618101119995, "eval_runtime": 62.2656, "eval_samples_per_second": 28.298, "eval_steps_per_second": 3.549, "step": 300 }, { "epoch": 0.30373360242179614, "grad_norm": 11.153295516967773, "learning_rate": 7.036809863190972e-06, "loss": 0.367, "step": 301 }, { "epoch": 0.30474268415741673, "grad_norm": 7.238802909851074, "learning_rate": 7.027611567331239e-06, "loss": 0.3071, "step": 302 }, { "epoch": 0.3057517658930373, "grad_norm": 5.204061031341553, "learning_rate": 7.018375631954384e-06, "loss": 0.1991, "step": 303 }, { "epoch": 0.3067608476286579, "grad_norm": 4.6721930503845215, "learning_rate": 7.0091021718823185e-06, "loss": 0.1833, "step": 304 }, { "epoch": 0.3077699293642785, "grad_norm": 4.843674182891846, "learning_rate": 6.999791302403463e-06, "loss": 0.2464, "step": 305 }, { "epoch": 0.3087790110998991, "grad_norm": 5.238533973693848, "learning_rate": 6.990443139271317e-06, "loss": 0.2551, "step": 306 }, { "epoch": 0.30978809283551967, "grad_norm": 20.923328399658203, "learning_rate": 6.981057798703019e-06, "loss": 0.7206, "step": 307 }, { "epoch": 0.31079717457114026, "grad_norm": 4.751791954040527, "learning_rate": 6.971635397377895e-06, "loss": 0.2031, "step": 308 }, { "epoch": 0.31180625630676084, "grad_norm": 9.50137996673584, "learning_rate": 6.962176052436019e-06, "loss": 0.4901, "step": 309 }, { "epoch": 0.31281533804238143, "grad_norm": 10.486673355102539, "learning_rate": 6.952679881476746e-06, "loss": 0.5644, "step": 310 }, { "epoch": 0.313824419778002, "grad_norm": 8.85893726348877, "learning_rate": 6.94314700255726e-06, "loss": 0.5932, "step": 311 }, { "epoch": 0.3148335015136226, "grad_norm": 33.33445358276367, "learning_rate": 6.933577534191101e-06, "loss": 1.3013, "step": 312 }, { "epoch": 0.3158425832492432, "grad_norm": 29.183631896972656, "learning_rate": 6.923971595346686e-06, "loss": 0.4947, "step": 313 }, { "epoch": 0.3168516649848638, "grad_norm": 7.223900318145752, "learning_rate": 6.914329305445844e-06, "loss": 0.3102, "step": 314 }, { "epoch": 0.31786074672048437, "grad_norm": 4.909074306488037, "learning_rate": 6.904650784362317e-06, "loss": 0.223, "step": 315 }, { "epoch": 0.31886982845610495, "grad_norm": 5.134761333465576, "learning_rate": 6.89493615242028e-06, "loss": 0.2572, "step": 316 }, { "epoch": 0.31987891019172554, "grad_norm": 4.783396244049072, "learning_rate": 6.885185530392841e-06, "loss": 0.2743, "step": 317 }, { "epoch": 0.32088799192734613, "grad_norm": 6.382033824920654, "learning_rate": 6.875399039500535e-06, "loss": 0.3093, "step": 318 }, { "epoch": 0.3218970736629667, "grad_norm": 7.4557318687438965, "learning_rate": 6.865576801409828e-06, "loss": 0.4611, "step": 319 }, { "epoch": 0.3229061553985873, "grad_norm": 9.1113862991333, "learning_rate": 6.855718938231597e-06, "loss": 0.5509, "step": 320 }, { "epoch": 0.3229061553985873, "eval_accuracy": 0.8422247446083996, "eval_loss": 0.2692955434322357, "eval_runtime": 62.6677, "eval_samples_per_second": 28.117, "eval_steps_per_second": 3.527, "step": 320 }, { "epoch": 0.3239152371342079, "grad_norm": 4.367307186126709, "learning_rate": 6.845825572519606e-06, "loss": 0.2284, "step": 321 }, { "epoch": 0.3249243188698285, "grad_norm": 4.980352878570557, "learning_rate": 6.8358968272689995e-06, "loss": 0.2197, "step": 322 }, { "epoch": 0.32593340060544906, "grad_norm": 3.6899354457855225, "learning_rate": 6.825932825914758e-06, "loss": 0.1665, "step": 323 }, { "epoch": 0.32694248234106965, "grad_norm": 5.287685394287109, "learning_rate": 6.815933692330168e-06, "loss": 0.2254, "step": 324 }, { "epoch": 0.32795156407669024, "grad_norm": 4.298030376434326, "learning_rate": 6.805899550825285e-06, "loss": 0.1869, "step": 325 }, { "epoch": 0.32896064581231077, "grad_norm": 6.648251533508301, "learning_rate": 6.795830526145385e-06, "loss": 0.438, "step": 326 }, { "epoch": 0.32996972754793136, "grad_norm": 8.279413223266602, "learning_rate": 6.785726743469415e-06, "loss": 0.3674, "step": 327 }, { "epoch": 0.33097880928355194, "grad_norm": 8.667623519897461, "learning_rate": 6.775588328408435e-06, "loss": 0.2876, "step": 328 }, { "epoch": 0.33198789101917253, "grad_norm": 4.858876705169678, "learning_rate": 6.765415407004061e-06, "loss": 0.2051, "step": 329 }, { "epoch": 0.3329969727547931, "grad_norm": 5.615779876708984, "learning_rate": 6.75520810572689e-06, "loss": 0.2162, "step": 330 }, { "epoch": 0.3340060544904137, "grad_norm": 5.255247116088867, "learning_rate": 6.744966551474935e-06, "loss": 0.2493, "step": 331 }, { "epoch": 0.3350151362260343, "grad_norm": 4.313990116119385, "learning_rate": 6.734690871572044e-06, "loss": 0.1736, "step": 332 }, { "epoch": 0.3360242179616549, "grad_norm": 6.205388069152832, "learning_rate": 6.72438119376632e-06, "loss": 0.1372, "step": 333 }, { "epoch": 0.33703329969727547, "grad_norm": 4.961912631988525, "learning_rate": 6.714037646228529e-06, "loss": 0.1083, "step": 334 }, { "epoch": 0.33804238143289606, "grad_norm": 11.366336822509766, "learning_rate": 6.703660357550507e-06, "loss": 0.1794, "step": 335 }, { "epoch": 0.33905146316851664, "grad_norm": 36.39216613769531, "learning_rate": 6.693249456743565e-06, "loss": 0.8015, "step": 336 }, { "epoch": 0.34006054490413723, "grad_norm": 26.968290328979492, "learning_rate": 6.682805073236883e-06, "loss": 0.5109, "step": 337 }, { "epoch": 0.3410696266397578, "grad_norm": 12.663142204284668, "learning_rate": 6.672327336875899e-06, "loss": 0.5224, "step": 338 }, { "epoch": 0.3420787083753784, "grad_norm": 7.804121017456055, "learning_rate": 6.661816377920695e-06, "loss": 0.2021, "step": 339 }, { "epoch": 0.343087790110999, "grad_norm": 8.033084869384766, "learning_rate": 6.651272327044385e-06, "loss": 0.2942, "step": 340 }, { "epoch": 0.343087790110999, "eval_accuracy": 0.840522133938706, "eval_loss": 0.2765989303588867, "eval_runtime": 62.3633, "eval_samples_per_second": 28.254, "eval_steps_per_second": 3.544, "step": 340 }, { "epoch": 0.3440968718466196, "grad_norm": 6.994205951690674, "learning_rate": 6.640695315331476e-06, "loss": 0.422, "step": 341 }, { "epoch": 0.34510595358224017, "grad_norm": 6.707622051239014, "learning_rate": 6.630085474276255e-06, "loss": 0.4127, "step": 342 }, { "epoch": 0.34611503531786075, "grad_norm": 2.8467178344726562, "learning_rate": 6.619442935781141e-06, "loss": 0.0992, "step": 343 }, { "epoch": 0.34712411705348134, "grad_norm": 3.68339467048645, "learning_rate": 6.608767832155051e-06, "loss": 0.1661, "step": 344 }, { "epoch": 0.3481331987891019, "grad_norm": 4.811255931854248, "learning_rate": 6.598060296111755e-06, "loss": 0.1805, "step": 345 }, { "epoch": 0.3491422805247225, "grad_norm": 7.682583332061768, "learning_rate": 6.58732046076823e-06, "loss": 0.4899, "step": 346 }, { "epoch": 0.3501513622603431, "grad_norm": 4.097416400909424, "learning_rate": 6.5765484596429905e-06, "loss": 0.1297, "step": 347 }, { "epoch": 0.3511604439959637, "grad_norm": 6.541982650756836, "learning_rate": 6.565744426654449e-06, "loss": 0.2545, "step": 348 }, { "epoch": 0.3521695257315843, "grad_norm": 6.773824214935303, "learning_rate": 6.554908496119232e-06, "loss": 0.2685, "step": 349 }, { "epoch": 0.35317860746720486, "grad_norm": 5.670194625854492, "learning_rate": 6.544040802750526e-06, "loss": 0.3245, "step": 350 }, { "epoch": 0.35418768920282545, "grad_norm": 5.502919673919678, "learning_rate": 6.5331414816563914e-06, "loss": 0.282, "step": 351 }, { "epoch": 0.35519677093844604, "grad_norm": 6.195550918579102, "learning_rate": 6.52221066833809e-06, "loss": 0.3791, "step": 352 }, { "epoch": 0.3562058526740666, "grad_norm": 6.497583866119385, "learning_rate": 6.511248498688395e-06, "loss": 0.3853, "step": 353 }, { "epoch": 0.35721493440968716, "grad_norm": 4.045083045959473, "learning_rate": 6.500255108989904e-06, "loss": 0.1618, "step": 354 }, { "epoch": 0.35822401614530774, "grad_norm": 4.328484058380127, "learning_rate": 6.489230635913346e-06, "loss": 0.2393, "step": 355 }, { "epoch": 0.35923309788092833, "grad_norm": 7.9082183837890625, "learning_rate": 6.478175216515884e-06, "loss": 0.2272, "step": 356 }, { "epoch": 0.3602421796165489, "grad_norm": 3.21086049079895, "learning_rate": 6.467088988239402e-06, "loss": 0.0689, "step": 357 }, { "epoch": 0.3612512613521695, "grad_norm": 5.211169242858887, "learning_rate": 6.455972088908807e-06, "loss": 0.1841, "step": 358 }, { "epoch": 0.3622603430877901, "grad_norm": 5.808556079864502, "learning_rate": 6.444824656730311e-06, "loss": 0.1076, "step": 359 }, { "epoch": 0.3632694248234107, "grad_norm": 16.821155548095703, "learning_rate": 6.43364683028971e-06, "loss": 0.4606, "step": 360 }, { "epoch": 0.3632694248234107, "eval_accuracy": 0.8484676503972758, "eval_loss": 0.3792904019355774, "eval_runtime": 62.8768, "eval_samples_per_second": 28.023, "eval_steps_per_second": 3.515, "step": 360 }, { "epoch": 0.36427850655903127, "grad_norm": 12.411088943481445, "learning_rate": 6.422438748550666e-06, "loss": 0.2406, "step": 361 }, { "epoch": 0.36528758829465185, "grad_norm": 13.036988258361816, "learning_rate": 6.411200550852978e-06, "loss": 0.3551, "step": 362 }, { "epoch": 0.36629667003027244, "grad_norm": 13.768930435180664, "learning_rate": 6.3999323769108485e-06, "loss": 0.3561, "step": 363 }, { "epoch": 0.36730575176589303, "grad_norm": 4.591919422149658, "learning_rate": 6.388634366811145e-06, "loss": 0.1434, "step": 364 }, { "epoch": 0.3683148335015136, "grad_norm": 9.278558731079102, "learning_rate": 6.377306661011664e-06, "loss": 0.445, "step": 365 }, { "epoch": 0.3693239152371342, "grad_norm": 12.307679176330566, "learning_rate": 6.365949400339378e-06, "loss": 0.3966, "step": 366 }, { "epoch": 0.3703329969727548, "grad_norm": 5.971735000610352, "learning_rate": 6.354562725988691e-06, "loss": 0.2508, "step": 367 }, { "epoch": 0.3713420787083754, "grad_norm": 5.919165134429932, "learning_rate": 6.343146779519681e-06, "loss": 0.287, "step": 368 }, { "epoch": 0.37235116044399597, "grad_norm": 4.112932205200195, "learning_rate": 6.331701702856335e-06, "loss": 0.1366, "step": 369 }, { "epoch": 0.37336024217961655, "grad_norm": 9.273783683776855, "learning_rate": 6.3202276382847925e-06, "loss": 0.4608, "step": 370 }, { "epoch": 0.37436932391523714, "grad_norm": 6.013513088226318, "learning_rate": 6.308724728451572e-06, "loss": 0.1759, "step": 371 }, { "epoch": 0.3753784056508577, "grad_norm": 4.636185169219971, "learning_rate": 6.2971931163618e-06, "loss": 0.1464, "step": 372 }, { "epoch": 0.3763874873864783, "grad_norm": 9.652426719665527, "learning_rate": 6.285632945377429e-06, "loss": 0.3786, "step": 373 }, { "epoch": 0.3773965691220989, "grad_norm": 4.286848545074463, "learning_rate": 6.274044359215461e-06, "loss": 0.1569, "step": 374 }, { "epoch": 0.3784056508577195, "grad_norm": 6.9270853996276855, "learning_rate": 6.2624275019461545e-06, "loss": 0.3606, "step": 375 }, { "epoch": 0.3794147325933401, "grad_norm": 4.69766092300415, "learning_rate": 6.250782517991241e-06, "loss": 0.1682, "step": 376 }, { "epoch": 0.38042381432896066, "grad_norm": 4.445621967315674, "learning_rate": 6.239109552122122e-06, "loss": 0.1959, "step": 377 }, { "epoch": 0.38143289606458125, "grad_norm": 8.952390670776367, "learning_rate": 6.227408749458073e-06, "loss": 0.3696, "step": 378 }, { "epoch": 0.38244197780020184, "grad_norm": 8.195367813110352, "learning_rate": 6.215680255464441e-06, "loss": 0.3198, "step": 379 }, { "epoch": 0.3834510595358224, "grad_norm": 14.417852401733398, "learning_rate": 6.203924215950831e-06, "loss": 0.6757, "step": 380 }, { "epoch": 0.3834510595358224, "eval_accuracy": 0.8513053348467651, "eval_loss": 0.29195961356163025, "eval_runtime": 62.6263, "eval_samples_per_second": 28.135, "eval_steps_per_second": 3.529, "step": 380 }, { "epoch": 0.384460141271443, "grad_norm": 10.577422142028809, "learning_rate": 6.192140777069298e-06, "loss": 0.3435, "step": 381 }, { "epoch": 0.3854692230070636, "grad_norm": 5.178398132324219, "learning_rate": 6.180330085312526e-06, "loss": 0.2577, "step": 382 }, { "epoch": 0.38647830474268413, "grad_norm": 5.634678363800049, "learning_rate": 6.168492287512014e-06, "loss": 0.2692, "step": 383 }, { "epoch": 0.3874873864783047, "grad_norm": 7.128296852111816, "learning_rate": 6.156627530836239e-06, "loss": 0.2777, "step": 384 }, { "epoch": 0.3884964682139253, "grad_norm": 6.315018653869629, "learning_rate": 6.144735962788837e-06, "loss": 0.3489, "step": 385 }, { "epoch": 0.3895055499495459, "grad_norm": 3.5851023197174072, "learning_rate": 6.132817731206765e-06, "loss": 0.1926, "step": 386 }, { "epoch": 0.3905146316851665, "grad_norm": 10.211600303649902, "learning_rate": 6.120872984258462e-06, "loss": 0.4683, "step": 387 }, { "epoch": 0.39152371342078707, "grad_norm": 12.310287475585938, "learning_rate": 6.108901870442009e-06, "loss": 0.3685, "step": 388 }, { "epoch": 0.39253279515640765, "grad_norm": 6.385642051696777, "learning_rate": 6.096904538583283e-06, "loss": 0.2559, "step": 389 }, { "epoch": 0.39354187689202824, "grad_norm": 7.468835830688477, "learning_rate": 6.084881137834103e-06, "loss": 0.4374, "step": 390 }, { "epoch": 0.39455095862764883, "grad_norm": 2.5986368656158447, "learning_rate": 6.072831817670382e-06, "loss": 0.0992, "step": 391 }, { "epoch": 0.3955600403632694, "grad_norm": 2.776925802230835, "learning_rate": 6.060756727890262e-06, "loss": 0.1379, "step": 392 }, { "epoch": 0.39656912209889, "grad_norm": 5.025798797607422, "learning_rate": 6.04865601861226e-06, "loss": 0.37, "step": 393 }, { "epoch": 0.3975782038345106, "grad_norm": 4.2647857666015625, "learning_rate": 6.036529840273388e-06, "loss": 0.2485, "step": 394 }, { "epoch": 0.3985872855701312, "grad_norm": 3.0298309326171875, "learning_rate": 6.0243783436273e-06, "loss": 0.1594, "step": 395 }, { "epoch": 0.39959636730575177, "grad_norm": 5.6464338302612305, "learning_rate": 6.012201679742408e-06, "loss": 0.3974, "step": 396 }, { "epoch": 0.40060544904137235, "grad_norm": 4.060618877410889, "learning_rate": 6e-06, "loss": 0.2286, "step": 397 }, { "epoch": 0.40161453077699294, "grad_norm": 4.622409820556641, "learning_rate": 5.987773456092368e-06, "loss": 0.315, "step": 398 }, { "epoch": 0.4026236125126135, "grad_norm": 5.471047401428223, "learning_rate": 5.9755222000209165e-06, "loss": 0.4354, "step": 399 }, { "epoch": 0.4036326942482341, "grad_norm": 3.929957628250122, "learning_rate": 5.963246384094273e-06, "loss": 0.2409, "step": 400 }, { "epoch": 0.4036326942482341, "eval_accuracy": 0.8376844494892168, "eval_loss": 0.28469741344451904, "eval_runtime": 62.4393, "eval_samples_per_second": 28.219, "eval_steps_per_second": 3.539, "step": 400 } ], "logging_steps": 1, "max_steps": 991, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 20, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 0.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }