diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,8330 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 4.0, + "eval_steps": 500, + "global_step": 1184, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0033783783783783786, + "grad_norm": 134.4763946533203, + "learning_rate": 0.0, + "loss": 4.6817, + "step": 1 + }, + { + "epoch": 0.006756756756756757, + "grad_norm": 125.12445068359375, + "learning_rate": 5.017166594399687e-06, + "loss": 4.5642, + "step": 2 + }, + { + "epoch": 0.010135135135135136, + "grad_norm": 133.296142578125, + "learning_rate": 7.952020911994375e-06, + "loss": 4.787, + "step": 3 + }, + { + "epoch": 0.013513513513513514, + "grad_norm": 72.2229995727539, + "learning_rate": 1.0034333188799373e-05, + "loss": 3.1422, + "step": 4 + }, + { + "epoch": 0.016891891891891893, + "grad_norm": 83.57042694091797, + "learning_rate": 1.164950007226698e-05, + "loss": 1.9827, + "step": 5 + }, + { + "epoch": 0.02027027027027027, + "grad_norm": 35.78062057495117, + "learning_rate": 1.2969187506394062e-05, + "loss": 1.0212, + "step": 6 + }, + { + "epoch": 0.02364864864864865, + "grad_norm": 48.31964111328125, + "learning_rate": 1.4084967333570947e-05, + "loss": 1.1823, + "step": 7 + }, + { + "epoch": 0.02702702702702703, + "grad_norm": 10.594072341918945, + "learning_rate": 1.505149978319906e-05, + "loss": 0.7368, + "step": 8 + }, + { + "epoch": 0.030405405405405407, + "grad_norm": 7.897088050842285, + "learning_rate": 1.590404182398875e-05, + "loss": 0.6276, + "step": 9 + }, + { + "epoch": 0.033783783783783786, + "grad_norm": 3.8260443210601807, + "learning_rate": 1.666666666666667e-05, + "loss": 0.5934, + "step": 10 + }, + { + "epoch": 0.037162162162162164, + "grad_norm": 11.113093376159668, + "learning_rate": 1.7356544752637084e-05, + "loss": 0.6987, + "step": 11 + }, + { + "epoch": 0.04054054054054054, + "grad_norm": 3.2817893028259277, + "learning_rate": 1.7986354100793748e-05, + "loss": 0.5722, + "step": 12 + }, + { + "epoch": 0.04391891891891892, + "grad_norm": 1.4319814443588257, + "learning_rate": 1.8565722538447282e-05, + "loss": 0.5199, + "step": 13 + }, + { + "epoch": 0.0472972972972973, + "grad_norm": 1.1898471117019653, + "learning_rate": 1.9102133927970633e-05, + "loss": 0.5063, + "step": 14 + }, + { + "epoch": 0.05067567567567568, + "grad_norm": 2.1912143230438232, + "learning_rate": 1.9601520984261358e-05, + "loss": 0.5368, + "step": 15 + }, + { + "epoch": 0.05405405405405406, + "grad_norm": 1.0015403032302856, + "learning_rate": 2.0068666377598747e-05, + "loss": 0.4682, + "step": 16 + }, + { + "epoch": 0.057432432432432436, + "grad_norm": 1.114931583404541, + "learning_rate": 2.0507482022971233e-05, + "loss": 0.4997, + "step": 17 + }, + { + "epoch": 0.060810810810810814, + "grad_norm": 1.0931981801986694, + "learning_rate": 2.0921208418388435e-05, + "loss": 0.4659, + "step": 18 + }, + { + "epoch": 0.06418918918918919, + "grad_norm": 0.8113773465156555, + "learning_rate": 2.1312560015880482e-05, + "loss": 0.4697, + "step": 19 + }, + { + "epoch": 0.06756756756756757, + "grad_norm": 1.0267772674560547, + "learning_rate": 2.1683833261066357e-05, + "loss": 0.4851, + "step": 20 + }, + { + "epoch": 0.07094594594594594, + "grad_norm": 0.8238614201545715, + "learning_rate": 2.2036988245565324e-05, + "loss": 0.4517, + "step": 21 + }, + { + "epoch": 0.07432432432432433, + "grad_norm": 0.9366945624351501, + "learning_rate": 2.2373711347036773e-05, + "loss": 0.4982, + "step": 22 + }, + { + "epoch": 0.0777027027027027, + "grad_norm": 0.8443475365638733, + "learning_rate": 2.269546393362655e-05, + "loss": 0.4128, + "step": 23 + }, + { + "epoch": 0.08108108108108109, + "grad_norm": 0.8733354806900024, + "learning_rate": 2.3003520695193437e-05, + "loss": 0.4281, + "step": 24 + }, + { + "epoch": 0.08445945945945946, + "grad_norm": 0.7766995429992676, + "learning_rate": 2.329900014453396e-05, + "loss": 0.4407, + "step": 25 + }, + { + "epoch": 0.08783783783783784, + "grad_norm": 0.8224064111709595, + "learning_rate": 2.3582889132846968e-05, + "loss": 0.4155, + "step": 26 + }, + { + "epoch": 0.09121621621621621, + "grad_norm": 0.7931010127067566, + "learning_rate": 2.3856062735983123e-05, + "loss": 0.4191, + "step": 27 + }, + { + "epoch": 0.0945945945945946, + "grad_norm": 0.6803894639015198, + "learning_rate": 2.4119300522370322e-05, + "loss": 0.4055, + "step": 28 + }, + { + "epoch": 0.09797297297297297, + "grad_norm": 0.748294472694397, + "learning_rate": 2.4373299964982603e-05, + "loss": 0.4443, + "step": 29 + }, + { + "epoch": 0.10135135135135136, + "grad_norm": 0.706881582736969, + "learning_rate": 2.4618687578661044e-05, + "loss": 0.4093, + "step": 30 + }, + { + "epoch": 0.10472972972972973, + "grad_norm": 0.7517805695533752, + "learning_rate": 2.4856028230571212e-05, + "loss": 0.4288, + "step": 31 + }, + { + "epoch": 0.10810810810810811, + "grad_norm": 0.6549198031425476, + "learning_rate": 2.5085832971998436e-05, + "loss": 0.4222, + "step": 32 + }, + { + "epoch": 0.11148648648648649, + "grad_norm": 0.9364942908287048, + "learning_rate": 2.530856566463146e-05, + "loss": 0.416, + "step": 33 + }, + { + "epoch": 0.11486486486486487, + "grad_norm": 0.624920129776001, + "learning_rate": 2.552464861737092e-05, + "loss": 0.3735, + "step": 34 + }, + { + "epoch": 0.11824324324324324, + "grad_norm": 0.6083908081054688, + "learning_rate": 2.5734467405837933e-05, + "loss": 0.3315, + "step": 35 + }, + { + "epoch": 0.12162162162162163, + "grad_norm": 0.6106983423233032, + "learning_rate": 2.5938375012788124e-05, + "loss": 0.3652, + "step": 36 + }, + { + "epoch": 0.125, + "grad_norm": 0.7572540044784546, + "learning_rate": 2.6136695401116585e-05, + "loss": 0.452, + "step": 37 + }, + { + "epoch": 0.12837837837837837, + "grad_norm": 0.6654285788536072, + "learning_rate": 2.6329726610280168e-05, + "loss": 0.3328, + "step": 38 + }, + { + "epoch": 0.13175675675675674, + "grad_norm": 0.676705002784729, + "learning_rate": 2.651774345044166e-05, + "loss": 0.3987, + "step": 39 + }, + { + "epoch": 0.13513513513513514, + "grad_norm": 0.6630686521530151, + "learning_rate": 2.6700999855466042e-05, + "loss": 0.3794, + "step": 40 + }, + { + "epoch": 0.13851351351351351, + "grad_norm": 0.8196331858634949, + "learning_rate": 2.687973094532893e-05, + "loss": 0.4218, + "step": 41 + }, + { + "epoch": 0.14189189189189189, + "grad_norm": 0.6006856560707092, + "learning_rate": 2.7054154839965013e-05, + "loss": 0.3788, + "step": 42 + }, + { + "epoch": 0.14527027027027026, + "grad_norm": 0.6180748343467712, + "learning_rate": 2.722447425965978e-05, + "loss": 0.3946, + "step": 43 + }, + { + "epoch": 0.14864864864864866, + "grad_norm": 0.7010207772254944, + "learning_rate": 2.739087794143646e-05, + "loss": 0.389, + "step": 44 + }, + { + "epoch": 0.15202702702702703, + "grad_norm": 0.681907594203949, + "learning_rate": 2.755354189625573e-05, + "loss": 0.399, + "step": 45 + }, + { + "epoch": 0.1554054054054054, + "grad_norm": 0.6506620645523071, + "learning_rate": 2.771263052802624e-05, + "loss": 0.3924, + "step": 46 + }, + { + "epoch": 0.15878378378378377, + "grad_norm": 0.6619815230369568, + "learning_rate": 2.7868297632261957e-05, + "loss": 0.4111, + "step": 47 + }, + { + "epoch": 0.16216216216216217, + "grad_norm": 0.6953612565994263, + "learning_rate": 2.8020687289593123e-05, + "loss": 0.3958, + "step": 48 + }, + { + "epoch": 0.16554054054054054, + "grad_norm": 0.761862576007843, + "learning_rate": 2.8169934667141895e-05, + "loss": 0.3844, + "step": 49 + }, + { + "epoch": 0.16891891891891891, + "grad_norm": 0.7193901538848877, + "learning_rate": 2.8316166738933646e-05, + "loss": 0.3847, + "step": 50 + }, + { + "epoch": 0.17229729729729729, + "grad_norm": 0.6792317032814026, + "learning_rate": 2.845950293496561e-05, + "loss": 0.4043, + "step": 51 + }, + { + "epoch": 0.17567567567567569, + "grad_norm": 0.6307753920555115, + "learning_rate": 2.8600055727246657e-05, + "loss": 0.3806, + "step": 52 + }, + { + "epoch": 0.17905405405405406, + "grad_norm": 0.7366315722465515, + "learning_rate": 2.8737931160013153e-05, + "loss": 0.416, + "step": 53 + }, + { + "epoch": 0.18243243243243243, + "grad_norm": 0.603863537311554, + "learning_rate": 2.8873229330382812e-05, + "loss": 0.3548, + "step": 54 + }, + { + "epoch": 0.1858108108108108, + "grad_norm": 0.7424588799476624, + "learning_rate": 2.9006044824904066e-05, + "loss": 0.3982, + "step": 55 + }, + { + "epoch": 0.1891891891891892, + "grad_norm": 0.5944585204124451, + "learning_rate": 2.913646711677001e-05, + "loss": 0.3536, + "step": 56 + }, + { + "epoch": 0.19256756756756757, + "grad_norm": 0.6465046405792236, + "learning_rate": 2.926458092787486e-05, + "loss": 0.3594, + "step": 57 + }, + { + "epoch": 0.19594594594594594, + "grad_norm": 0.7178723812103271, + "learning_rate": 2.939046655938229e-05, + "loss": 0.3725, + "step": 58 + }, + { + "epoch": 0.19932432432432431, + "grad_norm": 0.7240431308746338, + "learning_rate": 2.951420019403574e-05, + "loss": 0.3896, + "step": 59 + }, + { + "epoch": 0.20270270270270271, + "grad_norm": 0.7555009126663208, + "learning_rate": 2.963585417306073e-05, + "loss": 0.3633, + "step": 60 + }, + { + "epoch": 0.20608108108108109, + "grad_norm": 0.6279475092887878, + "learning_rate": 2.9755497250179453e-05, + "loss": 0.3447, + "step": 61 + }, + { + "epoch": 0.20945945945945946, + "grad_norm": 0.6943121552467346, + "learning_rate": 2.98731948249709e-05, + "loss": 0.3851, + "step": 62 + }, + { + "epoch": 0.21283783783783783, + "grad_norm": 0.6565203070640564, + "learning_rate": 2.9989009157559694e-05, + "loss": 0.3896, + "step": 63 + }, + { + "epoch": 0.21621621621621623, + "grad_norm": 4.912997722625732, + "learning_rate": 3.010299956639812e-05, + "loss": 0.4528, + "step": 64 + }, + { + "epoch": 0.2195945945945946, + "grad_norm": 0.7776851058006287, + "learning_rate": 3.021522261071426e-05, + "loss": 0.3151, + "step": 65 + }, + { + "epoch": 0.22297297297297297, + "grad_norm": 0.5963460206985474, + "learning_rate": 3.0325732259031143e-05, + "loss": 0.3501, + "step": 66 + }, + { + "epoch": 0.22635135135135134, + "grad_norm": 0.8160498142242432, + "learning_rate": 3.043458004501377e-05, + "loss": 0.3383, + "step": 67 + }, + { + "epoch": 0.22972972972972974, + "grad_norm": 0.6507856249809265, + "learning_rate": 3.054181521177061e-05, + "loss": 0.3559, + "step": 68 + }, + { + "epoch": 0.23310810810810811, + "grad_norm": 0.6282461881637573, + "learning_rate": 3.064748484562093e-05, + "loss": 0.3951, + "step": 69 + }, + { + "epoch": 0.23648648648648649, + "grad_norm": 0.6464638710021973, + "learning_rate": 3.0751634000237615e-05, + "loss": 0.3759, + "step": 70 + }, + { + "epoch": 0.23986486486486486, + "grad_norm": 0.621934175491333, + "learning_rate": 3.085430581198459e-05, + "loss": 0.409, + "step": 71 + }, + { + "epoch": 0.24324324324324326, + "grad_norm": 0.6526879668235779, + "learning_rate": 3.095554160718781e-05, + "loss": 0.3938, + "step": 72 + }, + { + "epoch": 0.24662162162162163, + "grad_norm": 0.6536738276481628, + "learning_rate": 3.10553810020076e-05, + "loss": 0.3657, + "step": 73 + }, + { + "epoch": 0.25, + "grad_norm": 0.6381665468215942, + "learning_rate": 3.115386199551628e-05, + "loss": 0.3743, + "step": 74 + }, + { + "epoch": 0.2533783783783784, + "grad_norm": 0.6818379759788513, + "learning_rate": 3.1251021056528336e-05, + "loss": 0.3188, + "step": 75 + }, + { + "epoch": 0.25675675675675674, + "grad_norm": 0.5829269886016846, + "learning_rate": 3.134689320467986e-05, + "loss": 0.3622, + "step": 76 + }, + { + "epoch": 0.26013513513513514, + "grad_norm": 0.6988232731819153, + "learning_rate": 3.144151208620804e-05, + "loss": 0.3768, + "step": 77 + }, + { + "epoch": 0.2635135135135135, + "grad_norm": 0.5981537699699402, + "learning_rate": 3.1534910044841344e-05, + "loss": 0.3655, + "step": 78 + }, + { + "epoch": 0.2668918918918919, + "grad_norm": 0.5821400284767151, + "learning_rate": 3.1627118188174024e-05, + "loss": 0.358, + "step": 79 + }, + { + "epoch": 0.2702702702702703, + "grad_norm": 0.620126485824585, + "learning_rate": 3.171816644986573e-05, + "loss": 0.3579, + "step": 80 + }, + { + "epoch": 0.27364864864864863, + "grad_norm": 0.6327139139175415, + "learning_rate": 3.18080836479775e-05, + "loss": 0.3916, + "step": 81 + }, + { + "epoch": 0.27702702702702703, + "grad_norm": 0.5900242328643799, + "learning_rate": 3.1896897539728616e-05, + "loss": 0.3768, + "step": 82 + }, + { + "epoch": 0.28040540540540543, + "grad_norm": 0.5920627117156982, + "learning_rate": 3.198463487293457e-05, + "loss": 0.3709, + "step": 83 + }, + { + "epoch": 0.28378378378378377, + "grad_norm": 0.5604241490364075, + "learning_rate": 3.207132143436469e-05, + "loss": 0.3871, + "step": 84 + }, + { + "epoch": 0.28716216216216217, + "grad_norm": 0.6093663573265076, + "learning_rate": 3.215698209523821e-05, + "loss": 0.3632, + "step": 85 + }, + { + "epoch": 0.2905405405405405, + "grad_norm": 0.568435549736023, + "learning_rate": 3.224164085405946e-05, + "loss": 0.3788, + "step": 86 + }, + { + "epoch": 0.2939189189189189, + "grad_norm": 0.5885617136955261, + "learning_rate": 3.232532087697698e-05, + "loss": 0.3472, + "step": 87 + }, + { + "epoch": 0.2972972972972973, + "grad_norm": 0.5474864840507507, + "learning_rate": 3.240804453583615e-05, + "loss": 0.3815, + "step": 88 + }, + { + "epoch": 0.30067567567567566, + "grad_norm": 0.6116411089897156, + "learning_rate": 3.248983344408188e-05, + "loss": 0.3374, + "step": 89 + }, + { + "epoch": 0.30405405405405406, + "grad_norm": 0.5523970127105713, + "learning_rate": 3.2570708490655414e-05, + "loss": 0.3503, + "step": 90 + }, + { + "epoch": 0.30743243243243246, + "grad_norm": 0.6035043597221375, + "learning_rate": 3.265068987201822e-05, + "loss": 0.3699, + "step": 91 + }, + { + "epoch": 0.3108108108108108, + "grad_norm": 0.4965035617351532, + "learning_rate": 3.2729797122425925e-05, + "loss": 0.3702, + "step": 92 + }, + { + "epoch": 0.3141891891891892, + "grad_norm": 0.5582573413848877, + "learning_rate": 3.280804914256559e-05, + "loss": 0.333, + "step": 93 + }, + { + "epoch": 0.31756756756756754, + "grad_norm": 0.5313113331794739, + "learning_rate": 3.288546422666164e-05, + "loss": 0.329, + "step": 94 + }, + { + "epoch": 0.32094594594594594, + "grad_norm": 0.6049978733062744, + "learning_rate": 3.2962060088147464e-05, + "loss": 0.3931, + "step": 95 + }, + { + "epoch": 0.32432432432432434, + "grad_norm": 0.5387487411499023, + "learning_rate": 3.3037853883992805e-05, + "loss": 0.3432, + "step": 96 + }, + { + "epoch": 0.3277027027027027, + "grad_norm": 0.588671088218689, + "learning_rate": 3.3112862237770756e-05, + "loss": 0.3508, + "step": 97 + }, + { + "epoch": 0.3310810810810811, + "grad_norm": 0.536973774433136, + "learning_rate": 3.3187101261541584e-05, + "loss": 0.372, + "step": 98 + }, + { + "epoch": 0.3344594594594595, + "grad_norm": 0.676276445388794, + "learning_rate": 3.326058657662584e-05, + "loss": 0.3832, + "step": 99 + }, + { + "epoch": 0.33783783783783783, + "grad_norm": 0.46296200156211853, + "learning_rate": 3.333333333333334e-05, + "loss": 0.3553, + "step": 100 + }, + { + "epoch": 0.34121621621621623, + "grad_norm": 0.674472451210022, + "learning_rate": 3.340535622971072e-05, + "loss": 0.3448, + "step": 101 + }, + { + "epoch": 0.34459459459459457, + "grad_norm": 0.5100315809249878, + "learning_rate": 3.3476669529365295e-05, + "loss": 0.3367, + "step": 102 + }, + { + "epoch": 0.34797297297297297, + "grad_norm": 0.5637315511703491, + "learning_rate": 3.3547287078419544e-05, + "loss": 0.3786, + "step": 103 + }, + { + "epoch": 0.35135135135135137, + "grad_norm": 0.5659502744674683, + "learning_rate": 3.361722232164634e-05, + "loss": 0.352, + "step": 104 + }, + { + "epoch": 0.3547297297297297, + "grad_norm": 0.5390239357948303, + "learning_rate": 3.3686488317832306e-05, + "loss": 0.3473, + "step": 105 + }, + { + "epoch": 0.3581081081081081, + "grad_norm": 0.5552096366882324, + "learning_rate": 3.375509775441284e-05, + "loss": 0.2945, + "step": 106 + }, + { + "epoch": 0.3614864864864865, + "grad_norm": 0.5016259551048279, + "learning_rate": 3.382306296142016e-05, + "loss": 0.313, + "step": 107 + }, + { + "epoch": 0.36486486486486486, + "grad_norm": 0.5755091905593872, + "learning_rate": 3.38903959247825e-05, + "loss": 0.3367, + "step": 108 + }, + { + "epoch": 0.36824324324324326, + "grad_norm": 0.5590441823005676, + "learning_rate": 3.395710829901039e-05, + "loss": 0.3639, + "step": 109 + }, + { + "epoch": 0.3716216216216216, + "grad_norm": 0.5077652335166931, + "learning_rate": 3.402321141930376e-05, + "loss": 0.3523, + "step": 110 + }, + { + "epoch": 0.375, + "grad_norm": 0.5614023804664612, + "learning_rate": 3.4088716313110955e-05, + "loss": 0.3418, + "step": 111 + }, + { + "epoch": 0.3783783783783784, + "grad_norm": 0.5325702428817749, + "learning_rate": 3.415363371116969e-05, + "loss": 0.3208, + "step": 112 + }, + { + "epoch": 0.38175675675675674, + "grad_norm": 0.5361766219139099, + "learning_rate": 3.4217974058057e-05, + "loss": 0.3412, + "step": 113 + }, + { + "epoch": 0.38513513513513514, + "grad_norm": 0.5173611044883728, + "learning_rate": 3.428174752227455e-05, + "loss": 0.3615, + "step": 114 + }, + { + "epoch": 0.3885135135135135, + "grad_norm": 0.5852875709533691, + "learning_rate": 3.434496400589353e-05, + "loss": 0.3171, + "step": 115 + }, + { + "epoch": 0.3918918918918919, + "grad_norm": 0.48835834860801697, + "learning_rate": 3.440763315378198e-05, + "loss": 0.315, + "step": 116 + }, + { + "epoch": 0.3952702702702703, + "grad_norm": 0.5958307385444641, + "learning_rate": 3.446976436243603e-05, + "loss": 0.3117, + "step": 117 + }, + { + "epoch": 0.39864864864864863, + "grad_norm": 0.5645899772644043, + "learning_rate": 3.4531366788435425e-05, + "loss": 0.3192, + "step": 118 + }, + { + "epoch": 0.40202702702702703, + "grad_norm": 0.5764517784118652, + "learning_rate": 3.459244935654219e-05, + "loss": 0.3353, + "step": 119 + }, + { + "epoch": 0.40540540540540543, + "grad_norm": 0.5075375437736511, + "learning_rate": 3.465302076746041e-05, + "loss": 0.3572, + "step": 120 + }, + { + "epoch": 0.40878378378378377, + "grad_norm": 0.48014962673187256, + "learning_rate": 3.471308950527417e-05, + "loss": 0.3217, + "step": 121 + }, + { + "epoch": 0.41216216216216217, + "grad_norm": 0.533892035484314, + "learning_rate": 3.477266384457914e-05, + "loss": 0.3214, + "step": 122 + }, + { + "epoch": 0.4155405405405405, + "grad_norm": 0.4845181405544281, + "learning_rate": 3.48317518573233e-05, + "loss": 0.3332, + "step": 123 + }, + { + "epoch": 0.4189189189189189, + "grad_norm": 0.5133293271064758, + "learning_rate": 3.489036141937059e-05, + "loss": 0.354, + "step": 124 + }, + { + "epoch": 0.4222972972972973, + "grad_norm": 0.5352755188941956, + "learning_rate": 3.494850021680094e-05, + "loss": 0.353, + "step": 125 + }, + { + "epoch": 0.42567567567567566, + "grad_norm": 0.5208103060722351, + "learning_rate": 3.500617575195938e-05, + "loss": 0.3065, + "step": 126 + }, + { + "epoch": 0.42905405405405406, + "grad_norm": 0.5982815623283386, + "learning_rate": 3.5063395349265945e-05, + "loss": 0.3329, + "step": 127 + }, + { + "epoch": 0.43243243243243246, + "grad_norm": 0.5241547226905823, + "learning_rate": 3.5120166160797804e-05, + "loss": 0.3476, + "step": 128 + }, + { + "epoch": 0.4358108108108108, + "grad_norm": 0.5543828010559082, + "learning_rate": 3.517649517165415e-05, + "loss": 0.3517, + "step": 129 + }, + { + "epoch": 0.4391891891891892, + "grad_norm": 0.5167660117149353, + "learning_rate": 3.523238920511395e-05, + "loss": 0.3417, + "step": 130 + }, + { + "epoch": 0.44256756756756754, + "grad_norm": 0.5201629400253296, + "learning_rate": 3.528785492759607e-05, + "loss": 0.3543, + "step": 131 + }, + { + "epoch": 0.44594594594594594, + "grad_norm": 0.5214123129844666, + "learning_rate": 3.5342898853430836e-05, + "loss": 0.3552, + "step": 132 + }, + { + "epoch": 0.44932432432432434, + "grad_norm": 0.526899516582489, + "learning_rate": 3.539752734945143e-05, + "loss": 0.3469, + "step": 133 + }, + { + "epoch": 0.4527027027027027, + "grad_norm": 0.55535888671875, + "learning_rate": 3.5451746639413466e-05, + "loss": 0.294, + "step": 134 + }, + { + "epoch": 0.4560810810810811, + "grad_norm": 0.46593329310417175, + "learning_rate": 3.550556280825011e-05, + "loss": 0.2557, + "step": 135 + }, + { + "epoch": 0.4594594594594595, + "grad_norm": 0.5066341757774353, + "learning_rate": 3.55589818061703e-05, + "loss": 0.2885, + "step": 136 + }, + { + "epoch": 0.46283783783783783, + "grad_norm": 0.4943198263645172, + "learning_rate": 3.561200945260678e-05, + "loss": 0.3436, + "step": 137 + }, + { + "epoch": 0.46621621621621623, + "grad_norm": 0.5013923048973083, + "learning_rate": 3.5664651440020616e-05, + "loss": 0.2865, + "step": 138 + }, + { + "epoch": 0.46959459459459457, + "grad_norm": 0.5165452361106873, + "learning_rate": 3.571691333756825e-05, + "loss": 0.3394, + "step": 139 + }, + { + "epoch": 0.47297297297297297, + "grad_norm": 0.5002908706665039, + "learning_rate": 3.5768800594637304e-05, + "loss": 0.3278, + "step": 140 + }, + { + "epoch": 0.47635135135135137, + "grad_norm": 0.5236304402351379, + "learning_rate": 3.582031854425634e-05, + "loss": 0.3343, + "step": 141 + }, + { + "epoch": 0.4797297297297297, + "grad_norm": 0.49403268098831177, + "learning_rate": 3.587147240638428e-05, + "loss": 0.3079, + "step": 142 + }, + { + "epoch": 0.4831081081081081, + "grad_norm": 0.4847567677497864, + "learning_rate": 3.5922267291084366e-05, + "loss": 0.3305, + "step": 143 + }, + { + "epoch": 0.4864864864864865, + "grad_norm": 0.5461480617523193, + "learning_rate": 3.5972708201587496e-05, + "loss": 0.3242, + "step": 144 + }, + { + "epoch": 0.48986486486486486, + "grad_norm": 0.5380986332893372, + "learning_rate": 3.6022800037249585e-05, + "loss": 0.3341, + "step": 145 + }, + { + "epoch": 0.49324324324324326, + "grad_norm": 0.49580562114715576, + "learning_rate": 3.607254759640729e-05, + "loss": 0.31, + "step": 146 + }, + { + "epoch": 0.4966216216216216, + "grad_norm": 0.5374506115913391, + "learning_rate": 3.612195557913627e-05, + "loss": 0.3048, + "step": 147 + }, + { + "epoch": 0.5, + "grad_norm": 0.535591721534729, + "learning_rate": 3.6171028589915954e-05, + "loss": 0.3266, + "step": 148 + }, + { + "epoch": 0.5033783783783784, + "grad_norm": 0.5777239799499512, + "learning_rate": 3.6219771140204575e-05, + "loss": 0.3509, + "step": 149 + }, + { + "epoch": 0.5067567567567568, + "grad_norm": 0.5570788383483887, + "learning_rate": 3.626818765092802e-05, + "loss": 0.3348, + "step": 150 + }, + { + "epoch": 0.5101351351351351, + "grad_norm": 0.5780688524246216, + "learning_rate": 3.6316282454886157e-05, + "loss": 0.3218, + "step": 151 + }, + { + "epoch": 0.5135135135135135, + "grad_norm": 0.5464823842048645, + "learning_rate": 3.636405979907955e-05, + "loss": 0.347, + "step": 152 + }, + { + "epoch": 0.5168918918918919, + "grad_norm": 0.46229949593544006, + "learning_rate": 3.6411523846959985e-05, + "loss": 0.2618, + "step": 153 + }, + { + "epoch": 0.5202702702702703, + "grad_norm": 0.5127717852592468, + "learning_rate": 3.645867868060772e-05, + "loss": 0.3364, + "step": 154 + }, + { + "epoch": 0.5236486486486487, + "grad_norm": 0.4915751814842224, + "learning_rate": 3.6505528302838193e-05, + "loss": 0.327, + "step": 155 + }, + { + "epoch": 0.527027027027027, + "grad_norm": 0.4658984839916229, + "learning_rate": 3.6552076639241027e-05, + "loss": 0.3008, + "step": 156 + }, + { + "epoch": 0.5304054054054054, + "grad_norm": 0.4983694553375244, + "learning_rate": 3.65983275401539e-05, + "loss": 0.3153, + "step": 157 + }, + { + "epoch": 0.5337837837837838, + "grad_norm": 0.49560025334358215, + "learning_rate": 3.664428478257371e-05, + "loss": 0.3409, + "step": 158 + }, + { + "epoch": 0.5371621621621622, + "grad_norm": 0.5617703199386597, + "learning_rate": 3.668995207200753e-05, + "loss": 0.3295, + "step": 159 + }, + { + "epoch": 0.5405405405405406, + "grad_norm": 0.5226999521255493, + "learning_rate": 3.673533304426541e-05, + "loss": 0.3522, + "step": 160 + }, + { + "epoch": 0.543918918918919, + "grad_norm": 0.49685564637184143, + "learning_rate": 3.67804312671975e-05, + "loss": 0.3322, + "step": 161 + }, + { + "epoch": 0.5472972972972973, + "grad_norm": 0.543743371963501, + "learning_rate": 3.682525024237719e-05, + "loss": 0.3283, + "step": 162 + }, + { + "epoch": 0.5506756756756757, + "grad_norm": 0.5247477889060974, + "learning_rate": 3.6869793406732636e-05, + "loss": 0.3104, + "step": 163 + }, + { + "epoch": 0.5540540540540541, + "grad_norm": 0.5228151679039001, + "learning_rate": 3.69140641341283e-05, + "loss": 0.3292, + "step": 164 + }, + { + "epoch": 0.5574324324324325, + "grad_norm": 0.48610207438468933, + "learning_rate": 3.695806573689844e-05, + "loss": 0.3183, + "step": 165 + }, + { + "epoch": 0.5608108108108109, + "grad_norm": 0.4703191816806793, + "learning_rate": 3.700180146733426e-05, + "loss": 0.2935, + "step": 166 + }, + { + "epoch": 0.5641891891891891, + "grad_norm": 0.5501482486724854, + "learning_rate": 3.704527451912639e-05, + "loss": 0.3249, + "step": 167 + }, + { + "epoch": 0.5675675675675675, + "grad_norm": 0.45025497674942017, + "learning_rate": 3.708848802876438e-05, + "loss": 0.3128, + "step": 168 + }, + { + "epoch": 0.5709459459459459, + "grad_norm": 0.5103681087493896, + "learning_rate": 3.7131445076894564e-05, + "loss": 0.2882, + "step": 169 + }, + { + "epoch": 0.5743243243243243, + "grad_norm": 0.4952690601348877, + "learning_rate": 3.717414868963791e-05, + "loss": 0.3268, + "step": 170 + }, + { + "epoch": 0.5777027027027027, + "grad_norm": 0.4887889325618744, + "learning_rate": 3.721660183986924e-05, + "loss": 0.3161, + "step": 171 + }, + { + "epoch": 0.581081081081081, + "grad_norm": 0.5142275094985962, + "learning_rate": 3.725880744845915e-05, + "loss": 0.2942, + "step": 172 + }, + { + "epoch": 0.5844594594594594, + "grad_norm": 0.518671452999115, + "learning_rate": 3.730076838547993e-05, + "loss": 0.3232, + "step": 173 + }, + { + "epoch": 0.5878378378378378, + "grad_norm": 0.5324185490608215, + "learning_rate": 3.734248747137666e-05, + "loss": 0.3115, + "step": 174 + }, + { + "epoch": 0.5912162162162162, + "grad_norm": 0.5307193398475647, + "learning_rate": 3.738396747810492e-05, + "loss": 0.3245, + "step": 175 + }, + { + "epoch": 0.5945945945945946, + "grad_norm": 0.47318169474601746, + "learning_rate": 3.7425211130235834e-05, + "loss": 0.3152, + "step": 176 + }, + { + "epoch": 0.597972972972973, + "grad_norm": 0.5233657956123352, + "learning_rate": 3.7466221106030115e-05, + "loss": 0.3115, + "step": 177 + }, + { + "epoch": 0.6013513513513513, + "grad_norm": 0.4699764847755432, + "learning_rate": 3.750700003848157e-05, + "loss": 0.2948, + "step": 178 + }, + { + "epoch": 0.6047297297297297, + "grad_norm": 0.5511758327484131, + "learning_rate": 3.7547550516331555e-05, + "loss": 0.337, + "step": 179 + }, + { + "epoch": 0.6081081081081081, + "grad_norm": 0.5334018468856812, + "learning_rate": 3.75878750850551e-05, + "loss": 0.3244, + "step": 180 + }, + { + "epoch": 0.6114864864864865, + "grad_norm": 0.5040500164031982, + "learning_rate": 3.7627976247819744e-05, + "loss": 0.3173, + "step": 181 + }, + { + "epoch": 0.6148648648648649, + "grad_norm": 0.5094459652900696, + "learning_rate": 3.766785646641792e-05, + "loss": 0.3087, + "step": 182 + }, + { + "epoch": 0.6182432432432432, + "grad_norm": 0.4679316282272339, + "learning_rate": 3.770751816217383e-05, + "loss": 0.3261, + "step": 183 + }, + { + "epoch": 0.6216216216216216, + "grad_norm": 0.5099210143089294, + "learning_rate": 3.7746963716825615e-05, + "loss": 0.3239, + "step": 184 + }, + { + "epoch": 0.625, + "grad_norm": 0.4242151379585266, + "learning_rate": 3.778619547338356e-05, + "loss": 0.3082, + "step": 185 + }, + { + "epoch": 0.6283783783783784, + "grad_norm": 0.4796642065048218, + "learning_rate": 3.782521573696528e-05, + "loss": 0.2753, + "step": 186 + }, + { + "epoch": 0.6317567567567568, + "grad_norm": 0.43798592686653137, + "learning_rate": 3.786402677560832e-05, + "loss": 0.3033, + "step": 187 + }, + { + "epoch": 0.6351351351351351, + "grad_norm": 0.4947024881839752, + "learning_rate": 3.790263082106134e-05, + "loss": 0.3171, + "step": 188 + }, + { + "epoch": 0.6385135135135135, + "grad_norm": 0.40923503041267395, + "learning_rate": 3.794103006955407e-05, + "loss": 0.2927, + "step": 189 + }, + { + "epoch": 0.6418918918918919, + "grad_norm": 0.6707382202148438, + "learning_rate": 3.797922668254715e-05, + "loss": 0.2885, + "step": 190 + }, + { + "epoch": 0.6452702702702703, + "grad_norm": 0.4996836185455322, + "learning_rate": 3.801722278746213e-05, + "loss": 0.3379, + "step": 191 + }, + { + "epoch": 0.6486486486486487, + "grad_norm": 0.4873179495334625, + "learning_rate": 3.8055020478392495e-05, + "loss": 0.3091, + "step": 192 + }, + { + "epoch": 0.652027027027027, + "grad_norm": 0.4989306330680847, + "learning_rate": 3.809262181679623e-05, + "loss": 0.3758, + "step": 193 + }, + { + "epoch": 0.6554054054054054, + "grad_norm": 0.5524582862854004, + "learning_rate": 3.813002883217044e-05, + "loss": 0.3044, + "step": 194 + }, + { + "epoch": 0.6587837837837838, + "grad_norm": 0.5253859162330627, + "learning_rate": 3.816724352270863e-05, + "loss": 0.3556, + "step": 195 + }, + { + "epoch": 0.6621621621621622, + "grad_norm": 0.5060839653015137, + "learning_rate": 3.8204267855941266e-05, + "loss": 0.3352, + "step": 196 + }, + { + "epoch": 0.6655405405405406, + "grad_norm": 0.5375229716300964, + "learning_rate": 3.824110376935989e-05, + "loss": 0.3328, + "step": 197 + }, + { + "epoch": 0.668918918918919, + "grad_norm": 0.4941028654575348, + "learning_rate": 3.827775317102552e-05, + "loss": 0.296, + "step": 198 + }, + { + "epoch": 0.6722972972972973, + "grad_norm": 0.5323918461799622, + "learning_rate": 3.831421794016178e-05, + "loss": 0.3049, + "step": 199 + }, + { + "epoch": 0.6756756756756757, + "grad_norm": 0.5118021965026855, + "learning_rate": 3.835049992773302e-05, + "loss": 0.3133, + "step": 200 + }, + { + "epoch": 0.6790540540540541, + "grad_norm": 0.46183663606643677, + "learning_rate": 3.838660095700815e-05, + "loss": 0.3185, + "step": 201 + }, + { + "epoch": 0.6824324324324325, + "grad_norm": 0.47574761509895325, + "learning_rate": 3.84225228241104e-05, + "loss": 0.302, + "step": 202 + }, + { + "epoch": 0.6858108108108109, + "grad_norm": 0.5129667520523071, + "learning_rate": 3.8458267298553554e-05, + "loss": 0.3407, + "step": 203 + }, + { + "epoch": 0.6891891891891891, + "grad_norm": 1.532895803451538, + "learning_rate": 3.8493836123764984e-05, + "loss": 0.2858, + "step": 204 + }, + { + "epoch": 0.6925675675675675, + "grad_norm": 0.5879459381103516, + "learning_rate": 3.852923101759591e-05, + "loss": 0.3638, + "step": 205 + }, + { + "epoch": 0.6959459459459459, + "grad_norm": 0.589292585849762, + "learning_rate": 3.856445367281923e-05, + "loss": 0.3077, + "step": 206 + }, + { + "epoch": 0.6993243243243243, + "grad_norm": 0.44842028617858887, + "learning_rate": 3.859950575761529e-05, + "loss": 0.2791, + "step": 207 + }, + { + "epoch": 0.7027027027027027, + "grad_norm": 0.6291372179985046, + "learning_rate": 3.8634388916046025e-05, + "loss": 0.3404, + "step": 208 + }, + { + "epoch": 0.706081081081081, + "grad_norm": 0.4508068561553955, + "learning_rate": 3.866910476851757e-05, + "loss": 0.2859, + "step": 209 + }, + { + "epoch": 0.7094594594594594, + "grad_norm": 0.7987334132194519, + "learning_rate": 3.870365491223199e-05, + "loss": 0.3224, + "step": 210 + }, + { + "epoch": 0.7128378378378378, + "grad_norm": 0.5030388236045837, + "learning_rate": 3.8738040921628215e-05, + "loss": 0.3225, + "step": 211 + }, + { + "epoch": 0.7162162162162162, + "grad_norm": 0.5433597564697266, + "learning_rate": 3.877226434881253e-05, + "loss": 0.3294, + "step": 212 + }, + { + "epoch": 0.7195945945945946, + "grad_norm": 0.48354920744895935, + "learning_rate": 3.880632672397897e-05, + "loss": 0.3035, + "step": 213 + }, + { + "epoch": 0.722972972972973, + "grad_norm": 0.5904508829116821, + "learning_rate": 3.884022955581985e-05, + "loss": 0.3261, + "step": 214 + }, + { + "epoch": 0.7263513513513513, + "grad_norm": 0.5152267813682556, + "learning_rate": 3.887397433192676e-05, + "loss": 0.2926, + "step": 215 + }, + { + "epoch": 0.7297297297297297, + "grad_norm": 0.46898239850997925, + "learning_rate": 3.890756251918219e-05, + "loss": 0.2808, + "step": 216 + }, + { + "epoch": 0.7331081081081081, + "grad_norm": 0.4848116338253021, + "learning_rate": 3.894099556414216e-05, + "loss": 0.2995, + "step": 217 + }, + { + "epoch": 0.7364864864864865, + "grad_norm": 0.48101919889450073, + "learning_rate": 3.897427489341009e-05, + "loss": 0.3054, + "step": 218 + }, + { + "epoch": 0.7398648648648649, + "grad_norm": 0.49954646825790405, + "learning_rate": 3.900740191400198e-05, + "loss": 0.3239, + "step": 219 + }, + { + "epoch": 0.7432432432432432, + "grad_norm": 0.46193623542785645, + "learning_rate": 3.904037801370344e-05, + "loss": 0.3215, + "step": 220 + }, + { + "epoch": 0.7466216216216216, + "grad_norm": 0.4544968605041504, + "learning_rate": 3.9073204561418514e-05, + "loss": 0.2829, + "step": 221 + }, + { + "epoch": 0.75, + "grad_norm": 0.4560447335243225, + "learning_rate": 3.9105882907510644e-05, + "loss": 0.3052, + "step": 222 + }, + { + "epoch": 0.7533783783783784, + "grad_norm": 0.4681329131126404, + "learning_rate": 3.913841438413601e-05, + "loss": 0.3259, + "step": 223 + }, + { + "epoch": 0.7567567567567568, + "grad_norm": 0.4784477949142456, + "learning_rate": 3.917080030556938e-05, + "loss": 0.3252, + "step": 224 + }, + { + "epoch": 0.7601351351351351, + "grad_norm": 0.4931364953517914, + "learning_rate": 3.9203041968522716e-05, + "loss": 0.3252, + "step": 225 + }, + { + "epoch": 0.7635135135135135, + "grad_norm": 0.4844168424606323, + "learning_rate": 3.923514065245669e-05, + "loss": 0.3185, + "step": 226 + }, + { + "epoch": 0.7668918918918919, + "grad_norm": 0.5114946961402893, + "learning_rate": 3.926709761988538e-05, + "loss": 0.3136, + "step": 227 + }, + { + "epoch": 0.7702702702702703, + "grad_norm": 0.4893558621406555, + "learning_rate": 3.929891411667424e-05, + "loss": 0.2977, + "step": 228 + }, + { + "epoch": 0.7736486486486487, + "grad_norm": 0.4515063762664795, + "learning_rate": 3.933059137233147e-05, + "loss": 0.3031, + "step": 229 + }, + { + "epoch": 0.777027027027027, + "grad_norm": 0.48304086923599243, + "learning_rate": 3.9362130600293214e-05, + "loss": 0.3292, + "step": 230 + }, + { + "epoch": 0.7804054054054054, + "grad_norm": 0.4545653760433197, + "learning_rate": 3.9393532998202405e-05, + "loss": 0.3014, + "step": 231 + }, + { + "epoch": 0.7837837837837838, + "grad_norm": 0.4236096739768982, + "learning_rate": 3.942479974818166e-05, + "loss": 0.2804, + "step": 232 + }, + { + "epoch": 0.7871621621621622, + "grad_norm": 0.4648077189922333, + "learning_rate": 3.945593201710032e-05, + "loss": 0.3243, + "step": 233 + }, + { + "epoch": 0.7905405405405406, + "grad_norm": 0.47247567772865295, + "learning_rate": 3.9486930956835724e-05, + "loss": 0.2907, + "step": 234 + }, + { + "epoch": 0.793918918918919, + "grad_norm": 0.4848794639110565, + "learning_rate": 3.951779770452894e-05, + "loss": 0.3265, + "step": 235 + }, + { + "epoch": 0.7972972972972973, + "grad_norm": 0.5114079713821411, + "learning_rate": 3.954853338283512e-05, + "loss": 0.3194, + "step": 236 + }, + { + "epoch": 0.8006756756756757, + "grad_norm": 0.4802190959453583, + "learning_rate": 3.9579139100168404e-05, + "loss": 0.2857, + "step": 237 + }, + { + "epoch": 0.8040540540540541, + "grad_norm": 0.522281289100647, + "learning_rate": 3.960961595094187e-05, + "loss": 0.2843, + "step": 238 + }, + { + "epoch": 0.8074324324324325, + "grad_norm": 0.4458360970020294, + "learning_rate": 3.96399650158023e-05, + "loss": 0.2755, + "step": 239 + }, + { + "epoch": 0.8108108108108109, + "grad_norm": 0.53693687915802, + "learning_rate": 3.96701873618601e-05, + "loss": 0.3015, + "step": 240 + }, + { + "epoch": 0.8141891891891891, + "grad_norm": 0.46697714924812317, + "learning_rate": 3.970028404291448e-05, + "loss": 0.2911, + "step": 241 + }, + { + "epoch": 0.8175675675675675, + "grad_norm": 0.5392897725105286, + "learning_rate": 3.9730256099673865e-05, + "loss": 0.3093, + "step": 242 + }, + { + "epoch": 0.8209459459459459, + "grad_norm": 0.5555176734924316, + "learning_rate": 3.976010455997187e-05, + "loss": 0.3078, + "step": 243 + }, + { + "epoch": 0.8243243243243243, + "grad_norm": 0.42975664138793945, + "learning_rate": 3.978983043897883e-05, + "loss": 0.2723, + "step": 244 + }, + { + "epoch": 0.8277027027027027, + "grad_norm": 0.6645532250404358, + "learning_rate": 3.981943473940888e-05, + "loss": 0.3161, + "step": 245 + }, + { + "epoch": 0.831081081081081, + "grad_norm": 0.47302716970443726, + "learning_rate": 3.984891845172299e-05, + "loss": 0.3002, + "step": 246 + }, + { + "epoch": 0.8344594594594594, + "grad_norm": 0.5395675301551819, + "learning_rate": 3.987828255432777e-05, + "loss": 0.3614, + "step": 247 + }, + { + "epoch": 0.8378378378378378, + "grad_norm": 0.5418782234191895, + "learning_rate": 3.9907528013770276e-05, + "loss": 0.3226, + "step": 248 + }, + { + "epoch": 0.8412162162162162, + "grad_norm": 0.4323344826698303, + "learning_rate": 3.993665578492894e-05, + "loss": 0.289, + "step": 249 + }, + { + "epoch": 0.8445945945945946, + "grad_norm": 0.5098227858543396, + "learning_rate": 3.9965666811200624e-05, + "loss": 0.3346, + "step": 250 + }, + { + "epoch": 0.847972972972973, + "grad_norm": 0.4577200412750244, + "learning_rate": 3.999456202468397e-05, + "loss": 0.3526, + "step": 251 + }, + { + "epoch": 0.8513513513513513, + "grad_norm": 0.48974189162254333, + "learning_rate": 4.002334234635907e-05, + "loss": 0.3086, + "step": 252 + }, + { + "epoch": 0.8547297297297297, + "grad_norm": 0.47080233693122864, + "learning_rate": 4.005200868626364e-05, + "loss": 0.3578, + "step": 253 + }, + { + "epoch": 0.8581081081081081, + "grad_norm": 0.5034478902816772, + "learning_rate": 4.008056194366564e-05, + "loss": 0.3166, + "step": 254 + }, + { + "epoch": 0.8614864864864865, + "grad_norm": 0.44603192806243896, + "learning_rate": 4.010900300723259e-05, + "loss": 0.2969, + "step": 255 + }, + { + "epoch": 0.8648648648648649, + "grad_norm": 0.5267269611358643, + "learning_rate": 4.013733275519749e-05, + "loss": 0.3341, + "step": 256 + }, + { + "epoch": 0.8682432432432432, + "grad_norm": 0.424925297498703, + "learning_rate": 4.016555205552158e-05, + "loss": 0.2942, + "step": 257 + }, + { + "epoch": 0.8716216216216216, + "grad_norm": 0.5034767389297485, + "learning_rate": 4.0193661766053834e-05, + "loss": 0.2668, + "step": 258 + }, + { + "epoch": 0.875, + "grad_norm": 0.5288417935371399, + "learning_rate": 4.022166273468753e-05, + "loss": 0.3424, + "step": 259 + }, + { + "epoch": 0.8783783783783784, + "grad_norm": 0.4726288616657257, + "learning_rate": 4.024955579951363e-05, + "loss": 0.2906, + "step": 260 + }, + { + "epoch": 0.8817567567567568, + "grad_norm": 0.5311090350151062, + "learning_rate": 4.027734178897136e-05, + "loss": 0.3307, + "step": 261 + }, + { + "epoch": 0.8851351351351351, + "grad_norm": 0.427002489566803, + "learning_rate": 4.030502152199576e-05, + "loss": 0.2569, + "step": 262 + }, + { + "epoch": 0.8885135135135135, + "grad_norm": 0.543899416923523, + "learning_rate": 4.033259580816264e-05, + "loss": 0.2925, + "step": 263 + }, + { + "epoch": 0.8918918918918919, + "grad_norm": 0.4932996928691864, + "learning_rate": 4.036006544783052e-05, + "loss": 0.3058, + "step": 264 + }, + { + "epoch": 0.8952702702702703, + "grad_norm": 0.5138298273086548, + "learning_rate": 4.0387431232280135e-05, + "loss": 0.2952, + "step": 265 + }, + { + "epoch": 0.8986486486486487, + "grad_norm": 0.4801797568798065, + "learning_rate": 4.041469394385112e-05, + "loss": 0.2873, + "step": 266 + }, + { + "epoch": 0.902027027027027, + "grad_norm": 0.453140527009964, + "learning_rate": 4.0441854356076257e-05, + "loss": 0.2747, + "step": 267 + }, + { + "epoch": 0.9054054054054054, + "grad_norm": 0.4941580593585968, + "learning_rate": 4.046891323381315e-05, + "loss": 0.3157, + "step": 268 + }, + { + "epoch": 0.9087837837837838, + "grad_norm": 0.6151431798934937, + "learning_rate": 4.049587133337347e-05, + "loss": 0.3139, + "step": 269 + }, + { + "epoch": 0.9121621621621622, + "grad_norm": 0.4674379229545593, + "learning_rate": 4.0522729402649793e-05, + "loss": 0.3122, + "step": 270 + }, + { + "epoch": 0.9155405405405406, + "grad_norm": 0.4719599187374115, + "learning_rate": 4.0549488181240096e-05, + "loss": 0.3101, + "step": 271 + }, + { + "epoch": 0.918918918918919, + "grad_norm": 0.483374685049057, + "learning_rate": 4.057614840056998e-05, + "loss": 0.3087, + "step": 272 + }, + { + "epoch": 0.9222972972972973, + "grad_norm": 0.48100754618644714, + "learning_rate": 4.06027107840126e-05, + "loss": 0.3044, + "step": 273 + }, + { + "epoch": 0.9256756756756757, + "grad_norm": 0.44335874915122986, + "learning_rate": 4.0629176047006474e-05, + "loss": 0.2929, + "step": 274 + }, + { + "epoch": 0.9290540540540541, + "grad_norm": 0.4475744962692261, + "learning_rate": 4.065554489717105e-05, + "loss": 0.321, + "step": 275 + }, + { + "epoch": 0.9324324324324325, + "grad_norm": 0.43009960651397705, + "learning_rate": 4.068181803442029e-05, + "loss": 0.3026, + "step": 276 + }, + { + "epoch": 0.9358108108108109, + "grad_norm": 0.48129919171333313, + "learning_rate": 4.0707996151074147e-05, + "loss": 0.2876, + "step": 277 + }, + { + "epoch": 0.9391891891891891, + "grad_norm": 0.423658549785614, + "learning_rate": 4.073407993196794e-05, + "loss": 0.2968, + "step": 278 + }, + { + "epoch": 0.9425675675675675, + "grad_norm": 0.485857218503952, + "learning_rate": 4.076007005455996e-05, + "loss": 0.3144, + "step": 279 + }, + { + "epoch": 0.9459459459459459, + "grad_norm": 0.4936545193195343, + "learning_rate": 4.0785967189036986e-05, + "loss": 0.3103, + "step": 280 + }, + { + "epoch": 0.9493243243243243, + "grad_norm": 0.47265681624412537, + "learning_rate": 4.0811771998418e-05, + "loss": 0.3136, + "step": 281 + }, + { + "epoch": 0.9527027027027027, + "grad_norm": 0.4872439503669739, + "learning_rate": 4.083748513865602e-05, + "loss": 0.335, + "step": 282 + }, + { + "epoch": 0.956081081081081, + "grad_norm": 0.4380621910095215, + "learning_rate": 4.086310725873818e-05, + "loss": 0.3036, + "step": 283 + }, + { + "epoch": 0.9594594594594594, + "grad_norm": 0.4759541451931, + "learning_rate": 4.0888639000783966e-05, + "loss": 0.2827, + "step": 284 + }, + { + "epoch": 0.9628378378378378, + "grad_norm": 0.995196521282196, + "learning_rate": 4.0914081000141844e-05, + "loss": 0.3162, + "step": 285 + }, + { + "epoch": 0.9662162162162162, + "grad_norm": 0.5686355829238892, + "learning_rate": 4.0939433885484055e-05, + "loss": 0.2992, + "step": 286 + }, + { + "epoch": 0.9695945945945946, + "grad_norm": 0.43086209893226624, + "learning_rate": 4.0964698278899874e-05, + "loss": 0.278, + "step": 287 + }, + { + "epoch": 0.972972972972973, + "grad_norm": 0.5099641680717468, + "learning_rate": 4.0989874795987185e-05, + "loss": 0.3023, + "step": 288 + }, + { + "epoch": 0.9763513513513513, + "grad_norm": 0.39634063839912415, + "learning_rate": 4.1014964045942465e-05, + "loss": 0.2806, + "step": 289 + }, + { + "epoch": 0.9797297297297297, + "grad_norm": 0.4685455858707428, + "learning_rate": 4.103996663164927e-05, + "loss": 0.3078, + "step": 290 + }, + { + "epoch": 0.9831081081081081, + "grad_norm": 0.4501771330833435, + "learning_rate": 4.106488314976513e-05, + "loss": 0.3048, + "step": 291 + }, + { + "epoch": 0.9864864864864865, + "grad_norm": 0.4352227747440338, + "learning_rate": 4.108971419080698e-05, + "loss": 0.305, + "step": 292 + }, + { + "epoch": 0.9898648648648649, + "grad_norm": 0.4946969151496887, + "learning_rate": 4.111446033923516e-05, + "loss": 0.3027, + "step": 293 + }, + { + "epoch": 0.9932432432432432, + "grad_norm": 0.463313490152359, + "learning_rate": 4.113912217353596e-05, + "loss": 0.3148, + "step": 294 + }, + { + "epoch": 0.9966216216216216, + "grad_norm": 0.42501333355903625, + "learning_rate": 4.116370026630272e-05, + "loss": 0.303, + "step": 295 + }, + { + "epoch": 1.0, + "grad_norm": 0.4174898862838745, + "learning_rate": 4.118819518431564e-05, + "loss": 0.2768, + "step": 296 + }, + { + "epoch": 1.0033783783783783, + "grad_norm": 0.5591414570808411, + "learning_rate": 4.121260748862021e-05, + "loss": 0.2319, + "step": 297 + }, + { + "epoch": 1.0067567567567568, + "grad_norm": 0.4628024101257324, + "learning_rate": 4.123693773460426e-05, + "loss": 0.2332, + "step": 298 + }, + { + "epoch": 1.010135135135135, + "grad_norm": 0.9493624567985535, + "learning_rate": 4.126118647207383e-05, + "loss": 0.2113, + "step": 299 + }, + { + "epoch": 1.0135135135135136, + "grad_norm": 0.6151244044303894, + "learning_rate": 4.1285354245327715e-05, + "loss": 0.2428, + "step": 300 + }, + { + "epoch": 1.0168918918918919, + "grad_norm": 0.4934402406215668, + "learning_rate": 4.1309441593230726e-05, + "loss": 0.2394, + "step": 301 + }, + { + "epoch": 1.0202702702702702, + "grad_norm": 0.5464348793029785, + "learning_rate": 4.133344904928585e-05, + "loss": 0.2584, + "step": 302 + }, + { + "epoch": 1.0236486486486487, + "grad_norm": 0.754717230796814, + "learning_rate": 4.1357377141705084e-05, + "loss": 0.2191, + "step": 303 + }, + { + "epoch": 1.027027027027027, + "grad_norm": 0.48623141646385193, + "learning_rate": 4.1381226393479236e-05, + "loss": 0.2185, + "step": 304 + }, + { + "epoch": 1.0304054054054055, + "grad_norm": 32.5340461730957, + "learning_rate": 4.1404997322446435e-05, + "loss": 0.2328, + "step": 305 + }, + { + "epoch": 1.0337837837837838, + "grad_norm": 0.7727927565574646, + "learning_rate": 4.142869044135967e-05, + "loss": 0.2197, + "step": 306 + }, + { + "epoch": 1.037162162162162, + "grad_norm": 0.47851279377937317, + "learning_rate": 4.145230625795311e-05, + "loss": 0.2093, + "step": 307 + }, + { + "epoch": 1.0405405405405406, + "grad_norm": 0.684687077999115, + "learning_rate": 4.14758452750074e-05, + "loss": 0.2145, + "step": 308 + }, + { + "epoch": 1.0439189189189189, + "grad_norm": 0.4834594428539276, + "learning_rate": 4.149930799041392e-05, + "loss": 0.222, + "step": 309 + }, + { + "epoch": 1.0472972972972974, + "grad_norm": 0.5479368567466736, + "learning_rate": 4.152269489723788e-05, + "loss": 0.2245, + "step": 310 + }, + { + "epoch": 1.0506756756756757, + "grad_norm": 0.48793625831604004, + "learning_rate": 4.1546006483780626e-05, + "loss": 0.2428, + "step": 311 + }, + { + "epoch": 1.054054054054054, + "grad_norm": 0.5712085962295532, + "learning_rate": 4.156924323364072e-05, + "loss": 0.2311, + "step": 312 + }, + { + "epoch": 1.0574324324324325, + "grad_norm": 0.46200135350227356, + "learning_rate": 4.1592405625774144e-05, + "loss": 0.2315, + "step": 313 + }, + { + "epoch": 1.0608108108108107, + "grad_norm": 0.5134051442146301, + "learning_rate": 4.161549413455358e-05, + "loss": 0.2205, + "step": 314 + }, + { + "epoch": 1.0641891891891893, + "grad_norm": 0.4155457019805908, + "learning_rate": 4.163850922982668e-05, + "loss": 0.1969, + "step": 315 + }, + { + "epoch": 1.0675675675675675, + "grad_norm": 0.47662997245788574, + "learning_rate": 4.16614513769734e-05, + "loss": 0.2183, + "step": 316 + }, + { + "epoch": 1.070945945945946, + "grad_norm": 0.4776044189929962, + "learning_rate": 4.1684321036962526e-05, + "loss": 0.2581, + "step": 317 + }, + { + "epoch": 1.0743243243243243, + "grad_norm": 0.5491393804550171, + "learning_rate": 4.170711866640721e-05, + "loss": 0.2454, + "step": 318 + }, + { + "epoch": 1.0777027027027026, + "grad_norm": 0.4745365381240845, + "learning_rate": 4.1729844717619684e-05, + "loss": 0.2303, + "step": 319 + }, + { + "epoch": 1.0810810810810811, + "grad_norm": 0.49116480350494385, + "learning_rate": 4.17524996386651e-05, + "loss": 0.2205, + "step": 320 + }, + { + "epoch": 1.0844594594594594, + "grad_norm": 0.47651857137680054, + "learning_rate": 4.177508387341454e-05, + "loss": 0.2538, + "step": 321 + }, + { + "epoch": 1.087837837837838, + "grad_norm": 24.205638885498047, + "learning_rate": 4.179759786159719e-05, + "loss": 0.5427, + "step": 322 + }, + { + "epoch": 1.0912162162162162, + "grad_norm": 0.7813121676445007, + "learning_rate": 4.182004203885172e-05, + "loss": 0.2445, + "step": 323 + }, + { + "epoch": 1.0945945945945945, + "grad_norm": 52.29485321044922, + "learning_rate": 4.184241683677687e-05, + "loss": 0.2577, + "step": 324 + }, + { + "epoch": 1.097972972972973, + "grad_norm": 0.8282439112663269, + "learning_rate": 4.1864722682981245e-05, + "loss": 0.2421, + "step": 325 + }, + { + "epoch": 1.1013513513513513, + "grad_norm": 0.4729321300983429, + "learning_rate": 4.188696000113232e-05, + "loss": 0.2122, + "step": 326 + }, + { + "epoch": 1.1047297297297298, + "grad_norm": 0.6732134819030762, + "learning_rate": 4.190912921100477e-05, + "loss": 0.2295, + "step": 327 + }, + { + "epoch": 1.1081081081081081, + "grad_norm": 0.567088782787323, + "learning_rate": 4.1931230728527994e-05, + "loss": 0.2352, + "step": 328 + }, + { + "epoch": 1.1114864864864864, + "grad_norm": 8.066313743591309, + "learning_rate": 4.195326496583291e-05, + "loss": 0.7099, + "step": 329 + }, + { + "epoch": 1.114864864864865, + "grad_norm": 1.0606003999710083, + "learning_rate": 4.1975232331298125e-05, + "loss": 0.2105, + "step": 330 + }, + { + "epoch": 1.1182432432432432, + "grad_norm": 0.4454852342605591, + "learning_rate": 4.1997133229595316e-05, + "loss": 0.2063, + "step": 331 + }, + { + "epoch": 1.1216216216216217, + "grad_norm": 0.873603880405426, + "learning_rate": 4.201896806173394e-05, + "loss": 0.2447, + "step": 332 + }, + { + "epoch": 1.125, + "grad_norm": 0.6670548319816589, + "learning_rate": 4.2040737225105335e-05, + "loss": 0.2298, + "step": 333 + }, + { + "epoch": 1.1283783783783783, + "grad_norm": 0.5834859609603882, + "learning_rate": 4.206244111352608e-05, + "loss": 0.1799, + "step": 334 + }, + { + "epoch": 1.1317567567567568, + "grad_norm": 0.8503464460372925, + "learning_rate": 4.2084080117280756e-05, + "loss": 0.2254, + "step": 335 + }, + { + "epoch": 1.135135135135135, + "grad_norm": 0.5728862285614014, + "learning_rate": 4.210565462316407e-05, + "loss": 0.2342, + "step": 336 + }, + { + "epoch": 1.1385135135135136, + "grad_norm": 0.7374505400657654, + "learning_rate": 4.2127165014522315e-05, + "loss": 0.2384, + "step": 337 + }, + { + "epoch": 1.1418918918918919, + "grad_norm": 0.5631051659584045, + "learning_rate": 4.214861167129425e-05, + "loss": 0.2213, + "step": 338 + }, + { + "epoch": 1.1452702702702702, + "grad_norm": 26.428640365600586, + "learning_rate": 4.2169994970051365e-05, + "loss": 0.8302, + "step": 339 + }, + { + "epoch": 1.1486486486486487, + "grad_norm": 1.1665905714035034, + "learning_rate": 4.219131528403759e-05, + "loss": 0.2568, + "step": 340 + }, + { + "epoch": 1.152027027027027, + "grad_norm": 0.6159250140190125, + "learning_rate": 4.22125729832083e-05, + "loss": 0.2574, + "step": 341 + }, + { + "epoch": 1.1554054054054055, + "grad_norm": 182.47439575195312, + "learning_rate": 4.2233768434268914e-05, + "loss": 0.8118, + "step": 342 + }, + { + "epoch": 1.1587837837837838, + "grad_norm": 1.685766339302063, + "learning_rate": 4.225490200071284e-05, + "loss": 0.2347, + "step": 343 + }, + { + "epoch": 1.1621621621621623, + "grad_norm": 0.9619600176811218, + "learning_rate": 4.227597404285883e-05, + "loss": 0.2453, + "step": 344 + }, + { + "epoch": 1.1655405405405406, + "grad_norm": 0.49926599860191345, + "learning_rate": 4.229698491788791e-05, + "loss": 0.2313, + "step": 345 + }, + { + "epoch": 1.1689189189189189, + "grad_norm": 0.9699143767356873, + "learning_rate": 4.231793497987961e-05, + "loss": 0.256, + "step": 346 + }, + { + "epoch": 1.1722972972972974, + "grad_norm": 0.7428460121154785, + "learning_rate": 4.2338824579847904e-05, + "loss": 0.2655, + "step": 347 + }, + { + "epoch": 1.1756756756756757, + "grad_norm": 0.6480849385261536, + "learning_rate": 4.235965406577636e-05, + "loss": 0.2385, + "step": 348 + }, + { + "epoch": 1.179054054054054, + "grad_norm": 0.6347863078117371, + "learning_rate": 4.2380423782653e-05, + "loss": 0.2196, + "step": 349 + }, + { + "epoch": 1.1824324324324325, + "grad_norm": 0.6471324563026428, + "learning_rate": 4.240113407250459e-05, + "loss": 0.2337, + "step": 350 + }, + { + "epoch": 1.1858108108108107, + "grad_norm": 0.5807657241821289, + "learning_rate": 4.24217852744304e-05, + "loss": 0.2438, + "step": 351 + }, + { + "epoch": 1.1891891891891893, + "grad_norm": 0.5200501084327698, + "learning_rate": 4.244237772463552e-05, + "loss": 0.2458, + "step": 352 + }, + { + "epoch": 1.1925675675675675, + "grad_norm": 0.5445655584335327, + "learning_rate": 4.246291175646371e-05, + "loss": 0.2391, + "step": 353 + }, + { + "epoch": 1.195945945945946, + "grad_norm": 0.6283414363861084, + "learning_rate": 4.24833877004298e-05, + "loss": 0.2299, + "step": 354 + }, + { + "epoch": 1.1993243243243243, + "grad_norm": 0.5581656694412231, + "learning_rate": 4.250380588425157e-05, + "loss": 0.2272, + "step": 355 + }, + { + "epoch": 1.2027027027027026, + "grad_norm": 0.5711120367050171, + "learning_rate": 4.2524166632881255e-05, + "loss": 0.2737, + "step": 356 + }, + { + "epoch": 1.2060810810810811, + "grad_norm": 64.4117660522461, + "learning_rate": 4.254447026853656e-05, + "loss": 0.2648, + "step": 357 + }, + { + "epoch": 1.2094594594594594, + "grad_norm": 0.84456467628479, + "learning_rate": 4.2564717110731244e-05, + "loss": 0.2643, + "step": 358 + }, + { + "epoch": 1.2128378378378377, + "grad_norm": 1.7171592712402344, + "learning_rate": 4.258490747630532e-05, + "loss": 0.6041, + "step": 359 + }, + { + "epoch": 1.2162162162162162, + "grad_norm": 0.6733882427215576, + "learning_rate": 4.260504167945479e-05, + "loss": 0.2226, + "step": 360 + }, + { + "epoch": 1.2195945945945945, + "grad_norm": 0.5079653859138489, + "learning_rate": 4.2625120031760965e-05, + "loss": 0.2205, + "step": 361 + }, + { + "epoch": 1.222972972972973, + "grad_norm": 0.5646266341209412, + "learning_rate": 4.264514284221944e-05, + "loss": 0.2272, + "step": 362 + }, + { + "epoch": 1.2263513513513513, + "grad_norm": 0.5336301922798157, + "learning_rate": 4.266511041726854e-05, + "loss": 0.2536, + "step": 363 + }, + { + "epoch": 1.2297297297297298, + "grad_norm": 0.5218775272369385, + "learning_rate": 4.26850230608176e-05, + "loss": 0.2582, + "step": 364 + }, + { + "epoch": 1.2331081081081081, + "grad_norm": 0.47051167488098145, + "learning_rate": 4.2704881074274584e-05, + "loss": 0.2288, + "step": 365 + }, + { + "epoch": 1.2364864864864864, + "grad_norm": 0.4820377826690674, + "learning_rate": 4.272468475657351e-05, + "loss": 0.2396, + "step": 366 + }, + { + "epoch": 1.239864864864865, + "grad_norm": 1219.8048095703125, + "learning_rate": 4.2744434404201497e-05, + "loss": 1.9213, + "step": 367 + }, + { + "epoch": 1.2432432432432432, + "grad_norm": 1.3034559488296509, + "learning_rate": 4.27641303112253e-05, + "loss": 0.2528, + "step": 368 + }, + { + "epoch": 1.2466216216216217, + "grad_norm": 0.6443231105804443, + "learning_rate": 4.278377276931767e-05, + "loss": 0.2235, + "step": 369 + }, + { + "epoch": 1.25, + "grad_norm": 0.4247282147407532, + "learning_rate": 4.2803362067783256e-05, + "loss": 0.2111, + "step": 370 + }, + { + "epoch": 1.2533783783783785, + "grad_norm": 0.5400434136390686, + "learning_rate": 4.2822898493584104e-05, + "loss": 0.2256, + "step": 371 + }, + { + "epoch": 1.2567567567567568, + "grad_norm": 0.43781599402427673, + "learning_rate": 4.284238233136496e-05, + "loss": 0.2295, + "step": 372 + }, + { + "epoch": 1.260135135135135, + "grad_norm": 0.5443878173828125, + "learning_rate": 4.286181386347813e-05, + "loss": 0.2394, + "step": 373 + }, + { + "epoch": 1.2635135135135136, + "grad_norm": 0.4616173207759857, + "learning_rate": 4.288119337000801e-05, + "loss": 0.2108, + "step": 374 + }, + { + "epoch": 1.2668918918918919, + "grad_norm": 0.4537034332752228, + "learning_rate": 4.2900521128795315e-05, + "loss": 0.2192, + "step": 375 + }, + { + "epoch": 1.2702702702702702, + "grad_norm": 0.5639699697494507, + "learning_rate": 4.291979741546102e-05, + "loss": 0.2403, + "step": 376 + }, + { + "epoch": 1.2736486486486487, + "grad_norm": 0.4133036434650421, + "learning_rate": 4.293902250342989e-05, + "loss": 0.2086, + "step": 377 + }, + { + "epoch": 1.277027027027027, + "grad_norm": 0.45922860503196716, + "learning_rate": 4.295819666395376e-05, + "loss": 0.236, + "step": 378 + }, + { + "epoch": 1.2804054054054055, + "grad_norm": 0.4423050284385681, + "learning_rate": 4.297732016613454e-05, + "loss": 0.2271, + "step": 379 + }, + { + "epoch": 1.2837837837837838, + "grad_norm": 0.4588952362537384, + "learning_rate": 4.299639327694684e-05, + "loss": 0.2356, + "step": 380 + }, + { + "epoch": 1.2871621621621623, + "grad_norm": 26.979991912841797, + "learning_rate": 4.3015416261260325e-05, + "loss": 0.2247, + "step": 381 + }, + { + "epoch": 1.2905405405405406, + "grad_norm": 0.6023054718971252, + "learning_rate": 4.303438938186182e-05, + "loss": 0.2477, + "step": 382 + }, + { + "epoch": 1.2939189189189189, + "grad_norm": 0.48854556679725647, + "learning_rate": 4.305331289947705e-05, + "loss": 0.221, + "step": 383 + }, + { + "epoch": 1.2972972972972974, + "grad_norm": 0.5875428915023804, + "learning_rate": 4.3072187072792184e-05, + "loss": 0.2443, + "step": 384 + }, + { + "epoch": 1.3006756756756757, + "grad_norm": 0.42068448662757874, + "learning_rate": 4.309101215847502e-05, + "loss": 0.2043, + "step": 385 + }, + { + "epoch": 1.304054054054054, + "grad_norm": 0.492660254240036, + "learning_rate": 4.3109788411195924e-05, + "loss": 0.246, + "step": 386 + }, + { + "epoch": 1.3074324324324325, + "grad_norm": 0.4522532820701599, + "learning_rate": 4.312851608364853e-05, + "loss": 0.2305, + "step": 387 + }, + { + "epoch": 1.3108108108108107, + "grad_norm": 0.4410349130630493, + "learning_rate": 4.314719542657013e-05, + "loss": 0.2146, + "step": 388 + }, + { + "epoch": 1.3141891891891893, + "grad_norm": 38.59849548339844, + "learning_rate": 4.3165826688761796e-05, + "loss": 1.1445, + "step": 389 + }, + { + "epoch": 1.3175675675675675, + "grad_norm": 0.6608863472938538, + "learning_rate": 4.318441011710833e-05, + "loss": 0.2591, + "step": 390 + }, + { + "epoch": 1.320945945945946, + "grad_norm": 0.7261571884155273, + "learning_rate": 4.3202945956597786e-05, + "loss": 0.2409, + "step": 391 + }, + { + "epoch": 1.3243243243243243, + "grad_norm": 13.75738525390625, + "learning_rate": 4.3221434450340956e-05, + "loss": 1.022, + "step": 392 + }, + { + "epoch": 1.3277027027027026, + "grad_norm": 1.4299050569534302, + "learning_rate": 4.323987583959045e-05, + "loss": 0.2287, + "step": 393 + }, + { + "epoch": 1.3310810810810811, + "grad_norm": 0.6660119295120239, + "learning_rate": 4.325827036375957e-05, + "loss": 0.2148, + "step": 394 + }, + { + "epoch": 1.3344594594594594, + "grad_norm": 3.13312029838562, + "learning_rate": 4.327661826044101e-05, + "loss": 0.6968, + "step": 395 + }, + { + "epoch": 1.3378378378378377, + "grad_norm": 29.60978126525879, + "learning_rate": 4.329491976542521e-05, + "loss": 0.6915, + "step": 396 + }, + { + "epoch": 1.3412162162162162, + "grad_norm": 0.8736880421638489, + "learning_rate": 4.331317511271859e-05, + "loss": 0.2388, + "step": 397 + }, + { + "epoch": 1.3445945945945945, + "grad_norm": 0.5873726606369019, + "learning_rate": 4.333138453456147e-05, + "loss": 0.2655, + "step": 398 + }, + { + "epoch": 1.347972972972973, + "grad_norm": 2.68512225151062, + "learning_rate": 4.334954826144581e-05, + "loss": 0.6129, + "step": 399 + }, + { + "epoch": 1.3513513513513513, + "grad_norm": 0.8919135928153992, + "learning_rate": 4.336766652213271e-05, + "loss": 0.2492, + "step": 400 + }, + { + "epoch": 1.3547297297297298, + "grad_norm": 2.113346576690674, + "learning_rate": 4.338573954366971e-05, + "loss": 0.2578, + "step": 401 + }, + { + "epoch": 1.3581081081081081, + "grad_norm": 2.666167736053467, + "learning_rate": 4.340376755140784e-05, + "loss": 0.6334, + "step": 402 + }, + { + "epoch": 1.3614864864864864, + "grad_norm": 0.8454524278640747, + "learning_rate": 4.342175076901849e-05, + "loss": 0.2426, + "step": 403 + }, + { + "epoch": 1.364864864864865, + "grad_norm": 0.7469472885131836, + "learning_rate": 4.343968941851009e-05, + "loss": 0.2715, + "step": 404 + }, + { + "epoch": 1.3682432432432432, + "grad_norm": 0.6965801119804382, + "learning_rate": 4.345758372024448e-05, + "loss": 0.2462, + "step": 405 + }, + { + "epoch": 1.3716216216216215, + "grad_norm": 1.5828707218170166, + "learning_rate": 4.347543389295324e-05, + "loss": 0.5967, + "step": 406 + }, + { + "epoch": 1.375, + "grad_norm": 0.5627582669258118, + "learning_rate": 4.3493240153753666e-05, + "loss": 0.226, + "step": 407 + }, + { + "epoch": 1.3783783783783785, + "grad_norm": 0.5101889967918396, + "learning_rate": 4.3511002718164666e-05, + "loss": 0.2266, + "step": 408 + }, + { + "epoch": 1.3817567567567568, + "grad_norm": 0.6242619156837463, + "learning_rate": 4.352872180012237e-05, + "loss": 0.2028, + "step": 409 + }, + { + "epoch": 1.385135135135135, + "grad_norm": 0.5768154263496399, + "learning_rate": 4.35463976119956e-05, + "loss": 0.2291, + "step": 410 + }, + { + "epoch": 1.3885135135135136, + "grad_norm": 0.5170788764953613, + "learning_rate": 4.356403036460115e-05, + "loss": 0.2124, + "step": 411 + }, + { + "epoch": 1.3918918918918919, + "grad_norm": 0.530846357345581, + "learning_rate": 4.3581620267218916e-05, + "loss": 0.2013, + "step": 412 + }, + { + "epoch": 1.3952702702702702, + "grad_norm": 0.5346778631210327, + "learning_rate": 4.359916752760669e-05, + "loss": 0.2439, + "step": 413 + }, + { + "epoch": 1.3986486486486487, + "grad_norm": 0.5345392227172852, + "learning_rate": 4.361667235201499e-05, + "loss": 0.2321, + "step": 414 + }, + { + "epoch": 1.402027027027027, + "grad_norm": 0.5348331928253174, + "learning_rate": 4.363413494520154e-05, + "loss": 0.2418, + "step": 415 + }, + { + "epoch": 1.4054054054054055, + "grad_norm": 0.9726234078407288, + "learning_rate": 4.365155551044572e-05, + "loss": 0.213, + "step": 416 + }, + { + "epoch": 1.4087837837837838, + "grad_norm": 2.116424083709717, + "learning_rate": 4.366893424956263e-05, + "loss": 0.5716, + "step": 417 + }, + { + "epoch": 1.4121621621621623, + "grad_norm": 0.6004624962806702, + "learning_rate": 4.368627136291726e-05, + "loss": 0.2404, + "step": 418 + }, + { + "epoch": 1.4155405405405406, + "grad_norm": 0.44696101546287537, + "learning_rate": 4.370356704943825e-05, + "loss": 0.2349, + "step": 419 + }, + { + "epoch": 1.4189189189189189, + "grad_norm": 0.4704833924770355, + "learning_rate": 4.372082150663168e-05, + "loss": 0.2365, + "step": 420 + }, + { + "epoch": 1.4222972972972974, + "grad_norm": 0.48295876383781433, + "learning_rate": 4.3738034930594475e-05, + "loss": 0.2256, + "step": 421 + }, + { + "epoch": 1.4256756756756757, + "grad_norm": 0.4726882576942444, + "learning_rate": 4.3755207516027904e-05, + "loss": 0.2283, + "step": 422 + }, + { + "epoch": 1.429054054054054, + "grad_norm": 0.5251411199569702, + "learning_rate": 4.377233945625071e-05, + "loss": 0.2374, + "step": 423 + }, + { + "epoch": 1.4324324324324325, + "grad_norm": 94.52925109863281, + "learning_rate": 4.378943094321221e-05, + "loss": 0.4887, + "step": 424 + }, + { + "epoch": 1.4358108108108107, + "grad_norm": 0.5848643183708191, + "learning_rate": 4.3806482167505196e-05, + "loss": 0.2373, + "step": 425 + }, + { + "epoch": 1.4391891891891893, + "grad_norm": 0.870020866394043, + "learning_rate": 4.382349331837866e-05, + "loss": 0.5312, + "step": 426 + }, + { + "epoch": 1.4425675675675675, + "grad_norm": 0.58022540807724, + "learning_rate": 4.3840464583750404e-05, + "loss": 0.2347, + "step": 427 + }, + { + "epoch": 1.445945945945946, + "grad_norm": 0.4780952036380768, + "learning_rate": 4.385739615021954e-05, + "loss": 0.2392, + "step": 428 + }, + { + "epoch": 1.4493243243243243, + "grad_norm": 0.4691276252269745, + "learning_rate": 4.387428820307874e-05, + "loss": 0.2346, + "step": 429 + }, + { + "epoch": 1.4527027027027026, + "grad_norm": 0.4635365903377533, + "learning_rate": 4.3891140926326446e-05, + "loss": 0.2242, + "step": 430 + }, + { + "epoch": 1.4560810810810811, + "grad_norm": 0.7928866147994995, + "learning_rate": 4.390795450267886e-05, + "loss": 0.2267, + "step": 431 + }, + { + "epoch": 1.4594594594594594, + "grad_norm": 0.5844922661781311, + "learning_rate": 4.3924729113581876e-05, + "loss": 0.237, + "step": 432 + }, + { + "epoch": 1.4628378378378377, + "grad_norm": 0.4539477229118347, + "learning_rate": 4.394146493922276e-05, + "loss": 0.2624, + "step": 433 + }, + { + "epoch": 1.4662162162162162, + "grad_norm": 0.4881402254104614, + "learning_rate": 4.395816215854185e-05, + "loss": 0.2076, + "step": 434 + }, + { + "epoch": 1.4695945945945945, + "grad_norm": 0.42912018299102783, + "learning_rate": 4.397482094924396e-05, + "loss": 0.224, + "step": 435 + }, + { + "epoch": 1.472972972972973, + "grad_norm": 8.578511238098145, + "learning_rate": 4.399144148780977e-05, + "loss": 0.5361, + "step": 436 + }, + { + "epoch": 1.4763513513513513, + "grad_norm": 0.4810888469219208, + "learning_rate": 4.400802394950703e-05, + "loss": 0.2274, + "step": 437 + }, + { + "epoch": 1.4797297297297298, + "grad_norm": 0.5107089281082153, + "learning_rate": 4.402456850840166e-05, + "loss": 0.2295, + "step": 438 + }, + { + "epoch": 1.4831081081081081, + "grad_norm": 0.4564654529094696, + "learning_rate": 4.4041075337368695e-05, + "loss": 0.2447, + "step": 439 + }, + { + "epoch": 1.4864864864864864, + "grad_norm": 1.4533860683441162, + "learning_rate": 4.405754460810312e-05, + "loss": 0.2522, + "step": 440 + }, + { + "epoch": 1.489864864864865, + "grad_norm": 2.172696828842163, + "learning_rate": 4.407397649113065e-05, + "loss": 0.8104, + "step": 441 + }, + { + "epoch": 1.4932432432432432, + "grad_norm": 0.4864687919616699, + "learning_rate": 4.40903711558182e-05, + "loss": 0.2657, + "step": 442 + }, + { + "epoch": 1.4966216216216215, + "grad_norm": 0.4211529791355133, + "learning_rate": 4.41067287703845e-05, + "loss": 0.246, + "step": 443 + }, + { + "epoch": 1.5, + "grad_norm": 0.5068204998970032, + "learning_rate": 4.412304950191033e-05, + "loss": 0.219, + "step": 444 + }, + { + "epoch": 1.5033783783783785, + "grad_norm": 0.52825927734375, + "learning_rate": 4.413933351634886e-05, + "loss": 0.2293, + "step": 445 + }, + { + "epoch": 1.5067567567567568, + "grad_norm": 0.45864129066467285, + "learning_rate": 4.4155580978535707e-05, + "loss": 0.218, + "step": 446 + }, + { + "epoch": 1.510135135135135, + "grad_norm": 1.0356566905975342, + "learning_rate": 4.417179205219895e-05, + "loss": 0.2278, + "step": 447 + }, + { + "epoch": 1.5135135135135136, + "grad_norm": 0.4032718241214752, + "learning_rate": 4.418796689996907e-05, + "loss": 0.2276, + "step": 448 + }, + { + "epoch": 1.5168918918918919, + "grad_norm": 0.4162694811820984, + "learning_rate": 4.420410568338872e-05, + "loss": 0.23, + "step": 449 + }, + { + "epoch": 1.5202702702702702, + "grad_norm": 0.4550395607948303, + "learning_rate": 4.42202085629224e-05, + "loss": 0.2445, + "step": 450 + }, + { + "epoch": 1.5236486486486487, + "grad_norm": 0.5222985148429871, + "learning_rate": 4.423627569796601e-05, + "loss": 0.2405, + "step": 451 + }, + { + "epoch": 1.527027027027027, + "grad_norm": 0.4746388792991638, + "learning_rate": 4.425230724685638e-05, + "loss": 0.2299, + "step": 452 + }, + { + "epoch": 1.5304054054054053, + "grad_norm": 0.47175803780555725, + "learning_rate": 4.4268303366880536e-05, + "loss": 0.234, + "step": 453 + }, + { + "epoch": 1.5337837837837838, + "grad_norm": 0.46304166316986084, + "learning_rate": 4.428426421428507e-05, + "loss": 0.2165, + "step": 454 + }, + { + "epoch": 1.5371621621621623, + "grad_norm": 0.8317951560020447, + "learning_rate": 4.430018994428521e-05, + "loss": 0.2648, + "step": 455 + }, + { + "epoch": 1.5405405405405406, + "grad_norm": 0.4764087498188019, + "learning_rate": 4.431608071107392e-05, + "loss": 0.2526, + "step": 456 + }, + { + "epoch": 1.5439189189189189, + "grad_norm": 0.39065808057785034, + "learning_rate": 4.433193666783084e-05, + "loss": 0.2428, + "step": 457 + }, + { + "epoch": 1.5472972972972974, + "grad_norm": 0.40975654125213623, + "learning_rate": 4.4347757966731156e-05, + "loss": 0.2437, + "step": 458 + }, + { + "epoch": 1.5506756756756757, + "grad_norm": 0.4318806231021881, + "learning_rate": 4.436354475895436e-05, + "loss": 0.233, + "step": 459 + }, + { + "epoch": 1.554054054054054, + "grad_norm": 0.4574219882488251, + "learning_rate": 4.437929719469291e-05, + "loss": 0.2441, + "step": 460 + }, + { + "epoch": 1.5574324324324325, + "grad_norm": 0.39155369997024536, + "learning_rate": 4.4395015423160807e-05, + "loss": 0.2233, + "step": 461 + }, + { + "epoch": 1.560810810810811, + "grad_norm": 0.4589376747608185, + "learning_rate": 4.4410699592602094e-05, + "loss": 0.225, + "step": 462 + }, + { + "epoch": 1.564189189189189, + "grad_norm": 0.8952370882034302, + "learning_rate": 4.442634985029922e-05, + "loss": 0.4926, + "step": 463 + }, + { + "epoch": 1.5675675675675675, + "grad_norm": 0.43023157119750977, + "learning_rate": 4.444196634258136e-05, + "loss": 0.2333, + "step": 464 + }, + { + "epoch": 1.570945945945946, + "grad_norm": 1.1993354558944702, + "learning_rate": 4.4457549214832566e-05, + "loss": 0.2457, + "step": 465 + }, + { + "epoch": 1.5743243243243243, + "grad_norm": 0.3993515074253082, + "learning_rate": 4.44730986115e-05, + "loss": 0.2457, + "step": 466 + }, + { + "epoch": 1.5777027027027026, + "grad_norm": 0.4565802812576294, + "learning_rate": 4.448861467610187e-05, + "loss": 0.2335, + "step": 467 + }, + { + "epoch": 1.5810810810810811, + "grad_norm": 0.4085776209831238, + "learning_rate": 4.4504097551235406e-05, + "loss": 0.237, + "step": 468 + }, + { + "epoch": 1.5844594594594594, + "grad_norm": 0.4212399125099182, + "learning_rate": 4.4519547378584725e-05, + "loss": 0.24, + "step": 469 + }, + { + "epoch": 1.5878378378378377, + "grad_norm": 0.408805251121521, + "learning_rate": 4.453496429892863e-05, + "loss": 0.2252, + "step": 470 + }, + { + "epoch": 1.5912162162162162, + "grad_norm": 0.42911630868911743, + "learning_rate": 4.455034845214827e-05, + "loss": 0.2099, + "step": 471 + }, + { + "epoch": 1.5945945945945947, + "grad_norm": 0.47564277052879333, + "learning_rate": 4.4565699977234796e-05, + "loss": 0.2135, + "step": 472 + }, + { + "epoch": 1.597972972972973, + "grad_norm": 0.3962784707546234, + "learning_rate": 4.458101901229686e-05, + "loss": 0.2346, + "step": 473 + }, + { + "epoch": 1.6013513513513513, + "grad_norm": 0.3911672532558441, + "learning_rate": 4.459630569456809e-05, + "loss": 0.2198, + "step": 474 + }, + { + "epoch": 1.6047297297297298, + "grad_norm": 1093.1121826171875, + "learning_rate": 4.461156016041444e-05, + "loss": 0.2989, + "step": 475 + }, + { + "epoch": 1.6081081081081081, + "grad_norm": 3.732886552810669, + "learning_rate": 4.462678254534156e-05, + "loss": 0.5155, + "step": 476 + }, + { + "epoch": 1.6114864864864864, + "grad_norm": 0.5420750975608826, + "learning_rate": 4.464197298400191e-05, + "loss": 0.2645, + "step": 477 + }, + { + "epoch": 1.614864864864865, + "grad_norm": 4.157422065734863, + "learning_rate": 4.4657131610201994e-05, + "loss": 0.5192, + "step": 478 + }, + { + "epoch": 1.6182432432432432, + "grad_norm": 0.4917280972003937, + "learning_rate": 4.467225855690939e-05, + "loss": 0.2187, + "step": 479 + }, + { + "epoch": 1.6216216216216215, + "grad_norm": 0.4480583071708679, + "learning_rate": 4.468735395625979e-05, + "loss": 0.2214, + "step": 480 + }, + { + "epoch": 1.625, + "grad_norm": 0.4467374384403229, + "learning_rate": 4.470241793956387e-05, + "loss": 0.2414, + "step": 481 + }, + { + "epoch": 1.6283783783783785, + "grad_norm": 0.48159006237983704, + "learning_rate": 4.471745063731416e-05, + "loss": 0.2386, + "step": 482 + }, + { + "epoch": 1.6317567567567568, + "grad_norm": 0.3858594000339508, + "learning_rate": 4.473245217919187e-05, + "loss": 0.1981, + "step": 483 + }, + { + "epoch": 1.635135135135135, + "grad_norm": 0.4672364592552185, + "learning_rate": 4.474742269407355e-05, + "loss": 0.2212, + "step": 484 + }, + { + "epoch": 1.6385135135135136, + "grad_norm": 0.5253187417984009, + "learning_rate": 4.476236231003773e-05, + "loss": 0.2753, + "step": 485 + }, + { + "epoch": 1.6418918918918919, + "grad_norm": 0.488766610622406, + "learning_rate": 4.477727115437156e-05, + "loss": 0.2571, + "step": 486 + }, + { + "epoch": 1.6452702702702702, + "grad_norm": 0.4526199996471405, + "learning_rate": 4.479214935357724e-05, + "loss": 0.2341, + "step": 487 + }, + { + "epoch": 1.6486486486486487, + "grad_norm": 0.4283704459667206, + "learning_rate": 4.480699703337852e-05, + "loss": 0.2533, + "step": 488 + }, + { + "epoch": 1.652027027027027, + "grad_norm": 0.45583993196487427, + "learning_rate": 4.4821814318727016e-05, + "loss": 0.2394, + "step": 489 + }, + { + "epoch": 1.6554054054054053, + "grad_norm": 0.3872677683830261, + "learning_rate": 4.483660133380856e-05, + "loss": 0.2005, + "step": 490 + }, + { + "epoch": 1.6587837837837838, + "grad_norm": 0.5054983496665955, + "learning_rate": 4.485135820204948e-05, + "loss": 0.231, + "step": 491 + }, + { + "epoch": 1.6621621621621623, + "grad_norm": 0.40404146909713745, + "learning_rate": 4.486608504612267e-05, + "loss": 0.2529, + "step": 492 + }, + { + "epoch": 1.6655405405405406, + "grad_norm": 0.4361649453639984, + "learning_rate": 4.488078198795383e-05, + "loss": 0.2485, + "step": 493 + }, + { + "epoch": 1.6689189189189189, + "grad_norm": 0.4425160586833954, + "learning_rate": 4.489544914872745e-05, + "loss": 0.2375, + "step": 494 + }, + { + "epoch": 1.6722972972972974, + "grad_norm": 0.4194391965866089, + "learning_rate": 4.4910086648892815e-05, + "loss": 0.2461, + "step": 495 + }, + { + "epoch": 1.6756756756756757, + "grad_norm": 0.42155027389526367, + "learning_rate": 4.4924694608169965e-05, + "loss": 0.2301, + "step": 496 + }, + { + "epoch": 1.679054054054054, + "grad_norm": 0.412643164396286, + "learning_rate": 4.4939273145555536e-05, + "loss": 0.235, + "step": 497 + }, + { + "epoch": 1.6824324324324325, + "grad_norm": 0.3827148973941803, + "learning_rate": 4.495382237932863e-05, + "loss": 0.23, + "step": 498 + }, + { + "epoch": 1.685810810810811, + "grad_norm": 0.4173565208911896, + "learning_rate": 4.4968342427056505e-05, + "loss": 0.2374, + "step": 499 + }, + { + "epoch": 1.689189189189189, + "grad_norm": 1055.430908203125, + "learning_rate": 4.498283340560031e-05, + "loss": 0.3647, + "step": 500 + }, + { + "epoch": 1.6925675675675675, + "grad_norm": 0.5496036410331726, + "learning_rate": 4.499729543112076e-05, + "loss": 0.2686, + "step": 501 + }, + { + "epoch": 1.695945945945946, + "grad_norm": 0.437640905380249, + "learning_rate": 4.501172861908366e-05, + "loss": 0.2521, + "step": 502 + }, + { + "epoch": 1.6993243243243243, + "grad_norm": 0.4216068685054779, + "learning_rate": 4.502613308426546e-05, + "loss": 0.2447, + "step": 503 + }, + { + "epoch": 1.7027027027027026, + "grad_norm": 0.4138765037059784, + "learning_rate": 4.504050894075876e-05, + "loss": 0.2379, + "step": 504 + }, + { + "epoch": 1.7060810810810811, + "grad_norm": 0.8677256107330322, + "learning_rate": 4.5054856301977696e-05, + "loss": 0.2399, + "step": 505 + }, + { + "epoch": 1.7094594594594594, + "grad_norm": 0.4616876542568207, + "learning_rate": 4.506917528066332e-05, + "loss": 0.229, + "step": 506 + }, + { + "epoch": 1.7128378378378377, + "grad_norm": 0.6452389359474182, + "learning_rate": 4.508346598888894e-05, + "loss": 0.2237, + "step": 507 + }, + { + "epoch": 1.7162162162162162, + "grad_norm": 0.4205520749092102, + "learning_rate": 4.509772853806532e-05, + "loss": 0.2175, + "step": 508 + }, + { + "epoch": 1.7195945945945947, + "grad_norm": 0.456882506608963, + "learning_rate": 4.511196303894598e-05, + "loss": 0.2444, + "step": 509 + }, + { + "epoch": 1.722972972972973, + "grad_norm": 0.43970227241516113, + "learning_rate": 4.512616960163227e-05, + "loss": 0.2357, + "step": 510 + }, + { + "epoch": 1.7263513513513513, + "grad_norm": 0.4025394320487976, + "learning_rate": 4.5140348335578547e-05, + "loss": 0.2375, + "step": 511 + }, + { + "epoch": 1.7297297297297298, + "grad_norm": 0.39442214369773865, + "learning_rate": 4.515449934959718e-05, + "loss": 0.2444, + "step": 512 + }, + { + "epoch": 1.7331081081081081, + "grad_norm": 0.44028565287590027, + "learning_rate": 4.516862275186361e-05, + "loss": 0.2324, + "step": 513 + }, + { + "epoch": 1.7364864864864864, + "grad_norm": 0.4466850161552429, + "learning_rate": 4.518271864992127e-05, + "loss": 0.2359, + "step": 514 + }, + { + "epoch": 1.739864864864865, + "grad_norm": 0.4365015923976898, + "learning_rate": 4.519678715068652e-05, + "loss": 0.2708, + "step": 515 + }, + { + "epoch": 1.7432432432432432, + "grad_norm": 0.4398203194141388, + "learning_rate": 4.521082836045353e-05, + "loss": 0.2452, + "step": 516 + }, + { + "epoch": 1.7466216216216215, + "grad_norm": 0.44056662917137146, + "learning_rate": 4.5224842384899045e-05, + "loss": 0.2344, + "step": 517 + }, + { + "epoch": 1.75, + "grad_norm": 0.43141037225723267, + "learning_rate": 4.523882932908722e-05, + "loss": 0.2511, + "step": 518 + }, + { + "epoch": 1.7533783783783785, + "grad_norm": 0.3922167122364044, + "learning_rate": 4.52527892974743e-05, + "loss": 0.2162, + "step": 519 + }, + { + "epoch": 1.7567567567567568, + "grad_norm": 0.40995466709136963, + "learning_rate": 4.526672239391333e-05, + "loss": 0.2281, + "step": 520 + }, + { + "epoch": 1.760135135135135, + "grad_norm": 0.4008403718471527, + "learning_rate": 4.528062872165875e-05, + "loss": 0.2257, + "step": 521 + }, + { + "epoch": 1.7635135135135136, + "grad_norm": 0.5048766136169434, + "learning_rate": 4.529450838337104e-05, + "loss": 0.2304, + "step": 522 + }, + { + "epoch": 1.7668918918918919, + "grad_norm": 0.40119826793670654, + "learning_rate": 4.530836148112124e-05, + "loss": 0.2341, + "step": 523 + }, + { + "epoch": 1.7702702702702702, + "grad_norm": 0.41344285011291504, + "learning_rate": 4.532218811639545e-05, + "loss": 0.272, + "step": 524 + }, + { + "epoch": 1.7736486486486487, + "grad_norm": 0.38902515172958374, + "learning_rate": 4.5335988390099284e-05, + "loss": 0.2195, + "step": 525 + }, + { + "epoch": 1.777027027027027, + "grad_norm": 0.5754374861717224, + "learning_rate": 4.534976240256232e-05, + "loss": 0.2335, + "step": 526 + }, + { + "epoch": 1.7804054054054053, + "grad_norm": 31.7464656829834, + "learning_rate": 4.536351025354245e-05, + "loss": 0.4532, + "step": 527 + }, + { + "epoch": 1.7837837837837838, + "grad_norm": 0.46500369906425476, + "learning_rate": 4.537723204223021e-05, + "loss": 0.2541, + "step": 528 + }, + { + "epoch": 1.7871621621621623, + "grad_norm": 0.41807517409324646, + "learning_rate": 4.53909278672531e-05, + "loss": 0.2287, + "step": 529 + }, + { + "epoch": 1.7905405405405406, + "grad_norm": 0.40613386034965515, + "learning_rate": 4.5404597826679824e-05, + "loss": 0.2196, + "step": 530 + }, + { + "epoch": 1.7939189189189189, + "grad_norm": 0.4200097322463989, + "learning_rate": 4.541824201802449e-05, + "loss": 0.2393, + "step": 531 + }, + { + "epoch": 1.7972972972972974, + "grad_norm": 0.48739388585090637, + "learning_rate": 4.543186053825081e-05, + "loss": 0.2469, + "step": 532 + }, + { + "epoch": 1.8006756756756757, + "grad_norm": 0.5199280381202698, + "learning_rate": 4.544545348377621e-05, + "loss": 0.2652, + "step": 533 + }, + { + "epoch": 1.804054054054054, + "grad_norm": 0.44481709599494934, + "learning_rate": 4.5459020950475946e-05, + "loss": 0.2386, + "step": 534 + }, + { + "epoch": 1.8074324324324325, + "grad_norm": 1.6978758573532104, + "learning_rate": 4.5472563033687145e-05, + "loss": 0.5287, + "step": 535 + }, + { + "epoch": 1.810810810810811, + "grad_norm": 0.3990303575992584, + "learning_rate": 4.548607982821284e-05, + "loss": 0.2102, + "step": 536 + }, + { + "epoch": 1.814189189189189, + "grad_norm": 0.4208585321903229, + "learning_rate": 4.5499571428325935e-05, + "loss": 0.2058, + "step": 537 + }, + { + "epoch": 1.8175675675675675, + "grad_norm": 0.41925048828125, + "learning_rate": 4.5513037927773155e-05, + "loss": 0.2349, + "step": 538 + }, + { + "epoch": 1.820945945945946, + "grad_norm": 0.43288394808769226, + "learning_rate": 4.5526479419778986e-05, + "loss": 0.2519, + "step": 539 + }, + { + "epoch": 1.8243243243243243, + "grad_norm": 0.4215989112854004, + "learning_rate": 4.553989599704948e-05, + "loss": 0.2518, + "step": 540 + }, + { + "epoch": 1.8277027027027026, + "grad_norm": 0.4134393334388733, + "learning_rate": 4.555328775177616e-05, + "loss": 0.2341, + "step": 541 + }, + { + "epoch": 1.8310810810810811, + "grad_norm": 0.4243868887424469, + "learning_rate": 4.5566654775639785e-05, + "loss": 0.2553, + "step": 542 + }, + { + "epoch": 1.8344594594594594, + "grad_norm": 0.40011635422706604, + "learning_rate": 4.5579997159814117e-05, + "loss": 0.2447, + "step": 543 + }, + { + "epoch": 1.8378378378378377, + "grad_norm": 0.38058069348335266, + "learning_rate": 4.5593314994969665e-05, + "loss": 0.2307, + "step": 544 + }, + { + "epoch": 1.8412162162162162, + "grad_norm": 0.47068503499031067, + "learning_rate": 4.560660837127738e-05, + "loss": 0.1971, + "step": 545 + }, + { + "epoch": 1.8445945945945947, + "grad_norm": 0.4493515193462372, + "learning_rate": 4.561987737841229e-05, + "loss": 0.2552, + "step": 546 + }, + { + "epoch": 1.847972972972973, + "grad_norm": 0.43196865916252136, + "learning_rate": 4.563312210555719e-05, + "loss": 0.2373, + "step": 547 + }, + { + "epoch": 1.8513513513513513, + "grad_norm": 0.4201470911502838, + "learning_rate": 4.564634264140616e-05, + "loss": 0.2304, + "step": 548 + }, + { + "epoch": 1.8547297297297298, + "grad_norm": 0.3680441975593567, + "learning_rate": 4.56595390741682e-05, + "loss": 0.2263, + "step": 549 + }, + { + "epoch": 1.8581081081081081, + "grad_norm": 0.3840002119541168, + "learning_rate": 4.567271149157073e-05, + "loss": 0.218, + "step": 550 + }, + { + "epoch": 1.8614864864864864, + "grad_norm": 0.39841052889823914, + "learning_rate": 4.5685859980863086e-05, + "loss": 0.2497, + "step": 551 + }, + { + "epoch": 1.864864864864865, + "grad_norm": 0.3956766128540039, + "learning_rate": 4.569898462881999e-05, + "loss": 0.2514, + "step": 552 + }, + { + "epoch": 1.8682432432432432, + "grad_norm": 0.4630524218082428, + "learning_rate": 4.571208552174497e-05, + "loss": 0.2744, + "step": 553 + }, + { + "epoch": 1.8716216216216215, + "grad_norm": 0.4056280851364136, + "learning_rate": 4.572516274547383e-05, + "loss": 0.2712, + "step": 554 + }, + { + "epoch": 1.875, + "grad_norm": 0.4127073585987091, + "learning_rate": 4.573821638537794e-05, + "loss": 0.2319, + "step": 555 + }, + { + "epoch": 1.8783783783783785, + "grad_norm": 0.38253894448280334, + "learning_rate": 4.575124652636763e-05, + "loss": 0.2302, + "step": 556 + }, + { + "epoch": 1.8817567567567568, + "grad_norm": 0.405472993850708, + "learning_rate": 4.5764253252895486e-05, + "loss": 0.2413, + "step": 557 + }, + { + "epoch": 1.885135135135135, + "grad_norm": 0.41931191086769104, + "learning_rate": 4.577723664895965e-05, + "loss": 0.2477, + "step": 558 + }, + { + "epoch": 1.8885135135135136, + "grad_norm": 0.3967287540435791, + "learning_rate": 4.579019679810706e-05, + "loss": 0.2525, + "step": 559 + }, + { + "epoch": 1.8918918918918919, + "grad_norm": 0.3867218792438507, + "learning_rate": 4.5803133783436676e-05, + "loss": 0.2226, + "step": 560 + }, + { + "epoch": 1.8952702702702702, + "grad_norm": 0.4048525094985962, + "learning_rate": 4.581604768760269e-05, + "loss": 0.2434, + "step": 561 + }, + { + "epoch": 1.8986486486486487, + "grad_norm": 0.38650572299957275, + "learning_rate": 4.582893859281769e-05, + "loss": 0.2287, + "step": 562 + }, + { + "epoch": 1.902027027027027, + "grad_norm": 0.3860650062561035, + "learning_rate": 4.584180658085578e-05, + "loss": 0.2285, + "step": 563 + }, + { + "epoch": 1.9054054054054053, + "grad_norm": 0.3902740478515625, + "learning_rate": 4.585465173305571e-05, + "loss": 0.2174, + "step": 564 + }, + { + "epoch": 1.9087837837837838, + "grad_norm": 0.4053342342376709, + "learning_rate": 4.5867474130323984e-05, + "loss": 0.2482, + "step": 565 + }, + { + "epoch": 1.9121621621621623, + "grad_norm": 0.35971924662590027, + "learning_rate": 4.588027385313786e-05, + "loss": 0.1858, + "step": 566 + }, + { + "epoch": 1.9155405405405406, + "grad_norm": 0.4442739486694336, + "learning_rate": 4.5893050981548446e-05, + "loss": 0.2847, + "step": 567 + }, + { + "epoch": 1.9189189189189189, + "grad_norm": 0.4430129826068878, + "learning_rate": 4.5905805595183656e-05, + "loss": 0.2265, + "step": 568 + }, + { + "epoch": 1.9222972972972974, + "grad_norm": 0.39476659893989563, + "learning_rate": 4.591853777325119e-05, + "loss": 0.258, + "step": 569 + }, + { + "epoch": 1.9256756756756757, + "grad_norm": 0.4573532044887543, + "learning_rate": 4.593124759454153e-05, + "loss": 0.2398, + "step": 570 + }, + { + "epoch": 1.929054054054054, + "grad_norm": 0.3871706426143646, + "learning_rate": 4.5943935137430806e-05, + "loss": 0.2292, + "step": 571 + }, + { + "epoch": 1.9324324324324325, + "grad_norm": 0.38287678360939026, + "learning_rate": 4.595660047988374e-05, + "loss": 0.2313, + "step": 572 + }, + { + "epoch": 1.935810810810811, + "grad_norm": 0.41778552532196045, + "learning_rate": 4.59692436994565e-05, + "loss": 0.2501, + "step": 573 + }, + { + "epoch": 1.939189189189189, + "grad_norm": 0.3646586537361145, + "learning_rate": 4.5981864873299563e-05, + "loss": 0.2099, + "step": 574 + }, + { + "epoch": 1.9425675675675675, + "grad_norm": 0.3954225778579712, + "learning_rate": 4.599446407816052e-05, + "loss": 0.2272, + "step": 575 + }, + { + "epoch": 1.945945945945946, + "grad_norm": 0.39867308735847473, + "learning_rate": 4.6007041390386874e-05, + "loss": 0.2352, + "step": 576 + }, + { + "epoch": 1.9493243243243243, + "grad_norm": 0.3616117238998413, + "learning_rate": 4.601959688592886e-05, + "loss": 0.2207, + "step": 577 + }, + { + "epoch": 1.9527027027027026, + "grad_norm": 0.42957910895347595, + "learning_rate": 4.603213064034216e-05, + "loss": 0.2437, + "step": 578 + }, + { + "epoch": 1.9560810810810811, + "grad_norm": 0.3899705708026886, + "learning_rate": 4.604464272879061e-05, + "loss": 0.2436, + "step": 579 + }, + { + "epoch": 1.9594594594594594, + "grad_norm": 0.3997238278388977, + "learning_rate": 4.605713322604896e-05, + "loss": 0.2065, + "step": 580 + }, + { + "epoch": 1.9628378378378377, + "grad_norm": 0.4001472592353821, + "learning_rate": 4.606960220650551e-05, + "loss": 0.2435, + "step": 581 + }, + { + "epoch": 1.9662162162162162, + "grad_norm": 0.4067547023296356, + "learning_rate": 4.608204974416481e-05, + "loss": 0.2456, + "step": 582 + }, + { + "epoch": 1.9695945945945947, + "grad_norm": 0.3919241428375244, + "learning_rate": 4.6094475912650234e-05, + "loss": 0.2214, + "step": 583 + }, + { + "epoch": 1.972972972972973, + "grad_norm": 0.3752463757991791, + "learning_rate": 4.610688078520666e-05, + "loss": 0.2148, + "step": 584 + }, + { + "epoch": 1.9763513513513513, + "grad_norm": 0.42536041140556335, + "learning_rate": 4.611926443470301e-05, + "loss": 0.2219, + "step": 585 + }, + { + "epoch": 1.9797297297297298, + "grad_norm": 0.3676905930042267, + "learning_rate": 4.6131626933634844e-05, + "loss": 0.2153, + "step": 586 + }, + { + "epoch": 1.9831081081081081, + "grad_norm": 0.3774767816066742, + "learning_rate": 4.6143968354126914e-05, + "loss": 0.2459, + "step": 587 + }, + { + "epoch": 1.9864864864864864, + "grad_norm": 0.38765907287597656, + "learning_rate": 4.6156288767935646e-05, + "loss": 0.2339, + "step": 588 + }, + { + "epoch": 1.989864864864865, + "grad_norm": 0.3474208116531372, + "learning_rate": 4.61685882464517e-05, + "loss": 0.2109, + "step": 589 + }, + { + "epoch": 1.9932432432432432, + "grad_norm": 0.39578577876091003, + "learning_rate": 4.61808668607024e-05, + "loss": 0.2232, + "step": 590 + }, + { + "epoch": 1.9966216216216215, + "grad_norm": 0.3756243884563446, + "learning_rate": 4.619312468135426e-05, + "loss": 0.2286, + "step": 591 + }, + { + "epoch": 2.0, + "grad_norm": 0.40366870164871216, + "learning_rate": 4.620536177871533e-05, + "loss": 0.1979, + "step": 592 + }, + { + "epoch": 2.0033783783783785, + "grad_norm": 0.5792023539543152, + "learning_rate": 4.621757822273772e-05, + "loss": 0.1664, + "step": 593 + }, + { + "epoch": 2.0067567567567566, + "grad_norm": 0.48093676567077637, + "learning_rate": 4.62297740830199e-05, + "loss": 0.1489, + "step": 594 + }, + { + "epoch": 2.010135135135135, + "grad_norm": 0.7113845348358154, + "learning_rate": 4.6241949428809165e-05, + "loss": 0.1666, + "step": 595 + }, + { + "epoch": 2.0135135135135136, + "grad_norm": 0.4888218343257904, + "learning_rate": 4.625410432900395e-05, + "loss": 0.1667, + "step": 596 + }, + { + "epoch": 2.016891891891892, + "grad_norm": 0.43272024393081665, + "learning_rate": 4.626623885215616e-05, + "loss": 0.1611, + "step": 597 + }, + { + "epoch": 2.02027027027027, + "grad_norm": 0.4639625549316406, + "learning_rate": 4.627835306647352e-05, + "loss": 0.183, + "step": 598 + }, + { + "epoch": 2.0236486486486487, + "grad_norm": 0.4807380735874176, + "learning_rate": 4.629044703982186e-05, + "loss": 0.1592, + "step": 599 + }, + { + "epoch": 2.027027027027027, + "grad_norm": 0.4514772891998291, + "learning_rate": 4.63025208397274e-05, + "loss": 0.1459, + "step": 600 + }, + { + "epoch": 2.0304054054054053, + "grad_norm": 0.4227406978607178, + "learning_rate": 4.6314574533379e-05, + "loss": 0.1632, + "step": 601 + }, + { + "epoch": 2.0337837837837838, + "grad_norm": 0.42751410603523254, + "learning_rate": 4.632660818763041e-05, + "loss": 0.1699, + "step": 602 + }, + { + "epoch": 2.0371621621621623, + "grad_norm": 0.45877569913864136, + "learning_rate": 4.633862186900253e-05, + "loss": 0.1725, + "step": 603 + }, + { + "epoch": 2.0405405405405403, + "grad_norm": 0.41468122601509094, + "learning_rate": 4.6350615643685535e-05, + "loss": 0.172, + "step": 604 + }, + { + "epoch": 2.043918918918919, + "grad_norm": 0.42624783515930176, + "learning_rate": 4.6362589577541154e-05, + "loss": 0.1591, + "step": 605 + }, + { + "epoch": 2.0472972972972974, + "grad_norm": 0.3854513466358185, + "learning_rate": 4.637454373610477e-05, + "loss": 0.1486, + "step": 606 + }, + { + "epoch": 2.050675675675676, + "grad_norm": 0.40797704458236694, + "learning_rate": 4.638647818458763e-05, + "loss": 0.1521, + "step": 607 + }, + { + "epoch": 2.054054054054054, + "grad_norm": 0.4326710104942322, + "learning_rate": 4.639839298787892e-05, + "loss": 0.1441, + "step": 608 + }, + { + "epoch": 2.0574324324324325, + "grad_norm": 0.39950937032699585, + "learning_rate": 4.641028821054793e-05, + "loss": 0.165, + "step": 609 + }, + { + "epoch": 2.060810810810811, + "grad_norm": 0.38399842381477356, + "learning_rate": 4.6422163916846124e-05, + "loss": 0.1464, + "step": 610 + }, + { + "epoch": 2.064189189189189, + "grad_norm": 0.40971097350120544, + "learning_rate": 4.643402017070924e-05, + "loss": 0.1604, + "step": 611 + }, + { + "epoch": 2.0675675675675675, + "grad_norm": 0.4519285261631012, + "learning_rate": 4.644585703575936e-05, + "loss": 0.1739, + "step": 612 + }, + { + "epoch": 2.070945945945946, + "grad_norm": 0.40968844294548035, + "learning_rate": 4.645767457530692e-05, + "loss": 0.1592, + "step": 613 + }, + { + "epoch": 2.074324324324324, + "grad_norm": 0.4271821975708008, + "learning_rate": 4.64694728523528e-05, + "loss": 0.1591, + "step": 614 + }, + { + "epoch": 2.0777027027027026, + "grad_norm": 0.44706177711486816, + "learning_rate": 4.648125192959028e-05, + "loss": 0.1745, + "step": 615 + }, + { + "epoch": 2.081081081081081, + "grad_norm": 0.3895583152770996, + "learning_rate": 4.649301186940709e-05, + "loss": 0.1541, + "step": 616 + }, + { + "epoch": 2.0844594594594597, + "grad_norm": 0.5112977623939514, + "learning_rate": 4.650475273388737e-05, + "loss": 0.1851, + "step": 617 + }, + { + "epoch": 2.0878378378378377, + "grad_norm": 0.4201546609401703, + "learning_rate": 4.651647458481359e-05, + "loss": 0.1502, + "step": 618 + }, + { + "epoch": 2.0912162162162162, + "grad_norm": 0.4547381103038788, + "learning_rate": 4.652817748366864e-05, + "loss": 0.1533, + "step": 619 + }, + { + "epoch": 2.0945945945945947, + "grad_norm": 0.48063594102859497, + "learning_rate": 4.653986149163757e-05, + "loss": 0.1704, + "step": 620 + }, + { + "epoch": 2.097972972972973, + "grad_norm": 0.3941943943500519, + "learning_rate": 4.655152666960967e-05, + "loss": 0.1338, + "step": 621 + }, + { + "epoch": 2.1013513513513513, + "grad_norm": 0.43712350726127625, + "learning_rate": 4.6563173078180315e-05, + "loss": 0.1573, + "step": 622 + }, + { + "epoch": 2.10472972972973, + "grad_norm": 0.43442657589912415, + "learning_rate": 4.657480077765283e-05, + "loss": 0.1672, + "step": 623 + }, + { + "epoch": 2.108108108108108, + "grad_norm": 0.45270854234695435, + "learning_rate": 4.6586409828040405e-05, + "loss": 0.1851, + "step": 624 + }, + { + "epoch": 2.1114864864864864, + "grad_norm": 0.5774021744728088, + "learning_rate": 4.659800028906792e-05, + "loss": 0.1451, + "step": 625 + }, + { + "epoch": 2.114864864864865, + "grad_norm": 0.4964922368526459, + "learning_rate": 4.660957222017383e-05, + "loss": 0.1656, + "step": 626 + }, + { + "epoch": 2.1182432432432434, + "grad_norm": 0.417864054441452, + "learning_rate": 4.662112568051194e-05, + "loss": 0.1672, + "step": 627 + }, + { + "epoch": 2.1216216216216215, + "grad_norm": 0.5328717231750488, + "learning_rate": 4.663266072895327e-05, + "loss": 0.1761, + "step": 628 + }, + { + "epoch": 2.125, + "grad_norm": 0.4234524369239807, + "learning_rate": 4.664417742408782e-05, + "loss": 0.1679, + "step": 629 + }, + { + "epoch": 2.1283783783783785, + "grad_norm": 0.5051260590553284, + "learning_rate": 4.665567582422637e-05, + "loss": 0.1763, + "step": 630 + }, + { + "epoch": 2.1317567567567566, + "grad_norm": 0.44952192902565, + "learning_rate": 4.666715598740224e-05, + "loss": 0.1535, + "step": 631 + }, + { + "epoch": 2.135135135135135, + "grad_norm": 0.44377925992012024, + "learning_rate": 4.667861797137309e-05, + "loss": 0.1456, + "step": 632 + }, + { + "epoch": 2.1385135135135136, + "grad_norm": 0.3786265254020691, + "learning_rate": 4.669006183362258e-05, + "loss": 0.1478, + "step": 633 + }, + { + "epoch": 2.141891891891892, + "grad_norm": 0.4277481436729431, + "learning_rate": 4.670148763136221e-05, + "loss": 0.157, + "step": 634 + }, + { + "epoch": 2.14527027027027, + "grad_norm": 0.43035444617271423, + "learning_rate": 4.671289542153293e-05, + "loss": 0.1552, + "step": 635 + }, + { + "epoch": 2.1486486486486487, + "grad_norm": 0.45837166905403137, + "learning_rate": 4.672428526080691e-05, + "loss": 0.1848, + "step": 636 + }, + { + "epoch": 2.152027027027027, + "grad_norm": 0.41413986682891846, + "learning_rate": 4.673565720558918e-05, + "loss": 0.1803, + "step": 637 + }, + { + "epoch": 2.1554054054054053, + "grad_norm": 0.4246644377708435, + "learning_rate": 4.6747011312019374e-05, + "loss": 0.177, + "step": 638 + }, + { + "epoch": 2.1587837837837838, + "grad_norm": 0.4299694001674652, + "learning_rate": 4.6758347635973334e-05, + "loss": 0.164, + "step": 639 + }, + { + "epoch": 2.1621621621621623, + "grad_norm": 0.4426731765270233, + "learning_rate": 4.676966623306479e-05, + "loss": 0.164, + "step": 640 + }, + { + "epoch": 2.1655405405405403, + "grad_norm": 0.4473170340061188, + "learning_rate": 4.678096715864696e-05, + "loss": 0.1642, + "step": 641 + }, + { + "epoch": 2.168918918918919, + "grad_norm": 0.44445449113845825, + "learning_rate": 4.679225046781422e-05, + "loss": 0.155, + "step": 642 + }, + { + "epoch": 2.1722972972972974, + "grad_norm": 0.42584556341171265, + "learning_rate": 4.68035162154037e-05, + "loss": 0.171, + "step": 643 + }, + { + "epoch": 2.175675675675676, + "grad_norm": 0.39437368512153625, + "learning_rate": 4.681476445599687e-05, + "loss": 0.1692, + "step": 644 + }, + { + "epoch": 2.179054054054054, + "grad_norm": 0.40476444363594055, + "learning_rate": 4.6825995243921137e-05, + "loss": 0.162, + "step": 645 + }, + { + "epoch": 2.1824324324324325, + "grad_norm": 0.42337456345558167, + "learning_rate": 4.683720863325141e-05, + "loss": 0.184, + "step": 646 + }, + { + "epoch": 2.185810810810811, + "grad_norm": 0.4256432056427002, + "learning_rate": 4.684840467781168e-05, + "loss": 0.175, + "step": 647 + }, + { + "epoch": 2.189189189189189, + "grad_norm": 0.40138664841651917, + "learning_rate": 4.685958343117656e-05, + "loss": 0.1577, + "step": 648 + }, + { + "epoch": 2.1925675675675675, + "grad_norm": 0.4015848636627197, + "learning_rate": 4.6870744946672826e-05, + "loss": 0.1521, + "step": 649 + }, + { + "epoch": 2.195945945945946, + "grad_norm": 0.4208281934261322, + "learning_rate": 4.688188927738093e-05, + "loss": 0.1685, + "step": 650 + }, + { + "epoch": 2.199324324324324, + "grad_norm": 0.4102751910686493, + "learning_rate": 4.689301647613653e-05, + "loss": 0.1837, + "step": 651 + }, + { + "epoch": 2.2027027027027026, + "grad_norm": 0.43685656785964966, + "learning_rate": 4.6904126595532014e-05, + "loss": 0.174, + "step": 652 + }, + { + "epoch": 2.206081081081081, + "grad_norm": 0.392456978559494, + "learning_rate": 4.69152196879179e-05, + "loss": 0.1741, + "step": 653 + }, + { + "epoch": 2.2094594594594597, + "grad_norm": 0.4086831212043762, + "learning_rate": 4.692629580540446e-05, + "loss": 0.1559, + "step": 654 + }, + { + "epoch": 2.2128378378378377, + "grad_norm": 0.38833364844322205, + "learning_rate": 4.693735499986305e-05, + "loss": 0.1618, + "step": 655 + }, + { + "epoch": 2.2162162162162162, + "grad_norm": 0.36838966608047485, + "learning_rate": 4.694839732292767e-05, + "loss": 0.1478, + "step": 656 + }, + { + "epoch": 2.2195945945945947, + "grad_norm": 0.40487056970596313, + "learning_rate": 4.6959422825996345e-05, + "loss": 0.1738, + "step": 657 + }, + { + "epoch": 2.222972972972973, + "grad_norm": 0.3815852999687195, + "learning_rate": 4.69704315602326e-05, + "loss": 0.1665, + "step": 658 + }, + { + "epoch": 2.2263513513513513, + "grad_norm": 0.400643914937973, + "learning_rate": 4.698142357656684e-05, + "loss": 0.1771, + "step": 659 + }, + { + "epoch": 2.22972972972973, + "grad_norm": 0.3691990077495575, + "learning_rate": 4.6992398925697814e-05, + "loss": 0.1591, + "step": 660 + }, + { + "epoch": 2.233108108108108, + "grad_norm": 0.4398171305656433, + "learning_rate": 4.7003357658094e-05, + "loss": 0.2062, + "step": 661 + }, + { + "epoch": 2.2364864864864864, + "grad_norm": 0.3774857521057129, + "learning_rate": 4.7014299823995005e-05, + "loss": 0.1502, + "step": 662 + }, + { + "epoch": 2.239864864864865, + "grad_norm": 0.4315558075904846, + "learning_rate": 4.702522547341289e-05, + "loss": 0.1636, + "step": 663 + }, + { + "epoch": 2.2432432432432434, + "grad_norm": 0.4040674865245819, + "learning_rate": 4.703613465613363e-05, + "loss": 0.1625, + "step": 664 + }, + { + "epoch": 2.2466216216216215, + "grad_norm": 0.46972906589508057, + "learning_rate": 4.704702742171841e-05, + "loss": 0.1833, + "step": 665 + }, + { + "epoch": 2.25, + "grad_norm": 0.41066527366638184, + "learning_rate": 4.7057903819505024e-05, + "loss": 0.1707, + "step": 666 + }, + { + "epoch": 2.2533783783783785, + "grad_norm": 0.3562461733818054, + "learning_rate": 4.7068763898609154e-05, + "loss": 0.1508, + "step": 667 + }, + { + "epoch": 2.2567567567567566, + "grad_norm": 0.4074662923812866, + "learning_rate": 4.707960770792576e-05, + "loss": 0.1786, + "step": 668 + }, + { + "epoch": 2.260135135135135, + "grad_norm": 0.3978296220302582, + "learning_rate": 4.709043529613039e-05, + "loss": 0.1797, + "step": 669 + }, + { + "epoch": 2.2635135135135136, + "grad_norm": 0.42612069845199585, + "learning_rate": 4.710124671168044e-05, + "loss": 0.1874, + "step": 670 + }, + { + "epoch": 2.266891891891892, + "grad_norm": 0.4137380123138428, + "learning_rate": 4.711204200281654e-05, + "loss": 0.1897, + "step": 671 + }, + { + "epoch": 2.27027027027027, + "grad_norm": 0.37814861536026, + "learning_rate": 4.712282121756376e-05, + "loss": 0.1617, + "step": 672 + }, + { + "epoch": 2.2736486486486487, + "grad_norm": 0.42515134811401367, + "learning_rate": 4.713358440373295e-05, + "loss": 0.1652, + "step": 673 + }, + { + "epoch": 2.277027027027027, + "grad_norm": 0.42217129468917847, + "learning_rate": 4.7144331608922e-05, + "loss": 0.1872, + "step": 674 + }, + { + "epoch": 2.2804054054054053, + "grad_norm": 0.3996788263320923, + "learning_rate": 4.715506288051709e-05, + "loss": 0.1603, + "step": 675 + }, + { + "epoch": 2.2837837837837838, + "grad_norm": 0.4035404622554779, + "learning_rate": 4.7165778265693935e-05, + "loss": 0.1689, + "step": 676 + }, + { + "epoch": 2.2871621621621623, + "grad_norm": 0.39791160821914673, + "learning_rate": 4.7176477811419076e-05, + "loss": 0.1804, + "step": 677 + }, + { + "epoch": 2.2905405405405403, + "grad_norm": 0.4175347685813904, + "learning_rate": 4.718716156445106e-05, + "loss": 0.1811, + "step": 678 + }, + { + "epoch": 2.293918918918919, + "grad_norm": 0.39992383122444153, + "learning_rate": 4.7197829571341704e-05, + "loss": 0.1805, + "step": 679 + }, + { + "epoch": 2.2972972972972974, + "grad_norm": 0.40433621406555176, + "learning_rate": 4.720848187843727e-05, + "loss": 0.1726, + "step": 680 + }, + { + "epoch": 2.3006756756756754, + "grad_norm": 0.38958102464675903, + "learning_rate": 4.721911853187975e-05, + "loss": 0.1663, + "step": 681 + }, + { + "epoch": 2.304054054054054, + "grad_norm": 0.40906044840812683, + "learning_rate": 4.722973957760799e-05, + "loss": 0.1637, + "step": 682 + }, + { + "epoch": 2.3074324324324325, + "grad_norm": 0.4296117424964905, + "learning_rate": 4.724034506135888e-05, + "loss": 0.178, + "step": 683 + }, + { + "epoch": 2.310810810810811, + "grad_norm": 0.40840208530426025, + "learning_rate": 4.725093502866861e-05, + "loss": 0.1725, + "step": 684 + }, + { + "epoch": 2.314189189189189, + "grad_norm": 0.3669198751449585, + "learning_rate": 4.7261509524873764e-05, + "loss": 0.161, + "step": 685 + }, + { + "epoch": 2.3175675675675675, + "grad_norm": 0.40982183814048767, + "learning_rate": 4.727206859511253e-05, + "loss": 0.1757, + "step": 686 + }, + { + "epoch": 2.320945945945946, + "grad_norm": 0.4490487575531006, + "learning_rate": 4.7282612284325846e-05, + "loss": 0.1975, + "step": 687 + }, + { + "epoch": 2.3243243243243246, + "grad_norm": 0.4303133487701416, + "learning_rate": 4.729314063725853e-05, + "loss": 0.1753, + "step": 688 + }, + { + "epoch": 2.3277027027027026, + "grad_norm": 0.5020397901535034, + "learning_rate": 4.730365369846044e-05, + "loss": 0.1959, + "step": 689 + }, + { + "epoch": 2.331081081081081, + "grad_norm": 0.4409542679786682, + "learning_rate": 4.7314151512287594e-05, + "loss": 0.1779, + "step": 690 + }, + { + "epoch": 2.3344594594594597, + "grad_norm": 0.41255807876586914, + "learning_rate": 4.732463412290331e-05, + "loss": 0.1647, + "step": 691 + }, + { + "epoch": 2.3378378378378377, + "grad_norm": 0.41047075390815735, + "learning_rate": 4.73351015742793e-05, + "loss": 0.1699, + "step": 692 + }, + { + "epoch": 2.3412162162162162, + "grad_norm": 0.4248691201210022, + "learning_rate": 4.7345553910196785e-05, + "loss": 0.1785, + "step": 693 + }, + { + "epoch": 2.3445945945945947, + "grad_norm": 0.41576990485191345, + "learning_rate": 4.735599117424759e-05, + "loss": 0.1806, + "step": 694 + }, + { + "epoch": 2.347972972972973, + "grad_norm": 0.399854838848114, + "learning_rate": 4.7366413409835235e-05, + "loss": 0.1689, + "step": 695 + }, + { + "epoch": 2.3513513513513513, + "grad_norm": 0.39361926913261414, + "learning_rate": 4.737682066017604e-05, + "loss": 0.1757, + "step": 696 + }, + { + "epoch": 2.35472972972973, + "grad_norm": 0.36152175068855286, + "learning_rate": 4.7387212968300166e-05, + "loss": 0.1421, + "step": 697 + }, + { + "epoch": 2.358108108108108, + "grad_norm": 0.4112738370895386, + "learning_rate": 4.7397590377052686e-05, + "loss": 0.156, + "step": 698 + }, + { + "epoch": 2.3614864864864864, + "grad_norm": 1.494449496269226, + "learning_rate": 4.74079529290947e-05, + "loss": 0.433, + "step": 699 + }, + { + "epoch": 2.364864864864865, + "grad_norm": 0.44618678092956543, + "learning_rate": 4.741830066690428e-05, + "loss": 0.1738, + "step": 700 + }, + { + "epoch": 2.3682432432432434, + "grad_norm": 0.4271227717399597, + "learning_rate": 4.742863363277765e-05, + "loss": 0.1859, + "step": 701 + }, + { + "epoch": 2.3716216216216215, + "grad_norm": 0.38577789068222046, + "learning_rate": 4.743895186883009e-05, + "loss": 0.1699, + "step": 702 + }, + { + "epoch": 2.375, + "grad_norm": 0.45399177074432373, + "learning_rate": 4.7449255416997075e-05, + "loss": 0.2042, + "step": 703 + }, + { + "epoch": 2.3783783783783785, + "grad_norm": 0.4084223806858063, + "learning_rate": 4.7459544319035206e-05, + "loss": 0.1626, + "step": 704 + }, + { + "epoch": 2.3817567567567566, + "grad_norm": 0.4444795846939087, + "learning_rate": 4.746981861652332e-05, + "loss": 0.1524, + "step": 705 + }, + { + "epoch": 2.385135135135135, + "grad_norm": 0.41611728072166443, + "learning_rate": 4.74800783508634e-05, + "loss": 0.1965, + "step": 706 + }, + { + "epoch": 2.3885135135135136, + "grad_norm": 0.4463081657886505, + "learning_rate": 4.7490323563281665e-05, + "loss": 0.1703, + "step": 707 + }, + { + "epoch": 2.391891891891892, + "grad_norm": 0.4274112582206726, + "learning_rate": 4.750055429482949e-05, + "loss": 0.1468, + "step": 708 + }, + { + "epoch": 2.39527027027027, + "grad_norm": 0.5068261623382568, + "learning_rate": 4.751077058638445e-05, + "loss": 0.1959, + "step": 709 + }, + { + "epoch": 2.3986486486486487, + "grad_norm": 0.46266356110572815, + "learning_rate": 4.752097247865126e-05, + "loss": 0.1876, + "step": 710 + }, + { + "epoch": 2.402027027027027, + "grad_norm": 0.4017108082771301, + "learning_rate": 4.753116001216277e-05, + "loss": 0.1567, + "step": 711 + }, + { + "epoch": 2.4054054054054053, + "grad_norm": 0.4414288103580475, + "learning_rate": 4.7541333227280944e-05, + "loss": 0.171, + "step": 712 + }, + { + "epoch": 2.4087837837837838, + "grad_norm": 0.47343650460243225, + "learning_rate": 4.755149216419776e-05, + "loss": 0.1973, + "step": 713 + }, + { + "epoch": 2.4121621621621623, + "grad_norm": 0.4446316361427307, + "learning_rate": 4.756163686293624e-05, + "loss": 0.1804, + "step": 714 + }, + { + "epoch": 2.4155405405405403, + "grad_norm": 0.4418063163757324, + "learning_rate": 4.7571767363351344e-05, + "loss": 0.1913, + "step": 715 + }, + { + "epoch": 2.418918918918919, + "grad_norm": 0.4041652977466583, + "learning_rate": 4.758188370513093e-05, + "loss": 0.1852, + "step": 716 + }, + { + "epoch": 2.4222972972972974, + "grad_norm": 0.41937291622161865, + "learning_rate": 4.759198592779667e-05, + "loss": 0.1925, + "step": 717 + }, + { + "epoch": 2.4256756756756754, + "grad_norm": 0.4121197462081909, + "learning_rate": 4.760207407070501e-05, + "loss": 0.1821, + "step": 718 + }, + { + "epoch": 2.429054054054054, + "grad_norm": 0.3996151387691498, + "learning_rate": 4.761214817304805e-05, + "loss": 0.1837, + "step": 719 + }, + { + "epoch": 2.4324324324324325, + "grad_norm": 0.38412001729011536, + "learning_rate": 4.762220827385448e-05, + "loss": 0.1685, + "step": 720 + }, + { + "epoch": 2.435810810810811, + "grad_norm": 1.5490686893463135, + "learning_rate": 4.763225441199049e-05, + "loss": 0.432, + "step": 721 + }, + { + "epoch": 2.439189189189189, + "grad_norm": 0.4263227880001068, + "learning_rate": 4.7642286626160654e-05, + "loss": 0.1847, + "step": 722 + }, + { + "epoch": 2.4425675675675675, + "grad_norm": 0.3754761815071106, + "learning_rate": 4.765230495490885e-05, + "loss": 0.1786, + "step": 723 + }, + { + "epoch": 2.445945945945946, + "grad_norm": 0.43871861696243286, + "learning_rate": 4.7662309436619115e-05, + "loss": 0.189, + "step": 724 + }, + { + "epoch": 2.4493243243243246, + "grad_norm": 0.4133760333061218, + "learning_rate": 4.7672300109516563e-05, + "loss": 0.1794, + "step": 725 + }, + { + "epoch": 2.4527027027027026, + "grad_norm": 0.38480374217033386, + "learning_rate": 4.768227701166823e-05, + "loss": 0.1666, + "step": 726 + }, + { + "epoch": 2.456081081081081, + "grad_norm": 0.4138960540294647, + "learning_rate": 4.7692240180983964e-05, + "loss": 0.1695, + "step": 727 + }, + { + "epoch": 2.4594594594594597, + "grad_norm": 0.3594178557395935, + "learning_rate": 4.770218965521729e-05, + "loss": 0.1519, + "step": 728 + }, + { + "epoch": 2.4628378378378377, + "grad_norm": 0.4005281627178192, + "learning_rate": 4.7712125471966245e-05, + "loss": 0.1696, + "step": 729 + }, + { + "epoch": 2.4662162162162162, + "grad_norm": 0.3780256509780884, + "learning_rate": 4.7722047668674267e-05, + "loss": 0.174, + "step": 730 + }, + { + "epoch": 2.4695945945945947, + "grad_norm": 0.4074977934360504, + "learning_rate": 4.7731956282631004e-05, + "loss": 0.1696, + "step": 731 + }, + { + "epoch": 2.472972972972973, + "grad_norm": 0.4063095152378082, + "learning_rate": 4.77418513509732e-05, + "loss": 0.1731, + "step": 732 + }, + { + "epoch": 2.4763513513513513, + "grad_norm": 0.40110018849372864, + "learning_rate": 4.775173291068547e-05, + "loss": 0.1787, + "step": 733 + }, + { + "epoch": 2.47972972972973, + "grad_norm": 0.3782990574836731, + "learning_rate": 4.776160099860117e-05, + "loss": 0.1753, + "step": 734 + }, + { + "epoch": 2.483108108108108, + "grad_norm": 0.3815288543701172, + "learning_rate": 4.777145565140325e-05, + "loss": 0.1766, + "step": 735 + }, + { + "epoch": 2.4864864864864864, + "grad_norm": 0.3779781758785248, + "learning_rate": 4.7781296905624986e-05, + "loss": 0.1857, + "step": 736 + }, + { + "epoch": 2.489864864864865, + "grad_norm": 0.33802899718284607, + "learning_rate": 4.779112479765086e-05, + "loss": 0.1515, + "step": 737 + }, + { + "epoch": 2.4932432432432434, + "grad_norm": 0.4050140082836151, + "learning_rate": 4.780093936371736e-05, + "loss": 0.1908, + "step": 738 + }, + { + "epoch": 2.4966216216216215, + "grad_norm": 0.4146076738834381, + "learning_rate": 4.781074063991376e-05, + "loss": 0.1812, + "step": 739 + }, + { + "epoch": 2.5, + "grad_norm": 0.34979405999183655, + "learning_rate": 4.782052866218294e-05, + "loss": 0.1505, + "step": 740 + }, + { + "epoch": 2.5033783783783785, + "grad_norm": 0.3686615526676178, + "learning_rate": 4.783030346632214e-05, + "loss": 0.1656, + "step": 741 + }, + { + "epoch": 2.506756756756757, + "grad_norm": 0.38285037875175476, + "learning_rate": 4.7840065087983786e-05, + "loss": 0.181, + "step": 742 + }, + { + "epoch": 2.510135135135135, + "grad_norm": 0.39826491475105286, + "learning_rate": 4.784981356267626e-05, + "loss": 0.201, + "step": 743 + }, + { + "epoch": 2.5135135135135136, + "grad_norm": 0.40031111240386963, + "learning_rate": 4.785954892576465e-05, + "loss": 0.1676, + "step": 744 + }, + { + "epoch": 2.516891891891892, + "grad_norm": 0.412266343832016, + "learning_rate": 4.7869271212471554e-05, + "loss": 0.1807, + "step": 745 + }, + { + "epoch": 2.52027027027027, + "grad_norm": 0.38458314538002014, + "learning_rate": 4.7878980457877814e-05, + "loss": 0.1639, + "step": 746 + }, + { + "epoch": 2.5236486486486487, + "grad_norm": 0.45593783259391785, + "learning_rate": 4.7888676696923315e-05, + "loss": 0.1844, + "step": 747 + }, + { + "epoch": 2.527027027027027, + "grad_norm": 0.36613929271698, + "learning_rate": 4.7898359964407695e-05, + "loss": 0.1536, + "step": 748 + }, + { + "epoch": 2.5304054054054053, + "grad_norm": 0.4267789125442505, + "learning_rate": 4.790803029499111e-05, + "loss": 0.1706, + "step": 749 + }, + { + "epoch": 2.5337837837837838, + "grad_norm": 1.5388500690460205, + "learning_rate": 4.7917687723195004e-05, + "loss": 0.4571, + "step": 750 + }, + { + "epoch": 2.5371621621621623, + "grad_norm": 0.42774277925491333, + "learning_rate": 4.792733228340281e-05, + "loss": 0.1907, + "step": 751 + }, + { + "epoch": 2.5405405405405403, + "grad_norm": 0.38267603516578674, + "learning_rate": 4.793696400986071e-05, + "loss": 0.163, + "step": 752 + }, + { + "epoch": 2.543918918918919, + "grad_norm": 0.4370862543582916, + "learning_rate": 4.7946582936678344e-05, + "loss": 0.2019, + "step": 753 + }, + { + "epoch": 2.5472972972972974, + "grad_norm": 0.3963480293750763, + "learning_rate": 4.795618909782957e-05, + "loss": 0.158, + "step": 754 + }, + { + "epoch": 2.5506756756756754, + "grad_norm": 0.4094291031360626, + "learning_rate": 4.796578252715314e-05, + "loss": 0.1769, + "step": 755 + }, + { + "epoch": 2.554054054054054, + "grad_norm": 0.41850391030311584, + "learning_rate": 4.797536325835345e-05, + "loss": 0.17, + "step": 756 + }, + { + "epoch": 2.5574324324324325, + "grad_norm": 0.4879460036754608, + "learning_rate": 4.7984931325001216e-05, + "loss": 0.2163, + "step": 757 + }, + { + "epoch": 2.560810810810811, + "grad_norm": 0.4456654489040375, + "learning_rate": 4.799448676053423e-05, + "loss": 0.2086, + "step": 758 + }, + { + "epoch": 2.564189189189189, + "grad_norm": 0.41925883293151855, + "learning_rate": 4.800402959825802e-05, + "loss": 0.1884, + "step": 759 + }, + { + "epoch": 2.5675675675675675, + "grad_norm": 0.4378669857978821, + "learning_rate": 4.801355987134653e-05, + "loss": 0.2024, + "step": 760 + }, + { + "epoch": 2.570945945945946, + "grad_norm": 0.5047132968902588, + "learning_rate": 4.802307761284289e-05, + "loss": 0.1961, + "step": 761 + }, + { + "epoch": 2.5743243243243246, + "grad_norm": 0.41319629549980164, + "learning_rate": 4.8032582855660014e-05, + "loss": 0.1957, + "step": 762 + }, + { + "epoch": 2.5777027027027026, + "grad_norm": 0.46079912781715393, + "learning_rate": 4.8042075632581346e-05, + "loss": 0.1931, + "step": 763 + }, + { + "epoch": 2.581081081081081, + "grad_norm": 0.431325763463974, + "learning_rate": 4.80515559762615e-05, + "loss": 0.1788, + "step": 764 + }, + { + "epoch": 2.5844594594594597, + "grad_norm": 0.40309104323387146, + "learning_rate": 4.8061023919226964e-05, + "loss": 0.1817, + "step": 765 + }, + { + "epoch": 2.5878378378378377, + "grad_norm": 0.434621661901474, + "learning_rate": 4.807047949387674e-05, + "loss": 0.1831, + "step": 766 + }, + { + "epoch": 2.5912162162162162, + "grad_norm": 0.35631585121154785, + "learning_rate": 4.807992273248302e-05, + "loss": 0.1683, + "step": 767 + }, + { + "epoch": 2.5945945945945947, + "grad_norm": 0.41151365637779236, + "learning_rate": 4.808935366719187e-05, + "loss": 0.1904, + "step": 768 + }, + { + "epoch": 2.597972972972973, + "grad_norm": 0.4089500904083252, + "learning_rate": 4.8098772330023855e-05, + "loss": 0.1742, + "step": 769 + }, + { + "epoch": 2.6013513513513513, + "grad_norm": 0.38123077154159546, + "learning_rate": 4.81081787528747e-05, + "loss": 0.1727, + "step": 770 + }, + { + "epoch": 2.60472972972973, + "grad_norm": 0.37730783224105835, + "learning_rate": 4.811757296751595e-05, + "loss": 0.1904, + "step": 771 + }, + { + "epoch": 2.608108108108108, + "grad_norm": 0.38781440258026123, + "learning_rate": 4.812695500559561e-05, + "loss": 0.1931, + "step": 772 + }, + { + "epoch": 2.6114864864864864, + "grad_norm": 0.4114435315132141, + "learning_rate": 4.8136324898638756e-05, + "loss": 0.2105, + "step": 773 + }, + { + "epoch": 2.614864864864865, + "grad_norm": 0.39372387528419495, + "learning_rate": 4.8145682678048214e-05, + "loss": 0.1831, + "step": 774 + }, + { + "epoch": 2.618243243243243, + "grad_norm": 0.40174078941345215, + "learning_rate": 4.815502837510518e-05, + "loss": 0.1975, + "step": 775 + }, + { + "epoch": 2.6216216216216215, + "grad_norm": 0.4061259627342224, + "learning_rate": 4.816436202096981e-05, + "loss": 0.1744, + "step": 776 + }, + { + "epoch": 2.625, + "grad_norm": 0.3997017741203308, + "learning_rate": 4.81736836466819e-05, + "loss": 0.1782, + "step": 777 + }, + { + "epoch": 2.6283783783783785, + "grad_norm": 0.4183482229709625, + "learning_rate": 4.8182993283161485e-05, + "loss": 0.1941, + "step": 778 + }, + { + "epoch": 2.631756756756757, + "grad_norm": 0.3742780387401581, + "learning_rate": 4.819229096120941e-05, + "loss": 0.1681, + "step": 779 + }, + { + "epoch": 2.635135135135135, + "grad_norm": 0.3982739746570587, + "learning_rate": 4.820157671150801e-05, + "loss": 0.1941, + "step": 780 + }, + { + "epoch": 2.6385135135135136, + "grad_norm": 0.3757505714893341, + "learning_rate": 4.821085056462168e-05, + "loss": 0.1744, + "step": 781 + }, + { + "epoch": 2.641891891891892, + "grad_norm": 0.39277997612953186, + "learning_rate": 4.822011255099747e-05, + "loss": 0.1803, + "step": 782 + }, + { + "epoch": 2.64527027027027, + "grad_norm": 0.35470277070999146, + "learning_rate": 4.8229362700965726e-05, + "loss": 0.1651, + "step": 783 + }, + { + "epoch": 2.6486486486486487, + "grad_norm": 0.381610631942749, + "learning_rate": 4.8238601044740645e-05, + "loss": 0.183, + "step": 784 + }, + { + "epoch": 2.652027027027027, + "grad_norm": 0.37985488772392273, + "learning_rate": 4.824782761242088e-05, + "loss": 0.1637, + "step": 785 + }, + { + "epoch": 2.6554054054054053, + "grad_norm": 0.4003821909427643, + "learning_rate": 4.8257042433990135e-05, + "loss": 0.18, + "step": 786 + }, + { + "epoch": 2.6587837837837838, + "grad_norm": 0.37772974371910095, + "learning_rate": 4.826624553931775e-05, + "loss": 0.1886, + "step": 787 + }, + { + "epoch": 2.6621621621621623, + "grad_norm": 0.3801021873950958, + "learning_rate": 4.827543695815926e-05, + "loss": 0.1718, + "step": 788 + }, + { + "epoch": 2.6655405405405403, + "grad_norm": 0.3759152591228485, + "learning_rate": 4.8284616720157006e-05, + "loss": 0.1698, + "step": 789 + }, + { + "epoch": 2.668918918918919, + "grad_norm": 0.38654980063438416, + "learning_rate": 4.82937848548407e-05, + "loss": 0.1885, + "step": 790 + }, + { + "epoch": 2.6722972972972974, + "grad_norm": 0.4393473267555237, + "learning_rate": 4.8302941391627947e-05, + "loss": 0.1891, + "step": 791 + }, + { + "epoch": 2.6756756756756754, + "grad_norm": 0.3831874132156372, + "learning_rate": 4.83120863598249e-05, + "loss": 0.1888, + "step": 792 + }, + { + "epoch": 2.679054054054054, + "grad_norm": 0.38641834259033203, + "learning_rate": 4.832121978862673e-05, + "loss": 0.1842, + "step": 793 + }, + { + "epoch": 2.6824324324324325, + "grad_norm": 0.39984583854675293, + "learning_rate": 4.8330341707118276e-05, + "loss": 0.1681, + "step": 794 + }, + { + "epoch": 2.685810810810811, + "grad_norm": 0.40252962708473206, + "learning_rate": 4.833945214427451e-05, + "loss": 0.1756, + "step": 795 + }, + { + "epoch": 2.689189189189189, + "grad_norm": 0.40013495087623596, + "learning_rate": 4.834855112896116e-05, + "loss": 0.2146, + "step": 796 + }, + { + "epoch": 2.6925675675675675, + "grad_norm": 0.385640412569046, + "learning_rate": 4.835763868993521e-05, + "loss": 0.1851, + "step": 797 + }, + { + "epoch": 2.695945945945946, + "grad_norm": 0.3977515697479248, + "learning_rate": 4.8366714855845496e-05, + "loss": 0.1809, + "step": 798 + }, + { + "epoch": 2.6993243243243246, + "grad_norm": 0.40790390968322754, + "learning_rate": 4.837577965523319e-05, + "loss": 0.1887, + "step": 799 + }, + { + "epoch": 2.7027027027027026, + "grad_norm": 0.3771408498287201, + "learning_rate": 4.8384833116532396e-05, + "loss": 0.1732, + "step": 800 + }, + { + "epoch": 2.706081081081081, + "grad_norm": 0.42990434169769287, + "learning_rate": 4.8393875268070636e-05, + "loss": 0.1952, + "step": 801 + }, + { + "epoch": 2.7094594594594597, + "grad_norm": 0.38725292682647705, + "learning_rate": 4.84029061380694e-05, + "loss": 0.1903, + "step": 802 + }, + { + "epoch": 2.7128378378378377, + "grad_norm": 0.4196310341358185, + "learning_rate": 4.841192575464469e-05, + "loss": 0.2035, + "step": 803 + }, + { + "epoch": 2.7162162162162162, + "grad_norm": 0.4179129898548126, + "learning_rate": 4.842093414580753e-05, + "loss": 0.2002, + "step": 804 + }, + { + "epoch": 2.7195945945945947, + "grad_norm": 0.44000861048698425, + "learning_rate": 4.842993133946448e-05, + "loss": 0.2299, + "step": 805 + }, + { + "epoch": 2.722972972972973, + "grad_norm": 0.4462367296218872, + "learning_rate": 4.843891736341818e-05, + "loss": 0.1931, + "step": 806 + }, + { + "epoch": 2.7263513513513513, + "grad_norm": 0.42056065797805786, + "learning_rate": 4.8447892245367846e-05, + "loss": 0.2012, + "step": 807 + }, + { + "epoch": 2.72972972972973, + "grad_norm": 0.48042038083076477, + "learning_rate": 4.845685601290977e-05, + "loss": 0.1988, + "step": 808 + }, + { + "epoch": 2.733108108108108, + "grad_norm": 0.39594507217407227, + "learning_rate": 4.846580869353787e-05, + "loss": 0.1966, + "step": 809 + }, + { + "epoch": 2.7364864864864864, + "grad_norm": 0.4334581792354584, + "learning_rate": 4.847475031464416e-05, + "loss": 0.1861, + "step": 810 + }, + { + "epoch": 2.739864864864865, + "grad_norm": 0.40071550011634827, + "learning_rate": 4.8483680903519274e-05, + "loss": 0.2009, + "step": 811 + }, + { + "epoch": 2.743243243243243, + "grad_norm": 0.40345826745033264, + "learning_rate": 4.8492600487352926e-05, + "loss": 0.1692, + "step": 812 + }, + { + "epoch": 2.7466216216216215, + "grad_norm": 0.41826799511909485, + "learning_rate": 4.850150909323447e-05, + "loss": 0.1904, + "step": 813 + }, + { + "epoch": 2.75, + "grad_norm": 0.3715426027774811, + "learning_rate": 4.8510406748153355e-05, + "loss": 0.1782, + "step": 814 + }, + { + "epoch": 2.7533783783783785, + "grad_norm": 0.39606815576553345, + "learning_rate": 4.8519293478999614e-05, + "loss": 0.1824, + "step": 815 + }, + { + "epoch": 2.756756756756757, + "grad_norm": 0.3894452452659607, + "learning_rate": 4.8528169312564355e-05, + "loss": 0.1799, + "step": 816 + }, + { + "epoch": 2.760135135135135, + "grad_norm": 0.38379618525505066, + "learning_rate": 4.8537034275540264e-05, + "loss": 0.1792, + "step": 817 + }, + { + "epoch": 2.7635135135135136, + "grad_norm": 0.4054020643234253, + "learning_rate": 4.854588839452205e-05, + "loss": 0.1842, + "step": 818 + }, + { + "epoch": 2.766891891891892, + "grad_norm": 0.424482524394989, + "learning_rate": 4.855473169600698e-05, + "loss": 0.1781, + "step": 819 + }, + { + "epoch": 2.77027027027027, + "grad_norm": 0.40162649750709534, + "learning_rate": 4.856356420639528e-05, + "loss": 0.2129, + "step": 820 + }, + { + "epoch": 2.7736486486486487, + "grad_norm": 0.4196532666683197, + "learning_rate": 4.857238595199068e-05, + "loss": 0.1844, + "step": 821 + }, + { + "epoch": 2.777027027027027, + "grad_norm": 0.39450812339782715, + "learning_rate": 4.858119695900084e-05, + "loss": 0.1917, + "step": 822 + }, + { + "epoch": 2.7804054054054053, + "grad_norm": 0.37791892886161804, + "learning_rate": 4.858999725353783e-05, + "loss": 0.1562, + "step": 823 + }, + { + "epoch": 2.7837837837837838, + "grad_norm": 0.4025024473667145, + "learning_rate": 4.8598786861618605e-05, + "loss": 0.1963, + "step": 824 + }, + { + "epoch": 2.7871621621621623, + "grad_norm": 0.382782518863678, + "learning_rate": 4.860756580916542e-05, + "loss": 0.1811, + "step": 825 + }, + { + "epoch": 2.7905405405405403, + "grad_norm": 0.40827038884162903, + "learning_rate": 4.861633412200637e-05, + "loss": 0.1914, + "step": 826 + }, + { + "epoch": 2.793918918918919, + "grad_norm": 0.35224664211273193, + "learning_rate": 4.862509182587578e-05, + "loss": 0.1587, + "step": 827 + }, + { + "epoch": 2.7972972972972974, + "grad_norm": 0.40827634930610657, + "learning_rate": 4.863383894641467e-05, + "loss": 0.1648, + "step": 828 + }, + { + "epoch": 2.8006756756756754, + "grad_norm": 0.3934107720851898, + "learning_rate": 4.864257550917123e-05, + "loss": 0.1867, + "step": 829 + }, + { + "epoch": 2.804054054054054, + "grad_norm": 0.4319005012512207, + "learning_rate": 4.865130153960124e-05, + "loss": 0.1789, + "step": 830 + }, + { + "epoch": 2.8074324324324325, + "grad_norm": 0.4071970582008362, + "learning_rate": 4.8660017063068526e-05, + "loss": 0.1718, + "step": 831 + }, + { + "epoch": 2.810810810810811, + "grad_norm": 0.4246816635131836, + "learning_rate": 4.8668722104845403e-05, + "loss": 0.1911, + "step": 832 + }, + { + "epoch": 2.814189189189189, + "grad_norm": 0.41370970010757446, + "learning_rate": 4.8677416690113134e-05, + "loss": 0.1968, + "step": 833 + }, + { + "epoch": 2.8175675675675675, + "grad_norm": 0.3547184467315674, + "learning_rate": 4.868610084396232e-05, + "loss": 0.1612, + "step": 834 + }, + { + "epoch": 2.820945945945946, + "grad_norm": 0.4511527121067047, + "learning_rate": 4.869477459139337e-05, + "loss": 0.1608, + "step": 835 + }, + { + "epoch": 2.8243243243243246, + "grad_norm": 0.394466757774353, + "learning_rate": 4.870343795731694e-05, + "loss": 0.1846, + "step": 836 + }, + { + "epoch": 2.8277027027027026, + "grad_norm": 0.3896351158618927, + "learning_rate": 4.8712090966554334e-05, + "loss": 0.1954, + "step": 837 + }, + { + "epoch": 2.831081081081081, + "grad_norm": 0.4012662172317505, + "learning_rate": 4.872073364383795e-05, + "loss": 0.1875, + "step": 838 + }, + { + "epoch": 2.8344594594594597, + "grad_norm": 0.39143961668014526, + "learning_rate": 4.8729366013811674e-05, + "loss": 0.2066, + "step": 839 + }, + { + "epoch": 2.8378378378378377, + "grad_norm": 0.4024941325187683, + "learning_rate": 4.8737988101031366e-05, + "loss": 0.1836, + "step": 840 + }, + { + "epoch": 2.8412162162162162, + "grad_norm": 0.3429911732673645, + "learning_rate": 4.874659992996521e-05, + "loss": 0.1687, + "step": 841 + }, + { + "epoch": 2.8445945945945947, + "grad_norm": 0.3772415816783905, + "learning_rate": 4.875520152499416e-05, + "loss": 0.2006, + "step": 842 + }, + { + "epoch": 2.847972972972973, + "grad_norm": 0.41397014260292053, + "learning_rate": 4.876379291041238e-05, + "loss": 0.2175, + "step": 843 + }, + { + "epoch": 2.8513513513513513, + "grad_norm": 0.41761255264282227, + "learning_rate": 4.8772374110427594e-05, + "loss": 0.1779, + "step": 844 + }, + { + "epoch": 2.85472972972973, + "grad_norm": 0.4116886854171753, + "learning_rate": 4.878094514916154e-05, + "loss": 0.2, + "step": 845 + }, + { + "epoch": 2.858108108108108, + "grad_norm": 0.3655768930912018, + "learning_rate": 4.8789506050650396e-05, + "loss": 0.1818, + "step": 846 + }, + { + "epoch": 2.8614864864864864, + "grad_norm": 0.43548357486724854, + "learning_rate": 4.879805683884512e-05, + "loss": 0.2029, + "step": 847 + }, + { + "epoch": 2.864864864864865, + "grad_norm": 0.3908335566520691, + "learning_rate": 4.8806597537611906e-05, + "loss": 0.1936, + "step": 848 + }, + { + "epoch": 2.868243243243243, + "grad_norm": 0.3648586869239807, + "learning_rate": 4.881512817073255e-05, + "loss": 0.1765, + "step": 849 + }, + { + "epoch": 2.8716216216216215, + "grad_norm": 0.39892107248306274, + "learning_rate": 4.882364876190489e-05, + "loss": 0.1594, + "step": 850 + }, + { + "epoch": 2.875, + "grad_norm": 0.3832027316093445, + "learning_rate": 4.8832159334743136e-05, + "loss": 0.1773, + "step": 851 + }, + { + "epoch": 2.8783783783783785, + "grad_norm": 0.4372403621673584, + "learning_rate": 4.884065991277833e-05, + "loss": 0.2057, + "step": 852 + }, + { + "epoch": 2.881756756756757, + "grad_norm": 0.36279168725013733, + "learning_rate": 4.8849150519458726e-05, + "loss": 0.1726, + "step": 853 + }, + { + "epoch": 2.885135135135135, + "grad_norm": 0.37340691685676575, + "learning_rate": 4.885763117815009e-05, + "loss": 0.178, + "step": 854 + }, + { + "epoch": 2.8885135135135136, + "grad_norm": 0.3998451828956604, + "learning_rate": 4.886610191213622e-05, + "loss": 0.1816, + "step": 855 + }, + { + "epoch": 2.891891891891892, + "grad_norm": 0.3779732584953308, + "learning_rate": 4.887456274461922e-05, + "loss": 0.2068, + "step": 856 + }, + { + "epoch": 2.89527027027027, + "grad_norm": 0.4435446262359619, + "learning_rate": 4.8883013698719973e-05, + "loss": 0.215, + "step": 857 + }, + { + "epoch": 2.8986486486486487, + "grad_norm": 0.37998583912849426, + "learning_rate": 4.889145479747843e-05, + "loss": 0.181, + "step": 858 + }, + { + "epoch": 2.902027027027027, + "grad_norm": 0.4032045900821686, + "learning_rate": 4.889988606385404e-05, + "loss": 0.1878, + "step": 859 + }, + { + "epoch": 2.9054054054054053, + "grad_norm": 0.4383464753627777, + "learning_rate": 4.8908307520726135e-05, + "loss": 0.1948, + "step": 860 + }, + { + "epoch": 2.9087837837837838, + "grad_norm": 0.38737547397613525, + "learning_rate": 4.891671919089425e-05, + "loss": 0.2007, + "step": 861 + }, + { + "epoch": 2.9121621621621623, + "grad_norm": 0.39179664850234985, + "learning_rate": 4.892512109707855e-05, + "loss": 0.1865, + "step": 862 + }, + { + "epoch": 2.9155405405405403, + "grad_norm": 0.3944191038608551, + "learning_rate": 4.893351326192016e-05, + "loss": 0.1831, + "step": 863 + }, + { + "epoch": 2.918918918918919, + "grad_norm": 0.3212383985519409, + "learning_rate": 4.894189570798156e-05, + "loss": 0.1494, + "step": 864 + }, + { + "epoch": 2.9222972972972974, + "grad_norm": 0.3861388564109802, + "learning_rate": 4.895026845774691e-05, + "loss": 0.2045, + "step": 865 + }, + { + "epoch": 2.9256756756756754, + "grad_norm": 0.37900376319885254, + "learning_rate": 4.895863153362244e-05, + "loss": 0.1818, + "step": 866 + }, + { + "epoch": 2.929054054054054, + "grad_norm": 0.4495367109775543, + "learning_rate": 4.896698495793684e-05, + "loss": 0.1936, + "step": 867 + }, + { + "epoch": 2.9324324324324325, + "grad_norm": 0.40020546317100525, + "learning_rate": 4.897532875294154e-05, + "loss": 0.1964, + "step": 868 + }, + { + "epoch": 2.935810810810811, + "grad_norm": 0.3625187277793884, + "learning_rate": 4.8983662940811115e-05, + "loss": 0.1737, + "step": 869 + }, + { + "epoch": 2.939189189189189, + "grad_norm": 0.3938811719417572, + "learning_rate": 4.899198754364365e-05, + "loss": 0.1488, + "step": 870 + }, + { + "epoch": 2.9425675675675675, + "grad_norm": 0.36418697237968445, + "learning_rate": 4.900030258346106e-05, + "loss": 0.1822, + "step": 871 + }, + { + "epoch": 2.945945945945946, + "grad_norm": 0.3703506588935852, + "learning_rate": 4.900860808220946e-05, + "loss": 0.1846, + "step": 872 + }, + { + "epoch": 2.9493243243243246, + "grad_norm": 0.4332616925239563, + "learning_rate": 4.90169040617595e-05, + "loss": 0.2105, + "step": 873 + }, + { + "epoch": 2.9527027027027026, + "grad_norm": 0.3770284354686737, + "learning_rate": 4.9025190543906715e-05, + "loss": 0.1815, + "step": 874 + }, + { + "epoch": 2.956081081081081, + "grad_norm": 0.3703818619251251, + "learning_rate": 4.903346755037189e-05, + "loss": 0.1713, + "step": 875 + }, + { + "epoch": 2.9594594594594597, + "grad_norm": 0.37928834557533264, + "learning_rate": 4.904173510280135e-05, + "loss": 0.1915, + "step": 876 + }, + { + "epoch": 2.9628378378378377, + "grad_norm": 0.4267534911632538, + "learning_rate": 4.904999322276735e-05, + "loss": 0.2157, + "step": 877 + }, + { + "epoch": 2.9662162162162162, + "grad_norm": 0.3936428725719452, + "learning_rate": 4.9058241931768385e-05, + "loss": 0.1947, + "step": 878 + }, + { + "epoch": 2.9695945945945947, + "grad_norm": 0.3503859043121338, + "learning_rate": 4.9066481251229535e-05, + "loss": 0.1836, + "step": 879 + }, + { + "epoch": 2.972972972972973, + "grad_norm": 1.8326294422149658, + "learning_rate": 4.907471120250281e-05, + "loss": 0.4757, + "step": 880 + }, + { + "epoch": 2.9763513513513513, + "grad_norm": 0.4133761525154114, + "learning_rate": 4.9082931806867474e-05, + "loss": 0.2003, + "step": 881 + }, + { + "epoch": 2.97972972972973, + "grad_norm": 0.3848228454589844, + "learning_rate": 4.909114308553033e-05, + "loss": 0.19, + "step": 882 + }, + { + "epoch": 2.983108108108108, + "grad_norm": 0.38853901624679565, + "learning_rate": 4.909934505962615e-05, + "loss": 0.2071, + "step": 883 + }, + { + "epoch": 2.9864864864864864, + "grad_norm": 0.35326841473579407, + "learning_rate": 4.9107537750217886e-05, + "loss": 0.1742, + "step": 884 + }, + { + "epoch": 2.989864864864865, + "grad_norm": 0.33690720796585083, + "learning_rate": 4.9115721178297093e-05, + "loss": 0.1734, + "step": 885 + }, + { + "epoch": 2.993243243243243, + "grad_norm": 0.35074329376220703, + "learning_rate": 4.9123895364784184e-05, + "loss": 0.1623, + "step": 886 + }, + { + "epoch": 2.9966216216216215, + "grad_norm": 0.35647299885749817, + "learning_rate": 4.913206033052877e-05, + "loss": 0.1647, + "step": 887 + }, + { + "epoch": 3.0, + "grad_norm": 0.4068041145801544, + "learning_rate": 4.914021609631002e-05, + "loss": 0.1831, + "step": 888 + }, + { + "epoch": 3.0033783783783785, + "grad_norm": 0.4528217017650604, + "learning_rate": 4.91483626828369e-05, + "loss": 0.128, + "step": 889 + }, + { + "epoch": 3.0067567567567566, + "grad_norm": 0.3695124089717865, + "learning_rate": 4.915650011074855e-05, + "loss": 0.1077, + "step": 890 + }, + { + "epoch": 3.010135135135135, + "grad_norm": 0.570489227771759, + "learning_rate": 4.916462840061458e-05, + "loss": 0.11, + "step": 891 + }, + { + "epoch": 3.0135135135135136, + "grad_norm": 0.43204137682914734, + "learning_rate": 4.917274757293539e-05, + "loss": 0.1032, + "step": 892 + }, + { + "epoch": 3.016891891891892, + "grad_norm": 0.5314778089523315, + "learning_rate": 4.918085764814244e-05, + "loss": 0.1304, + "step": 893 + }, + { + "epoch": 3.02027027027027, + "grad_norm": 0.3921876847743988, + "learning_rate": 4.9188958646598624e-05, + "loss": 0.0949, + "step": 894 + }, + { + "epoch": 3.0236486486486487, + "grad_norm": 0.35549548268318176, + "learning_rate": 4.919705058859854e-05, + "loss": 0.0989, + "step": 895 + }, + { + "epoch": 3.027027027027027, + "grad_norm": 0.4394858181476593, + "learning_rate": 4.920513349436875e-05, + "loss": 0.1152, + "step": 896 + }, + { + "epoch": 3.0304054054054053, + "grad_norm": 0.4386696219444275, + "learning_rate": 4.92132073840682e-05, + "loss": 0.1163, + "step": 897 + }, + { + "epoch": 3.0337837837837838, + "grad_norm": 0.38169318437576294, + "learning_rate": 4.922127227778841e-05, + "loss": 0.1078, + "step": 898 + }, + { + "epoch": 3.0371621621621623, + "grad_norm": 0.410858690738678, + "learning_rate": 4.9229328195553815e-05, + "loss": 0.1165, + "step": 899 + }, + { + "epoch": 3.0405405405405403, + "grad_norm": 0.4674147069454193, + "learning_rate": 4.923737515732209e-05, + "loss": 0.12, + "step": 900 + }, + { + "epoch": 3.043918918918919, + "grad_norm": 0.3922157883644104, + "learning_rate": 4.924541318298438e-05, + "loss": 0.1131, + "step": 901 + }, + { + "epoch": 3.0472972972972974, + "grad_norm": 0.3732883632183075, + "learning_rate": 4.92534422923657e-05, + "loss": 0.1025, + "step": 902 + }, + { + "epoch": 3.050675675675676, + "grad_norm": 0.4513222277164459, + "learning_rate": 4.9261462505225106e-05, + "loss": 0.1207, + "step": 903 + }, + { + "epoch": 3.054054054054054, + "grad_norm": 0.39665260910987854, + "learning_rate": 4.926947384125606e-05, + "loss": 0.1115, + "step": 904 + }, + { + "epoch": 3.0574324324324325, + "grad_norm": 0.3784550130367279, + "learning_rate": 4.927747632008672e-05, + "loss": 0.1151, + "step": 905 + }, + { + "epoch": 3.060810810810811, + "grad_norm": 0.44802817702293396, + "learning_rate": 4.9285469961280226e-05, + "loss": 0.113, + "step": 906 + }, + { + "epoch": 3.064189189189189, + "grad_norm": 0.37687918543815613, + "learning_rate": 4.9293454784334924e-05, + "loss": 0.1157, + "step": 907 + }, + { + "epoch": 3.0675675675675675, + "grad_norm": 0.3582609295845032, + "learning_rate": 4.9301430808684754e-05, + "loss": 0.1031, + "step": 908 + }, + { + "epoch": 3.070945945945946, + "grad_norm": 0.3885122239589691, + "learning_rate": 4.930939805369946e-05, + "loss": 0.1043, + "step": 909 + }, + { + "epoch": 3.074324324324324, + "grad_norm": 0.5625414252281189, + "learning_rate": 4.93173565386849e-05, + "loss": 0.1109, + "step": 910 + }, + { + "epoch": 3.0777027027027026, + "grad_norm": 0.3783611059188843, + "learning_rate": 4.932530628288331e-05, + "loss": 0.1047, + "step": 911 + }, + { + "epoch": 3.081081081081081, + "grad_norm": 0.48341429233551025, + "learning_rate": 4.933324730547361e-05, + "loss": 0.0961, + "step": 912 + }, + { + "epoch": 3.0844594594594597, + "grad_norm": 0.399813711643219, + "learning_rate": 4.934117962557165e-05, + "loss": 0.1178, + "step": 913 + }, + { + "epoch": 3.0878378378378377, + "grad_norm": 0.40703117847442627, + "learning_rate": 4.9349103262230524e-05, + "loss": 0.1179, + "step": 914 + }, + { + "epoch": 3.0912162162162162, + "grad_norm": 0.39595040678977966, + "learning_rate": 4.935701823444081e-05, + "loss": 0.1161, + "step": 915 + }, + { + "epoch": 3.0945945945945947, + "grad_norm": 0.3974362313747406, + "learning_rate": 4.9364924561130845e-05, + "loss": 0.1167, + "step": 916 + }, + { + "epoch": 3.097972972972973, + "grad_norm": 0.426070898771286, + "learning_rate": 4.937282226116702e-05, + "loss": 0.1305, + "step": 917 + }, + { + "epoch": 3.1013513513513513, + "grad_norm": 0.38482606410980225, + "learning_rate": 4.938071135335405e-05, + "loss": 0.1018, + "step": 918 + }, + { + "epoch": 3.10472972972973, + "grad_norm": 0.4263748228549957, + "learning_rate": 4.938859185643519e-05, + "loss": 0.1156, + "step": 919 + }, + { + "epoch": 3.108108108108108, + "grad_norm": 0.39638155698776245, + "learning_rate": 4.939646378909259e-05, + "loss": 0.1157, + "step": 920 + }, + { + "epoch": 3.1114864864864864, + "grad_norm": 0.40607044100761414, + "learning_rate": 4.940432716994748e-05, + "loss": 0.1197, + "step": 921 + }, + { + "epoch": 3.114864864864865, + "grad_norm": 0.4135359823703766, + "learning_rate": 4.9412182017560496e-05, + "loss": 0.1202, + "step": 922 + }, + { + "epoch": 3.1182432432432434, + "grad_norm": 0.38252270221710205, + "learning_rate": 4.942002835043187e-05, + "loss": 0.1126, + "step": 923 + }, + { + "epoch": 3.1216216216216215, + "grad_norm": 0.3727289140224457, + "learning_rate": 4.942786618700178e-05, + "loss": 0.1111, + "step": 924 + }, + { + "epoch": 3.125, + "grad_norm": 0.380953848361969, + "learning_rate": 4.9435695545650545e-05, + "loss": 0.1151, + "step": 925 + }, + { + "epoch": 3.1283783783783785, + "grad_norm": 0.37086960673332214, + "learning_rate": 4.944351644469891e-05, + "loss": 0.1106, + "step": 926 + }, + { + "epoch": 3.1317567567567566, + "grad_norm": 0.4138876497745514, + "learning_rate": 4.945132890240829e-05, + "loss": 0.1239, + "step": 927 + }, + { + "epoch": 3.135135135135135, + "grad_norm": 0.4137141704559326, + "learning_rate": 4.945913293698104e-05, + "loss": 0.1252, + "step": 928 + }, + { + "epoch": 3.1385135135135136, + "grad_norm": 0.3904976546764374, + "learning_rate": 4.9466928566560696e-05, + "loss": 0.107, + "step": 929 + }, + { + "epoch": 3.141891891891892, + "grad_norm": 0.3888641595840454, + "learning_rate": 4.9474715809232256e-05, + "loss": 0.113, + "step": 930 + }, + { + "epoch": 3.14527027027027, + "grad_norm": 0.39176711440086365, + "learning_rate": 4.948249468302239e-05, + "loss": 0.1135, + "step": 931 + }, + { + "epoch": 3.1486486486486487, + "grad_norm": 0.3809608817100525, + "learning_rate": 4.9490265205899697e-05, + "loss": 0.1051, + "step": 932 + }, + { + "epoch": 3.152027027027027, + "grad_norm": 0.4019430875778198, + "learning_rate": 4.9498027395775006e-05, + "loss": 0.121, + "step": 933 + }, + { + "epoch": 3.1554054054054053, + "grad_norm": 0.3715710937976837, + "learning_rate": 4.950578127050156e-05, + "loss": 0.1225, + "step": 934 + }, + { + "epoch": 3.1587837837837838, + "grad_norm": 0.4521862864494324, + "learning_rate": 4.95135268478753e-05, + "loss": 0.1294, + "step": 935 + }, + { + "epoch": 3.1621621621621623, + "grad_norm": 0.4279133379459381, + "learning_rate": 4.952126414563509e-05, + "loss": 0.1229, + "step": 936 + }, + { + "epoch": 3.1655405405405403, + "grad_norm": 0.4043607711791992, + "learning_rate": 4.952899318146297e-05, + "loss": 0.1117, + "step": 937 + }, + { + "epoch": 3.168918918918919, + "grad_norm": 0.34410539269447327, + "learning_rate": 4.9536713972984414e-05, + "loss": 0.0972, + "step": 938 + }, + { + "epoch": 3.1722972972972974, + "grad_norm": 0.43061432242393494, + "learning_rate": 4.954442653776852e-05, + "loss": 0.1333, + "step": 939 + }, + { + "epoch": 3.175675675675676, + "grad_norm": 0.36851924657821655, + "learning_rate": 4.955213089332832e-05, + "loss": 0.1114, + "step": 940 + }, + { + "epoch": 3.179054054054054, + "grad_norm": 0.37471747398376465, + "learning_rate": 4.955982705712095e-05, + "loss": 0.1169, + "step": 941 + }, + { + "epoch": 3.1824324324324325, + "grad_norm": 0.35380908846855164, + "learning_rate": 4.956751504654796e-05, + "loss": 0.1005, + "step": 942 + }, + { + "epoch": 3.185810810810811, + "grad_norm": 0.38561439514160156, + "learning_rate": 4.957519487895548e-05, + "loss": 0.1187, + "step": 943 + }, + { + "epoch": 3.189189189189189, + "grad_norm": 0.4120042324066162, + "learning_rate": 4.9582866571634485e-05, + "loss": 0.1273, + "step": 944 + }, + { + "epoch": 3.1925675675675675, + "grad_norm": 0.39863529801368713, + "learning_rate": 4.959053014182106e-05, + "loss": 0.1125, + "step": 945 + }, + { + "epoch": 3.195945945945946, + "grad_norm": 0.38955143094062805, + "learning_rate": 4.959818560669655e-05, + "loss": 0.1175, + "step": 946 + }, + { + "epoch": 3.199324324324324, + "grad_norm": 0.3573088049888611, + "learning_rate": 4.96058329833879e-05, + "loss": 0.1131, + "step": 947 + }, + { + "epoch": 3.2027027027027026, + "grad_norm": 0.4304436147212982, + "learning_rate": 4.961347228896777e-05, + "loss": 0.1262, + "step": 948 + }, + { + "epoch": 3.206081081081081, + "grad_norm": 0.39087799191474915, + "learning_rate": 4.962110354045488e-05, + "loss": 0.1191, + "step": 949 + }, + { + "epoch": 3.2094594594594597, + "grad_norm": 0.3944397568702698, + "learning_rate": 4.962872675481414e-05, + "loss": 0.1182, + "step": 950 + }, + { + "epoch": 3.2128378378378377, + "grad_norm": 0.40278294682502747, + "learning_rate": 4.9636341948956906e-05, + "loss": 0.1211, + "step": 951 + }, + { + "epoch": 3.2162162162162162, + "grad_norm": 0.41054508090019226, + "learning_rate": 4.964394913974124e-05, + "loss": 0.1195, + "step": 952 + }, + { + "epoch": 3.2195945945945947, + "grad_norm": 0.4175059199333191, + "learning_rate": 4.965154834397211e-05, + "loss": 0.1353, + "step": 953 + }, + { + "epoch": 3.222972972972973, + "grad_norm": 0.3865712583065033, + "learning_rate": 4.965913957840159e-05, + "loss": 0.1111, + "step": 954 + }, + { + "epoch": 3.2263513513513513, + "grad_norm": 0.39877578616142273, + "learning_rate": 4.966672285972911e-05, + "loss": 0.1256, + "step": 955 + }, + { + "epoch": 3.22972972972973, + "grad_norm": 0.34512677788734436, + "learning_rate": 4.967429820460167e-05, + "loss": 0.1078, + "step": 956 + }, + { + "epoch": 3.233108108108108, + "grad_norm": 0.4256219267845154, + "learning_rate": 4.9681865629614064e-05, + "loss": 0.1313, + "step": 957 + }, + { + "epoch": 3.2364864864864864, + "grad_norm": 0.46931833028793335, + "learning_rate": 4.9689425151309074e-05, + "loss": 0.1396, + "step": 958 + }, + { + "epoch": 3.239864864864865, + "grad_norm": 0.40905606746673584, + "learning_rate": 4.969697678617773e-05, + "loss": 0.1242, + "step": 959 + }, + { + "epoch": 3.2432432432432434, + "grad_norm": 0.36054709553718567, + "learning_rate": 4.970452055065948e-05, + "loss": 0.1149, + "step": 960 + }, + { + "epoch": 3.2466216216216215, + "grad_norm": 0.3850068747997284, + "learning_rate": 4.9712056461142423e-05, + "loss": 0.1074, + "step": 961 + }, + { + "epoch": 3.25, + "grad_norm": 0.42875200510025024, + "learning_rate": 4.971958453396355e-05, + "loss": 0.1152, + "step": 962 + }, + { + "epoch": 3.2533783783783785, + "grad_norm": 0.413173109292984, + "learning_rate": 4.972710478540891e-05, + "loss": 0.1273, + "step": 963 + }, + { + "epoch": 3.2567567567567566, + "grad_norm": 0.3743407428264618, + "learning_rate": 4.973461723171385e-05, + "loss": 0.1149, + "step": 964 + }, + { + "epoch": 3.260135135135135, + "grad_norm": 0.44052329659461975, + "learning_rate": 4.9742121889063213e-05, + "loss": 0.1301, + "step": 965 + }, + { + "epoch": 3.2635135135135136, + "grad_norm": 0.38358744978904724, + "learning_rate": 4.974961877359156e-05, + "loss": 0.1133, + "step": 966 + }, + { + "epoch": 3.266891891891892, + "grad_norm": 0.423977792263031, + "learning_rate": 4.975710790138336e-05, + "loss": 0.1346, + "step": 967 + }, + { + "epoch": 3.27027027027027, + "grad_norm": 1.5824426412582397, + "learning_rate": 4.976458928847323e-05, + "loss": 0.3884, + "step": 968 + }, + { + "epoch": 3.2736486486486487, + "grad_norm": 0.41981515288352966, + "learning_rate": 4.977206295084609e-05, + "loss": 0.1164, + "step": 969 + }, + { + "epoch": 3.277027027027027, + "grad_norm": 0.42901647090911865, + "learning_rate": 4.9779528904437424e-05, + "loss": 0.1269, + "step": 970 + }, + { + "epoch": 3.2804054054054053, + "grad_norm": 0.3595876693725586, + "learning_rate": 4.978698716513342e-05, + "loss": 0.1121, + "step": 971 + }, + { + "epoch": 3.2837837837837838, + "grad_norm": 0.39047256112098694, + "learning_rate": 4.9794437748771244e-05, + "loss": 0.1219, + "step": 972 + }, + { + "epoch": 3.2871621621621623, + "grad_norm": 0.41640281677246094, + "learning_rate": 4.9801880671139204e-05, + "loss": 0.1303, + "step": 973 + }, + { + "epoch": 3.2905405405405403, + "grad_norm": 0.371383398771286, + "learning_rate": 4.980931594797693e-05, + "loss": 0.112, + "step": 974 + }, + { + "epoch": 3.293918918918919, + "grad_norm": 0.4297455847263336, + "learning_rate": 4.981674359497562e-05, + "loss": 0.1326, + "step": 975 + }, + { + "epoch": 3.2972972972972974, + "grad_norm": 0.39308154582977295, + "learning_rate": 4.98241636277782e-05, + "loss": 0.1244, + "step": 976 + }, + { + "epoch": 3.3006756756756754, + "grad_norm": 0.3956491947174072, + "learning_rate": 4.983157606197955e-05, + "loss": 0.1203, + "step": 977 + }, + { + "epoch": 3.304054054054054, + "grad_norm": 0.41044941544532776, + "learning_rate": 4.98389809131267e-05, + "loss": 0.1247, + "step": 978 + }, + { + "epoch": 3.3074324324324325, + "grad_norm": 0.4159061312675476, + "learning_rate": 4.984637819671897e-05, + "loss": 0.1321, + "step": 979 + }, + { + "epoch": 3.310810810810811, + "grad_norm": 0.3757896423339844, + "learning_rate": 4.985376792820825e-05, + "loss": 0.1157, + "step": 980 + }, + { + "epoch": 3.314189189189189, + "grad_norm": 0.3729749321937561, + "learning_rate": 4.986115012299915e-05, + "loss": 0.1103, + "step": 981 + }, + { + "epoch": 3.3175675675675675, + "grad_norm": 0.3994838297367096, + "learning_rate": 4.986852479644916e-05, + "loss": 0.1284, + "step": 982 + }, + { + "epoch": 3.320945945945946, + "grad_norm": 0.3987760841846466, + "learning_rate": 4.987589196386893e-05, + "loss": 0.1126, + "step": 983 + }, + { + "epoch": 3.3243243243243246, + "grad_norm": 0.3847789168357849, + "learning_rate": 4.988325164052236e-05, + "loss": 0.1199, + "step": 984 + }, + { + "epoch": 3.3277027027027026, + "grad_norm": 0.41571831703186035, + "learning_rate": 4.9890603841626866e-05, + "loss": 0.1295, + "step": 985 + }, + { + "epoch": 3.331081081081081, + "grad_norm": 0.3801005184650421, + "learning_rate": 4.989794858235352e-05, + "loss": 0.1276, + "step": 986 + }, + { + "epoch": 3.3344594594594597, + "grad_norm": 0.4302000403404236, + "learning_rate": 4.990528587782729e-05, + "loss": 0.1355, + "step": 987 + }, + { + "epoch": 3.3378378378378377, + "grad_norm": 0.3750097155570984, + "learning_rate": 4.9912615743127146e-05, + "loss": 0.1159, + "step": 988 + }, + { + "epoch": 3.3412162162162162, + "grad_norm": 0.43669402599334717, + "learning_rate": 4.991993819328633e-05, + "loss": 0.1349, + "step": 989 + }, + { + "epoch": 3.3445945945945947, + "grad_norm": 0.40606677532196045, + "learning_rate": 4.9927253243292505e-05, + "loss": 0.123, + "step": 990 + }, + { + "epoch": 3.347972972972973, + "grad_norm": 0.40521538257598877, + "learning_rate": 4.993456090808793e-05, + "loss": 0.1212, + "step": 991 + }, + { + "epoch": 3.3513513513513513, + "grad_norm": 0.38929492235183716, + "learning_rate": 4.994186120256965e-05, + "loss": 0.1204, + "step": 992 + }, + { + "epoch": 3.35472972972973, + "grad_norm": 0.3639843761920929, + "learning_rate": 4.9949154141589696e-05, + "loss": 0.1175, + "step": 993 + }, + { + "epoch": 3.358108108108108, + "grad_norm": 0.3805122971534729, + "learning_rate": 4.995643973995523e-05, + "loss": 0.1198, + "step": 994 + }, + { + "epoch": 3.3614864864864864, + "grad_norm": 0.395796000957489, + "learning_rate": 4.9963718012428765e-05, + "loss": 0.1348, + "step": 995 + }, + { + "epoch": 3.364864864864865, + "grad_norm": 0.3744940459728241, + "learning_rate": 4.9970988973728314e-05, + "loss": 0.1104, + "step": 996 + }, + { + "epoch": 3.3682432432432434, + "grad_norm": 1.3542141914367676, + "learning_rate": 4.99782526385276e-05, + "loss": 0.3864, + "step": 997 + }, + { + "epoch": 3.3716216216216215, + "grad_norm": 0.42771685123443604, + "learning_rate": 4.998550902145619e-05, + "loss": 0.1383, + "step": 998 + }, + { + "epoch": 3.375, + "grad_norm": 0.3889855444431305, + "learning_rate": 4.999275813709971e-05, + "loss": 0.1318, + "step": 999 + }, + { + "epoch": 3.3783783783783785, + "grad_norm": 0.37391164898872375, + "learning_rate": 5e-05, + "loss": 0.1285, + "step": 1000 + }, + { + "epoch": 3.3817567567567566, + "grad_norm": 0.3580280542373657, + "learning_rate": 4.9996356488619556e-05, + "loss": 0.1125, + "step": 1001 + }, + { + "epoch": 3.385135135135135, + "grad_norm": 0.3707447052001953, + "learning_rate": 4.9985427016598435e-05, + "loss": 0.1221, + "step": 1002 + }, + { + "epoch": 3.3885135135135136, + "grad_norm": 0.3772697150707245, + "learning_rate": 4.996721476998771e-05, + "loss": 0.1215, + "step": 1003 + }, + { + "epoch": 3.391891891891892, + "grad_norm": 0.40340831875801086, + "learning_rate": 4.9941725057840504e-05, + "loss": 0.1233, + "step": 1004 + }, + { + "epoch": 3.39527027027027, + "grad_norm": 0.3812180161476135, + "learning_rate": 4.9908965310664374e-05, + "loss": 0.1149, + "step": 1005 + }, + { + "epoch": 3.3986486486486487, + "grad_norm": 0.4367975890636444, + "learning_rate": 4.986894507825522e-05, + "loss": 0.1358, + "step": 1006 + }, + { + "epoch": 3.402027027027027, + "grad_norm": 0.4075244963169098, + "learning_rate": 4.9821676026913475e-05, + "loss": 0.132, + "step": 1007 + }, + { + "epoch": 3.4054054054054053, + "grad_norm": 0.392610639333725, + "learning_rate": 4.9767171936043175e-05, + "loss": 0.1169, + "step": 1008 + }, + { + "epoch": 3.4087837837837838, + "grad_norm": 0.3933659791946411, + "learning_rate": 4.970544869413522e-05, + "loss": 0.1197, + "step": 1009 + }, + { + "epoch": 3.4121621621621623, + "grad_norm": 0.3939339816570282, + "learning_rate": 4.963652429413563e-05, + "loss": 0.1315, + "step": 1010 + }, + { + "epoch": 3.4155405405405403, + "grad_norm": 0.37384697794914246, + "learning_rate": 4.9560418828200494e-05, + "loss": 0.1187, + "step": 1011 + }, + { + "epoch": 3.418918918918919, + "grad_norm": 0.3945925831794739, + "learning_rate": 4.9477154481838875e-05, + "loss": 0.1353, + "step": 1012 + }, + { + "epoch": 3.4222972972972974, + "grad_norm": 0.42826682329177856, + "learning_rate": 4.9386755527445475e-05, + "loss": 0.1459, + "step": 1013 + }, + { + "epoch": 3.4256756756756754, + "grad_norm": 0.4206187427043915, + "learning_rate": 4.928924831722504e-05, + "loss": 0.1292, + "step": 1014 + }, + { + "epoch": 3.429054054054054, + "grad_norm": 0.38580143451690674, + "learning_rate": 4.9184661275510446e-05, + "loss": 0.1153, + "step": 1015 + }, + { + "epoch": 3.4324324324324325, + "grad_norm": 0.393794983625412, + "learning_rate": 4.907302489047662e-05, + "loss": 0.1179, + "step": 1016 + }, + { + "epoch": 3.435810810810811, + "grad_norm": 0.44305288791656494, + "learning_rate": 4.895437170525303e-05, + "loss": 0.1302, + "step": 1017 + }, + { + "epoch": 3.439189189189189, + "grad_norm": 0.4056559205055237, + "learning_rate": 4.882873630843699e-05, + "loss": 0.1302, + "step": 1018 + }, + { + "epoch": 3.4425675675675675, + "grad_norm": 0.40912213921546936, + "learning_rate": 4.869615532401074e-05, + "loss": 0.1327, + "step": 1019 + }, + { + "epoch": 3.445945945945946, + "grad_norm": 0.38370949029922485, + "learning_rate": 4.855666740066522e-05, + "loss": 0.129, + "step": 1020 + }, + { + "epoch": 3.4493243243243246, + "grad_norm": 0.41348859667778015, + "learning_rate": 4.841031320053351e-05, + "loss": 0.1243, + "step": 1021 + }, + { + "epoch": 3.4527027027027026, + "grad_norm": 0.40817150473594666, + "learning_rate": 4.825713538733748e-05, + "loss": 0.1321, + "step": 1022 + }, + { + "epoch": 3.456081081081081, + "grad_norm": 31.2849063873291, + "learning_rate": 4.80971786139509e-05, + "loss": 0.186, + "step": 1023 + }, + { + "epoch": 3.4594594594594597, + "grad_norm": 8.919435501098633, + "learning_rate": 4.793048950938256e-05, + "loss": 0.6012, + "step": 1024 + }, + { + "epoch": 3.4628378378378377, + "grad_norm": 1.2597789764404297, + "learning_rate": 4.7757116665183614e-05, + "loss": 0.373, + "step": 1025 + }, + { + "epoch": 3.4662162162162162, + "grad_norm": 144.5867156982422, + "learning_rate": 4.757711062128251e-05, + "loss": 0.6334, + "step": 1026 + }, + { + "epoch": 3.4695945945945947, + "grad_norm": 42.916805267333984, + "learning_rate": 4.739052385125216e-05, + "loss": 0.6214, + "step": 1027 + }, + { + "epoch": 3.472972972972973, + "grad_norm": 0.7827467322349548, + "learning_rate": 4.7197410747013376e-05, + "loss": 0.308, + "step": 1028 + }, + { + "epoch": 3.4763513513513513, + "grad_norm": 0.7671857476234436, + "learning_rate": 4.6997827602979024e-05, + "loss": 0.3552, + "step": 1029 + }, + { + "epoch": 3.47972972972973, + "grad_norm": 3.2267792224884033, + "learning_rate": 4.67918325996437e-05, + "loss": 0.3405, + "step": 1030 + }, + { + "epoch": 3.483108108108108, + "grad_norm": 13.3894624710083, + "learning_rate": 4.6579485786623475e-05, + "loss": 0.4361, + "step": 1031 + }, + { + "epoch": 3.4864864864864864, + "grad_norm": 2.525721311569214, + "learning_rate": 4.636084906515085e-05, + "loss": 0.1359, + "step": 1032 + }, + { + "epoch": 3.489864864864865, + "grad_norm": 52.08081817626953, + "learning_rate": 4.6135986170029947e-05, + "loss": 0.5506, + "step": 1033 + }, + { + "epoch": 3.4932432432432434, + "grad_norm": 1.6132444143295288, + "learning_rate": 4.5904962651057134e-05, + "loss": 0.1839, + "step": 1034 + }, + { + "epoch": 3.4966216216216215, + "grad_norm": 1.116671085357666, + "learning_rate": 4.566784585391263e-05, + "loss": 0.1534, + "step": 1035 + }, + { + "epoch": 3.5, + "grad_norm": 0.6348862051963806, + "learning_rate": 4.542470490052853e-05, + "loss": 0.1429, + "step": 1036 + }, + { + "epoch": 3.5033783783783785, + "grad_norm": 0.6647199392318726, + "learning_rate": 4.517561066893909e-05, + "loss": 0.1351, + "step": 1037 + }, + { + "epoch": 3.506756756756757, + "grad_norm": 0.4823889136314392, + "learning_rate": 4.492063577261908e-05, + "loss": 0.1312, + "step": 1038 + }, + { + "epoch": 3.510135135135135, + "grad_norm": 0.48596346378326416, + "learning_rate": 4.4659854539316174e-05, + "loss": 0.1418, + "step": 1039 + }, + { + "epoch": 3.5135135135135136, + "grad_norm": 0.7251044511795044, + "learning_rate": 4.439334298938374e-05, + "loss": 0.1465, + "step": 1040 + }, + { + "epoch": 3.516891891891892, + "grad_norm": 0.4920099973678589, + "learning_rate": 4.4121178813620046e-05, + "loss": 0.1481, + "step": 1041 + }, + { + "epoch": 3.52027027027027, + "grad_norm": 0.42203307151794434, + "learning_rate": 4.384344135062071e-05, + "loss": 0.1219, + "step": 1042 + }, + { + "epoch": 3.5236486486486487, + "grad_norm": 0.47877076268196106, + "learning_rate": 4.3560211563650635e-05, + "loss": 0.1282, + "step": 1043 + }, + { + "epoch": 3.527027027027027, + "grad_norm": 0.4383416175842285, + "learning_rate": 4.327157201704241e-05, + "loss": 0.133, + "step": 1044 + }, + { + "epoch": 3.5304054054054053, + "grad_norm": 0.49161508679389954, + "learning_rate": 4.297760685212801e-05, + "loss": 0.1268, + "step": 1045 + }, + { + "epoch": 3.5337837837837838, + "grad_norm": 0.44632676243782043, + "learning_rate": 4.2678401762710726e-05, + "loss": 0.1304, + "step": 1046 + }, + { + "epoch": 3.5371621621621623, + "grad_norm": 0.48019739985466003, + "learning_rate": 4.237404397008455e-05, + "loss": 0.134, + "step": 1047 + }, + { + "epoch": 3.5405405405405403, + "grad_norm": 0.3879307508468628, + "learning_rate": 4.206462219760831e-05, + "loss": 0.1106, + "step": 1048 + }, + { + "epoch": 3.543918918918919, + "grad_norm": 8.76919937133789, + "learning_rate": 4.175022664484191e-05, + "loss": 0.1268, + "step": 1049 + }, + { + "epoch": 3.5472972972972974, + "grad_norm": 0.5539897680282593, + "learning_rate": 4.143094896125227e-05, + "loss": 0.134, + "step": 1050 + }, + { + "epoch": 3.5506756756756754, + "grad_norm": 0.41708147525787354, + "learning_rate": 4.1106882219496535e-05, + "loss": 0.1185, + "step": 1051 + }, + { + "epoch": 3.554054054054054, + "grad_norm": 0.7360028028488159, + "learning_rate": 4.077812088829051e-05, + "loss": 0.1251, + "step": 1052 + }, + { + "epoch": 3.5574324324324325, + "grad_norm": 0.49543508887290955, + "learning_rate": 4.044476080487005e-05, + "loss": 0.1405, + "step": 1053 + }, + { + "epoch": 3.560810810810811, + "grad_norm": 0.40071457624435425, + "learning_rate": 4.0106899147053564e-05, + "loss": 0.1162, + "step": 1054 + }, + { + "epoch": 3.564189189189189, + "grad_norm": 0.6165151596069336, + "learning_rate": 3.976463440491367e-05, + "loss": 0.123, + "step": 1055 + }, + { + "epoch": 3.5675675675675675, + "grad_norm": 0.4160919487476349, + "learning_rate": 3.941806635206639e-05, + "loss": 0.1276, + "step": 1056 + }, + { + "epoch": 3.570945945945946, + "grad_norm": 0.4072069525718689, + "learning_rate": 3.9067296016586105e-05, + "loss": 0.1336, + "step": 1057 + }, + { + "epoch": 3.5743243243243246, + "grad_norm": 0.4126531183719635, + "learning_rate": 3.871242565155485e-05, + "loss": 0.1267, + "step": 1058 + }, + { + "epoch": 3.5777027027027026, + "grad_norm": 2.380211353302002, + "learning_rate": 3.835355870525455e-05, + "loss": 0.4067, + "step": 1059 + }, + { + "epoch": 3.581081081081081, + "grad_norm": 0.3900314271450043, + "learning_rate": 3.799079979101075e-05, + "loss": 0.1231, + "step": 1060 + }, + { + "epoch": 3.5844594594594597, + "grad_norm": 0.4255238473415375, + "learning_rate": 3.7624254656696864e-05, + "loss": 0.1405, + "step": 1061 + }, + { + "epoch": 3.5878378378378377, + "grad_norm": 0.36614516377449036, + "learning_rate": 3.7254030153907494e-05, + "loss": 0.118, + "step": 1062 + }, + { + "epoch": 3.5912162162162162, + "grad_norm": 0.40764889121055603, + "learning_rate": 3.688023420681019e-05, + "loss": 0.127, + "step": 1063 + }, + { + "epoch": 3.5945945945945947, + "grad_norm": 0.35974109172821045, + "learning_rate": 3.650297578068448e-05, + "loss": 0.1238, + "step": 1064 + }, + { + "epoch": 3.597972972972973, + "grad_norm": 0.4197885990142822, + "learning_rate": 3.6122364850157326e-05, + "loss": 0.1342, + "step": 1065 + }, + { + "epoch": 3.6013513513513513, + "grad_norm": 0.3706846237182617, + "learning_rate": 3.573851236714447e-05, + "loss": 0.1108, + "step": 1066 + }, + { + "epoch": 3.60472972972973, + "grad_norm": 0.3801131844520569, + "learning_rate": 3.535153022850684e-05, + "loss": 0.1279, + "step": 1067 + }, + { + "epoch": 3.608108108108108, + "grad_norm": 0.35572677850723267, + "learning_rate": 3.496153124343142e-05, + "loss": 0.1263, + "step": 1068 + }, + { + "epoch": 3.6114864864864864, + "grad_norm": 0.3624805510044098, + "learning_rate": 3.4568629100546333e-05, + "loss": 0.1185, + "step": 1069 + }, + { + "epoch": 3.614864864864865, + "grad_norm": 0.36778637766838074, + "learning_rate": 3.417293833477938e-05, + "loss": 0.1222, + "step": 1070 + }, + { + "epoch": 3.618243243243243, + "grad_norm": 0.34212857484817505, + "learning_rate": 3.377457429397001e-05, + "loss": 0.1196, + "step": 1071 + }, + { + "epoch": 3.6216216216216215, + "grad_norm": 0.3891725540161133, + "learning_rate": 3.337365310524423e-05, + "loss": 0.1291, + "step": 1072 + }, + { + "epoch": 3.625, + "grad_norm": 0.399291068315506, + "learning_rate": 3.2970291641162396e-05, + "loss": 0.1267, + "step": 1073 + }, + { + "epoch": 3.6283783783783785, + "grad_norm": 0.346253901720047, + "learning_rate": 3.25646074856497e-05, + "loss": 0.1207, + "step": 1074 + }, + { + "epoch": 3.631756756756757, + "grad_norm": 0.7178367376327515, + "learning_rate": 3.2156718899719216e-05, + "loss": 0.372, + "step": 1075 + }, + { + "epoch": 3.635135135135135, + "grad_norm": 0.41096293926239014, + "learning_rate": 3.174674478699772e-05, + "loss": 0.1273, + "step": 1076 + }, + { + "epoch": 3.6385135135135136, + "grad_norm": 0.3635697066783905, + "learning_rate": 3.133480465906398e-05, + "loss": 0.1256, + "step": 1077 + }, + { + "epoch": 3.641891891891892, + "grad_norm": 0.3569808006286621, + "learning_rate": 3.092101860061e-05, + "loss": 0.1334, + "step": 1078 + }, + { + "epoch": 3.64527027027027, + "grad_norm": 0.339813768863678, + "learning_rate": 3.0505507234435122e-05, + "loss": 0.111, + "step": 1079 + }, + { + "epoch": 3.6486486486486487, + "grad_norm": 0.4040168225765228, + "learning_rate": 3.0088391686283214e-05, + "loss": 0.1348, + "step": 1080 + }, + { + "epoch": 3.652027027027027, + "grad_norm": 0.37235626578330994, + "learning_rate": 2.966979354953336e-05, + "loss": 0.1262, + "step": 1081 + }, + { + "epoch": 3.6554054054054053, + "grad_norm": 0.3623943030834198, + "learning_rate": 2.92498348497541e-05, + "loss": 0.1185, + "step": 1082 + }, + { + "epoch": 3.6587837837837838, + "grad_norm": 0.36152735352516174, + "learning_rate": 2.882863800913182e-05, + "loss": 0.1248, + "step": 1083 + }, + { + "epoch": 3.6621621621621623, + "grad_norm": 0.38995814323425293, + "learning_rate": 2.8406325810783425e-05, + "loss": 0.1282, + "step": 1084 + }, + { + "epoch": 3.6655405405405403, + "grad_norm": 0.35114696621894836, + "learning_rate": 2.798302136296379e-05, + "loss": 0.1098, + "step": 1085 + }, + { + "epoch": 3.668918918918919, + "grad_norm": 0.35527390241622925, + "learning_rate": 2.7558848063178506e-05, + "loss": 0.1129, + "step": 1086 + }, + { + "epoch": 3.6722972972972974, + "grad_norm": 0.3661573827266693, + "learning_rate": 2.713392956221225e-05, + "loss": 0.1187, + "step": 1087 + }, + { + "epoch": 3.6756756756756754, + "grad_norm": 0.3659035265445709, + "learning_rate": 2.6708389728083372e-05, + "loss": 0.1311, + "step": 1088 + }, + { + "epoch": 3.679054054054054, + "grad_norm": 0.36093148589134216, + "learning_rate": 2.6282352609935028e-05, + "loss": 0.1294, + "step": 1089 + }, + { + "epoch": 3.6824324324324325, + "grad_norm": 0.36720040440559387, + "learning_rate": 2.5855942401873734e-05, + "loss": 0.128, + "step": 1090 + }, + { + "epoch": 3.685810810810811, + "grad_norm": 0.40525415539741516, + "learning_rate": 2.5429283406765436e-05, + "loss": 0.1373, + "step": 1091 + }, + { + "epoch": 3.689189189189189, + "grad_norm": 0.3309617042541504, + "learning_rate": 2.50025e-05, + "loss": 0.115, + "step": 1092 + }, + { + "epoch": 3.6925675675675675, + "grad_norm": 0.3174631893634796, + "learning_rate": 2.4575716593234574e-05, + "loss": 0.105, + "step": 1093 + }, + { + "epoch": 3.695945945945946, + "grad_norm": 0.36353635787963867, + "learning_rate": 2.414905759812627e-05, + "loss": 0.1215, + "step": 1094 + }, + { + "epoch": 3.6993243243243246, + "grad_norm": 0.3367186188697815, + "learning_rate": 2.372264739006497e-05, + "loss": 0.1121, + "step": 1095 + }, + { + "epoch": 3.7027027027027026, + "grad_norm": 0.3455405831336975, + "learning_rate": 2.329661027191664e-05, + "loss": 0.1142, + "step": 1096 + }, + { + "epoch": 3.706081081081081, + "grad_norm": 0.3431136906147003, + "learning_rate": 2.287107043778775e-05, + "loss": 0.1132, + "step": 1097 + }, + { + "epoch": 3.7094594594594597, + "grad_norm": 0.34689489006996155, + "learning_rate": 2.2446151936821504e-05, + "loss": 0.1129, + "step": 1098 + }, + { + "epoch": 3.7128378378378377, + "grad_norm": 0.3749043941497803, + "learning_rate": 2.2021978637036214e-05, + "loss": 0.1268, + "step": 1099 + }, + { + "epoch": 3.7162162162162162, + "grad_norm": 0.34978923201560974, + "learning_rate": 2.1598674189216575e-05, + "loss": 0.1285, + "step": 1100 + }, + { + "epoch": 3.7195945945945947, + "grad_norm": 0.3356480300426483, + "learning_rate": 2.117636199086818e-05, + "loss": 0.116, + "step": 1101 + }, + { + "epoch": 3.722972972972973, + "grad_norm": 0.3172987699508667, + "learning_rate": 2.0755165150245906e-05, + "loss": 0.1137, + "step": 1102 + }, + { + "epoch": 3.7263513513513513, + "grad_norm": 0.34791630506515503, + "learning_rate": 2.0335206450466658e-05, + "loss": 0.1153, + "step": 1103 + }, + { + "epoch": 3.72972972972973, + "grad_norm": 0.3284004330635071, + "learning_rate": 1.9916608313716786e-05, + "loss": 0.1131, + "step": 1104 + }, + { + "epoch": 3.733108108108108, + "grad_norm": 0.3191678822040558, + "learning_rate": 1.9499492765564878e-05, + "loss": 0.1123, + "step": 1105 + }, + { + "epoch": 3.7364864864864864, + "grad_norm": 0.33098334074020386, + "learning_rate": 1.9083981399389997e-05, + "loss": 0.1156, + "step": 1106 + }, + { + "epoch": 3.739864864864865, + "grad_norm": 0.3604113459587097, + "learning_rate": 1.8670195340936026e-05, + "loss": 0.1164, + "step": 1107 + }, + { + "epoch": 3.743243243243243, + "grad_norm": 0.33711400628089905, + "learning_rate": 1.825825521300229e-05, + "loss": 0.1077, + "step": 1108 + }, + { + "epoch": 3.7466216216216215, + "grad_norm": 0.3241155743598938, + "learning_rate": 1.7848281100280788e-05, + "loss": 0.1053, + "step": 1109 + }, + { + "epoch": 3.75, + "grad_norm": 0.33422669768333435, + "learning_rate": 1.7440392514350303e-05, + "loss": 0.1226, + "step": 1110 + }, + { + "epoch": 3.7533783783783785, + "grad_norm": 0.30073583126068115, + "learning_rate": 1.7034708358837607e-05, + "loss": 0.0983, + "step": 1111 + }, + { + "epoch": 3.756756756756757, + "grad_norm": 0.37151581048965454, + "learning_rate": 1.6631346894755773e-05, + "loss": 0.1274, + "step": 1112 + }, + { + "epoch": 3.760135135135135, + "grad_norm": 0.3346865773200989, + "learning_rate": 1.6230425706029996e-05, + "loss": 0.1203, + "step": 1113 + }, + { + "epoch": 3.7635135135135136, + "grad_norm": 0.4401615560054779, + "learning_rate": 1.5832061665220622e-05, + "loss": 0.1122, + "step": 1114 + }, + { + "epoch": 3.766891891891892, + "grad_norm": 0.3216763436794281, + "learning_rate": 1.5436370899453666e-05, + "loss": 0.1139, + "step": 1115 + }, + { + "epoch": 3.77027027027027, + "grad_norm": 0.34183549880981445, + "learning_rate": 1.504346875656858e-05, + "loss": 0.1195, + "step": 1116 + }, + { + "epoch": 3.7736486486486487, + "grad_norm": 0.3108338415622711, + "learning_rate": 1.4653469771493166e-05, + "loss": 0.1018, + "step": 1117 + }, + { + "epoch": 3.777027027027027, + "grad_norm": 0.3019558787345886, + "learning_rate": 1.426648763285553e-05, + "loss": 0.1044, + "step": 1118 + }, + { + "epoch": 3.7804054054054053, + "grad_norm": 0.30197712779045105, + "learning_rate": 1.3882635149842685e-05, + "loss": 0.0976, + "step": 1119 + }, + { + "epoch": 3.7837837837837838, + "grad_norm": 15.628427505493164, + "learning_rate": 1.3502024219315524e-05, + "loss": 0.1159, + "step": 1120 + }, + { + "epoch": 3.7871621621621623, + "grad_norm": 0.32293814420700073, + "learning_rate": 1.3124765793189808e-05, + "loss": 0.108, + "step": 1121 + }, + { + "epoch": 3.7905405405405403, + "grad_norm": 0.31840014457702637, + "learning_rate": 1.2750969846092514e-05, + "loss": 0.1011, + "step": 1122 + }, + { + "epoch": 3.793918918918919, + "grad_norm": 0.33972540497779846, + "learning_rate": 1.2380745343303146e-05, + "loss": 0.115, + "step": 1123 + }, + { + "epoch": 3.7972972972972974, + "grad_norm": 0.3430902063846588, + "learning_rate": 1.2014200208989255e-05, + "loss": 0.1136, + "step": 1124 + }, + { + "epoch": 3.8006756756756754, + "grad_norm": 0.34559592604637146, + "learning_rate": 1.165144129474546e-05, + "loss": 0.1172, + "step": 1125 + }, + { + "epoch": 3.804054054054054, + "grad_norm": 0.33134159445762634, + "learning_rate": 1.1292574348445151e-05, + "loss": 0.1107, + "step": 1126 + }, + { + "epoch": 3.8074324324324325, + "grad_norm": 0.2891589403152466, + "learning_rate": 1.0937703983413898e-05, + "loss": 0.0962, + "step": 1127 + }, + { + "epoch": 3.810810810810811, + "grad_norm": 0.34017154574394226, + "learning_rate": 1.058693364793361e-05, + "loss": 0.1168, + "step": 1128 + }, + { + "epoch": 3.814189189189189, + "grad_norm": 0.3454993963241577, + "learning_rate": 1.0240365595086335e-05, + "loss": 0.1167, + "step": 1129 + }, + { + "epoch": 3.8175675675675675, + "grad_norm": 0.33566930890083313, + "learning_rate": 9.898100852946441e-06, + "loss": 0.1209, + "step": 1130 + }, + { + "epoch": 3.820945945945946, + "grad_norm": 0.31290584802627563, + "learning_rate": 9.560239195129949e-06, + "loss": 0.1062, + "step": 1131 + }, + { + "epoch": 3.8243243243243246, + "grad_norm": 0.33595094084739685, + "learning_rate": 9.226879111709494e-06, + "loss": 0.1091, + "step": 1132 + }, + { + "epoch": 3.8277027027027026, + "grad_norm": 0.3021032512187958, + "learning_rate": 8.898117780503471e-06, + "loss": 0.1029, + "step": 1133 + }, + { + "epoch": 3.831081081081081, + "grad_norm": 0.3187675476074219, + "learning_rate": 8.574051038747738e-06, + "loss": 0.1121, + "step": 1134 + }, + { + "epoch": 3.8344594594594597, + "grad_norm": 0.2892165780067444, + "learning_rate": 8.254773355158088e-06, + "loss": 0.1007, + "step": 1135 + }, + { + "epoch": 3.8378378378378377, + "grad_norm": 0.33364662528038025, + "learning_rate": 7.94037780239169e-06, + "loss": 0.1146, + "step": 1136 + }, + { + "epoch": 3.8412162162162162, + "grad_norm": 0.32407206296920776, + "learning_rate": 7.630956029915455e-06, + "loss": 0.1016, + "step": 1137 + }, + { + "epoch": 3.8445945945945947, + "grad_norm": 0.29210153222084045, + "learning_rate": 7.32659823728928e-06, + "loss": 0.0989, + "step": 1138 + }, + { + "epoch": 3.847972972972973, + "grad_norm": 0.3027746081352234, + "learning_rate": 7.027393147871994e-06, + "loss": 0.1031, + "step": 1139 + }, + { + "epoch": 3.8513513513513513, + "grad_norm": 0.34020087122917175, + "learning_rate": 6.733427982957593e-06, + "loss": 0.1105, + "step": 1140 + }, + { + "epoch": 3.85472972972973, + "grad_norm": 0.3309917747974396, + "learning_rate": 6.444788436349374e-06, + "loss": 0.1155, + "step": 1141 + }, + { + "epoch": 3.858108108108108, + "grad_norm": 0.3191674053668976, + "learning_rate": 6.16155864937929e-06, + "loss": 0.1113, + "step": 1142 + }, + { + "epoch": 3.8614864864864864, + "grad_norm": 0.3304139971733093, + "learning_rate": 5.883821186379955e-06, + "loss": 0.1113, + "step": 1143 + }, + { + "epoch": 3.864864864864865, + "grad_norm": 0.3479062616825104, + "learning_rate": 5.6116570106162695e-06, + "loss": 0.1173, + "step": 1144 + }, + { + "epoch": 3.868243243243243, + "grad_norm": 0.3519810140132904, + "learning_rate": 5.345145460683825e-06, + "loss": 0.1076, + "step": 1145 + }, + { + "epoch": 3.8716216216216215, + "grad_norm": 0.4121153652667999, + "learning_rate": 5.0843642273809276e-06, + "loss": 0.1073, + "step": 1146 + }, + { + "epoch": 3.875, + "grad_norm": 0.3174130618572235, + "learning_rate": 4.82938933106091e-06, + "loss": 0.1016, + "step": 1147 + }, + { + "epoch": 3.8783783783783785, + "grad_norm": 0.2952731251716614, + "learning_rate": 4.5802950994714764e-06, + "loss": 0.1038, + "step": 1148 + }, + { + "epoch": 3.881756756756757, + "grad_norm": 0.304749071598053, + "learning_rate": 4.337154146087377e-06, + "loss": 0.1083, + "step": 1149 + }, + { + "epoch": 3.885135135135135, + "grad_norm": 0.3059677481651306, + "learning_rate": 4.100037348942866e-06, + "loss": 0.099, + "step": 1150 + }, + { + "epoch": 3.8885135135135136, + "grad_norm": 0.3029370605945587, + "learning_rate": 3.869013829970057e-06, + "loss": 0.1017, + "step": 1151 + }, + { + "epoch": 3.891891891891892, + "grad_norm": 0.28835681080818176, + "learning_rate": 3.644150934849158e-06, + "loss": 0.1, + "step": 1152 + }, + { + "epoch": 3.89527027027027, + "grad_norm": 0.2965335249900818, + "learning_rate": 3.425514213376533e-06, + "loss": 0.1023, + "step": 1153 + }, + { + "epoch": 3.8986486486486487, + "grad_norm": 0.2934282422065735, + "learning_rate": 3.213167400356303e-06, + "loss": 0.0937, + "step": 1154 + }, + { + "epoch": 3.902027027027027, + "grad_norm": 0.32294800877571106, + "learning_rate": 3.007172397020972e-06, + "loss": 0.1108, + "step": 1155 + }, + { + "epoch": 3.9054054054054053, + "grad_norm": 0.2884935438632965, + "learning_rate": 2.8075892529866257e-06, + "loss": 0.1025, + "step": 1156 + }, + { + "epoch": 3.9087837837837838, + "grad_norm": 0.31093308329582214, + "learning_rate": 2.6144761487478416e-06, + "loss": 0.107, + "step": 1157 + }, + { + "epoch": 3.9121621621621623, + "grad_norm": 0.3277430236339569, + "learning_rate": 2.4278893787174935e-06, + "loss": 0.1216, + "step": 1158 + }, + { + "epoch": 3.9155405405405403, + "grad_norm": 0.467430979013443, + "learning_rate": 2.2478833348163886e-06, + "loss": 0.3528, + "step": 1159 + }, + { + "epoch": 3.918918918918919, + "grad_norm": 0.2848818302154541, + "learning_rate": 2.074510490617438e-06, + "loss": 0.0981, + "step": 1160 + }, + { + "epoch": 3.9222972972972974, + "grad_norm": 0.2947328984737396, + "learning_rate": 1.9078213860491097e-06, + "loss": 0.1003, + "step": 1161 + }, + { + "epoch": 3.9256756756756754, + "grad_norm": 0.28310471773147583, + "learning_rate": 1.7478646126625187e-06, + "loss": 0.096, + "step": 1162 + }, + { + "epoch": 3.929054054054054, + "grad_norm": 0.31671369075775146, + "learning_rate": 1.5946867994665007e-06, + "loss": 0.1062, + "step": 1163 + }, + { + "epoch": 3.9324324324324325, + "grad_norm": 0.2964227795600891, + "learning_rate": 1.4483325993347872e-06, + "loss": 0.102, + "step": 1164 + }, + { + "epoch": 3.935810810810811, + "grad_norm": 0.309048593044281, + "learning_rate": 1.308844675989258e-06, + "loss": 0.1049, + "step": 1165 + }, + { + "epoch": 3.939189189189189, + "grad_norm": 0.31580665707588196, + "learning_rate": 1.1762636915630122e-06, + "loss": 0.1019, + "step": 1166 + }, + { + "epoch": 3.9425675675675675, + "grad_norm": 0.27929919958114624, + "learning_rate": 1.0506282947469768e-06, + "loss": 0.0962, + "step": 1167 + }, + { + "epoch": 3.945945945945946, + "grad_norm": 0.44457200169563293, + "learning_rate": 9.319751095233885e-07, + "loss": 0.3502, + "step": 1168 + }, + { + "epoch": 3.9493243243243246, + "grad_norm": 0.26814886927604675, + "learning_rate": 8.203387244895631e-07, + "loss": 0.0968, + "step": 1169 + }, + { + "epoch": 3.9527027027027026, + "grad_norm": 0.31180819869041443, + "learning_rate": 7.157516827749585e-07, + "loss": 0.1097, + "step": 1170 + }, + { + "epoch": 3.956081081081081, + "grad_norm": 0.30869176983833313, + "learning_rate": 6.182444725545342e-07, + "loss": 0.1067, + "step": 1171 + }, + { + "epoch": 3.9594594594594597, + "grad_norm": 0.29772305488586426, + "learning_rate": 5.278455181611357e-07, + "loss": 0.1061, + "step": 1172 + }, + { + "epoch": 3.9628378378378377, + "grad_norm": 0.3115202784538269, + "learning_rate": 4.445811717995063e-07, + "loss": 0.1081, + "step": 1173 + }, + { + "epoch": 3.9662162162162162, + "grad_norm": 0.31144705414772034, + "learning_rate": 3.684757058643714e-07, + "loss": 0.097, + "step": 1174 + }, + { + "epoch": 3.9695945945945947, + "grad_norm": 0.3056892156600952, + "learning_rate": 2.995513058647855e-07, + "loss": 0.1069, + "step": 1175 + }, + { + "epoch": 3.972972972972973, + "grad_norm": 0.327035129070282, + "learning_rate": 2.3782806395682474e-07, + "loss": 0.1168, + "step": 1176 + }, + { + "epoch": 3.9763513513513513, + "grad_norm": 0.31039026379585266, + "learning_rate": 1.8332397308652485e-07, + "loss": 0.105, + "step": 1177 + }, + { + "epoch": 3.97972972972973, + "grad_norm": 0.2924991548061371, + "learning_rate": 1.3605492174477425e-07, + "loss": 0.0984, + "step": 1178 + }, + { + "epoch": 3.983108108108108, + "grad_norm": 0.2952897250652313, + "learning_rate": 9.603468933562955e-08, + "loss": 0.1039, + "step": 1179 + }, + { + "epoch": 3.9864864864864864, + "grad_norm": 0.28377214074134827, + "learning_rate": 6.32749421594948e-08, + "loss": 0.098, + "step": 1180 + }, + { + "epoch": 3.989864864864865, + "grad_norm": 0.3225543200969696, + "learning_rate": 3.778523001229054e-08, + "loss": 0.1196, + "step": 1181 + }, + { + "epoch": 3.993243243243243, + "grad_norm": 0.3456767201423645, + "learning_rate": 1.957298340156484e-08, + "loss": 0.1071, + "step": 1182 + }, + { + "epoch": 3.9966216216216215, + "grad_norm": 0.30064505338668823, + "learning_rate": 8.64351138044836e-09, + "loss": 0.1043, + "step": 1183 + }, + { + "epoch": 4.0, + "grad_norm": 0.30672967433929443, + "learning_rate": 5e-09, + "loss": 0.0813, + "step": 1184 + }, + { + "epoch": 4.0, + "step": 1184, + "total_flos": 1.048251868267099e+19, + "train_loss": 0.24899632067771982, + "train_runtime": 10196.9918, + "train_samples_per_second": 2.78, + "train_steps_per_second": 0.116 + } + ], + "logging_steps": 1.0, + "max_steps": 1184, + "num_input_tokens_seen": 0, + "num_train_epochs": 4, + "save_steps": 250, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 1.048251868267099e+19, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}