{ "best_metric": null, "best_model_checkpoint": null, "epoch": 4.0, "eval_steps": 500, "global_step": 1184, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0033783783783783786, "grad_norm": 134.4763946533203, "learning_rate": 0.0, "loss": 4.6817, "step": 1 }, { "epoch": 0.006756756756756757, "grad_norm": 125.12445068359375, "learning_rate": 5.017166594399687e-06, "loss": 4.5642, "step": 2 }, { "epoch": 0.010135135135135136, "grad_norm": 133.296142578125, "learning_rate": 7.952020911994375e-06, "loss": 4.787, "step": 3 }, { "epoch": 0.013513513513513514, "grad_norm": 72.2229995727539, "learning_rate": 1.0034333188799373e-05, "loss": 3.1422, "step": 4 }, { "epoch": 0.016891891891891893, "grad_norm": 83.57042694091797, "learning_rate": 1.164950007226698e-05, "loss": 1.9827, "step": 5 }, { "epoch": 0.02027027027027027, "grad_norm": 35.78062057495117, "learning_rate": 1.2969187506394062e-05, "loss": 1.0212, "step": 6 }, { "epoch": 0.02364864864864865, "grad_norm": 48.31964111328125, "learning_rate": 1.4084967333570947e-05, "loss": 1.1823, "step": 7 }, { "epoch": 0.02702702702702703, "grad_norm": 10.594072341918945, "learning_rate": 1.505149978319906e-05, "loss": 0.7368, "step": 8 }, { "epoch": 0.030405405405405407, "grad_norm": 7.897088050842285, "learning_rate": 1.590404182398875e-05, "loss": 0.6276, "step": 9 }, { "epoch": 0.033783783783783786, "grad_norm": 3.8260443210601807, "learning_rate": 1.666666666666667e-05, "loss": 0.5934, "step": 10 }, { "epoch": 0.037162162162162164, "grad_norm": 11.113093376159668, "learning_rate": 1.7356544752637084e-05, "loss": 0.6987, "step": 11 }, { "epoch": 0.04054054054054054, "grad_norm": 3.2817893028259277, "learning_rate": 1.7986354100793748e-05, "loss": 0.5722, "step": 12 }, { "epoch": 0.04391891891891892, "grad_norm": 1.4319814443588257, "learning_rate": 1.8565722538447282e-05, "loss": 0.5199, "step": 13 }, { "epoch": 0.0472972972972973, "grad_norm": 1.1898471117019653, "learning_rate": 1.9102133927970633e-05, "loss": 0.5063, "step": 14 }, { "epoch": 0.05067567567567568, "grad_norm": 2.1912143230438232, "learning_rate": 1.9601520984261358e-05, "loss": 0.5368, "step": 15 }, { "epoch": 0.05405405405405406, "grad_norm": 1.0015403032302856, "learning_rate": 2.0068666377598747e-05, "loss": 0.4682, "step": 16 }, { "epoch": 0.057432432432432436, "grad_norm": 1.114931583404541, "learning_rate": 2.0507482022971233e-05, "loss": 0.4997, "step": 17 }, { "epoch": 0.060810810810810814, "grad_norm": 1.0931981801986694, "learning_rate": 2.0921208418388435e-05, "loss": 0.4659, "step": 18 }, { "epoch": 0.06418918918918919, "grad_norm": 0.8113773465156555, "learning_rate": 2.1312560015880482e-05, "loss": 0.4697, "step": 19 }, { "epoch": 0.06756756756756757, "grad_norm": 1.0267772674560547, "learning_rate": 2.1683833261066357e-05, "loss": 0.4851, "step": 20 }, { "epoch": 0.07094594594594594, "grad_norm": 0.8238614201545715, "learning_rate": 2.2036988245565324e-05, "loss": 0.4517, "step": 21 }, { "epoch": 0.07432432432432433, "grad_norm": 0.9366945624351501, "learning_rate": 2.2373711347036773e-05, "loss": 0.4982, "step": 22 }, { "epoch": 0.0777027027027027, "grad_norm": 0.8443475365638733, "learning_rate": 2.269546393362655e-05, "loss": 0.4128, "step": 23 }, { "epoch": 0.08108108108108109, "grad_norm": 0.8733354806900024, "learning_rate": 2.3003520695193437e-05, "loss": 0.4281, "step": 24 }, { "epoch": 0.08445945945945946, "grad_norm": 0.7766995429992676, "learning_rate": 2.329900014453396e-05, "loss": 0.4407, "step": 25 }, { "epoch": 0.08783783783783784, "grad_norm": 0.8224064111709595, "learning_rate": 2.3582889132846968e-05, "loss": 0.4155, "step": 26 }, { "epoch": 0.09121621621621621, "grad_norm": 0.7931010127067566, "learning_rate": 2.3856062735983123e-05, "loss": 0.4191, "step": 27 }, { "epoch": 0.0945945945945946, "grad_norm": 0.6803894639015198, "learning_rate": 2.4119300522370322e-05, "loss": 0.4055, "step": 28 }, { "epoch": 0.09797297297297297, "grad_norm": 0.748294472694397, "learning_rate": 2.4373299964982603e-05, "loss": 0.4443, "step": 29 }, { "epoch": 0.10135135135135136, "grad_norm": 0.706881582736969, "learning_rate": 2.4618687578661044e-05, "loss": 0.4093, "step": 30 }, { "epoch": 0.10472972972972973, "grad_norm": 0.7517805695533752, "learning_rate": 2.4856028230571212e-05, "loss": 0.4288, "step": 31 }, { "epoch": 0.10810810810810811, "grad_norm": 0.6549198031425476, "learning_rate": 2.5085832971998436e-05, "loss": 0.4222, "step": 32 }, { "epoch": 0.11148648648648649, "grad_norm": 0.9364942908287048, "learning_rate": 2.530856566463146e-05, "loss": 0.416, "step": 33 }, { "epoch": 0.11486486486486487, "grad_norm": 0.624920129776001, "learning_rate": 2.552464861737092e-05, "loss": 0.3735, "step": 34 }, { "epoch": 0.11824324324324324, "grad_norm": 0.6083908081054688, "learning_rate": 2.5734467405837933e-05, "loss": 0.3315, "step": 35 }, { "epoch": 0.12162162162162163, "grad_norm": 0.6106983423233032, "learning_rate": 2.5938375012788124e-05, "loss": 0.3652, "step": 36 }, { "epoch": 0.125, "grad_norm": 0.7572540044784546, "learning_rate": 2.6136695401116585e-05, "loss": 0.452, "step": 37 }, { "epoch": 0.12837837837837837, "grad_norm": 0.6654285788536072, "learning_rate": 2.6329726610280168e-05, "loss": 0.3328, "step": 38 }, { "epoch": 0.13175675675675674, "grad_norm": 0.676705002784729, "learning_rate": 2.651774345044166e-05, "loss": 0.3987, "step": 39 }, { "epoch": 0.13513513513513514, "grad_norm": 0.6630686521530151, "learning_rate": 2.6700999855466042e-05, "loss": 0.3794, "step": 40 }, { "epoch": 0.13851351351351351, "grad_norm": 0.8196331858634949, "learning_rate": 2.687973094532893e-05, "loss": 0.4218, "step": 41 }, { "epoch": 0.14189189189189189, "grad_norm": 0.6006856560707092, "learning_rate": 2.7054154839965013e-05, "loss": 0.3788, "step": 42 }, { "epoch": 0.14527027027027026, "grad_norm": 0.6180748343467712, "learning_rate": 2.722447425965978e-05, "loss": 0.3946, "step": 43 }, { "epoch": 0.14864864864864866, "grad_norm": 0.7010207772254944, "learning_rate": 2.739087794143646e-05, "loss": 0.389, "step": 44 }, { "epoch": 0.15202702702702703, "grad_norm": 0.681907594203949, "learning_rate": 2.755354189625573e-05, "loss": 0.399, "step": 45 }, { "epoch": 0.1554054054054054, "grad_norm": 0.6506620645523071, "learning_rate": 2.771263052802624e-05, "loss": 0.3924, "step": 46 }, { "epoch": 0.15878378378378377, "grad_norm": 0.6619815230369568, "learning_rate": 2.7868297632261957e-05, "loss": 0.4111, "step": 47 }, { "epoch": 0.16216216216216217, "grad_norm": 0.6953612565994263, "learning_rate": 2.8020687289593123e-05, "loss": 0.3958, "step": 48 }, { "epoch": 0.16554054054054054, "grad_norm": 0.761862576007843, "learning_rate": 2.8169934667141895e-05, "loss": 0.3844, "step": 49 }, { "epoch": 0.16891891891891891, "grad_norm": 0.7193901538848877, "learning_rate": 2.8316166738933646e-05, "loss": 0.3847, "step": 50 }, { "epoch": 0.17229729729729729, "grad_norm": 0.6792317032814026, "learning_rate": 2.845950293496561e-05, "loss": 0.4043, "step": 51 }, { "epoch": 0.17567567567567569, "grad_norm": 0.6307753920555115, "learning_rate": 2.8600055727246657e-05, "loss": 0.3806, "step": 52 }, { "epoch": 0.17905405405405406, "grad_norm": 0.7366315722465515, "learning_rate": 2.8737931160013153e-05, "loss": 0.416, "step": 53 }, { "epoch": 0.18243243243243243, "grad_norm": 0.603863537311554, "learning_rate": 2.8873229330382812e-05, "loss": 0.3548, "step": 54 }, { "epoch": 0.1858108108108108, "grad_norm": 0.7424588799476624, "learning_rate": 2.9006044824904066e-05, "loss": 0.3982, "step": 55 }, { "epoch": 0.1891891891891892, "grad_norm": 0.5944585204124451, "learning_rate": 2.913646711677001e-05, "loss": 0.3536, "step": 56 }, { "epoch": 0.19256756756756757, "grad_norm": 0.6465046405792236, "learning_rate": 2.926458092787486e-05, "loss": 0.3594, "step": 57 }, { "epoch": 0.19594594594594594, "grad_norm": 0.7178723812103271, "learning_rate": 2.939046655938229e-05, "loss": 0.3725, "step": 58 }, { "epoch": 0.19932432432432431, "grad_norm": 0.7240431308746338, "learning_rate": 2.951420019403574e-05, "loss": 0.3896, "step": 59 }, { "epoch": 0.20270270270270271, "grad_norm": 0.7555009126663208, "learning_rate": 2.963585417306073e-05, "loss": 0.3633, "step": 60 }, { "epoch": 0.20608108108108109, "grad_norm": 0.6279475092887878, "learning_rate": 2.9755497250179453e-05, "loss": 0.3447, "step": 61 }, { "epoch": 0.20945945945945946, "grad_norm": 0.6943121552467346, "learning_rate": 2.98731948249709e-05, "loss": 0.3851, "step": 62 }, { "epoch": 0.21283783783783783, "grad_norm": 0.6565203070640564, "learning_rate": 2.9989009157559694e-05, "loss": 0.3896, "step": 63 }, { "epoch": 0.21621621621621623, "grad_norm": 4.912997722625732, "learning_rate": 3.010299956639812e-05, "loss": 0.4528, "step": 64 }, { "epoch": 0.2195945945945946, "grad_norm": 0.7776851058006287, "learning_rate": 3.021522261071426e-05, "loss": 0.3151, "step": 65 }, { "epoch": 0.22297297297297297, "grad_norm": 0.5963460206985474, "learning_rate": 3.0325732259031143e-05, "loss": 0.3501, "step": 66 }, { "epoch": 0.22635135135135134, "grad_norm": 0.8160498142242432, "learning_rate": 3.043458004501377e-05, "loss": 0.3383, "step": 67 }, { "epoch": 0.22972972972972974, "grad_norm": 0.6507856249809265, "learning_rate": 3.054181521177061e-05, "loss": 0.3559, "step": 68 }, { "epoch": 0.23310810810810811, "grad_norm": 0.6282461881637573, "learning_rate": 3.064748484562093e-05, "loss": 0.3951, "step": 69 }, { "epoch": 0.23648648648648649, "grad_norm": 0.6464638710021973, "learning_rate": 3.0751634000237615e-05, "loss": 0.3759, "step": 70 }, { "epoch": 0.23986486486486486, "grad_norm": 0.621934175491333, "learning_rate": 3.085430581198459e-05, "loss": 0.409, "step": 71 }, { "epoch": 0.24324324324324326, "grad_norm": 0.6526879668235779, "learning_rate": 3.095554160718781e-05, "loss": 0.3938, "step": 72 }, { "epoch": 0.24662162162162163, "grad_norm": 0.6536738276481628, "learning_rate": 3.10553810020076e-05, "loss": 0.3657, "step": 73 }, { "epoch": 0.25, "grad_norm": 0.6381665468215942, "learning_rate": 3.115386199551628e-05, "loss": 0.3743, "step": 74 }, { "epoch": 0.2533783783783784, "grad_norm": 0.6818379759788513, "learning_rate": 3.1251021056528336e-05, "loss": 0.3188, "step": 75 }, { "epoch": 0.25675675675675674, "grad_norm": 0.5829269886016846, "learning_rate": 3.134689320467986e-05, "loss": 0.3622, "step": 76 }, { "epoch": 0.26013513513513514, "grad_norm": 0.6988232731819153, "learning_rate": 3.144151208620804e-05, "loss": 0.3768, "step": 77 }, { "epoch": 0.2635135135135135, "grad_norm": 0.5981537699699402, "learning_rate": 3.1534910044841344e-05, "loss": 0.3655, "step": 78 }, { "epoch": 0.2668918918918919, "grad_norm": 0.5821400284767151, "learning_rate": 3.1627118188174024e-05, "loss": 0.358, "step": 79 }, { "epoch": 0.2702702702702703, "grad_norm": 0.620126485824585, "learning_rate": 3.171816644986573e-05, "loss": 0.3579, "step": 80 }, { "epoch": 0.27364864864864863, "grad_norm": 0.6327139139175415, "learning_rate": 3.18080836479775e-05, "loss": 0.3916, "step": 81 }, { "epoch": 0.27702702702702703, "grad_norm": 0.5900242328643799, "learning_rate": 3.1896897539728616e-05, "loss": 0.3768, "step": 82 }, { "epoch": 0.28040540540540543, "grad_norm": 0.5920627117156982, "learning_rate": 3.198463487293457e-05, "loss": 0.3709, "step": 83 }, { "epoch": 0.28378378378378377, "grad_norm": 0.5604241490364075, "learning_rate": 3.207132143436469e-05, "loss": 0.3871, "step": 84 }, { "epoch": 0.28716216216216217, "grad_norm": 0.6093663573265076, "learning_rate": 3.215698209523821e-05, "loss": 0.3632, "step": 85 }, { "epoch": 0.2905405405405405, "grad_norm": 0.568435549736023, "learning_rate": 3.224164085405946e-05, "loss": 0.3788, "step": 86 }, { "epoch": 0.2939189189189189, "grad_norm": 0.5885617136955261, "learning_rate": 3.232532087697698e-05, "loss": 0.3472, "step": 87 }, { "epoch": 0.2972972972972973, "grad_norm": 0.5474864840507507, "learning_rate": 3.240804453583615e-05, "loss": 0.3815, "step": 88 }, { "epoch": 0.30067567567567566, "grad_norm": 0.6116411089897156, "learning_rate": 3.248983344408188e-05, "loss": 0.3374, "step": 89 }, { "epoch": 0.30405405405405406, "grad_norm": 0.5523970127105713, "learning_rate": 3.2570708490655414e-05, "loss": 0.3503, "step": 90 }, { "epoch": 0.30743243243243246, "grad_norm": 0.6035043597221375, "learning_rate": 3.265068987201822e-05, "loss": 0.3699, "step": 91 }, { "epoch": 0.3108108108108108, "grad_norm": 0.4965035617351532, "learning_rate": 3.2729797122425925e-05, "loss": 0.3702, "step": 92 }, { "epoch": 0.3141891891891892, "grad_norm": 0.5582573413848877, "learning_rate": 3.280804914256559e-05, "loss": 0.333, "step": 93 }, { "epoch": 0.31756756756756754, "grad_norm": 0.5313113331794739, "learning_rate": 3.288546422666164e-05, "loss": 0.329, "step": 94 }, { "epoch": 0.32094594594594594, "grad_norm": 0.6049978733062744, "learning_rate": 3.2962060088147464e-05, "loss": 0.3931, "step": 95 }, { "epoch": 0.32432432432432434, "grad_norm": 0.5387487411499023, "learning_rate": 3.3037853883992805e-05, "loss": 0.3432, "step": 96 }, { "epoch": 0.3277027027027027, "grad_norm": 0.588671088218689, "learning_rate": 3.3112862237770756e-05, "loss": 0.3508, "step": 97 }, { "epoch": 0.3310810810810811, "grad_norm": 0.536973774433136, "learning_rate": 3.3187101261541584e-05, "loss": 0.372, "step": 98 }, { "epoch": 0.3344594594594595, "grad_norm": 0.676276445388794, "learning_rate": 3.326058657662584e-05, "loss": 0.3832, "step": 99 }, { "epoch": 0.33783783783783783, "grad_norm": 0.46296200156211853, "learning_rate": 3.333333333333334e-05, "loss": 0.3553, "step": 100 }, { "epoch": 0.34121621621621623, "grad_norm": 0.674472451210022, "learning_rate": 3.340535622971072e-05, "loss": 0.3448, "step": 101 }, { "epoch": 0.34459459459459457, "grad_norm": 0.5100315809249878, "learning_rate": 3.3476669529365295e-05, "loss": 0.3367, "step": 102 }, { "epoch": 0.34797297297297297, "grad_norm": 0.5637315511703491, "learning_rate": 3.3547287078419544e-05, "loss": 0.3786, "step": 103 }, { "epoch": 0.35135135135135137, "grad_norm": 0.5659502744674683, "learning_rate": 3.361722232164634e-05, "loss": 0.352, "step": 104 }, { "epoch": 0.3547297297297297, "grad_norm": 0.5390239357948303, "learning_rate": 3.3686488317832306e-05, "loss": 0.3473, "step": 105 }, { "epoch": 0.3581081081081081, "grad_norm": 0.5552096366882324, "learning_rate": 3.375509775441284e-05, "loss": 0.2945, "step": 106 }, { "epoch": 0.3614864864864865, "grad_norm": 0.5016259551048279, "learning_rate": 3.382306296142016e-05, "loss": 0.313, "step": 107 }, { "epoch": 0.36486486486486486, "grad_norm": 0.5755091905593872, "learning_rate": 3.38903959247825e-05, "loss": 0.3367, "step": 108 }, { "epoch": 0.36824324324324326, "grad_norm": 0.5590441823005676, "learning_rate": 3.395710829901039e-05, "loss": 0.3639, "step": 109 }, { "epoch": 0.3716216216216216, "grad_norm": 0.5077652335166931, "learning_rate": 3.402321141930376e-05, "loss": 0.3523, "step": 110 }, { "epoch": 0.375, "grad_norm": 0.5614023804664612, "learning_rate": 3.4088716313110955e-05, "loss": 0.3418, "step": 111 }, { "epoch": 0.3783783783783784, "grad_norm": 0.5325702428817749, "learning_rate": 3.415363371116969e-05, "loss": 0.3208, "step": 112 }, { "epoch": 0.38175675675675674, "grad_norm": 0.5361766219139099, "learning_rate": 3.4217974058057e-05, "loss": 0.3412, "step": 113 }, { "epoch": 0.38513513513513514, "grad_norm": 0.5173611044883728, "learning_rate": 3.428174752227455e-05, "loss": 0.3615, "step": 114 }, { "epoch": 0.3885135135135135, "grad_norm": 0.5852875709533691, "learning_rate": 3.434496400589353e-05, "loss": 0.3171, "step": 115 }, { "epoch": 0.3918918918918919, "grad_norm": 0.48835834860801697, "learning_rate": 3.440763315378198e-05, "loss": 0.315, "step": 116 }, { "epoch": 0.3952702702702703, "grad_norm": 0.5958307385444641, "learning_rate": 3.446976436243603e-05, "loss": 0.3117, "step": 117 }, { "epoch": 0.39864864864864863, "grad_norm": 0.5645899772644043, "learning_rate": 3.4531366788435425e-05, "loss": 0.3192, "step": 118 }, { "epoch": 0.40202702702702703, "grad_norm": 0.5764517784118652, "learning_rate": 3.459244935654219e-05, "loss": 0.3353, "step": 119 }, { "epoch": 0.40540540540540543, "grad_norm": 0.5075375437736511, "learning_rate": 3.465302076746041e-05, "loss": 0.3572, "step": 120 }, { "epoch": 0.40878378378378377, "grad_norm": 0.48014962673187256, "learning_rate": 3.471308950527417e-05, "loss": 0.3217, "step": 121 }, { "epoch": 0.41216216216216217, "grad_norm": 0.533892035484314, "learning_rate": 3.477266384457914e-05, "loss": 0.3214, "step": 122 }, { "epoch": 0.4155405405405405, "grad_norm": 0.4845181405544281, "learning_rate": 3.48317518573233e-05, "loss": 0.3332, "step": 123 }, { "epoch": 0.4189189189189189, "grad_norm": 0.5133293271064758, "learning_rate": 3.489036141937059e-05, "loss": 0.354, "step": 124 }, { "epoch": 0.4222972972972973, "grad_norm": 0.5352755188941956, "learning_rate": 3.494850021680094e-05, "loss": 0.353, "step": 125 }, { "epoch": 0.42567567567567566, "grad_norm": 0.5208103060722351, "learning_rate": 3.500617575195938e-05, "loss": 0.3065, "step": 126 }, { "epoch": 0.42905405405405406, "grad_norm": 0.5982815623283386, "learning_rate": 3.5063395349265945e-05, "loss": 0.3329, "step": 127 }, { "epoch": 0.43243243243243246, "grad_norm": 0.5241547226905823, "learning_rate": 3.5120166160797804e-05, "loss": 0.3476, "step": 128 }, { "epoch": 0.4358108108108108, "grad_norm": 0.5543828010559082, "learning_rate": 3.517649517165415e-05, "loss": 0.3517, "step": 129 }, { "epoch": 0.4391891891891892, "grad_norm": 0.5167660117149353, "learning_rate": 3.523238920511395e-05, "loss": 0.3417, "step": 130 }, { "epoch": 0.44256756756756754, "grad_norm": 0.5201629400253296, "learning_rate": 3.528785492759607e-05, "loss": 0.3543, "step": 131 }, { "epoch": 0.44594594594594594, "grad_norm": 0.5214123129844666, "learning_rate": 3.5342898853430836e-05, "loss": 0.3552, "step": 132 }, { "epoch": 0.44932432432432434, "grad_norm": 0.526899516582489, "learning_rate": 3.539752734945143e-05, "loss": 0.3469, "step": 133 }, { "epoch": 0.4527027027027027, "grad_norm": 0.55535888671875, "learning_rate": 3.5451746639413466e-05, "loss": 0.294, "step": 134 }, { "epoch": 0.4560810810810811, "grad_norm": 0.46593329310417175, "learning_rate": 3.550556280825011e-05, "loss": 0.2557, "step": 135 }, { "epoch": 0.4594594594594595, "grad_norm": 0.5066341757774353, "learning_rate": 3.55589818061703e-05, "loss": 0.2885, "step": 136 }, { "epoch": 0.46283783783783783, "grad_norm": 0.4943198263645172, "learning_rate": 3.561200945260678e-05, "loss": 0.3436, "step": 137 }, { "epoch": 0.46621621621621623, "grad_norm": 0.5013923048973083, "learning_rate": 3.5664651440020616e-05, "loss": 0.2865, "step": 138 }, { "epoch": 0.46959459459459457, "grad_norm": 0.5165452361106873, "learning_rate": 3.571691333756825e-05, "loss": 0.3394, "step": 139 }, { "epoch": 0.47297297297297297, "grad_norm": 0.5002908706665039, "learning_rate": 3.5768800594637304e-05, "loss": 0.3278, "step": 140 }, { "epoch": 0.47635135135135137, "grad_norm": 0.5236304402351379, "learning_rate": 3.582031854425634e-05, "loss": 0.3343, "step": 141 }, { "epoch": 0.4797297297297297, "grad_norm": 0.49403268098831177, "learning_rate": 3.587147240638428e-05, "loss": 0.3079, "step": 142 }, { "epoch": 0.4831081081081081, "grad_norm": 0.4847567677497864, "learning_rate": 3.5922267291084366e-05, "loss": 0.3305, "step": 143 }, { "epoch": 0.4864864864864865, "grad_norm": 0.5461480617523193, "learning_rate": 3.5972708201587496e-05, "loss": 0.3242, "step": 144 }, { "epoch": 0.48986486486486486, "grad_norm": 0.5380986332893372, "learning_rate": 3.6022800037249585e-05, "loss": 0.3341, "step": 145 }, { "epoch": 0.49324324324324326, "grad_norm": 0.49580562114715576, "learning_rate": 3.607254759640729e-05, "loss": 0.31, "step": 146 }, { "epoch": 0.4966216216216216, "grad_norm": 0.5374506115913391, "learning_rate": 3.612195557913627e-05, "loss": 0.3048, "step": 147 }, { "epoch": 0.5, "grad_norm": 0.535591721534729, "learning_rate": 3.6171028589915954e-05, "loss": 0.3266, "step": 148 }, { "epoch": 0.5033783783783784, "grad_norm": 0.5777239799499512, "learning_rate": 3.6219771140204575e-05, "loss": 0.3509, "step": 149 }, { "epoch": 0.5067567567567568, "grad_norm": 0.5570788383483887, "learning_rate": 3.626818765092802e-05, "loss": 0.3348, "step": 150 }, { "epoch": 0.5101351351351351, "grad_norm": 0.5780688524246216, "learning_rate": 3.6316282454886157e-05, "loss": 0.3218, "step": 151 }, { "epoch": 0.5135135135135135, "grad_norm": 0.5464823842048645, "learning_rate": 3.636405979907955e-05, "loss": 0.347, "step": 152 }, { "epoch": 0.5168918918918919, "grad_norm": 0.46229949593544006, "learning_rate": 3.6411523846959985e-05, "loss": 0.2618, "step": 153 }, { "epoch": 0.5202702702702703, "grad_norm": 0.5127717852592468, "learning_rate": 3.645867868060772e-05, "loss": 0.3364, "step": 154 }, { "epoch": 0.5236486486486487, "grad_norm": 0.4915751814842224, "learning_rate": 3.6505528302838193e-05, "loss": 0.327, "step": 155 }, { "epoch": 0.527027027027027, "grad_norm": 0.4658984839916229, "learning_rate": 3.6552076639241027e-05, "loss": 0.3008, "step": 156 }, { "epoch": 0.5304054054054054, "grad_norm": 0.4983694553375244, "learning_rate": 3.65983275401539e-05, "loss": 0.3153, "step": 157 }, { "epoch": 0.5337837837837838, "grad_norm": 0.49560025334358215, "learning_rate": 3.664428478257371e-05, "loss": 0.3409, "step": 158 }, { "epoch": 0.5371621621621622, "grad_norm": 0.5617703199386597, "learning_rate": 3.668995207200753e-05, "loss": 0.3295, "step": 159 }, { "epoch": 0.5405405405405406, "grad_norm": 0.5226999521255493, "learning_rate": 3.673533304426541e-05, "loss": 0.3522, "step": 160 }, { "epoch": 0.543918918918919, "grad_norm": 0.49685564637184143, "learning_rate": 3.67804312671975e-05, "loss": 0.3322, "step": 161 }, { "epoch": 0.5472972972972973, "grad_norm": 0.543743371963501, "learning_rate": 3.682525024237719e-05, "loss": 0.3283, "step": 162 }, { "epoch": 0.5506756756756757, "grad_norm": 0.5247477889060974, "learning_rate": 3.6869793406732636e-05, "loss": 0.3104, "step": 163 }, { "epoch": 0.5540540540540541, "grad_norm": 0.5228151679039001, "learning_rate": 3.69140641341283e-05, "loss": 0.3292, "step": 164 }, { "epoch": 0.5574324324324325, "grad_norm": 0.48610207438468933, "learning_rate": 3.695806573689844e-05, "loss": 0.3183, "step": 165 }, { "epoch": 0.5608108108108109, "grad_norm": 0.4703191816806793, "learning_rate": 3.700180146733426e-05, "loss": 0.2935, "step": 166 }, { "epoch": 0.5641891891891891, "grad_norm": 0.5501482486724854, "learning_rate": 3.704527451912639e-05, "loss": 0.3249, "step": 167 }, { "epoch": 0.5675675675675675, "grad_norm": 0.45025497674942017, "learning_rate": 3.708848802876438e-05, "loss": 0.3128, "step": 168 }, { "epoch": 0.5709459459459459, "grad_norm": 0.5103681087493896, "learning_rate": 3.7131445076894564e-05, "loss": 0.2882, "step": 169 }, { "epoch": 0.5743243243243243, "grad_norm": 0.4952690601348877, "learning_rate": 3.717414868963791e-05, "loss": 0.3268, "step": 170 }, { "epoch": 0.5777027027027027, "grad_norm": 0.4887889325618744, "learning_rate": 3.721660183986924e-05, "loss": 0.3161, "step": 171 }, { "epoch": 0.581081081081081, "grad_norm": 0.5142275094985962, "learning_rate": 3.725880744845915e-05, "loss": 0.2942, "step": 172 }, { "epoch": 0.5844594594594594, "grad_norm": 0.518671452999115, "learning_rate": 3.730076838547993e-05, "loss": 0.3232, "step": 173 }, { "epoch": 0.5878378378378378, "grad_norm": 0.5324185490608215, "learning_rate": 3.734248747137666e-05, "loss": 0.3115, "step": 174 }, { "epoch": 0.5912162162162162, "grad_norm": 0.5307193398475647, "learning_rate": 3.738396747810492e-05, "loss": 0.3245, "step": 175 }, { "epoch": 0.5945945945945946, "grad_norm": 0.47318169474601746, "learning_rate": 3.7425211130235834e-05, "loss": 0.3152, "step": 176 }, { "epoch": 0.597972972972973, "grad_norm": 0.5233657956123352, "learning_rate": 3.7466221106030115e-05, "loss": 0.3115, "step": 177 }, { "epoch": 0.6013513513513513, "grad_norm": 0.4699764847755432, "learning_rate": 3.750700003848157e-05, "loss": 0.2948, "step": 178 }, { "epoch": 0.6047297297297297, "grad_norm": 0.5511758327484131, "learning_rate": 3.7547550516331555e-05, "loss": 0.337, "step": 179 }, { "epoch": 0.6081081081081081, "grad_norm": 0.5334018468856812, "learning_rate": 3.75878750850551e-05, "loss": 0.3244, "step": 180 }, { "epoch": 0.6114864864864865, "grad_norm": 0.5040500164031982, "learning_rate": 3.7627976247819744e-05, "loss": 0.3173, "step": 181 }, { "epoch": 0.6148648648648649, "grad_norm": 0.5094459652900696, "learning_rate": 3.766785646641792e-05, "loss": 0.3087, "step": 182 }, { "epoch": 0.6182432432432432, "grad_norm": 0.4679316282272339, "learning_rate": 3.770751816217383e-05, "loss": 0.3261, "step": 183 }, { "epoch": 0.6216216216216216, "grad_norm": 0.5099210143089294, "learning_rate": 3.7746963716825615e-05, "loss": 0.3239, "step": 184 }, { "epoch": 0.625, "grad_norm": 0.4242151379585266, "learning_rate": 3.778619547338356e-05, "loss": 0.3082, "step": 185 }, { "epoch": 0.6283783783783784, "grad_norm": 0.4796642065048218, "learning_rate": 3.782521573696528e-05, "loss": 0.2753, "step": 186 }, { "epoch": 0.6317567567567568, "grad_norm": 0.43798592686653137, "learning_rate": 3.786402677560832e-05, "loss": 0.3033, "step": 187 }, { "epoch": 0.6351351351351351, "grad_norm": 0.4947024881839752, "learning_rate": 3.790263082106134e-05, "loss": 0.3171, "step": 188 }, { "epoch": 0.6385135135135135, "grad_norm": 0.40923503041267395, "learning_rate": 3.794103006955407e-05, "loss": 0.2927, "step": 189 }, { "epoch": 0.6418918918918919, "grad_norm": 0.6707382202148438, "learning_rate": 3.797922668254715e-05, "loss": 0.2885, "step": 190 }, { "epoch": 0.6452702702702703, "grad_norm": 0.4996836185455322, "learning_rate": 3.801722278746213e-05, "loss": 0.3379, "step": 191 }, { "epoch": 0.6486486486486487, "grad_norm": 0.4873179495334625, "learning_rate": 3.8055020478392495e-05, "loss": 0.3091, "step": 192 }, { "epoch": 0.652027027027027, "grad_norm": 0.4989306330680847, "learning_rate": 3.809262181679623e-05, "loss": 0.3758, "step": 193 }, { "epoch": 0.6554054054054054, "grad_norm": 0.5524582862854004, "learning_rate": 3.813002883217044e-05, "loss": 0.3044, "step": 194 }, { "epoch": 0.6587837837837838, "grad_norm": 0.5253859162330627, "learning_rate": 3.816724352270863e-05, "loss": 0.3556, "step": 195 }, { "epoch": 0.6621621621621622, "grad_norm": 0.5060839653015137, "learning_rate": 3.8204267855941266e-05, "loss": 0.3352, "step": 196 }, { "epoch": 0.6655405405405406, "grad_norm": 0.5375229716300964, "learning_rate": 3.824110376935989e-05, "loss": 0.3328, "step": 197 }, { "epoch": 0.668918918918919, "grad_norm": 0.4941028654575348, "learning_rate": 3.827775317102552e-05, "loss": 0.296, "step": 198 }, { "epoch": 0.6722972972972973, "grad_norm": 0.5323918461799622, "learning_rate": 3.831421794016178e-05, "loss": 0.3049, "step": 199 }, { "epoch": 0.6756756756756757, "grad_norm": 0.5118021965026855, "learning_rate": 3.835049992773302e-05, "loss": 0.3133, "step": 200 }, { "epoch": 0.6790540540540541, "grad_norm": 0.46183663606643677, "learning_rate": 3.838660095700815e-05, "loss": 0.3185, "step": 201 }, { "epoch": 0.6824324324324325, "grad_norm": 0.47574761509895325, "learning_rate": 3.84225228241104e-05, "loss": 0.302, "step": 202 }, { "epoch": 0.6858108108108109, "grad_norm": 0.5129667520523071, "learning_rate": 3.8458267298553554e-05, "loss": 0.3407, "step": 203 }, { "epoch": 0.6891891891891891, "grad_norm": 1.532895803451538, "learning_rate": 3.8493836123764984e-05, "loss": 0.2858, "step": 204 }, { "epoch": 0.6925675675675675, "grad_norm": 0.5879459381103516, "learning_rate": 3.852923101759591e-05, "loss": 0.3638, "step": 205 }, { "epoch": 0.6959459459459459, "grad_norm": 0.589292585849762, "learning_rate": 3.856445367281923e-05, "loss": 0.3077, "step": 206 }, { "epoch": 0.6993243243243243, "grad_norm": 0.44842028617858887, "learning_rate": 3.859950575761529e-05, "loss": 0.2791, "step": 207 }, { "epoch": 0.7027027027027027, "grad_norm": 0.6291372179985046, "learning_rate": 3.8634388916046025e-05, "loss": 0.3404, "step": 208 }, { "epoch": 0.706081081081081, "grad_norm": 0.4508068561553955, "learning_rate": 3.866910476851757e-05, "loss": 0.2859, "step": 209 }, { "epoch": 0.7094594594594594, "grad_norm": 0.7987334132194519, "learning_rate": 3.870365491223199e-05, "loss": 0.3224, "step": 210 }, { "epoch": 0.7128378378378378, "grad_norm": 0.5030388236045837, "learning_rate": 3.8738040921628215e-05, "loss": 0.3225, "step": 211 }, { "epoch": 0.7162162162162162, "grad_norm": 0.5433597564697266, "learning_rate": 3.877226434881253e-05, "loss": 0.3294, "step": 212 }, { "epoch": 0.7195945945945946, "grad_norm": 0.48354920744895935, "learning_rate": 3.880632672397897e-05, "loss": 0.3035, "step": 213 }, { "epoch": 0.722972972972973, "grad_norm": 0.5904508829116821, "learning_rate": 3.884022955581985e-05, "loss": 0.3261, "step": 214 }, { "epoch": 0.7263513513513513, "grad_norm": 0.5152267813682556, "learning_rate": 3.887397433192676e-05, "loss": 0.2926, "step": 215 }, { "epoch": 0.7297297297297297, "grad_norm": 0.46898239850997925, "learning_rate": 3.890756251918219e-05, "loss": 0.2808, "step": 216 }, { "epoch": 0.7331081081081081, "grad_norm": 0.4848116338253021, "learning_rate": 3.894099556414216e-05, "loss": 0.2995, "step": 217 }, { "epoch": 0.7364864864864865, "grad_norm": 0.48101919889450073, "learning_rate": 3.897427489341009e-05, "loss": 0.3054, "step": 218 }, { "epoch": 0.7398648648648649, "grad_norm": 0.49954646825790405, "learning_rate": 3.900740191400198e-05, "loss": 0.3239, "step": 219 }, { "epoch": 0.7432432432432432, "grad_norm": 0.46193623542785645, "learning_rate": 3.904037801370344e-05, "loss": 0.3215, "step": 220 }, { "epoch": 0.7466216216216216, "grad_norm": 0.4544968605041504, "learning_rate": 3.9073204561418514e-05, "loss": 0.2829, "step": 221 }, { "epoch": 0.75, "grad_norm": 0.4560447335243225, "learning_rate": 3.9105882907510644e-05, "loss": 0.3052, "step": 222 }, { "epoch": 0.7533783783783784, "grad_norm": 0.4681329131126404, "learning_rate": 3.913841438413601e-05, "loss": 0.3259, "step": 223 }, { "epoch": 0.7567567567567568, "grad_norm": 0.4784477949142456, "learning_rate": 3.917080030556938e-05, "loss": 0.3252, "step": 224 }, { "epoch": 0.7601351351351351, "grad_norm": 0.4931364953517914, "learning_rate": 3.9203041968522716e-05, "loss": 0.3252, "step": 225 }, { "epoch": 0.7635135135135135, "grad_norm": 0.4844168424606323, "learning_rate": 3.923514065245669e-05, "loss": 0.3185, "step": 226 }, { "epoch": 0.7668918918918919, "grad_norm": 0.5114946961402893, "learning_rate": 3.926709761988538e-05, "loss": 0.3136, "step": 227 }, { "epoch": 0.7702702702702703, "grad_norm": 0.4893558621406555, "learning_rate": 3.929891411667424e-05, "loss": 0.2977, "step": 228 }, { "epoch": 0.7736486486486487, "grad_norm": 0.4515063762664795, "learning_rate": 3.933059137233147e-05, "loss": 0.3031, "step": 229 }, { "epoch": 0.777027027027027, "grad_norm": 0.48304086923599243, "learning_rate": 3.9362130600293214e-05, "loss": 0.3292, "step": 230 }, { "epoch": 0.7804054054054054, "grad_norm": 0.4545653760433197, "learning_rate": 3.9393532998202405e-05, "loss": 0.3014, "step": 231 }, { "epoch": 0.7837837837837838, "grad_norm": 0.4236096739768982, "learning_rate": 3.942479974818166e-05, "loss": 0.2804, "step": 232 }, { "epoch": 0.7871621621621622, "grad_norm": 0.4648077189922333, "learning_rate": 3.945593201710032e-05, "loss": 0.3243, "step": 233 }, { "epoch": 0.7905405405405406, "grad_norm": 0.47247567772865295, "learning_rate": 3.9486930956835724e-05, "loss": 0.2907, "step": 234 }, { "epoch": 0.793918918918919, "grad_norm": 0.4848794639110565, "learning_rate": 3.951779770452894e-05, "loss": 0.3265, "step": 235 }, { "epoch": 0.7972972972972973, "grad_norm": 0.5114079713821411, "learning_rate": 3.954853338283512e-05, "loss": 0.3194, "step": 236 }, { "epoch": 0.8006756756756757, "grad_norm": 0.4802190959453583, "learning_rate": 3.9579139100168404e-05, "loss": 0.2857, "step": 237 }, { "epoch": 0.8040540540540541, "grad_norm": 0.522281289100647, "learning_rate": 3.960961595094187e-05, "loss": 0.2843, "step": 238 }, { "epoch": 0.8074324324324325, "grad_norm": 0.4458360970020294, "learning_rate": 3.96399650158023e-05, "loss": 0.2755, "step": 239 }, { "epoch": 0.8108108108108109, "grad_norm": 0.53693687915802, "learning_rate": 3.96701873618601e-05, "loss": 0.3015, "step": 240 }, { "epoch": 0.8141891891891891, "grad_norm": 0.46697714924812317, "learning_rate": 3.970028404291448e-05, "loss": 0.2911, "step": 241 }, { "epoch": 0.8175675675675675, "grad_norm": 0.5392897725105286, "learning_rate": 3.9730256099673865e-05, "loss": 0.3093, "step": 242 }, { "epoch": 0.8209459459459459, "grad_norm": 0.5555176734924316, "learning_rate": 3.976010455997187e-05, "loss": 0.3078, "step": 243 }, { "epoch": 0.8243243243243243, "grad_norm": 0.42975664138793945, "learning_rate": 3.978983043897883e-05, "loss": 0.2723, "step": 244 }, { "epoch": 0.8277027027027027, "grad_norm": 0.6645532250404358, "learning_rate": 3.981943473940888e-05, "loss": 0.3161, "step": 245 }, { "epoch": 0.831081081081081, "grad_norm": 0.47302716970443726, "learning_rate": 3.984891845172299e-05, "loss": 0.3002, "step": 246 }, { "epoch": 0.8344594594594594, "grad_norm": 0.5395675301551819, "learning_rate": 3.987828255432777e-05, "loss": 0.3614, "step": 247 }, { "epoch": 0.8378378378378378, "grad_norm": 0.5418782234191895, "learning_rate": 3.9907528013770276e-05, "loss": 0.3226, "step": 248 }, { "epoch": 0.8412162162162162, "grad_norm": 0.4323344826698303, "learning_rate": 3.993665578492894e-05, "loss": 0.289, "step": 249 }, { "epoch": 0.8445945945945946, "grad_norm": 0.5098227858543396, "learning_rate": 3.9965666811200624e-05, "loss": 0.3346, "step": 250 }, { "epoch": 0.847972972972973, "grad_norm": 0.4577200412750244, "learning_rate": 3.999456202468397e-05, "loss": 0.3526, "step": 251 }, { "epoch": 0.8513513513513513, "grad_norm": 0.48974189162254333, "learning_rate": 4.002334234635907e-05, "loss": 0.3086, "step": 252 }, { "epoch": 0.8547297297297297, "grad_norm": 0.47080233693122864, "learning_rate": 4.005200868626364e-05, "loss": 0.3578, "step": 253 }, { "epoch": 0.8581081081081081, "grad_norm": 0.5034478902816772, "learning_rate": 4.008056194366564e-05, "loss": 0.3166, "step": 254 }, { "epoch": 0.8614864864864865, "grad_norm": 0.44603192806243896, "learning_rate": 4.010900300723259e-05, "loss": 0.2969, "step": 255 }, { "epoch": 0.8648648648648649, "grad_norm": 0.5267269611358643, "learning_rate": 4.013733275519749e-05, "loss": 0.3341, "step": 256 }, { "epoch": 0.8682432432432432, "grad_norm": 0.424925297498703, "learning_rate": 4.016555205552158e-05, "loss": 0.2942, "step": 257 }, { "epoch": 0.8716216216216216, "grad_norm": 0.5034767389297485, "learning_rate": 4.0193661766053834e-05, "loss": 0.2668, "step": 258 }, { "epoch": 0.875, "grad_norm": 0.5288417935371399, "learning_rate": 4.022166273468753e-05, "loss": 0.3424, "step": 259 }, { "epoch": 0.8783783783783784, "grad_norm": 0.4726288616657257, "learning_rate": 4.024955579951363e-05, "loss": 0.2906, "step": 260 }, { "epoch": 0.8817567567567568, "grad_norm": 0.5311090350151062, "learning_rate": 4.027734178897136e-05, "loss": 0.3307, "step": 261 }, { "epoch": 0.8851351351351351, "grad_norm": 0.427002489566803, "learning_rate": 4.030502152199576e-05, "loss": 0.2569, "step": 262 }, { "epoch": 0.8885135135135135, "grad_norm": 0.543899416923523, "learning_rate": 4.033259580816264e-05, "loss": 0.2925, "step": 263 }, { "epoch": 0.8918918918918919, "grad_norm": 0.4932996928691864, "learning_rate": 4.036006544783052e-05, "loss": 0.3058, "step": 264 }, { "epoch": 0.8952702702702703, "grad_norm": 0.5138298273086548, "learning_rate": 4.0387431232280135e-05, "loss": 0.2952, "step": 265 }, { "epoch": 0.8986486486486487, "grad_norm": 0.4801797568798065, "learning_rate": 4.041469394385112e-05, "loss": 0.2873, "step": 266 }, { "epoch": 0.902027027027027, "grad_norm": 0.453140527009964, "learning_rate": 4.0441854356076257e-05, "loss": 0.2747, "step": 267 }, { "epoch": 0.9054054054054054, "grad_norm": 0.4941580593585968, "learning_rate": 4.046891323381315e-05, "loss": 0.3157, "step": 268 }, { "epoch": 0.9087837837837838, "grad_norm": 0.6151431798934937, "learning_rate": 4.049587133337347e-05, "loss": 0.3139, "step": 269 }, { "epoch": 0.9121621621621622, "grad_norm": 0.4674379229545593, "learning_rate": 4.0522729402649793e-05, "loss": 0.3122, "step": 270 }, { "epoch": 0.9155405405405406, "grad_norm": 0.4719599187374115, "learning_rate": 4.0549488181240096e-05, "loss": 0.3101, "step": 271 }, { "epoch": 0.918918918918919, "grad_norm": 0.483374685049057, "learning_rate": 4.057614840056998e-05, "loss": 0.3087, "step": 272 }, { "epoch": 0.9222972972972973, "grad_norm": 0.48100754618644714, "learning_rate": 4.06027107840126e-05, "loss": 0.3044, "step": 273 }, { "epoch": 0.9256756756756757, "grad_norm": 0.44335874915122986, "learning_rate": 4.0629176047006474e-05, "loss": 0.2929, "step": 274 }, { "epoch": 0.9290540540540541, "grad_norm": 0.4475744962692261, "learning_rate": 4.065554489717105e-05, "loss": 0.321, "step": 275 }, { "epoch": 0.9324324324324325, "grad_norm": 0.43009960651397705, "learning_rate": 4.068181803442029e-05, "loss": 0.3026, "step": 276 }, { "epoch": 0.9358108108108109, "grad_norm": 0.48129919171333313, "learning_rate": 4.0707996151074147e-05, "loss": 0.2876, "step": 277 }, { "epoch": 0.9391891891891891, "grad_norm": 0.423658549785614, "learning_rate": 4.073407993196794e-05, "loss": 0.2968, "step": 278 }, { "epoch": 0.9425675675675675, "grad_norm": 0.485857218503952, "learning_rate": 4.076007005455996e-05, "loss": 0.3144, "step": 279 }, { "epoch": 0.9459459459459459, "grad_norm": 0.4936545193195343, "learning_rate": 4.0785967189036986e-05, "loss": 0.3103, "step": 280 }, { "epoch": 0.9493243243243243, "grad_norm": 0.47265681624412537, "learning_rate": 4.0811771998418e-05, "loss": 0.3136, "step": 281 }, { "epoch": 0.9527027027027027, "grad_norm": 0.4872439503669739, "learning_rate": 4.083748513865602e-05, "loss": 0.335, "step": 282 }, { "epoch": 0.956081081081081, "grad_norm": 0.4380621910095215, "learning_rate": 4.086310725873818e-05, "loss": 0.3036, "step": 283 }, { "epoch": 0.9594594594594594, "grad_norm": 0.4759541451931, "learning_rate": 4.0888639000783966e-05, "loss": 0.2827, "step": 284 }, { "epoch": 0.9628378378378378, "grad_norm": 0.995196521282196, "learning_rate": 4.0914081000141844e-05, "loss": 0.3162, "step": 285 }, { "epoch": 0.9662162162162162, "grad_norm": 0.5686355829238892, "learning_rate": 4.0939433885484055e-05, "loss": 0.2992, "step": 286 }, { "epoch": 0.9695945945945946, "grad_norm": 0.43086209893226624, "learning_rate": 4.0964698278899874e-05, "loss": 0.278, "step": 287 }, { "epoch": 0.972972972972973, "grad_norm": 0.5099641680717468, "learning_rate": 4.0989874795987185e-05, "loss": 0.3023, "step": 288 }, { "epoch": 0.9763513513513513, "grad_norm": 0.39634063839912415, "learning_rate": 4.1014964045942465e-05, "loss": 0.2806, "step": 289 }, { "epoch": 0.9797297297297297, "grad_norm": 0.4685455858707428, "learning_rate": 4.103996663164927e-05, "loss": 0.3078, "step": 290 }, { "epoch": 0.9831081081081081, "grad_norm": 0.4501771330833435, "learning_rate": 4.106488314976513e-05, "loss": 0.3048, "step": 291 }, { "epoch": 0.9864864864864865, "grad_norm": 0.4352227747440338, "learning_rate": 4.108971419080698e-05, "loss": 0.305, "step": 292 }, { "epoch": 0.9898648648648649, "grad_norm": 0.4946969151496887, "learning_rate": 4.111446033923516e-05, "loss": 0.3027, "step": 293 }, { "epoch": 0.9932432432432432, "grad_norm": 0.463313490152359, "learning_rate": 4.113912217353596e-05, "loss": 0.3148, "step": 294 }, { "epoch": 0.9966216216216216, "grad_norm": 0.42501333355903625, "learning_rate": 4.116370026630272e-05, "loss": 0.303, "step": 295 }, { "epoch": 1.0, "grad_norm": 0.4174898862838745, "learning_rate": 4.118819518431564e-05, "loss": 0.2768, "step": 296 }, { "epoch": 1.0033783783783783, "grad_norm": 0.5591414570808411, "learning_rate": 4.121260748862021e-05, "loss": 0.2319, "step": 297 }, { "epoch": 1.0067567567567568, "grad_norm": 0.4628024101257324, "learning_rate": 4.123693773460426e-05, "loss": 0.2332, "step": 298 }, { "epoch": 1.010135135135135, "grad_norm": 0.9493624567985535, "learning_rate": 4.126118647207383e-05, "loss": 0.2113, "step": 299 }, { "epoch": 1.0135135135135136, "grad_norm": 0.6151244044303894, "learning_rate": 4.1285354245327715e-05, "loss": 0.2428, "step": 300 }, { "epoch": 1.0168918918918919, "grad_norm": 0.4934402406215668, "learning_rate": 4.1309441593230726e-05, "loss": 0.2394, "step": 301 }, { "epoch": 1.0202702702702702, "grad_norm": 0.5464348793029785, "learning_rate": 4.133344904928585e-05, "loss": 0.2584, "step": 302 }, { "epoch": 1.0236486486486487, "grad_norm": 0.754717230796814, "learning_rate": 4.1357377141705084e-05, "loss": 0.2191, "step": 303 }, { "epoch": 1.027027027027027, "grad_norm": 0.48623141646385193, "learning_rate": 4.1381226393479236e-05, "loss": 0.2185, "step": 304 }, { "epoch": 1.0304054054054055, "grad_norm": 32.5340461730957, "learning_rate": 4.1404997322446435e-05, "loss": 0.2328, "step": 305 }, { "epoch": 1.0337837837837838, "grad_norm": 0.7727927565574646, "learning_rate": 4.142869044135967e-05, "loss": 0.2197, "step": 306 }, { "epoch": 1.037162162162162, "grad_norm": 0.47851279377937317, "learning_rate": 4.145230625795311e-05, "loss": 0.2093, "step": 307 }, { "epoch": 1.0405405405405406, "grad_norm": 0.684687077999115, "learning_rate": 4.14758452750074e-05, "loss": 0.2145, "step": 308 }, { "epoch": 1.0439189189189189, "grad_norm": 0.4834594428539276, "learning_rate": 4.149930799041392e-05, "loss": 0.222, "step": 309 }, { "epoch": 1.0472972972972974, "grad_norm": 0.5479368567466736, "learning_rate": 4.152269489723788e-05, "loss": 0.2245, "step": 310 }, { "epoch": 1.0506756756756757, "grad_norm": 0.48793625831604004, "learning_rate": 4.1546006483780626e-05, "loss": 0.2428, "step": 311 }, { "epoch": 1.054054054054054, "grad_norm": 0.5712085962295532, "learning_rate": 4.156924323364072e-05, "loss": 0.2311, "step": 312 }, { "epoch": 1.0574324324324325, "grad_norm": 0.46200135350227356, "learning_rate": 4.1592405625774144e-05, "loss": 0.2315, "step": 313 }, { "epoch": 1.0608108108108107, "grad_norm": 0.5134051442146301, "learning_rate": 4.161549413455358e-05, "loss": 0.2205, "step": 314 }, { "epoch": 1.0641891891891893, "grad_norm": 0.4155457019805908, "learning_rate": 4.163850922982668e-05, "loss": 0.1969, "step": 315 }, { "epoch": 1.0675675675675675, "grad_norm": 0.47662997245788574, "learning_rate": 4.16614513769734e-05, "loss": 0.2183, "step": 316 }, { "epoch": 1.070945945945946, "grad_norm": 0.4776044189929962, "learning_rate": 4.1684321036962526e-05, "loss": 0.2581, "step": 317 }, { "epoch": 1.0743243243243243, "grad_norm": 0.5491393804550171, "learning_rate": 4.170711866640721e-05, "loss": 0.2454, "step": 318 }, { "epoch": 1.0777027027027026, "grad_norm": 0.4745365381240845, "learning_rate": 4.1729844717619684e-05, "loss": 0.2303, "step": 319 }, { "epoch": 1.0810810810810811, "grad_norm": 0.49116480350494385, "learning_rate": 4.17524996386651e-05, "loss": 0.2205, "step": 320 }, { "epoch": 1.0844594594594594, "grad_norm": 0.47651857137680054, "learning_rate": 4.177508387341454e-05, "loss": 0.2538, "step": 321 }, { "epoch": 1.087837837837838, "grad_norm": 24.205638885498047, "learning_rate": 4.179759786159719e-05, "loss": 0.5427, "step": 322 }, { "epoch": 1.0912162162162162, "grad_norm": 0.7813121676445007, "learning_rate": 4.182004203885172e-05, "loss": 0.2445, "step": 323 }, { "epoch": 1.0945945945945945, "grad_norm": 52.29485321044922, "learning_rate": 4.184241683677687e-05, "loss": 0.2577, "step": 324 }, { "epoch": 1.097972972972973, "grad_norm": 0.8282439112663269, "learning_rate": 4.1864722682981245e-05, "loss": 0.2421, "step": 325 }, { "epoch": 1.1013513513513513, "grad_norm": 0.4729321300983429, "learning_rate": 4.188696000113232e-05, "loss": 0.2122, "step": 326 }, { "epoch": 1.1047297297297298, "grad_norm": 0.6732134819030762, "learning_rate": 4.190912921100477e-05, "loss": 0.2295, "step": 327 }, { "epoch": 1.1081081081081081, "grad_norm": 0.567088782787323, "learning_rate": 4.1931230728527994e-05, "loss": 0.2352, "step": 328 }, { "epoch": 1.1114864864864864, "grad_norm": 8.066313743591309, "learning_rate": 4.195326496583291e-05, "loss": 0.7099, "step": 329 }, { "epoch": 1.114864864864865, "grad_norm": 1.0606003999710083, "learning_rate": 4.1975232331298125e-05, "loss": 0.2105, "step": 330 }, { "epoch": 1.1182432432432432, "grad_norm": 0.4454852342605591, "learning_rate": 4.1997133229595316e-05, "loss": 0.2063, "step": 331 }, { "epoch": 1.1216216216216217, "grad_norm": 0.873603880405426, "learning_rate": 4.201896806173394e-05, "loss": 0.2447, "step": 332 }, { "epoch": 1.125, "grad_norm": 0.6670548319816589, "learning_rate": 4.2040737225105335e-05, "loss": 0.2298, "step": 333 }, { "epoch": 1.1283783783783783, "grad_norm": 0.5834859609603882, "learning_rate": 4.206244111352608e-05, "loss": 0.1799, "step": 334 }, { "epoch": 1.1317567567567568, "grad_norm": 0.8503464460372925, "learning_rate": 4.2084080117280756e-05, "loss": 0.2254, "step": 335 }, { "epoch": 1.135135135135135, "grad_norm": 0.5728862285614014, "learning_rate": 4.210565462316407e-05, "loss": 0.2342, "step": 336 }, { "epoch": 1.1385135135135136, "grad_norm": 0.7374505400657654, "learning_rate": 4.2127165014522315e-05, "loss": 0.2384, "step": 337 }, { "epoch": 1.1418918918918919, "grad_norm": 0.5631051659584045, "learning_rate": 4.214861167129425e-05, "loss": 0.2213, "step": 338 }, { "epoch": 1.1452702702702702, "grad_norm": 26.428640365600586, "learning_rate": 4.2169994970051365e-05, "loss": 0.8302, "step": 339 }, { "epoch": 1.1486486486486487, "grad_norm": 1.1665905714035034, "learning_rate": 4.219131528403759e-05, "loss": 0.2568, "step": 340 }, { "epoch": 1.152027027027027, "grad_norm": 0.6159250140190125, "learning_rate": 4.22125729832083e-05, "loss": 0.2574, "step": 341 }, { "epoch": 1.1554054054054055, "grad_norm": 182.47439575195312, "learning_rate": 4.2233768434268914e-05, "loss": 0.8118, "step": 342 }, { "epoch": 1.1587837837837838, "grad_norm": 1.685766339302063, "learning_rate": 4.225490200071284e-05, "loss": 0.2347, "step": 343 }, { "epoch": 1.1621621621621623, "grad_norm": 0.9619600176811218, "learning_rate": 4.227597404285883e-05, "loss": 0.2453, "step": 344 }, { "epoch": 1.1655405405405406, "grad_norm": 0.49926599860191345, "learning_rate": 4.229698491788791e-05, "loss": 0.2313, "step": 345 }, { "epoch": 1.1689189189189189, "grad_norm": 0.9699143767356873, "learning_rate": 4.231793497987961e-05, "loss": 0.256, "step": 346 }, { "epoch": 1.1722972972972974, "grad_norm": 0.7428460121154785, "learning_rate": 4.2338824579847904e-05, "loss": 0.2655, "step": 347 }, { "epoch": 1.1756756756756757, "grad_norm": 0.6480849385261536, "learning_rate": 4.235965406577636e-05, "loss": 0.2385, "step": 348 }, { "epoch": 1.179054054054054, "grad_norm": 0.6347863078117371, "learning_rate": 4.2380423782653e-05, "loss": 0.2196, "step": 349 }, { "epoch": 1.1824324324324325, "grad_norm": 0.6471324563026428, "learning_rate": 4.240113407250459e-05, "loss": 0.2337, "step": 350 }, { "epoch": 1.1858108108108107, "grad_norm": 0.5807657241821289, "learning_rate": 4.24217852744304e-05, "loss": 0.2438, "step": 351 }, { "epoch": 1.1891891891891893, "grad_norm": 0.5200501084327698, "learning_rate": 4.244237772463552e-05, "loss": 0.2458, "step": 352 }, { "epoch": 1.1925675675675675, "grad_norm": 0.5445655584335327, "learning_rate": 4.246291175646371e-05, "loss": 0.2391, "step": 353 }, { "epoch": 1.195945945945946, "grad_norm": 0.6283414363861084, "learning_rate": 4.24833877004298e-05, "loss": 0.2299, "step": 354 }, { "epoch": 1.1993243243243243, "grad_norm": 0.5581656694412231, "learning_rate": 4.250380588425157e-05, "loss": 0.2272, "step": 355 }, { "epoch": 1.2027027027027026, "grad_norm": 0.5711120367050171, "learning_rate": 4.2524166632881255e-05, "loss": 0.2737, "step": 356 }, { "epoch": 1.2060810810810811, "grad_norm": 64.4117660522461, "learning_rate": 4.254447026853656e-05, "loss": 0.2648, "step": 357 }, { "epoch": 1.2094594594594594, "grad_norm": 0.84456467628479, "learning_rate": 4.2564717110731244e-05, "loss": 0.2643, "step": 358 }, { "epoch": 1.2128378378378377, "grad_norm": 1.7171592712402344, "learning_rate": 4.258490747630532e-05, "loss": 0.6041, "step": 359 }, { "epoch": 1.2162162162162162, "grad_norm": 0.6733882427215576, "learning_rate": 4.260504167945479e-05, "loss": 0.2226, "step": 360 }, { "epoch": 1.2195945945945945, "grad_norm": 0.5079653859138489, "learning_rate": 4.2625120031760965e-05, "loss": 0.2205, "step": 361 }, { "epoch": 1.222972972972973, "grad_norm": 0.5646266341209412, "learning_rate": 4.264514284221944e-05, "loss": 0.2272, "step": 362 }, { "epoch": 1.2263513513513513, "grad_norm": 0.5336301922798157, "learning_rate": 4.266511041726854e-05, "loss": 0.2536, "step": 363 }, { "epoch": 1.2297297297297298, "grad_norm": 0.5218775272369385, "learning_rate": 4.26850230608176e-05, "loss": 0.2582, "step": 364 }, { "epoch": 1.2331081081081081, "grad_norm": 0.47051167488098145, "learning_rate": 4.2704881074274584e-05, "loss": 0.2288, "step": 365 }, { "epoch": 1.2364864864864864, "grad_norm": 0.4820377826690674, "learning_rate": 4.272468475657351e-05, "loss": 0.2396, "step": 366 }, { "epoch": 1.239864864864865, "grad_norm": 1219.8048095703125, "learning_rate": 4.2744434404201497e-05, "loss": 1.9213, "step": 367 }, { "epoch": 1.2432432432432432, "grad_norm": 1.3034559488296509, "learning_rate": 4.27641303112253e-05, "loss": 0.2528, "step": 368 }, { "epoch": 1.2466216216216217, "grad_norm": 0.6443231105804443, "learning_rate": 4.278377276931767e-05, "loss": 0.2235, "step": 369 }, { "epoch": 1.25, "grad_norm": 0.4247282147407532, "learning_rate": 4.2803362067783256e-05, "loss": 0.2111, "step": 370 }, { "epoch": 1.2533783783783785, "grad_norm": 0.5400434136390686, "learning_rate": 4.2822898493584104e-05, "loss": 0.2256, "step": 371 }, { "epoch": 1.2567567567567568, "grad_norm": 0.43781599402427673, "learning_rate": 4.284238233136496e-05, "loss": 0.2295, "step": 372 }, { "epoch": 1.260135135135135, "grad_norm": 0.5443878173828125, "learning_rate": 4.286181386347813e-05, "loss": 0.2394, "step": 373 }, { "epoch": 1.2635135135135136, "grad_norm": 0.4616173207759857, "learning_rate": 4.288119337000801e-05, "loss": 0.2108, "step": 374 }, { "epoch": 1.2668918918918919, "grad_norm": 0.4537034332752228, "learning_rate": 4.2900521128795315e-05, "loss": 0.2192, "step": 375 }, { "epoch": 1.2702702702702702, "grad_norm": 0.5639699697494507, "learning_rate": 4.291979741546102e-05, "loss": 0.2403, "step": 376 }, { "epoch": 1.2736486486486487, "grad_norm": 0.4133036434650421, "learning_rate": 4.293902250342989e-05, "loss": 0.2086, "step": 377 }, { "epoch": 1.277027027027027, "grad_norm": 0.45922860503196716, "learning_rate": 4.295819666395376e-05, "loss": 0.236, "step": 378 }, { "epoch": 1.2804054054054055, "grad_norm": 0.4423050284385681, "learning_rate": 4.297732016613454e-05, "loss": 0.2271, "step": 379 }, { "epoch": 1.2837837837837838, "grad_norm": 0.4588952362537384, "learning_rate": 4.299639327694684e-05, "loss": 0.2356, "step": 380 }, { "epoch": 1.2871621621621623, "grad_norm": 26.979991912841797, "learning_rate": 4.3015416261260325e-05, "loss": 0.2247, "step": 381 }, { "epoch": 1.2905405405405406, "grad_norm": 0.6023054718971252, "learning_rate": 4.303438938186182e-05, "loss": 0.2477, "step": 382 }, { "epoch": 1.2939189189189189, "grad_norm": 0.48854556679725647, "learning_rate": 4.305331289947705e-05, "loss": 0.221, "step": 383 }, { "epoch": 1.2972972972972974, "grad_norm": 0.5875428915023804, "learning_rate": 4.3072187072792184e-05, "loss": 0.2443, "step": 384 }, { "epoch": 1.3006756756756757, "grad_norm": 0.42068448662757874, "learning_rate": 4.309101215847502e-05, "loss": 0.2043, "step": 385 }, { "epoch": 1.304054054054054, "grad_norm": 0.492660254240036, "learning_rate": 4.3109788411195924e-05, "loss": 0.246, "step": 386 }, { "epoch": 1.3074324324324325, "grad_norm": 0.4522532820701599, "learning_rate": 4.312851608364853e-05, "loss": 0.2305, "step": 387 }, { "epoch": 1.3108108108108107, "grad_norm": 0.4410349130630493, "learning_rate": 4.314719542657013e-05, "loss": 0.2146, "step": 388 }, { "epoch": 1.3141891891891893, "grad_norm": 38.59849548339844, "learning_rate": 4.3165826688761796e-05, "loss": 1.1445, "step": 389 }, { "epoch": 1.3175675675675675, "grad_norm": 0.6608863472938538, "learning_rate": 4.318441011710833e-05, "loss": 0.2591, "step": 390 }, { "epoch": 1.320945945945946, "grad_norm": 0.7261571884155273, "learning_rate": 4.3202945956597786e-05, "loss": 0.2409, "step": 391 }, { "epoch": 1.3243243243243243, "grad_norm": 13.75738525390625, "learning_rate": 4.3221434450340956e-05, "loss": 1.022, "step": 392 }, { "epoch": 1.3277027027027026, "grad_norm": 1.4299050569534302, "learning_rate": 4.323987583959045e-05, "loss": 0.2287, "step": 393 }, { "epoch": 1.3310810810810811, "grad_norm": 0.6660119295120239, "learning_rate": 4.325827036375957e-05, "loss": 0.2148, "step": 394 }, { "epoch": 1.3344594594594594, "grad_norm": 3.13312029838562, "learning_rate": 4.327661826044101e-05, "loss": 0.6968, "step": 395 }, { "epoch": 1.3378378378378377, "grad_norm": 29.60978126525879, "learning_rate": 4.329491976542521e-05, "loss": 0.6915, "step": 396 }, { "epoch": 1.3412162162162162, "grad_norm": 0.8736880421638489, "learning_rate": 4.331317511271859e-05, "loss": 0.2388, "step": 397 }, { "epoch": 1.3445945945945945, "grad_norm": 0.5873726606369019, "learning_rate": 4.333138453456147e-05, "loss": 0.2655, "step": 398 }, { "epoch": 1.347972972972973, "grad_norm": 2.68512225151062, "learning_rate": 4.334954826144581e-05, "loss": 0.6129, "step": 399 }, { "epoch": 1.3513513513513513, "grad_norm": 0.8919135928153992, "learning_rate": 4.336766652213271e-05, "loss": 0.2492, "step": 400 }, { "epoch": 1.3547297297297298, "grad_norm": 2.113346576690674, "learning_rate": 4.338573954366971e-05, "loss": 0.2578, "step": 401 }, { "epoch": 1.3581081081081081, "grad_norm": 2.666167736053467, "learning_rate": 4.340376755140784e-05, "loss": 0.6334, "step": 402 }, { "epoch": 1.3614864864864864, "grad_norm": 0.8454524278640747, "learning_rate": 4.342175076901849e-05, "loss": 0.2426, "step": 403 }, { "epoch": 1.364864864864865, "grad_norm": 0.7469472885131836, "learning_rate": 4.343968941851009e-05, "loss": 0.2715, "step": 404 }, { "epoch": 1.3682432432432432, "grad_norm": 0.6965801119804382, "learning_rate": 4.345758372024448e-05, "loss": 0.2462, "step": 405 }, { "epoch": 1.3716216216216215, "grad_norm": 1.5828707218170166, "learning_rate": 4.347543389295324e-05, "loss": 0.5967, "step": 406 }, { "epoch": 1.375, "grad_norm": 0.5627582669258118, "learning_rate": 4.3493240153753666e-05, "loss": 0.226, "step": 407 }, { "epoch": 1.3783783783783785, "grad_norm": 0.5101889967918396, "learning_rate": 4.3511002718164666e-05, "loss": 0.2266, "step": 408 }, { "epoch": 1.3817567567567568, "grad_norm": 0.6242619156837463, "learning_rate": 4.352872180012237e-05, "loss": 0.2028, "step": 409 }, { "epoch": 1.385135135135135, "grad_norm": 0.5768154263496399, "learning_rate": 4.35463976119956e-05, "loss": 0.2291, "step": 410 }, { "epoch": 1.3885135135135136, "grad_norm": 0.5170788764953613, "learning_rate": 4.356403036460115e-05, "loss": 0.2124, "step": 411 }, { "epoch": 1.3918918918918919, "grad_norm": 0.530846357345581, "learning_rate": 4.3581620267218916e-05, "loss": 0.2013, "step": 412 }, { "epoch": 1.3952702702702702, "grad_norm": 0.5346778631210327, "learning_rate": 4.359916752760669e-05, "loss": 0.2439, "step": 413 }, { "epoch": 1.3986486486486487, "grad_norm": 0.5345392227172852, "learning_rate": 4.361667235201499e-05, "loss": 0.2321, "step": 414 }, { "epoch": 1.402027027027027, "grad_norm": 0.5348331928253174, "learning_rate": 4.363413494520154e-05, "loss": 0.2418, "step": 415 }, { "epoch": 1.4054054054054055, "grad_norm": 0.9726234078407288, "learning_rate": 4.365155551044572e-05, "loss": 0.213, "step": 416 }, { "epoch": 1.4087837837837838, "grad_norm": 2.116424083709717, "learning_rate": 4.366893424956263e-05, "loss": 0.5716, "step": 417 }, { "epoch": 1.4121621621621623, "grad_norm": 0.6004624962806702, "learning_rate": 4.368627136291726e-05, "loss": 0.2404, "step": 418 }, { "epoch": 1.4155405405405406, "grad_norm": 0.44696101546287537, "learning_rate": 4.370356704943825e-05, "loss": 0.2349, "step": 419 }, { "epoch": 1.4189189189189189, "grad_norm": 0.4704833924770355, "learning_rate": 4.372082150663168e-05, "loss": 0.2365, "step": 420 }, { "epoch": 1.4222972972972974, "grad_norm": 0.48295876383781433, "learning_rate": 4.3738034930594475e-05, "loss": 0.2256, "step": 421 }, { "epoch": 1.4256756756756757, "grad_norm": 0.4726882576942444, "learning_rate": 4.3755207516027904e-05, "loss": 0.2283, "step": 422 }, { "epoch": 1.429054054054054, "grad_norm": 0.5251411199569702, "learning_rate": 4.377233945625071e-05, "loss": 0.2374, "step": 423 }, { "epoch": 1.4324324324324325, "grad_norm": 94.52925109863281, "learning_rate": 4.378943094321221e-05, "loss": 0.4887, "step": 424 }, { "epoch": 1.4358108108108107, "grad_norm": 0.5848643183708191, "learning_rate": 4.3806482167505196e-05, "loss": 0.2373, "step": 425 }, { "epoch": 1.4391891891891893, "grad_norm": 0.870020866394043, "learning_rate": 4.382349331837866e-05, "loss": 0.5312, "step": 426 }, { "epoch": 1.4425675675675675, "grad_norm": 0.58022540807724, "learning_rate": 4.3840464583750404e-05, "loss": 0.2347, "step": 427 }, { "epoch": 1.445945945945946, "grad_norm": 0.4780952036380768, "learning_rate": 4.385739615021954e-05, "loss": 0.2392, "step": 428 }, { "epoch": 1.4493243243243243, "grad_norm": 0.4691276252269745, "learning_rate": 4.387428820307874e-05, "loss": 0.2346, "step": 429 }, { "epoch": 1.4527027027027026, "grad_norm": 0.4635365903377533, "learning_rate": 4.3891140926326446e-05, "loss": 0.2242, "step": 430 }, { "epoch": 1.4560810810810811, "grad_norm": 0.7928866147994995, "learning_rate": 4.390795450267886e-05, "loss": 0.2267, "step": 431 }, { "epoch": 1.4594594594594594, "grad_norm": 0.5844922661781311, "learning_rate": 4.3924729113581876e-05, "loss": 0.237, "step": 432 }, { "epoch": 1.4628378378378377, "grad_norm": 0.4539477229118347, "learning_rate": 4.394146493922276e-05, "loss": 0.2624, "step": 433 }, { "epoch": 1.4662162162162162, "grad_norm": 0.4881402254104614, "learning_rate": 4.395816215854185e-05, "loss": 0.2076, "step": 434 }, { "epoch": 1.4695945945945945, "grad_norm": 0.42912018299102783, "learning_rate": 4.397482094924396e-05, "loss": 0.224, "step": 435 }, { "epoch": 1.472972972972973, "grad_norm": 8.578511238098145, "learning_rate": 4.399144148780977e-05, "loss": 0.5361, "step": 436 }, { "epoch": 1.4763513513513513, "grad_norm": 0.4810888469219208, "learning_rate": 4.400802394950703e-05, "loss": 0.2274, "step": 437 }, { "epoch": 1.4797297297297298, "grad_norm": 0.5107089281082153, "learning_rate": 4.402456850840166e-05, "loss": 0.2295, "step": 438 }, { "epoch": 1.4831081081081081, "grad_norm": 0.4564654529094696, "learning_rate": 4.4041075337368695e-05, "loss": 0.2447, "step": 439 }, { "epoch": 1.4864864864864864, "grad_norm": 1.4533860683441162, "learning_rate": 4.405754460810312e-05, "loss": 0.2522, "step": 440 }, { "epoch": 1.489864864864865, "grad_norm": 2.172696828842163, "learning_rate": 4.407397649113065e-05, "loss": 0.8104, "step": 441 }, { "epoch": 1.4932432432432432, "grad_norm": 0.4864687919616699, "learning_rate": 4.40903711558182e-05, "loss": 0.2657, "step": 442 }, { "epoch": 1.4966216216216215, "grad_norm": 0.4211529791355133, "learning_rate": 4.41067287703845e-05, "loss": 0.246, "step": 443 }, { "epoch": 1.5, "grad_norm": 0.5068204998970032, "learning_rate": 4.412304950191033e-05, "loss": 0.219, "step": 444 }, { "epoch": 1.5033783783783785, "grad_norm": 0.52825927734375, "learning_rate": 4.413933351634886e-05, "loss": 0.2293, "step": 445 }, { "epoch": 1.5067567567567568, "grad_norm": 0.45864129066467285, "learning_rate": 4.4155580978535707e-05, "loss": 0.218, "step": 446 }, { "epoch": 1.510135135135135, "grad_norm": 1.0356566905975342, "learning_rate": 4.417179205219895e-05, "loss": 0.2278, "step": 447 }, { "epoch": 1.5135135135135136, "grad_norm": 0.4032718241214752, "learning_rate": 4.418796689996907e-05, "loss": 0.2276, "step": 448 }, { "epoch": 1.5168918918918919, "grad_norm": 0.4162694811820984, "learning_rate": 4.420410568338872e-05, "loss": 0.23, "step": 449 }, { "epoch": 1.5202702702702702, "grad_norm": 0.4550395607948303, "learning_rate": 4.42202085629224e-05, "loss": 0.2445, "step": 450 }, { "epoch": 1.5236486486486487, "grad_norm": 0.5222985148429871, "learning_rate": 4.423627569796601e-05, "loss": 0.2405, "step": 451 }, { "epoch": 1.527027027027027, "grad_norm": 0.4746388792991638, "learning_rate": 4.425230724685638e-05, "loss": 0.2299, "step": 452 }, { "epoch": 1.5304054054054053, "grad_norm": 0.47175803780555725, "learning_rate": 4.4268303366880536e-05, "loss": 0.234, "step": 453 }, { "epoch": 1.5337837837837838, "grad_norm": 0.46304166316986084, "learning_rate": 4.428426421428507e-05, "loss": 0.2165, "step": 454 }, { "epoch": 1.5371621621621623, "grad_norm": 0.8317951560020447, "learning_rate": 4.430018994428521e-05, "loss": 0.2648, "step": 455 }, { "epoch": 1.5405405405405406, "grad_norm": 0.4764087498188019, "learning_rate": 4.431608071107392e-05, "loss": 0.2526, "step": 456 }, { "epoch": 1.5439189189189189, "grad_norm": 0.39065808057785034, "learning_rate": 4.433193666783084e-05, "loss": 0.2428, "step": 457 }, { "epoch": 1.5472972972972974, "grad_norm": 0.40975654125213623, "learning_rate": 4.4347757966731156e-05, "loss": 0.2437, "step": 458 }, { "epoch": 1.5506756756756757, "grad_norm": 0.4318806231021881, "learning_rate": 4.436354475895436e-05, "loss": 0.233, "step": 459 }, { "epoch": 1.554054054054054, "grad_norm": 0.4574219882488251, "learning_rate": 4.437929719469291e-05, "loss": 0.2441, "step": 460 }, { "epoch": 1.5574324324324325, "grad_norm": 0.39155369997024536, "learning_rate": 4.4395015423160807e-05, "loss": 0.2233, "step": 461 }, { "epoch": 1.560810810810811, "grad_norm": 0.4589376747608185, "learning_rate": 4.4410699592602094e-05, "loss": 0.225, "step": 462 }, { "epoch": 1.564189189189189, "grad_norm": 0.8952370882034302, "learning_rate": 4.442634985029922e-05, "loss": 0.4926, "step": 463 }, { "epoch": 1.5675675675675675, "grad_norm": 0.43023157119750977, "learning_rate": 4.444196634258136e-05, "loss": 0.2333, "step": 464 }, { "epoch": 1.570945945945946, "grad_norm": 1.1993354558944702, "learning_rate": 4.4457549214832566e-05, "loss": 0.2457, "step": 465 }, { "epoch": 1.5743243243243243, "grad_norm": 0.3993515074253082, "learning_rate": 4.44730986115e-05, "loss": 0.2457, "step": 466 }, { "epoch": 1.5777027027027026, "grad_norm": 0.4565802812576294, "learning_rate": 4.448861467610187e-05, "loss": 0.2335, "step": 467 }, { "epoch": 1.5810810810810811, "grad_norm": 0.4085776209831238, "learning_rate": 4.4504097551235406e-05, "loss": 0.237, "step": 468 }, { "epoch": 1.5844594594594594, "grad_norm": 0.4212399125099182, "learning_rate": 4.4519547378584725e-05, "loss": 0.24, "step": 469 }, { "epoch": 1.5878378378378377, "grad_norm": 0.408805251121521, "learning_rate": 4.453496429892863e-05, "loss": 0.2252, "step": 470 }, { "epoch": 1.5912162162162162, "grad_norm": 0.42911630868911743, "learning_rate": 4.455034845214827e-05, "loss": 0.2099, "step": 471 }, { "epoch": 1.5945945945945947, "grad_norm": 0.47564277052879333, "learning_rate": 4.4565699977234796e-05, "loss": 0.2135, "step": 472 }, { "epoch": 1.597972972972973, "grad_norm": 0.3962784707546234, "learning_rate": 4.458101901229686e-05, "loss": 0.2346, "step": 473 }, { "epoch": 1.6013513513513513, "grad_norm": 0.3911672532558441, "learning_rate": 4.459630569456809e-05, "loss": 0.2198, "step": 474 }, { "epoch": 1.6047297297297298, "grad_norm": 1093.1121826171875, "learning_rate": 4.461156016041444e-05, "loss": 0.2989, "step": 475 }, { "epoch": 1.6081081081081081, "grad_norm": 3.732886552810669, "learning_rate": 4.462678254534156e-05, "loss": 0.5155, "step": 476 }, { "epoch": 1.6114864864864864, "grad_norm": 0.5420750975608826, "learning_rate": 4.464197298400191e-05, "loss": 0.2645, "step": 477 }, { "epoch": 1.614864864864865, "grad_norm": 4.157422065734863, "learning_rate": 4.4657131610201994e-05, "loss": 0.5192, "step": 478 }, { "epoch": 1.6182432432432432, "grad_norm": 0.4917280972003937, "learning_rate": 4.467225855690939e-05, "loss": 0.2187, "step": 479 }, { "epoch": 1.6216216216216215, "grad_norm": 0.4480583071708679, "learning_rate": 4.468735395625979e-05, "loss": 0.2214, "step": 480 }, { "epoch": 1.625, "grad_norm": 0.4467374384403229, "learning_rate": 4.470241793956387e-05, "loss": 0.2414, "step": 481 }, { "epoch": 1.6283783783783785, "grad_norm": 0.48159006237983704, "learning_rate": 4.471745063731416e-05, "loss": 0.2386, "step": 482 }, { "epoch": 1.6317567567567568, "grad_norm": 0.3858594000339508, "learning_rate": 4.473245217919187e-05, "loss": 0.1981, "step": 483 }, { "epoch": 1.635135135135135, "grad_norm": 0.4672364592552185, "learning_rate": 4.474742269407355e-05, "loss": 0.2212, "step": 484 }, { "epoch": 1.6385135135135136, "grad_norm": 0.5253187417984009, "learning_rate": 4.476236231003773e-05, "loss": 0.2753, "step": 485 }, { "epoch": 1.6418918918918919, "grad_norm": 0.488766610622406, "learning_rate": 4.477727115437156e-05, "loss": 0.2571, "step": 486 }, { "epoch": 1.6452702702702702, "grad_norm": 0.4526199996471405, "learning_rate": 4.479214935357724e-05, "loss": 0.2341, "step": 487 }, { "epoch": 1.6486486486486487, "grad_norm": 0.4283704459667206, "learning_rate": 4.480699703337852e-05, "loss": 0.2533, "step": 488 }, { "epoch": 1.652027027027027, "grad_norm": 0.45583993196487427, "learning_rate": 4.4821814318727016e-05, "loss": 0.2394, "step": 489 }, { "epoch": 1.6554054054054053, "grad_norm": 0.3872677683830261, "learning_rate": 4.483660133380856e-05, "loss": 0.2005, "step": 490 }, { "epoch": 1.6587837837837838, "grad_norm": 0.5054983496665955, "learning_rate": 4.485135820204948e-05, "loss": 0.231, "step": 491 }, { "epoch": 1.6621621621621623, "grad_norm": 0.40404146909713745, "learning_rate": 4.486608504612267e-05, "loss": 0.2529, "step": 492 }, { "epoch": 1.6655405405405406, "grad_norm": 0.4361649453639984, "learning_rate": 4.488078198795383e-05, "loss": 0.2485, "step": 493 }, { "epoch": 1.6689189189189189, "grad_norm": 0.4425160586833954, "learning_rate": 4.489544914872745e-05, "loss": 0.2375, "step": 494 }, { "epoch": 1.6722972972972974, "grad_norm": 0.4194391965866089, "learning_rate": 4.4910086648892815e-05, "loss": 0.2461, "step": 495 }, { "epoch": 1.6756756756756757, "grad_norm": 0.42155027389526367, "learning_rate": 4.4924694608169965e-05, "loss": 0.2301, "step": 496 }, { "epoch": 1.679054054054054, "grad_norm": 0.412643164396286, "learning_rate": 4.4939273145555536e-05, "loss": 0.235, "step": 497 }, { "epoch": 1.6824324324324325, "grad_norm": 0.3827148973941803, "learning_rate": 4.495382237932863e-05, "loss": 0.23, "step": 498 }, { "epoch": 1.685810810810811, "grad_norm": 0.4173565208911896, "learning_rate": 4.4968342427056505e-05, "loss": 0.2374, "step": 499 }, { "epoch": 1.689189189189189, "grad_norm": 1055.430908203125, "learning_rate": 4.498283340560031e-05, "loss": 0.3647, "step": 500 }, { "epoch": 1.6925675675675675, "grad_norm": 0.5496036410331726, "learning_rate": 4.499729543112076e-05, "loss": 0.2686, "step": 501 }, { "epoch": 1.695945945945946, "grad_norm": 0.437640905380249, "learning_rate": 4.501172861908366e-05, "loss": 0.2521, "step": 502 }, { "epoch": 1.6993243243243243, "grad_norm": 0.4216068685054779, "learning_rate": 4.502613308426546e-05, "loss": 0.2447, "step": 503 }, { "epoch": 1.7027027027027026, "grad_norm": 0.4138765037059784, "learning_rate": 4.504050894075876e-05, "loss": 0.2379, "step": 504 }, { "epoch": 1.7060810810810811, "grad_norm": 0.8677256107330322, "learning_rate": 4.5054856301977696e-05, "loss": 0.2399, "step": 505 }, { "epoch": 1.7094594594594594, "grad_norm": 0.4616876542568207, "learning_rate": 4.506917528066332e-05, "loss": 0.229, "step": 506 }, { "epoch": 1.7128378378378377, "grad_norm": 0.6452389359474182, "learning_rate": 4.508346598888894e-05, "loss": 0.2237, "step": 507 }, { "epoch": 1.7162162162162162, "grad_norm": 0.4205520749092102, "learning_rate": 4.509772853806532e-05, "loss": 0.2175, "step": 508 }, { "epoch": 1.7195945945945947, "grad_norm": 0.456882506608963, "learning_rate": 4.511196303894598e-05, "loss": 0.2444, "step": 509 }, { "epoch": 1.722972972972973, "grad_norm": 0.43970227241516113, "learning_rate": 4.512616960163227e-05, "loss": 0.2357, "step": 510 }, { "epoch": 1.7263513513513513, "grad_norm": 0.4025394320487976, "learning_rate": 4.5140348335578547e-05, "loss": 0.2375, "step": 511 }, { "epoch": 1.7297297297297298, "grad_norm": 0.39442214369773865, "learning_rate": 4.515449934959718e-05, "loss": 0.2444, "step": 512 }, { "epoch": 1.7331081081081081, "grad_norm": 0.44028565287590027, "learning_rate": 4.516862275186361e-05, "loss": 0.2324, "step": 513 }, { "epoch": 1.7364864864864864, "grad_norm": 0.4466850161552429, "learning_rate": 4.518271864992127e-05, "loss": 0.2359, "step": 514 }, { "epoch": 1.739864864864865, "grad_norm": 0.4365015923976898, "learning_rate": 4.519678715068652e-05, "loss": 0.2708, "step": 515 }, { "epoch": 1.7432432432432432, "grad_norm": 0.4398203194141388, "learning_rate": 4.521082836045353e-05, "loss": 0.2452, "step": 516 }, { "epoch": 1.7466216216216215, "grad_norm": 0.44056662917137146, "learning_rate": 4.5224842384899045e-05, "loss": 0.2344, "step": 517 }, { "epoch": 1.75, "grad_norm": 0.43141037225723267, "learning_rate": 4.523882932908722e-05, "loss": 0.2511, "step": 518 }, { "epoch": 1.7533783783783785, "grad_norm": 0.3922167122364044, "learning_rate": 4.52527892974743e-05, "loss": 0.2162, "step": 519 }, { "epoch": 1.7567567567567568, "grad_norm": 0.40995466709136963, "learning_rate": 4.526672239391333e-05, "loss": 0.2281, "step": 520 }, { "epoch": 1.760135135135135, "grad_norm": 0.4008403718471527, "learning_rate": 4.528062872165875e-05, "loss": 0.2257, "step": 521 }, { "epoch": 1.7635135135135136, "grad_norm": 0.5048766136169434, "learning_rate": 4.529450838337104e-05, "loss": 0.2304, "step": 522 }, { "epoch": 1.7668918918918919, "grad_norm": 0.40119826793670654, "learning_rate": 4.530836148112124e-05, "loss": 0.2341, "step": 523 }, { "epoch": 1.7702702702702702, "grad_norm": 0.41344285011291504, "learning_rate": 4.532218811639545e-05, "loss": 0.272, "step": 524 }, { "epoch": 1.7736486486486487, "grad_norm": 0.38902515172958374, "learning_rate": 4.5335988390099284e-05, "loss": 0.2195, "step": 525 }, { "epoch": 1.777027027027027, "grad_norm": 0.5754374861717224, "learning_rate": 4.534976240256232e-05, "loss": 0.2335, "step": 526 }, { "epoch": 1.7804054054054053, "grad_norm": 31.7464656829834, "learning_rate": 4.536351025354245e-05, "loss": 0.4532, "step": 527 }, { "epoch": 1.7837837837837838, "grad_norm": 0.46500369906425476, "learning_rate": 4.537723204223021e-05, "loss": 0.2541, "step": 528 }, { "epoch": 1.7871621621621623, "grad_norm": 0.41807517409324646, "learning_rate": 4.53909278672531e-05, "loss": 0.2287, "step": 529 }, { "epoch": 1.7905405405405406, "grad_norm": 0.40613386034965515, "learning_rate": 4.5404597826679824e-05, "loss": 0.2196, "step": 530 }, { "epoch": 1.7939189189189189, "grad_norm": 0.4200097322463989, "learning_rate": 4.541824201802449e-05, "loss": 0.2393, "step": 531 }, { "epoch": 1.7972972972972974, "grad_norm": 0.48739388585090637, "learning_rate": 4.543186053825081e-05, "loss": 0.2469, "step": 532 }, { "epoch": 1.8006756756756757, "grad_norm": 0.5199280381202698, "learning_rate": 4.544545348377621e-05, "loss": 0.2652, "step": 533 }, { "epoch": 1.804054054054054, "grad_norm": 0.44481709599494934, "learning_rate": 4.5459020950475946e-05, "loss": 0.2386, "step": 534 }, { "epoch": 1.8074324324324325, "grad_norm": 1.6978758573532104, "learning_rate": 4.5472563033687145e-05, "loss": 0.5287, "step": 535 }, { "epoch": 1.810810810810811, "grad_norm": 0.3990303575992584, "learning_rate": 4.548607982821284e-05, "loss": 0.2102, "step": 536 }, { "epoch": 1.814189189189189, "grad_norm": 0.4208585321903229, "learning_rate": 4.5499571428325935e-05, "loss": 0.2058, "step": 537 }, { "epoch": 1.8175675675675675, "grad_norm": 0.41925048828125, "learning_rate": 4.5513037927773155e-05, "loss": 0.2349, "step": 538 }, { "epoch": 1.820945945945946, "grad_norm": 0.43288394808769226, "learning_rate": 4.5526479419778986e-05, "loss": 0.2519, "step": 539 }, { "epoch": 1.8243243243243243, "grad_norm": 0.4215989112854004, "learning_rate": 4.553989599704948e-05, "loss": 0.2518, "step": 540 }, { "epoch": 1.8277027027027026, "grad_norm": 0.4134393334388733, "learning_rate": 4.555328775177616e-05, "loss": 0.2341, "step": 541 }, { "epoch": 1.8310810810810811, "grad_norm": 0.4243868887424469, "learning_rate": 4.5566654775639785e-05, "loss": 0.2553, "step": 542 }, { "epoch": 1.8344594594594594, "grad_norm": 0.40011635422706604, "learning_rate": 4.5579997159814117e-05, "loss": 0.2447, "step": 543 }, { "epoch": 1.8378378378378377, "grad_norm": 0.38058069348335266, "learning_rate": 4.5593314994969665e-05, "loss": 0.2307, "step": 544 }, { "epoch": 1.8412162162162162, "grad_norm": 0.47068503499031067, "learning_rate": 4.560660837127738e-05, "loss": 0.1971, "step": 545 }, { "epoch": 1.8445945945945947, "grad_norm": 0.4493515193462372, "learning_rate": 4.561987737841229e-05, "loss": 0.2552, "step": 546 }, { "epoch": 1.847972972972973, "grad_norm": 0.43196865916252136, "learning_rate": 4.563312210555719e-05, "loss": 0.2373, "step": 547 }, { "epoch": 1.8513513513513513, "grad_norm": 0.4201470911502838, "learning_rate": 4.564634264140616e-05, "loss": 0.2304, "step": 548 }, { "epoch": 1.8547297297297298, "grad_norm": 0.3680441975593567, "learning_rate": 4.56595390741682e-05, "loss": 0.2263, "step": 549 }, { "epoch": 1.8581081081081081, "grad_norm": 0.3840002119541168, "learning_rate": 4.567271149157073e-05, "loss": 0.218, "step": 550 }, { "epoch": 1.8614864864864864, "grad_norm": 0.39841052889823914, "learning_rate": 4.5685859980863086e-05, "loss": 0.2497, "step": 551 }, { "epoch": 1.864864864864865, "grad_norm": 0.3956766128540039, "learning_rate": 4.569898462881999e-05, "loss": 0.2514, "step": 552 }, { "epoch": 1.8682432432432432, "grad_norm": 0.4630524218082428, "learning_rate": 4.571208552174497e-05, "loss": 0.2744, "step": 553 }, { "epoch": 1.8716216216216215, "grad_norm": 0.4056280851364136, "learning_rate": 4.572516274547383e-05, "loss": 0.2712, "step": 554 }, { "epoch": 1.875, "grad_norm": 0.4127073585987091, "learning_rate": 4.573821638537794e-05, "loss": 0.2319, "step": 555 }, { "epoch": 1.8783783783783785, "grad_norm": 0.38253894448280334, "learning_rate": 4.575124652636763e-05, "loss": 0.2302, "step": 556 }, { "epoch": 1.8817567567567568, "grad_norm": 0.405472993850708, "learning_rate": 4.5764253252895486e-05, "loss": 0.2413, "step": 557 }, { "epoch": 1.885135135135135, "grad_norm": 0.41931191086769104, "learning_rate": 4.577723664895965e-05, "loss": 0.2477, "step": 558 }, { "epoch": 1.8885135135135136, "grad_norm": 0.3967287540435791, "learning_rate": 4.579019679810706e-05, "loss": 0.2525, "step": 559 }, { "epoch": 1.8918918918918919, "grad_norm": 0.3867218792438507, "learning_rate": 4.5803133783436676e-05, "loss": 0.2226, "step": 560 }, { "epoch": 1.8952702702702702, "grad_norm": 0.4048525094985962, "learning_rate": 4.581604768760269e-05, "loss": 0.2434, "step": 561 }, { "epoch": 1.8986486486486487, "grad_norm": 0.38650572299957275, "learning_rate": 4.582893859281769e-05, "loss": 0.2287, "step": 562 }, { "epoch": 1.902027027027027, "grad_norm": 0.3860650062561035, "learning_rate": 4.584180658085578e-05, "loss": 0.2285, "step": 563 }, { "epoch": 1.9054054054054053, "grad_norm": 0.3902740478515625, "learning_rate": 4.585465173305571e-05, "loss": 0.2174, "step": 564 }, { "epoch": 1.9087837837837838, "grad_norm": 0.4053342342376709, "learning_rate": 4.5867474130323984e-05, "loss": 0.2482, "step": 565 }, { "epoch": 1.9121621621621623, "grad_norm": 0.35971924662590027, "learning_rate": 4.588027385313786e-05, "loss": 0.1858, "step": 566 }, { "epoch": 1.9155405405405406, "grad_norm": 0.4442739486694336, "learning_rate": 4.5893050981548446e-05, "loss": 0.2847, "step": 567 }, { "epoch": 1.9189189189189189, "grad_norm": 0.4430129826068878, "learning_rate": 4.5905805595183656e-05, "loss": 0.2265, "step": 568 }, { "epoch": 1.9222972972972974, "grad_norm": 0.39476659893989563, "learning_rate": 4.591853777325119e-05, "loss": 0.258, "step": 569 }, { "epoch": 1.9256756756756757, "grad_norm": 0.4573532044887543, "learning_rate": 4.593124759454153e-05, "loss": 0.2398, "step": 570 }, { "epoch": 1.929054054054054, "grad_norm": 0.3871706426143646, "learning_rate": 4.5943935137430806e-05, "loss": 0.2292, "step": 571 }, { "epoch": 1.9324324324324325, "grad_norm": 0.38287678360939026, "learning_rate": 4.595660047988374e-05, "loss": 0.2313, "step": 572 }, { "epoch": 1.935810810810811, "grad_norm": 0.41778552532196045, "learning_rate": 4.59692436994565e-05, "loss": 0.2501, "step": 573 }, { "epoch": 1.939189189189189, "grad_norm": 0.3646586537361145, "learning_rate": 4.5981864873299563e-05, "loss": 0.2099, "step": 574 }, { "epoch": 1.9425675675675675, "grad_norm": 0.3954225778579712, "learning_rate": 4.599446407816052e-05, "loss": 0.2272, "step": 575 }, { "epoch": 1.945945945945946, "grad_norm": 0.39867308735847473, "learning_rate": 4.6007041390386874e-05, "loss": 0.2352, "step": 576 }, { "epoch": 1.9493243243243243, "grad_norm": 0.3616117238998413, "learning_rate": 4.601959688592886e-05, "loss": 0.2207, "step": 577 }, { "epoch": 1.9527027027027026, "grad_norm": 0.42957910895347595, "learning_rate": 4.603213064034216e-05, "loss": 0.2437, "step": 578 }, { "epoch": 1.9560810810810811, "grad_norm": 0.3899705708026886, "learning_rate": 4.604464272879061e-05, "loss": 0.2436, "step": 579 }, { "epoch": 1.9594594594594594, "grad_norm": 0.3997238278388977, "learning_rate": 4.605713322604896e-05, "loss": 0.2065, "step": 580 }, { "epoch": 1.9628378378378377, "grad_norm": 0.4001472592353821, "learning_rate": 4.606960220650551e-05, "loss": 0.2435, "step": 581 }, { "epoch": 1.9662162162162162, "grad_norm": 0.4067547023296356, "learning_rate": 4.608204974416481e-05, "loss": 0.2456, "step": 582 }, { "epoch": 1.9695945945945947, "grad_norm": 0.3919241428375244, "learning_rate": 4.6094475912650234e-05, "loss": 0.2214, "step": 583 }, { "epoch": 1.972972972972973, "grad_norm": 0.3752463757991791, "learning_rate": 4.610688078520666e-05, "loss": 0.2148, "step": 584 }, { "epoch": 1.9763513513513513, "grad_norm": 0.42536041140556335, "learning_rate": 4.611926443470301e-05, "loss": 0.2219, "step": 585 }, { "epoch": 1.9797297297297298, "grad_norm": 0.3676905930042267, "learning_rate": 4.6131626933634844e-05, "loss": 0.2153, "step": 586 }, { "epoch": 1.9831081081081081, "grad_norm": 0.3774767816066742, "learning_rate": 4.6143968354126914e-05, "loss": 0.2459, "step": 587 }, { "epoch": 1.9864864864864864, "grad_norm": 0.38765907287597656, "learning_rate": 4.6156288767935646e-05, "loss": 0.2339, "step": 588 }, { "epoch": 1.989864864864865, "grad_norm": 0.3474208116531372, "learning_rate": 4.61685882464517e-05, "loss": 0.2109, "step": 589 }, { "epoch": 1.9932432432432432, "grad_norm": 0.39578577876091003, "learning_rate": 4.61808668607024e-05, "loss": 0.2232, "step": 590 }, { "epoch": 1.9966216216216215, "grad_norm": 0.3756243884563446, "learning_rate": 4.619312468135426e-05, "loss": 0.2286, "step": 591 }, { "epoch": 2.0, "grad_norm": 0.40366870164871216, "learning_rate": 4.620536177871533e-05, "loss": 0.1979, "step": 592 }, { "epoch": 2.0033783783783785, "grad_norm": 0.5792023539543152, "learning_rate": 4.621757822273772e-05, "loss": 0.1664, "step": 593 }, { "epoch": 2.0067567567567566, "grad_norm": 0.48093676567077637, "learning_rate": 4.62297740830199e-05, "loss": 0.1489, "step": 594 }, { "epoch": 2.010135135135135, "grad_norm": 0.7113845348358154, "learning_rate": 4.6241949428809165e-05, "loss": 0.1666, "step": 595 }, { "epoch": 2.0135135135135136, "grad_norm": 0.4888218343257904, "learning_rate": 4.625410432900395e-05, "loss": 0.1667, "step": 596 }, { "epoch": 2.016891891891892, "grad_norm": 0.43272024393081665, "learning_rate": 4.626623885215616e-05, "loss": 0.1611, "step": 597 }, { "epoch": 2.02027027027027, "grad_norm": 0.4639625549316406, "learning_rate": 4.627835306647352e-05, "loss": 0.183, "step": 598 }, { "epoch": 2.0236486486486487, "grad_norm": 0.4807380735874176, "learning_rate": 4.629044703982186e-05, "loss": 0.1592, "step": 599 }, { "epoch": 2.027027027027027, "grad_norm": 0.4514772891998291, "learning_rate": 4.63025208397274e-05, "loss": 0.1459, "step": 600 }, { "epoch": 2.0304054054054053, "grad_norm": 0.4227406978607178, "learning_rate": 4.6314574533379e-05, "loss": 0.1632, "step": 601 }, { "epoch": 2.0337837837837838, "grad_norm": 0.42751410603523254, "learning_rate": 4.632660818763041e-05, "loss": 0.1699, "step": 602 }, { "epoch": 2.0371621621621623, "grad_norm": 0.45877569913864136, "learning_rate": 4.633862186900253e-05, "loss": 0.1725, "step": 603 }, { "epoch": 2.0405405405405403, "grad_norm": 0.41468122601509094, "learning_rate": 4.6350615643685535e-05, "loss": 0.172, "step": 604 }, { "epoch": 2.043918918918919, "grad_norm": 0.42624783515930176, "learning_rate": 4.6362589577541154e-05, "loss": 0.1591, "step": 605 }, { "epoch": 2.0472972972972974, "grad_norm": 0.3854513466358185, "learning_rate": 4.637454373610477e-05, "loss": 0.1486, "step": 606 }, { "epoch": 2.050675675675676, "grad_norm": 0.40797704458236694, "learning_rate": 4.638647818458763e-05, "loss": 0.1521, "step": 607 }, { "epoch": 2.054054054054054, "grad_norm": 0.4326710104942322, "learning_rate": 4.639839298787892e-05, "loss": 0.1441, "step": 608 }, { "epoch": 2.0574324324324325, "grad_norm": 0.39950937032699585, "learning_rate": 4.641028821054793e-05, "loss": 0.165, "step": 609 }, { "epoch": 2.060810810810811, "grad_norm": 0.38399842381477356, "learning_rate": 4.6422163916846124e-05, "loss": 0.1464, "step": 610 }, { "epoch": 2.064189189189189, "grad_norm": 0.40971097350120544, "learning_rate": 4.643402017070924e-05, "loss": 0.1604, "step": 611 }, { "epoch": 2.0675675675675675, "grad_norm": 0.4519285261631012, "learning_rate": 4.644585703575936e-05, "loss": 0.1739, "step": 612 }, { "epoch": 2.070945945945946, "grad_norm": 0.40968844294548035, "learning_rate": 4.645767457530692e-05, "loss": 0.1592, "step": 613 }, { "epoch": 2.074324324324324, "grad_norm": 0.4271821975708008, "learning_rate": 4.64694728523528e-05, "loss": 0.1591, "step": 614 }, { "epoch": 2.0777027027027026, "grad_norm": 0.44706177711486816, "learning_rate": 4.648125192959028e-05, "loss": 0.1745, "step": 615 }, { "epoch": 2.081081081081081, "grad_norm": 0.3895583152770996, "learning_rate": 4.649301186940709e-05, "loss": 0.1541, "step": 616 }, { "epoch": 2.0844594594594597, "grad_norm": 0.5112977623939514, "learning_rate": 4.650475273388737e-05, "loss": 0.1851, "step": 617 }, { "epoch": 2.0878378378378377, "grad_norm": 0.4201546609401703, "learning_rate": 4.651647458481359e-05, "loss": 0.1502, "step": 618 }, { "epoch": 2.0912162162162162, "grad_norm": 0.4547381103038788, "learning_rate": 4.652817748366864e-05, "loss": 0.1533, "step": 619 }, { "epoch": 2.0945945945945947, "grad_norm": 0.48063594102859497, "learning_rate": 4.653986149163757e-05, "loss": 0.1704, "step": 620 }, { "epoch": 2.097972972972973, "grad_norm": 0.3941943943500519, "learning_rate": 4.655152666960967e-05, "loss": 0.1338, "step": 621 }, { "epoch": 2.1013513513513513, "grad_norm": 0.43712350726127625, "learning_rate": 4.6563173078180315e-05, "loss": 0.1573, "step": 622 }, { "epoch": 2.10472972972973, "grad_norm": 0.43442657589912415, "learning_rate": 4.657480077765283e-05, "loss": 0.1672, "step": 623 }, { "epoch": 2.108108108108108, "grad_norm": 0.45270854234695435, "learning_rate": 4.6586409828040405e-05, "loss": 0.1851, "step": 624 }, { "epoch": 2.1114864864864864, "grad_norm": 0.5774021744728088, "learning_rate": 4.659800028906792e-05, "loss": 0.1451, "step": 625 }, { "epoch": 2.114864864864865, "grad_norm": 0.4964922368526459, "learning_rate": 4.660957222017383e-05, "loss": 0.1656, "step": 626 }, { "epoch": 2.1182432432432434, "grad_norm": 0.417864054441452, "learning_rate": 4.662112568051194e-05, "loss": 0.1672, "step": 627 }, { "epoch": 2.1216216216216215, "grad_norm": 0.5328717231750488, "learning_rate": 4.663266072895327e-05, "loss": 0.1761, "step": 628 }, { "epoch": 2.125, "grad_norm": 0.4234524369239807, "learning_rate": 4.664417742408782e-05, "loss": 0.1679, "step": 629 }, { "epoch": 2.1283783783783785, "grad_norm": 0.5051260590553284, "learning_rate": 4.665567582422637e-05, "loss": 0.1763, "step": 630 }, { "epoch": 2.1317567567567566, "grad_norm": 0.44952192902565, "learning_rate": 4.666715598740224e-05, "loss": 0.1535, "step": 631 }, { "epoch": 2.135135135135135, "grad_norm": 0.44377925992012024, "learning_rate": 4.667861797137309e-05, "loss": 0.1456, "step": 632 }, { "epoch": 2.1385135135135136, "grad_norm": 0.3786265254020691, "learning_rate": 4.669006183362258e-05, "loss": 0.1478, "step": 633 }, { "epoch": 2.141891891891892, "grad_norm": 0.4277481436729431, "learning_rate": 4.670148763136221e-05, "loss": 0.157, "step": 634 }, { "epoch": 2.14527027027027, "grad_norm": 0.43035444617271423, "learning_rate": 4.671289542153293e-05, "loss": 0.1552, "step": 635 }, { "epoch": 2.1486486486486487, "grad_norm": 0.45837166905403137, "learning_rate": 4.672428526080691e-05, "loss": 0.1848, "step": 636 }, { "epoch": 2.152027027027027, "grad_norm": 0.41413986682891846, "learning_rate": 4.673565720558918e-05, "loss": 0.1803, "step": 637 }, { "epoch": 2.1554054054054053, "grad_norm": 0.4246644377708435, "learning_rate": 4.6747011312019374e-05, "loss": 0.177, "step": 638 }, { "epoch": 2.1587837837837838, "grad_norm": 0.4299694001674652, "learning_rate": 4.6758347635973334e-05, "loss": 0.164, "step": 639 }, { "epoch": 2.1621621621621623, "grad_norm": 0.4426731765270233, "learning_rate": 4.676966623306479e-05, "loss": 0.164, "step": 640 }, { "epoch": 2.1655405405405403, "grad_norm": 0.4473170340061188, "learning_rate": 4.678096715864696e-05, "loss": 0.1642, "step": 641 }, { "epoch": 2.168918918918919, "grad_norm": 0.44445449113845825, "learning_rate": 4.679225046781422e-05, "loss": 0.155, "step": 642 }, { "epoch": 2.1722972972972974, "grad_norm": 0.42584556341171265, "learning_rate": 4.68035162154037e-05, "loss": 0.171, "step": 643 }, { "epoch": 2.175675675675676, "grad_norm": 0.39437368512153625, "learning_rate": 4.681476445599687e-05, "loss": 0.1692, "step": 644 }, { "epoch": 2.179054054054054, "grad_norm": 0.40476444363594055, "learning_rate": 4.6825995243921137e-05, "loss": 0.162, "step": 645 }, { "epoch": 2.1824324324324325, "grad_norm": 0.42337456345558167, "learning_rate": 4.683720863325141e-05, "loss": 0.184, "step": 646 }, { "epoch": 2.185810810810811, "grad_norm": 0.4256432056427002, "learning_rate": 4.684840467781168e-05, "loss": 0.175, "step": 647 }, { "epoch": 2.189189189189189, "grad_norm": 0.40138664841651917, "learning_rate": 4.685958343117656e-05, "loss": 0.1577, "step": 648 }, { "epoch": 2.1925675675675675, "grad_norm": 0.4015848636627197, "learning_rate": 4.6870744946672826e-05, "loss": 0.1521, "step": 649 }, { "epoch": 2.195945945945946, "grad_norm": 0.4208281934261322, "learning_rate": 4.688188927738093e-05, "loss": 0.1685, "step": 650 }, { "epoch": 2.199324324324324, "grad_norm": 0.4102751910686493, "learning_rate": 4.689301647613653e-05, "loss": 0.1837, "step": 651 }, { "epoch": 2.2027027027027026, "grad_norm": 0.43685656785964966, "learning_rate": 4.6904126595532014e-05, "loss": 0.174, "step": 652 }, { "epoch": 2.206081081081081, "grad_norm": 0.392456978559494, "learning_rate": 4.69152196879179e-05, "loss": 0.1741, "step": 653 }, { "epoch": 2.2094594594594597, "grad_norm": 0.4086831212043762, "learning_rate": 4.692629580540446e-05, "loss": 0.1559, "step": 654 }, { "epoch": 2.2128378378378377, "grad_norm": 0.38833364844322205, "learning_rate": 4.693735499986305e-05, "loss": 0.1618, "step": 655 }, { "epoch": 2.2162162162162162, "grad_norm": 0.36838966608047485, "learning_rate": 4.694839732292767e-05, "loss": 0.1478, "step": 656 }, { "epoch": 2.2195945945945947, "grad_norm": 0.40487056970596313, "learning_rate": 4.6959422825996345e-05, "loss": 0.1738, "step": 657 }, { "epoch": 2.222972972972973, "grad_norm": 0.3815852999687195, "learning_rate": 4.69704315602326e-05, "loss": 0.1665, "step": 658 }, { "epoch": 2.2263513513513513, "grad_norm": 0.400643914937973, "learning_rate": 4.698142357656684e-05, "loss": 0.1771, "step": 659 }, { "epoch": 2.22972972972973, "grad_norm": 0.3691990077495575, "learning_rate": 4.6992398925697814e-05, "loss": 0.1591, "step": 660 }, { "epoch": 2.233108108108108, "grad_norm": 0.4398171305656433, "learning_rate": 4.7003357658094e-05, "loss": 0.2062, "step": 661 }, { "epoch": 2.2364864864864864, "grad_norm": 0.3774857521057129, "learning_rate": 4.7014299823995005e-05, "loss": 0.1502, "step": 662 }, { "epoch": 2.239864864864865, "grad_norm": 0.4315558075904846, "learning_rate": 4.702522547341289e-05, "loss": 0.1636, "step": 663 }, { "epoch": 2.2432432432432434, "grad_norm": 0.4040674865245819, "learning_rate": 4.703613465613363e-05, "loss": 0.1625, "step": 664 }, { "epoch": 2.2466216216216215, "grad_norm": 0.46972906589508057, "learning_rate": 4.704702742171841e-05, "loss": 0.1833, "step": 665 }, { "epoch": 2.25, "grad_norm": 0.41066527366638184, "learning_rate": 4.7057903819505024e-05, "loss": 0.1707, "step": 666 }, { "epoch": 2.2533783783783785, "grad_norm": 0.3562461733818054, "learning_rate": 4.7068763898609154e-05, "loss": 0.1508, "step": 667 }, { "epoch": 2.2567567567567566, "grad_norm": 0.4074662923812866, "learning_rate": 4.707960770792576e-05, "loss": 0.1786, "step": 668 }, { "epoch": 2.260135135135135, "grad_norm": 0.3978296220302582, "learning_rate": 4.709043529613039e-05, "loss": 0.1797, "step": 669 }, { "epoch": 2.2635135135135136, "grad_norm": 0.42612069845199585, "learning_rate": 4.710124671168044e-05, "loss": 0.1874, "step": 670 }, { "epoch": 2.266891891891892, "grad_norm": 0.4137380123138428, "learning_rate": 4.711204200281654e-05, "loss": 0.1897, "step": 671 }, { "epoch": 2.27027027027027, "grad_norm": 0.37814861536026, "learning_rate": 4.712282121756376e-05, "loss": 0.1617, "step": 672 }, { "epoch": 2.2736486486486487, "grad_norm": 0.42515134811401367, "learning_rate": 4.713358440373295e-05, "loss": 0.1652, "step": 673 }, { "epoch": 2.277027027027027, "grad_norm": 0.42217129468917847, "learning_rate": 4.7144331608922e-05, "loss": 0.1872, "step": 674 }, { "epoch": 2.2804054054054053, "grad_norm": 0.3996788263320923, "learning_rate": 4.715506288051709e-05, "loss": 0.1603, "step": 675 }, { "epoch": 2.2837837837837838, "grad_norm": 0.4035404622554779, "learning_rate": 4.7165778265693935e-05, "loss": 0.1689, "step": 676 }, { "epoch": 2.2871621621621623, "grad_norm": 0.39791160821914673, "learning_rate": 4.7176477811419076e-05, "loss": 0.1804, "step": 677 }, { "epoch": 2.2905405405405403, "grad_norm": 0.4175347685813904, "learning_rate": 4.718716156445106e-05, "loss": 0.1811, "step": 678 }, { "epoch": 2.293918918918919, "grad_norm": 0.39992383122444153, "learning_rate": 4.7197829571341704e-05, "loss": 0.1805, "step": 679 }, { "epoch": 2.2972972972972974, "grad_norm": 0.40433621406555176, "learning_rate": 4.720848187843727e-05, "loss": 0.1726, "step": 680 }, { "epoch": 2.3006756756756754, "grad_norm": 0.38958102464675903, "learning_rate": 4.721911853187975e-05, "loss": 0.1663, "step": 681 }, { "epoch": 2.304054054054054, "grad_norm": 0.40906044840812683, "learning_rate": 4.722973957760799e-05, "loss": 0.1637, "step": 682 }, { "epoch": 2.3074324324324325, "grad_norm": 0.4296117424964905, "learning_rate": 4.724034506135888e-05, "loss": 0.178, "step": 683 }, { "epoch": 2.310810810810811, "grad_norm": 0.40840208530426025, "learning_rate": 4.725093502866861e-05, "loss": 0.1725, "step": 684 }, { "epoch": 2.314189189189189, "grad_norm": 0.3669198751449585, "learning_rate": 4.7261509524873764e-05, "loss": 0.161, "step": 685 }, { "epoch": 2.3175675675675675, "grad_norm": 0.40982183814048767, "learning_rate": 4.727206859511253e-05, "loss": 0.1757, "step": 686 }, { "epoch": 2.320945945945946, "grad_norm": 0.4490487575531006, "learning_rate": 4.7282612284325846e-05, "loss": 0.1975, "step": 687 }, { "epoch": 2.3243243243243246, "grad_norm": 0.4303133487701416, "learning_rate": 4.729314063725853e-05, "loss": 0.1753, "step": 688 }, { "epoch": 2.3277027027027026, "grad_norm": 0.5020397901535034, "learning_rate": 4.730365369846044e-05, "loss": 0.1959, "step": 689 }, { "epoch": 2.331081081081081, "grad_norm": 0.4409542679786682, "learning_rate": 4.7314151512287594e-05, "loss": 0.1779, "step": 690 }, { "epoch": 2.3344594594594597, "grad_norm": 0.41255807876586914, "learning_rate": 4.732463412290331e-05, "loss": 0.1647, "step": 691 }, { "epoch": 2.3378378378378377, "grad_norm": 0.41047075390815735, "learning_rate": 4.73351015742793e-05, "loss": 0.1699, "step": 692 }, { "epoch": 2.3412162162162162, "grad_norm": 0.4248691201210022, "learning_rate": 4.7345553910196785e-05, "loss": 0.1785, "step": 693 }, { "epoch": 2.3445945945945947, "grad_norm": 0.41576990485191345, "learning_rate": 4.735599117424759e-05, "loss": 0.1806, "step": 694 }, { "epoch": 2.347972972972973, "grad_norm": 0.399854838848114, "learning_rate": 4.7366413409835235e-05, "loss": 0.1689, "step": 695 }, { "epoch": 2.3513513513513513, "grad_norm": 0.39361926913261414, "learning_rate": 4.737682066017604e-05, "loss": 0.1757, "step": 696 }, { "epoch": 2.35472972972973, "grad_norm": 0.36152175068855286, "learning_rate": 4.7387212968300166e-05, "loss": 0.1421, "step": 697 }, { "epoch": 2.358108108108108, "grad_norm": 0.4112738370895386, "learning_rate": 4.7397590377052686e-05, "loss": 0.156, "step": 698 }, { "epoch": 2.3614864864864864, "grad_norm": 1.494449496269226, "learning_rate": 4.74079529290947e-05, "loss": 0.433, "step": 699 }, { "epoch": 2.364864864864865, "grad_norm": 0.44618678092956543, "learning_rate": 4.741830066690428e-05, "loss": 0.1738, "step": 700 }, { "epoch": 2.3682432432432434, "grad_norm": 0.4271227717399597, "learning_rate": 4.742863363277765e-05, "loss": 0.1859, "step": 701 }, { "epoch": 2.3716216216216215, "grad_norm": 0.38577789068222046, "learning_rate": 4.743895186883009e-05, "loss": 0.1699, "step": 702 }, { "epoch": 2.375, "grad_norm": 0.45399177074432373, "learning_rate": 4.7449255416997075e-05, "loss": 0.2042, "step": 703 }, { "epoch": 2.3783783783783785, "grad_norm": 0.4084223806858063, "learning_rate": 4.7459544319035206e-05, "loss": 0.1626, "step": 704 }, { "epoch": 2.3817567567567566, "grad_norm": 0.4444795846939087, "learning_rate": 4.746981861652332e-05, "loss": 0.1524, "step": 705 }, { "epoch": 2.385135135135135, "grad_norm": 0.41611728072166443, "learning_rate": 4.74800783508634e-05, "loss": 0.1965, "step": 706 }, { "epoch": 2.3885135135135136, "grad_norm": 0.4463081657886505, "learning_rate": 4.7490323563281665e-05, "loss": 0.1703, "step": 707 }, { "epoch": 2.391891891891892, "grad_norm": 0.4274112582206726, "learning_rate": 4.750055429482949e-05, "loss": 0.1468, "step": 708 }, { "epoch": 2.39527027027027, "grad_norm": 0.5068261623382568, "learning_rate": 4.751077058638445e-05, "loss": 0.1959, "step": 709 }, { "epoch": 2.3986486486486487, "grad_norm": 0.46266356110572815, "learning_rate": 4.752097247865126e-05, "loss": 0.1876, "step": 710 }, { "epoch": 2.402027027027027, "grad_norm": 0.4017108082771301, "learning_rate": 4.753116001216277e-05, "loss": 0.1567, "step": 711 }, { "epoch": 2.4054054054054053, "grad_norm": 0.4414288103580475, "learning_rate": 4.7541333227280944e-05, "loss": 0.171, "step": 712 }, { "epoch": 2.4087837837837838, "grad_norm": 0.47343650460243225, "learning_rate": 4.755149216419776e-05, "loss": 0.1973, "step": 713 }, { "epoch": 2.4121621621621623, "grad_norm": 0.4446316361427307, "learning_rate": 4.756163686293624e-05, "loss": 0.1804, "step": 714 }, { "epoch": 2.4155405405405403, "grad_norm": 0.4418063163757324, "learning_rate": 4.7571767363351344e-05, "loss": 0.1913, "step": 715 }, { "epoch": 2.418918918918919, "grad_norm": 0.4041652977466583, "learning_rate": 4.758188370513093e-05, "loss": 0.1852, "step": 716 }, { "epoch": 2.4222972972972974, "grad_norm": 0.41937291622161865, "learning_rate": 4.759198592779667e-05, "loss": 0.1925, "step": 717 }, { "epoch": 2.4256756756756754, "grad_norm": 0.4121197462081909, "learning_rate": 4.760207407070501e-05, "loss": 0.1821, "step": 718 }, { "epoch": 2.429054054054054, "grad_norm": 0.3996151387691498, "learning_rate": 4.761214817304805e-05, "loss": 0.1837, "step": 719 }, { "epoch": 2.4324324324324325, "grad_norm": 0.38412001729011536, "learning_rate": 4.762220827385448e-05, "loss": 0.1685, "step": 720 }, { "epoch": 2.435810810810811, "grad_norm": 1.5490686893463135, "learning_rate": 4.763225441199049e-05, "loss": 0.432, "step": 721 }, { "epoch": 2.439189189189189, "grad_norm": 0.4263227880001068, "learning_rate": 4.7642286626160654e-05, "loss": 0.1847, "step": 722 }, { "epoch": 2.4425675675675675, "grad_norm": 0.3754761815071106, "learning_rate": 4.765230495490885e-05, "loss": 0.1786, "step": 723 }, { "epoch": 2.445945945945946, "grad_norm": 0.43871861696243286, "learning_rate": 4.7662309436619115e-05, "loss": 0.189, "step": 724 }, { "epoch": 2.4493243243243246, "grad_norm": 0.4133760333061218, "learning_rate": 4.7672300109516563e-05, "loss": 0.1794, "step": 725 }, { "epoch": 2.4527027027027026, "grad_norm": 0.38480374217033386, "learning_rate": 4.768227701166823e-05, "loss": 0.1666, "step": 726 }, { "epoch": 2.456081081081081, "grad_norm": 0.4138960540294647, "learning_rate": 4.7692240180983964e-05, "loss": 0.1695, "step": 727 }, { "epoch": 2.4594594594594597, "grad_norm": 0.3594178557395935, "learning_rate": 4.770218965521729e-05, "loss": 0.1519, "step": 728 }, { "epoch": 2.4628378378378377, "grad_norm": 0.4005281627178192, "learning_rate": 4.7712125471966245e-05, "loss": 0.1696, "step": 729 }, { "epoch": 2.4662162162162162, "grad_norm": 0.3780256509780884, "learning_rate": 4.7722047668674267e-05, "loss": 0.174, "step": 730 }, { "epoch": 2.4695945945945947, "grad_norm": 0.4074977934360504, "learning_rate": 4.7731956282631004e-05, "loss": 0.1696, "step": 731 }, { "epoch": 2.472972972972973, "grad_norm": 0.4063095152378082, "learning_rate": 4.77418513509732e-05, "loss": 0.1731, "step": 732 }, { "epoch": 2.4763513513513513, "grad_norm": 0.40110018849372864, "learning_rate": 4.775173291068547e-05, "loss": 0.1787, "step": 733 }, { "epoch": 2.47972972972973, "grad_norm": 0.3782990574836731, "learning_rate": 4.776160099860117e-05, "loss": 0.1753, "step": 734 }, { "epoch": 2.483108108108108, "grad_norm": 0.3815288543701172, "learning_rate": 4.777145565140325e-05, "loss": 0.1766, "step": 735 }, { "epoch": 2.4864864864864864, "grad_norm": 0.3779781758785248, "learning_rate": 4.7781296905624986e-05, "loss": 0.1857, "step": 736 }, { "epoch": 2.489864864864865, "grad_norm": 0.33802899718284607, "learning_rate": 4.779112479765086e-05, "loss": 0.1515, "step": 737 }, { "epoch": 2.4932432432432434, "grad_norm": 0.4050140082836151, "learning_rate": 4.780093936371736e-05, "loss": 0.1908, "step": 738 }, { "epoch": 2.4966216216216215, "grad_norm": 0.4146076738834381, "learning_rate": 4.781074063991376e-05, "loss": 0.1812, "step": 739 }, { "epoch": 2.5, "grad_norm": 0.34979405999183655, "learning_rate": 4.782052866218294e-05, "loss": 0.1505, "step": 740 }, { "epoch": 2.5033783783783785, "grad_norm": 0.3686615526676178, "learning_rate": 4.783030346632214e-05, "loss": 0.1656, "step": 741 }, { "epoch": 2.506756756756757, "grad_norm": 0.38285037875175476, "learning_rate": 4.7840065087983786e-05, "loss": 0.181, "step": 742 }, { "epoch": 2.510135135135135, "grad_norm": 0.39826491475105286, "learning_rate": 4.784981356267626e-05, "loss": 0.201, "step": 743 }, { "epoch": 2.5135135135135136, "grad_norm": 0.40031111240386963, "learning_rate": 4.785954892576465e-05, "loss": 0.1676, "step": 744 }, { "epoch": 2.516891891891892, "grad_norm": 0.412266343832016, "learning_rate": 4.7869271212471554e-05, "loss": 0.1807, "step": 745 }, { "epoch": 2.52027027027027, "grad_norm": 0.38458314538002014, "learning_rate": 4.7878980457877814e-05, "loss": 0.1639, "step": 746 }, { "epoch": 2.5236486486486487, "grad_norm": 0.45593783259391785, "learning_rate": 4.7888676696923315e-05, "loss": 0.1844, "step": 747 }, { "epoch": 2.527027027027027, "grad_norm": 0.36613929271698, "learning_rate": 4.7898359964407695e-05, "loss": 0.1536, "step": 748 }, { "epoch": 2.5304054054054053, "grad_norm": 0.4267789125442505, "learning_rate": 4.790803029499111e-05, "loss": 0.1706, "step": 749 }, { "epoch": 2.5337837837837838, "grad_norm": 1.5388500690460205, "learning_rate": 4.7917687723195004e-05, "loss": 0.4571, "step": 750 }, { "epoch": 2.5371621621621623, "grad_norm": 0.42774277925491333, "learning_rate": 4.792733228340281e-05, "loss": 0.1907, "step": 751 }, { "epoch": 2.5405405405405403, "grad_norm": 0.38267603516578674, "learning_rate": 4.793696400986071e-05, "loss": 0.163, "step": 752 }, { "epoch": 2.543918918918919, "grad_norm": 0.4370862543582916, "learning_rate": 4.7946582936678344e-05, "loss": 0.2019, "step": 753 }, { "epoch": 2.5472972972972974, "grad_norm": 0.3963480293750763, "learning_rate": 4.795618909782957e-05, "loss": 0.158, "step": 754 }, { "epoch": 2.5506756756756754, "grad_norm": 0.4094291031360626, "learning_rate": 4.796578252715314e-05, "loss": 0.1769, "step": 755 }, { "epoch": 2.554054054054054, "grad_norm": 0.41850391030311584, "learning_rate": 4.797536325835345e-05, "loss": 0.17, "step": 756 }, { "epoch": 2.5574324324324325, "grad_norm": 0.4879460036754608, "learning_rate": 4.7984931325001216e-05, "loss": 0.2163, "step": 757 }, { "epoch": 2.560810810810811, "grad_norm": 0.4456654489040375, "learning_rate": 4.799448676053423e-05, "loss": 0.2086, "step": 758 }, { "epoch": 2.564189189189189, "grad_norm": 0.41925883293151855, "learning_rate": 4.800402959825802e-05, "loss": 0.1884, "step": 759 }, { "epoch": 2.5675675675675675, "grad_norm": 0.4378669857978821, "learning_rate": 4.801355987134653e-05, "loss": 0.2024, "step": 760 }, { "epoch": 2.570945945945946, "grad_norm": 0.5047132968902588, "learning_rate": 4.802307761284289e-05, "loss": 0.1961, "step": 761 }, { "epoch": 2.5743243243243246, "grad_norm": 0.41319629549980164, "learning_rate": 4.8032582855660014e-05, "loss": 0.1957, "step": 762 }, { "epoch": 2.5777027027027026, "grad_norm": 0.46079912781715393, "learning_rate": 4.8042075632581346e-05, "loss": 0.1931, "step": 763 }, { "epoch": 2.581081081081081, "grad_norm": 0.431325763463974, "learning_rate": 4.80515559762615e-05, "loss": 0.1788, "step": 764 }, { "epoch": 2.5844594594594597, "grad_norm": 0.40309104323387146, "learning_rate": 4.8061023919226964e-05, "loss": 0.1817, "step": 765 }, { "epoch": 2.5878378378378377, "grad_norm": 0.434621661901474, "learning_rate": 4.807047949387674e-05, "loss": 0.1831, "step": 766 }, { "epoch": 2.5912162162162162, "grad_norm": 0.35631585121154785, "learning_rate": 4.807992273248302e-05, "loss": 0.1683, "step": 767 }, { "epoch": 2.5945945945945947, "grad_norm": 0.41151365637779236, "learning_rate": 4.808935366719187e-05, "loss": 0.1904, "step": 768 }, { "epoch": 2.597972972972973, "grad_norm": 0.4089500904083252, "learning_rate": 4.8098772330023855e-05, "loss": 0.1742, "step": 769 }, { "epoch": 2.6013513513513513, "grad_norm": 0.38123077154159546, "learning_rate": 4.81081787528747e-05, "loss": 0.1727, "step": 770 }, { "epoch": 2.60472972972973, "grad_norm": 0.37730783224105835, "learning_rate": 4.811757296751595e-05, "loss": 0.1904, "step": 771 }, { "epoch": 2.608108108108108, "grad_norm": 0.38781440258026123, "learning_rate": 4.812695500559561e-05, "loss": 0.1931, "step": 772 }, { "epoch": 2.6114864864864864, "grad_norm": 0.4114435315132141, "learning_rate": 4.8136324898638756e-05, "loss": 0.2105, "step": 773 }, { "epoch": 2.614864864864865, "grad_norm": 0.39372387528419495, "learning_rate": 4.8145682678048214e-05, "loss": 0.1831, "step": 774 }, { "epoch": 2.618243243243243, "grad_norm": 0.40174078941345215, "learning_rate": 4.815502837510518e-05, "loss": 0.1975, "step": 775 }, { "epoch": 2.6216216216216215, "grad_norm": 0.4061259627342224, "learning_rate": 4.816436202096981e-05, "loss": 0.1744, "step": 776 }, { "epoch": 2.625, "grad_norm": 0.3997017741203308, "learning_rate": 4.81736836466819e-05, "loss": 0.1782, "step": 777 }, { "epoch": 2.6283783783783785, "grad_norm": 0.4183482229709625, "learning_rate": 4.8182993283161485e-05, "loss": 0.1941, "step": 778 }, { "epoch": 2.631756756756757, "grad_norm": 0.3742780387401581, "learning_rate": 4.819229096120941e-05, "loss": 0.1681, "step": 779 }, { "epoch": 2.635135135135135, "grad_norm": 0.3982739746570587, "learning_rate": 4.820157671150801e-05, "loss": 0.1941, "step": 780 }, { "epoch": 2.6385135135135136, "grad_norm": 0.3757505714893341, "learning_rate": 4.821085056462168e-05, "loss": 0.1744, "step": 781 }, { "epoch": 2.641891891891892, "grad_norm": 0.39277997612953186, "learning_rate": 4.822011255099747e-05, "loss": 0.1803, "step": 782 }, { "epoch": 2.64527027027027, "grad_norm": 0.35470277070999146, "learning_rate": 4.8229362700965726e-05, "loss": 0.1651, "step": 783 }, { "epoch": 2.6486486486486487, "grad_norm": 0.381610631942749, "learning_rate": 4.8238601044740645e-05, "loss": 0.183, "step": 784 }, { "epoch": 2.652027027027027, "grad_norm": 0.37985488772392273, "learning_rate": 4.824782761242088e-05, "loss": 0.1637, "step": 785 }, { "epoch": 2.6554054054054053, "grad_norm": 0.4003821909427643, "learning_rate": 4.8257042433990135e-05, "loss": 0.18, "step": 786 }, { "epoch": 2.6587837837837838, "grad_norm": 0.37772974371910095, "learning_rate": 4.826624553931775e-05, "loss": 0.1886, "step": 787 }, { "epoch": 2.6621621621621623, "grad_norm": 0.3801021873950958, "learning_rate": 4.827543695815926e-05, "loss": 0.1718, "step": 788 }, { "epoch": 2.6655405405405403, "grad_norm": 0.3759152591228485, "learning_rate": 4.8284616720157006e-05, "loss": 0.1698, "step": 789 }, { "epoch": 2.668918918918919, "grad_norm": 0.38654980063438416, "learning_rate": 4.82937848548407e-05, "loss": 0.1885, "step": 790 }, { "epoch": 2.6722972972972974, "grad_norm": 0.4393473267555237, "learning_rate": 4.8302941391627947e-05, "loss": 0.1891, "step": 791 }, { "epoch": 2.6756756756756754, "grad_norm": 0.3831874132156372, "learning_rate": 4.83120863598249e-05, "loss": 0.1888, "step": 792 }, { "epoch": 2.679054054054054, "grad_norm": 0.38641834259033203, "learning_rate": 4.832121978862673e-05, "loss": 0.1842, "step": 793 }, { "epoch": 2.6824324324324325, "grad_norm": 0.39984583854675293, "learning_rate": 4.8330341707118276e-05, "loss": 0.1681, "step": 794 }, { "epoch": 2.685810810810811, "grad_norm": 0.40252962708473206, "learning_rate": 4.833945214427451e-05, "loss": 0.1756, "step": 795 }, { "epoch": 2.689189189189189, "grad_norm": 0.40013495087623596, "learning_rate": 4.834855112896116e-05, "loss": 0.2146, "step": 796 }, { "epoch": 2.6925675675675675, "grad_norm": 0.385640412569046, "learning_rate": 4.835763868993521e-05, "loss": 0.1851, "step": 797 }, { "epoch": 2.695945945945946, "grad_norm": 0.3977515697479248, "learning_rate": 4.8366714855845496e-05, "loss": 0.1809, "step": 798 }, { "epoch": 2.6993243243243246, "grad_norm": 0.40790390968322754, "learning_rate": 4.837577965523319e-05, "loss": 0.1887, "step": 799 }, { "epoch": 2.7027027027027026, "grad_norm": 0.3771408498287201, "learning_rate": 4.8384833116532396e-05, "loss": 0.1732, "step": 800 }, { "epoch": 2.706081081081081, "grad_norm": 0.42990434169769287, "learning_rate": 4.8393875268070636e-05, "loss": 0.1952, "step": 801 }, { "epoch": 2.7094594594594597, "grad_norm": 0.38725292682647705, "learning_rate": 4.84029061380694e-05, "loss": 0.1903, "step": 802 }, { "epoch": 2.7128378378378377, "grad_norm": 0.4196310341358185, "learning_rate": 4.841192575464469e-05, "loss": 0.2035, "step": 803 }, { "epoch": 2.7162162162162162, "grad_norm": 0.4179129898548126, "learning_rate": 4.842093414580753e-05, "loss": 0.2002, "step": 804 }, { "epoch": 2.7195945945945947, "grad_norm": 0.44000861048698425, "learning_rate": 4.842993133946448e-05, "loss": 0.2299, "step": 805 }, { "epoch": 2.722972972972973, "grad_norm": 0.4462367296218872, "learning_rate": 4.843891736341818e-05, "loss": 0.1931, "step": 806 }, { "epoch": 2.7263513513513513, "grad_norm": 0.42056065797805786, "learning_rate": 4.8447892245367846e-05, "loss": 0.2012, "step": 807 }, { "epoch": 2.72972972972973, "grad_norm": 0.48042038083076477, "learning_rate": 4.845685601290977e-05, "loss": 0.1988, "step": 808 }, { "epoch": 2.733108108108108, "grad_norm": 0.39594507217407227, "learning_rate": 4.846580869353787e-05, "loss": 0.1966, "step": 809 }, { "epoch": 2.7364864864864864, "grad_norm": 0.4334581792354584, "learning_rate": 4.847475031464416e-05, "loss": 0.1861, "step": 810 }, { "epoch": 2.739864864864865, "grad_norm": 0.40071550011634827, "learning_rate": 4.8483680903519274e-05, "loss": 0.2009, "step": 811 }, { "epoch": 2.743243243243243, "grad_norm": 0.40345826745033264, "learning_rate": 4.8492600487352926e-05, "loss": 0.1692, "step": 812 }, { "epoch": 2.7466216216216215, "grad_norm": 0.41826799511909485, "learning_rate": 4.850150909323447e-05, "loss": 0.1904, "step": 813 }, { "epoch": 2.75, "grad_norm": 0.3715426027774811, "learning_rate": 4.8510406748153355e-05, "loss": 0.1782, "step": 814 }, { "epoch": 2.7533783783783785, "grad_norm": 0.39606815576553345, "learning_rate": 4.8519293478999614e-05, "loss": 0.1824, "step": 815 }, { "epoch": 2.756756756756757, "grad_norm": 0.3894452452659607, "learning_rate": 4.8528169312564355e-05, "loss": 0.1799, "step": 816 }, { "epoch": 2.760135135135135, "grad_norm": 0.38379618525505066, "learning_rate": 4.8537034275540264e-05, "loss": 0.1792, "step": 817 }, { "epoch": 2.7635135135135136, "grad_norm": 0.4054020643234253, "learning_rate": 4.854588839452205e-05, "loss": 0.1842, "step": 818 }, { "epoch": 2.766891891891892, "grad_norm": 0.424482524394989, "learning_rate": 4.855473169600698e-05, "loss": 0.1781, "step": 819 }, { "epoch": 2.77027027027027, "grad_norm": 0.40162649750709534, "learning_rate": 4.856356420639528e-05, "loss": 0.2129, "step": 820 }, { "epoch": 2.7736486486486487, "grad_norm": 0.4196532666683197, "learning_rate": 4.857238595199068e-05, "loss": 0.1844, "step": 821 }, { "epoch": 2.777027027027027, "grad_norm": 0.39450812339782715, "learning_rate": 4.858119695900084e-05, "loss": 0.1917, "step": 822 }, { "epoch": 2.7804054054054053, "grad_norm": 0.37791892886161804, "learning_rate": 4.858999725353783e-05, "loss": 0.1562, "step": 823 }, { "epoch": 2.7837837837837838, "grad_norm": 0.4025024473667145, "learning_rate": 4.8598786861618605e-05, "loss": 0.1963, "step": 824 }, { "epoch": 2.7871621621621623, "grad_norm": 0.382782518863678, "learning_rate": 4.860756580916542e-05, "loss": 0.1811, "step": 825 }, { "epoch": 2.7905405405405403, "grad_norm": 0.40827038884162903, "learning_rate": 4.861633412200637e-05, "loss": 0.1914, "step": 826 }, { "epoch": 2.793918918918919, "grad_norm": 0.35224664211273193, "learning_rate": 4.862509182587578e-05, "loss": 0.1587, "step": 827 }, { "epoch": 2.7972972972972974, "grad_norm": 0.40827634930610657, "learning_rate": 4.863383894641467e-05, "loss": 0.1648, "step": 828 }, { "epoch": 2.8006756756756754, "grad_norm": 0.3934107720851898, "learning_rate": 4.864257550917123e-05, "loss": 0.1867, "step": 829 }, { "epoch": 2.804054054054054, "grad_norm": 0.4319005012512207, "learning_rate": 4.865130153960124e-05, "loss": 0.1789, "step": 830 }, { "epoch": 2.8074324324324325, "grad_norm": 0.4071970582008362, "learning_rate": 4.8660017063068526e-05, "loss": 0.1718, "step": 831 }, { "epoch": 2.810810810810811, "grad_norm": 0.4246816635131836, "learning_rate": 4.8668722104845403e-05, "loss": 0.1911, "step": 832 }, { "epoch": 2.814189189189189, "grad_norm": 0.41370970010757446, "learning_rate": 4.8677416690113134e-05, "loss": 0.1968, "step": 833 }, { "epoch": 2.8175675675675675, "grad_norm": 0.3547184467315674, "learning_rate": 4.868610084396232e-05, "loss": 0.1612, "step": 834 }, { "epoch": 2.820945945945946, "grad_norm": 0.4511527121067047, "learning_rate": 4.869477459139337e-05, "loss": 0.1608, "step": 835 }, { "epoch": 2.8243243243243246, "grad_norm": 0.394466757774353, "learning_rate": 4.870343795731694e-05, "loss": 0.1846, "step": 836 }, { "epoch": 2.8277027027027026, "grad_norm": 0.3896351158618927, "learning_rate": 4.8712090966554334e-05, "loss": 0.1954, "step": 837 }, { "epoch": 2.831081081081081, "grad_norm": 0.4012662172317505, "learning_rate": 4.872073364383795e-05, "loss": 0.1875, "step": 838 }, { "epoch": 2.8344594594594597, "grad_norm": 0.39143961668014526, "learning_rate": 4.8729366013811674e-05, "loss": 0.2066, "step": 839 }, { "epoch": 2.8378378378378377, "grad_norm": 0.4024941325187683, "learning_rate": 4.8737988101031366e-05, "loss": 0.1836, "step": 840 }, { "epoch": 2.8412162162162162, "grad_norm": 0.3429911732673645, "learning_rate": 4.874659992996521e-05, "loss": 0.1687, "step": 841 }, { "epoch": 2.8445945945945947, "grad_norm": 0.3772415816783905, "learning_rate": 4.875520152499416e-05, "loss": 0.2006, "step": 842 }, { "epoch": 2.847972972972973, "grad_norm": 0.41397014260292053, "learning_rate": 4.876379291041238e-05, "loss": 0.2175, "step": 843 }, { "epoch": 2.8513513513513513, "grad_norm": 0.41761255264282227, "learning_rate": 4.8772374110427594e-05, "loss": 0.1779, "step": 844 }, { "epoch": 2.85472972972973, "grad_norm": 0.4116886854171753, "learning_rate": 4.878094514916154e-05, "loss": 0.2, "step": 845 }, { "epoch": 2.858108108108108, "grad_norm": 0.3655768930912018, "learning_rate": 4.8789506050650396e-05, "loss": 0.1818, "step": 846 }, { "epoch": 2.8614864864864864, "grad_norm": 0.43548357486724854, "learning_rate": 4.879805683884512e-05, "loss": 0.2029, "step": 847 }, { "epoch": 2.864864864864865, "grad_norm": 0.3908335566520691, "learning_rate": 4.8806597537611906e-05, "loss": 0.1936, "step": 848 }, { "epoch": 2.868243243243243, "grad_norm": 0.3648586869239807, "learning_rate": 4.881512817073255e-05, "loss": 0.1765, "step": 849 }, { "epoch": 2.8716216216216215, "grad_norm": 0.39892107248306274, "learning_rate": 4.882364876190489e-05, "loss": 0.1594, "step": 850 }, { "epoch": 2.875, "grad_norm": 0.3832027316093445, "learning_rate": 4.8832159334743136e-05, "loss": 0.1773, "step": 851 }, { "epoch": 2.8783783783783785, "grad_norm": 0.4372403621673584, "learning_rate": 4.884065991277833e-05, "loss": 0.2057, "step": 852 }, { "epoch": 2.881756756756757, "grad_norm": 0.36279168725013733, "learning_rate": 4.8849150519458726e-05, "loss": 0.1726, "step": 853 }, { "epoch": 2.885135135135135, "grad_norm": 0.37340691685676575, "learning_rate": 4.885763117815009e-05, "loss": 0.178, "step": 854 }, { "epoch": 2.8885135135135136, "grad_norm": 0.3998451828956604, "learning_rate": 4.886610191213622e-05, "loss": 0.1816, "step": 855 }, { "epoch": 2.891891891891892, "grad_norm": 0.3779732584953308, "learning_rate": 4.887456274461922e-05, "loss": 0.2068, "step": 856 }, { "epoch": 2.89527027027027, "grad_norm": 0.4435446262359619, "learning_rate": 4.8883013698719973e-05, "loss": 0.215, "step": 857 }, { "epoch": 2.8986486486486487, "grad_norm": 0.37998583912849426, "learning_rate": 4.889145479747843e-05, "loss": 0.181, "step": 858 }, { "epoch": 2.902027027027027, "grad_norm": 0.4032045900821686, "learning_rate": 4.889988606385404e-05, "loss": 0.1878, "step": 859 }, { "epoch": 2.9054054054054053, "grad_norm": 0.4383464753627777, "learning_rate": 4.8908307520726135e-05, "loss": 0.1948, "step": 860 }, { "epoch": 2.9087837837837838, "grad_norm": 0.38737547397613525, "learning_rate": 4.891671919089425e-05, "loss": 0.2007, "step": 861 }, { "epoch": 2.9121621621621623, "grad_norm": 0.39179664850234985, "learning_rate": 4.892512109707855e-05, "loss": 0.1865, "step": 862 }, { "epoch": 2.9155405405405403, "grad_norm": 0.3944191038608551, "learning_rate": 4.893351326192016e-05, "loss": 0.1831, "step": 863 }, { "epoch": 2.918918918918919, "grad_norm": 0.3212383985519409, "learning_rate": 4.894189570798156e-05, "loss": 0.1494, "step": 864 }, { "epoch": 2.9222972972972974, "grad_norm": 0.3861388564109802, "learning_rate": 4.895026845774691e-05, "loss": 0.2045, "step": 865 }, { "epoch": 2.9256756756756754, "grad_norm": 0.37900376319885254, "learning_rate": 4.895863153362244e-05, "loss": 0.1818, "step": 866 }, { "epoch": 2.929054054054054, "grad_norm": 0.4495367109775543, "learning_rate": 4.896698495793684e-05, "loss": 0.1936, "step": 867 }, { "epoch": 2.9324324324324325, "grad_norm": 0.40020546317100525, "learning_rate": 4.897532875294154e-05, "loss": 0.1964, "step": 868 }, { "epoch": 2.935810810810811, "grad_norm": 0.3625187277793884, "learning_rate": 4.8983662940811115e-05, "loss": 0.1737, "step": 869 }, { "epoch": 2.939189189189189, "grad_norm": 0.3938811719417572, "learning_rate": 4.899198754364365e-05, "loss": 0.1488, "step": 870 }, { "epoch": 2.9425675675675675, "grad_norm": 0.36418697237968445, "learning_rate": 4.900030258346106e-05, "loss": 0.1822, "step": 871 }, { "epoch": 2.945945945945946, "grad_norm": 0.3703506588935852, "learning_rate": 4.900860808220946e-05, "loss": 0.1846, "step": 872 }, { "epoch": 2.9493243243243246, "grad_norm": 0.4332616925239563, "learning_rate": 4.90169040617595e-05, "loss": 0.2105, "step": 873 }, { "epoch": 2.9527027027027026, "grad_norm": 0.3770284354686737, "learning_rate": 4.9025190543906715e-05, "loss": 0.1815, "step": 874 }, { "epoch": 2.956081081081081, "grad_norm": 0.3703818619251251, "learning_rate": 4.903346755037189e-05, "loss": 0.1713, "step": 875 }, { "epoch": 2.9594594594594597, "grad_norm": 0.37928834557533264, "learning_rate": 4.904173510280135e-05, "loss": 0.1915, "step": 876 }, { "epoch": 2.9628378378378377, "grad_norm": 0.4267534911632538, "learning_rate": 4.904999322276735e-05, "loss": 0.2157, "step": 877 }, { "epoch": 2.9662162162162162, "grad_norm": 0.3936428725719452, "learning_rate": 4.9058241931768385e-05, "loss": 0.1947, "step": 878 }, { "epoch": 2.9695945945945947, "grad_norm": 0.3503859043121338, "learning_rate": 4.9066481251229535e-05, "loss": 0.1836, "step": 879 }, { "epoch": 2.972972972972973, "grad_norm": 1.8326294422149658, "learning_rate": 4.907471120250281e-05, "loss": 0.4757, "step": 880 }, { "epoch": 2.9763513513513513, "grad_norm": 0.4133761525154114, "learning_rate": 4.9082931806867474e-05, "loss": 0.2003, "step": 881 }, { "epoch": 2.97972972972973, "grad_norm": 0.3848228454589844, "learning_rate": 4.909114308553033e-05, "loss": 0.19, "step": 882 }, { "epoch": 2.983108108108108, "grad_norm": 0.38853901624679565, "learning_rate": 4.909934505962615e-05, "loss": 0.2071, "step": 883 }, { "epoch": 2.9864864864864864, "grad_norm": 0.35326841473579407, "learning_rate": 4.9107537750217886e-05, "loss": 0.1742, "step": 884 }, { "epoch": 2.989864864864865, "grad_norm": 0.33690720796585083, "learning_rate": 4.9115721178297093e-05, "loss": 0.1734, "step": 885 }, { "epoch": 2.993243243243243, "grad_norm": 0.35074329376220703, "learning_rate": 4.9123895364784184e-05, "loss": 0.1623, "step": 886 }, { "epoch": 2.9966216216216215, "grad_norm": 0.35647299885749817, "learning_rate": 4.913206033052877e-05, "loss": 0.1647, "step": 887 }, { "epoch": 3.0, "grad_norm": 0.4068041145801544, "learning_rate": 4.914021609631002e-05, "loss": 0.1831, "step": 888 }, { "epoch": 3.0033783783783785, "grad_norm": 0.4528217017650604, "learning_rate": 4.91483626828369e-05, "loss": 0.128, "step": 889 }, { "epoch": 3.0067567567567566, "grad_norm": 0.3695124089717865, "learning_rate": 4.915650011074855e-05, "loss": 0.1077, "step": 890 }, { "epoch": 3.010135135135135, "grad_norm": 0.570489227771759, "learning_rate": 4.916462840061458e-05, "loss": 0.11, "step": 891 }, { "epoch": 3.0135135135135136, "grad_norm": 0.43204137682914734, "learning_rate": 4.917274757293539e-05, "loss": 0.1032, "step": 892 }, { "epoch": 3.016891891891892, "grad_norm": 0.5314778089523315, "learning_rate": 4.918085764814244e-05, "loss": 0.1304, "step": 893 }, { "epoch": 3.02027027027027, "grad_norm": 0.3921876847743988, "learning_rate": 4.9188958646598624e-05, "loss": 0.0949, "step": 894 }, { "epoch": 3.0236486486486487, "grad_norm": 0.35549548268318176, "learning_rate": 4.919705058859854e-05, "loss": 0.0989, "step": 895 }, { "epoch": 3.027027027027027, "grad_norm": 0.4394858181476593, "learning_rate": 4.920513349436875e-05, "loss": 0.1152, "step": 896 }, { "epoch": 3.0304054054054053, "grad_norm": 0.4386696219444275, "learning_rate": 4.92132073840682e-05, "loss": 0.1163, "step": 897 }, { "epoch": 3.0337837837837838, "grad_norm": 0.38169318437576294, "learning_rate": 4.922127227778841e-05, "loss": 0.1078, "step": 898 }, { "epoch": 3.0371621621621623, "grad_norm": 0.410858690738678, "learning_rate": 4.9229328195553815e-05, "loss": 0.1165, "step": 899 }, { "epoch": 3.0405405405405403, "grad_norm": 0.4674147069454193, "learning_rate": 4.923737515732209e-05, "loss": 0.12, "step": 900 }, { "epoch": 3.043918918918919, "grad_norm": 0.3922157883644104, "learning_rate": 4.924541318298438e-05, "loss": 0.1131, "step": 901 }, { "epoch": 3.0472972972972974, "grad_norm": 0.3732883632183075, "learning_rate": 4.92534422923657e-05, "loss": 0.1025, "step": 902 }, { "epoch": 3.050675675675676, "grad_norm": 0.4513222277164459, "learning_rate": 4.9261462505225106e-05, "loss": 0.1207, "step": 903 }, { "epoch": 3.054054054054054, "grad_norm": 0.39665260910987854, "learning_rate": 4.926947384125606e-05, "loss": 0.1115, "step": 904 }, { "epoch": 3.0574324324324325, "grad_norm": 0.3784550130367279, "learning_rate": 4.927747632008672e-05, "loss": 0.1151, "step": 905 }, { "epoch": 3.060810810810811, "grad_norm": 0.44802817702293396, "learning_rate": 4.9285469961280226e-05, "loss": 0.113, "step": 906 }, { "epoch": 3.064189189189189, "grad_norm": 0.37687918543815613, "learning_rate": 4.9293454784334924e-05, "loss": 0.1157, "step": 907 }, { "epoch": 3.0675675675675675, "grad_norm": 0.3582609295845032, "learning_rate": 4.9301430808684754e-05, "loss": 0.1031, "step": 908 }, { "epoch": 3.070945945945946, "grad_norm": 0.3885122239589691, "learning_rate": 4.930939805369946e-05, "loss": 0.1043, "step": 909 }, { "epoch": 3.074324324324324, "grad_norm": 0.5625414252281189, "learning_rate": 4.93173565386849e-05, "loss": 0.1109, "step": 910 }, { "epoch": 3.0777027027027026, "grad_norm": 0.3783611059188843, "learning_rate": 4.932530628288331e-05, "loss": 0.1047, "step": 911 }, { "epoch": 3.081081081081081, "grad_norm": 0.48341429233551025, "learning_rate": 4.933324730547361e-05, "loss": 0.0961, "step": 912 }, { "epoch": 3.0844594594594597, "grad_norm": 0.399813711643219, "learning_rate": 4.934117962557165e-05, "loss": 0.1178, "step": 913 }, { "epoch": 3.0878378378378377, "grad_norm": 0.40703117847442627, "learning_rate": 4.9349103262230524e-05, "loss": 0.1179, "step": 914 }, { "epoch": 3.0912162162162162, "grad_norm": 0.39595040678977966, "learning_rate": 4.935701823444081e-05, "loss": 0.1161, "step": 915 }, { "epoch": 3.0945945945945947, "grad_norm": 0.3974362313747406, "learning_rate": 4.9364924561130845e-05, "loss": 0.1167, "step": 916 }, { "epoch": 3.097972972972973, "grad_norm": 0.426070898771286, "learning_rate": 4.937282226116702e-05, "loss": 0.1305, "step": 917 }, { "epoch": 3.1013513513513513, "grad_norm": 0.38482606410980225, "learning_rate": 4.938071135335405e-05, "loss": 0.1018, "step": 918 }, { "epoch": 3.10472972972973, "grad_norm": 0.4263748228549957, "learning_rate": 4.938859185643519e-05, "loss": 0.1156, "step": 919 }, { "epoch": 3.108108108108108, "grad_norm": 0.39638155698776245, "learning_rate": 4.939646378909259e-05, "loss": 0.1157, "step": 920 }, { "epoch": 3.1114864864864864, "grad_norm": 0.40607044100761414, "learning_rate": 4.940432716994748e-05, "loss": 0.1197, "step": 921 }, { "epoch": 3.114864864864865, "grad_norm": 0.4135359823703766, "learning_rate": 4.9412182017560496e-05, "loss": 0.1202, "step": 922 }, { "epoch": 3.1182432432432434, "grad_norm": 0.38252270221710205, "learning_rate": 4.942002835043187e-05, "loss": 0.1126, "step": 923 }, { "epoch": 3.1216216216216215, "grad_norm": 0.3727289140224457, "learning_rate": 4.942786618700178e-05, "loss": 0.1111, "step": 924 }, { "epoch": 3.125, "grad_norm": 0.380953848361969, "learning_rate": 4.9435695545650545e-05, "loss": 0.1151, "step": 925 }, { "epoch": 3.1283783783783785, "grad_norm": 0.37086960673332214, "learning_rate": 4.944351644469891e-05, "loss": 0.1106, "step": 926 }, { "epoch": 3.1317567567567566, "grad_norm": 0.4138876497745514, "learning_rate": 4.945132890240829e-05, "loss": 0.1239, "step": 927 }, { "epoch": 3.135135135135135, "grad_norm": 0.4137141704559326, "learning_rate": 4.945913293698104e-05, "loss": 0.1252, "step": 928 }, { "epoch": 3.1385135135135136, "grad_norm": 0.3904976546764374, "learning_rate": 4.9466928566560696e-05, "loss": 0.107, "step": 929 }, { "epoch": 3.141891891891892, "grad_norm": 0.3888641595840454, "learning_rate": 4.9474715809232256e-05, "loss": 0.113, "step": 930 }, { "epoch": 3.14527027027027, "grad_norm": 0.39176711440086365, "learning_rate": 4.948249468302239e-05, "loss": 0.1135, "step": 931 }, { "epoch": 3.1486486486486487, "grad_norm": 0.3809608817100525, "learning_rate": 4.9490265205899697e-05, "loss": 0.1051, "step": 932 }, { "epoch": 3.152027027027027, "grad_norm": 0.4019430875778198, "learning_rate": 4.9498027395775006e-05, "loss": 0.121, "step": 933 }, { "epoch": 3.1554054054054053, "grad_norm": 0.3715710937976837, "learning_rate": 4.950578127050156e-05, "loss": 0.1225, "step": 934 }, { "epoch": 3.1587837837837838, "grad_norm": 0.4521862864494324, "learning_rate": 4.95135268478753e-05, "loss": 0.1294, "step": 935 }, { "epoch": 3.1621621621621623, "grad_norm": 0.4279133379459381, "learning_rate": 4.952126414563509e-05, "loss": 0.1229, "step": 936 }, { "epoch": 3.1655405405405403, "grad_norm": 0.4043607711791992, "learning_rate": 4.952899318146297e-05, "loss": 0.1117, "step": 937 }, { "epoch": 3.168918918918919, "grad_norm": 0.34410539269447327, "learning_rate": 4.9536713972984414e-05, "loss": 0.0972, "step": 938 }, { "epoch": 3.1722972972972974, "grad_norm": 0.43061432242393494, "learning_rate": 4.954442653776852e-05, "loss": 0.1333, "step": 939 }, { "epoch": 3.175675675675676, "grad_norm": 0.36851924657821655, "learning_rate": 4.955213089332832e-05, "loss": 0.1114, "step": 940 }, { "epoch": 3.179054054054054, "grad_norm": 0.37471747398376465, "learning_rate": 4.955982705712095e-05, "loss": 0.1169, "step": 941 }, { "epoch": 3.1824324324324325, "grad_norm": 0.35380908846855164, "learning_rate": 4.956751504654796e-05, "loss": 0.1005, "step": 942 }, { "epoch": 3.185810810810811, "grad_norm": 0.38561439514160156, "learning_rate": 4.957519487895548e-05, "loss": 0.1187, "step": 943 }, { "epoch": 3.189189189189189, "grad_norm": 0.4120042324066162, "learning_rate": 4.9582866571634485e-05, "loss": 0.1273, "step": 944 }, { "epoch": 3.1925675675675675, "grad_norm": 0.39863529801368713, "learning_rate": 4.959053014182106e-05, "loss": 0.1125, "step": 945 }, { "epoch": 3.195945945945946, "grad_norm": 0.38955143094062805, "learning_rate": 4.959818560669655e-05, "loss": 0.1175, "step": 946 }, { "epoch": 3.199324324324324, "grad_norm": 0.3573088049888611, "learning_rate": 4.96058329833879e-05, "loss": 0.1131, "step": 947 }, { "epoch": 3.2027027027027026, "grad_norm": 0.4304436147212982, "learning_rate": 4.961347228896777e-05, "loss": 0.1262, "step": 948 }, { "epoch": 3.206081081081081, "grad_norm": 0.39087799191474915, "learning_rate": 4.962110354045488e-05, "loss": 0.1191, "step": 949 }, { "epoch": 3.2094594594594597, "grad_norm": 0.3944397568702698, "learning_rate": 4.962872675481414e-05, "loss": 0.1182, "step": 950 }, { "epoch": 3.2128378378378377, "grad_norm": 0.40278294682502747, "learning_rate": 4.9636341948956906e-05, "loss": 0.1211, "step": 951 }, { "epoch": 3.2162162162162162, "grad_norm": 0.41054508090019226, "learning_rate": 4.964394913974124e-05, "loss": 0.1195, "step": 952 }, { "epoch": 3.2195945945945947, "grad_norm": 0.4175059199333191, "learning_rate": 4.965154834397211e-05, "loss": 0.1353, "step": 953 }, { "epoch": 3.222972972972973, "grad_norm": 0.3865712583065033, "learning_rate": 4.965913957840159e-05, "loss": 0.1111, "step": 954 }, { "epoch": 3.2263513513513513, "grad_norm": 0.39877578616142273, "learning_rate": 4.966672285972911e-05, "loss": 0.1256, "step": 955 }, { "epoch": 3.22972972972973, "grad_norm": 0.34512677788734436, "learning_rate": 4.967429820460167e-05, "loss": 0.1078, "step": 956 }, { "epoch": 3.233108108108108, "grad_norm": 0.4256219267845154, "learning_rate": 4.9681865629614064e-05, "loss": 0.1313, "step": 957 }, { "epoch": 3.2364864864864864, "grad_norm": 0.46931833028793335, "learning_rate": 4.9689425151309074e-05, "loss": 0.1396, "step": 958 }, { "epoch": 3.239864864864865, "grad_norm": 0.40905606746673584, "learning_rate": 4.969697678617773e-05, "loss": 0.1242, "step": 959 }, { "epoch": 3.2432432432432434, "grad_norm": 0.36054709553718567, "learning_rate": 4.970452055065948e-05, "loss": 0.1149, "step": 960 }, { "epoch": 3.2466216216216215, "grad_norm": 0.3850068747997284, "learning_rate": 4.9712056461142423e-05, "loss": 0.1074, "step": 961 }, { "epoch": 3.25, "grad_norm": 0.42875200510025024, "learning_rate": 4.971958453396355e-05, "loss": 0.1152, "step": 962 }, { "epoch": 3.2533783783783785, "grad_norm": 0.413173109292984, "learning_rate": 4.972710478540891e-05, "loss": 0.1273, "step": 963 }, { "epoch": 3.2567567567567566, "grad_norm": 0.3743407428264618, "learning_rate": 4.973461723171385e-05, "loss": 0.1149, "step": 964 }, { "epoch": 3.260135135135135, "grad_norm": 0.44052329659461975, "learning_rate": 4.9742121889063213e-05, "loss": 0.1301, "step": 965 }, { "epoch": 3.2635135135135136, "grad_norm": 0.38358744978904724, "learning_rate": 4.974961877359156e-05, "loss": 0.1133, "step": 966 }, { "epoch": 3.266891891891892, "grad_norm": 0.423977792263031, "learning_rate": 4.975710790138336e-05, "loss": 0.1346, "step": 967 }, { "epoch": 3.27027027027027, "grad_norm": 1.5824426412582397, "learning_rate": 4.976458928847323e-05, "loss": 0.3884, "step": 968 }, { "epoch": 3.2736486486486487, "grad_norm": 0.41981515288352966, "learning_rate": 4.977206295084609e-05, "loss": 0.1164, "step": 969 }, { "epoch": 3.277027027027027, "grad_norm": 0.42901647090911865, "learning_rate": 4.9779528904437424e-05, "loss": 0.1269, "step": 970 }, { "epoch": 3.2804054054054053, "grad_norm": 0.3595876693725586, "learning_rate": 4.978698716513342e-05, "loss": 0.1121, "step": 971 }, { "epoch": 3.2837837837837838, "grad_norm": 0.39047256112098694, "learning_rate": 4.9794437748771244e-05, "loss": 0.1219, "step": 972 }, { "epoch": 3.2871621621621623, "grad_norm": 0.41640281677246094, "learning_rate": 4.9801880671139204e-05, "loss": 0.1303, "step": 973 }, { "epoch": 3.2905405405405403, "grad_norm": 0.371383398771286, "learning_rate": 4.980931594797693e-05, "loss": 0.112, "step": 974 }, { "epoch": 3.293918918918919, "grad_norm": 0.4297455847263336, "learning_rate": 4.981674359497562e-05, "loss": 0.1326, "step": 975 }, { "epoch": 3.2972972972972974, "grad_norm": 0.39308154582977295, "learning_rate": 4.98241636277782e-05, "loss": 0.1244, "step": 976 }, { "epoch": 3.3006756756756754, "grad_norm": 0.3956491947174072, "learning_rate": 4.983157606197955e-05, "loss": 0.1203, "step": 977 }, { "epoch": 3.304054054054054, "grad_norm": 0.41044941544532776, "learning_rate": 4.98389809131267e-05, "loss": 0.1247, "step": 978 }, { "epoch": 3.3074324324324325, "grad_norm": 0.4159061312675476, "learning_rate": 4.984637819671897e-05, "loss": 0.1321, "step": 979 }, { "epoch": 3.310810810810811, "grad_norm": 0.3757896423339844, "learning_rate": 4.985376792820825e-05, "loss": 0.1157, "step": 980 }, { "epoch": 3.314189189189189, "grad_norm": 0.3729749321937561, "learning_rate": 4.986115012299915e-05, "loss": 0.1103, "step": 981 }, { "epoch": 3.3175675675675675, "grad_norm": 0.3994838297367096, "learning_rate": 4.986852479644916e-05, "loss": 0.1284, "step": 982 }, { "epoch": 3.320945945945946, "grad_norm": 0.3987760841846466, "learning_rate": 4.987589196386893e-05, "loss": 0.1126, "step": 983 }, { "epoch": 3.3243243243243246, "grad_norm": 0.3847789168357849, "learning_rate": 4.988325164052236e-05, "loss": 0.1199, "step": 984 }, { "epoch": 3.3277027027027026, "grad_norm": 0.41571831703186035, "learning_rate": 4.9890603841626866e-05, "loss": 0.1295, "step": 985 }, { "epoch": 3.331081081081081, "grad_norm": 0.3801005184650421, "learning_rate": 4.989794858235352e-05, "loss": 0.1276, "step": 986 }, { "epoch": 3.3344594594594597, "grad_norm": 0.4302000403404236, "learning_rate": 4.990528587782729e-05, "loss": 0.1355, "step": 987 }, { "epoch": 3.3378378378378377, "grad_norm": 0.3750097155570984, "learning_rate": 4.9912615743127146e-05, "loss": 0.1159, "step": 988 }, { "epoch": 3.3412162162162162, "grad_norm": 0.43669402599334717, "learning_rate": 4.991993819328633e-05, "loss": 0.1349, "step": 989 }, { "epoch": 3.3445945945945947, "grad_norm": 0.40606677532196045, "learning_rate": 4.9927253243292505e-05, "loss": 0.123, "step": 990 }, { "epoch": 3.347972972972973, "grad_norm": 0.40521538257598877, "learning_rate": 4.993456090808793e-05, "loss": 0.1212, "step": 991 }, { "epoch": 3.3513513513513513, "grad_norm": 0.38929492235183716, "learning_rate": 4.994186120256965e-05, "loss": 0.1204, "step": 992 }, { "epoch": 3.35472972972973, "grad_norm": 0.3639843761920929, "learning_rate": 4.9949154141589696e-05, "loss": 0.1175, "step": 993 }, { "epoch": 3.358108108108108, "grad_norm": 0.3805122971534729, "learning_rate": 4.995643973995523e-05, "loss": 0.1198, "step": 994 }, { "epoch": 3.3614864864864864, "grad_norm": 0.395796000957489, "learning_rate": 4.9963718012428765e-05, "loss": 0.1348, "step": 995 }, { "epoch": 3.364864864864865, "grad_norm": 0.3744940459728241, "learning_rate": 4.9970988973728314e-05, "loss": 0.1104, "step": 996 }, { "epoch": 3.3682432432432434, "grad_norm": 1.3542141914367676, "learning_rate": 4.99782526385276e-05, "loss": 0.3864, "step": 997 }, { "epoch": 3.3716216216216215, "grad_norm": 0.42771685123443604, "learning_rate": 4.998550902145619e-05, "loss": 0.1383, "step": 998 }, { "epoch": 3.375, "grad_norm": 0.3889855444431305, "learning_rate": 4.999275813709971e-05, "loss": 0.1318, "step": 999 }, { "epoch": 3.3783783783783785, "grad_norm": 0.37391164898872375, "learning_rate": 5e-05, "loss": 0.1285, "step": 1000 }, { "epoch": 3.3817567567567566, "grad_norm": 0.3580280542373657, "learning_rate": 4.9996356488619556e-05, "loss": 0.1125, "step": 1001 }, { "epoch": 3.385135135135135, "grad_norm": 0.3707447052001953, "learning_rate": 4.9985427016598435e-05, "loss": 0.1221, "step": 1002 }, { "epoch": 3.3885135135135136, "grad_norm": 0.3772697150707245, "learning_rate": 4.996721476998771e-05, "loss": 0.1215, "step": 1003 }, { "epoch": 3.391891891891892, "grad_norm": 0.40340831875801086, "learning_rate": 4.9941725057840504e-05, "loss": 0.1233, "step": 1004 }, { "epoch": 3.39527027027027, "grad_norm": 0.3812180161476135, "learning_rate": 4.9908965310664374e-05, "loss": 0.1149, "step": 1005 }, { "epoch": 3.3986486486486487, "grad_norm": 0.4367975890636444, "learning_rate": 4.986894507825522e-05, "loss": 0.1358, "step": 1006 }, { "epoch": 3.402027027027027, "grad_norm": 0.4075244963169098, "learning_rate": 4.9821676026913475e-05, "loss": 0.132, "step": 1007 }, { "epoch": 3.4054054054054053, "grad_norm": 0.392610639333725, "learning_rate": 4.9767171936043175e-05, "loss": 0.1169, "step": 1008 }, { "epoch": 3.4087837837837838, "grad_norm": 0.3933659791946411, "learning_rate": 4.970544869413522e-05, "loss": 0.1197, "step": 1009 }, { "epoch": 3.4121621621621623, "grad_norm": 0.3939339816570282, "learning_rate": 4.963652429413563e-05, "loss": 0.1315, "step": 1010 }, { "epoch": 3.4155405405405403, "grad_norm": 0.37384697794914246, "learning_rate": 4.9560418828200494e-05, "loss": 0.1187, "step": 1011 }, { "epoch": 3.418918918918919, "grad_norm": 0.3945925831794739, "learning_rate": 4.9477154481838875e-05, "loss": 0.1353, "step": 1012 }, { "epoch": 3.4222972972972974, "grad_norm": 0.42826682329177856, "learning_rate": 4.9386755527445475e-05, "loss": 0.1459, "step": 1013 }, { "epoch": 3.4256756756756754, "grad_norm": 0.4206187427043915, "learning_rate": 4.928924831722504e-05, "loss": 0.1292, "step": 1014 }, { "epoch": 3.429054054054054, "grad_norm": 0.38580143451690674, "learning_rate": 4.9184661275510446e-05, "loss": 0.1153, "step": 1015 }, { "epoch": 3.4324324324324325, "grad_norm": 0.393794983625412, "learning_rate": 4.907302489047662e-05, "loss": 0.1179, "step": 1016 }, { "epoch": 3.435810810810811, "grad_norm": 0.44305288791656494, "learning_rate": 4.895437170525303e-05, "loss": 0.1302, "step": 1017 }, { "epoch": 3.439189189189189, "grad_norm": 0.4056559205055237, "learning_rate": 4.882873630843699e-05, "loss": 0.1302, "step": 1018 }, { "epoch": 3.4425675675675675, "grad_norm": 0.40912213921546936, "learning_rate": 4.869615532401074e-05, "loss": 0.1327, "step": 1019 }, { "epoch": 3.445945945945946, "grad_norm": 0.38370949029922485, "learning_rate": 4.855666740066522e-05, "loss": 0.129, "step": 1020 }, { "epoch": 3.4493243243243246, "grad_norm": 0.41348859667778015, "learning_rate": 4.841031320053351e-05, "loss": 0.1243, "step": 1021 }, { "epoch": 3.4527027027027026, "grad_norm": 0.40817150473594666, "learning_rate": 4.825713538733748e-05, "loss": 0.1321, "step": 1022 }, { "epoch": 3.456081081081081, "grad_norm": 31.2849063873291, "learning_rate": 4.80971786139509e-05, "loss": 0.186, "step": 1023 }, { "epoch": 3.4594594594594597, "grad_norm": 8.919435501098633, "learning_rate": 4.793048950938256e-05, "loss": 0.6012, "step": 1024 }, { "epoch": 3.4628378378378377, "grad_norm": 1.2597789764404297, "learning_rate": 4.7757116665183614e-05, "loss": 0.373, "step": 1025 }, { "epoch": 3.4662162162162162, "grad_norm": 144.5867156982422, "learning_rate": 4.757711062128251e-05, "loss": 0.6334, "step": 1026 }, { "epoch": 3.4695945945945947, "grad_norm": 42.916805267333984, "learning_rate": 4.739052385125216e-05, "loss": 0.6214, "step": 1027 }, { "epoch": 3.472972972972973, "grad_norm": 0.7827467322349548, "learning_rate": 4.7197410747013376e-05, "loss": 0.308, "step": 1028 }, { "epoch": 3.4763513513513513, "grad_norm": 0.7671857476234436, "learning_rate": 4.6997827602979024e-05, "loss": 0.3552, "step": 1029 }, { "epoch": 3.47972972972973, "grad_norm": 3.2267792224884033, "learning_rate": 4.67918325996437e-05, "loss": 0.3405, "step": 1030 }, { "epoch": 3.483108108108108, "grad_norm": 13.3894624710083, "learning_rate": 4.6579485786623475e-05, "loss": 0.4361, "step": 1031 }, { "epoch": 3.4864864864864864, "grad_norm": 2.525721311569214, "learning_rate": 4.636084906515085e-05, "loss": 0.1359, "step": 1032 }, { "epoch": 3.489864864864865, "grad_norm": 52.08081817626953, "learning_rate": 4.6135986170029947e-05, "loss": 0.5506, "step": 1033 }, { "epoch": 3.4932432432432434, "grad_norm": 1.6132444143295288, "learning_rate": 4.5904962651057134e-05, "loss": 0.1839, "step": 1034 }, { "epoch": 3.4966216216216215, "grad_norm": 1.116671085357666, "learning_rate": 4.566784585391263e-05, "loss": 0.1534, "step": 1035 }, { "epoch": 3.5, "grad_norm": 0.6348862051963806, "learning_rate": 4.542470490052853e-05, "loss": 0.1429, "step": 1036 }, { "epoch": 3.5033783783783785, "grad_norm": 0.6647199392318726, "learning_rate": 4.517561066893909e-05, "loss": 0.1351, "step": 1037 }, { "epoch": 3.506756756756757, "grad_norm": 0.4823889136314392, "learning_rate": 4.492063577261908e-05, "loss": 0.1312, "step": 1038 }, { "epoch": 3.510135135135135, "grad_norm": 0.48596346378326416, "learning_rate": 4.4659854539316174e-05, "loss": 0.1418, "step": 1039 }, { "epoch": 3.5135135135135136, "grad_norm": 0.7251044511795044, "learning_rate": 4.439334298938374e-05, "loss": 0.1465, "step": 1040 }, { "epoch": 3.516891891891892, "grad_norm": 0.4920099973678589, "learning_rate": 4.4121178813620046e-05, "loss": 0.1481, "step": 1041 }, { "epoch": 3.52027027027027, "grad_norm": 0.42203307151794434, "learning_rate": 4.384344135062071e-05, "loss": 0.1219, "step": 1042 }, { "epoch": 3.5236486486486487, "grad_norm": 0.47877076268196106, "learning_rate": 4.3560211563650635e-05, "loss": 0.1282, "step": 1043 }, { "epoch": 3.527027027027027, "grad_norm": 0.4383416175842285, "learning_rate": 4.327157201704241e-05, "loss": 0.133, "step": 1044 }, { "epoch": 3.5304054054054053, "grad_norm": 0.49161508679389954, "learning_rate": 4.297760685212801e-05, "loss": 0.1268, "step": 1045 }, { "epoch": 3.5337837837837838, "grad_norm": 0.44632676243782043, "learning_rate": 4.2678401762710726e-05, "loss": 0.1304, "step": 1046 }, { "epoch": 3.5371621621621623, "grad_norm": 0.48019739985466003, "learning_rate": 4.237404397008455e-05, "loss": 0.134, "step": 1047 }, { "epoch": 3.5405405405405403, "grad_norm": 0.3879307508468628, "learning_rate": 4.206462219760831e-05, "loss": 0.1106, "step": 1048 }, { "epoch": 3.543918918918919, "grad_norm": 8.76919937133789, "learning_rate": 4.175022664484191e-05, "loss": 0.1268, "step": 1049 }, { "epoch": 3.5472972972972974, "grad_norm": 0.5539897680282593, "learning_rate": 4.143094896125227e-05, "loss": 0.134, "step": 1050 }, { "epoch": 3.5506756756756754, "grad_norm": 0.41708147525787354, "learning_rate": 4.1106882219496535e-05, "loss": 0.1185, "step": 1051 }, { "epoch": 3.554054054054054, "grad_norm": 0.7360028028488159, "learning_rate": 4.077812088829051e-05, "loss": 0.1251, "step": 1052 }, { "epoch": 3.5574324324324325, "grad_norm": 0.49543508887290955, "learning_rate": 4.044476080487005e-05, "loss": 0.1405, "step": 1053 }, { "epoch": 3.560810810810811, "grad_norm": 0.40071457624435425, "learning_rate": 4.0106899147053564e-05, "loss": 0.1162, "step": 1054 }, { "epoch": 3.564189189189189, "grad_norm": 0.6165151596069336, "learning_rate": 3.976463440491367e-05, "loss": 0.123, "step": 1055 }, { "epoch": 3.5675675675675675, "grad_norm": 0.4160919487476349, "learning_rate": 3.941806635206639e-05, "loss": 0.1276, "step": 1056 }, { "epoch": 3.570945945945946, "grad_norm": 0.4072069525718689, "learning_rate": 3.9067296016586105e-05, "loss": 0.1336, "step": 1057 }, { "epoch": 3.5743243243243246, "grad_norm": 0.4126531183719635, "learning_rate": 3.871242565155485e-05, "loss": 0.1267, "step": 1058 }, { "epoch": 3.5777027027027026, "grad_norm": 2.380211353302002, "learning_rate": 3.835355870525455e-05, "loss": 0.4067, "step": 1059 }, { "epoch": 3.581081081081081, "grad_norm": 0.3900314271450043, "learning_rate": 3.799079979101075e-05, "loss": 0.1231, "step": 1060 }, { "epoch": 3.5844594594594597, "grad_norm": 0.4255238473415375, "learning_rate": 3.7624254656696864e-05, "loss": 0.1405, "step": 1061 }, { "epoch": 3.5878378378378377, "grad_norm": 0.36614516377449036, "learning_rate": 3.7254030153907494e-05, "loss": 0.118, "step": 1062 }, { "epoch": 3.5912162162162162, "grad_norm": 0.40764889121055603, "learning_rate": 3.688023420681019e-05, "loss": 0.127, "step": 1063 }, { "epoch": 3.5945945945945947, "grad_norm": 0.35974109172821045, "learning_rate": 3.650297578068448e-05, "loss": 0.1238, "step": 1064 }, { "epoch": 3.597972972972973, "grad_norm": 0.4197885990142822, "learning_rate": 3.6122364850157326e-05, "loss": 0.1342, "step": 1065 }, { "epoch": 3.6013513513513513, "grad_norm": 0.3706846237182617, "learning_rate": 3.573851236714447e-05, "loss": 0.1108, "step": 1066 }, { "epoch": 3.60472972972973, "grad_norm": 0.3801131844520569, "learning_rate": 3.535153022850684e-05, "loss": 0.1279, "step": 1067 }, { "epoch": 3.608108108108108, "grad_norm": 0.35572677850723267, "learning_rate": 3.496153124343142e-05, "loss": 0.1263, "step": 1068 }, { "epoch": 3.6114864864864864, "grad_norm": 0.3624805510044098, "learning_rate": 3.4568629100546333e-05, "loss": 0.1185, "step": 1069 }, { "epoch": 3.614864864864865, "grad_norm": 0.36778637766838074, "learning_rate": 3.417293833477938e-05, "loss": 0.1222, "step": 1070 }, { "epoch": 3.618243243243243, "grad_norm": 0.34212857484817505, "learning_rate": 3.377457429397001e-05, "loss": 0.1196, "step": 1071 }, { "epoch": 3.6216216216216215, "grad_norm": 0.3891725540161133, "learning_rate": 3.337365310524423e-05, "loss": 0.1291, "step": 1072 }, { "epoch": 3.625, "grad_norm": 0.399291068315506, "learning_rate": 3.2970291641162396e-05, "loss": 0.1267, "step": 1073 }, { "epoch": 3.6283783783783785, "grad_norm": 0.346253901720047, "learning_rate": 3.25646074856497e-05, "loss": 0.1207, "step": 1074 }, { "epoch": 3.631756756756757, "grad_norm": 0.7178367376327515, "learning_rate": 3.2156718899719216e-05, "loss": 0.372, "step": 1075 }, { "epoch": 3.635135135135135, "grad_norm": 0.41096293926239014, "learning_rate": 3.174674478699772e-05, "loss": 0.1273, "step": 1076 }, { "epoch": 3.6385135135135136, "grad_norm": 0.3635697066783905, "learning_rate": 3.133480465906398e-05, "loss": 0.1256, "step": 1077 }, { "epoch": 3.641891891891892, "grad_norm": 0.3569808006286621, "learning_rate": 3.092101860061e-05, "loss": 0.1334, "step": 1078 }, { "epoch": 3.64527027027027, "grad_norm": 0.339813768863678, "learning_rate": 3.0505507234435122e-05, "loss": 0.111, "step": 1079 }, { "epoch": 3.6486486486486487, "grad_norm": 0.4040168225765228, "learning_rate": 3.0088391686283214e-05, "loss": 0.1348, "step": 1080 }, { "epoch": 3.652027027027027, "grad_norm": 0.37235626578330994, "learning_rate": 2.966979354953336e-05, "loss": 0.1262, "step": 1081 }, { "epoch": 3.6554054054054053, "grad_norm": 0.3623943030834198, "learning_rate": 2.92498348497541e-05, "loss": 0.1185, "step": 1082 }, { "epoch": 3.6587837837837838, "grad_norm": 0.36152735352516174, "learning_rate": 2.882863800913182e-05, "loss": 0.1248, "step": 1083 }, { "epoch": 3.6621621621621623, "grad_norm": 0.38995814323425293, "learning_rate": 2.8406325810783425e-05, "loss": 0.1282, "step": 1084 }, { "epoch": 3.6655405405405403, "grad_norm": 0.35114696621894836, "learning_rate": 2.798302136296379e-05, "loss": 0.1098, "step": 1085 }, { "epoch": 3.668918918918919, "grad_norm": 0.35527390241622925, "learning_rate": 2.7558848063178506e-05, "loss": 0.1129, "step": 1086 }, { "epoch": 3.6722972972972974, "grad_norm": 0.3661573827266693, "learning_rate": 2.713392956221225e-05, "loss": 0.1187, "step": 1087 }, { "epoch": 3.6756756756756754, "grad_norm": 0.3659035265445709, "learning_rate": 2.6708389728083372e-05, "loss": 0.1311, "step": 1088 }, { "epoch": 3.679054054054054, "grad_norm": 0.36093148589134216, "learning_rate": 2.6282352609935028e-05, "loss": 0.1294, "step": 1089 }, { "epoch": 3.6824324324324325, "grad_norm": 0.36720040440559387, "learning_rate": 2.5855942401873734e-05, "loss": 0.128, "step": 1090 }, { "epoch": 3.685810810810811, "grad_norm": 0.40525415539741516, "learning_rate": 2.5429283406765436e-05, "loss": 0.1373, "step": 1091 }, { "epoch": 3.689189189189189, "grad_norm": 0.3309617042541504, "learning_rate": 2.50025e-05, "loss": 0.115, "step": 1092 }, { "epoch": 3.6925675675675675, "grad_norm": 0.3174631893634796, "learning_rate": 2.4575716593234574e-05, "loss": 0.105, "step": 1093 }, { "epoch": 3.695945945945946, "grad_norm": 0.36353635787963867, "learning_rate": 2.414905759812627e-05, "loss": 0.1215, "step": 1094 }, { "epoch": 3.6993243243243246, "grad_norm": 0.3367186188697815, "learning_rate": 2.372264739006497e-05, "loss": 0.1121, "step": 1095 }, { "epoch": 3.7027027027027026, "grad_norm": 0.3455405831336975, "learning_rate": 2.329661027191664e-05, "loss": 0.1142, "step": 1096 }, { "epoch": 3.706081081081081, "grad_norm": 0.3431136906147003, "learning_rate": 2.287107043778775e-05, "loss": 0.1132, "step": 1097 }, { "epoch": 3.7094594594594597, "grad_norm": 0.34689489006996155, "learning_rate": 2.2446151936821504e-05, "loss": 0.1129, "step": 1098 }, { "epoch": 3.7128378378378377, "grad_norm": 0.3749043941497803, "learning_rate": 2.2021978637036214e-05, "loss": 0.1268, "step": 1099 }, { "epoch": 3.7162162162162162, "grad_norm": 0.34978923201560974, "learning_rate": 2.1598674189216575e-05, "loss": 0.1285, "step": 1100 }, { "epoch": 3.7195945945945947, "grad_norm": 0.3356480300426483, "learning_rate": 2.117636199086818e-05, "loss": 0.116, "step": 1101 }, { "epoch": 3.722972972972973, "grad_norm": 0.3172987699508667, "learning_rate": 2.0755165150245906e-05, "loss": 0.1137, "step": 1102 }, { "epoch": 3.7263513513513513, "grad_norm": 0.34791630506515503, "learning_rate": 2.0335206450466658e-05, "loss": 0.1153, "step": 1103 }, { "epoch": 3.72972972972973, "grad_norm": 0.3284004330635071, "learning_rate": 1.9916608313716786e-05, "loss": 0.1131, "step": 1104 }, { "epoch": 3.733108108108108, "grad_norm": 0.3191678822040558, "learning_rate": 1.9499492765564878e-05, "loss": 0.1123, "step": 1105 }, { "epoch": 3.7364864864864864, "grad_norm": 0.33098334074020386, "learning_rate": 1.9083981399389997e-05, "loss": 0.1156, "step": 1106 }, { "epoch": 3.739864864864865, "grad_norm": 0.3604113459587097, "learning_rate": 1.8670195340936026e-05, "loss": 0.1164, "step": 1107 }, { "epoch": 3.743243243243243, "grad_norm": 0.33711400628089905, "learning_rate": 1.825825521300229e-05, "loss": 0.1077, "step": 1108 }, { "epoch": 3.7466216216216215, "grad_norm": 0.3241155743598938, "learning_rate": 1.7848281100280788e-05, "loss": 0.1053, "step": 1109 }, { "epoch": 3.75, "grad_norm": 0.33422669768333435, "learning_rate": 1.7440392514350303e-05, "loss": 0.1226, "step": 1110 }, { "epoch": 3.7533783783783785, "grad_norm": 0.30073583126068115, "learning_rate": 1.7034708358837607e-05, "loss": 0.0983, "step": 1111 }, { "epoch": 3.756756756756757, "grad_norm": 0.37151581048965454, "learning_rate": 1.6631346894755773e-05, "loss": 0.1274, "step": 1112 }, { "epoch": 3.760135135135135, "grad_norm": 0.3346865773200989, "learning_rate": 1.6230425706029996e-05, "loss": 0.1203, "step": 1113 }, { "epoch": 3.7635135135135136, "grad_norm": 0.4401615560054779, "learning_rate": 1.5832061665220622e-05, "loss": 0.1122, "step": 1114 }, { "epoch": 3.766891891891892, "grad_norm": 0.3216763436794281, "learning_rate": 1.5436370899453666e-05, "loss": 0.1139, "step": 1115 }, { "epoch": 3.77027027027027, "grad_norm": 0.34183549880981445, "learning_rate": 1.504346875656858e-05, "loss": 0.1195, "step": 1116 }, { "epoch": 3.7736486486486487, "grad_norm": 0.3108338415622711, "learning_rate": 1.4653469771493166e-05, "loss": 0.1018, "step": 1117 }, { "epoch": 3.777027027027027, "grad_norm": 0.3019558787345886, "learning_rate": 1.426648763285553e-05, "loss": 0.1044, "step": 1118 }, { "epoch": 3.7804054054054053, "grad_norm": 0.30197712779045105, "learning_rate": 1.3882635149842685e-05, "loss": 0.0976, "step": 1119 }, { "epoch": 3.7837837837837838, "grad_norm": 15.628427505493164, "learning_rate": 1.3502024219315524e-05, "loss": 0.1159, "step": 1120 }, { "epoch": 3.7871621621621623, "grad_norm": 0.32293814420700073, "learning_rate": 1.3124765793189808e-05, "loss": 0.108, "step": 1121 }, { "epoch": 3.7905405405405403, "grad_norm": 0.31840014457702637, "learning_rate": 1.2750969846092514e-05, "loss": 0.1011, "step": 1122 }, { "epoch": 3.793918918918919, "grad_norm": 0.33972540497779846, "learning_rate": 1.2380745343303146e-05, "loss": 0.115, "step": 1123 }, { "epoch": 3.7972972972972974, "grad_norm": 0.3430902063846588, "learning_rate": 1.2014200208989255e-05, "loss": 0.1136, "step": 1124 }, { "epoch": 3.8006756756756754, "grad_norm": 0.34559592604637146, "learning_rate": 1.165144129474546e-05, "loss": 0.1172, "step": 1125 }, { "epoch": 3.804054054054054, "grad_norm": 0.33134159445762634, "learning_rate": 1.1292574348445151e-05, "loss": 0.1107, "step": 1126 }, { "epoch": 3.8074324324324325, "grad_norm": 0.2891589403152466, "learning_rate": 1.0937703983413898e-05, "loss": 0.0962, "step": 1127 }, { "epoch": 3.810810810810811, "grad_norm": 0.34017154574394226, "learning_rate": 1.058693364793361e-05, "loss": 0.1168, "step": 1128 }, { "epoch": 3.814189189189189, "grad_norm": 0.3454993963241577, "learning_rate": 1.0240365595086335e-05, "loss": 0.1167, "step": 1129 }, { "epoch": 3.8175675675675675, "grad_norm": 0.33566930890083313, "learning_rate": 9.898100852946441e-06, "loss": 0.1209, "step": 1130 }, { "epoch": 3.820945945945946, "grad_norm": 0.31290584802627563, "learning_rate": 9.560239195129949e-06, "loss": 0.1062, "step": 1131 }, { "epoch": 3.8243243243243246, "grad_norm": 0.33595094084739685, "learning_rate": 9.226879111709494e-06, "loss": 0.1091, "step": 1132 }, { "epoch": 3.8277027027027026, "grad_norm": 0.3021032512187958, "learning_rate": 8.898117780503471e-06, "loss": 0.1029, "step": 1133 }, { "epoch": 3.831081081081081, "grad_norm": 0.3187675476074219, "learning_rate": 8.574051038747738e-06, "loss": 0.1121, "step": 1134 }, { "epoch": 3.8344594594594597, "grad_norm": 0.2892165780067444, "learning_rate": 8.254773355158088e-06, "loss": 0.1007, "step": 1135 }, { "epoch": 3.8378378378378377, "grad_norm": 0.33364662528038025, "learning_rate": 7.94037780239169e-06, "loss": 0.1146, "step": 1136 }, { "epoch": 3.8412162162162162, "grad_norm": 0.32407206296920776, "learning_rate": 7.630956029915455e-06, "loss": 0.1016, "step": 1137 }, { "epoch": 3.8445945945945947, "grad_norm": 0.29210153222084045, "learning_rate": 7.32659823728928e-06, "loss": 0.0989, "step": 1138 }, { "epoch": 3.847972972972973, "grad_norm": 0.3027746081352234, "learning_rate": 7.027393147871994e-06, "loss": 0.1031, "step": 1139 }, { "epoch": 3.8513513513513513, "grad_norm": 0.34020087122917175, "learning_rate": 6.733427982957593e-06, "loss": 0.1105, "step": 1140 }, { "epoch": 3.85472972972973, "grad_norm": 0.3309917747974396, "learning_rate": 6.444788436349374e-06, "loss": 0.1155, "step": 1141 }, { "epoch": 3.858108108108108, "grad_norm": 0.3191674053668976, "learning_rate": 6.16155864937929e-06, "loss": 0.1113, "step": 1142 }, { "epoch": 3.8614864864864864, "grad_norm": 0.3304139971733093, "learning_rate": 5.883821186379955e-06, "loss": 0.1113, "step": 1143 }, { "epoch": 3.864864864864865, "grad_norm": 0.3479062616825104, "learning_rate": 5.6116570106162695e-06, "loss": 0.1173, "step": 1144 }, { "epoch": 3.868243243243243, "grad_norm": 0.3519810140132904, "learning_rate": 5.345145460683825e-06, "loss": 0.1076, "step": 1145 }, { "epoch": 3.8716216216216215, "grad_norm": 0.4121153652667999, "learning_rate": 5.0843642273809276e-06, "loss": 0.1073, "step": 1146 }, { "epoch": 3.875, "grad_norm": 0.3174130618572235, "learning_rate": 4.82938933106091e-06, "loss": 0.1016, "step": 1147 }, { "epoch": 3.8783783783783785, "grad_norm": 0.2952731251716614, "learning_rate": 4.5802950994714764e-06, "loss": 0.1038, "step": 1148 }, { "epoch": 3.881756756756757, "grad_norm": 0.304749071598053, "learning_rate": 4.337154146087377e-06, "loss": 0.1083, "step": 1149 }, { "epoch": 3.885135135135135, "grad_norm": 0.3059677481651306, "learning_rate": 4.100037348942866e-06, "loss": 0.099, "step": 1150 }, { "epoch": 3.8885135135135136, "grad_norm": 0.3029370605945587, "learning_rate": 3.869013829970057e-06, "loss": 0.1017, "step": 1151 }, { "epoch": 3.891891891891892, "grad_norm": 0.28835681080818176, "learning_rate": 3.644150934849158e-06, "loss": 0.1, "step": 1152 }, { "epoch": 3.89527027027027, "grad_norm": 0.2965335249900818, "learning_rate": 3.425514213376533e-06, "loss": 0.1023, "step": 1153 }, { "epoch": 3.8986486486486487, "grad_norm": 0.2934282422065735, "learning_rate": 3.213167400356303e-06, "loss": 0.0937, "step": 1154 }, { "epoch": 3.902027027027027, "grad_norm": 0.32294800877571106, "learning_rate": 3.007172397020972e-06, "loss": 0.1108, "step": 1155 }, { "epoch": 3.9054054054054053, "grad_norm": 0.2884935438632965, "learning_rate": 2.8075892529866257e-06, "loss": 0.1025, "step": 1156 }, { "epoch": 3.9087837837837838, "grad_norm": 0.31093308329582214, "learning_rate": 2.6144761487478416e-06, "loss": 0.107, "step": 1157 }, { "epoch": 3.9121621621621623, "grad_norm": 0.3277430236339569, "learning_rate": 2.4278893787174935e-06, "loss": 0.1216, "step": 1158 }, { "epoch": 3.9155405405405403, "grad_norm": 0.467430979013443, "learning_rate": 2.2478833348163886e-06, "loss": 0.3528, "step": 1159 }, { "epoch": 3.918918918918919, "grad_norm": 0.2848818302154541, "learning_rate": 2.074510490617438e-06, "loss": 0.0981, "step": 1160 }, { "epoch": 3.9222972972972974, "grad_norm": 0.2947328984737396, "learning_rate": 1.9078213860491097e-06, "loss": 0.1003, "step": 1161 }, { "epoch": 3.9256756756756754, "grad_norm": 0.28310471773147583, "learning_rate": 1.7478646126625187e-06, "loss": 0.096, "step": 1162 }, { "epoch": 3.929054054054054, "grad_norm": 0.31671369075775146, "learning_rate": 1.5946867994665007e-06, "loss": 0.1062, "step": 1163 }, { "epoch": 3.9324324324324325, "grad_norm": 0.2964227795600891, "learning_rate": 1.4483325993347872e-06, "loss": 0.102, "step": 1164 }, { "epoch": 3.935810810810811, "grad_norm": 0.309048593044281, "learning_rate": 1.308844675989258e-06, "loss": 0.1049, "step": 1165 }, { "epoch": 3.939189189189189, "grad_norm": 0.31580665707588196, "learning_rate": 1.1762636915630122e-06, "loss": 0.1019, "step": 1166 }, { "epoch": 3.9425675675675675, "grad_norm": 0.27929919958114624, "learning_rate": 1.0506282947469768e-06, "loss": 0.0962, "step": 1167 }, { "epoch": 3.945945945945946, "grad_norm": 0.44457200169563293, "learning_rate": 9.319751095233885e-07, "loss": 0.3502, "step": 1168 }, { "epoch": 3.9493243243243246, "grad_norm": 0.26814886927604675, "learning_rate": 8.203387244895631e-07, "loss": 0.0968, "step": 1169 }, { "epoch": 3.9527027027027026, "grad_norm": 0.31180819869041443, "learning_rate": 7.157516827749585e-07, "loss": 0.1097, "step": 1170 }, { "epoch": 3.956081081081081, "grad_norm": 0.30869176983833313, "learning_rate": 6.182444725545342e-07, "loss": 0.1067, "step": 1171 }, { "epoch": 3.9594594594594597, "grad_norm": 0.29772305488586426, "learning_rate": 5.278455181611357e-07, "loss": 0.1061, "step": 1172 }, { "epoch": 3.9628378378378377, "grad_norm": 0.3115202784538269, "learning_rate": 4.445811717995063e-07, "loss": 0.1081, "step": 1173 }, { "epoch": 3.9662162162162162, "grad_norm": 0.31144705414772034, "learning_rate": 3.684757058643714e-07, "loss": 0.097, "step": 1174 }, { "epoch": 3.9695945945945947, "grad_norm": 0.3056892156600952, "learning_rate": 2.995513058647855e-07, "loss": 0.1069, "step": 1175 }, { "epoch": 3.972972972972973, "grad_norm": 0.327035129070282, "learning_rate": 2.3782806395682474e-07, "loss": 0.1168, "step": 1176 }, { "epoch": 3.9763513513513513, "grad_norm": 0.31039026379585266, "learning_rate": 1.8332397308652485e-07, "loss": 0.105, "step": 1177 }, { "epoch": 3.97972972972973, "grad_norm": 0.2924991548061371, "learning_rate": 1.3605492174477425e-07, "loss": 0.0984, "step": 1178 }, { "epoch": 3.983108108108108, "grad_norm": 0.2952897250652313, "learning_rate": 9.603468933562955e-08, "loss": 0.1039, "step": 1179 }, { "epoch": 3.9864864864864864, "grad_norm": 0.28377214074134827, "learning_rate": 6.32749421594948e-08, "loss": 0.098, "step": 1180 }, { "epoch": 3.989864864864865, "grad_norm": 0.3225543200969696, "learning_rate": 3.778523001229054e-08, "loss": 0.1196, "step": 1181 }, { "epoch": 3.993243243243243, "grad_norm": 0.3456767201423645, "learning_rate": 1.957298340156484e-08, "loss": 0.1071, "step": 1182 }, { "epoch": 3.9966216216216215, "grad_norm": 0.30064505338668823, "learning_rate": 8.64351138044836e-09, "loss": 0.1043, "step": 1183 }, { "epoch": 4.0, "grad_norm": 0.30672967433929443, "learning_rate": 5e-09, "loss": 0.0813, "step": 1184 }, { "epoch": 4.0, "step": 1184, "total_flos": 1.048251868267099e+19, "train_loss": 0.24899632067771982, "train_runtime": 10196.9918, "train_samples_per_second": 2.78, "train_steps_per_second": 0.116 } ], "logging_steps": 1.0, "max_steps": 1184, "num_input_tokens_seen": 0, "num_train_epochs": 4, "save_steps": 250, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.048251868267099e+19, "train_batch_size": 1, "trial_name": null, "trial_params": null }