{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.9790351098762313, "eval_steps": 495, "global_step": 1978, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0, "grad_norm": 53.622260681945534, "learning_rate": 5.000000000000001e-07, "loss": 1.1058, "step": 1 }, { "epoch": 0.0, "eval_loss": 1.1559669971466064, "eval_runtime": 134.2853, "eval_samples_per_second": 9.733, "eval_steps_per_second": 0.276, "step": 1 }, { "epoch": 0.0, "grad_norm": 40.68249059631914, "learning_rate": 1.0000000000000002e-06, "loss": 1.1361, "step": 2 }, { "epoch": 0.0, "grad_norm": 43.27766914416912, "learning_rate": 1.5e-06, "loss": 1.1577, "step": 3 }, { "epoch": 0.0, "grad_norm": 64.72729420725338, "learning_rate": 2.0000000000000003e-06, "loss": 1.1292, "step": 4 }, { "epoch": 0.01, "grad_norm": 56.372636612148945, "learning_rate": 2.5e-06, "loss": 1.1337, "step": 5 }, { "epoch": 0.01, "grad_norm": 50.14861511180959, "learning_rate": 3e-06, "loss": 1.1778, "step": 6 }, { "epoch": 0.01, "grad_norm": 62.95626490942665, "learning_rate": 3.5e-06, "loss": 1.1317, "step": 7 }, { "epoch": 0.01, "grad_norm": 25.825764536536617, "learning_rate": 4.000000000000001e-06, "loss": 1.0504, "step": 8 }, { "epoch": 0.01, "grad_norm": 30.983005109311666, "learning_rate": 4.5e-06, "loss": 1.0636, "step": 9 }, { "epoch": 0.01, "grad_norm": 42.69870581147392, "learning_rate": 5e-06, "loss": 1.1061, "step": 10 }, { "epoch": 0.01, "grad_norm": 30.4102208779435, "learning_rate": 4.9999968146329895e-06, "loss": 1.0577, "step": 11 }, { "epoch": 0.01, "grad_norm": 19.477219524924408, "learning_rate": 4.999987258540075e-06, "loss": 1.0341, "step": 12 }, { "epoch": 0.01, "grad_norm": 16.065115351967986, "learning_rate": 4.9999713317456065e-06, "loss": 1.0212, "step": 13 }, { "epoch": 0.01, "grad_norm": 20.305999270879685, "learning_rate": 4.999949034290173e-06, "loss": 0.9934, "step": 14 }, { "epoch": 0.02, "grad_norm": 15.296134261870147, "learning_rate": 4.999920366230593e-06, "loss": 1.0452, "step": 15 }, { "epoch": 0.02, "grad_norm": 21.688573186878063, "learning_rate": 4.9998853276399215e-06, "loss": 1.0699, "step": 16 }, { "epoch": 0.02, "grad_norm": 14.335847546461085, "learning_rate": 4.999843918607447e-06, "loss": 1.0098, "step": 17 }, { "epoch": 0.02, "grad_norm": 22.3952228510413, "learning_rate": 4.999796139238694e-06, "loss": 1.0408, "step": 18 }, { "epoch": 0.02, "grad_norm": 9.166221867386932, "learning_rate": 4.999741989655415e-06, "loss": 1.0028, "step": 19 }, { "epoch": 0.02, "grad_norm": 11.838379188476093, "learning_rate": 4.999681469995601e-06, "loss": 1.0009, "step": 20 }, { "epoch": 0.02, "grad_norm": 10.419092295341123, "learning_rate": 4.9996145804134735e-06, "loss": 1.022, "step": 21 }, { "epoch": 0.02, "grad_norm": 9.35336523909701, "learning_rate": 4.999541321079486e-06, "loss": 0.9748, "step": 22 }, { "epoch": 0.02, "grad_norm": 9.198682807281516, "learning_rate": 4.999461692180326e-06, "loss": 1.0549, "step": 23 }, { "epoch": 0.02, "grad_norm": 12.547373233238517, "learning_rate": 4.999375693918911e-06, "loss": 0.9437, "step": 24 }, { "epoch": 0.03, "grad_norm": 8.595672331068808, "learning_rate": 4.99928332651439e-06, "loss": 1.0163, "step": 25 }, { "epoch": 0.03, "grad_norm": 12.993405908580561, "learning_rate": 4.999184590202142e-06, "loss": 1.0146, "step": 26 }, { "epoch": 0.03, "grad_norm": 8.549924685378986, "learning_rate": 4.999079485233775e-06, "loss": 0.9542, "step": 27 }, { "epoch": 0.03, "grad_norm": 10.198320823624972, "learning_rate": 4.998968011877129e-06, "loss": 0.9641, "step": 28 }, { "epoch": 0.03, "grad_norm": 8.2545266805555, "learning_rate": 4.99885017041627e-06, "loss": 0.9832, "step": 29 }, { "epoch": 0.03, "grad_norm": 7.9313407875475646, "learning_rate": 4.998725961151493e-06, "loss": 0.9685, "step": 30 }, { "epoch": 0.03, "grad_norm": 9.807588698454, "learning_rate": 4.998595384399319e-06, "loss": 0.9193, "step": 31 }, { "epoch": 0.03, "grad_norm": 9.876327124281431, "learning_rate": 4.998458440492498e-06, "loss": 0.9616, "step": 32 }, { "epoch": 0.03, "grad_norm": 8.98501521873747, "learning_rate": 4.998315129780001e-06, "loss": 1.0099, "step": 33 }, { "epoch": 0.03, "grad_norm": 10.044564125219294, "learning_rate": 4.998165452627025e-06, "loss": 0.9615, "step": 34 }, { "epoch": 0.04, "grad_norm": 11.210345590334603, "learning_rate": 4.9980094094149945e-06, "loss": 1.1204, "step": 35 }, { "epoch": 0.04, "grad_norm": 7.886709354787903, "learning_rate": 4.997847000541551e-06, "loss": 0.9783, "step": 36 }, { "epoch": 0.04, "grad_norm": 7.5254666516991335, "learning_rate": 4.997678226420561e-06, "loss": 0.9161, "step": 37 }, { "epoch": 0.04, "grad_norm": 8.272181825775892, "learning_rate": 4.99750308748211e-06, "loss": 0.9425, "step": 38 }, { "epoch": 0.04, "grad_norm": 6.553520865941246, "learning_rate": 4.997321584172504e-06, "loss": 0.9962, "step": 39 }, { "epoch": 0.04, "grad_norm": 8.229506711694187, "learning_rate": 4.997133716954266e-06, "loss": 0.9347, "step": 40 }, { "epoch": 0.04, "grad_norm": 8.567484614549636, "learning_rate": 4.996939486306138e-06, "loss": 0.9485, "step": 41 }, { "epoch": 0.04, "grad_norm": 6.61521839575945, "learning_rate": 4.996738892723076e-06, "loss": 0.883, "step": 42 }, { "epoch": 0.04, "grad_norm": 8.10212907196996, "learning_rate": 4.99653193671625e-06, "loss": 1.1048, "step": 43 }, { "epoch": 0.04, "grad_norm": 8.80462692647593, "learning_rate": 4.996318618813046e-06, "loss": 0.9816, "step": 44 }, { "epoch": 0.05, "grad_norm": 7.097343998505614, "learning_rate": 4.996098939557062e-06, "loss": 0.9561, "step": 45 }, { "epoch": 0.05, "grad_norm": 7.282799226817679, "learning_rate": 4.995872899508103e-06, "loss": 0.9617, "step": 46 }, { "epoch": 0.05, "grad_norm": 12.216537774733089, "learning_rate": 4.995640499242187e-06, "loss": 0.9243, "step": 47 }, { "epoch": 0.05, "grad_norm": 6.359198206461496, "learning_rate": 4.995401739351537e-06, "loss": 0.9436, "step": 48 }, { "epoch": 0.05, "grad_norm": 6.346073589036576, "learning_rate": 4.995156620444584e-06, "loss": 0.8964, "step": 49 }, { "epoch": 0.05, "grad_norm": 5.688812989274403, "learning_rate": 4.9949051431459615e-06, "loss": 0.9005, "step": 50 }, { "epoch": 0.05, "grad_norm": 7.002996068199546, "learning_rate": 4.994647308096509e-06, "loss": 0.9334, "step": 51 }, { "epoch": 0.05, "grad_norm": 7.083307002817586, "learning_rate": 4.994383115953266e-06, "loss": 0.8988, "step": 52 }, { "epoch": 0.05, "grad_norm": 6.468005918770268, "learning_rate": 4.9941125673894705e-06, "loss": 0.9222, "step": 53 }, { "epoch": 0.05, "grad_norm": 6.444789411001062, "learning_rate": 4.993835663094562e-06, "loss": 0.9338, "step": 54 }, { "epoch": 0.06, "grad_norm": 7.644539402695577, "learning_rate": 4.9935524037741705e-06, "loss": 1.0552, "step": 55 }, { "epoch": 0.06, "grad_norm": 6.40049014930346, "learning_rate": 4.9932627901501265e-06, "loss": 0.9328, "step": 56 }, { "epoch": 0.06, "grad_norm": 8.906097856477972, "learning_rate": 4.99296682296045e-06, "loss": 1.1228, "step": 57 }, { "epoch": 0.06, "grad_norm": 7.079585029259878, "learning_rate": 4.992664502959351e-06, "loss": 1.0413, "step": 58 }, { "epoch": 0.06, "grad_norm": 10.34845460315542, "learning_rate": 4.992355830917232e-06, "loss": 0.9309, "step": 59 }, { "epoch": 0.06, "grad_norm": 6.2525622909780765, "learning_rate": 4.992040807620678e-06, "loss": 0.9742, "step": 60 }, { "epoch": 0.06, "grad_norm": 6.750158577512716, "learning_rate": 4.991719433872461e-06, "loss": 0.9132, "step": 61 }, { "epoch": 0.06, "grad_norm": 6.2784065217091145, "learning_rate": 4.9913917104915374e-06, "loss": 0.9586, "step": 62 }, { "epoch": 0.06, "grad_norm": 6.747179104322235, "learning_rate": 4.991057638313042e-06, "loss": 0.8983, "step": 63 }, { "epoch": 0.06, "grad_norm": 13.39546333182503, "learning_rate": 4.990717218188286e-06, "loss": 0.9282, "step": 64 }, { "epoch": 0.07, "grad_norm": 6.225226219601699, "learning_rate": 4.990370450984764e-06, "loss": 0.8964, "step": 65 }, { "epoch": 0.07, "grad_norm": 7.297944379719202, "learning_rate": 4.990017337586137e-06, "loss": 0.886, "step": 66 }, { "epoch": 0.07, "grad_norm": 6.684537834909945, "learning_rate": 4.989657878892245e-06, "loss": 0.9474, "step": 67 }, { "epoch": 0.07, "grad_norm": 6.800644526106279, "learning_rate": 4.989292075819091e-06, "loss": 0.9222, "step": 68 }, { "epoch": 0.07, "grad_norm": 5.935752803831776, "learning_rate": 4.988919929298851e-06, "loss": 0.9675, "step": 69 }, { "epoch": 0.07, "grad_norm": 6.0962518911658385, "learning_rate": 4.988541440279862e-06, "loss": 0.8945, "step": 70 }, { "epoch": 0.07, "grad_norm": 8.739181681630852, "learning_rate": 4.988156609726628e-06, "loss": 0.9216, "step": 71 }, { "epoch": 0.07, "grad_norm": 6.531531611955777, "learning_rate": 4.987765438619806e-06, "loss": 0.9153, "step": 72 }, { "epoch": 0.07, "grad_norm": 11.36936049546851, "learning_rate": 4.987367927956218e-06, "loss": 1.2286, "step": 73 }, { "epoch": 0.07, "grad_norm": 6.073878979737398, "learning_rate": 4.9869640787488375e-06, "loss": 0.9115, "step": 74 }, { "epoch": 0.08, "grad_norm": 5.8272340999466445, "learning_rate": 4.98655389202679e-06, "loss": 0.9509, "step": 75 }, { "epoch": 0.08, "grad_norm": 6.397354616805974, "learning_rate": 4.986137368835351e-06, "loss": 1.0375, "step": 76 }, { "epoch": 0.08, "grad_norm": 7.966255876371102, "learning_rate": 4.985714510235945e-06, "loss": 1.0293, "step": 77 }, { "epoch": 0.08, "grad_norm": 6.185986548775906, "learning_rate": 4.985285317306141e-06, "loss": 0.9109, "step": 78 }, { "epoch": 0.08, "grad_norm": 7.024522011830165, "learning_rate": 4.984849791139647e-06, "loss": 0.9061, "step": 79 }, { "epoch": 0.08, "grad_norm": 5.995863723081675, "learning_rate": 4.984407932846311e-06, "loss": 0.8925, "step": 80 }, { "epoch": 0.08, "grad_norm": 5.590869203505634, "learning_rate": 4.983959743552118e-06, "loss": 0.9194, "step": 81 }, { "epoch": 0.08, "grad_norm": 6.037924790619887, "learning_rate": 4.983505224399188e-06, "loss": 0.9598, "step": 82 }, { "epoch": 0.08, "grad_norm": 6.517206233677988, "learning_rate": 4.983044376545767e-06, "loss": 0.8645, "step": 83 }, { "epoch": 0.08, "grad_norm": 6.230061445243754, "learning_rate": 4.982577201166232e-06, "loss": 0.9021, "step": 84 }, { "epoch": 0.09, "grad_norm": 6.911990663487462, "learning_rate": 4.9821036994510816e-06, "loss": 0.858, "step": 85 }, { "epoch": 0.09, "grad_norm": 5.984021384400987, "learning_rate": 4.981623872606938e-06, "loss": 0.8723, "step": 86 }, { "epoch": 0.09, "grad_norm": 6.703129385119859, "learning_rate": 4.981137721856541e-06, "loss": 0.9518, "step": 87 }, { "epoch": 0.09, "grad_norm": 7.12479649165451, "learning_rate": 4.980645248438746e-06, "loss": 0.9136, "step": 88 }, { "epoch": 0.09, "grad_norm": 11.122916834992573, "learning_rate": 4.9801464536085184e-06, "loss": 0.9066, "step": 89 }, { "epoch": 0.09, "grad_norm": 7.841215482759505, "learning_rate": 4.9796413386369344e-06, "loss": 1.0641, "step": 90 }, { "epoch": 0.09, "grad_norm": 5.958073787403037, "learning_rate": 4.979129904811177e-06, "loss": 0.9263, "step": 91 }, { "epoch": 0.09, "grad_norm": 7.020080058824307, "learning_rate": 4.978612153434527e-06, "loss": 0.9603, "step": 92 }, { "epoch": 0.09, "grad_norm": 5.574162423590218, "learning_rate": 4.9780880858263684e-06, "loss": 0.8855, "step": 93 }, { "epoch": 0.09, "grad_norm": 5.834495325366524, "learning_rate": 4.977557703322178e-06, "loss": 0.8873, "step": 94 }, { "epoch": 0.1, "grad_norm": 5.461734817525809, "learning_rate": 4.9770210072735286e-06, "loss": 0.8965, "step": 95 }, { "epoch": 0.1, "grad_norm": 6.653047237885576, "learning_rate": 4.976477999048078e-06, "loss": 1.0051, "step": 96 }, { "epoch": 0.1, "grad_norm": 6.092048158281413, "learning_rate": 4.975928680029571e-06, "loss": 0.8779, "step": 97 }, { "epoch": 0.1, "grad_norm": 5.730921274763532, "learning_rate": 4.975373051617831e-06, "loss": 0.8886, "step": 98 }, { "epoch": 0.1, "grad_norm": 7.999386874766411, "learning_rate": 4.974811115228767e-06, "loss": 0.8932, "step": 99 }, { "epoch": 0.1, "grad_norm": 5.569281154720038, "learning_rate": 4.974242872294354e-06, "loss": 0.9093, "step": 100 }, { "epoch": 0.1, "grad_norm": 6.243416936742017, "learning_rate": 4.973668324262645e-06, "loss": 0.8787, "step": 101 }, { "epoch": 0.1, "grad_norm": 6.425532736177959, "learning_rate": 4.973087472597754e-06, "loss": 0.8689, "step": 102 }, { "epoch": 0.1, "grad_norm": 5.834945735385297, "learning_rate": 4.972500318779864e-06, "loss": 1.0439, "step": 103 }, { "epoch": 0.11, "grad_norm": 7.010749117953628, "learning_rate": 4.971906864305214e-06, "loss": 1.0111, "step": 104 }, { "epoch": 0.11, "grad_norm": 5.399008383715632, "learning_rate": 4.9713071106861e-06, "loss": 0.9488, "step": 105 }, { "epoch": 0.11, "grad_norm": 6.852009850723007, "learning_rate": 4.970701059450872e-06, "loss": 0.9165, "step": 106 }, { "epoch": 0.11, "grad_norm": 5.490567287919795, "learning_rate": 4.970088712143924e-06, "loss": 0.93, "step": 107 }, { "epoch": 0.11, "grad_norm": 5.466174338824931, "learning_rate": 4.969470070325699e-06, "loss": 0.8258, "step": 108 }, { "epoch": 0.11, "grad_norm": 5.3052755303637325, "learning_rate": 4.968845135572678e-06, "loss": 0.8643, "step": 109 }, { "epoch": 0.11, "grad_norm": 5.79116786116443, "learning_rate": 4.968213909477376e-06, "loss": 0.9092, "step": 110 }, { "epoch": 0.11, "grad_norm": 6.243372776357591, "learning_rate": 4.967576393648344e-06, "loss": 0.8854, "step": 111 }, { "epoch": 0.11, "grad_norm": 5.554172933410867, "learning_rate": 4.96693258971016e-06, "loss": 0.9373, "step": 112 }, { "epoch": 0.11, "grad_norm": 14.37748982425115, "learning_rate": 4.9662824993034245e-06, "loss": 0.8937, "step": 113 }, { "epoch": 0.12, "grad_norm": 5.41048524226914, "learning_rate": 4.965626124084759e-06, "loss": 0.9385, "step": 114 }, { "epoch": 0.12, "grad_norm": 5.3566507966065355, "learning_rate": 4.9649634657268e-06, "loss": 0.9047, "step": 115 }, { "epoch": 0.12, "grad_norm": 6.532323859056279, "learning_rate": 4.964294525918196e-06, "loss": 0.8841, "step": 116 }, { "epoch": 0.12, "grad_norm": 8.816489663062468, "learning_rate": 4.963619306363602e-06, "loss": 0.8959, "step": 117 }, { "epoch": 0.12, "grad_norm": 7.372965529585357, "learning_rate": 4.962937808783675e-06, "loss": 1.0141, "step": 118 }, { "epoch": 0.12, "grad_norm": 5.68129098809545, "learning_rate": 4.962250034915072e-06, "loss": 0.8776, "step": 119 }, { "epoch": 0.12, "grad_norm": 6.138429945353462, "learning_rate": 4.961555986510443e-06, "loss": 0.9064, "step": 120 }, { "epoch": 0.12, "grad_norm": 6.2074388893831625, "learning_rate": 4.960855665338424e-06, "loss": 0.8984, "step": 121 }, { "epoch": 0.12, "grad_norm": 6.423407890211476, "learning_rate": 4.960149073183643e-06, "loss": 0.8478, "step": 122 }, { "epoch": 0.12, "grad_norm": 15.084543572578946, "learning_rate": 4.959436211846702e-06, "loss": 1.0345, "step": 123 }, { "epoch": 0.13, "grad_norm": 5.093406275304441, "learning_rate": 4.958717083144182e-06, "loss": 0.8912, "step": 124 }, { "epoch": 0.13, "grad_norm": 5.398558925299228, "learning_rate": 4.9579916889086336e-06, "loss": 0.9241, "step": 125 }, { "epoch": 0.13, "grad_norm": 5.573650372167155, "learning_rate": 4.957260030988575e-06, "loss": 0.9476, "step": 126 }, { "epoch": 0.13, "grad_norm": 5.430534040644023, "learning_rate": 4.956522111248483e-06, "loss": 0.9208, "step": 127 }, { "epoch": 0.13, "grad_norm": 5.56067970590197, "learning_rate": 4.955777931568797e-06, "loss": 0.9274, "step": 128 }, { "epoch": 0.13, "grad_norm": 5.601548738488328, "learning_rate": 4.955027493845903e-06, "loss": 0.8426, "step": 129 }, { "epoch": 0.13, "grad_norm": 8.927604688533856, "learning_rate": 4.954270799992138e-06, "loss": 0.8618, "step": 130 }, { "epoch": 0.13, "grad_norm": 5.6362861369729895, "learning_rate": 4.953507851935779e-06, "loss": 0.921, "step": 131 }, { "epoch": 0.13, "grad_norm": 7.966230987922584, "learning_rate": 4.952738651621043e-06, "loss": 0.92, "step": 132 }, { "epoch": 0.13, "grad_norm": 5.8283544064228145, "learning_rate": 4.9519632010080765e-06, "loss": 0.8719, "step": 133 }, { "epoch": 0.14, "grad_norm": 6.733524485804168, "learning_rate": 4.951181502072957e-06, "loss": 0.8664, "step": 134 }, { "epoch": 0.14, "grad_norm": 4.93365687921664, "learning_rate": 4.950393556807682e-06, "loss": 0.9109, "step": 135 }, { "epoch": 0.14, "grad_norm": 5.720266452276143, "learning_rate": 4.9495993672201675e-06, "loss": 1.0141, "step": 136 }, { "epoch": 0.14, "grad_norm": 6.577392904905377, "learning_rate": 4.948798935334242e-06, "loss": 0.8947, "step": 137 }, { "epoch": 0.14, "grad_norm": 5.248018611453421, "learning_rate": 4.947992263189641e-06, "loss": 0.8519, "step": 138 }, { "epoch": 0.14, "grad_norm": 4.964178402688073, "learning_rate": 4.947179352842001e-06, "loss": 0.8732, "step": 139 }, { "epoch": 0.14, "grad_norm": 7.727649852322783, "learning_rate": 4.946360206362858e-06, "loss": 0.9003, "step": 140 }, { "epoch": 0.14, "grad_norm": 5.678031094542374, "learning_rate": 4.945534825839637e-06, "loss": 0.931, "step": 141 }, { "epoch": 0.14, "grad_norm": 4.8975198222340195, "learning_rate": 4.944703213375648e-06, "loss": 0.9022, "step": 142 }, { "epoch": 0.14, "grad_norm": 5.2460458320933485, "learning_rate": 4.943865371090087e-06, "loss": 0.8835, "step": 143 }, { "epoch": 0.15, "grad_norm": 5.080524254102179, "learning_rate": 4.94302130111802e-06, "loss": 0.8957, "step": 144 }, { "epoch": 0.15, "grad_norm": 5.614189755845358, "learning_rate": 4.942171005610385e-06, "loss": 0.8961, "step": 145 }, { "epoch": 0.15, "grad_norm": 5.2283236967224, "learning_rate": 4.941314486733986e-06, "loss": 0.82, "step": 146 }, { "epoch": 0.15, "grad_norm": 4.903937090069375, "learning_rate": 4.940451746671484e-06, "loss": 0.8733, "step": 147 }, { "epoch": 0.15, "grad_norm": 5.044592476065338, "learning_rate": 4.939582787621394e-06, "loss": 0.9423, "step": 148 }, { "epoch": 0.15, "grad_norm": 10.766060880843327, "learning_rate": 4.938707611798078e-06, "loss": 0.8449, "step": 149 }, { "epoch": 0.15, "grad_norm": 4.623203037666776, "learning_rate": 4.937826221431742e-06, "loss": 0.8623, "step": 150 }, { "epoch": 0.15, "grad_norm": 5.0474657841525525, "learning_rate": 4.936938618768427e-06, "loss": 0.9026, "step": 151 }, { "epoch": 0.15, "grad_norm": 31.531685747916384, "learning_rate": 4.936044806070005e-06, "loss": 0.8545, "step": 152 }, { "epoch": 0.15, "grad_norm": 5.2640420937536, "learning_rate": 4.935144785614173e-06, "loss": 0.8793, "step": 153 }, { "epoch": 0.16, "grad_norm": 4.645683842200697, "learning_rate": 4.934238559694448e-06, "loss": 0.8472, "step": 154 }, { "epoch": 0.16, "grad_norm": 6.108168033210612, "learning_rate": 4.93332613062016e-06, "loss": 1.004, "step": 155 }, { "epoch": 0.16, "grad_norm": 5.832130176498325, "learning_rate": 4.932407500716445e-06, "loss": 0.9278, "step": 156 }, { "epoch": 0.16, "grad_norm": 6.2388094461996575, "learning_rate": 4.9314826723242425e-06, "loss": 1.0358, "step": 157 }, { "epoch": 0.16, "grad_norm": 6.954370983448125, "learning_rate": 4.930551647800287e-06, "loss": 0.9919, "step": 158 }, { "epoch": 0.16, "grad_norm": 4.997616854364736, "learning_rate": 4.9296144295171025e-06, "loss": 0.8445, "step": 159 }, { "epoch": 0.16, "grad_norm": 7.115062150515696, "learning_rate": 4.928671019862995e-06, "loss": 0.8546, "step": 160 }, { "epoch": 0.16, "grad_norm": 6.146468237609964, "learning_rate": 4.927721421242051e-06, "loss": 0.8879, "step": 161 }, { "epoch": 0.16, "grad_norm": 5.269100479953049, "learning_rate": 4.926765636074124e-06, "loss": 0.8928, "step": 162 }, { "epoch": 0.16, "grad_norm": 4.829206775414763, "learning_rate": 4.925803666794839e-06, "loss": 0.8912, "step": 163 }, { "epoch": 0.17, "grad_norm": 13.500834231735407, "learning_rate": 4.924835515855573e-06, "loss": 0.89, "step": 164 }, { "epoch": 0.17, "grad_norm": 5.876726093761002, "learning_rate": 4.923861185723461e-06, "loss": 1.0295, "step": 165 }, { "epoch": 0.17, "grad_norm": 6.552028600594945, "learning_rate": 4.92288067888138e-06, "loss": 0.8523, "step": 166 }, { "epoch": 0.17, "grad_norm": 9.604721277839626, "learning_rate": 4.921893997827951e-06, "loss": 0.8793, "step": 167 }, { "epoch": 0.17, "grad_norm": 19.05047261104596, "learning_rate": 4.920901145077527e-06, "loss": 1.0183, "step": 168 }, { "epoch": 0.17, "grad_norm": 8.670208792155977, "learning_rate": 4.919902123160187e-06, "loss": 0.916, "step": 169 }, { "epoch": 0.17, "grad_norm": 5.963579222763698, "learning_rate": 4.918896934621734e-06, "loss": 0.8927, "step": 170 }, { "epoch": 0.17, "grad_norm": 15.698308686475166, "learning_rate": 4.917885582023683e-06, "loss": 0.973, "step": 171 }, { "epoch": 0.17, "grad_norm": 15.158791715235347, "learning_rate": 4.9168680679432565e-06, "loss": 0.8786, "step": 172 }, { "epoch": 0.17, "grad_norm": 8.935267549779024, "learning_rate": 4.915844394973379e-06, "loss": 0.9105, "step": 173 }, { "epoch": 0.18, "grad_norm": 7.85639856911323, "learning_rate": 4.914814565722671e-06, "loss": 0.8709, "step": 174 }, { "epoch": 0.18, "grad_norm": 7.730019511721717, "learning_rate": 4.913778582815439e-06, "loss": 0.9203, "step": 175 }, { "epoch": 0.18, "grad_norm": 30.04084618027315, "learning_rate": 4.912736448891672e-06, "loss": 0.8835, "step": 176 }, { "epoch": 0.18, "grad_norm": 11.53316056998077, "learning_rate": 4.9116881666070325e-06, "loss": 0.8722, "step": 177 }, { "epoch": 0.18, "grad_norm": 36.52217384345542, "learning_rate": 4.9106337386328524e-06, "loss": 0.8615, "step": 178 }, { "epoch": 0.18, "grad_norm": 115.42434006092905, "learning_rate": 4.9095731676561246e-06, "loss": 0.8587, "step": 179 }, { "epoch": 0.18, "grad_norm": 15.54123495731584, "learning_rate": 4.908506456379493e-06, "loss": 0.863, "step": 180 }, { "epoch": 0.18, "grad_norm": 85.99384572854959, "learning_rate": 4.907433607521252e-06, "loss": 0.894, "step": 181 }, { "epoch": 0.18, "grad_norm": 36.13462335675736, "learning_rate": 4.906354623815336e-06, "loss": 0.9055, "step": 182 }, { "epoch": 0.18, "grad_norm": 27.02795167506559, "learning_rate": 4.905269508011312e-06, "loss": 0.9135, "step": 183 }, { "epoch": 0.19, "grad_norm": 78.08453850338199, "learning_rate": 4.904178262874374e-06, "loss": 0.9165, "step": 184 }, { "epoch": 0.19, "grad_norm": 36.25089616666813, "learning_rate": 4.903080891185335e-06, "loss": 0.9778, "step": 185 }, { "epoch": 0.19, "grad_norm": 49.253048863400004, "learning_rate": 4.9019773957406194e-06, "loss": 0.8498, "step": 186 }, { "epoch": 0.19, "grad_norm": 32.09689217815892, "learning_rate": 4.900867779352258e-06, "loss": 0.9056, "step": 187 }, { "epoch": 0.19, "grad_norm": 26.828381083665327, "learning_rate": 4.899752044847881e-06, "loss": 0.8266, "step": 188 }, { "epoch": 0.19, "grad_norm": 144.02426662483788, "learning_rate": 4.898630195070705e-06, "loss": 0.8645, "step": 189 }, { "epoch": 0.19, "grad_norm": 47.52439763925523, "learning_rate": 4.897502232879533e-06, "loss": 0.8619, "step": 190 }, { "epoch": 0.19, "grad_norm": 74.68661681875047, "learning_rate": 4.896368161148744e-06, "loss": 0.8985, "step": 191 }, { "epoch": 0.19, "grad_norm": 75.68493630646316, "learning_rate": 4.895227982768287e-06, "loss": 0.8809, "step": 192 }, { "epoch": 0.19, "grad_norm": 29.706946472395224, "learning_rate": 4.89408170064367e-06, "loss": 0.8396, "step": 193 }, { "epoch": 0.2, "grad_norm": 161.48487196225275, "learning_rate": 4.892929317695957e-06, "loss": 0.8857, "step": 194 }, { "epoch": 0.2, "grad_norm": 41.09638985280378, "learning_rate": 4.891770836861757e-06, "loss": 0.8633, "step": 195 }, { "epoch": 0.2, "grad_norm": 69.69119197887811, "learning_rate": 4.890606261093221e-06, "loss": 1.0396, "step": 196 }, { "epoch": 0.2, "grad_norm": 88.82274883602727, "learning_rate": 4.889435593358029e-06, "loss": 0.8651, "step": 197 }, { "epoch": 0.2, "grad_norm": 81.79928830463821, "learning_rate": 4.888258836639386e-06, "loss": 0.8389, "step": 198 }, { "epoch": 0.2, "grad_norm": 29.87578836447596, "learning_rate": 4.887075993936014e-06, "loss": 0.8775, "step": 199 }, { "epoch": 0.2, "grad_norm": 108.15216101207292, "learning_rate": 4.885887068262143e-06, "loss": 0.8656, "step": 200 }, { "epoch": 0.2, "grad_norm": 37.34405700410443, "learning_rate": 4.884692062647506e-06, "loss": 0.8682, "step": 201 }, { "epoch": 0.2, "grad_norm": 18.005378230428164, "learning_rate": 4.883490980137327e-06, "loss": 0.9104, "step": 202 }, { "epoch": 0.21, "grad_norm": 13.182657069815662, "learning_rate": 4.8822838237923164e-06, "loss": 0.8802, "step": 203 }, { "epoch": 0.21, "grad_norm": 20.126124496397114, "learning_rate": 4.881070596688664e-06, "loss": 0.9229, "step": 204 }, { "epoch": 0.21, "grad_norm": 8.025782615380805, "learning_rate": 4.8798513019180295e-06, "loss": 1.0137, "step": 205 }, { "epoch": 0.21, "grad_norm": 11.303199949810612, "learning_rate": 4.878625942587531e-06, "loss": 0.8844, "step": 206 }, { "epoch": 0.21, "grad_norm": 4.929950753578784, "learning_rate": 4.8773945218197475e-06, "loss": 0.8617, "step": 207 }, { "epoch": 0.21, "grad_norm": 10.422846526317487, "learning_rate": 4.876157042752698e-06, "loss": 0.8628, "step": 208 }, { "epoch": 0.21, "grad_norm": 5.406124330714249, "learning_rate": 4.874913508539844e-06, "loss": 0.8978, "step": 209 }, { "epoch": 0.21, "grad_norm": 10.297329933831254, "learning_rate": 4.8736639223500734e-06, "loss": 0.9213, "step": 210 }, { "epoch": 0.21, "grad_norm": 39.92063204593234, "learning_rate": 4.872408287367702e-06, "loss": 0.9151, "step": 211 }, { "epoch": 0.21, "grad_norm": 12.707518091656361, "learning_rate": 4.871146606792455e-06, "loss": 0.8726, "step": 212 }, { "epoch": 0.22, "grad_norm": 12.451656943060126, "learning_rate": 4.869878883839464e-06, "loss": 0.8831, "step": 213 }, { "epoch": 0.22, "grad_norm": 11.149775907376688, "learning_rate": 4.868605121739261e-06, "loss": 0.7839, "step": 214 }, { "epoch": 0.22, "grad_norm": 6.069136050487129, "learning_rate": 4.867325323737765e-06, "loss": 0.8095, "step": 215 }, { "epoch": 0.22, "grad_norm": 33.341330536587115, "learning_rate": 4.866039493096276e-06, "loss": 0.8652, "step": 216 }, { "epoch": 0.22, "grad_norm": 12.327283349644702, "learning_rate": 4.86474763309147e-06, "loss": 0.9332, "step": 217 }, { "epoch": 0.22, "grad_norm": 14.886718679591452, "learning_rate": 4.863449747015384e-06, "loss": 0.9123, "step": 218 }, { "epoch": 0.22, "grad_norm": 18.439054151579196, "learning_rate": 4.862145838175413e-06, "loss": 0.9952, "step": 219 }, { "epoch": 0.22, "grad_norm": 6.450032781643884, "learning_rate": 4.8608359098943014e-06, "loss": 0.8619, "step": 220 }, { "epoch": 0.22, "grad_norm": 5.76591848048908, "learning_rate": 4.859519965510129e-06, "loss": 0.8904, "step": 221 }, { "epoch": 0.22, "grad_norm": 19.230556702347112, "learning_rate": 4.858198008376308e-06, "loss": 0.8454, "step": 222 }, { "epoch": 0.23, "grad_norm": 12.812798200753768, "learning_rate": 4.856870041861576e-06, "loss": 0.8541, "step": 223 }, { "epoch": 0.23, "grad_norm": 20.396035600104717, "learning_rate": 4.8555360693499786e-06, "loss": 0.8172, "step": 224 }, { "epoch": 0.23, "grad_norm": 28.324370260593682, "learning_rate": 4.854196094240872e-06, "loss": 0.8665, "step": 225 }, { "epoch": 0.23, "grad_norm": 24.090190708297367, "learning_rate": 4.8528501199489045e-06, "loss": 0.8964, "step": 226 }, { "epoch": 0.23, "grad_norm": 51.69442396543579, "learning_rate": 4.8514981499040146e-06, "loss": 1.0263, "step": 227 }, { "epoch": 0.23, "grad_norm": 13.682340026316318, "learning_rate": 4.850140187551417e-06, "loss": 0.8913, "step": 228 }, { "epoch": 0.23, "grad_norm": 10.568651019154684, "learning_rate": 4.848776236351602e-06, "loss": 0.8711, "step": 229 }, { "epoch": 0.23, "grad_norm": 17.95282322327484, "learning_rate": 4.847406299780316e-06, "loss": 0.8534, "step": 230 }, { "epoch": 0.23, "grad_norm": 11.086531755411718, "learning_rate": 4.846030381328559e-06, "loss": 0.8577, "step": 231 }, { "epoch": 0.23, "grad_norm": 20.43851904662153, "learning_rate": 4.8446484845025754e-06, "loss": 0.8711, "step": 232 }, { "epoch": 0.24, "grad_norm": 9.019286512020804, "learning_rate": 4.843260612823844e-06, "loss": 0.8493, "step": 233 }, { "epoch": 0.24, "grad_norm": 28.047694724485186, "learning_rate": 4.84186676982907e-06, "loss": 0.8913, "step": 234 }, { "epoch": 0.24, "grad_norm": 12.170251493093959, "learning_rate": 4.840466959070174e-06, "loss": 0.8561, "step": 235 }, { "epoch": 0.24, "grad_norm": 8.931963450098133, "learning_rate": 4.839061184114286e-06, "loss": 0.831, "step": 236 }, { "epoch": 0.24, "grad_norm": 11.80446928845208, "learning_rate": 4.837649448543731e-06, "loss": 0.8314, "step": 237 }, { "epoch": 0.24, "grad_norm": 6.844616516126078, "learning_rate": 4.836231755956028e-06, "loss": 0.8366, "step": 238 }, { "epoch": 0.24, "grad_norm": 36.62950097854142, "learning_rate": 4.834808109963873e-06, "loss": 0.8674, "step": 239 }, { "epoch": 0.24, "grad_norm": 7.123248446431571, "learning_rate": 4.8333785141951335e-06, "loss": 0.8705, "step": 240 }, { "epoch": 0.24, "grad_norm": 53.32802137920739, "learning_rate": 4.83194297229284e-06, "loss": 0.8973, "step": 241 }, { "epoch": 0.24, "grad_norm": 17.134182129435587, "learning_rate": 4.830501487915175e-06, "loss": 0.8552, "step": 242 }, { "epoch": 0.25, "grad_norm": 18.056985775383417, "learning_rate": 4.829054064735463e-06, "loss": 0.8743, "step": 243 }, { "epoch": 0.25, "grad_norm": 23.339383649662626, "learning_rate": 4.827600706442164e-06, "loss": 0.8401, "step": 244 }, { "epoch": 0.25, "grad_norm": 44.31150837853898, "learning_rate": 4.826141416738861e-06, "loss": 0.8045, "step": 245 }, { "epoch": 0.25, "grad_norm": 46.76596172244923, "learning_rate": 4.824676199344253e-06, "loss": 0.8392, "step": 246 }, { "epoch": 0.25, "grad_norm": 19.466255218481894, "learning_rate": 4.823205057992145e-06, "loss": 0.8422, "step": 247 }, { "epoch": 0.25, "grad_norm": 31.8224396652641, "learning_rate": 4.821727996431435e-06, "loss": 1.0119, "step": 248 }, { "epoch": 0.25, "grad_norm": 28.30370717907097, "learning_rate": 4.8202450184261114e-06, "loss": 0.8858, "step": 249 }, { "epoch": 0.25, "grad_norm": 14.967566855603655, "learning_rate": 4.8187561277552376e-06, "loss": 0.8458, "step": 250 }, { "epoch": 0.25, "grad_norm": 40.45152173099994, "learning_rate": 4.817261328212942e-06, "loss": 0.8414, "step": 251 }, { "epoch": 0.25, "grad_norm": 44.04696789627551, "learning_rate": 4.815760623608415e-06, "loss": 0.8429, "step": 252 }, { "epoch": 0.26, "grad_norm": 22.669028244807055, "learning_rate": 4.8142540177658925e-06, "loss": 0.8377, "step": 253 }, { "epoch": 0.26, "grad_norm": 36.679299334418545, "learning_rate": 4.812741514524647e-06, "loss": 0.8908, "step": 254 }, { "epoch": 0.26, "grad_norm": 19.281368535984196, "learning_rate": 4.811223117738981e-06, "loss": 0.8768, "step": 255 }, { "epoch": 0.26, "grad_norm": 44.69042639681561, "learning_rate": 4.809698831278217e-06, "loss": 0.8919, "step": 256 }, { "epoch": 0.26, "grad_norm": 20.917629234011702, "learning_rate": 4.808168659026683e-06, "loss": 0.8443, "step": 257 }, { "epoch": 0.26, "grad_norm": 20.01311483179014, "learning_rate": 4.806632604883709e-06, "loss": 0.8775, "step": 258 }, { "epoch": 0.26, "grad_norm": 68.90691376267046, "learning_rate": 4.805090672763609e-06, "loss": 0.8913, "step": 259 }, { "epoch": 0.26, "grad_norm": 41.87111801337849, "learning_rate": 4.803542866595681e-06, "loss": 0.823, "step": 260 }, { "epoch": 0.26, "grad_norm": 25.071970545031796, "learning_rate": 4.801989190324187e-06, "loss": 0.8796, "step": 261 }, { "epoch": 0.26, "grad_norm": 109.413625000989, "learning_rate": 4.800429647908354e-06, "loss": 0.858, "step": 262 }, { "epoch": 0.27, "grad_norm": 43.38280987527835, "learning_rate": 4.7988642433223535e-06, "loss": 0.8521, "step": 263 }, { "epoch": 0.27, "grad_norm": 32.13753250435139, "learning_rate": 4.797292980555293e-06, "loss": 0.8757, "step": 264 }, { "epoch": 0.27, "grad_norm": 46.78789327684748, "learning_rate": 4.795715863611212e-06, "loss": 0.895, "step": 265 }, { "epoch": 0.27, "grad_norm": 43.281917729979604, "learning_rate": 4.7941328965090705e-06, "loss": 0.8568, "step": 266 }, { "epoch": 0.27, "grad_norm": 41.384686856921995, "learning_rate": 4.7925440832827305e-06, "loss": 0.9944, "step": 267 }, { "epoch": 0.27, "grad_norm": 40.92527648821274, "learning_rate": 4.790949427980956e-06, "loss": 0.8338, "step": 268 }, { "epoch": 0.27, "grad_norm": 82.6200683793308, "learning_rate": 4.789348934667397e-06, "loss": 0.844, "step": 269 }, { "epoch": 0.27, "grad_norm": 46.48434653058032, "learning_rate": 4.787742607420579e-06, "loss": 0.8408, "step": 270 }, { "epoch": 0.27, "grad_norm": 100.35892484768685, "learning_rate": 4.786130450333897e-06, "loss": 0.8429, "step": 271 }, { "epoch": 0.27, "grad_norm": 104.75380555405746, "learning_rate": 4.784512467515599e-06, "loss": 0.8469, "step": 272 }, { "epoch": 0.28, "grad_norm": 62.43051068810147, "learning_rate": 4.782888663088781e-06, "loss": 0.8564, "step": 273 }, { "epoch": 0.28, "grad_norm": 93.44666911645837, "learning_rate": 4.7812590411913755e-06, "loss": 0.8957, "step": 274 }, { "epoch": 0.28, "grad_norm": 19.862765945899945, "learning_rate": 4.779623605976135e-06, "loss": 1.0236, "step": 275 }, { "epoch": 0.28, "grad_norm": 46.26361057838037, "learning_rate": 4.777982361610629e-06, "loss": 0.9026, "step": 276 }, { "epoch": 0.28, "grad_norm": 52.97493760905037, "learning_rate": 4.77633531227723e-06, "loss": 0.871, "step": 277 }, { "epoch": 0.28, "grad_norm": 34.041391067268314, "learning_rate": 4.774682462173105e-06, "loss": 0.8587, "step": 278 }, { "epoch": 0.28, "grad_norm": 83.06466502493622, "learning_rate": 4.773023815510199e-06, "loss": 0.9954, "step": 279 }, { "epoch": 0.28, "grad_norm": 32.778966883694586, "learning_rate": 4.771359376515231e-06, "loss": 0.8576, "step": 280 }, { "epoch": 0.28, "grad_norm": 58.90680833070677, "learning_rate": 4.769689149429683e-06, "loss": 0.8974, "step": 281 }, { "epoch": 0.28, "grad_norm": 18.71277300441414, "learning_rate": 4.768013138509781e-06, "loss": 0.804, "step": 282 }, { "epoch": 0.29, "grad_norm": 107.32301554273414, "learning_rate": 4.766331348026493e-06, "loss": 0.7948, "step": 283 }, { "epoch": 0.29, "grad_norm": 32.37059166820261, "learning_rate": 4.764643782265517e-06, "loss": 0.7972, "step": 284 }, { "epoch": 0.29, "grad_norm": 60.12326441678363, "learning_rate": 4.762950445527264e-06, "loss": 0.8201, "step": 285 }, { "epoch": 0.29, "grad_norm": 53.85163139056427, "learning_rate": 4.7612513421268546e-06, "loss": 0.8695, "step": 286 }, { "epoch": 0.29, "grad_norm": 62.74196467038289, "learning_rate": 4.759546476394103e-06, "loss": 0.8713, "step": 287 }, { "epoch": 0.29, "grad_norm": 39.83211475405053, "learning_rate": 4.757835852673506e-06, "loss": 0.7924, "step": 288 }, { "epoch": 0.29, "grad_norm": 44.36619825791928, "learning_rate": 4.756119475324237e-06, "loss": 0.9933, "step": 289 }, { "epoch": 0.29, "grad_norm": 37.10810986202249, "learning_rate": 4.754397348720129e-06, "loss": 0.8461, "step": 290 }, { "epoch": 0.29, "grad_norm": 168.41339372943395, "learning_rate": 4.752669477249666e-06, "loss": 1.0045, "step": 291 }, { "epoch": 0.3, "grad_norm": 38.375750184514935, "learning_rate": 4.750935865315972e-06, "loss": 0.8597, "step": 292 }, { "epoch": 0.3, "grad_norm": 139.29153601089564, "learning_rate": 4.749196517336798e-06, "loss": 0.8526, "step": 293 }, { "epoch": 0.3, "grad_norm": 84.96134212632315, "learning_rate": 4.7474514377445155e-06, "loss": 0.8705, "step": 294 }, { "epoch": 0.3, "grad_norm": 79.39624824781504, "learning_rate": 4.745700630986097e-06, "loss": 0.8023, "step": 295 }, { "epoch": 0.3, "grad_norm": 20.962475475808905, "learning_rate": 4.743944101523116e-06, "loss": 0.9936, "step": 296 }, { "epoch": 0.3, "grad_norm": 18.755287063878363, "learning_rate": 4.742181853831721e-06, "loss": 0.822, "step": 297 }, { "epoch": 0.3, "grad_norm": 14.452455674915326, "learning_rate": 4.740413892402639e-06, "loss": 0.8518, "step": 298 }, { "epoch": 0.3, "grad_norm": 123.72945221996332, "learning_rate": 4.7386402217411555e-06, "loss": 0.846, "step": 299 }, { "epoch": 0.3, "grad_norm": 114.85788665830623, "learning_rate": 4.7368608463671015e-06, "loss": 0.8264, "step": 300 }, { "epoch": 0.3, "grad_norm": 47.28480009336326, "learning_rate": 4.73507577081485e-06, "loss": 0.8695, "step": 301 }, { "epoch": 0.31, "grad_norm": 41.276139747783475, "learning_rate": 4.733284999633297e-06, "loss": 0.8486, "step": 302 }, { "epoch": 0.31, "grad_norm": 17.925091199690936, "learning_rate": 4.731488537385853e-06, "loss": 0.8492, "step": 303 }, { "epoch": 0.31, "grad_norm": 61.187483963098465, "learning_rate": 4.7296863886504315e-06, "loss": 0.8728, "step": 304 }, { "epoch": 0.31, "grad_norm": 13.651808271590621, "learning_rate": 4.7278785580194365e-06, "loss": 0.8495, "step": 305 }, { "epoch": 0.31, "grad_norm": 27.644885260501276, "learning_rate": 4.726065050099752e-06, "loss": 0.848, "step": 306 }, { "epoch": 0.31, "grad_norm": 53.55797464707186, "learning_rate": 4.7242458695127275e-06, "loss": 0.894, "step": 307 }, { "epoch": 0.31, "grad_norm": 48.26707421432808, "learning_rate": 4.72242102089417e-06, "loss": 0.8572, "step": 308 }, { "epoch": 0.31, "grad_norm": 35.94106221579397, "learning_rate": 4.720590508894329e-06, "loss": 0.8329, "step": 309 }, { "epoch": 0.31, "grad_norm": 118.69134781259584, "learning_rate": 4.718754338177887e-06, "loss": 0.8335, "step": 310 }, { "epoch": 0.31, "grad_norm": 10.673509155086721, "learning_rate": 4.7169125134239456e-06, "loss": 0.8335, "step": 311 }, { "epoch": 0.32, "grad_norm": 20.703878872323784, "learning_rate": 4.715065039326015e-06, "loss": 0.8741, "step": 312 }, { "epoch": 0.32, "grad_norm": 42.51236987142145, "learning_rate": 4.713211920592003e-06, "loss": 0.8686, "step": 313 }, { "epoch": 0.32, "grad_norm": 31.669223786445617, "learning_rate": 4.711353161944199e-06, "loss": 0.844, "step": 314 }, { "epoch": 0.32, "grad_norm": 55.15593720360342, "learning_rate": 4.709488768119266e-06, "loss": 0.8519, "step": 315 }, { "epoch": 0.32, "grad_norm": 95.39927875998099, "learning_rate": 4.707618743868226e-06, "loss": 0.8131, "step": 316 }, { "epoch": 0.32, "grad_norm": 10.654454638290147, "learning_rate": 4.705743093956452e-06, "loss": 0.9639, "step": 317 }, { "epoch": 0.32, "grad_norm": 38.23375131308745, "learning_rate": 4.703861823163649e-06, "loss": 0.8365, "step": 318 }, { "epoch": 0.32, "grad_norm": 33.691960306044486, "learning_rate": 4.701974936283848e-06, "loss": 0.9664, "step": 319 }, { "epoch": 0.32, "grad_norm": 10.611314812796477, "learning_rate": 4.700082438125391e-06, "loss": 0.8283, "step": 320 }, { "epoch": 0.32, "grad_norm": 20.55619262108289, "learning_rate": 4.6981843335109176e-06, "loss": 0.8272, "step": 321 }, { "epoch": 0.33, "grad_norm": 34.03935919948644, "learning_rate": 4.696280627277356e-06, "loss": 0.8481, "step": 322 }, { "epoch": 0.33, "grad_norm": 20.169566617502596, "learning_rate": 4.69437132427591e-06, "loss": 0.8419, "step": 323 }, { "epoch": 0.33, "grad_norm": 40.583282897440746, "learning_rate": 4.692456429372044e-06, "loss": 0.8535, "step": 324 }, { "epoch": 0.33, "grad_norm": 8.056427978564315, "learning_rate": 4.690535947445471e-06, "loss": 0.851, "step": 325 }, { "epoch": 0.33, "grad_norm": 24.00351867462924, "learning_rate": 4.688609883390144e-06, "loss": 0.8446, "step": 326 }, { "epoch": 0.33, "grad_norm": 16.11523409967761, "learning_rate": 4.686678242114239e-06, "loss": 0.7832, "step": 327 }, { "epoch": 0.33, "grad_norm": 8.214685204880942, "learning_rate": 4.6847410285401465e-06, "loss": 0.839, "step": 328 }, { "epoch": 0.33, "grad_norm": 52.51672691072579, "learning_rate": 4.682798247604453e-06, "loss": 0.8327, "step": 329 }, { "epoch": 0.33, "grad_norm": 8.016142819670371, "learning_rate": 4.680849904257938e-06, "loss": 0.8135, "step": 330 }, { "epoch": 0.33, "grad_norm": 46.655885402782395, "learning_rate": 4.67889600346555e-06, "loss": 0.7983, "step": 331 }, { "epoch": 0.34, "grad_norm": 16.138965226785817, "learning_rate": 4.676936550206402e-06, "loss": 0.8542, "step": 332 }, { "epoch": 0.34, "grad_norm": 55.643105621600114, "learning_rate": 4.674971549473757e-06, "loss": 0.9038, "step": 333 }, { "epoch": 0.34, "grad_norm": 16.31989957270255, "learning_rate": 4.673001006275013e-06, "loss": 0.8399, "step": 334 }, { "epoch": 0.34, "grad_norm": 14.138316979026195, "learning_rate": 4.671024925631694e-06, "loss": 0.7812, "step": 335 }, { "epoch": 0.34, "grad_norm": 50.2685770405374, "learning_rate": 4.669043312579433e-06, "loss": 0.8358, "step": 336 }, { "epoch": 0.34, "grad_norm": 9.047670202154327, "learning_rate": 4.667056172167962e-06, "loss": 0.8361, "step": 337 }, { "epoch": 0.34, "grad_norm": 15.329346838170585, "learning_rate": 4.665063509461098e-06, "loss": 0.8112, "step": 338 }, { "epoch": 0.34, "grad_norm": 34.581957147308756, "learning_rate": 4.6630653295367286e-06, "loss": 0.8819, "step": 339 }, { "epoch": 0.34, "grad_norm": 70.00837894867692, "learning_rate": 4.6610616374868066e-06, "loss": 0.8316, "step": 340 }, { "epoch": 0.34, "grad_norm": 19.54707036059632, "learning_rate": 4.659052438417326e-06, "loss": 0.8171, "step": 341 }, { "epoch": 0.35, "grad_norm": 122.51019371942445, "learning_rate": 4.6570377374483155e-06, "loss": 0.8908, "step": 342 }, { "epoch": 0.35, "grad_norm": 24.88314319301423, "learning_rate": 4.655017539713826e-06, "loss": 0.8623, "step": 343 }, { "epoch": 0.35, "grad_norm": 17.959493645537147, "learning_rate": 4.652991850361912e-06, "loss": 0.8354, "step": 344 }, { "epoch": 0.35, "grad_norm": 13.1801020702163, "learning_rate": 4.650960674554627e-06, "loss": 0.8247, "step": 345 }, { "epoch": 0.35, "grad_norm": 14.455355688503463, "learning_rate": 4.648924017468003e-06, "loss": 0.8286, "step": 346 }, { "epoch": 0.35, "grad_norm": 8.731820345524651, "learning_rate": 4.64688188429204e-06, "loss": 0.8327, "step": 347 }, { "epoch": 0.35, "grad_norm": 42.87517105426398, "learning_rate": 4.644834280230693e-06, "loss": 0.8393, "step": 348 }, { "epoch": 0.35, "grad_norm": 50.868804589197836, "learning_rate": 4.642781210501858e-06, "loss": 0.8598, "step": 349 }, { "epoch": 0.35, "grad_norm": 5.919995360552276, "learning_rate": 4.640722680337358e-06, "loss": 0.962, "step": 350 }, { "epoch": 0.35, "grad_norm": 26.64669124982461, "learning_rate": 4.638658694982936e-06, "loss": 0.798, "step": 351 }, { "epoch": 0.36, "grad_norm": 26.1192454778817, "learning_rate": 4.6365892596982295e-06, "loss": 0.8762, "step": 352 }, { "epoch": 0.36, "grad_norm": 18.9062598091514, "learning_rate": 4.634514379756769e-06, "loss": 0.7937, "step": 353 }, { "epoch": 0.36, "grad_norm": 11.482713649160976, "learning_rate": 4.632434060445956e-06, "loss": 0.8616, "step": 354 }, { "epoch": 0.36, "grad_norm": 8.370599147559595, "learning_rate": 4.6303483070670574e-06, "loss": 0.7956, "step": 355 }, { "epoch": 0.36, "grad_norm": 12.925585913640592, "learning_rate": 4.628257124935183e-06, "loss": 0.8246, "step": 356 }, { "epoch": 0.36, "grad_norm": 13.662180432298918, "learning_rate": 4.626160519379279e-06, "loss": 0.8376, "step": 357 }, { "epoch": 0.36, "grad_norm": 30.308232889166852, "learning_rate": 4.624058495742115e-06, "loss": 0.8591, "step": 358 }, { "epoch": 0.36, "grad_norm": 64.26811094645683, "learning_rate": 4.621951059380259e-06, "loss": 0.8079, "step": 359 }, { "epoch": 0.36, "grad_norm": 34.37337832128252, "learning_rate": 4.619838215664082e-06, "loss": 0.7918, "step": 360 }, { "epoch": 0.36, "grad_norm": 15.079822107556895, "learning_rate": 4.617719969977729e-06, "loss": 0.8811, "step": 361 }, { "epoch": 0.37, "grad_norm": 11.388476367649812, "learning_rate": 4.615596327719111e-06, "loss": 0.8445, "step": 362 }, { "epoch": 0.37, "grad_norm": 8.881147580394376, "learning_rate": 4.613467294299893e-06, "loss": 0.8489, "step": 363 }, { "epoch": 0.37, "grad_norm": 7.3102827631277885, "learning_rate": 4.611332875145476e-06, "loss": 0.8141, "step": 364 }, { "epoch": 0.37, "grad_norm": 19.341931669078043, "learning_rate": 4.609193075694989e-06, "loss": 0.8301, "step": 365 }, { "epoch": 0.37, "grad_norm": 16.293526833129757, "learning_rate": 4.607047901401267e-06, "loss": 0.8472, "step": 366 }, { "epoch": 0.37, "grad_norm": 12.701861210256677, "learning_rate": 4.604897357730846e-06, "loss": 0.8258, "step": 367 }, { "epoch": 0.37, "grad_norm": 12.022062250305899, "learning_rate": 4.60274145016394e-06, "loss": 0.7865, "step": 368 }, { "epoch": 0.37, "grad_norm": 8.942347186149718, "learning_rate": 4.600580184194436e-06, "loss": 0.8538, "step": 369 }, { "epoch": 0.37, "grad_norm": 7.084710601938015, "learning_rate": 4.598413565329876e-06, "loss": 0.8213, "step": 370 }, { "epoch": 0.37, "grad_norm": 8.773416021429492, "learning_rate": 4.596241599091438e-06, "loss": 0.8281, "step": 371 }, { "epoch": 0.38, "grad_norm": 11.412681654994278, "learning_rate": 4.59406429101393e-06, "loss": 0.8073, "step": 372 }, { "epoch": 0.38, "grad_norm": 33.38456658772077, "learning_rate": 4.591881646645775e-06, "loss": 0.8078, "step": 373 }, { "epoch": 0.38, "grad_norm": 20.116008377655895, "learning_rate": 4.589693671548989e-06, "loss": 0.8791, "step": 374 }, { "epoch": 0.38, "grad_norm": 25.441358730169238, "learning_rate": 4.587500371299176e-06, "loss": 0.8096, "step": 375 }, { "epoch": 0.38, "grad_norm": 10.632165430252845, "learning_rate": 4.585301751485508e-06, "loss": 0.8124, "step": 376 }, { "epoch": 0.38, "grad_norm": 24.69121638703892, "learning_rate": 4.583097817710716e-06, "loss": 0.8292, "step": 377 }, { "epoch": 0.38, "grad_norm": 42.373835243000144, "learning_rate": 4.580888575591068e-06, "loss": 0.8124, "step": 378 }, { "epoch": 0.38, "grad_norm": 15.688601624623248, "learning_rate": 4.578674030756364e-06, "loss": 0.834, "step": 379 }, { "epoch": 0.38, "grad_norm": 12.872910424438674, "learning_rate": 4.5764541888499116e-06, "loss": 0.8293, "step": 380 }, { "epoch": 0.38, "grad_norm": 11.085141619803203, "learning_rate": 4.574229055528522e-06, "loss": 0.8419, "step": 381 }, { "epoch": 0.39, "grad_norm": 20.128058817237306, "learning_rate": 4.571998636462487e-06, "loss": 0.8072, "step": 382 }, { "epoch": 0.39, "grad_norm": 45.8693129268715, "learning_rate": 4.5697629373355694e-06, "loss": 0.8516, "step": 383 }, { "epoch": 0.39, "grad_norm": 12.540832336612032, "learning_rate": 4.567521963844987e-06, "loss": 0.7809, "step": 384 }, { "epoch": 0.39, "grad_norm": 7.272618009907343, "learning_rate": 4.5652757217014e-06, "loss": 0.8836, "step": 385 }, { "epoch": 0.39, "grad_norm": 25.542789288767313, "learning_rate": 4.56302421662889e-06, "loss": 0.8266, "step": 386 }, { "epoch": 0.39, "grad_norm": 21.60322316347993, "learning_rate": 4.560767454364955e-06, "loss": 0.8591, "step": 387 }, { "epoch": 0.39, "grad_norm": 23.667006186382807, "learning_rate": 4.5585054406604865e-06, "loss": 0.8391, "step": 388 }, { "epoch": 0.39, "grad_norm": 9.485503231699525, "learning_rate": 4.556238181279761e-06, "loss": 0.8561, "step": 389 }, { "epoch": 0.39, "grad_norm": 24.57297216692732, "learning_rate": 4.55396568200042e-06, "loss": 0.8277, "step": 390 }, { "epoch": 0.4, "grad_norm": 13.729895392458525, "learning_rate": 4.551687948613459e-06, "loss": 0.8032, "step": 391 }, { "epoch": 0.4, "grad_norm": 18.865208120754964, "learning_rate": 4.549404986923213e-06, "loss": 0.863, "step": 392 }, { "epoch": 0.4, "grad_norm": 22.190589070773914, "learning_rate": 4.547116802747335e-06, "loss": 0.8989, "step": 393 }, { "epoch": 0.4, "grad_norm": 21.450617982266408, "learning_rate": 4.544823401916794e-06, "loss": 0.792, "step": 394 }, { "epoch": 0.4, "grad_norm": 35.65422229133744, "learning_rate": 4.542524790275848e-06, "loss": 0.845, "step": 395 }, { "epoch": 0.4, "grad_norm": 11.364569307665759, "learning_rate": 4.5402209736820325e-06, "loss": 0.829, "step": 396 }, { "epoch": 0.4, "grad_norm": 21.94205192387575, "learning_rate": 4.537911958006149e-06, "loss": 0.8234, "step": 397 }, { "epoch": 0.4, "grad_norm": 10.63181070102972, "learning_rate": 4.5355977491322485e-06, "loss": 0.9657, "step": 398 }, { "epoch": 0.4, "grad_norm": 98.56328297192321, "learning_rate": 4.5332783529576145e-06, "loss": 0.851, "step": 399 }, { "epoch": 0.4, "grad_norm": 15.486561420545545, "learning_rate": 4.530953775392749e-06, "loss": 1.0024, "step": 400 }, { "epoch": 0.41, "grad_norm": 5.36867796146416, "learning_rate": 4.528624022361359e-06, "loss": 0.8203, "step": 401 }, { "epoch": 0.41, "grad_norm": 23.16373737065163, "learning_rate": 4.5262890998003374e-06, "loss": 0.8405, "step": 402 }, { "epoch": 0.41, "grad_norm": 24.557753611805698, "learning_rate": 4.523949013659754e-06, "loss": 0.8594, "step": 403 }, { "epoch": 0.41, "grad_norm": 239.577821661523, "learning_rate": 4.521603769902835e-06, "loss": 0.8801, "step": 404 }, { "epoch": 0.41, "grad_norm": 17.722681590802623, "learning_rate": 4.5192533745059494e-06, "loss": 0.7956, "step": 405 }, { "epoch": 0.41, "grad_norm": 21.064904857136227, "learning_rate": 4.5168978334585955e-06, "loss": 0.849, "step": 406 }, { "epoch": 0.41, "grad_norm": 29.204434577367632, "learning_rate": 4.514537152763384e-06, "loss": 0.8081, "step": 407 }, { "epoch": 0.41, "grad_norm": 31.677494628738625, "learning_rate": 4.512171338436022e-06, "loss": 0.8123, "step": 408 }, { "epoch": 0.41, "grad_norm": 28.259106443263146, "learning_rate": 4.509800396505298e-06, "loss": 0.7939, "step": 409 }, { "epoch": 0.41, "grad_norm": 31.436330986692848, "learning_rate": 4.507424333013069e-06, "loss": 0.9276, "step": 410 }, { "epoch": 0.42, "grad_norm": 62.15027720090623, "learning_rate": 4.505043154014243e-06, "loss": 0.8111, "step": 411 }, { "epoch": 0.42, "grad_norm": 30.339486412023945, "learning_rate": 4.502656865576762e-06, "loss": 0.8102, "step": 412 }, { "epoch": 0.42, "grad_norm": 29.14289188449159, "learning_rate": 4.500265473781591e-06, "loss": 0.8526, "step": 413 }, { "epoch": 0.42, "grad_norm": 16.33958767689576, "learning_rate": 4.497868984722698e-06, "loss": 0.7931, "step": 414 }, { "epoch": 0.42, "grad_norm": 6.946767762788769, "learning_rate": 4.495467404507039e-06, "loss": 0.7723, "step": 415 }, { "epoch": 0.42, "grad_norm": 16.808686330416283, "learning_rate": 4.493060739254548e-06, "loss": 0.7941, "step": 416 }, { "epoch": 0.42, "grad_norm": 11.193622425241314, "learning_rate": 4.4906489950981125e-06, "loss": 0.8121, "step": 417 }, { "epoch": 0.42, "grad_norm": 7.172258662371558, "learning_rate": 4.4882321781835666e-06, "loss": 0.8399, "step": 418 }, { "epoch": 0.42, "grad_norm": 7.172757454398357, "learning_rate": 4.485810294669668e-06, "loss": 0.7951, "step": 419 }, { "epoch": 0.42, "grad_norm": 12.912002056626209, "learning_rate": 4.4833833507280884e-06, "loss": 0.8724, "step": 420 }, { "epoch": 0.43, "grad_norm": 9.5299105957465, "learning_rate": 4.4809513525433925e-06, "loss": 0.8688, "step": 421 }, { "epoch": 0.43, "grad_norm": 5.942401217715412, "learning_rate": 4.478514306313026e-06, "loss": 0.8349, "step": 422 }, { "epoch": 0.43, "grad_norm": 14.70118213793697, "learning_rate": 4.476072218247297e-06, "loss": 0.8126, "step": 423 }, { "epoch": 0.43, "grad_norm": 35.64062421380103, "learning_rate": 4.473625094569366e-06, "loss": 0.8502, "step": 424 }, { "epoch": 0.43, "grad_norm": 9.515237433838145, "learning_rate": 4.471172941515219e-06, "loss": 0.8318, "step": 425 }, { "epoch": 0.43, "grad_norm": 4.52525000568301, "learning_rate": 4.4687157653336645e-06, "loss": 0.836, "step": 426 }, { "epoch": 0.43, "grad_norm": 14.93156121415628, "learning_rate": 4.466253572286308e-06, "loss": 0.8386, "step": 427 }, { "epoch": 0.43, "grad_norm": 5.403603957950335, "learning_rate": 4.463786368647541e-06, "loss": 0.8026, "step": 428 }, { "epoch": 0.43, "grad_norm": 5.101795351603165, "learning_rate": 4.461314160704521e-06, "loss": 0.8424, "step": 429 }, { "epoch": 0.43, "grad_norm": 4.657146697628601, "learning_rate": 4.458836954757161e-06, "loss": 0.8853, "step": 430 }, { "epoch": 0.44, "grad_norm": 4.8266562887019315, "learning_rate": 4.456354757118109e-06, "loss": 0.8311, "step": 431 }, { "epoch": 0.44, "grad_norm": 4.734850079452724, "learning_rate": 4.453867574112733e-06, "loss": 0.8143, "step": 432 }, { "epoch": 0.44, "grad_norm": 8.241211166354187, "learning_rate": 4.4513754120791065e-06, "loss": 0.8309, "step": 433 }, { "epoch": 0.44, "grad_norm": 4.984336730679723, "learning_rate": 4.448878277367988e-06, "loss": 0.834, "step": 434 }, { "epoch": 0.44, "grad_norm": 4.435146207156021, "learning_rate": 4.446376176342812e-06, "loss": 0.8344, "step": 435 }, { "epoch": 0.44, "grad_norm": 5.354979176494093, "learning_rate": 4.443869115379667e-06, "loss": 0.8822, "step": 436 }, { "epoch": 0.44, "grad_norm": 4.740333651131116, "learning_rate": 4.441357100867278e-06, "loss": 0.7742, "step": 437 }, { "epoch": 0.44, "grad_norm": 6.7966105580368605, "learning_rate": 4.438840139206998e-06, "loss": 0.8741, "step": 438 }, { "epoch": 0.44, "grad_norm": 5.13633031473771, "learning_rate": 4.436318236812782e-06, "loss": 0.8654, "step": 439 }, { "epoch": 0.44, "grad_norm": 5.791216706993845, "learning_rate": 4.433791400111179e-06, "loss": 0.8262, "step": 440 }, { "epoch": 0.45, "grad_norm": 6.234562179308168, "learning_rate": 4.431259635541312e-06, "loss": 0.8392, "step": 441 }, { "epoch": 0.45, "grad_norm": 5.150150040963309, "learning_rate": 4.428722949554858e-06, "loss": 0.8191, "step": 442 }, { "epoch": 0.45, "grad_norm": 4.563262237056235, "learning_rate": 4.426181348616038e-06, "loss": 0.7961, "step": 443 }, { "epoch": 0.45, "grad_norm": 4.409500438314574, "learning_rate": 4.423634839201601e-06, "loss": 0.8554, "step": 444 }, { "epoch": 0.45, "grad_norm": 5.103854158767785, "learning_rate": 4.421083427800795e-06, "loss": 0.8073, "step": 445 }, { "epoch": 0.45, "grad_norm": 4.956272764605345, "learning_rate": 4.41852712091537e-06, "loss": 0.822, "step": 446 }, { "epoch": 0.45, "grad_norm": 47.36121108696217, "learning_rate": 4.415965925059544e-06, "loss": 0.8117, "step": 447 }, { "epoch": 0.45, "grad_norm": 7.6663655272761515, "learning_rate": 4.413399846759998e-06, "loss": 0.8469, "step": 448 }, { "epoch": 0.45, "grad_norm": 7.344767060016519, "learning_rate": 4.4108288925558505e-06, "loss": 0.8716, "step": 449 }, { "epoch": 0.45, "grad_norm": 4.520414583397527, "learning_rate": 4.40825306899865e-06, "loss": 0.8154, "step": 450 }, { "epoch": 0.46, "grad_norm": 5.337724239493111, "learning_rate": 4.405672382652349e-06, "loss": 0.8204, "step": 451 }, { "epoch": 0.46, "grad_norm": 6.931095236367972, "learning_rate": 4.403086840093297e-06, "loss": 0.7721, "step": 452 }, { "epoch": 0.46, "grad_norm": 4.283750585511646, "learning_rate": 4.400496447910212e-06, "loss": 0.8099, "step": 453 }, { "epoch": 0.46, "grad_norm": 4.885379997216326, "learning_rate": 4.397901212704176e-06, "loss": 0.7897, "step": 454 }, { "epoch": 0.46, "grad_norm": 6.083991051490237, "learning_rate": 4.3953011410886105e-06, "loss": 0.8497, "step": 455 }, { "epoch": 0.46, "grad_norm": 4.86609034632746, "learning_rate": 4.392696239689261e-06, "loss": 0.8271, "step": 456 }, { "epoch": 0.46, "grad_norm": 7.4379794775118615, "learning_rate": 4.390086515144179e-06, "loss": 0.8416, "step": 457 }, { "epoch": 0.46, "grad_norm": 10.915187189394999, "learning_rate": 4.387471974103713e-06, "loss": 0.7873, "step": 458 }, { "epoch": 0.46, "grad_norm": 5.839190871758675, "learning_rate": 4.384852623230478e-06, "loss": 0.7898, "step": 459 }, { "epoch": 0.46, "grad_norm": 5.670517697152507, "learning_rate": 4.38222846919935e-06, "loss": 0.8431, "step": 460 }, { "epoch": 0.47, "grad_norm": 6.4010524847241115, "learning_rate": 4.379599518697444e-06, "loss": 0.8489, "step": 461 }, { "epoch": 0.47, "grad_norm": 5.505346933106831, "learning_rate": 4.3769657784240975e-06, "loss": 0.8056, "step": 462 }, { "epoch": 0.47, "grad_norm": 4.519181163960888, "learning_rate": 4.3743272550908545e-06, "loss": 0.82, "step": 463 }, { "epoch": 0.47, "grad_norm": 5.2720001800167475, "learning_rate": 4.3716839554214475e-06, "loss": 0.8261, "step": 464 }, { "epoch": 0.47, "grad_norm": 5.48180004445423, "learning_rate": 4.369035886151778e-06, "loss": 0.7666, "step": 465 }, { "epoch": 0.47, "grad_norm": 10.355727305381658, "learning_rate": 4.366383054029907e-06, "loss": 0.8203, "step": 466 }, { "epoch": 0.47, "grad_norm": 5.185243645713917, "learning_rate": 4.363725465816028e-06, "loss": 0.8336, "step": 467 }, { "epoch": 0.47, "grad_norm": 4.537447056649908, "learning_rate": 4.3610631282824556e-06, "loss": 0.8141, "step": 468 }, { "epoch": 0.47, "grad_norm": 10.059335937073103, "learning_rate": 4.358396048213609e-06, "loss": 0.9323, "step": 469 }, { "epoch": 0.47, "grad_norm": 7.558752810891884, "learning_rate": 4.355724232405989e-06, "loss": 0.8052, "step": 470 }, { "epoch": 0.48, "grad_norm": 8.901083371299169, "learning_rate": 4.35304768766817e-06, "loss": 0.8681, "step": 471 }, { "epoch": 0.48, "grad_norm": 5.749432223564655, "learning_rate": 4.350366420820771e-06, "loss": 0.7636, "step": 472 }, { "epoch": 0.48, "grad_norm": 5.467983174988701, "learning_rate": 4.3476804386964486e-06, "loss": 0.798, "step": 473 }, { "epoch": 0.48, "grad_norm": 14.340080127005345, "learning_rate": 4.3449897481398735e-06, "loss": 0.8325, "step": 474 }, { "epoch": 0.48, "grad_norm": 10.900246873425738, "learning_rate": 4.342294356007715e-06, "loss": 0.8191, "step": 475 }, { "epoch": 0.48, "grad_norm": 27.212678856649585, "learning_rate": 4.339594269168624e-06, "loss": 0.7983, "step": 476 }, { "epoch": 0.48, "grad_norm": 12.291890124401506, "learning_rate": 4.336889494503215e-06, "loss": 0.8223, "step": 477 }, { "epoch": 0.48, "grad_norm": 64.09126736889749, "learning_rate": 4.3341800389040465e-06, "loss": 0.7935, "step": 478 }, { "epoch": 0.48, "grad_norm": 5.384829430193068, "learning_rate": 4.331465909275608e-06, "loss": 0.8063, "step": 479 }, { "epoch": 0.48, "grad_norm": 5.980479175502184, "learning_rate": 4.3287471125342994e-06, "loss": 0.8264, "step": 480 }, { "epoch": 0.49, "grad_norm": 14.891763292361196, "learning_rate": 4.326023655608412e-06, "loss": 0.8724, "step": 481 }, { "epoch": 0.49, "grad_norm": 47.12998834101147, "learning_rate": 4.3232955454381126e-06, "loss": 0.9869, "step": 482 }, { "epoch": 0.49, "grad_norm": 4.471610169915462, "learning_rate": 4.320562788975429e-06, "loss": 0.8504, "step": 483 }, { "epoch": 0.49, "grad_norm": 5.687055238382911, "learning_rate": 4.317825393184226e-06, "loss": 0.8192, "step": 484 }, { "epoch": 0.49, "grad_norm": 6.0149060191653625, "learning_rate": 4.315083365040193e-06, "loss": 0.8474, "step": 485 }, { "epoch": 0.49, "grad_norm": 5.497573287322376, "learning_rate": 4.31233671153082e-06, "loss": 0.7821, "step": 486 }, { "epoch": 0.49, "grad_norm": 5.877574667166528, "learning_rate": 4.30958543965539e-06, "loss": 0.7886, "step": 487 }, { "epoch": 0.49, "grad_norm": 8.602706507929737, "learning_rate": 4.306829556424949e-06, "loss": 0.922, "step": 488 }, { "epoch": 0.49, "grad_norm": 5.056672401512086, "learning_rate": 4.3040690688622965e-06, "loss": 0.824, "step": 489 }, { "epoch": 0.5, "grad_norm": 19.382968104583764, "learning_rate": 4.3013039840019675e-06, "loss": 0.8193, "step": 490 }, { "epoch": 0.5, "grad_norm": 4.880915789014604, "learning_rate": 4.2985343088902096e-06, "loss": 0.8291, "step": 491 }, { "epoch": 0.5, "grad_norm": 12.940372419412574, "learning_rate": 4.295760050584966e-06, "loss": 0.8503, "step": 492 }, { "epoch": 0.5, "grad_norm": 5.731004069006167, "learning_rate": 4.292981216155864e-06, "loss": 0.7975, "step": 493 }, { "epoch": 0.5, "grad_norm": 14.301842614329003, "learning_rate": 4.2901978126841885e-06, "loss": 0.8179, "step": 494 }, { "epoch": 0.5, "grad_norm": 4.304262507046851, "learning_rate": 4.287409847262868e-06, "loss": 0.8559, "step": 495 }, { "epoch": 0.5, "eval_loss": 0.8057979345321655, "eval_runtime": 132.6219, "eval_samples_per_second": 9.855, "eval_steps_per_second": 0.279, "step": 495 }, { "epoch": 0.5, "grad_norm": 12.900039338569156, "learning_rate": 4.284617326996458e-06, "loss": 0.8245, "step": 496 }, { "epoch": 0.5, "grad_norm": 27.720291530372055, "learning_rate": 4.281820259001121e-06, "loss": 0.8123, "step": 497 }, { "epoch": 0.5, "grad_norm": 9.875124909540267, "learning_rate": 4.2790186504046045e-06, "loss": 0.8077, "step": 498 }, { "epoch": 0.5, "grad_norm": 4.959102865654325, "learning_rate": 4.276212508346232e-06, "loss": 0.8102, "step": 499 }, { "epoch": 0.51, "grad_norm": 8.445468566132282, "learning_rate": 4.273401839976877e-06, "loss": 0.9495, "step": 500 }, { "epoch": 0.51, "grad_norm": 7.168758336315809, "learning_rate": 4.2705866524589475e-06, "loss": 0.794, "step": 501 }, { "epoch": 0.51, "grad_norm": 23.960316842270476, "learning_rate": 4.267766952966369e-06, "loss": 0.8234, "step": 502 }, { "epoch": 0.51, "grad_norm": 18.37836287213558, "learning_rate": 4.264942748684563e-06, "loss": 0.8037, "step": 503 }, { "epoch": 0.51, "grad_norm": 5.337874181371052, "learning_rate": 4.262114046810429e-06, "loss": 0.8341, "step": 504 }, { "epoch": 0.51, "grad_norm": 23.756712822011433, "learning_rate": 4.2592808545523335e-06, "loss": 0.7835, "step": 505 }, { "epoch": 0.51, "grad_norm": 6.5799458933423365, "learning_rate": 4.256443179130081e-06, "loss": 0.823, "step": 506 }, { "epoch": 0.51, "grad_norm": 5.340830389502352, "learning_rate": 4.2536010277748995e-06, "loss": 0.7958, "step": 507 }, { "epoch": 0.51, "grad_norm": 4.274138909771709, "learning_rate": 4.250754407729428e-06, "loss": 0.806, "step": 508 }, { "epoch": 0.51, "grad_norm": 5.0306349490940985, "learning_rate": 4.2479033262476885e-06, "loss": 0.8728, "step": 509 }, { "epoch": 0.52, "grad_norm": 4.644338434000799, "learning_rate": 4.2450477905950745e-06, "loss": 0.8304, "step": 510 }, { "epoch": 0.52, "grad_norm": 4.447934903535319, "learning_rate": 4.242187808048329e-06, "loss": 0.8294, "step": 511 }, { "epoch": 0.52, "grad_norm": 4.468749772112203, "learning_rate": 4.239323385895527e-06, "loss": 0.8134, "step": 512 }, { "epoch": 0.52, "grad_norm": 4.595150327241665, "learning_rate": 4.236454531436058e-06, "loss": 0.9786, "step": 513 }, { "epoch": 0.52, "grad_norm": 3.7601508175241807, "learning_rate": 4.233581251980604e-06, "loss": 0.8269, "step": 514 }, { "epoch": 0.52, "grad_norm": 4.86366757480535, "learning_rate": 4.230703554851127e-06, "loss": 0.8258, "step": 515 }, { "epoch": 0.52, "grad_norm": 4.352222075706524, "learning_rate": 4.227821447380842e-06, "loss": 0.9652, "step": 516 }, { "epoch": 0.52, "grad_norm": 8.73857112989018, "learning_rate": 4.224934936914206e-06, "loss": 0.8383, "step": 517 }, { "epoch": 0.52, "grad_norm": 4.044749221503628, "learning_rate": 4.222044030806894e-06, "loss": 0.8207, "step": 518 }, { "epoch": 0.52, "grad_norm": 4.613463894710261, "learning_rate": 4.219148736425785e-06, "loss": 0.8116, "step": 519 }, { "epoch": 0.53, "grad_norm": 5.8777142171863686, "learning_rate": 4.216249061148939e-06, "loss": 0.7927, "step": 520 }, { "epoch": 0.53, "grad_norm": 4.297186700651947, "learning_rate": 4.2133450123655805e-06, "loss": 0.8784, "step": 521 }, { "epoch": 0.53, "grad_norm": 4.9580457958209765, "learning_rate": 4.210436597476077e-06, "loss": 0.8232, "step": 522 }, { "epoch": 0.53, "grad_norm": 4.156257841735334, "learning_rate": 4.207523823891924e-06, "loss": 0.8446, "step": 523 }, { "epoch": 0.53, "grad_norm": 4.364319132512279, "learning_rate": 4.204606699035724e-06, "loss": 0.8001, "step": 524 }, { "epoch": 0.53, "grad_norm": 4.626437632049346, "learning_rate": 4.201685230341168e-06, "loss": 0.8488, "step": 525 }, { "epoch": 0.53, "grad_norm": 3.6979503756764536, "learning_rate": 4.198759425253015e-06, "loss": 0.8231, "step": 526 }, { "epoch": 0.53, "grad_norm": 3.9375406125469117, "learning_rate": 4.195829291227076e-06, "loss": 0.8262, "step": 527 }, { "epoch": 0.53, "grad_norm": 3.757592718600369, "learning_rate": 4.192894835730193e-06, "loss": 0.8325, "step": 528 }, { "epoch": 0.53, "grad_norm": 5.166098354237156, "learning_rate": 4.1899560662402204e-06, "loss": 0.8207, "step": 529 }, { "epoch": 0.54, "grad_norm": 3.5626403449402146, "learning_rate": 4.187012990246005e-06, "loss": 0.848, "step": 530 }, { "epoch": 0.54, "grad_norm": 4.405523794955747, "learning_rate": 4.18406561524737e-06, "loss": 0.8153, "step": 531 }, { "epoch": 0.54, "grad_norm": 4.461492151159491, "learning_rate": 4.18111394875509e-06, "loss": 0.7961, "step": 532 }, { "epoch": 0.54, "grad_norm": 3.9579080031218945, "learning_rate": 4.178157998290879e-06, "loss": 0.7961, "step": 533 }, { "epoch": 0.54, "grad_norm": 3.4157753139788114, "learning_rate": 4.175197771387368e-06, "loss": 0.77, "step": 534 }, { "epoch": 0.54, "grad_norm": 3.6556329052007372, "learning_rate": 4.172233275588082e-06, "loss": 0.757, "step": 535 }, { "epoch": 0.54, "grad_norm": 3.56739734349662, "learning_rate": 4.169264518447428e-06, "loss": 0.816, "step": 536 }, { "epoch": 0.54, "grad_norm": 3.929236437983629, "learning_rate": 4.16629150753067e-06, "loss": 0.7774, "step": 537 }, { "epoch": 0.54, "grad_norm": 4.162206048648253, "learning_rate": 4.163314250413913e-06, "loss": 0.8242, "step": 538 }, { "epoch": 0.54, "grad_norm": 3.5348802562769683, "learning_rate": 4.160332754684085e-06, "loss": 0.8292, "step": 539 }, { "epoch": 0.55, "grad_norm": 4.026476793505602, "learning_rate": 4.157347027938907e-06, "loss": 0.8438, "step": 540 }, { "epoch": 0.55, "grad_norm": 3.6419095683671205, "learning_rate": 4.154357077786892e-06, "loss": 0.8358, "step": 541 }, { "epoch": 0.55, "grad_norm": 4.410825648674678, "learning_rate": 4.1513629118473095e-06, "loss": 0.7966, "step": 542 }, { "epoch": 0.55, "grad_norm": 3.7192835271993423, "learning_rate": 4.1483645377501726e-06, "loss": 0.7922, "step": 543 }, { "epoch": 0.55, "grad_norm": 4.2093727504165575, "learning_rate": 4.145361963136219e-06, "loss": 0.8179, "step": 544 }, { "epoch": 0.55, "grad_norm": 3.81543786003894, "learning_rate": 4.1423551956568916e-06, "loss": 0.8072, "step": 545 }, { "epoch": 0.55, "grad_norm": 4.0092000330441335, "learning_rate": 4.139344242974317e-06, "loss": 0.811, "step": 546 }, { "epoch": 0.55, "grad_norm": 3.927721495131119, "learning_rate": 4.136329112761285e-06, "loss": 0.8587, "step": 547 }, { "epoch": 0.55, "grad_norm": 3.6480183964959076, "learning_rate": 4.133309812701233e-06, "loss": 0.7872, "step": 548 }, { "epoch": 0.55, "grad_norm": 3.603038562895384, "learning_rate": 4.130286350488224e-06, "loss": 0.8211, "step": 549 }, { "epoch": 0.56, "grad_norm": 4.502055880585956, "learning_rate": 4.127258733826929e-06, "loss": 0.7947, "step": 550 }, { "epoch": 0.56, "grad_norm": 4.0596180713400845, "learning_rate": 4.124226970432602e-06, "loss": 0.8371, "step": 551 }, { "epoch": 0.56, "grad_norm": 5.32641556527492, "learning_rate": 4.121191068031067e-06, "loss": 0.9996, "step": 552 }, { "epoch": 0.56, "grad_norm": 3.657535832118382, "learning_rate": 4.118151034358696e-06, "loss": 0.7846, "step": 553 }, { "epoch": 0.56, "grad_norm": 4.542970770312805, "learning_rate": 4.1151068771623864e-06, "loss": 0.9695, "step": 554 }, { "epoch": 0.56, "grad_norm": 3.4165335352607418, "learning_rate": 4.112058604199544e-06, "loss": 0.7936, "step": 555 }, { "epoch": 0.56, "grad_norm": 4.38297960645218, "learning_rate": 4.109006223238064e-06, "loss": 0.7862, "step": 556 }, { "epoch": 0.56, "grad_norm": 4.597014556198219, "learning_rate": 4.105949742056309e-06, "loss": 0.9645, "step": 557 }, { "epoch": 0.56, "grad_norm": 3.431709372470043, "learning_rate": 4.102889168443091e-06, "loss": 0.8077, "step": 558 }, { "epoch": 0.56, "grad_norm": 3.787025575201097, "learning_rate": 4.0998245101976495e-06, "loss": 0.811, "step": 559 }, { "epoch": 0.57, "grad_norm": 4.1176444935534535, "learning_rate": 4.096755775129634e-06, "loss": 0.824, "step": 560 }, { "epoch": 0.57, "grad_norm": 3.874994826180069, "learning_rate": 4.093682971059081e-06, "loss": 0.8207, "step": 561 }, { "epoch": 0.57, "grad_norm": 5.615505880427777, "learning_rate": 4.0906061058164e-06, "loss": 0.9337, "step": 562 }, { "epoch": 0.57, "grad_norm": 4.159261391680892, "learning_rate": 4.087525187242345e-06, "loss": 0.7938, "step": 563 }, { "epoch": 0.57, "grad_norm": 3.982328427595764, "learning_rate": 4.084440223188002e-06, "loss": 0.8353, "step": 564 }, { "epoch": 0.57, "grad_norm": 3.501905560670443, "learning_rate": 4.081351221514765e-06, "loss": 0.8038, "step": 565 }, { "epoch": 0.57, "grad_norm": 4.4798436725132795, "learning_rate": 4.078258190094318e-06, "loss": 0.8106, "step": 566 }, { "epoch": 0.57, "grad_norm": 4.630604033125049, "learning_rate": 4.0751611368086115e-06, "loss": 0.8368, "step": 567 }, { "epoch": 0.57, "grad_norm": 4.247722886052707, "learning_rate": 4.072060069549848e-06, "loss": 0.7804, "step": 568 }, { "epoch": 0.57, "grad_norm": 3.976719242252589, "learning_rate": 4.068954996220457e-06, "loss": 0.8517, "step": 569 }, { "epoch": 0.58, "grad_norm": 3.7378777926374647, "learning_rate": 4.065845924733077e-06, "loss": 0.7879, "step": 570 }, { "epoch": 0.58, "grad_norm": 3.377152336223136, "learning_rate": 4.062732863010534e-06, "loss": 0.8151, "step": 571 }, { "epoch": 0.58, "grad_norm": 3.958932705104774, "learning_rate": 4.059615818985826e-06, "loss": 0.8261, "step": 572 }, { "epoch": 0.58, "grad_norm": 4.056870191904715, "learning_rate": 4.056494800602093e-06, "loss": 0.8036, "step": 573 }, { "epoch": 0.58, "grad_norm": 4.276330614300562, "learning_rate": 4.053369815812608e-06, "loss": 0.7653, "step": 574 }, { "epoch": 0.58, "grad_norm": 3.7171393121481557, "learning_rate": 4.050240872580749e-06, "loss": 0.811, "step": 575 }, { "epoch": 0.58, "grad_norm": 3.8515114977006384, "learning_rate": 4.047107978879985e-06, "loss": 0.837, "step": 576 }, { "epoch": 0.58, "grad_norm": 3.990125985702027, "learning_rate": 4.043971142693845e-06, "loss": 0.7478, "step": 577 }, { "epoch": 0.58, "grad_norm": 5.038946309180022, "learning_rate": 4.040830372015909e-06, "loss": 0.8186, "step": 578 }, { "epoch": 0.58, "grad_norm": 5.347765467237379, "learning_rate": 4.0376856748497865e-06, "loss": 0.8224, "step": 579 }, { "epoch": 0.59, "grad_norm": 3.8056078948982957, "learning_rate": 4.034537059209085e-06, "loss": 0.7922, "step": 580 }, { "epoch": 0.59, "grad_norm": 4.699749057539843, "learning_rate": 4.031384533117404e-06, "loss": 0.8004, "step": 581 }, { "epoch": 0.59, "grad_norm": 3.985653883322575, "learning_rate": 4.0282281046083045e-06, "loss": 0.8352, "step": 582 }, { "epoch": 0.59, "grad_norm": 3.6031461215401115, "learning_rate": 4.025067781725294e-06, "loss": 0.8557, "step": 583 }, { "epoch": 0.59, "grad_norm": 4.374834805628173, "learning_rate": 4.021903572521802e-06, "loss": 0.8473, "step": 584 }, { "epoch": 0.59, "grad_norm": 4.706489580366399, "learning_rate": 4.0187354850611634e-06, "loss": 0.8707, "step": 585 }, { "epoch": 0.59, "grad_norm": 3.6196942519877844, "learning_rate": 4.015563527416596e-06, "loss": 0.7828, "step": 586 }, { "epoch": 0.59, "grad_norm": 4.021419041448392, "learning_rate": 4.012387707671177e-06, "loss": 0.8212, "step": 587 }, { "epoch": 0.59, "grad_norm": 7.5051269502991484, "learning_rate": 4.00920803391783e-06, "loss": 0.8125, "step": 588 }, { "epoch": 0.6, "grad_norm": 4.816524818864674, "learning_rate": 4.006024514259295e-06, "loss": 0.8458, "step": 589 }, { "epoch": 0.6, "grad_norm": 3.999047814247483, "learning_rate": 4.002837156808116e-06, "loss": 0.9794, "step": 590 }, { "epoch": 0.6, "grad_norm": 4.053935034209712, "learning_rate": 3.999645969686616e-06, "loss": 0.8172, "step": 591 }, { "epoch": 0.6, "grad_norm": 3.4988389719653408, "learning_rate": 3.996450961026876e-06, "loss": 0.8358, "step": 592 }, { "epoch": 0.6, "grad_norm": 3.7828610862275447, "learning_rate": 3.993252138970716e-06, "loss": 0.8472, "step": 593 }, { "epoch": 0.6, "grad_norm": 4.30179599391661, "learning_rate": 3.990049511669675e-06, "loss": 0.9382, "step": 594 }, { "epoch": 0.6, "grad_norm": 3.7610184281848373, "learning_rate": 3.986843087284986e-06, "loss": 0.8122, "step": 595 }, { "epoch": 0.6, "grad_norm": 3.7656350305584976, "learning_rate": 3.983632873987562e-06, "loss": 0.7727, "step": 596 }, { "epoch": 0.6, "grad_norm": 3.760295447041278, "learning_rate": 3.980418879957967e-06, "loss": 0.7896, "step": 597 }, { "epoch": 0.6, "grad_norm": 3.1049226948000386, "learning_rate": 3.977201113386402e-06, "loss": 0.7927, "step": 598 }, { "epoch": 0.61, "grad_norm": 11.952856680061187, "learning_rate": 3.973979582472681e-06, "loss": 0.82, "step": 599 }, { "epoch": 0.61, "grad_norm": 3.6775385257137136, "learning_rate": 3.970754295426211e-06, "loss": 0.7982, "step": 600 }, { "epoch": 0.61, "grad_norm": 4.126563331457935, "learning_rate": 3.96752526046597e-06, "loss": 0.7973, "step": 601 }, { "epoch": 0.61, "grad_norm": 3.9969733729191637, "learning_rate": 3.964292485820487e-06, "loss": 0.7848, "step": 602 }, { "epoch": 0.61, "grad_norm": 3.843494951843216, "learning_rate": 3.961055979727822e-06, "loss": 0.7809, "step": 603 }, { "epoch": 0.61, "grad_norm": 3.722737193377863, "learning_rate": 3.957815750435542e-06, "loss": 0.8272, "step": 604 }, { "epoch": 0.61, "grad_norm": 3.505878864813977, "learning_rate": 3.954571806200702e-06, "loss": 0.7547, "step": 605 }, { "epoch": 0.61, "grad_norm": 3.816576897584943, "learning_rate": 3.951324155289825e-06, "loss": 0.8178, "step": 606 }, { "epoch": 0.61, "grad_norm": 3.6494681155339252, "learning_rate": 3.948072805978879e-06, "loss": 0.8163, "step": 607 }, { "epoch": 0.61, "grad_norm": 3.31954721595174, "learning_rate": 3.9448177665532575e-06, "loss": 0.8036, "step": 608 }, { "epoch": 0.62, "grad_norm": 3.99601790046458, "learning_rate": 3.941559045307756e-06, "loss": 0.8052, "step": 609 }, { "epoch": 0.62, "grad_norm": 3.256204345344998, "learning_rate": 3.938296650546552e-06, "loss": 0.8062, "step": 610 }, { "epoch": 0.62, "grad_norm": 3.6522897880772778, "learning_rate": 3.935030590583186e-06, "loss": 0.8461, "step": 611 }, { "epoch": 0.62, "grad_norm": 3.5724478570414706, "learning_rate": 3.931760873740539e-06, "loss": 0.8173, "step": 612 }, { "epoch": 0.62, "grad_norm": 4.369010546999307, "learning_rate": 3.928487508350808e-06, "loss": 0.8355, "step": 613 }, { "epoch": 0.62, "grad_norm": 3.867196950400295, "learning_rate": 3.9252105027554885e-06, "loss": 0.7982, "step": 614 }, { "epoch": 0.62, "grad_norm": 3.757894555280206, "learning_rate": 3.9219298653053544e-06, "loss": 0.744, "step": 615 }, { "epoch": 0.62, "grad_norm": 4.483773308176931, "learning_rate": 3.918645604360434e-06, "loss": 0.8264, "step": 616 }, { "epoch": 0.62, "grad_norm": 3.5936542631360995, "learning_rate": 3.915357728289985e-06, "loss": 0.773, "step": 617 }, { "epoch": 0.62, "grad_norm": 3.7125514529042447, "learning_rate": 3.912066245472484e-06, "loss": 0.789, "step": 618 }, { "epoch": 0.63, "grad_norm": 3.994678394733242, "learning_rate": 3.908771164295595e-06, "loss": 0.8054, "step": 619 }, { "epoch": 0.63, "grad_norm": 3.440863065927586, "learning_rate": 3.9054724931561516e-06, "loss": 0.7667, "step": 620 }, { "epoch": 0.63, "grad_norm": 3.6354877018750003, "learning_rate": 3.902170240460137e-06, "loss": 0.8434, "step": 621 }, { "epoch": 0.63, "grad_norm": 3.6697310301169783, "learning_rate": 3.898864414622661e-06, "loss": 0.8277, "step": 622 }, { "epoch": 0.63, "grad_norm": 3.4657019048153543, "learning_rate": 3.895555024067937e-06, "loss": 0.8058, "step": 623 }, { "epoch": 0.63, "grad_norm": 3.569249201300816, "learning_rate": 3.8922420772292645e-06, "loss": 0.8646, "step": 624 }, { "epoch": 0.63, "grad_norm": 3.6960654260718675, "learning_rate": 3.888925582549006e-06, "loss": 0.8439, "step": 625 }, { "epoch": 0.63, "grad_norm": 3.206703534075406, "learning_rate": 3.8856055484785625e-06, "loss": 0.958, "step": 626 }, { "epoch": 0.63, "grad_norm": 3.8024335105854963, "learning_rate": 3.8822819834783556e-06, "loss": 0.7933, "step": 627 }, { "epoch": 0.63, "grad_norm": 3.311892134941326, "learning_rate": 3.878954896017804e-06, "loss": 0.7679, "step": 628 }, { "epoch": 0.64, "grad_norm": 3.6892320743049383, "learning_rate": 3.875624294575306e-06, "loss": 0.8663, "step": 629 }, { "epoch": 0.64, "grad_norm": 5.747838560381203, "learning_rate": 3.872290187638208e-06, "loss": 0.8028, "step": 630 }, { "epoch": 0.64, "grad_norm": 4.130227898119578, "learning_rate": 3.868952583702798e-06, "loss": 0.7775, "step": 631 }, { "epoch": 0.64, "grad_norm": 3.8622011098862004, "learning_rate": 3.865611491274267e-06, "loss": 0.8362, "step": 632 }, { "epoch": 0.64, "grad_norm": 4.076018210248722, "learning_rate": 3.862266918866702e-06, "loss": 0.9772, "step": 633 }, { "epoch": 0.64, "grad_norm": 4.361556277412654, "learning_rate": 3.858918875003053e-06, "loss": 0.7957, "step": 634 }, { "epoch": 0.64, "grad_norm": 5.386167012643809, "learning_rate": 3.855567368215122e-06, "loss": 0.9516, "step": 635 }, { "epoch": 0.64, "grad_norm": 3.9439193106997736, "learning_rate": 3.852212407043528e-06, "loss": 0.8231, "step": 636 }, { "epoch": 0.64, "grad_norm": 4.465518470978496, "learning_rate": 3.848854000037702e-06, "loss": 0.8476, "step": 637 }, { "epoch": 0.64, "grad_norm": 4.675424607415688, "learning_rate": 3.845492155755848e-06, "loss": 0.7827, "step": 638 }, { "epoch": 0.65, "grad_norm": 6.031861648386004, "learning_rate": 3.8421268827649325e-06, "loss": 0.8044, "step": 639 }, { "epoch": 0.65, "grad_norm": 5.884738538466192, "learning_rate": 3.83875818964066e-06, "loss": 0.9754, "step": 640 }, { "epoch": 0.65, "grad_norm": 6.050577830000915, "learning_rate": 3.835386084967451e-06, "loss": 0.7774, "step": 641 }, { "epoch": 0.65, "grad_norm": 4.992604354329953, "learning_rate": 3.832010577338414e-06, "loss": 0.8663, "step": 642 }, { "epoch": 0.65, "grad_norm": 4.137703965401395, "learning_rate": 3.828631675355338e-06, "loss": 0.8247, "step": 643 }, { "epoch": 0.65, "grad_norm": 3.9250818325319647, "learning_rate": 3.825249387628654e-06, "loss": 0.9427, "step": 644 }, { "epoch": 0.65, "grad_norm": 5.876925468778531, "learning_rate": 3.821863722777427e-06, "loss": 0.7972, "step": 645 }, { "epoch": 0.65, "grad_norm": 4.029506273549229, "learning_rate": 3.818474689429324e-06, "loss": 0.8085, "step": 646 }, { "epoch": 0.65, "grad_norm": 4.095701024225992, "learning_rate": 3.815082296220596e-06, "loss": 0.8278, "step": 647 }, { "epoch": 0.65, "grad_norm": 4.0688198508439255, "learning_rate": 3.8116865517960584e-06, "loss": 0.7869, "step": 648 }, { "epoch": 0.66, "grad_norm": 4.679593702074773, "learning_rate": 3.808287464809063e-06, "loss": 0.8103, "step": 649 }, { "epoch": 0.66, "grad_norm": 3.980055385102619, "learning_rate": 3.8048850439214844e-06, "loss": 0.9489, "step": 650 }, { "epoch": 0.66, "grad_norm": 3.89441546745002, "learning_rate": 3.8014792978036874e-06, "loss": 0.8053, "step": 651 }, { "epoch": 0.66, "grad_norm": 3.495335889379114, "learning_rate": 3.7980702351345146e-06, "loss": 0.796, "step": 652 }, { "epoch": 0.66, "grad_norm": 3.69936931190018, "learning_rate": 3.7946578646012578e-06, "loss": 0.8388, "step": 653 }, { "epoch": 0.66, "grad_norm": 3.711786581597773, "learning_rate": 3.7912421948996394e-06, "loss": 0.772, "step": 654 }, { "epoch": 0.66, "grad_norm": 4.3115598706174225, "learning_rate": 3.787823234733788e-06, "loss": 0.8086, "step": 655 }, { "epoch": 0.66, "grad_norm": 4.206524036573733, "learning_rate": 3.7844009928162195e-06, "loss": 0.8422, "step": 656 }, { "epoch": 0.66, "grad_norm": 4.581806945598464, "learning_rate": 3.78097547786781e-06, "loss": 0.8299, "step": 657 }, { "epoch": 0.66, "grad_norm": 3.649063963125477, "learning_rate": 3.7775466986177763e-06, "loss": 0.8233, "step": 658 }, { "epoch": 0.67, "grad_norm": 3.9874069159693497, "learning_rate": 3.7741146638036567e-06, "loss": 0.836, "step": 659 }, { "epoch": 0.67, "grad_norm": 5.101352434056481, "learning_rate": 3.770679382171283e-06, "loss": 0.8053, "step": 660 }, { "epoch": 0.67, "grad_norm": 4.057518103549027, "learning_rate": 3.7672408624747598e-06, "loss": 0.8119, "step": 661 }, { "epoch": 0.67, "grad_norm": 4.134292768219472, "learning_rate": 3.7637991134764475e-06, "loss": 0.8441, "step": 662 }, { "epoch": 0.67, "grad_norm": 4.037293558864936, "learning_rate": 3.7603541439469314e-06, "loss": 0.8475, "step": 663 }, { "epoch": 0.67, "grad_norm": 4.067939888641445, "learning_rate": 3.756905962665005e-06, "loss": 0.8176, "step": 664 }, { "epoch": 0.67, "grad_norm": 4.741871018197297, "learning_rate": 3.7534545784176486e-06, "loss": 0.8252, "step": 665 }, { "epoch": 0.67, "grad_norm": 3.7687469109452536, "learning_rate": 3.7500000000000005e-06, "loss": 0.8222, "step": 666 }, { "epoch": 0.67, "grad_norm": 4.0489323193480855, "learning_rate": 3.7465422362153416e-06, "loss": 0.8325, "step": 667 }, { "epoch": 0.67, "grad_norm": 4.049191319739393, "learning_rate": 3.7430812958750695e-06, "loss": 0.8391, "step": 668 }, { "epoch": 0.68, "grad_norm": 3.601920010559379, "learning_rate": 3.7396171877986766e-06, "loss": 0.7752, "step": 669 }, { "epoch": 0.68, "grad_norm": 4.353923235687504, "learning_rate": 3.736149920813726e-06, "loss": 0.7507, "step": 670 }, { "epoch": 0.68, "grad_norm": 4.022804558800863, "learning_rate": 3.7326795037558335e-06, "loss": 0.8216, "step": 671 }, { "epoch": 0.68, "grad_norm": 4.056771582273963, "learning_rate": 3.72920594546864e-06, "loss": 0.7952, "step": 672 }, { "epoch": 0.68, "grad_norm": 3.418690092445234, "learning_rate": 3.7257292548037917e-06, "loss": 0.7834, "step": 673 }, { "epoch": 0.68, "grad_norm": 4.027419729266051, "learning_rate": 3.7222494406209174e-06, "loss": 0.8238, "step": 674 }, { "epoch": 0.68, "grad_norm": 4.557745951925816, "learning_rate": 3.718766511787606e-06, "loss": 0.8367, "step": 675 }, { "epoch": 0.68, "grad_norm": 4.16754573699968, "learning_rate": 3.715280477179382e-06, "loss": 0.8561, "step": 676 }, { "epoch": 0.68, "grad_norm": 4.213296187055132, "learning_rate": 3.7117913456796855e-06, "loss": 0.8006, "step": 677 }, { "epoch": 0.69, "grad_norm": 3.895541456046058, "learning_rate": 3.7082991261798473e-06, "loss": 0.8019, "step": 678 }, { "epoch": 0.69, "grad_norm": 4.225646010065031, "learning_rate": 3.7048038275790695e-06, "loss": 0.7799, "step": 679 }, { "epoch": 0.69, "grad_norm": 3.793719931690171, "learning_rate": 3.701305458784397e-06, "loss": 0.7877, "step": 680 }, { "epoch": 0.69, "grad_norm": 3.6918538059697505, "learning_rate": 3.697804028710703e-06, "loss": 0.8472, "step": 681 }, { "epoch": 0.69, "grad_norm": 3.7943218099807967, "learning_rate": 3.6942995462806574e-06, "loss": 0.84, "step": 682 }, { "epoch": 0.69, "grad_norm": 3.603972411456462, "learning_rate": 3.6907920204247122e-06, "loss": 0.8205, "step": 683 }, { "epoch": 0.69, "grad_norm": 3.755592793493984, "learning_rate": 3.6872814600810716e-06, "loss": 0.8354, "step": 684 }, { "epoch": 0.69, "grad_norm": 4.658007950410123, "learning_rate": 3.6837678741956747e-06, "loss": 0.8767, "step": 685 }, { "epoch": 0.69, "grad_norm": 3.4964527787354416, "learning_rate": 3.6802512717221694e-06, "loss": 0.7814, "step": 686 }, { "epoch": 0.69, "grad_norm": 3.434255209389191, "learning_rate": 3.6767316616218927e-06, "loss": 0.8071, "step": 687 }, { "epoch": 0.7, "grad_norm": 4.272321481007903, "learning_rate": 3.6732090528638432e-06, "loss": 0.7787, "step": 688 }, { "epoch": 0.7, "grad_norm": 4.221528130430344, "learning_rate": 3.669683454424663e-06, "loss": 0.7729, "step": 689 }, { "epoch": 0.7, "grad_norm": 3.9597068009542915, "learning_rate": 3.6661548752886113e-06, "loss": 0.7689, "step": 690 }, { "epoch": 0.7, "grad_norm": 3.5568349195672253, "learning_rate": 3.6626233244475445e-06, "loss": 0.9255, "step": 691 }, { "epoch": 0.7, "grad_norm": 4.719503943417636, "learning_rate": 3.6590888109008904e-06, "loss": 0.8281, "step": 692 }, { "epoch": 0.7, "grad_norm": 3.6514574646812568, "learning_rate": 3.6555513436556285e-06, "loss": 0.8008, "step": 693 }, { "epoch": 0.7, "grad_norm": 4.548664005213854, "learning_rate": 3.6520109317262624e-06, "loss": 0.7797, "step": 694 }, { "epoch": 0.7, "grad_norm": 3.704773651343669, "learning_rate": 3.6484675841348024e-06, "loss": 0.813, "step": 695 }, { "epoch": 0.7, "grad_norm": 3.6582715454571657, "learning_rate": 3.6449213099107376e-06, "loss": 0.8007, "step": 696 }, { "epoch": 0.7, "grad_norm": 4.094141645704761, "learning_rate": 3.6413721180910165e-06, "loss": 0.837, "step": 697 }, { "epoch": 0.71, "grad_norm": 4.618837723050664, "learning_rate": 3.6378200177200223e-06, "loss": 0.8316, "step": 698 }, { "epoch": 0.71, "grad_norm": 3.38898352006268, "learning_rate": 3.6342650178495488e-06, "loss": 0.8103, "step": 699 }, { "epoch": 0.71, "grad_norm": 3.7976368474709896, "learning_rate": 3.6307071275387807e-06, "loss": 0.8059, "step": 700 }, { "epoch": 0.71, "grad_norm": 3.5189322963766205, "learning_rate": 3.6271463558542646e-06, "loss": 0.7959, "step": 701 }, { "epoch": 0.71, "grad_norm": 3.3331229796313027, "learning_rate": 3.623582711869895e-06, "loss": 0.7882, "step": 702 }, { "epoch": 0.71, "grad_norm": 5.530082944684263, "learning_rate": 3.6200162046668826e-06, "loss": 0.79, "step": 703 }, { "epoch": 0.71, "grad_norm": 3.3051251617528568, "learning_rate": 3.616446843333733e-06, "loss": 0.7773, "step": 704 }, { "epoch": 0.71, "grad_norm": 3.3812653747476427, "learning_rate": 3.6128746369662283e-06, "loss": 0.7819, "step": 705 }, { "epoch": 0.71, "grad_norm": 3.5371020189256805, "learning_rate": 3.6092995946673996e-06, "loss": 0.8518, "step": 706 }, { "epoch": 0.71, "grad_norm": 3.834778332537793, "learning_rate": 3.6057217255475034e-06, "loss": 0.8418, "step": 707 }, { "epoch": 0.72, "grad_norm": 4.415666375807929, "learning_rate": 3.602141038724001e-06, "loss": 0.8146, "step": 708 }, { "epoch": 0.72, "grad_norm": 3.641170161321054, "learning_rate": 3.5985575433215345e-06, "loss": 0.8347, "step": 709 }, { "epoch": 0.72, "grad_norm": 3.7635739548814207, "learning_rate": 3.5949712484719013e-06, "loss": 0.8207, "step": 710 }, { "epoch": 0.72, "grad_norm": 4.1180161320176225, "learning_rate": 3.591382163314034e-06, "loss": 0.9224, "step": 711 }, { "epoch": 0.72, "grad_norm": 4.130810975608334, "learning_rate": 3.587790296993976e-06, "loss": 0.843, "step": 712 }, { "epoch": 0.72, "grad_norm": 4.691833335291648, "learning_rate": 3.5841956586648553e-06, "loss": 0.9047, "step": 713 }, { "epoch": 0.72, "grad_norm": 4.657933022135858, "learning_rate": 3.5805982574868673e-06, "loss": 1.1555, "step": 714 }, { "epoch": 0.72, "grad_norm": 3.2260069416532207, "learning_rate": 3.5769981026272477e-06, "loss": 0.7869, "step": 715 }, { "epoch": 0.72, "grad_norm": 3.4291174313472643, "learning_rate": 3.5733952032602454e-06, "loss": 0.7723, "step": 716 }, { "epoch": 0.72, "grad_norm": 3.919942403046828, "learning_rate": 3.5697895685671076e-06, "loss": 0.8133, "step": 717 }, { "epoch": 0.73, "grad_norm": 3.9420409702367167, "learning_rate": 3.5661812077360496e-06, "loss": 0.8135, "step": 718 }, { "epoch": 0.73, "grad_norm": 4.062763811575889, "learning_rate": 3.562570129962234e-06, "loss": 0.8098, "step": 719 }, { "epoch": 0.73, "grad_norm": 3.1692597156645306, "learning_rate": 3.5589563444477477e-06, "loss": 0.7864, "step": 720 }, { "epoch": 0.73, "grad_norm": 4.529414675977005, "learning_rate": 3.5553398604015777e-06, "loss": 0.8194, "step": 721 }, { "epoch": 0.73, "grad_norm": 3.6410323820649144, "learning_rate": 3.5517206870395852e-06, "loss": 0.8061, "step": 722 }, { "epoch": 0.73, "grad_norm": 3.5825284359470673, "learning_rate": 3.548098833584489e-06, "loss": 0.8055, "step": 723 }, { "epoch": 0.73, "grad_norm": 3.6994250667973314, "learning_rate": 3.544474309265834e-06, "loss": 0.8014, "step": 724 }, { "epoch": 0.73, "grad_norm": 4.609457196655879, "learning_rate": 3.5408471233199713e-06, "loss": 0.8953, "step": 725 }, { "epoch": 0.73, "grad_norm": 4.357338921046579, "learning_rate": 3.5372172849900377e-06, "loss": 0.815, "step": 726 }, { "epoch": 0.73, "grad_norm": 4.534756505244773, "learning_rate": 3.5335848035259257e-06, "loss": 0.8281, "step": 727 }, { "epoch": 0.74, "grad_norm": 4.46200388603068, "learning_rate": 3.5299496881842654e-06, "loss": 0.8979, "step": 728 }, { "epoch": 0.74, "grad_norm": 3.8353237250725893, "learning_rate": 3.526311948228397e-06, "loss": 0.8392, "step": 729 }, { "epoch": 0.74, "grad_norm": 3.513564550196794, "learning_rate": 3.5226715929283507e-06, "loss": 0.8276, "step": 730 }, { "epoch": 0.74, "grad_norm": 4.840869909676071, "learning_rate": 3.5190286315608196e-06, "loss": 0.7808, "step": 731 }, { "epoch": 0.74, "grad_norm": 3.6376115911500326, "learning_rate": 3.5153830734091404e-06, "loss": 0.7985, "step": 732 }, { "epoch": 0.74, "grad_norm": 8.73646238891082, "learning_rate": 3.511734927763265e-06, "loss": 0.8393, "step": 733 }, { "epoch": 0.74, "grad_norm": 4.215260596616087, "learning_rate": 3.508084203919739e-06, "loss": 0.8183, "step": 734 }, { "epoch": 0.74, "grad_norm": 4.0886422935222395, "learning_rate": 3.50443091118168e-06, "loss": 0.7957, "step": 735 }, { "epoch": 0.74, "grad_norm": 3.7885638513567383, "learning_rate": 3.5007750588587495e-06, "loss": 0.8112, "step": 736 }, { "epoch": 0.74, "grad_norm": 3.9547192279690653, "learning_rate": 3.497116656267133e-06, "loss": 0.7731, "step": 737 }, { "epoch": 0.75, "grad_norm": 3.86013814393367, "learning_rate": 3.493455712729514e-06, "loss": 0.8547, "step": 738 }, { "epoch": 0.75, "grad_norm": 3.9264148109718997, "learning_rate": 3.4897922375750517e-06, "loss": 0.7858, "step": 739 }, { "epoch": 0.75, "grad_norm": 4.5652346553585925, "learning_rate": 3.4861262401393563e-06, "loss": 0.785, "step": 740 }, { "epoch": 0.75, "grad_norm": 4.1139632776522586, "learning_rate": 3.4824577297644663e-06, "loss": 0.9653, "step": 741 }, { "epoch": 0.75, "grad_norm": 5.588885807787667, "learning_rate": 3.478786715798823e-06, "loss": 0.8312, "step": 742 }, { "epoch": 0.75, "grad_norm": 6.383737000046943, "learning_rate": 3.4751132075972473e-06, "loss": 0.7787, "step": 743 }, { "epoch": 0.75, "grad_norm": 4.038979136777571, "learning_rate": 3.471437214520917e-06, "loss": 0.8058, "step": 744 }, { "epoch": 0.75, "grad_norm": 4.227693679919142, "learning_rate": 3.4677587459373417e-06, "loss": 0.9012, "step": 745 }, { "epoch": 0.75, "grad_norm": 3.985607862312997, "learning_rate": 3.46407781122034e-06, "loss": 0.819, "step": 746 }, { "epoch": 0.75, "grad_norm": 3.922668922254852, "learning_rate": 3.4603944197500126e-06, "loss": 0.7354, "step": 747 }, { "epoch": 0.76, "grad_norm": 4.343492283466782, "learning_rate": 3.4567085809127247e-06, "loss": 0.786, "step": 748 }, { "epoch": 0.76, "grad_norm": 4.484528470192613, "learning_rate": 3.4530203041010745e-06, "loss": 0.7849, "step": 749 }, { "epoch": 0.76, "grad_norm": 4.367273259116789, "learning_rate": 3.449329598713874e-06, "loss": 0.7921, "step": 750 }, { "epoch": 0.76, "grad_norm": 3.927650734068294, "learning_rate": 3.4456364741561256e-06, "loss": 0.7746, "step": 751 }, { "epoch": 0.76, "grad_norm": 4.586041331240587, "learning_rate": 3.4419409398389937e-06, "loss": 0.8183, "step": 752 }, { "epoch": 0.76, "grad_norm": 3.7164145526774224, "learning_rate": 3.4382430051797844e-06, "loss": 0.8202, "step": 753 }, { "epoch": 0.76, "grad_norm": 4.394679226693479, "learning_rate": 3.434542679601922e-06, "loss": 0.7859, "step": 754 }, { "epoch": 0.76, "grad_norm": 4.096618466435277, "learning_rate": 3.430839972534923e-06, "loss": 0.7583, "step": 755 }, { "epoch": 0.76, "grad_norm": 4.013175572729404, "learning_rate": 3.4271348934143707e-06, "loss": 0.8071, "step": 756 }, { "epoch": 0.76, "grad_norm": 4.292337945500214, "learning_rate": 3.423427451681895e-06, "loss": 0.8085, "step": 757 }, { "epoch": 0.77, "grad_norm": 3.9544343369431463, "learning_rate": 3.4197176567851463e-06, "loss": 0.8177, "step": 758 }, { "epoch": 0.77, "grad_norm": 4.433418332095438, "learning_rate": 3.4160055181777714e-06, "loss": 0.7884, "step": 759 }, { "epoch": 0.77, "grad_norm": 4.956803986907128, "learning_rate": 3.4122910453193885e-06, "loss": 0.7959, "step": 760 }, { "epoch": 0.77, "grad_norm": 5.450063143284857, "learning_rate": 3.4085742476755657e-06, "loss": 0.9269, "step": 761 }, { "epoch": 0.77, "grad_norm": 5.196193261350649, "learning_rate": 3.404855134717795e-06, "loss": 1.0828, "step": 762 }, { "epoch": 0.77, "grad_norm": 8.093995434969097, "learning_rate": 3.4011337159234674e-06, "loss": 0.7985, "step": 763 }, { "epoch": 0.77, "grad_norm": 4.575208484444267, "learning_rate": 3.3974100007758514e-06, "loss": 0.7914, "step": 764 }, { "epoch": 0.77, "grad_norm": 4.88561287010472, "learning_rate": 3.3936839987640665e-06, "loss": 0.8413, "step": 765 }, { "epoch": 0.77, "grad_norm": 4.10258246501254, "learning_rate": 3.3899557193830585e-06, "loss": 0.7827, "step": 766 }, { "epoch": 0.77, "grad_norm": 4.134612322280866, "learning_rate": 3.3862251721335793e-06, "loss": 0.7846, "step": 767 }, { "epoch": 0.78, "grad_norm": 4.127521869620768, "learning_rate": 3.382492366522158e-06, "loss": 0.8187, "step": 768 }, { "epoch": 0.78, "grad_norm": 4.694265995175311, "learning_rate": 3.3787573120610794e-06, "loss": 0.9095, "step": 769 }, { "epoch": 0.78, "grad_norm": 4.832053974686459, "learning_rate": 3.375020018268359e-06, "loss": 0.8422, "step": 770 }, { "epoch": 0.78, "grad_norm": 3.615649772558988, "learning_rate": 3.371280494667719e-06, "loss": 0.7797, "step": 771 }, { "epoch": 0.78, "grad_norm": 4.377263570264409, "learning_rate": 3.367538750788563e-06, "loss": 0.9017, "step": 772 }, { "epoch": 0.78, "grad_norm": 5.6300240178979815, "learning_rate": 3.3637947961659533e-06, "loss": 0.8064, "step": 773 }, { "epoch": 0.78, "grad_norm": 4.977615821468784, "learning_rate": 3.360048640340585e-06, "loss": 0.7614, "step": 774 }, { "epoch": 0.78, "grad_norm": 3.9739872782044285, "learning_rate": 3.356300292858763e-06, "loss": 0.8155, "step": 775 }, { "epoch": 0.78, "grad_norm": 4.365429767411158, "learning_rate": 3.3525497632723786e-06, "loss": 0.7931, "step": 776 }, { "epoch": 0.79, "grad_norm": 4.03510482219512, "learning_rate": 3.348797061138881e-06, "loss": 0.7533, "step": 777 }, { "epoch": 0.79, "grad_norm": 4.423899720091732, "learning_rate": 3.345042196021257e-06, "loss": 0.849, "step": 778 }, { "epoch": 0.79, "grad_norm": 4.3577483066057745, "learning_rate": 3.3412851774880066e-06, "loss": 0.914, "step": 779 }, { "epoch": 0.79, "grad_norm": 4.8429976699063095, "learning_rate": 3.3375260151131156e-06, "loss": 0.8208, "step": 780 }, { "epoch": 0.79, "grad_norm": 3.846433730762683, "learning_rate": 3.333764718476032e-06, "loss": 0.7667, "step": 781 }, { "epoch": 0.79, "grad_norm": 7.424150572222642, "learning_rate": 3.3300012971616467e-06, "loss": 0.7906, "step": 782 }, { "epoch": 0.79, "grad_norm": 3.78110316938379, "learning_rate": 3.3262357607602596e-06, "loss": 0.8568, "step": 783 }, { "epoch": 0.79, "grad_norm": 4.484285048821905, "learning_rate": 3.3224681188675643e-06, "loss": 0.7874, "step": 784 }, { "epoch": 0.79, "grad_norm": 4.710007453384313, "learning_rate": 3.318698381084619e-06, "loss": 0.812, "step": 785 }, { "epoch": 0.79, "grad_norm": 4.303370184571771, "learning_rate": 3.3149265570178215e-06, "loss": 0.9035, "step": 786 }, { "epoch": 0.8, "grad_norm": 4.26213714508879, "learning_rate": 3.3111526562788864e-06, "loss": 0.8047, "step": 787 }, { "epoch": 0.8, "grad_norm": 3.9733132790695183, "learning_rate": 3.3073766884848235e-06, "loss": 0.7961, "step": 788 }, { "epoch": 0.8, "grad_norm": 5.156615645995947, "learning_rate": 3.303598663257904e-06, "loss": 0.824, "step": 789 }, { "epoch": 0.8, "grad_norm": 4.736017510667661, "learning_rate": 3.2998185902256475e-06, "loss": 0.8016, "step": 790 }, { "epoch": 0.8, "grad_norm": 5.180043733793966, "learning_rate": 3.2960364790207897e-06, "loss": 0.7707, "step": 791 }, { "epoch": 0.8, "grad_norm": 4.905861811752156, "learning_rate": 3.2922523392812605e-06, "loss": 0.7466, "step": 792 }, { "epoch": 0.8, "grad_norm": 3.923996642696014, "learning_rate": 3.2884661806501576e-06, "loss": 0.8046, "step": 793 }, { "epoch": 0.8, "grad_norm": 3.9631467827534426, "learning_rate": 3.284678012775727e-06, "loss": 0.9511, "step": 794 }, { "epoch": 0.8, "grad_norm": 3.965325234874838, "learning_rate": 3.2808878453113317e-06, "loss": 0.8636, "step": 795 }, { "epoch": 0.8, "grad_norm": 3.642633499379332, "learning_rate": 3.2770956879154305e-06, "loss": 0.7702, "step": 796 }, { "epoch": 0.81, "grad_norm": 4.492042123923082, "learning_rate": 3.273301550251555e-06, "loss": 0.8537, "step": 797 }, { "epoch": 0.81, "grad_norm": 3.8876988328768602, "learning_rate": 3.269505441988281e-06, "loss": 0.8605, "step": 798 }, { "epoch": 0.81, "grad_norm": 3.766606063581812, "learning_rate": 3.2657073727992078e-06, "loss": 0.7985, "step": 799 }, { "epoch": 0.81, "grad_norm": 4.267264555929007, "learning_rate": 3.2619073523629304e-06, "loss": 0.7971, "step": 800 }, { "epoch": 0.81, "grad_norm": 4.919439020397232, "learning_rate": 3.258105390363016e-06, "loss": 0.8164, "step": 801 }, { "epoch": 0.81, "grad_norm": 4.24583647910275, "learning_rate": 3.2543014964879814e-06, "loss": 0.8043, "step": 802 }, { "epoch": 0.81, "grad_norm": 3.4364112443855985, "learning_rate": 3.250495680431264e-06, "loss": 0.769, "step": 803 }, { "epoch": 0.81, "grad_norm": 4.40163318603266, "learning_rate": 3.246687951891201e-06, "loss": 0.7878, "step": 804 }, { "epoch": 0.81, "grad_norm": 4.273441543771241, "learning_rate": 3.2428783205710023e-06, "loss": 0.8331, "step": 805 }, { "epoch": 0.81, "grad_norm": 3.7031687463067633, "learning_rate": 3.2390667961787276e-06, "loss": 0.7873, "step": 806 }, { "epoch": 0.82, "grad_norm": 3.5284627603373284, "learning_rate": 3.2352533884272596e-06, "loss": 0.7826, "step": 807 }, { "epoch": 0.82, "grad_norm": 3.901059309743076, "learning_rate": 3.2314381070342815e-06, "loss": 0.9248, "step": 808 }, { "epoch": 0.82, "grad_norm": 13.116776373765031, "learning_rate": 3.2276209617222497e-06, "loss": 0.8109, "step": 809 }, { "epoch": 0.82, "grad_norm": 3.5825950028846907, "learning_rate": 3.223801962218372e-06, "loss": 0.7666, "step": 810 }, { "epoch": 0.82, "grad_norm": 3.500505244324813, "learning_rate": 3.21998111825458e-06, "loss": 0.7398, "step": 811 }, { "epoch": 0.82, "grad_norm": 3.496092976488147, "learning_rate": 3.216158439567506e-06, "loss": 0.7631, "step": 812 }, { "epoch": 0.82, "grad_norm": 4.556135014522858, "learning_rate": 3.2123339358984573e-06, "loss": 0.8207, "step": 813 }, { "epoch": 0.82, "grad_norm": 3.7277520517697136, "learning_rate": 3.208507616993393e-06, "loss": 0.8516, "step": 814 }, { "epoch": 0.82, "grad_norm": 4.423036921270627, "learning_rate": 3.2046794926028966e-06, "loss": 0.791, "step": 815 }, { "epoch": 0.82, "grad_norm": 3.235611807092007, "learning_rate": 3.200849572482153e-06, "loss": 0.7558, "step": 816 }, { "epoch": 0.83, "grad_norm": 3.824187344341294, "learning_rate": 3.1970178663909233e-06, "loss": 0.7658, "step": 817 }, { "epoch": 0.83, "grad_norm": 4.275625768452698, "learning_rate": 3.19318438409352e-06, "loss": 0.8375, "step": 818 }, { "epoch": 0.83, "grad_norm": 4.566672221414914, "learning_rate": 3.189349135358781e-06, "loss": 0.762, "step": 819 }, { "epoch": 0.83, "grad_norm": 4.020446057440514, "learning_rate": 3.1855121299600454e-06, "loss": 0.8817, "step": 820 }, { "epoch": 0.83, "grad_norm": 4.2881734956366095, "learning_rate": 3.1816733776751313e-06, "loss": 0.8106, "step": 821 }, { "epoch": 0.83, "grad_norm": 4.673485294928623, "learning_rate": 3.1778328882863058e-06, "loss": 0.9695, "step": 822 }, { "epoch": 0.83, "grad_norm": 7.339180964057777, "learning_rate": 3.173990671580263e-06, "loss": 0.7922, "step": 823 }, { "epoch": 0.83, "grad_norm": 3.7378322020107713, "learning_rate": 3.170146737348099e-06, "loss": 0.8213, "step": 824 }, { "epoch": 0.83, "grad_norm": 4.440411381382036, "learning_rate": 3.166301095385288e-06, "loss": 0.8058, "step": 825 }, { "epoch": 0.83, "grad_norm": 3.645734306429232, "learning_rate": 3.162453755491655e-06, "loss": 0.7686, "step": 826 }, { "epoch": 0.84, "grad_norm": 4.646322329035178, "learning_rate": 3.1586047274713493e-06, "loss": 0.8386, "step": 827 }, { "epoch": 0.84, "grad_norm": 4.205705304738026, "learning_rate": 3.154754021132827e-06, "loss": 0.857, "step": 828 }, { "epoch": 0.84, "grad_norm": 49.63215769717914, "learning_rate": 3.1509016462888175e-06, "loss": 0.8218, "step": 829 }, { "epoch": 0.84, "grad_norm": 4.780984494435262, "learning_rate": 3.147047612756302e-06, "loss": 0.8394, "step": 830 }, { "epoch": 0.84, "grad_norm": 4.58164822302537, "learning_rate": 3.143191930356491e-06, "loss": 0.837, "step": 831 }, { "epoch": 0.84, "grad_norm": 4.852494144118555, "learning_rate": 3.139334608914795e-06, "loss": 0.9018, "step": 832 }, { "epoch": 0.84, "grad_norm": 4.145761545326323, "learning_rate": 3.135475658260801e-06, "loss": 0.8081, "step": 833 }, { "epoch": 0.84, "grad_norm": 3.7323925690162945, "learning_rate": 3.1316150882282486e-06, "loss": 0.7904, "step": 834 }, { "epoch": 0.84, "grad_norm": 3.828580217813585, "learning_rate": 3.1277529086550044e-06, "loss": 0.8209, "step": 835 }, { "epoch": 0.84, "grad_norm": 7.169193160513471, "learning_rate": 3.1238891293830344e-06, "loss": 0.8518, "step": 836 }, { "epoch": 0.85, "grad_norm": 3.7890433555958705, "learning_rate": 3.120023760258384e-06, "loss": 0.81, "step": 837 }, { "epoch": 0.85, "grad_norm": 5.4100072300835915, "learning_rate": 3.1161568111311487e-06, "loss": 0.8143, "step": 838 }, { "epoch": 0.85, "grad_norm": 4.745257896570691, "learning_rate": 3.112288291855449e-06, "loss": 0.7812, "step": 839 }, { "epoch": 0.85, "grad_norm": 3.989422633795527, "learning_rate": 3.108418212289408e-06, "loss": 0.7939, "step": 840 }, { "epoch": 0.85, "grad_norm": 3.69468243636268, "learning_rate": 3.1045465822951265e-06, "loss": 0.7719, "step": 841 }, { "epoch": 0.85, "grad_norm": 3.734646448984784, "learning_rate": 3.1006734117386517e-06, "loss": 0.8206, "step": 842 }, { "epoch": 0.85, "grad_norm": 4.406895307041761, "learning_rate": 3.0967987104899623e-06, "loss": 0.849, "step": 843 }, { "epoch": 0.85, "grad_norm": 5.120926190738434, "learning_rate": 3.092922488422933e-06, "loss": 0.8067, "step": 844 }, { "epoch": 0.85, "grad_norm": 4.395257481206765, "learning_rate": 3.0890447554153153e-06, "loss": 0.7899, "step": 845 }, { "epoch": 0.85, "grad_norm": 4.381884034429759, "learning_rate": 3.0851655213487123e-06, "loss": 0.8286, "step": 846 }, { "epoch": 0.86, "grad_norm": 3.4432938469880736, "learning_rate": 3.0812847961085527e-06, "loss": 0.7555, "step": 847 }, { "epoch": 0.86, "grad_norm": 5.491662807975452, "learning_rate": 3.077402589584061e-06, "loss": 0.7953, "step": 848 }, { "epoch": 0.86, "grad_norm": 5.109785316548567, "learning_rate": 3.073518911668241e-06, "loss": 0.8068, "step": 849 }, { "epoch": 0.86, "grad_norm": 4.816255710860229, "learning_rate": 3.0696337722578444e-06, "loss": 0.8116, "step": 850 }, { "epoch": 0.86, "grad_norm": 4.0550180873058945, "learning_rate": 3.0657471812533464e-06, "loss": 0.7445, "step": 851 }, { "epoch": 0.86, "grad_norm": 6.428409369009171, "learning_rate": 3.061859148558922e-06, "loss": 0.7809, "step": 852 }, { "epoch": 0.86, "grad_norm": 7.399300412933422, "learning_rate": 3.057969684082421e-06, "loss": 0.9132, "step": 853 }, { "epoch": 0.86, "grad_norm": 4.438275269071275, "learning_rate": 3.05407879773534e-06, "loss": 0.7677, "step": 854 }, { "epoch": 0.86, "grad_norm": 4.029683232488459, "learning_rate": 3.0501864994328002e-06, "loss": 0.8277, "step": 855 }, { "epoch": 0.86, "grad_norm": 6.985086578254067, "learning_rate": 3.04629279909352e-06, "loss": 0.7709, "step": 856 }, { "epoch": 0.87, "grad_norm": 3.989233419159114, "learning_rate": 3.0423977066397913e-06, "loss": 0.7942, "step": 857 }, { "epoch": 0.87, "grad_norm": 5.59818214654168, "learning_rate": 3.038501231997454e-06, "loss": 0.7889, "step": 858 }, { "epoch": 0.87, "grad_norm": 4.738840603418049, "learning_rate": 3.0346033850958685e-06, "loss": 0.8176, "step": 859 }, { "epoch": 0.87, "grad_norm": 4.316647635809497, "learning_rate": 3.0307041758678933e-06, "loss": 0.771, "step": 860 }, { "epoch": 0.87, "grad_norm": 5.364622380247118, "learning_rate": 3.0268036142498596e-06, "loss": 0.7959, "step": 861 }, { "epoch": 0.87, "grad_norm": 9.785481265487443, "learning_rate": 3.0229017101815424e-06, "loss": 0.8661, "step": 862 }, { "epoch": 0.87, "grad_norm": 4.766052528284133, "learning_rate": 3.018998473606139e-06, "loss": 0.7826, "step": 863 }, { "epoch": 0.87, "grad_norm": 4.701099515941216, "learning_rate": 3.0150939144702425e-06, "loss": 0.8236, "step": 864 }, { "epoch": 0.87, "grad_norm": 9.461997482098507, "learning_rate": 3.011188042723816e-06, "loss": 0.7458, "step": 865 }, { "epoch": 0.87, "grad_norm": 5.380245649553912, "learning_rate": 3.007280868320167e-06, "loss": 0.9145, "step": 866 }, { "epoch": 0.88, "grad_norm": 10.163712466924556, "learning_rate": 3.0033724012159244e-06, "loss": 0.8479, "step": 867 }, { "epoch": 0.88, "grad_norm": 4.2825214697150455, "learning_rate": 2.9994626513710085e-06, "loss": 0.8418, "step": 868 }, { "epoch": 0.88, "grad_norm": 4.133634973081057, "learning_rate": 2.9955516287486103e-06, "loss": 0.7698, "step": 869 }, { "epoch": 0.88, "grad_norm": 4.448695379643913, "learning_rate": 2.9916393433151636e-06, "loss": 0.8066, "step": 870 }, { "epoch": 0.88, "grad_norm": 4.466338581251602, "learning_rate": 2.9877258050403214e-06, "loss": 0.8007, "step": 871 }, { "epoch": 0.88, "grad_norm": 10.810936084229239, "learning_rate": 2.9838110238969266e-06, "loss": 0.8186, "step": 872 }, { "epoch": 0.88, "grad_norm": 4.803654636244916, "learning_rate": 2.9798950098609925e-06, "loss": 0.8928, "step": 873 }, { "epoch": 0.88, "grad_norm": 4.603349966398751, "learning_rate": 2.975977772911671e-06, "loss": 0.82, "step": 874 }, { "epoch": 0.88, "grad_norm": 4.1551088640505, "learning_rate": 2.9720593230312337e-06, "loss": 0.8218, "step": 875 }, { "epoch": 0.89, "grad_norm": 4.340225183652387, "learning_rate": 2.968139670205041e-06, "loss": 0.8037, "step": 876 }, { "epoch": 0.89, "grad_norm": 4.864128739357516, "learning_rate": 2.964218824421518e-06, "loss": 0.7639, "step": 877 }, { "epoch": 0.89, "grad_norm": 3.9960489196439664, "learning_rate": 2.960296795672132e-06, "loss": 0.7699, "step": 878 }, { "epoch": 0.89, "grad_norm": 4.852346612250402, "learning_rate": 2.9563735939513637e-06, "loss": 0.8014, "step": 879 }, { "epoch": 0.89, "grad_norm": 4.3848451966395325, "learning_rate": 2.9524492292566824e-06, "loss": 0.8093, "step": 880 }, { "epoch": 0.89, "grad_norm": 8.205235412527301, "learning_rate": 2.9485237115885223e-06, "loss": 0.8619, "step": 881 }, { "epoch": 0.89, "grad_norm": 6.176360237119004, "learning_rate": 2.9445970509502547e-06, "loss": 0.7676, "step": 882 }, { "epoch": 0.89, "grad_norm": 4.630007100247436, "learning_rate": 2.9406692573481634e-06, "loss": 0.8479, "step": 883 }, { "epoch": 0.89, "grad_norm": 4.947598794426371, "learning_rate": 2.93674034079142e-06, "loss": 0.8159, "step": 884 }, { "epoch": 0.89, "grad_norm": 5.208912980619626, "learning_rate": 2.932810311292058e-06, "loss": 0.7876, "step": 885 }, { "epoch": 0.9, "grad_norm": 5.419301085882204, "learning_rate": 2.928879178864946e-06, "loss": 0.8171, "step": 886 }, { "epoch": 0.9, "grad_norm": 4.453129993869502, "learning_rate": 2.9249469535277635e-06, "loss": 0.8192, "step": 887 }, { "epoch": 0.9, "grad_norm": 3.978894479574167, "learning_rate": 2.9210136453009753e-06, "loss": 0.8177, "step": 888 }, { "epoch": 0.9, "grad_norm": 4.085215406941065, "learning_rate": 2.9170792642078057e-06, "loss": 0.7986, "step": 889 }, { "epoch": 0.9, "grad_norm": 4.6198554119473725, "learning_rate": 2.9131438202742123e-06, "loss": 0.7632, "step": 890 }, { "epoch": 0.9, "grad_norm": 4.722723234078222, "learning_rate": 2.9092073235288633e-06, "loss": 0.8149, "step": 891 }, { "epoch": 0.9, "grad_norm": 5.181515177263234, "learning_rate": 2.9052697840031065e-06, "loss": 0.8288, "step": 892 }, { "epoch": 0.9, "grad_norm": 4.794572021110422, "learning_rate": 2.901331211730949e-06, "loss": 0.7967, "step": 893 }, { "epoch": 0.9, "grad_norm": 4.424733629629019, "learning_rate": 2.897391616749031e-06, "loss": 0.7658, "step": 894 }, { "epoch": 0.9, "grad_norm": 4.217478880240176, "learning_rate": 2.8934510090965943e-06, "loss": 0.7973, "step": 895 }, { "epoch": 0.91, "grad_norm": 8.450381133458173, "learning_rate": 2.889509398815467e-06, "loss": 0.8968, "step": 896 }, { "epoch": 0.91, "grad_norm": 6.678550127443945, "learning_rate": 2.885566795950028e-06, "loss": 0.7672, "step": 897 }, { "epoch": 0.91, "grad_norm": 4.847140586125922, "learning_rate": 2.8816232105471864e-06, "loss": 0.7939, "step": 898 }, { "epoch": 0.91, "grad_norm": 4.027451532927153, "learning_rate": 2.877678652656358e-06, "loss": 0.7753, "step": 899 }, { "epoch": 0.91, "grad_norm": 5.483294011279615, "learning_rate": 2.8737331323294315e-06, "loss": 0.7848, "step": 900 }, { "epoch": 0.91, "grad_norm": 4.358769311669604, "learning_rate": 2.8697866596207524e-06, "loss": 0.9225, "step": 901 }, { "epoch": 0.91, "grad_norm": 3.9764053893031757, "learning_rate": 2.865839244587093e-06, "loss": 0.8153, "step": 902 }, { "epoch": 0.91, "grad_norm": 4.817588176503937, "learning_rate": 2.861890897287625e-06, "loss": 0.8854, "step": 903 }, { "epoch": 0.91, "grad_norm": 5.88032298358066, "learning_rate": 2.8579416277838952e-06, "loss": 0.8062, "step": 904 }, { "epoch": 0.91, "grad_norm": 5.4231075227119225, "learning_rate": 2.8539914461398043e-06, "loss": 0.8082, "step": 905 }, { "epoch": 0.92, "grad_norm": 4.255932637689089, "learning_rate": 2.8500403624215733e-06, "loss": 0.801, "step": 906 }, { "epoch": 0.92, "grad_norm": 3.8397836719406153, "learning_rate": 2.846088386697723e-06, "loss": 0.8418, "step": 907 }, { "epoch": 0.92, "grad_norm": 3.9932629786289753, "learning_rate": 2.8421355290390506e-06, "loss": 0.8162, "step": 908 }, { "epoch": 0.92, "grad_norm": 4.487065147178443, "learning_rate": 2.838181799518595e-06, "loss": 0.8014, "step": 909 }, { "epoch": 0.92, "grad_norm": 3.914529347080895, "learning_rate": 2.8342272082116214e-06, "loss": 0.8356, "step": 910 }, { "epoch": 0.92, "grad_norm": 4.004022127968303, "learning_rate": 2.83027176519559e-06, "loss": 0.8389, "step": 911 }, { "epoch": 0.92, "grad_norm": 8.828551661954858, "learning_rate": 2.82631548055013e-06, "loss": 0.7885, "step": 912 }, { "epoch": 0.92, "grad_norm": 4.492552440308808, "learning_rate": 2.822358364357015e-06, "loss": 0.8, "step": 913 }, { "epoch": 0.92, "grad_norm": 4.342294518537859, "learning_rate": 2.8184004267001427e-06, "loss": 0.7626, "step": 914 }, { "epoch": 0.92, "grad_norm": 4.14513404654822, "learning_rate": 2.8144416776654964e-06, "loss": 0.7403, "step": 915 }, { "epoch": 0.93, "grad_norm": 4.207420214999504, "learning_rate": 2.8104821273411333e-06, "loss": 0.7719, "step": 916 }, { "epoch": 0.93, "grad_norm": 5.76849216288104, "learning_rate": 2.8065217858171495e-06, "loss": 0.772, "step": 917 }, { "epoch": 0.93, "grad_norm": 5.087940784747258, "learning_rate": 2.802560663185658e-06, "loss": 0.755, "step": 918 }, { "epoch": 0.93, "grad_norm": 5.804412743927739, "learning_rate": 2.7985987695407618e-06, "loss": 0.7937, "step": 919 }, { "epoch": 0.93, "grad_norm": 4.3563337300652725, "learning_rate": 2.7946361149785304e-06, "loss": 0.7953, "step": 920 }, { "epoch": 0.93, "grad_norm": 5.353344075424233, "learning_rate": 2.79067270959697e-06, "loss": 0.7835, "step": 921 }, { "epoch": 0.93, "grad_norm": 4.606865353160219, "learning_rate": 2.786708563496002e-06, "loss": 0.8217, "step": 922 }, { "epoch": 0.93, "grad_norm": 5.151168827461153, "learning_rate": 2.7827436867774334e-06, "loss": 0.7994, "step": 923 }, { "epoch": 0.93, "grad_norm": 4.3077719748381815, "learning_rate": 2.7787780895449353e-06, "loss": 0.81, "step": 924 }, { "epoch": 0.93, "grad_norm": 5.939054267141848, "learning_rate": 2.774811781904013e-06, "loss": 0.8043, "step": 925 }, { "epoch": 0.94, "grad_norm": 3.6482393833284537, "learning_rate": 2.7708447739619833e-06, "loss": 0.7984, "step": 926 }, { "epoch": 0.94, "grad_norm": 4.469643319922747, "learning_rate": 2.7668770758279473e-06, "loss": 0.7579, "step": 927 }, { "epoch": 0.94, "grad_norm": 4.297430505606569, "learning_rate": 2.762908697612765e-06, "loss": 0.7396, "step": 928 }, { "epoch": 0.94, "grad_norm": 5.299365762769891, "learning_rate": 2.7589396494290287e-06, "loss": 0.9516, "step": 929 }, { "epoch": 0.94, "grad_norm": 4.820980097501582, "learning_rate": 2.7549699413910387e-06, "loss": 0.7753, "step": 930 }, { "epoch": 0.94, "grad_norm": 4.070459691106533, "learning_rate": 2.750999583614777e-06, "loss": 0.7792, "step": 931 }, { "epoch": 0.94, "grad_norm": 3.9507862249719974, "learning_rate": 2.7470285862178804e-06, "loss": 0.7307, "step": 932 }, { "epoch": 0.94, "grad_norm": 3.6716263593517255, "learning_rate": 2.743056959319616e-06, "loss": 0.7952, "step": 933 }, { "epoch": 0.94, "grad_norm": 4.098216751623958, "learning_rate": 2.739084713040856e-06, "loss": 0.7741, "step": 934 }, { "epoch": 0.94, "grad_norm": 5.192569762900567, "learning_rate": 2.73511185750405e-06, "loss": 0.7985, "step": 935 }, { "epoch": 0.95, "grad_norm": 4.261338885304019, "learning_rate": 2.7311384028332e-06, "loss": 0.7996, "step": 936 }, { "epoch": 0.95, "grad_norm": 4.104647555320273, "learning_rate": 2.7271643591538355e-06, "loss": 0.7984, "step": 937 }, { "epoch": 0.95, "grad_norm": 4.119079048596377, "learning_rate": 2.723189736592986e-06, "loss": 0.8057, "step": 938 }, { "epoch": 0.95, "grad_norm": 15.046890388512228, "learning_rate": 2.719214545279158e-06, "loss": 0.805, "step": 939 }, { "epoch": 0.95, "grad_norm": 5.22626992572264, "learning_rate": 2.7152387953423047e-06, "loss": 0.8173, "step": 940 }, { "epoch": 0.95, "grad_norm": 5.450359321360147, "learning_rate": 2.711262496913805e-06, "loss": 0.7907, "step": 941 }, { "epoch": 0.95, "grad_norm": 4.084842715465518, "learning_rate": 2.707285660126435e-06, "loss": 0.7497, "step": 942 }, { "epoch": 0.95, "grad_norm": 4.0680641158312865, "learning_rate": 2.703308295114342e-06, "loss": 0.8914, "step": 943 }, { "epoch": 0.95, "grad_norm": 3.9396881683951595, "learning_rate": 2.6993304120130197e-06, "loss": 0.7543, "step": 944 }, { "epoch": 0.95, "grad_norm": 8.168996657657983, "learning_rate": 2.6953520209592827e-06, "loss": 0.832, "step": 945 }, { "epoch": 0.96, "grad_norm": 4.615116753830786, "learning_rate": 2.69137313209124e-06, "loss": 0.7986, "step": 946 }, { "epoch": 0.96, "grad_norm": 4.1733727377866865, "learning_rate": 2.6873937555482664e-06, "loss": 0.8034, "step": 947 }, { "epoch": 0.96, "grad_norm": 13.47432961457841, "learning_rate": 2.683413901470984e-06, "loss": 0.8469, "step": 948 }, { "epoch": 0.96, "grad_norm": 4.634549734894035, "learning_rate": 2.6794335800012294e-06, "loss": 0.7886, "step": 949 }, { "epoch": 0.96, "grad_norm": 5.128741319695401, "learning_rate": 2.6754528012820288e-06, "loss": 0.7701, "step": 950 }, { "epoch": 0.96, "grad_norm": 4.1198392744805545, "learning_rate": 2.6714715754575757e-06, "loss": 0.803, "step": 951 }, { "epoch": 0.96, "grad_norm": 4.337325262757384, "learning_rate": 2.6674899126732045e-06, "loss": 0.8442, "step": 952 }, { "epoch": 0.96, "grad_norm": 4.143567365031371, "learning_rate": 2.663507823075358e-06, "loss": 0.7851, "step": 953 }, { "epoch": 0.96, "grad_norm": 5.745829541899124, "learning_rate": 2.6595253168115707e-06, "loss": 0.7785, "step": 954 }, { "epoch": 0.96, "grad_norm": 4.933563947271949, "learning_rate": 2.65554240403044e-06, "loss": 0.8497, "step": 955 }, { "epoch": 0.97, "grad_norm": 3.798457207492107, "learning_rate": 2.6515590948815934e-06, "loss": 0.7611, "step": 956 }, { "epoch": 0.97, "grad_norm": 4.660299502636275, "learning_rate": 2.6475753995156746e-06, "loss": 0.8, "step": 957 }, { "epoch": 0.97, "grad_norm": 4.930258601091212, "learning_rate": 2.643591328084309e-06, "loss": 1.0316, "step": 958 }, { "epoch": 0.97, "grad_norm": 4.17822237801153, "learning_rate": 2.6396068907400784e-06, "loss": 0.8839, "step": 959 }, { "epoch": 0.97, "grad_norm": 4.338544848126124, "learning_rate": 2.635622097636501e-06, "loss": 0.7836, "step": 960 }, { "epoch": 0.97, "grad_norm": 4.40163809982145, "learning_rate": 2.631636958928e-06, "loss": 0.7965, "step": 961 }, { "epoch": 0.97, "grad_norm": 4.411528020705482, "learning_rate": 2.6276514847698763e-06, "loss": 0.745, "step": 962 }, { "epoch": 0.97, "grad_norm": 4.007219388292756, "learning_rate": 2.623665685318291e-06, "loss": 0.7881, "step": 963 }, { "epoch": 0.97, "grad_norm": 3.914836989848749, "learning_rate": 2.6196795707302304e-06, "loss": 0.7929, "step": 964 }, { "epoch": 0.97, "grad_norm": 4.063241782135146, "learning_rate": 2.6156931511634838e-06, "loss": 0.801, "step": 965 }, { "epoch": 0.98, "grad_norm": 4.213904333151254, "learning_rate": 2.61170643677662e-06, "loss": 0.8638, "step": 966 }, { "epoch": 0.98, "grad_norm": 4.285745911664518, "learning_rate": 2.607719437728957e-06, "loss": 0.77, "step": 967 }, { "epoch": 0.98, "grad_norm": 4.689303741924254, "learning_rate": 2.603732164180539e-06, "loss": 0.7846, "step": 968 }, { "epoch": 0.98, "grad_norm": 4.666129815883496, "learning_rate": 2.5997446262921105e-06, "loss": 0.7803, "step": 969 }, { "epoch": 0.98, "grad_norm": 4.477070260564313, "learning_rate": 2.595756834225089e-06, "loss": 0.7998, "step": 970 }, { "epoch": 0.98, "grad_norm": 4.184569743823347, "learning_rate": 2.5917687981415375e-06, "loss": 0.8212, "step": 971 }, { "epoch": 0.98, "grad_norm": 4.242849593041545, "learning_rate": 2.5877805282041456e-06, "loss": 0.7881, "step": 972 }, { "epoch": 0.98, "grad_norm": 3.735902947064209, "learning_rate": 2.583792034576194e-06, "loss": 0.7876, "step": 973 }, { "epoch": 0.98, "grad_norm": 4.46938280982608, "learning_rate": 2.5798033274215363e-06, "loss": 0.8288, "step": 974 }, { "epoch": 0.99, "grad_norm": 4.033690870298263, "learning_rate": 2.575814416904569e-06, "loss": 0.8134, "step": 975 }, { "epoch": 0.99, "grad_norm": 4.597709987316143, "learning_rate": 2.5718253131902084e-06, "loss": 0.823, "step": 976 }, { "epoch": 0.99, "grad_norm": 4.142434190114618, "learning_rate": 2.5678360264438608e-06, "loss": 0.797, "step": 977 }, { "epoch": 0.99, "grad_norm": 4.1203821819631115, "learning_rate": 2.5638465668314006e-06, "loss": 0.7622, "step": 978 }, { "epoch": 0.99, "grad_norm": 10.857001786231981, "learning_rate": 2.5598569445191418e-06, "loss": 0.7962, "step": 979 }, { "epoch": 0.99, "grad_norm": 5.560092742577767, "learning_rate": 2.5558671696738145e-06, "loss": 0.7627, "step": 980 }, { "epoch": 0.99, "grad_norm": 4.847492024501584, "learning_rate": 2.5518772524625356e-06, "loss": 0.812, "step": 981 }, { "epoch": 0.99, "grad_norm": 4.635222745766512, "learning_rate": 2.547887203052786e-06, "loss": 0.8278, "step": 982 }, { "epoch": 0.99, "grad_norm": 4.436016580970295, "learning_rate": 2.543897031612382e-06, "loss": 0.7846, "step": 983 }, { "epoch": 0.99, "grad_norm": 3.7357290599674355, "learning_rate": 2.539906748309454e-06, "loss": 0.8013, "step": 984 }, { "epoch": 1.0, "grad_norm": 4.288805373285202, "learning_rate": 2.535916363312414e-06, "loss": 0.8952, "step": 985 }, { "epoch": 1.0, "grad_norm": 4.097132254798519, "learning_rate": 2.531925886789935e-06, "loss": 0.7853, "step": 986 }, { "epoch": 1.0, "grad_norm": 4.933595710527326, "learning_rate": 2.5279353289109227e-06, "loss": 0.7475, "step": 987 }, { "epoch": 1.0, "grad_norm": 3.626882278982255, "learning_rate": 2.52394469984449e-06, "loss": 0.7467, "step": 988 }, { "epoch": 1.0, "grad_norm": 4.637647959979263, "learning_rate": 2.5199540097599316e-06, "loss": 0.8511, "step": 989 }, { "epoch": 1.0, "grad_norm": 7.680724379345019, "learning_rate": 2.515963268826698e-06, "loss": 0.8126, "step": 990 }, { "epoch": 1.0, "eval_loss": 0.7825744152069092, "eval_runtime": 132.3278, "eval_samples_per_second": 9.877, "eval_steps_per_second": 0.28, "step": 990 }, { "epoch": 1.0, "grad_norm": 9.435150049315459, "learning_rate": 2.5119724872143693e-06, "loss": 0.7793, "step": 991 }, { "epoch": 1.0, "grad_norm": 4.428839463678267, "learning_rate": 2.5079816750926265e-06, "loss": 0.7835, "step": 992 }, { "epoch": 1.0, "grad_norm": 4.812178708587603, "learning_rate": 2.503990842631233e-06, "loss": 0.7713, "step": 993 }, { "epoch": 1.0, "grad_norm": 5.310119503968455, "learning_rate": 2.5e-06, "loss": 0.8086, "step": 994 }, { "epoch": 1.01, "grad_norm": 4.430226941353578, "learning_rate": 2.4960091573687676e-06, "loss": 0.8097, "step": 995 }, { "epoch": 1.01, "grad_norm": 3.610908985148999, "learning_rate": 2.4920183249073747e-06, "loss": 0.7767, "step": 996 }, { "epoch": 1.01, "grad_norm": 3.968405660668456, "learning_rate": 2.4880275127856324e-06, "loss": 0.8019, "step": 997 }, { "epoch": 1.01, "grad_norm": 4.619570202471707, "learning_rate": 2.4840367311733023e-06, "loss": 0.8227, "step": 998 }, { "epoch": 1.01, "grad_norm": 4.015964225489614, "learning_rate": 2.4800459902400688e-06, "loss": 0.791, "step": 999 }, { "epoch": 1.01, "grad_norm": 4.320043568159517, "learning_rate": 2.4760553001555105e-06, "loss": 0.8228, "step": 1000 }, { "epoch": 1.01, "grad_norm": 4.2300506490405, "learning_rate": 2.472064671089078e-06, "loss": 0.7791, "step": 1001 }, { "epoch": 1.01, "grad_norm": 6.379414978112713, "learning_rate": 2.4680741132100664e-06, "loss": 0.773, "step": 1002 }, { "epoch": 1.01, "grad_norm": 4.601479074672601, "learning_rate": 2.4640836366875872e-06, "loss": 0.7605, "step": 1003 }, { "epoch": 1.01, "grad_norm": 4.995926197437788, "learning_rate": 2.4600932516905467e-06, "loss": 0.7493, "step": 1004 }, { "epoch": 1.02, "grad_norm": 4.249778963774547, "learning_rate": 2.456102968387618e-06, "loss": 0.9018, "step": 1005 }, { "epoch": 1.02, "grad_norm": 4.615966431082592, "learning_rate": 2.452112796947215e-06, "loss": 0.8157, "step": 1006 }, { "epoch": 1.02, "grad_norm": 4.065873680148196, "learning_rate": 2.4481227475374652e-06, "loss": 0.8493, "step": 1007 }, { "epoch": 1.02, "grad_norm": 3.870593463530035, "learning_rate": 2.4441328303261868e-06, "loss": 0.7854, "step": 1008 }, { "epoch": 1.02, "grad_norm": 3.7564760221433473, "learning_rate": 2.440143055480859e-06, "loss": 0.7845, "step": 1009 }, { "epoch": 1.0, "grad_norm": 4.870446937033956, "learning_rate": 2.4361534331686002e-06, "loss": 0.7661, "step": 1010 }, { "epoch": 1.0, "grad_norm": 4.934194790933073, "learning_rate": 2.4321639735561405e-06, "loss": 0.8733, "step": 1011 }, { "epoch": 1.0, "grad_norm": 3.7264907087626744, "learning_rate": 2.428174686809793e-06, "loss": 0.7993, "step": 1012 }, { "epoch": 1.0, "grad_norm": 5.274510042949237, "learning_rate": 2.4241855830954317e-06, "loss": 0.7672, "step": 1013 }, { "epoch": 1.01, "grad_norm": 4.437170490276932, "learning_rate": 2.420196672578465e-06, "loss": 0.8142, "step": 1014 }, { "epoch": 1.01, "grad_norm": 4.09490786990086, "learning_rate": 2.4162079654238073e-06, "loss": 0.7932, "step": 1015 }, { "epoch": 1.01, "grad_norm": 5.5992170923610605, "learning_rate": 2.412219471795855e-06, "loss": 0.8087, "step": 1016 }, { "epoch": 1.01, "grad_norm": 3.917077566627417, "learning_rate": 2.4082312018584625e-06, "loss": 0.8343, "step": 1017 }, { "epoch": 1.01, "grad_norm": 3.590177673574075, "learning_rate": 2.404243165774912e-06, "loss": 0.7575, "step": 1018 }, { "epoch": 1.01, "grad_norm": 4.421734538692513, "learning_rate": 2.4002553737078903e-06, "loss": 0.7678, "step": 1019 }, { "epoch": 1.01, "grad_norm": 4.0726019413582115, "learning_rate": 2.3962678358194617e-06, "loss": 0.7847, "step": 1020 }, { "epoch": 1.01, "grad_norm": 4.0414910476886785, "learning_rate": 2.392280562271044e-06, "loss": 0.8243, "step": 1021 }, { "epoch": 1.01, "grad_norm": 5.731918515166845, "learning_rate": 2.3882935632233805e-06, "loss": 0.7994, "step": 1022 }, { "epoch": 1.01, "grad_norm": 4.121022807470675, "learning_rate": 2.3843068488365167e-06, "loss": 0.7891, "step": 1023 }, { "epoch": 1.02, "grad_norm": 5.0069617325461495, "learning_rate": 2.3803204292697705e-06, "loss": 0.7859, "step": 1024 }, { "epoch": 1.02, "grad_norm": 4.066969700299868, "learning_rate": 2.3763343146817097e-06, "loss": 0.8024, "step": 1025 }, { "epoch": 1.02, "grad_norm": 4.076868215541214, "learning_rate": 2.372348515230124e-06, "loss": 0.781, "step": 1026 }, { "epoch": 1.02, "grad_norm": 5.44580327804881, "learning_rate": 2.3683630410720013e-06, "loss": 0.8014, "step": 1027 }, { "epoch": 1.02, "grad_norm": 14.576771764345674, "learning_rate": 2.364377902363499e-06, "loss": 0.7823, "step": 1028 }, { "epoch": 1.02, "grad_norm": 5.887292497093561, "learning_rate": 2.3603931092599216e-06, "loss": 0.9164, "step": 1029 }, { "epoch": 1.02, "grad_norm": 4.012162638866484, "learning_rate": 2.356408671915692e-06, "loss": 0.7961, "step": 1030 }, { "epoch": 1.02, "grad_norm": 3.5825184439978055, "learning_rate": 2.3524246004843262e-06, "loss": 0.7606, "step": 1031 }, { "epoch": 1.02, "grad_norm": 3.9637088817860744, "learning_rate": 2.3484409051184075e-06, "loss": 0.7834, "step": 1032 }, { "epoch": 1.02, "grad_norm": 4.9537549849251885, "learning_rate": 2.3444575959695615e-06, "loss": 0.8011, "step": 1033 }, { "epoch": 1.03, "grad_norm": 5.309949433262764, "learning_rate": 2.340474683188429e-06, "loss": 0.7852, "step": 1034 }, { "epoch": 1.03, "grad_norm": 17.986917941395983, "learning_rate": 2.3364921769246423e-06, "loss": 0.7669, "step": 1035 }, { "epoch": 1.03, "grad_norm": 3.716592584279815, "learning_rate": 2.3325100873267963e-06, "loss": 0.8255, "step": 1036 }, { "epoch": 1.03, "grad_norm": 8.241271251632352, "learning_rate": 2.3285284245424247e-06, "loss": 0.9107, "step": 1037 }, { "epoch": 1.03, "grad_norm": 7.8919471190913715, "learning_rate": 2.324547198717972e-06, "loss": 0.8093, "step": 1038 }, { "epoch": 1.03, "grad_norm": 5.0992986992131755, "learning_rate": 2.320566419998772e-06, "loss": 0.9247, "step": 1039 }, { "epoch": 1.03, "grad_norm": 5.047018545060358, "learning_rate": 2.316586098529017e-06, "loss": 0.7533, "step": 1040 }, { "epoch": 1.03, "grad_norm": 8.084822452580836, "learning_rate": 2.3126062444517336e-06, "loss": 0.9119, "step": 1041 }, { "epoch": 1.03, "grad_norm": 3.949333709906309, "learning_rate": 2.308626867908761e-06, "loss": 0.8266, "step": 1042 }, { "epoch": 1.03, "grad_norm": 18.354447357233287, "learning_rate": 2.3046479790407177e-06, "loss": 0.8305, "step": 1043 }, { "epoch": 1.04, "grad_norm": 4.594839026426293, "learning_rate": 2.3006695879869807e-06, "loss": 0.7534, "step": 1044 }, { "epoch": 1.04, "grad_norm": 3.680606770155669, "learning_rate": 2.296691704885659e-06, "loss": 0.7923, "step": 1045 }, { "epoch": 1.04, "grad_norm": 8.046065375407188, "learning_rate": 2.2927143398735664e-06, "loss": 0.7597, "step": 1046 }, { "epoch": 1.04, "grad_norm": 4.146496528855196, "learning_rate": 2.288737503086195e-06, "loss": 0.7685, "step": 1047 }, { "epoch": 1.04, "grad_norm": 4.55862230284598, "learning_rate": 2.284761204657696e-06, "loss": 0.7566, "step": 1048 }, { "epoch": 1.04, "grad_norm": 4.04730648082624, "learning_rate": 2.280785454720843e-06, "loss": 0.8026, "step": 1049 }, { "epoch": 1.04, "grad_norm": 4.489592584024217, "learning_rate": 2.2768102634070146e-06, "loss": 0.7582, "step": 1050 }, { "epoch": 1.04, "grad_norm": 8.921156817154868, "learning_rate": 2.2728356408461653e-06, "loss": 0.8798, "step": 1051 }, { "epoch": 1.04, "grad_norm": 5.180118036451976, "learning_rate": 2.2688615971668015e-06, "loss": 0.7906, "step": 1052 }, { "epoch": 1.04, "grad_norm": 101.04577612993059, "learning_rate": 2.26488814249595e-06, "loss": 0.8442, "step": 1053 }, { "epoch": 1.05, "grad_norm": 3.8214055724383433, "learning_rate": 2.2609152869591445e-06, "loss": 0.7478, "step": 1054 }, { "epoch": 1.05, "grad_norm": 3.91039758025345, "learning_rate": 2.2569430406803845e-06, "loss": 0.7679, "step": 1055 }, { "epoch": 1.05, "grad_norm": 4.319013551096071, "learning_rate": 2.252971413782121e-06, "loss": 0.7805, "step": 1056 }, { "epoch": 1.05, "grad_norm": 3.713120069572838, "learning_rate": 2.249000416385224e-06, "loss": 0.7822, "step": 1057 }, { "epoch": 1.05, "grad_norm": 8.266262461899506, "learning_rate": 2.2450300586089626e-06, "loss": 0.7635, "step": 1058 }, { "epoch": 1.05, "grad_norm": 8.110418184386617, "learning_rate": 2.2410603505709718e-06, "loss": 0.7842, "step": 1059 }, { "epoch": 1.05, "grad_norm": 3.7012875199461814, "learning_rate": 2.2370913023872357e-06, "loss": 0.8072, "step": 1060 }, { "epoch": 1.05, "grad_norm": 5.109464899932251, "learning_rate": 2.233122924172053e-06, "loss": 0.7954, "step": 1061 }, { "epoch": 1.05, "grad_norm": 3.6551235701500056, "learning_rate": 2.229155226038017e-06, "loss": 0.7844, "step": 1062 }, { "epoch": 1.05, "grad_norm": 3.552118816759354, "learning_rate": 2.2251882180959876e-06, "loss": 0.7587, "step": 1063 }, { "epoch": 1.06, "grad_norm": 3.4989984130132776, "learning_rate": 2.2212219104550664e-06, "loss": 0.7415, "step": 1064 }, { "epoch": 1.06, "grad_norm": 8.129927700947665, "learning_rate": 2.217256313222567e-06, "loss": 0.8136, "step": 1065 }, { "epoch": 1.06, "grad_norm": 4.03690382596103, "learning_rate": 2.2132914365039993e-06, "loss": 0.8152, "step": 1066 }, { "epoch": 1.06, "grad_norm": 4.344164682839045, "learning_rate": 2.209327290403031e-06, "loss": 0.8527, "step": 1067 }, { "epoch": 1.06, "grad_norm": 3.6614084580424757, "learning_rate": 2.2053638850214704e-06, "loss": 0.838, "step": 1068 }, { "epoch": 1.06, "grad_norm": 3.6154515357735537, "learning_rate": 2.201401230459239e-06, "loss": 0.7948, "step": 1069 }, { "epoch": 1.06, "grad_norm": 3.5759738118075366, "learning_rate": 2.1974393368143433e-06, "loss": 0.7659, "step": 1070 }, { "epoch": 1.06, "grad_norm": 3.7878379791745784, "learning_rate": 2.1934782141828505e-06, "loss": 0.8244, "step": 1071 }, { "epoch": 1.06, "grad_norm": 3.6308819407635715, "learning_rate": 2.189517872658867e-06, "loss": 0.7867, "step": 1072 }, { "epoch": 1.06, "grad_norm": 3.545699911216471, "learning_rate": 2.185558322334504e-06, "loss": 0.7414, "step": 1073 }, { "epoch": 1.07, "grad_norm": 3.4249349721328133, "learning_rate": 2.1815995732998586e-06, "loss": 0.7875, "step": 1074 }, { "epoch": 1.07, "grad_norm": 3.7445973808388877, "learning_rate": 2.1776416356429857e-06, "loss": 0.7967, "step": 1075 }, { "epoch": 1.07, "grad_norm": 3.806824840899025, "learning_rate": 2.173684519449872e-06, "loss": 0.7567, "step": 1076 }, { "epoch": 1.07, "grad_norm": 5.714674335901116, "learning_rate": 2.1697282348044114e-06, "loss": 0.8067, "step": 1077 }, { "epoch": 1.07, "grad_norm": 3.942109362960434, "learning_rate": 2.165772791788379e-06, "loss": 0.7451, "step": 1078 }, { "epoch": 1.07, "grad_norm": 4.1834882284478105, "learning_rate": 2.1618182004814054e-06, "loss": 0.774, "step": 1079 }, { "epoch": 1.07, "grad_norm": 6.738772076152307, "learning_rate": 2.15786447096095e-06, "loss": 0.7863, "step": 1080 }, { "epoch": 1.07, "grad_norm": 3.8040659737323472, "learning_rate": 2.1539116133022776e-06, "loss": 0.7929, "step": 1081 }, { "epoch": 1.07, "grad_norm": 4.2567073361175956, "learning_rate": 2.1499596375784284e-06, "loss": 0.7797, "step": 1082 }, { "epoch": 1.07, "grad_norm": 3.6508224715147737, "learning_rate": 2.146008553860197e-06, "loss": 0.8428, "step": 1083 }, { "epoch": 1.08, "grad_norm": 4.118725441028583, "learning_rate": 2.142058372216105e-06, "loss": 0.8169, "step": 1084 }, { "epoch": 1.08, "grad_norm": 3.8029807374968394, "learning_rate": 2.138109102712376e-06, "loss": 0.7441, "step": 1085 }, { "epoch": 1.08, "grad_norm": 5.1275783668978026, "learning_rate": 2.1341607554129073e-06, "loss": 0.8086, "step": 1086 }, { "epoch": 1.08, "grad_norm": 3.568601984552534, "learning_rate": 2.130213340379248e-06, "loss": 0.7918, "step": 1087 }, { "epoch": 1.08, "grad_norm": 4.184400688441942, "learning_rate": 2.12626686767057e-06, "loss": 0.8118, "step": 1088 }, { "epoch": 1.08, "grad_norm": 4.049362573544909, "learning_rate": 2.122321347343644e-06, "loss": 0.7631, "step": 1089 }, { "epoch": 1.08, "grad_norm": 4.051471615817484, "learning_rate": 2.1183767894528135e-06, "loss": 0.7722, "step": 1090 }, { "epoch": 1.08, "grad_norm": 3.949276080324605, "learning_rate": 2.1144332040499726e-06, "loss": 0.9389, "step": 1091 }, { "epoch": 1.08, "grad_norm": 4.628846053370376, "learning_rate": 2.1104906011845334e-06, "loss": 0.8214, "step": 1092 }, { "epoch": 1.08, "grad_norm": 4.400806048971509, "learning_rate": 2.1065489909034065e-06, "loss": 0.7807, "step": 1093 }, { "epoch": 1.09, "grad_norm": 3.547050134373076, "learning_rate": 2.1026083832509704e-06, "loss": 0.7613, "step": 1094 }, { "epoch": 1.09, "grad_norm": 4.645783433581276, "learning_rate": 2.0986687882690513e-06, "loss": 0.7958, "step": 1095 }, { "epoch": 1.09, "grad_norm": 6.3835868961525275, "learning_rate": 2.094730215996894e-06, "loss": 0.7776, "step": 1096 }, { "epoch": 1.09, "grad_norm": 4.399374827332895, "learning_rate": 2.090792676471137e-06, "loss": 0.7436, "step": 1097 }, { "epoch": 1.09, "grad_norm": 5.891592653655997, "learning_rate": 2.086856179725788e-06, "loss": 0.7627, "step": 1098 }, { "epoch": 1.09, "grad_norm": 3.6928161950981067, "learning_rate": 2.082920735792195e-06, "loss": 0.8058, "step": 1099 }, { "epoch": 1.09, "grad_norm": 3.4692201695028495, "learning_rate": 2.0789863546990256e-06, "loss": 0.7599, "step": 1100 }, { "epoch": 1.09, "grad_norm": 3.8841365172861, "learning_rate": 2.0750530464722374e-06, "loss": 0.9646, "step": 1101 }, { "epoch": 1.09, "grad_norm": 3.3833964281231395, "learning_rate": 2.0711208211350543e-06, "loss": 0.7489, "step": 1102 }, { "epoch": 1.09, "grad_norm": 3.894728811938954, "learning_rate": 2.0671896887079423e-06, "loss": 0.7591, "step": 1103 }, { "epoch": 1.1, "grad_norm": 4.304258118651057, "learning_rate": 2.0632596592085803e-06, "loss": 0.798, "step": 1104 }, { "epoch": 1.1, "grad_norm": 4.432978625782166, "learning_rate": 2.059330742651837e-06, "loss": 0.7924, "step": 1105 }, { "epoch": 1.1, "grad_norm": 3.848164092994974, "learning_rate": 2.055402949049746e-06, "loss": 0.8019, "step": 1106 }, { "epoch": 1.1, "grad_norm": 3.6604126323152966, "learning_rate": 2.0514762884114786e-06, "loss": 0.7608, "step": 1107 }, { "epoch": 1.1, "grad_norm": 3.950478491916118, "learning_rate": 2.047550770743318e-06, "loss": 0.788, "step": 1108 }, { "epoch": 1.1, "grad_norm": 4.831215312170034, "learning_rate": 2.0436264060486367e-06, "loss": 0.806, "step": 1109 }, { "epoch": 1.1, "grad_norm": 3.337136636008226, "learning_rate": 2.0397032043278686e-06, "loss": 0.823, "step": 1110 }, { "epoch": 1.1, "grad_norm": 4.3006636934514315, "learning_rate": 2.035781175578483e-06, "loss": 0.8051, "step": 1111 }, { "epoch": 1.1, "grad_norm": 3.547079984712057, "learning_rate": 2.0318603297949603e-06, "loss": 0.784, "step": 1112 }, { "epoch": 1.11, "grad_norm": 4.024179858161144, "learning_rate": 2.0279406769687667e-06, "loss": 0.7925, "step": 1113 }, { "epoch": 1.11, "grad_norm": 3.8987335356394506, "learning_rate": 2.024022227088329e-06, "loss": 0.7806, "step": 1114 }, { "epoch": 1.11, "grad_norm": 4.519470933018817, "learning_rate": 2.020104990139008e-06, "loss": 0.8417, "step": 1115 }, { "epoch": 1.11, "grad_norm": 3.688717858602124, "learning_rate": 2.016188976103074e-06, "loss": 0.7851, "step": 1116 }, { "epoch": 1.11, "grad_norm": 3.62789393638703, "learning_rate": 2.01227419495968e-06, "loss": 0.7773, "step": 1117 }, { "epoch": 1.11, "grad_norm": 4.031458993916617, "learning_rate": 2.008360656684837e-06, "loss": 0.7937, "step": 1118 }, { "epoch": 1.11, "grad_norm": 3.6363334134466254, "learning_rate": 2.004448371251391e-06, "loss": 0.7478, "step": 1119 }, { "epoch": 1.11, "grad_norm": 3.666719552295652, "learning_rate": 2.0005373486289932e-06, "loss": 0.8022, "step": 1120 }, { "epoch": 1.11, "grad_norm": 4.282769746933799, "learning_rate": 1.996627598784077e-06, "loss": 0.8285, "step": 1121 }, { "epoch": 1.11, "grad_norm": 3.6035899935531144, "learning_rate": 1.9927191316798333e-06, "loss": 0.7804, "step": 1122 }, { "epoch": 1.12, "grad_norm": 3.7437751327926527, "learning_rate": 1.988811957276185e-06, "loss": 0.8747, "step": 1123 }, { "epoch": 1.12, "grad_norm": 3.496198696593862, "learning_rate": 1.984906085529758e-06, "loss": 0.7914, "step": 1124 }, { "epoch": 1.12, "grad_norm": 3.7947369193970037, "learning_rate": 1.9810015263938623e-06, "loss": 0.7313, "step": 1125 }, { "epoch": 1.12, "grad_norm": 3.827284933028767, "learning_rate": 1.977098289818459e-06, "loss": 0.7859, "step": 1126 }, { "epoch": 1.12, "grad_norm": 4.028706232972695, "learning_rate": 1.973196385750141e-06, "loss": 0.8222, "step": 1127 }, { "epoch": 1.12, "grad_norm": 3.989445099470434, "learning_rate": 1.969295824132107e-06, "loss": 0.789, "step": 1128 }, { "epoch": 1.12, "grad_norm": 3.6596114594139397, "learning_rate": 1.9653966149041323e-06, "loss": 0.771, "step": 1129 }, { "epoch": 1.12, "grad_norm": 5.744658687029394, "learning_rate": 1.961498768002547e-06, "loss": 0.8807, "step": 1130 }, { "epoch": 1.12, "grad_norm": 3.7994060988032325, "learning_rate": 1.95760229336021e-06, "loss": 0.7813, "step": 1131 }, { "epoch": 1.12, "grad_norm": 4.733387856083046, "learning_rate": 1.9537072009064814e-06, "loss": 0.8054, "step": 1132 }, { "epoch": 1.13, "grad_norm": 5.160732849382161, "learning_rate": 1.949813500567201e-06, "loss": 0.7826, "step": 1133 }, { "epoch": 1.13, "grad_norm": 3.6698152412533482, "learning_rate": 1.9459212022646604e-06, "loss": 0.7928, "step": 1134 }, { "epoch": 1.13, "grad_norm": 3.8485882545383125, "learning_rate": 1.9420303159175795e-06, "loss": 0.7628, "step": 1135 }, { "epoch": 1.13, "grad_norm": 3.4928222960830757, "learning_rate": 1.938140851441078e-06, "loss": 0.7684, "step": 1136 }, { "epoch": 1.13, "grad_norm": 4.922970349794866, "learning_rate": 1.934252818746655e-06, "loss": 0.8043, "step": 1137 }, { "epoch": 1.13, "grad_norm": 4.919144399500264, "learning_rate": 1.930366227742157e-06, "loss": 0.7604, "step": 1138 }, { "epoch": 1.13, "grad_norm": 3.8550949628998876, "learning_rate": 1.926481088331759e-06, "loss": 0.8024, "step": 1139 }, { "epoch": 1.13, "grad_norm": 3.6371608301334604, "learning_rate": 1.922597410415939e-06, "loss": 0.8159, "step": 1140 }, { "epoch": 1.13, "grad_norm": 3.706285194656798, "learning_rate": 1.918715203891448e-06, "loss": 0.8673, "step": 1141 }, { "epoch": 1.13, "grad_norm": 3.839321564709285, "learning_rate": 1.914834478651288e-06, "loss": 0.7816, "step": 1142 }, { "epoch": 1.14, "grad_norm": 4.396549119678207, "learning_rate": 1.9109552445846855e-06, "loss": 0.8096, "step": 1143 }, { "epoch": 1.14, "grad_norm": 3.5716576529351816, "learning_rate": 1.907077511577068e-06, "loss": 0.7904, "step": 1144 }, { "epoch": 1.14, "grad_norm": 3.7724496927798765, "learning_rate": 1.9032012895100383e-06, "loss": 0.7897, "step": 1145 }, { "epoch": 1.14, "grad_norm": 28.743143211969254, "learning_rate": 1.8993265882613483e-06, "loss": 0.7505, "step": 1146 }, { "epoch": 1.14, "grad_norm": 3.9949722818830797, "learning_rate": 1.8954534177048744e-06, "loss": 0.8136, "step": 1147 }, { "epoch": 1.14, "grad_norm": 3.9077454378371184, "learning_rate": 1.8915817877105928e-06, "loss": 0.7981, "step": 1148 }, { "epoch": 1.14, "grad_norm": 4.051523604760233, "learning_rate": 1.8877117081445524e-06, "loss": 0.8098, "step": 1149 }, { "epoch": 1.14, "grad_norm": 4.231365448478539, "learning_rate": 1.8838431888688528e-06, "loss": 0.7534, "step": 1150 }, { "epoch": 1.14, "grad_norm": 3.9151667425945975, "learning_rate": 1.879976239741616e-06, "loss": 0.8768, "step": 1151 }, { "epoch": 1.14, "grad_norm": 3.6891243508446996, "learning_rate": 1.8761108706169656e-06, "loss": 0.7969, "step": 1152 }, { "epoch": 1.15, "grad_norm": 3.4500592454107535, "learning_rate": 1.8722470913449962e-06, "loss": 0.8005, "step": 1153 }, { "epoch": 1.15, "grad_norm": 4.35570443225781, "learning_rate": 1.8683849117717518e-06, "loss": 0.7448, "step": 1154 }, { "epoch": 1.15, "grad_norm": 3.9957135368473056, "learning_rate": 1.8645243417391995e-06, "loss": 0.7546, "step": 1155 }, { "epoch": 1.15, "grad_norm": 4.692192911794636, "learning_rate": 1.8606653910852058e-06, "loss": 0.7858, "step": 1156 }, { "epoch": 1.15, "grad_norm": 3.4799460404628606, "learning_rate": 1.8568080696435092e-06, "loss": 0.8008, "step": 1157 }, { "epoch": 1.15, "grad_norm": 5.22653864055669, "learning_rate": 1.852952387243698e-06, "loss": 0.7804, "step": 1158 }, { "epoch": 1.15, "grad_norm": 4.099176404298372, "learning_rate": 1.8490983537111831e-06, "loss": 0.7687, "step": 1159 }, { "epoch": 1.15, "grad_norm": 4.641865328492008, "learning_rate": 1.845245978867174e-06, "loss": 0.8224, "step": 1160 }, { "epoch": 1.15, "grad_norm": 3.555897283803014, "learning_rate": 1.8413952725286511e-06, "loss": 0.8879, "step": 1161 }, { "epoch": 1.15, "grad_norm": 6.7756038974592006, "learning_rate": 1.8375462445083464e-06, "loss": 0.8469, "step": 1162 }, { "epoch": 1.16, "grad_norm": 3.2826785527610363, "learning_rate": 1.833698904614713e-06, "loss": 0.9068, "step": 1163 }, { "epoch": 1.16, "grad_norm": 4.137361445765123, "learning_rate": 1.8298532626519008e-06, "loss": 0.8145, "step": 1164 }, { "epoch": 1.16, "grad_norm": 3.5150365351448816, "learning_rate": 1.8260093284197371e-06, "loss": 0.8143, "step": 1165 }, { "epoch": 1.16, "grad_norm": 3.277784244458752, "learning_rate": 1.8221671117136953e-06, "loss": 0.8426, "step": 1166 }, { "epoch": 1.16, "grad_norm": 4.219472826853829, "learning_rate": 1.8183266223248691e-06, "loss": 0.8207, "step": 1167 }, { "epoch": 1.16, "grad_norm": 3.2492653923789256, "learning_rate": 1.814487870039955e-06, "loss": 0.753, "step": 1168 }, { "epoch": 1.16, "grad_norm": 3.534420327263411, "learning_rate": 1.8106508646412202e-06, "loss": 0.8326, "step": 1169 }, { "epoch": 1.16, "grad_norm": 3.653314259165623, "learning_rate": 1.80681561590648e-06, "loss": 0.8206, "step": 1170 }, { "epoch": 1.16, "grad_norm": 3.965314818874547, "learning_rate": 1.8029821336090769e-06, "loss": 0.8001, "step": 1171 }, { "epoch": 1.16, "grad_norm": 3.327615983509177, "learning_rate": 1.7991504275178473e-06, "loss": 0.794, "step": 1172 }, { "epoch": 1.17, "grad_norm": 4.621765492023544, "learning_rate": 1.7953205073971036e-06, "loss": 0.8176, "step": 1173 }, { "epoch": 1.17, "grad_norm": 4.448850768126181, "learning_rate": 1.7914923830066074e-06, "loss": 0.7788, "step": 1174 }, { "epoch": 1.17, "grad_norm": 3.682663534875354, "learning_rate": 1.7876660641015436e-06, "loss": 0.8039, "step": 1175 }, { "epoch": 1.17, "grad_norm": 3.564524917829612, "learning_rate": 1.7838415604324943e-06, "loss": 0.7707, "step": 1176 }, { "epoch": 1.17, "grad_norm": 3.858831903763962, "learning_rate": 1.7800188817454209e-06, "loss": 0.8035, "step": 1177 }, { "epoch": 1.17, "grad_norm": 3.876072938853657, "learning_rate": 1.7761980377816287e-06, "loss": 0.796, "step": 1178 }, { "epoch": 1.17, "grad_norm": 4.242333258312018, "learning_rate": 1.7723790382777512e-06, "loss": 0.7838, "step": 1179 }, { "epoch": 1.17, "grad_norm": 4.107140742026446, "learning_rate": 1.7685618929657193e-06, "loss": 0.8289, "step": 1180 }, { "epoch": 1.17, "grad_norm": 3.2444132513512387, "learning_rate": 1.7647466115727419e-06, "loss": 0.7764, "step": 1181 }, { "epoch": 1.17, "grad_norm": 3.3757668997700923, "learning_rate": 1.7609332038212728e-06, "loss": 0.761, "step": 1182 }, { "epoch": 1.18, "grad_norm": 3.8104664088613824, "learning_rate": 1.7571216794289985e-06, "loss": 0.8179, "step": 1183 }, { "epoch": 1.18, "grad_norm": 5.404708705654263, "learning_rate": 1.7533120481088e-06, "loss": 0.8138, "step": 1184 }, { "epoch": 1.18, "grad_norm": 5.5634648748141275, "learning_rate": 1.7495043195687367e-06, "loss": 0.769, "step": 1185 }, { "epoch": 1.18, "grad_norm": 3.823406082840297, "learning_rate": 1.7456985035120194e-06, "loss": 0.771, "step": 1186 }, { "epoch": 1.18, "grad_norm": 3.396357328716178, "learning_rate": 1.7418946096369851e-06, "loss": 0.7901, "step": 1187 }, { "epoch": 1.18, "grad_norm": 3.6149120166515147, "learning_rate": 1.7380926476370702e-06, "loss": 0.818, "step": 1188 }, { "epoch": 1.18, "grad_norm": 4.016406555407024, "learning_rate": 1.734292627200793e-06, "loss": 0.7774, "step": 1189 }, { "epoch": 1.18, "grad_norm": 4.069937677697211, "learning_rate": 1.7304945580117195e-06, "loss": 0.8024, "step": 1190 }, { "epoch": 1.18, "grad_norm": 3.7571930508926266, "learning_rate": 1.7266984497484456e-06, "loss": 0.828, "step": 1191 }, { "epoch": 1.18, "grad_norm": 5.011758826537584, "learning_rate": 1.7229043120845705e-06, "loss": 0.7971, "step": 1192 }, { "epoch": 1.19, "grad_norm": 3.498245416062928, "learning_rate": 1.7191121546886696e-06, "loss": 0.8066, "step": 1193 }, { "epoch": 1.19, "grad_norm": 23.613550503117423, "learning_rate": 1.7153219872242727e-06, "loss": 0.7898, "step": 1194 }, { "epoch": 1.19, "grad_norm": 3.6105513963197753, "learning_rate": 1.711533819349842e-06, "loss": 0.7905, "step": 1195 }, { "epoch": 1.19, "grad_norm": 4.394808857561618, "learning_rate": 1.70774766071874e-06, "loss": 0.9266, "step": 1196 }, { "epoch": 1.19, "grad_norm": 3.645685780003328, "learning_rate": 1.7039635209792105e-06, "loss": 0.8306, "step": 1197 }, { "epoch": 1.19, "grad_norm": 3.1769035919480237, "learning_rate": 1.700181409774353e-06, "loss": 0.7573, "step": 1198 }, { "epoch": 1.19, "grad_norm": 4.007016128805905, "learning_rate": 1.6964013367420967e-06, "loss": 0.8264, "step": 1199 }, { "epoch": 1.19, "grad_norm": 3.652857290992286, "learning_rate": 1.692623311515178e-06, "loss": 0.7825, "step": 1200 }, { "epoch": 1.19, "grad_norm": 4.061308195392651, "learning_rate": 1.6888473437211134e-06, "loss": 0.7616, "step": 1201 }, { "epoch": 1.19, "grad_norm": 3.593093915248048, "learning_rate": 1.6850734429821792e-06, "loss": 0.7847, "step": 1202 }, { "epoch": 1.2, "grad_norm": 3.3270693155942554, "learning_rate": 1.6813016189153814e-06, "loss": 0.791, "step": 1203 }, { "epoch": 1.2, "grad_norm": 6.471153365626217, "learning_rate": 1.6775318811324365e-06, "loss": 0.832, "step": 1204 }, { "epoch": 1.2, "grad_norm": 5.15102543406431, "learning_rate": 1.6737642392397414e-06, "loss": 0.7926, "step": 1205 }, { "epoch": 1.2, "grad_norm": 4.830266878718885, "learning_rate": 1.6699987028383547e-06, "loss": 0.923, "step": 1206 }, { "epoch": 1.2, "grad_norm": 4.373947779555348, "learning_rate": 1.666235281523968e-06, "loss": 0.7593, "step": 1207 }, { "epoch": 1.2, "grad_norm": 3.5334531773938984, "learning_rate": 1.6624739848868855e-06, "loss": 0.7927, "step": 1208 }, { "epoch": 1.2, "grad_norm": 3.5270070172868406, "learning_rate": 1.6587148225119936e-06, "loss": 0.817, "step": 1209 }, { "epoch": 1.2, "grad_norm": 3.545468025507147, "learning_rate": 1.6549578039787436e-06, "loss": 0.8167, "step": 1210 }, { "epoch": 1.2, "grad_norm": 3.6200521347737715, "learning_rate": 1.65120293886112e-06, "loss": 0.784, "step": 1211 }, { "epoch": 1.21, "grad_norm": 3.52361659150107, "learning_rate": 1.6474502367276225e-06, "loss": 0.8069, "step": 1212 }, { "epoch": 1.21, "grad_norm": 3.479322784348679, "learning_rate": 1.643699707141237e-06, "loss": 0.7436, "step": 1213 }, { "epoch": 1.21, "grad_norm": 3.5315375632966934, "learning_rate": 1.6399513596594158e-06, "loss": 0.8233, "step": 1214 }, { "epoch": 1.21, "grad_norm": 3.611945989555201, "learning_rate": 1.6362052038340475e-06, "loss": 0.7784, "step": 1215 }, { "epoch": 1.21, "grad_norm": 3.5348710478024916, "learning_rate": 1.6324612492114378e-06, "loss": 0.7589, "step": 1216 }, { "epoch": 1.21, "grad_norm": 3.6025562281067125, "learning_rate": 1.6287195053322818e-06, "loss": 0.7676, "step": 1217 }, { "epoch": 1.21, "grad_norm": 3.373468886784458, "learning_rate": 1.6249799817316415e-06, "loss": 0.8341, "step": 1218 }, { "epoch": 1.21, "grad_norm": 3.307261302532676, "learning_rate": 1.6212426879389205e-06, "loss": 0.9442, "step": 1219 }, { "epoch": 1.21, "grad_norm": 3.330972570996881, "learning_rate": 1.6175076334778422e-06, "loss": 0.8083, "step": 1220 }, { "epoch": 1.21, "grad_norm": 3.36194389037731, "learning_rate": 1.6137748278664215e-06, "loss": 0.863, "step": 1221 }, { "epoch": 1.22, "grad_norm": 3.537762295639098, "learning_rate": 1.6100442806169423e-06, "loss": 0.8168, "step": 1222 }, { "epoch": 1.22, "grad_norm": 3.907737848792073, "learning_rate": 1.6063160012359348e-06, "loss": 0.7655, "step": 1223 }, { "epoch": 1.22, "grad_norm": 3.7679768612021305, "learning_rate": 1.602589999224149e-06, "loss": 0.7963, "step": 1224 }, { "epoch": 1.22, "grad_norm": 3.6331145077725147, "learning_rate": 1.5988662840765323e-06, "loss": 0.7564, "step": 1225 }, { "epoch": 1.22, "grad_norm": 3.049535918562501, "learning_rate": 1.595144865282205e-06, "loss": 0.7651, "step": 1226 }, { "epoch": 1.22, "grad_norm": 3.5895644396964084, "learning_rate": 1.5914257523244347e-06, "loss": 0.8152, "step": 1227 }, { "epoch": 1.22, "grad_norm": 3.760718365250507, "learning_rate": 1.5877089546806123e-06, "loss": 0.7758, "step": 1228 }, { "epoch": 1.22, "grad_norm": 4.270761772300484, "learning_rate": 1.5839944818222297e-06, "loss": 0.7891, "step": 1229 }, { "epoch": 1.22, "grad_norm": 4.106198014773157, "learning_rate": 1.5802823432148545e-06, "loss": 0.7823, "step": 1230 }, { "epoch": 1.22, "grad_norm": 3.41953830994807, "learning_rate": 1.5765725483181056e-06, "loss": 0.7674, "step": 1231 }, { "epoch": 1.23, "grad_norm": 3.3602845525058807, "learning_rate": 1.5728651065856297e-06, "loss": 0.7917, "step": 1232 }, { "epoch": 1.23, "grad_norm": 3.305442646731639, "learning_rate": 1.5691600274650782e-06, "loss": 0.793, "step": 1233 }, { "epoch": 1.23, "grad_norm": 4.074710797103209, "learning_rate": 1.5654573203980782e-06, "loss": 0.7733, "step": 1234 }, { "epoch": 1.23, "grad_norm": 3.1010678326424856, "learning_rate": 1.561756994820216e-06, "loss": 0.8473, "step": 1235 }, { "epoch": 1.23, "grad_norm": 3.2144306485082335, "learning_rate": 1.5580590601610073e-06, "loss": 0.7797, "step": 1236 }, { "epoch": 1.23, "grad_norm": 3.641816041253921, "learning_rate": 1.5543635258438746e-06, "loss": 0.7841, "step": 1237 }, { "epoch": 1.23, "grad_norm": 4.065692721638411, "learning_rate": 1.5506704012861257e-06, "loss": 0.7909, "step": 1238 }, { "epoch": 1.23, "grad_norm": 3.985065806446307, "learning_rate": 1.5469796958989261e-06, "loss": 0.7463, "step": 1239 }, { "epoch": 1.23, "grad_norm": 3.31477046748139, "learning_rate": 1.5432914190872757e-06, "loss": 0.8051, "step": 1240 }, { "epoch": 1.23, "grad_norm": 3.4858190924989554, "learning_rate": 1.5396055802499875e-06, "loss": 0.802, "step": 1241 }, { "epoch": 1.24, "grad_norm": 3.9693950061804038, "learning_rate": 1.5359221887796616e-06, "loss": 0.7777, "step": 1242 }, { "epoch": 1.24, "grad_norm": 3.5080596339343075, "learning_rate": 1.5322412540626593e-06, "loss": 0.8248, "step": 1243 }, { "epoch": 1.24, "grad_norm": 4.213235299969382, "learning_rate": 1.5285627854790838e-06, "loss": 0.9048, "step": 1244 }, { "epoch": 1.24, "grad_norm": 3.2865978532535016, "learning_rate": 1.5248867924027535e-06, "loss": 0.7639, "step": 1245 }, { "epoch": 1.24, "grad_norm": 3.770822782926791, "learning_rate": 1.5212132842011778e-06, "loss": 0.8124, "step": 1246 }, { "epoch": 1.24, "grad_norm": 3.305618667897665, "learning_rate": 1.5175422702355343e-06, "loss": 0.7763, "step": 1247 }, { "epoch": 1.24, "grad_norm": 3.6688542366977526, "learning_rate": 1.5138737598606448e-06, "loss": 0.7586, "step": 1248 }, { "epoch": 1.24, "grad_norm": 3.0345800112988806, "learning_rate": 1.5102077624249498e-06, "loss": 0.807, "step": 1249 }, { "epoch": 1.24, "grad_norm": 4.453039225237123, "learning_rate": 1.5065442872704872e-06, "loss": 0.7698, "step": 1250 }, { "epoch": 1.24, "grad_norm": 3.586136439566641, "learning_rate": 1.5028833437328682e-06, "loss": 0.7624, "step": 1251 }, { "epoch": 1.25, "grad_norm": 3.597290989179828, "learning_rate": 1.4992249411412514e-06, "loss": 0.7304, "step": 1252 }, { "epoch": 1.25, "grad_norm": 3.4157859025400903, "learning_rate": 1.4955690888183205e-06, "loss": 0.8128, "step": 1253 }, { "epoch": 1.25, "grad_norm": 3.647807337159193, "learning_rate": 1.4919157960802619e-06, "loss": 0.785, "step": 1254 }, { "epoch": 1.25, "grad_norm": 3.2106985341930145, "learning_rate": 1.4882650722367365e-06, "loss": 0.7943, "step": 1255 }, { "epoch": 1.25, "grad_norm": 3.7886887206250917, "learning_rate": 1.4846169265908603e-06, "loss": 0.7641, "step": 1256 }, { "epoch": 1.25, "grad_norm": 3.549369002532161, "learning_rate": 1.4809713684391809e-06, "loss": 0.804, "step": 1257 }, { "epoch": 1.25, "grad_norm": 3.8914453648280705, "learning_rate": 1.4773284070716504e-06, "loss": 0.7604, "step": 1258 }, { "epoch": 1.25, "grad_norm": 3.2370248369000927, "learning_rate": 1.473688051771604e-06, "loss": 0.794, "step": 1259 }, { "epoch": 1.25, "grad_norm": 3.2393232540988337, "learning_rate": 1.4700503118157361e-06, "loss": 0.7827, "step": 1260 }, { "epoch": 1.25, "grad_norm": 3.793022975623574, "learning_rate": 1.4664151964740754e-06, "loss": 0.8919, "step": 1261 }, { "epoch": 1.26, "grad_norm": 3.844591667127874, "learning_rate": 1.4627827150099627e-06, "loss": 0.7898, "step": 1262 }, { "epoch": 1.26, "grad_norm": 3.2138511202525555, "learning_rate": 1.4591528766800284e-06, "loss": 0.7659, "step": 1263 }, { "epoch": 1.26, "grad_norm": 3.7383275151744844, "learning_rate": 1.4555256907341668e-06, "loss": 0.9551, "step": 1264 }, { "epoch": 1.26, "grad_norm": 3.567091330313542, "learning_rate": 1.451901166415512e-06, "loss": 0.9573, "step": 1265 }, { "epoch": 1.26, "grad_norm": 3.654567841866817, "learning_rate": 1.4482793129604148e-06, "loss": 0.7528, "step": 1266 }, { "epoch": 1.26, "grad_norm": 3.383170340677262, "learning_rate": 1.4446601395984233e-06, "loss": 0.806, "step": 1267 }, { "epoch": 1.26, "grad_norm": 4.0574559949294215, "learning_rate": 1.4410436555522523e-06, "loss": 0.757, "step": 1268 }, { "epoch": 1.26, "grad_norm": 4.27060892250174, "learning_rate": 1.4374298700377668e-06, "loss": 0.8036, "step": 1269 }, { "epoch": 1.26, "grad_norm": 3.7563197256269567, "learning_rate": 1.4338187922639506e-06, "loss": 0.7864, "step": 1270 }, { "epoch": 1.26, "grad_norm": 3.8269488706822448, "learning_rate": 1.4302104314328935e-06, "loss": 0.802, "step": 1271 }, { "epoch": 1.27, "grad_norm": 3.594486489007269, "learning_rate": 1.426604796739755e-06, "loss": 0.7885, "step": 1272 }, { "epoch": 1.27, "grad_norm": 4.12307721886989, "learning_rate": 1.4230018973727535e-06, "loss": 0.8013, "step": 1273 }, { "epoch": 1.27, "grad_norm": 3.672851694617319, "learning_rate": 1.4194017425131324e-06, "loss": 0.7832, "step": 1274 }, { "epoch": 1.27, "grad_norm": 3.702233972675767, "learning_rate": 1.4158043413351456e-06, "loss": 0.7695, "step": 1275 }, { "epoch": 1.27, "grad_norm": 3.805706647243348, "learning_rate": 1.412209703006025e-06, "loss": 0.8052, "step": 1276 }, { "epoch": 1.27, "grad_norm": 5.564313055909103, "learning_rate": 1.4086178366859672e-06, "loss": 0.8886, "step": 1277 }, { "epoch": 1.27, "grad_norm": 3.807040052388912, "learning_rate": 1.405028751528099e-06, "loss": 0.7873, "step": 1278 }, { "epoch": 1.27, "grad_norm": 3.8439968937597944, "learning_rate": 1.4014424566784663e-06, "loss": 0.7591, "step": 1279 }, { "epoch": 1.27, "grad_norm": 6.524172187800951, "learning_rate": 1.397858961275999e-06, "loss": 0.7604, "step": 1280 }, { "epoch": 1.27, "grad_norm": 3.3058060282969324, "learning_rate": 1.3942782744524974e-06, "loss": 0.77, "step": 1281 }, { "epoch": 1.28, "grad_norm": 3.9203029629900117, "learning_rate": 1.3907004053326006e-06, "loss": 0.9166, "step": 1282 }, { "epoch": 1.28, "grad_norm": 3.1507492792221385, "learning_rate": 1.3871253630337723e-06, "loss": 0.751, "step": 1283 }, { "epoch": 1.28, "grad_norm": 3.7684674100065934, "learning_rate": 1.3835531566662675e-06, "loss": 0.7583, "step": 1284 }, { "epoch": 1.28, "grad_norm": 3.977510260077561, "learning_rate": 1.3799837953331191e-06, "loss": 0.7506, "step": 1285 }, { "epoch": 1.28, "grad_norm": 3.909640872992809, "learning_rate": 1.3764172881301063e-06, "loss": 0.8081, "step": 1286 }, { "epoch": 1.28, "grad_norm": 4.011995506453752, "learning_rate": 1.3728536441457358e-06, "loss": 0.8342, "step": 1287 }, { "epoch": 1.28, "grad_norm": 3.2929553881418236, "learning_rate": 1.3692928724612204e-06, "loss": 0.7711, "step": 1288 }, { "epoch": 1.28, "grad_norm": 3.4483797827863367, "learning_rate": 1.3657349821504519e-06, "loss": 0.8312, "step": 1289 }, { "epoch": 1.28, "grad_norm": 3.2866802492644207, "learning_rate": 1.362179982279979e-06, "loss": 0.7735, "step": 1290 }, { "epoch": 1.28, "grad_norm": 3.403254635220054, "learning_rate": 1.3586278819089837e-06, "loss": 0.7948, "step": 1291 }, { "epoch": 1.29, "grad_norm": 3.608975097980645, "learning_rate": 1.3550786900892634e-06, "loss": 0.7768, "step": 1292 }, { "epoch": 1.29, "grad_norm": 3.8812991584631367, "learning_rate": 1.3515324158651983e-06, "loss": 0.7959, "step": 1293 }, { "epoch": 1.29, "grad_norm": 3.6869207535997357, "learning_rate": 1.347989068273738e-06, "loss": 0.7968, "step": 1294 }, { "epoch": 1.29, "grad_norm": 3.5018285662542072, "learning_rate": 1.3444486563443724e-06, "loss": 0.8048, "step": 1295 }, { "epoch": 1.29, "grad_norm": 3.4889061289227548, "learning_rate": 1.3409111890991105e-06, "loss": 0.7991, "step": 1296 }, { "epoch": 1.29, "grad_norm": 3.6128227326659714, "learning_rate": 1.3373766755524564e-06, "loss": 0.7781, "step": 1297 }, { "epoch": 1.29, "grad_norm": 3.73862026332466, "learning_rate": 1.3338451247113898e-06, "loss": 0.7818, "step": 1298 }, { "epoch": 1.29, "grad_norm": 3.415255753449033, "learning_rate": 1.330316545575338e-06, "loss": 0.7859, "step": 1299 }, { "epoch": 1.29, "grad_norm": 3.386406043748882, "learning_rate": 1.3267909471361574e-06, "loss": 0.8001, "step": 1300 }, { "epoch": 1.3, "grad_norm": 3.7307539885929213, "learning_rate": 1.3232683383781081e-06, "loss": 0.7673, "step": 1301 }, { "epoch": 1.3, "grad_norm": 3.2598347784157977, "learning_rate": 1.3197487282778315e-06, "loss": 0.8339, "step": 1302 }, { "epoch": 1.3, "grad_norm": 4.1636431088716535, "learning_rate": 1.3162321258043261e-06, "loss": 0.8587, "step": 1303 }, { "epoch": 1.3, "grad_norm": 3.2453445556040785, "learning_rate": 1.3127185399189297e-06, "loss": 0.7565, "step": 1304 }, { "epoch": 1.3, "grad_norm": 3.633277497076972, "learning_rate": 1.3092079795752888e-06, "loss": 0.762, "step": 1305 }, { "epoch": 1.3, "grad_norm": 3.6811807682699933, "learning_rate": 1.3057004537193424e-06, "loss": 0.9463, "step": 1306 }, { "epoch": 1.3, "grad_norm": 3.3766249993372663, "learning_rate": 1.3021959712892979e-06, "loss": 0.7361, "step": 1307 }, { "epoch": 1.3, "grad_norm": 3.6176610628726946, "learning_rate": 1.2986945412156037e-06, "loss": 0.8059, "step": 1308 }, { "epoch": 1.3, "grad_norm": 3.4634860169189774, "learning_rate": 1.2951961724209317e-06, "loss": 0.8148, "step": 1309 }, { "epoch": 1.3, "grad_norm": 3.4439243994879054, "learning_rate": 1.2917008738201536e-06, "loss": 0.8151, "step": 1310 }, { "epoch": 1.31, "grad_norm": 3.5357406044426103, "learning_rate": 1.2882086543203153e-06, "loss": 0.7389, "step": 1311 }, { "epoch": 1.31, "grad_norm": 3.798914676426891, "learning_rate": 1.284719522820618e-06, "loss": 0.9176, "step": 1312 }, { "epoch": 1.31, "grad_norm": 4.189307507269919, "learning_rate": 1.2812334882123944e-06, "loss": 0.9112, "step": 1313 }, { "epoch": 1.31, "grad_norm": 3.18849024148244, "learning_rate": 1.2777505593790834e-06, "loss": 0.7606, "step": 1314 }, { "epoch": 1.31, "grad_norm": 3.5502388567497905, "learning_rate": 1.274270745196209e-06, "loss": 0.7836, "step": 1315 }, { "epoch": 1.31, "grad_norm": 3.6992555926931345, "learning_rate": 1.2707940545313613e-06, "loss": 0.7893, "step": 1316 }, { "epoch": 1.31, "grad_norm": 3.4913023675924455, "learning_rate": 1.2673204962441671e-06, "loss": 0.8029, "step": 1317 }, { "epoch": 1.31, "grad_norm": 3.6280940469830885, "learning_rate": 1.263850079186274e-06, "loss": 0.7657, "step": 1318 }, { "epoch": 1.31, "grad_norm": 3.7332162233713224, "learning_rate": 1.2603828122013244e-06, "loss": 0.7918, "step": 1319 }, { "epoch": 1.31, "grad_norm": 3.1444371049395308, "learning_rate": 1.2569187041249315e-06, "loss": 0.7736, "step": 1320 }, { "epoch": 1.32, "grad_norm": 3.850124974139382, "learning_rate": 1.253457763784659e-06, "loss": 0.779, "step": 1321 }, { "epoch": 1.32, "grad_norm": 3.233732491110932, "learning_rate": 1.2500000000000007e-06, "loss": 0.7591, "step": 1322 }, { "epoch": 1.32, "grad_norm": 3.5943619900327963, "learning_rate": 1.2465454215823531e-06, "loss": 0.7574, "step": 1323 }, { "epoch": 1.32, "grad_norm": 3.538133758535698, "learning_rate": 1.2430940373349944e-06, "loss": 0.8815, "step": 1324 }, { "epoch": 1.32, "grad_norm": 3.5201966934916937, "learning_rate": 1.2396458560530694e-06, "loss": 0.7726, "step": 1325 }, { "epoch": 1.32, "grad_norm": 3.5701798733374503, "learning_rate": 1.2362008865235536e-06, "loss": 0.8251, "step": 1326 }, { "epoch": 1.32, "grad_norm": 3.7608599293366645, "learning_rate": 1.2327591375252404e-06, "loss": 0.7693, "step": 1327 }, { "epoch": 1.32, "grad_norm": 3.654749013039256, "learning_rate": 1.2293206178287184e-06, "loss": 0.7894, "step": 1328 }, { "epoch": 1.32, "grad_norm": 3.5974097318261102, "learning_rate": 1.225885336196345e-06, "loss": 0.7686, "step": 1329 }, { "epoch": 1.32, "grad_norm": 4.450363868169627, "learning_rate": 1.2224533013822237e-06, "loss": 0.7796, "step": 1330 }, { "epoch": 1.33, "grad_norm": 3.5141448045992942, "learning_rate": 1.219024522132191e-06, "loss": 0.8055, "step": 1331 }, { "epoch": 1.33, "grad_norm": 4.4031679543570394, "learning_rate": 1.2155990071837817e-06, "loss": 0.7631, "step": 1332 }, { "epoch": 1.33, "grad_norm": 3.6273266299444127, "learning_rate": 1.2121767652662122e-06, "loss": 0.759, "step": 1333 }, { "epoch": 1.33, "grad_norm": 11.043021044780021, "learning_rate": 1.2087578051003619e-06, "loss": 0.762, "step": 1334 }, { "epoch": 1.33, "grad_norm": 3.1130989971877647, "learning_rate": 1.2053421353987437e-06, "loss": 0.7879, "step": 1335 }, { "epoch": 1.33, "grad_norm": 3.8680168037093527, "learning_rate": 1.2019297648654856e-06, "loss": 0.8136, "step": 1336 }, { "epoch": 1.33, "grad_norm": 3.498293369468184, "learning_rate": 1.198520702196313e-06, "loss": 0.7961, "step": 1337 }, { "epoch": 1.33, "grad_norm": 3.4074998387077637, "learning_rate": 1.1951149560785166e-06, "loss": 0.7831, "step": 1338 }, { "epoch": 1.33, "grad_norm": 3.068189411598379, "learning_rate": 1.191712535190937e-06, "loss": 0.7468, "step": 1339 }, { "epoch": 1.33, "grad_norm": 3.3882773456818414, "learning_rate": 1.1883134482039429e-06, "loss": 0.7839, "step": 1340 }, { "epoch": 1.34, "grad_norm": 3.805597945416793, "learning_rate": 1.1849177037794052e-06, "loss": 0.8915, "step": 1341 }, { "epoch": 1.34, "grad_norm": 3.323082396246443, "learning_rate": 1.181525310570677e-06, "loss": 0.7379, "step": 1342 }, { "epoch": 1.34, "grad_norm": 3.3460733379461782, "learning_rate": 1.178136277222573e-06, "loss": 0.7939, "step": 1343 }, { "epoch": 1.34, "grad_norm": 3.8127971165454784, "learning_rate": 1.174750612371346e-06, "loss": 0.741, "step": 1344 }, { "epoch": 1.34, "grad_norm": 4.399459204511115, "learning_rate": 1.1713683246446622e-06, "loss": 0.7689, "step": 1345 }, { "epoch": 1.34, "grad_norm": 3.9405586575262963, "learning_rate": 1.1679894226615861e-06, "loss": 0.809, "step": 1346 }, { "epoch": 1.34, "grad_norm": 7.742517122387614, "learning_rate": 1.1646139150325509e-06, "loss": 0.7601, "step": 1347 }, { "epoch": 1.34, "grad_norm": 3.691444290644045, "learning_rate": 1.16124181035934e-06, "loss": 0.8064, "step": 1348 }, { "epoch": 1.34, "grad_norm": 3.3219867744503584, "learning_rate": 1.157873117235067e-06, "loss": 0.7491, "step": 1349 }, { "epoch": 1.34, "grad_norm": 3.187089872214097, "learning_rate": 1.1545078442441526e-06, "loss": 0.7781, "step": 1350 }, { "epoch": 1.35, "grad_norm": 3.0768022509277158, "learning_rate": 1.1511459999622982e-06, "loss": 0.7604, "step": 1351 }, { "epoch": 1.35, "grad_norm": 3.4393576763182936, "learning_rate": 1.1477875929564719e-06, "loss": 0.7792, "step": 1352 }, { "epoch": 1.35, "grad_norm": 3.65351626181679, "learning_rate": 1.14443263178488e-06, "loss": 0.7811, "step": 1353 }, { "epoch": 1.35, "grad_norm": 3.322343243133562, "learning_rate": 1.1410811249969475e-06, "loss": 0.7763, "step": 1354 }, { "epoch": 1.35, "grad_norm": 3.4584313417950967, "learning_rate": 1.1377330811332988e-06, "loss": 0.8107, "step": 1355 }, { "epoch": 1.35, "grad_norm": 3.4353154789576688, "learning_rate": 1.1343885087257337e-06, "loss": 0.7723, "step": 1356 }, { "epoch": 1.35, "grad_norm": 3.4552724449743484, "learning_rate": 1.1310474162972026e-06, "loss": 0.7771, "step": 1357 }, { "epoch": 1.35, "grad_norm": 3.460219109569181, "learning_rate": 1.1277098123617922e-06, "loss": 0.7636, "step": 1358 }, { "epoch": 1.35, "grad_norm": 3.523448157612515, "learning_rate": 1.124375705424696e-06, "loss": 0.7462, "step": 1359 }, { "epoch": 1.35, "grad_norm": 3.424560990004581, "learning_rate": 1.1210451039821965e-06, "loss": 0.8027, "step": 1360 }, { "epoch": 1.36, "grad_norm": 4.652741731414873, "learning_rate": 1.117718016521645e-06, "loss": 0.7742, "step": 1361 }, { "epoch": 1.36, "grad_norm": 3.937180761382406, "learning_rate": 1.1143944515214385e-06, "loss": 0.7696, "step": 1362 }, { "epoch": 1.36, "grad_norm": 3.4757745372358455, "learning_rate": 1.1110744174509952e-06, "loss": 0.7384, "step": 1363 }, { "epoch": 1.36, "grad_norm": 3.641170072794961, "learning_rate": 1.1077579227707359e-06, "loss": 0.8981, "step": 1364 }, { "epoch": 1.36, "grad_norm": 3.6319058647281692, "learning_rate": 1.1044449759320642e-06, "loss": 0.7671, "step": 1365 }, { "epoch": 1.36, "grad_norm": 3.631306137670091, "learning_rate": 1.10113558537734e-06, "loss": 0.7509, "step": 1366 }, { "epoch": 1.36, "grad_norm": 3.808855605778371, "learning_rate": 1.0978297595398632e-06, "loss": 0.8332, "step": 1367 }, { "epoch": 1.36, "grad_norm": 3.6386463011966104, "learning_rate": 1.094527506843849e-06, "loss": 0.7681, "step": 1368 }, { "epoch": 1.36, "grad_norm": 3.5462268476262286, "learning_rate": 1.0912288357044063e-06, "loss": 0.7841, "step": 1369 }, { "epoch": 1.36, "grad_norm": 3.701501889272336, "learning_rate": 1.0879337545275165e-06, "loss": 0.7895, "step": 1370 }, { "epoch": 1.37, "grad_norm": 3.7494658510788414, "learning_rate": 1.0846422717100161e-06, "loss": 0.7807, "step": 1371 }, { "epoch": 1.37, "grad_norm": 3.2692856294194446, "learning_rate": 1.0813543956395675e-06, "loss": 0.785, "step": 1372 }, { "epoch": 1.37, "grad_norm": 3.0845583618770163, "learning_rate": 1.0780701346946453e-06, "loss": 0.8924, "step": 1373 }, { "epoch": 1.37, "grad_norm": 3.215984926683097, "learning_rate": 1.074789497244512e-06, "loss": 0.7701, "step": 1374 }, { "epoch": 1.37, "grad_norm": 3.957979514359011, "learning_rate": 1.0715124916491937e-06, "loss": 0.8176, "step": 1375 }, { "epoch": 1.37, "grad_norm": 3.2537986277830027, "learning_rate": 1.068239126259462e-06, "loss": 0.7728, "step": 1376 }, { "epoch": 1.37, "grad_norm": 3.50634653973343, "learning_rate": 1.0649694094168148e-06, "loss": 0.7767, "step": 1377 }, { "epoch": 1.37, "grad_norm": 3.0303187703983703, "learning_rate": 1.0617033494534486e-06, "loss": 0.7455, "step": 1378 }, { "epoch": 1.37, "grad_norm": 3.647550760135323, "learning_rate": 1.0584409546922445e-06, "loss": 0.7595, "step": 1379 }, { "epoch": 1.37, "grad_norm": 3.3323285585453357, "learning_rate": 1.0551822334467429e-06, "loss": 0.8867, "step": 1380 }, { "epoch": 1.38, "grad_norm": 3.2373920875142153, "learning_rate": 1.0519271940211214e-06, "loss": 0.7561, "step": 1381 }, { "epoch": 1.38, "grad_norm": 3.4336615901451744, "learning_rate": 1.0486758447101753e-06, "loss": 0.8313, "step": 1382 }, { "epoch": 1.38, "grad_norm": 3.4642260593747456, "learning_rate": 1.045428193799299e-06, "loss": 0.7711, "step": 1383 }, { "epoch": 1.38, "grad_norm": 4.227930679565343, "learning_rate": 1.0421842495644588e-06, "loss": 0.7481, "step": 1384 }, { "epoch": 1.38, "grad_norm": 3.642348770065061, "learning_rate": 1.0389440202721779e-06, "loss": 0.7941, "step": 1385 }, { "epoch": 1.38, "grad_norm": 3.5675794723407455, "learning_rate": 1.035707514179513e-06, "loss": 0.7625, "step": 1386 }, { "epoch": 1.38, "grad_norm": 3.62754690887538, "learning_rate": 1.032474739534031e-06, "loss": 0.8042, "step": 1387 }, { "epoch": 1.38, "grad_norm": 3.6077876348985423, "learning_rate": 1.0292457045737895e-06, "loss": 0.7732, "step": 1388 }, { "epoch": 1.38, "grad_norm": 3.367550422325174, "learning_rate": 1.02602041752732e-06, "loss": 0.7958, "step": 1389 }, { "epoch": 1.38, "grad_norm": 3.505928191381052, "learning_rate": 1.0227988866135995e-06, "loss": 0.7748, "step": 1390 }, { "epoch": 1.39, "grad_norm": 3.183043581273117, "learning_rate": 1.0195811200420332e-06, "loss": 0.8153, "step": 1391 }, { "epoch": 1.39, "grad_norm": 3.4823956162540646, "learning_rate": 1.0163671260124386e-06, "loss": 0.7771, "step": 1392 }, { "epoch": 1.39, "grad_norm": 3.108735194248819, "learning_rate": 1.0131569127150143e-06, "loss": 0.7693, "step": 1393 }, { "epoch": 1.39, "grad_norm": 3.892788349604029, "learning_rate": 1.0099504883303255e-06, "loss": 0.7977, "step": 1394 }, { "epoch": 1.39, "grad_norm": 4.014475613146025, "learning_rate": 1.0067478610292846e-06, "loss": 0.7941, "step": 1395 }, { "epoch": 1.39, "grad_norm": 3.3116973410380623, "learning_rate": 1.0035490389731257e-06, "loss": 0.7948, "step": 1396 }, { "epoch": 1.39, "grad_norm": 3.7902261440021214, "learning_rate": 1.0003540303133843e-06, "loss": 0.8941, "step": 1397 }, { "epoch": 1.39, "grad_norm": 4.240637725947338, "learning_rate": 9.971628431918845e-07, "loss": 0.742, "step": 1398 }, { "epoch": 1.39, "grad_norm": 3.2768607415284876, "learning_rate": 9.939754857407064e-07, "loss": 0.8011, "step": 1399 }, { "epoch": 1.4, "grad_norm": 3.5291913806476405, "learning_rate": 9.90791966082171e-07, "loss": 0.7891, "step": 1400 }, { "epoch": 1.4, "grad_norm": 3.2909363328810506, "learning_rate": 9.87612292328824e-07, "loss": 0.7557, "step": 1401 }, { "epoch": 1.4, "grad_norm": 3.422720752548584, "learning_rate": 9.844364725834058e-07, "loss": 0.7615, "step": 1402 }, { "epoch": 1.4, "grad_norm": 3.4897995884986743, "learning_rate": 9.812645149388363e-07, "loss": 0.7687, "step": 1403 }, { "epoch": 1.4, "grad_norm": 3.199379145909971, "learning_rate": 9.780964274781984e-07, "loss": 0.7615, "step": 1404 }, { "epoch": 1.4, "grad_norm": 3.2685783569349995, "learning_rate": 9.749322182747074e-07, "loss": 0.9173, "step": 1405 }, { "epoch": 1.4, "grad_norm": 3.3299273077447036, "learning_rate": 9.71771895391696e-07, "loss": 0.7975, "step": 1406 }, { "epoch": 1.4, "grad_norm": 3.635202310680024, "learning_rate": 9.686154668825972e-07, "loss": 0.8437, "step": 1407 }, { "epoch": 1.4, "grad_norm": 3.5033795350156, "learning_rate": 9.654629407909163e-07, "loss": 0.7784, "step": 1408 }, { "epoch": 1.4, "grad_norm": 3.1193406631079394, "learning_rate": 9.623143251502148e-07, "loss": 0.8012, "step": 1409 }, { "epoch": 1.41, "grad_norm": 3.1403302615425717, "learning_rate": 9.591696279840906e-07, "loss": 0.7891, "step": 1410 }, { "epoch": 1.41, "grad_norm": 3.428860058218006, "learning_rate": 9.560288573061562e-07, "loss": 0.8055, "step": 1411 }, { "epoch": 1.41, "grad_norm": 3.3092379856295002, "learning_rate": 9.52892021120016e-07, "loss": 0.742, "step": 1412 }, { "epoch": 1.41, "grad_norm": 3.263559084727425, "learning_rate": 9.497591274192508e-07, "loss": 0.7798, "step": 1413 }, { "epoch": 1.41, "grad_norm": 3.6296575460771883, "learning_rate": 9.466301841873929e-07, "loss": 0.743, "step": 1414 }, { "epoch": 1.41, "grad_norm": 3.15847592118413, "learning_rate": 9.435051993979078e-07, "loss": 0.7927, "step": 1415 }, { "epoch": 1.41, "grad_norm": 3.5758773271627717, "learning_rate": 9.403841810141748e-07, "loss": 0.814, "step": 1416 }, { "epoch": 1.41, "grad_norm": 3.7266404683560586, "learning_rate": 9.372671369894662e-07, "loss": 0.8801, "step": 1417 }, { "epoch": 1.41, "grad_norm": 3.290372979589686, "learning_rate": 9.341540752669234e-07, "loss": 0.782, "step": 1418 }, { "epoch": 1.41, "grad_norm": 2.891132859409568, "learning_rate": 9.310450037795435e-07, "loss": 0.76, "step": 1419 }, { "epoch": 1.42, "grad_norm": 3.3795890816366594, "learning_rate": 9.279399304501526e-07, "loss": 0.7799, "step": 1420 }, { "epoch": 1.42, "grad_norm": 3.3816506208317, "learning_rate": 9.248388631913888e-07, "loss": 0.7787, "step": 1421 }, { "epoch": 1.42, "grad_norm": 3.4614212326924267, "learning_rate": 9.217418099056821e-07, "loss": 0.7771, "step": 1422 }, { "epoch": 1.42, "grad_norm": 3.858579435506388, "learning_rate": 9.186487784852349e-07, "loss": 0.7655, "step": 1423 }, { "epoch": 1.42, "grad_norm": 4.24602517267995, "learning_rate": 9.155597768119978e-07, "loss": 0.7278, "step": 1424 }, { "epoch": 1.42, "grad_norm": 3.1086216366935715, "learning_rate": 9.124748127576553e-07, "loss": 0.8024, "step": 1425 }, { "epoch": 1.42, "grad_norm": 2.8188868457650824, "learning_rate": 9.093938941836012e-07, "loss": 0.779, "step": 1426 }, { "epoch": 1.42, "grad_norm": 3.1041422304428856, "learning_rate": 9.063170289409193e-07, "loss": 0.7508, "step": 1427 }, { "epoch": 1.42, "grad_norm": 3.17434210961302, "learning_rate": 9.032442248703666e-07, "loss": 0.7846, "step": 1428 }, { "epoch": 1.42, "grad_norm": 3.391106774584389, "learning_rate": 9.001754898023512e-07, "loss": 0.7425, "step": 1429 }, { "epoch": 1.43, "grad_norm": 3.3088661186663244, "learning_rate": 8.971108315569094e-07, "loss": 0.7679, "step": 1430 }, { "epoch": 1.43, "grad_norm": 3.2433916865346895, "learning_rate": 8.940502579436913e-07, "loss": 0.7645, "step": 1431 }, { "epoch": 1.43, "grad_norm": 3.6369567746602036, "learning_rate": 8.909937767619369e-07, "loss": 0.7722, "step": 1432 }, { "epoch": 1.43, "grad_norm": 3.7221101256588804, "learning_rate": 8.879413958004565e-07, "loss": 0.8094, "step": 1433 }, { "epoch": 1.43, "grad_norm": 3.368855704953316, "learning_rate": 8.848931228376137e-07, "loss": 0.7712, "step": 1434 }, { "epoch": 1.43, "grad_norm": 3.3682957368569637, "learning_rate": 8.818489656413042e-07, "loss": 0.9297, "step": 1435 }, { "epoch": 1.43, "grad_norm": 4.841492816662529, "learning_rate": 8.788089319689325e-07, "loss": 0.7871, "step": 1436 }, { "epoch": 1.43, "grad_norm": 3.4659976318232517, "learning_rate": 8.757730295673986e-07, "loss": 0.7531, "step": 1437 }, { "epoch": 1.43, "grad_norm": 3.430613832869515, "learning_rate": 8.727412661730724e-07, "loss": 0.7665, "step": 1438 }, { "epoch": 1.43, "grad_norm": 3.1363272288634696, "learning_rate": 8.697136495117764e-07, "loss": 0.7479, "step": 1439 }, { "epoch": 1.44, "grad_norm": 3.3770033396680095, "learning_rate": 8.666901872987677e-07, "loss": 0.7666, "step": 1440 }, { "epoch": 1.44, "grad_norm": 3.594025510230481, "learning_rate": 8.636708872387162e-07, "loss": 0.7875, "step": 1441 }, { "epoch": 1.44, "grad_norm": 2.8843280129736075, "learning_rate": 8.606557570256844e-07, "loss": 0.7676, "step": 1442 }, { "epoch": 1.44, "grad_norm": 3.1512239404887286, "learning_rate": 8.576448043431082e-07, "loss": 0.7982, "step": 1443 }, { "epoch": 1.44, "grad_norm": 3.831019356023553, "learning_rate": 8.546380368637813e-07, "loss": 0.7782, "step": 1444 }, { "epoch": 1.44, "grad_norm": 3.474394824915077, "learning_rate": 8.516354622498279e-07, "loss": 0.7689, "step": 1445 }, { "epoch": 1.44, "grad_norm": 2.992472448478769, "learning_rate": 8.486370881526917e-07, "loss": 0.8199, "step": 1446 }, { "epoch": 1.44, "grad_norm": 3.9096524667646406, "learning_rate": 8.456429222131083e-07, "loss": 0.7932, "step": 1447 }, { "epoch": 1.44, "grad_norm": 3.075246224442502, "learning_rate": 8.426529720610935e-07, "loss": 0.7497, "step": 1448 }, { "epoch": 1.44, "grad_norm": 3.6445648620889473, "learning_rate": 8.396672453159164e-07, "loss": 0.7581, "step": 1449 }, { "epoch": 1.45, "grad_norm": 3.2208677402283996, "learning_rate": 8.366857495860869e-07, "loss": 0.7933, "step": 1450 }, { "epoch": 1.45, "grad_norm": 3.751236108812793, "learning_rate": 8.337084924693303e-07, "loss": 0.8085, "step": 1451 }, { "epoch": 1.45, "grad_norm": 3.5240132297532223, "learning_rate": 8.307354815525731e-07, "loss": 0.7558, "step": 1452 }, { "epoch": 1.45, "grad_norm": 3.2893942265495917, "learning_rate": 8.277667244119186e-07, "loss": 0.8508, "step": 1453 }, { "epoch": 1.45, "grad_norm": 3.5435469777461246, "learning_rate": 8.24802228612633e-07, "loss": 0.8234, "step": 1454 }, { "epoch": 1.45, "grad_norm": 3.1791311488878176, "learning_rate": 8.218420017091209e-07, "loss": 0.7765, "step": 1455 }, { "epoch": 1.45, "grad_norm": 3.793753874992174, "learning_rate": 8.188860512449107e-07, "loss": 0.7626, "step": 1456 }, { "epoch": 1.45, "grad_norm": 4.448891639922815, "learning_rate": 8.159343847526308e-07, "loss": 0.8217, "step": 1457 }, { "epoch": 1.45, "grad_norm": 4.216766804633887, "learning_rate": 8.129870097539951e-07, "loss": 0.8178, "step": 1458 }, { "epoch": 1.45, "grad_norm": 3.272364791054044, "learning_rate": 8.100439337597799e-07, "loss": 0.7803, "step": 1459 }, { "epoch": 1.46, "grad_norm": 2.9457799937452993, "learning_rate": 8.071051642698075e-07, "loss": 0.7593, "step": 1460 }, { "epoch": 1.46, "grad_norm": 3.459105910482318, "learning_rate": 8.041707087729245e-07, "loss": 0.7715, "step": 1461 }, { "epoch": 1.46, "grad_norm": 2.9658810425775672, "learning_rate": 8.012405747469861e-07, "loss": 0.7675, "step": 1462 }, { "epoch": 1.46, "grad_norm": 2.8688377521091275, "learning_rate": 7.983147696588339e-07, "loss": 0.7487, "step": 1463 }, { "epoch": 1.46, "grad_norm": 3.422266345146778, "learning_rate": 7.953933009642772e-07, "loss": 0.7379, "step": 1464 }, { "epoch": 1.46, "grad_norm": 3.3212977552659, "learning_rate": 7.924761761080769e-07, "loss": 0.8916, "step": 1465 }, { "epoch": 1.46, "grad_norm": 4.819589260463883, "learning_rate": 7.895634025239243e-07, "loss": 0.8141, "step": 1466 }, { "epoch": 1.46, "grad_norm": 3.1663730418300244, "learning_rate": 7.866549876344201e-07, "loss": 0.8092, "step": 1467 }, { "epoch": 1.46, "grad_norm": 3.136072642945885, "learning_rate": 7.837509388510611e-07, "loss": 0.7905, "step": 1468 }, { "epoch": 1.46, "grad_norm": 3.9325540650221846, "learning_rate": 7.808512635742157e-07, "loss": 0.8215, "step": 1469 }, { "epoch": 1.47, "grad_norm": 3.5950029662320055, "learning_rate": 7.779559691931065e-07, "loss": 0.7761, "step": 1470 }, { "epoch": 1.47, "grad_norm": 3.1437693044783694, "learning_rate": 7.750650630857948e-07, "loss": 0.7492, "step": 1471 }, { "epoch": 1.47, "grad_norm": 2.9669781863860414, "learning_rate": 7.721785526191589e-07, "loss": 0.7603, "step": 1472 }, { "epoch": 1.47, "grad_norm": 3.291338283146987, "learning_rate": 7.692964451488736e-07, "loss": 0.8417, "step": 1473 }, { "epoch": 1.47, "grad_norm": 3.1456118229240997, "learning_rate": 7.66418748019396e-07, "loss": 0.8107, "step": 1474 }, { "epoch": 1.47, "grad_norm": 3.498228192798247, "learning_rate": 7.635454685639429e-07, "loss": 0.7692, "step": 1475 }, { "epoch": 1.47, "grad_norm": 3.5159714426234374, "learning_rate": 7.606766141044733e-07, "loss": 0.747, "step": 1476 }, { "epoch": 1.47, "grad_norm": 3.1948277787276163, "learning_rate": 7.578121919516712e-07, "loss": 0.8195, "step": 1477 }, { "epoch": 1.47, "grad_norm": 4.201124202458152, "learning_rate": 7.54952209404926e-07, "loss": 0.9011, "step": 1478 }, { "epoch": 1.47, "grad_norm": 3.2319645718539505, "learning_rate": 7.520966737523117e-07, "loss": 0.778, "step": 1479 }, { "epoch": 1.48, "grad_norm": 3.165296568149437, "learning_rate": 7.492455922705727e-07, "loss": 0.809, "step": 1480 }, { "epoch": 1.48, "grad_norm": 3.643918972374975, "learning_rate": 7.463989722251014e-07, "loss": 0.7649, "step": 1481 }, { "epoch": 1.48, "grad_norm": 3.068824178341287, "learning_rate": 7.435568208699203e-07, "loss": 0.7661, "step": 1482 }, { "epoch": 1.48, "grad_norm": 5.220754830582587, "learning_rate": 7.407191454476667e-07, "loss": 0.9245, "step": 1483 }, { "epoch": 1.48, "grad_norm": 3.4623672614135583, "learning_rate": 7.37885953189571e-07, "loss": 0.8074, "step": 1484 }, { "epoch": 1.48, "grad_norm": 3.9596735679732706, "learning_rate": 7.350572513154377e-07, "loss": 0.9133, "step": 1485 }, { "epoch": 1.48, "eval_loss": 0.7714810967445374, "eval_runtime": 132.243, "eval_samples_per_second": 9.883, "eval_steps_per_second": 0.28, "step": 1485 }, { "epoch": 1.48, "grad_norm": 16.21067244523335, "learning_rate": 7.322330470336314e-07, "loss": 0.7437, "step": 1486 }, { "epoch": 1.48, "grad_norm": 3.5800765380379085, "learning_rate": 7.294133475410528e-07, "loss": 0.8008, "step": 1487 }, { "epoch": 1.48, "grad_norm": 3.7440861946582467, "learning_rate": 7.265981600231234e-07, "loss": 0.795, "step": 1488 }, { "epoch": 1.48, "grad_norm": 3.1486176134117834, "learning_rate": 7.23787491653769e-07, "loss": 0.782, "step": 1489 }, { "epoch": 1.49, "grad_norm": 3.2496188791469387, "learning_rate": 7.209813495953963e-07, "loss": 0.7812, "step": 1490 }, { "epoch": 1.49, "grad_norm": 3.165185657173258, "learning_rate": 7.181797409988803e-07, "loss": 0.7839, "step": 1491 }, { "epoch": 1.49, "grad_norm": 3.144201973996701, "learning_rate": 7.153826730035424e-07, "loss": 0.7747, "step": 1492 }, { "epoch": 1.49, "grad_norm": 3.4597437374187217, "learning_rate": 7.125901527371329e-07, "loss": 0.8611, "step": 1493 }, { "epoch": 1.49, "grad_norm": 2.952619750642094, "learning_rate": 7.098021873158126e-07, "loss": 0.7436, "step": 1494 }, { "epoch": 1.49, "grad_norm": 3.116824212374202, "learning_rate": 7.07018783844137e-07, "loss": 0.7795, "step": 1495 }, { "epoch": 1.49, "grad_norm": 3.2547785737791584, "learning_rate": 7.042399494150342e-07, "loss": 0.7683, "step": 1496 }, { "epoch": 1.49, "grad_norm": 3.5300090306280163, "learning_rate": 7.014656911097909e-07, "loss": 0.7762, "step": 1497 }, { "epoch": 1.49, "grad_norm": 7.99053927440605, "learning_rate": 6.986960159980327e-07, "loss": 0.7786, "step": 1498 }, { "epoch": 1.5, "grad_norm": 5.283118212626346, "learning_rate": 6.959309311377038e-07, "loss": 0.8195, "step": 1499 }, { "epoch": 1.5, "grad_norm": 3.245545957369501, "learning_rate": 6.931704435750522e-07, "loss": 0.8692, "step": 1500 }, { "epoch": 1.5, "grad_norm": 6.318217113098099, "learning_rate": 6.904145603446116e-07, "loss": 0.7823, "step": 1501 }, { "epoch": 1.5, "grad_norm": 3.5768560468167907, "learning_rate": 6.876632884691803e-07, "loss": 0.9201, "step": 1502 }, { "epoch": 1.5, "grad_norm": 3.2514539489681393, "learning_rate": 6.84916634959808e-07, "loss": 0.769, "step": 1503 }, { "epoch": 1.5, "grad_norm": 3.1325846278882334, "learning_rate": 6.821746068157742e-07, "loss": 0.8101, "step": 1504 }, { "epoch": 1.5, "grad_norm": 3.8437110271734745, "learning_rate": 6.794372110245717e-07, "loss": 0.7683, "step": 1505 }, { "epoch": 1.5, "grad_norm": 4.0583337671108195, "learning_rate": 6.767044545618879e-07, "loss": 0.7359, "step": 1506 }, { "epoch": 1.5, "grad_norm": 3.4467168260350394, "learning_rate": 6.739763443915894e-07, "loss": 0.7958, "step": 1507 }, { "epoch": 1.5, "grad_norm": 3.335243571214721, "learning_rate": 6.712528874657012e-07, "loss": 0.7513, "step": 1508 }, { "epoch": 1.51, "grad_norm": 3.5062931739295595, "learning_rate": 6.685340907243915e-07, "loss": 0.8187, "step": 1509 }, { "epoch": 1.51, "grad_norm": 3.3543365564989087, "learning_rate": 6.658199610959537e-07, "loss": 0.7641, "step": 1510 }, { "epoch": 1.51, "grad_norm": 3.5094301991405636, "learning_rate": 6.631105054967859e-07, "loss": 0.7987, "step": 1511 }, { "epoch": 1.51, "grad_norm": 3.491389394205674, "learning_rate": 6.604057308313763e-07, "loss": 0.925, "step": 1512 }, { "epoch": 1.51, "grad_norm": 3.292738546769483, "learning_rate": 6.577056439922858e-07, "loss": 0.7651, "step": 1513 }, { "epoch": 1.51, "grad_norm": 3.4761798327388296, "learning_rate": 6.55010251860127e-07, "loss": 0.9854, "step": 1514 }, { "epoch": 1.51, "grad_norm": 3.526936641143009, "learning_rate": 6.523195613035522e-07, "loss": 0.7815, "step": 1515 }, { "epoch": 1.51, "grad_norm": 3.3370049015756393, "learning_rate": 6.496335791792294e-07, "loss": 0.8003, "step": 1516 }, { "epoch": 1.51, "grad_norm": 3.0791626467317843, "learning_rate": 6.469523123318308e-07, "loss": 0.8099, "step": 1517 }, { "epoch": 1.51, "grad_norm": 2.9923179915874876, "learning_rate": 6.442757675940109e-07, "loss": 0.7798, "step": 1518 }, { "epoch": 1.52, "grad_norm": 3.4647937601523537, "learning_rate": 6.41603951786392e-07, "loss": 0.7565, "step": 1519 }, { "epoch": 1.52, "grad_norm": 3.3125038280454215, "learning_rate": 6.389368717175449e-07, "loss": 0.761, "step": 1520 }, { "epoch": 1.52, "grad_norm": 3.4381942950202284, "learning_rate": 6.362745341839729e-07, "loss": 0.864, "step": 1521 }, { "epoch": 1.52, "grad_norm": 3.118549631024769, "learning_rate": 6.336169459700933e-07, "loss": 0.7513, "step": 1522 }, { "epoch": 1.52, "grad_norm": 3.022721960414149, "learning_rate": 6.309641138482223e-07, "loss": 0.8259, "step": 1523 }, { "epoch": 1.52, "grad_norm": 3.882786834428429, "learning_rate": 6.283160445785533e-07, "loss": 0.758, "step": 1524 }, { "epoch": 1.52, "grad_norm": 3.517627437074961, "learning_rate": 6.25672744909146e-07, "loss": 0.8175, "step": 1525 }, { "epoch": 1.52, "grad_norm": 3.2101847910161725, "learning_rate": 6.230342215759027e-07, "loss": 0.8272, "step": 1526 }, { "epoch": 1.52, "grad_norm": 4.353281924550767, "learning_rate": 6.204004813025569e-07, "loss": 0.885, "step": 1527 }, { "epoch": 1.52, "grad_norm": 3.6961265323357306, "learning_rate": 6.177715308006505e-07, "loss": 0.7936, "step": 1528 }, { "epoch": 1.53, "grad_norm": 4.044436931434244, "learning_rate": 6.151473767695229e-07, "loss": 0.7706, "step": 1529 }, { "epoch": 1.53, "grad_norm": 3.513248654832656, "learning_rate": 6.125280258962874e-07, "loss": 0.7758, "step": 1530 }, { "epoch": 1.53, "grad_norm": 3.331760445345406, "learning_rate": 6.099134848558208e-07, "loss": 0.7305, "step": 1531 }, { "epoch": 1.53, "grad_norm": 3.210528692980601, "learning_rate": 6.073037603107404e-07, "loss": 0.7628, "step": 1532 }, { "epoch": 1.53, "grad_norm": 3.1957725026843646, "learning_rate": 6.0469885891139e-07, "loss": 0.7709, "step": 1533 }, { "epoch": 1.53, "grad_norm": 3.451690589829594, "learning_rate": 6.020987872958237e-07, "loss": 0.8934, "step": 1534 }, { "epoch": 1.53, "grad_norm": 5.354162042664116, "learning_rate": 5.995035520897883e-07, "loss": 0.905, "step": 1535 }, { "epoch": 1.53, "grad_norm": 3.1280851178956963, "learning_rate": 5.969131599067043e-07, "loss": 0.7612, "step": 1536 }, { "epoch": 1.53, "grad_norm": 3.3619694211157323, "learning_rate": 5.943276173476509e-07, "loss": 0.8134, "step": 1537 }, { "epoch": 1.53, "grad_norm": 3.12495073641787, "learning_rate": 5.91746931001351e-07, "loss": 0.8394, "step": 1538 }, { "epoch": 1.54, "grad_norm": 3.6467123238235852, "learning_rate": 5.891711074441495e-07, "loss": 0.9064, "step": 1539 }, { "epoch": 1.54, "grad_norm": 3.763597818250878, "learning_rate": 5.866001532400023e-07, "loss": 0.7685, "step": 1540 }, { "epoch": 1.54, "grad_norm": 3.425498558130636, "learning_rate": 5.84034074940456e-07, "loss": 0.7478, "step": 1541 }, { "epoch": 1.54, "grad_norm": 3.5534813827238754, "learning_rate": 5.814728790846308e-07, "loss": 0.8062, "step": 1542 }, { "epoch": 1.54, "grad_norm": 3.133965609326465, "learning_rate": 5.789165721992052e-07, "loss": 0.7537, "step": 1543 }, { "epoch": 1.54, "grad_norm": 3.185613641333031, "learning_rate": 5.763651607984008e-07, "loss": 0.8008, "step": 1544 }, { "epoch": 1.54, "grad_norm": 3.4098961543034405, "learning_rate": 5.73818651383962e-07, "loss": 0.7602, "step": 1545 }, { "epoch": 1.54, "grad_norm": 3.41502064836941, "learning_rate": 5.712770504451426e-07, "loss": 0.7927, "step": 1546 }, { "epoch": 1.54, "grad_norm": 4.065982309049072, "learning_rate": 5.687403644586891e-07, "loss": 0.7619, "step": 1547 }, { "epoch": 1.54, "grad_norm": 3.5405979893866117, "learning_rate": 5.662085998888215e-07, "loss": 0.7409, "step": 1548 }, { "epoch": 1.55, "grad_norm": 3.364106722628429, "learning_rate": 5.636817631872185e-07, "loss": 0.8034, "step": 1549 }, { "epoch": 1.55, "grad_norm": 3.076885150013064, "learning_rate": 5.611598607930033e-07, "loss": 0.7608, "step": 1550 }, { "epoch": 1.55, "grad_norm": 3.293115531510627, "learning_rate": 5.586428991327222e-07, "loss": 0.7982, "step": 1551 }, { "epoch": 1.55, "grad_norm": 3.1697744832142973, "learning_rate": 5.561308846203333e-07, "loss": 0.8024, "step": 1552 }, { "epoch": 1.55, "grad_norm": 5.928997695083327, "learning_rate": 5.536238236571878e-07, "loss": 0.7811, "step": 1553 }, { "epoch": 1.55, "grad_norm": 3.5539229784700157, "learning_rate": 5.511217226320126e-07, "loss": 0.7898, "step": 1554 }, { "epoch": 1.55, "grad_norm": 3.075918181098289, "learning_rate": 5.486245879208946e-07, "loss": 0.7846, "step": 1555 }, { "epoch": 1.55, "grad_norm": 3.0913037666780245, "learning_rate": 5.46132425887268e-07, "loss": 0.8126, "step": 1556 }, { "epoch": 1.55, "grad_norm": 3.230962473117362, "learning_rate": 5.436452428818919e-07, "loss": 0.7576, "step": 1557 }, { "epoch": 1.55, "grad_norm": 3.438209258571105, "learning_rate": 5.411630452428396e-07, "loss": 0.7761, "step": 1558 }, { "epoch": 1.56, "grad_norm": 3.3721724793499486, "learning_rate": 5.386858392954799e-07, "loss": 0.7804, "step": 1559 }, { "epoch": 1.56, "grad_norm": 3.2256936678207575, "learning_rate": 5.362136313524607e-07, "loss": 0.7505, "step": 1560 }, { "epoch": 1.56, "grad_norm": 3.4659234295299894, "learning_rate": 5.337464277136925e-07, "loss": 0.8187, "step": 1561 }, { "epoch": 1.56, "grad_norm": 3.154353986456024, "learning_rate": 5.312842346663363e-07, "loss": 0.7856, "step": 1562 }, { "epoch": 1.56, "grad_norm": 3.6649530036211244, "learning_rate": 5.288270584847812e-07, "loss": 0.801, "step": 1563 }, { "epoch": 1.56, "grad_norm": 3.5891062417702373, "learning_rate": 5.263749054306347e-07, "loss": 0.7557, "step": 1564 }, { "epoch": 1.56, "grad_norm": 3.4284156040982348, "learning_rate": 5.239277817527031e-07, "loss": 0.8173, "step": 1565 }, { "epoch": 1.56, "grad_norm": 3.1131647399188562, "learning_rate": 5.214856936869753e-07, "loss": 0.7978, "step": 1566 }, { "epoch": 1.56, "grad_norm": 3.444857692149311, "learning_rate": 5.190486474566084e-07, "loss": 0.803, "step": 1567 }, { "epoch": 1.56, "grad_norm": 3.277265929358663, "learning_rate": 5.166166492719124e-07, "loss": 0.7943, "step": 1568 }, { "epoch": 1.57, "grad_norm": 4.474745056122077, "learning_rate": 5.141897053303327e-07, "loss": 0.9053, "step": 1569 }, { "epoch": 1.57, "grad_norm": 2.772693698393096, "learning_rate": 5.117678218164337e-07, "loss": 0.746, "step": 1570 }, { "epoch": 1.57, "grad_norm": 3.5088271605351133, "learning_rate": 5.093510049018879e-07, "loss": 0.746, "step": 1571 }, { "epoch": 1.57, "grad_norm": 3.2100113699047053, "learning_rate": 5.069392607454532e-07, "loss": 0.7867, "step": 1572 }, { "epoch": 1.57, "grad_norm": 3.7626312587287627, "learning_rate": 5.045325954929614e-07, "loss": 0.7541, "step": 1573 }, { "epoch": 1.57, "grad_norm": 3.7856437862914576, "learning_rate": 5.021310152773034e-07, "loss": 0.7884, "step": 1574 }, { "epoch": 1.57, "grad_norm": 3.6139283944268743, "learning_rate": 4.9973452621841e-07, "loss": 0.7408, "step": 1575 }, { "epoch": 1.57, "grad_norm": 4.971630250672382, "learning_rate": 4.973431344232377e-07, "loss": 0.7871, "step": 1576 }, { "epoch": 1.57, "grad_norm": 3.197066417562861, "learning_rate": 4.949568459857573e-07, "loss": 0.8069, "step": 1577 }, { "epoch": 1.57, "grad_norm": 3.1133709470352984, "learning_rate": 4.925756669869313e-07, "loss": 0.7262, "step": 1578 }, { "epoch": 1.58, "grad_norm": 3.5221431904419314, "learning_rate": 4.901996034947026e-07, "loss": 0.7778, "step": 1579 }, { "epoch": 1.58, "grad_norm": 4.745333255894113, "learning_rate": 4.878286615639791e-07, "loss": 0.8292, "step": 1580 }, { "epoch": 1.58, "grad_norm": 3.3406671353126076, "learning_rate": 4.854628472366171e-07, "loss": 0.7257, "step": 1581 }, { "epoch": 1.58, "grad_norm": 2.9411396567878882, "learning_rate": 4.831021665414043e-07, "loss": 0.7833, "step": 1582 }, { "epoch": 1.58, "grad_norm": 4.9534116064398575, "learning_rate": 4.80746625494051e-07, "loss": 0.7589, "step": 1583 }, { "epoch": 1.58, "grad_norm": 3.4117038298072475, "learning_rate": 4.783962300971661e-07, "loss": 0.8283, "step": 1584 }, { "epoch": 1.58, "grad_norm": 3.555149916343373, "learning_rate": 4.7605098634024684e-07, "loss": 0.7823, "step": 1585 }, { "epoch": 1.58, "grad_norm": 3.2686548916773153, "learning_rate": 4.737109001996637e-07, "loss": 0.7674, "step": 1586 }, { "epoch": 1.58, "grad_norm": 2.9146054479685546, "learning_rate": 4.713759776386428e-07, "loss": 0.8066, "step": 1587 }, { "epoch": 1.58, "grad_norm": 4.294939978474458, "learning_rate": 4.6904622460725163e-07, "loss": 0.7337, "step": 1588 }, { "epoch": 1.59, "grad_norm": 3.325676901230692, "learning_rate": 4.6672164704238577e-07, "loss": 0.8045, "step": 1589 }, { "epoch": 1.59, "grad_norm": 3.093121732379683, "learning_rate": 4.6440225086775176e-07, "loss": 0.7723, "step": 1590 }, { "epoch": 1.59, "grad_norm": 3.8048635027301483, "learning_rate": 4.6208804199385114e-07, "loss": 0.7537, "step": 1591 }, { "epoch": 1.59, "grad_norm": 3.226842314471975, "learning_rate": 4.5977902631796856e-07, "loss": 0.7905, "step": 1592 }, { "epoch": 1.59, "grad_norm": 3.4515382475230614, "learning_rate": 4.574752097241533e-07, "loss": 0.7547, "step": 1593 }, { "epoch": 1.59, "grad_norm": 2.8802554004269987, "learning_rate": 4.55176598083206e-07, "loss": 0.7463, "step": 1594 }, { "epoch": 1.59, "grad_norm": 3.295642185338833, "learning_rate": 4.5288319725266455e-07, "loss": 0.7722, "step": 1595 }, { "epoch": 1.59, "grad_norm": 3.384221838591506, "learning_rate": 4.5059501307678834e-07, "loss": 0.8351, "step": 1596 }, { "epoch": 1.59, "grad_norm": 3.335257773212835, "learning_rate": 4.4831205138654105e-07, "loss": 0.7677, "step": 1597 }, { "epoch": 1.6, "grad_norm": 3.0856431709827388, "learning_rate": 4.4603431799958074e-07, "loss": 0.8005, "step": 1598 }, { "epoch": 1.6, "grad_norm": 3.0983422163305065, "learning_rate": 4.4376181872024e-07, "loss": 0.8012, "step": 1599 }, { "epoch": 1.6, "grad_norm": 3.635035484938005, "learning_rate": 4.41494559339514e-07, "loss": 0.7798, "step": 1600 }, { "epoch": 1.6, "grad_norm": 3.414643036000891, "learning_rate": 4.3923254563504545e-07, "loss": 0.802, "step": 1601 }, { "epoch": 1.6, "grad_norm": 3.6653431256366855, "learning_rate": 4.3697578337111053e-07, "loss": 0.7559, "step": 1602 }, { "epoch": 1.6, "grad_norm": 3.186527235869702, "learning_rate": 4.347242782986008e-07, "loss": 0.7982, "step": 1603 }, { "epoch": 1.6, "grad_norm": 3.0471761808532922, "learning_rate": 4.3247803615501295e-07, "loss": 0.788, "step": 1604 }, { "epoch": 1.6, "grad_norm": 3.184691989589024, "learning_rate": 4.3023706266443144e-07, "loss": 0.7604, "step": 1605 }, { "epoch": 1.6, "grad_norm": 3.071099714327993, "learning_rate": 4.280013635375138e-07, "loss": 0.8144, "step": 1606 }, { "epoch": 1.6, "grad_norm": 3.5838150753268585, "learning_rate": 4.257709444714786e-07, "loss": 0.757, "step": 1607 }, { "epoch": 1.61, "grad_norm": 3.221818411552769, "learning_rate": 4.2354581115008895e-07, "loss": 0.8197, "step": 1608 }, { "epoch": 1.61, "grad_norm": 3.104918684917902, "learning_rate": 4.2132596924363666e-07, "loss": 0.7578, "step": 1609 }, { "epoch": 1.61, "grad_norm": 3.550352212627237, "learning_rate": 4.1911142440893197e-07, "loss": 0.8188, "step": 1610 }, { "epoch": 1.61, "grad_norm": 3.528081972604805, "learning_rate": 4.169021822892849e-07, "loss": 0.7673, "step": 1611 }, { "epoch": 1.61, "grad_norm": 3.0538608179890665, "learning_rate": 4.1469824851449207e-07, "loss": 0.7947, "step": 1612 }, { "epoch": 1.61, "grad_norm": 3.1622381615684776, "learning_rate": 4.124996287008246e-07, "loss": 0.7532, "step": 1613 }, { "epoch": 1.61, "grad_norm": 3.3650598586585456, "learning_rate": 4.103063284510117e-07, "loss": 0.7561, "step": 1614 }, { "epoch": 1.61, "grad_norm": 3.1185409143834892, "learning_rate": 4.081183533542263e-07, "loss": 0.7931, "step": 1615 }, { "epoch": 1.61, "grad_norm": 3.3611765503522197, "learning_rate": 4.059357089860702e-07, "loss": 0.7861, "step": 1616 }, { "epoch": 1.61, "grad_norm": 3.086919702360594, "learning_rate": 4.037584009085635e-07, "loss": 0.767, "step": 1617 }, { "epoch": 1.62, "grad_norm": 3.156083256025172, "learning_rate": 4.015864346701251e-07, "loss": 0.7533, "step": 1618 }, { "epoch": 1.62, "grad_norm": 3.2747917809166647, "learning_rate": 3.994198158055637e-07, "loss": 0.9259, "step": 1619 }, { "epoch": 1.62, "grad_norm": 3.234704016671206, "learning_rate": 3.9725854983606067e-07, "loss": 0.7715, "step": 1620 }, { "epoch": 1.62, "grad_norm": 3.4117187601849635, "learning_rate": 3.951026422691556e-07, "loss": 0.748, "step": 1621 }, { "epoch": 1.62, "grad_norm": 3.181287553209798, "learning_rate": 3.929520985987334e-07, "loss": 0.8294, "step": 1622 }, { "epoch": 1.62, "grad_norm": 3.459496426770352, "learning_rate": 3.908069243050122e-07, "loss": 0.9346, "step": 1623 }, { "epoch": 1.62, "grad_norm": 3.490849031317357, "learning_rate": 3.886671248545243e-07, "loss": 0.7581, "step": 1624 }, { "epoch": 1.62, "grad_norm": 3.627997767283677, "learning_rate": 3.865327057001078e-07, "loss": 0.8074, "step": 1625 }, { "epoch": 1.62, "grad_norm": 3.0578011249766206, "learning_rate": 3.8440367228089e-07, "loss": 0.7703, "step": 1626 }, { "epoch": 1.62, "grad_norm": 3.1419946216199155, "learning_rate": 3.822800300222726e-07, "loss": 0.7878, "step": 1627 }, { "epoch": 1.63, "grad_norm": 3.49848688847224, "learning_rate": 3.801617843359187e-07, "loss": 0.8015, "step": 1628 }, { "epoch": 1.63, "grad_norm": 3.2453797389915, "learning_rate": 3.7804894061974187e-07, "loss": 0.7976, "step": 1629 }, { "epoch": 1.63, "grad_norm": 4.7075979016252685, "learning_rate": 3.7594150425788677e-07, "loss": 0.7992, "step": 1630 }, { "epoch": 1.63, "grad_norm": 3.4072101100417798, "learning_rate": 3.738394806207207e-07, "loss": 0.8159, "step": 1631 }, { "epoch": 1.63, "grad_norm": 3.783642577924318, "learning_rate": 3.717428750648178e-07, "loss": 0.793, "step": 1632 }, { "epoch": 1.63, "grad_norm": 3.5934431266679607, "learning_rate": 3.6965169293294356e-07, "loss": 0.7779, "step": 1633 }, { "epoch": 1.63, "grad_norm": 3.390439860088858, "learning_rate": 3.6756593955404403e-07, "loss": 0.7663, "step": 1634 }, { "epoch": 1.63, "grad_norm": 3.304090040254168, "learning_rate": 3.6548562024323194e-07, "loss": 0.8038, "step": 1635 }, { "epoch": 1.63, "grad_norm": 3.1903997879993615, "learning_rate": 3.6341074030177114e-07, "loss": 0.794, "step": 1636 }, { "epoch": 1.63, "grad_norm": 3.3581718593820105, "learning_rate": 3.6134130501706417e-07, "loss": 0.7454, "step": 1637 }, { "epoch": 1.64, "grad_norm": 3.0863500065965335, "learning_rate": 3.5927731966264175e-07, "loss": 0.7817, "step": 1638 }, { "epoch": 1.64, "grad_norm": 2.9764558407052104, "learning_rate": 3.5721878949814327e-07, "loss": 0.7723, "step": 1639 }, { "epoch": 1.64, "grad_norm": 3.0358199311357787, "learning_rate": 3.551657197693079e-07, "loss": 0.7888, "step": 1640 }, { "epoch": 1.64, "grad_norm": 3.362836098758586, "learning_rate": 3.5311811570796053e-07, "loss": 0.7312, "step": 1641 }, { "epoch": 1.64, "grad_norm": 4.040846408628533, "learning_rate": 3.510759825319976e-07, "loss": 0.7638, "step": 1642 }, { "epoch": 1.64, "grad_norm": 3.2498705184122465, "learning_rate": 3.490393254453728e-07, "loss": 0.7722, "step": 1643 }, { "epoch": 1.64, "grad_norm": 3.1486758392836567, "learning_rate": 3.470081496380881e-07, "loss": 0.8083, "step": 1644 }, { "epoch": 1.64, "grad_norm": 3.5946194100274687, "learning_rate": 3.4498246028617536e-07, "loss": 0.7766, "step": 1645 }, { "epoch": 1.64, "grad_norm": 4.4070229518753985, "learning_rate": 3.4296226255168484e-07, "loss": 0.7974, "step": 1646 }, { "epoch": 1.64, "grad_norm": 3.70639217022746, "learning_rate": 3.4094756158267464e-07, "loss": 0.784, "step": 1647 }, { "epoch": 1.65, "grad_norm": 3.4334379142070897, "learning_rate": 3.389383625131942e-07, "loss": 0.8437, "step": 1648 }, { "epoch": 1.65, "grad_norm": 2.972948546094183, "learning_rate": 3.3693467046327116e-07, "loss": 0.7895, "step": 1649 }, { "epoch": 1.65, "grad_norm": 2.8941850553736295, "learning_rate": 3.3493649053890325e-07, "loss": 0.7539, "step": 1650 }, { "epoch": 1.65, "grad_norm": 3.2354312316280156, "learning_rate": 3.3294382783203906e-07, "loss": 0.7698, "step": 1651 }, { "epoch": 1.65, "grad_norm": 3.0408113035806648, "learning_rate": 3.309566874205672e-07, "loss": 0.8118, "step": 1652 }, { "epoch": 1.65, "grad_norm": 3.835480383735409, "learning_rate": 3.2897507436830623e-07, "loss": 0.7403, "step": 1653 }, { "epoch": 1.65, "grad_norm": 3.4911989048186496, "learning_rate": 3.2699899372498736e-07, "loss": 0.8103, "step": 1654 }, { "epoch": 1.65, "grad_norm": 3.2427222800366167, "learning_rate": 3.2502845052624355e-07, "loss": 0.9361, "step": 1655 }, { "epoch": 1.65, "grad_norm": 3.7118171710225702, "learning_rate": 3.2306344979359835e-07, "loss": 0.7662, "step": 1656 }, { "epoch": 1.65, "grad_norm": 2.96640801990311, "learning_rate": 3.211039965344512e-07, "loss": 0.7189, "step": 1657 }, { "epoch": 1.66, "grad_norm": 4.1888496147804, "learning_rate": 3.191500957420626e-07, "loss": 0.9366, "step": 1658 }, { "epoch": 1.66, "grad_norm": 2.901094608792275, "learning_rate": 3.1720175239554703e-07, "loss": 0.7444, "step": 1659 }, { "epoch": 1.66, "grad_norm": 3.5971496940026255, "learning_rate": 3.152589714598547e-07, "loss": 0.7896, "step": 1660 }, { "epoch": 1.66, "grad_norm": 2.9471023993828105, "learning_rate": 3.133217578857611e-07, "loss": 0.7398, "step": 1661 }, { "epoch": 1.66, "grad_norm": 3.779392095385681, "learning_rate": 3.1139011660985624e-07, "loss": 0.7633, "step": 1662 }, { "epoch": 1.66, "grad_norm": 3.8354083115803808, "learning_rate": 3.094640525545295e-07, "loss": 0.7753, "step": 1663 }, { "epoch": 1.66, "grad_norm": 3.5381798374408486, "learning_rate": 3.075435706279567e-07, "loss": 0.8435, "step": 1664 }, { "epoch": 1.66, "grad_norm": 3.20259895981585, "learning_rate": 3.056286757240903e-07, "loss": 0.7628, "step": 1665 }, { "epoch": 1.66, "grad_norm": 3.471404139094473, "learning_rate": 3.0371937272264454e-07, "loss": 0.751, "step": 1666 }, { "epoch": 1.66, "grad_norm": 3.6505077935947465, "learning_rate": 3.018156664890834e-07, "loss": 0.8772, "step": 1667 }, { "epoch": 1.67, "grad_norm": 3.0717598916760087, "learning_rate": 2.9991756187461e-07, "loss": 0.7886, "step": 1668 }, { "epoch": 1.67, "grad_norm": 4.057224701035791, "learning_rate": 2.980250637161525e-07, "loss": 0.8362, "step": 1669 }, { "epoch": 1.67, "grad_norm": 3.0614735455152267, "learning_rate": 2.961381768363511e-07, "loss": 0.7427, "step": 1670 }, { "epoch": 1.67, "grad_norm": 3.1072655631884847, "learning_rate": 2.9425690604354823e-07, "loss": 0.7782, "step": 1671 }, { "epoch": 1.67, "grad_norm": 3.1052103312298014, "learning_rate": 2.9238125613177406e-07, "loss": 0.7943, "step": 1672 }, { "epoch": 1.67, "grad_norm": 3.206758172352944, "learning_rate": 2.905112318807346e-07, "loss": 0.7571, "step": 1673 }, { "epoch": 1.67, "grad_norm": 2.8639824360406534, "learning_rate": 2.8864683805580136e-07, "loss": 0.811, "step": 1674 }, { "epoch": 1.67, "grad_norm": 3.191091807544574, "learning_rate": 2.8678807940799746e-07, "loss": 0.7924, "step": 1675 }, { "epoch": 1.67, "grad_norm": 3.3049237921790287, "learning_rate": 2.8493496067398485e-07, "loss": 0.776, "step": 1676 }, { "epoch": 1.67, "grad_norm": 2.89411652410967, "learning_rate": 2.830874865760552e-07, "loss": 0.8072, "step": 1677 }, { "epoch": 1.68, "grad_norm": 2.904237056783457, "learning_rate": 2.8124566182211434e-07, "loss": 0.7489, "step": 1678 }, { "epoch": 1.68, "grad_norm": 3.7408313796978514, "learning_rate": 2.794094911056719e-07, "loss": 0.9513, "step": 1679 }, { "epoch": 1.68, "grad_norm": 3.476319707973487, "learning_rate": 2.775789791058306e-07, "loss": 0.8024, "step": 1680 }, { "epoch": 1.68, "grad_norm": 3.011766232008437, "learning_rate": 2.7575413048727324e-07, "loss": 0.7126, "step": 1681 }, { "epoch": 1.68, "grad_norm": 3.1598432158778356, "learning_rate": 2.7393494990024835e-07, "loss": 0.8037, "step": 1682 }, { "epoch": 1.68, "grad_norm": 2.896812066061299, "learning_rate": 2.7212144198056374e-07, "loss": 0.7755, "step": 1683 }, { "epoch": 1.68, "grad_norm": 3.2512506442944455, "learning_rate": 2.7031361134956913e-07, "loss": 0.7601, "step": 1684 }, { "epoch": 1.68, "grad_norm": 3.8371311336746716, "learning_rate": 2.6851146261414746e-07, "loss": 0.815, "step": 1685 }, { "epoch": 1.68, "grad_norm": 3.2492343362612273, "learning_rate": 2.6671500036670324e-07, "loss": 0.8046, "step": 1686 }, { "epoch": 1.69, "grad_norm": 3.8596677952688987, "learning_rate": 2.649242291851503e-07, "loss": 0.785, "step": 1687 }, { "epoch": 1.69, "grad_norm": 3.3956580144833763, "learning_rate": 2.631391536328992e-07, "loss": 0.7594, "step": 1688 }, { "epoch": 1.69, "grad_norm": 3.0870126325390177, "learning_rate": 2.613597782588453e-07, "loss": 0.804, "step": 1689 }, { "epoch": 1.69, "grad_norm": 3.6471715644568747, "learning_rate": 2.5958610759736133e-07, "loss": 0.9639, "step": 1690 }, { "epoch": 1.69, "grad_norm": 3.846890381346315, "learning_rate": 2.5781814616827936e-07, "loss": 0.7625, "step": 1691 }, { "epoch": 1.69, "grad_norm": 3.9337654735546543, "learning_rate": 2.560558984768852e-07, "loss": 0.8878, "step": 1692 }, { "epoch": 1.69, "grad_norm": 3.631486377461807, "learning_rate": 2.5429936901390286e-07, "loss": 0.8992, "step": 1693 }, { "epoch": 1.69, "grad_norm": 3.203561054480043, "learning_rate": 2.5254856225548546e-07, "loss": 0.7485, "step": 1694 }, { "epoch": 1.69, "grad_norm": 3.3932785473759433, "learning_rate": 2.508034826632022e-07, "loss": 0.7654, "step": 1695 }, { "epoch": 1.69, "grad_norm": 3.0634895526308683, "learning_rate": 2.490641346840292e-07, "loss": 0.7659, "step": 1696 }, { "epoch": 1.7, "grad_norm": 3.1395289024889976, "learning_rate": 2.4733052275033447e-07, "loss": 0.7691, "step": 1697 }, { "epoch": 1.7, "grad_norm": 2.9381417900374824, "learning_rate": 2.456026512798715e-07, "loss": 0.7603, "step": 1698 }, { "epoch": 1.7, "grad_norm": 3.899133791732701, "learning_rate": 2.438805246757631e-07, "loss": 0.8049, "step": 1699 }, { "epoch": 1.7, "grad_norm": 3.3142413759629696, "learning_rate": 2.4216414732649436e-07, "loss": 0.7966, "step": 1700 }, { "epoch": 1.7, "grad_norm": 3.234927490341126, "learning_rate": 2.40453523605898e-07, "loss": 0.8895, "step": 1701 }, { "epoch": 1.7, "grad_norm": 4.366740240931983, "learning_rate": 2.38748657873146e-07, "loss": 0.8022, "step": 1702 }, { "epoch": 1.7, "grad_norm": 3.013678091920663, "learning_rate": 2.3704955447273636e-07, "loss": 0.7893, "step": 1703 }, { "epoch": 1.7, "grad_norm": 4.227998210437276, "learning_rate": 2.3535621773448397e-07, "loss": 0.75, "step": 1704 }, { "epoch": 1.7, "grad_norm": 3.308155194145714, "learning_rate": 2.3366865197350736e-07, "loss": 0.7856, "step": 1705 }, { "epoch": 1.7, "grad_norm": 3.2614079860954037, "learning_rate": 2.3198686149022016e-07, "loss": 0.8478, "step": 1706 }, { "epoch": 1.71, "grad_norm": 3.346270822159464, "learning_rate": 2.3031085057031778e-07, "loss": 0.7956, "step": 1707 }, { "epoch": 1.71, "grad_norm": 2.9462263160740068, "learning_rate": 2.2864062348476908e-07, "loss": 0.7866, "step": 1708 }, { "epoch": 1.71, "grad_norm": 3.479000959345854, "learning_rate": 2.269761844898022e-07, "loss": 0.7822, "step": 1709 }, { "epoch": 1.71, "grad_norm": 3.103094381716531, "learning_rate": 2.2531753782689598e-07, "loss": 0.7714, "step": 1710 }, { "epoch": 1.71, "grad_norm": 3.2213519749710247, "learning_rate": 2.2366468772276995e-07, "loss": 0.8211, "step": 1711 }, { "epoch": 1.71, "grad_norm": 3.214361550436147, "learning_rate": 2.2201763838937184e-07, "loss": 0.797, "step": 1712 }, { "epoch": 1.71, "grad_norm": 3.202481375017903, "learning_rate": 2.2037639402386567e-07, "loss": 0.7615, "step": 1713 }, { "epoch": 1.71, "grad_norm": 3.0066810426421493, "learning_rate": 2.1874095880862505e-07, "loss": 0.8022, "step": 1714 }, { "epoch": 1.71, "grad_norm": 3.1192932057492553, "learning_rate": 2.1711133691121906e-07, "loss": 0.7795, "step": 1715 }, { "epoch": 1.71, "grad_norm": 3.2337076171314902, "learning_rate": 2.1548753248440163e-07, "loss": 0.7893, "step": 1716 }, { "epoch": 1.72, "grad_norm": 3.8881122026217376, "learning_rate": 2.138695496661039e-07, "loss": 0.7675, "step": 1717 }, { "epoch": 1.72, "grad_norm": 4.371885150117092, "learning_rate": 2.122573925794219e-07, "loss": 0.8183, "step": 1718 }, { "epoch": 1.72, "grad_norm": 3.041592050171884, "learning_rate": 2.1065106533260381e-07, "loss": 0.9008, "step": 1719 }, { "epoch": 1.72, "grad_norm": 3.3334030541509274, "learning_rate": 2.0905057201904444e-07, "loss": 0.7654, "step": 1720 }, { "epoch": 1.72, "grad_norm": 5.243704377501293, "learning_rate": 2.0745591671727016e-07, "loss": 0.7433, "step": 1721 }, { "epoch": 1.72, "grad_norm": 5.136183540609473, "learning_rate": 2.0586710349093013e-07, "loss": 0.7753, "step": 1722 }, { "epoch": 1.72, "grad_norm": 2.9655062235941805, "learning_rate": 2.0428413638878764e-07, "loss": 0.7843, "step": 1723 }, { "epoch": 1.72, "grad_norm": 3.326781096783119, "learning_rate": 2.027070194447081e-07, "loss": 0.7902, "step": 1724 }, { "epoch": 1.72, "grad_norm": 3.7559417743969554, "learning_rate": 2.0113575667764756e-07, "loss": 0.785, "step": 1725 }, { "epoch": 1.72, "grad_norm": 2.81831220691812, "learning_rate": 1.9957035209164562e-07, "loss": 0.7688, "step": 1726 }, { "epoch": 1.73, "grad_norm": 3.040059720611323, "learning_rate": 1.9801080967581266e-07, "loss": 0.829, "step": 1727 }, { "epoch": 1.73, "grad_norm": 3.2395774476565404, "learning_rate": 1.9645713340431998e-07, "loss": 0.7778, "step": 1728 }, { "epoch": 1.73, "grad_norm": 4.898251356857874, "learning_rate": 1.9490932723639166e-07, "loss": 0.7537, "step": 1729 }, { "epoch": 1.73, "grad_norm": 3.163832568668124, "learning_rate": 1.9336739511629232e-07, "loss": 0.7774, "step": 1730 }, { "epoch": 1.73, "grad_norm": 4.200716508796324, "learning_rate": 1.918313409733169e-07, "loss": 0.8052, "step": 1731 }, { "epoch": 1.73, "grad_norm": 3.149971049529621, "learning_rate": 1.9030116872178317e-07, "loss": 0.7649, "step": 1732 }, { "epoch": 1.73, "grad_norm": 3.2037076396839814, "learning_rate": 1.887768822610192e-07, "loss": 0.7694, "step": 1733 }, { "epoch": 1.73, "grad_norm": 3.2638535448432138, "learning_rate": 1.8725848547535368e-07, "loss": 0.8179, "step": 1734 }, { "epoch": 1.73, "grad_norm": 3.0250436969757777, "learning_rate": 1.8574598223410873e-07, "loss": 0.8067, "step": 1735 }, { "epoch": 1.73, "grad_norm": 3.0441806393660253, "learning_rate": 1.8423937639158534e-07, "loss": 0.7716, "step": 1736 }, { "epoch": 1.74, "grad_norm": 7.514058371391118, "learning_rate": 1.8273867178705802e-07, "loss": 0.7895, "step": 1737 }, { "epoch": 1.74, "grad_norm": 2.785048866784707, "learning_rate": 1.8124387224476347e-07, "loss": 0.7755, "step": 1738 }, { "epoch": 1.74, "grad_norm": 3.441481749298209, "learning_rate": 1.7975498157388914e-07, "loss": 0.7745, "step": 1739 }, { "epoch": 1.74, "grad_norm": 4.390358042050544, "learning_rate": 1.7827200356856534e-07, "loss": 0.7995, "step": 1740 }, { "epoch": 1.74, "grad_norm": 2.9132339598904626, "learning_rate": 1.76794942007856e-07, "loss": 0.7506, "step": 1741 }, { "epoch": 1.74, "grad_norm": 3.2028721434486296, "learning_rate": 1.753238006557473e-07, "loss": 0.7576, "step": 1742 }, { "epoch": 1.74, "grad_norm": 2.9702972488322232, "learning_rate": 1.738585832611392e-07, "loss": 0.8188, "step": 1743 }, { "epoch": 1.74, "grad_norm": 3.317581287778142, "learning_rate": 1.723992935578367e-07, "loss": 0.8003, "step": 1744 }, { "epoch": 1.74, "grad_norm": 3.3785963887636, "learning_rate": 1.709459352645379e-07, "loss": 0.7963, "step": 1745 }, { "epoch": 1.74, "grad_norm": 4.001001353548268, "learning_rate": 1.6949851208482583e-07, "loss": 0.7396, "step": 1746 }, { "epoch": 1.75, "grad_norm": 6.956619509237825, "learning_rate": 1.6805702770716054e-07, "loss": 0.7985, "step": 1747 }, { "epoch": 1.75, "grad_norm": 3.234626914455938, "learning_rate": 1.6662148580486704e-07, "loss": 0.7634, "step": 1748 }, { "epoch": 1.75, "grad_norm": 3.756890651331269, "learning_rate": 1.651918900361277e-07, "loss": 0.7905, "step": 1749 }, { "epoch": 1.75, "grad_norm": 3.148258125023457, "learning_rate": 1.6376824404397252e-07, "loss": 0.7636, "step": 1750 }, { "epoch": 1.75, "grad_norm": 2.900501285520038, "learning_rate": 1.6235055145626955e-07, "loss": 0.7507, "step": 1751 }, { "epoch": 1.75, "grad_norm": 3.212475854355331, "learning_rate": 1.6093881588571504e-07, "loss": 0.7974, "step": 1752 }, { "epoch": 1.75, "grad_norm": 3.1994271035653794, "learning_rate": 1.5953304092982625e-07, "loss": 0.7583, "step": 1753 }, { "epoch": 1.75, "grad_norm": 3.1114214116598746, "learning_rate": 1.5813323017093042e-07, "loss": 0.7557, "step": 1754 }, { "epoch": 1.75, "grad_norm": 2.894989843567494, "learning_rate": 1.5673938717615605e-07, "loss": 0.7662, "step": 1755 }, { "epoch": 1.75, "grad_norm": 3.471435833155311, "learning_rate": 1.5535151549742527e-07, "loss": 0.8084, "step": 1756 }, { "epoch": 1.76, "grad_norm": 3.3175253832425433, "learning_rate": 1.5396961867144205e-07, "loss": 0.7567, "step": 1757 }, { "epoch": 1.76, "grad_norm": 3.35191564761709, "learning_rate": 1.525937002196845e-07, "loss": 0.8303, "step": 1758 }, { "epoch": 1.76, "grad_norm": 3.4495898694179705, "learning_rate": 1.5122376364839818e-07, "loss": 0.7795, "step": 1759 }, { "epoch": 1.76, "grad_norm": 3.078280729655168, "learning_rate": 1.4985981244858256e-07, "loss": 0.7416, "step": 1760 }, { "epoch": 1.76, "grad_norm": 3.251770807046257, "learning_rate": 1.4850185009598644e-07, "loss": 0.807, "step": 1761 }, { "epoch": 1.76, "grad_norm": 2.9938130892811463, "learning_rate": 1.471498800510962e-07, "loss": 0.78, "step": 1762 }, { "epoch": 1.76, "grad_norm": 3.3751507949055575, "learning_rate": 1.4580390575912874e-07, "loss": 0.7783, "step": 1763 }, { "epoch": 1.76, "grad_norm": 3.3041604542980663, "learning_rate": 1.4446393065002145e-07, "loss": 0.7871, "step": 1764 }, { "epoch": 1.76, "grad_norm": 3.4442148181169316, "learning_rate": 1.431299581384249e-07, "loss": 0.7624, "step": 1765 }, { "epoch": 1.76, "grad_norm": 3.1625848068155875, "learning_rate": 1.4180199162369207e-07, "loss": 0.7749, "step": 1766 }, { "epoch": 1.77, "grad_norm": 8.321954119777603, "learning_rate": 1.4048003448987212e-07, "loss": 0.7873, "step": 1767 }, { "epoch": 1.77, "grad_norm": 3.195932262992761, "learning_rate": 1.3916409010569925e-07, "loss": 0.776, "step": 1768 }, { "epoch": 1.77, "grad_norm": 3.7128812237319906, "learning_rate": 1.3785416182458712e-07, "loss": 0.9338, "step": 1769 }, { "epoch": 1.77, "grad_norm": 3.2169070835383313, "learning_rate": 1.3655025298461662e-07, "loss": 0.7125, "step": 1770 }, { "epoch": 1.77, "grad_norm": 3.2305937541880763, "learning_rate": 1.3525236690853093e-07, "loss": 0.7762, "step": 1771 }, { "epoch": 1.77, "grad_norm": 3.813602393235758, "learning_rate": 1.339605069037242e-07, "loss": 0.8267, "step": 1772 }, { "epoch": 1.77, "grad_norm": 3.489306104879544, "learning_rate": 1.3267467626223606e-07, "loss": 0.8159, "step": 1773 }, { "epoch": 1.77, "grad_norm": 3.2100951969245406, "learning_rate": 1.3139487826073937e-07, "loss": 0.7648, "step": 1774 }, { "epoch": 1.77, "grad_norm": 3.171267821383128, "learning_rate": 1.301211161605362e-07, "loss": 0.7684, "step": 1775 }, { "epoch": 1.77, "grad_norm": 3.3201626677791443, "learning_rate": 1.288533932075453e-07, "loss": 0.7935, "step": 1776 }, { "epoch": 1.78, "grad_norm": 3.7527822708606076, "learning_rate": 1.2759171263229814e-07, "loss": 0.79, "step": 1777 }, { "epoch": 1.78, "grad_norm": 3.1570440408944807, "learning_rate": 1.2633607764992673e-07, "loss": 0.7545, "step": 1778 }, { "epoch": 1.78, "grad_norm": 3.620294350313225, "learning_rate": 1.250864914601571e-07, "loss": 0.8017, "step": 1779 }, { "epoch": 1.78, "grad_norm": 3.4579772349445133, "learning_rate": 1.2384295724730266e-07, "loss": 0.9085, "step": 1780 }, { "epoch": 1.78, "grad_norm": 3.3747730659622297, "learning_rate": 1.2260547818025326e-07, "loss": 0.8328, "step": 1781 }, { "epoch": 1.78, "grad_norm": 3.2329947570868756, "learning_rate": 1.2137405741246916e-07, "loss": 0.7801, "step": 1782 }, { "epoch": 1.78, "grad_norm": 3.346114332182655, "learning_rate": 1.201486980819716e-07, "loss": 0.7973, "step": 1783 }, { "epoch": 1.78, "grad_norm": 3.7623379607899157, "learning_rate": 1.1892940331133612e-07, "loss": 0.7132, "step": 1784 }, { "epoch": 1.78, "grad_norm": 3.66956091328404, "learning_rate": 1.1771617620768394e-07, "loss": 0.8146, "step": 1785 }, { "epoch": 1.79, "grad_norm": 3.5408958111016746, "learning_rate": 1.1650901986267365e-07, "loss": 0.7359, "step": 1786 }, { "epoch": 1.79, "grad_norm": 6.414472725550592, "learning_rate": 1.1530793735249457e-07, "loss": 0.8086, "step": 1787 }, { "epoch": 1.79, "grad_norm": 3.394428137117122, "learning_rate": 1.1411293173785725e-07, "loss": 0.7716, "step": 1788 }, { "epoch": 1.79, "grad_norm": 2.9270356575921177, "learning_rate": 1.1292400606398635e-07, "loss": 0.7903, "step": 1789 }, { "epoch": 1.79, "grad_norm": 3.104550712124825, "learning_rate": 1.1174116336061469e-07, "loss": 0.8024, "step": 1790 }, { "epoch": 1.79, "grad_norm": 3.518556658371052, "learning_rate": 1.1056440664197143e-07, "loss": 0.7625, "step": 1791 }, { "epoch": 1.79, "grad_norm": 3.1054550549446183, "learning_rate": 1.0939373890677923e-07, "loss": 0.7517, "step": 1792 }, { "epoch": 1.79, "grad_norm": 3.3517057095864646, "learning_rate": 1.0822916313824317e-07, "loss": 0.7931, "step": 1793 }, { "epoch": 1.79, "grad_norm": 3.1664321885491766, "learning_rate": 1.0707068230404406e-07, "loss": 0.7366, "step": 1794 }, { "epoch": 1.79, "grad_norm": 3.353006236344301, "learning_rate": 1.0591829935633041e-07, "loss": 0.789, "step": 1795 }, { "epoch": 1.8, "grad_norm": 3.18594945849499, "learning_rate": 1.0477201723171377e-07, "loss": 0.816, "step": 1796 }, { "epoch": 1.8, "grad_norm": 3.161208136077106, "learning_rate": 1.036318388512561e-07, "loss": 0.8232, "step": 1797 }, { "epoch": 1.8, "grad_norm": 3.28597912059524, "learning_rate": 1.0249776712046745e-07, "loss": 0.7734, "step": 1798 }, { "epoch": 1.8, "grad_norm": 3.4084255254881657, "learning_rate": 1.0136980492929605e-07, "loss": 0.799, "step": 1799 }, { "epoch": 1.8, "grad_norm": 26.572221899801438, "learning_rate": 1.0024795515211989e-07, "loss": 0.7612, "step": 1800 }, { "epoch": 1.8, "grad_norm": 4.530124673582639, "learning_rate": 9.913222064774159e-08, "loss": 0.7616, "step": 1801 }, { "epoch": 1.8, "grad_norm": 3.3426378259214236, "learning_rate": 9.802260425938098e-08, "loss": 0.8112, "step": 1802 }, { "epoch": 1.8, "grad_norm": 3.068279032461048, "learning_rate": 9.691910881466565e-08, "loss": 0.7699, "step": 1803 }, { "epoch": 1.8, "grad_norm": 2.996408317688526, "learning_rate": 9.58217371256262e-08, "loss": 0.7396, "step": 1804 }, { "epoch": 1.8, "grad_norm": 2.94305532213708, "learning_rate": 9.473049198868822e-08, "loss": 0.7454, "step": 1805 }, { "epoch": 1.81, "grad_norm": 3.241301363252064, "learning_rate": 9.364537618466451e-08, "loss": 0.8254, "step": 1806 }, { "epoch": 1.81, "grad_norm": 2.8699192922977077, "learning_rate": 9.256639247874872e-08, "loss": 0.8174, "step": 1807 }, { "epoch": 1.81, "grad_norm": 3.4930627019454956, "learning_rate": 9.149354362050805e-08, "loss": 0.805, "step": 1808 }, { "epoch": 1.81, "grad_norm": 3.185989103506399, "learning_rate": 9.042683234387644e-08, "loss": 0.7769, "step": 1809 }, { "epoch": 1.81, "grad_norm": 3.493811500457918, "learning_rate": 8.936626136714754e-08, "loss": 0.8029, "step": 1810 }, { "epoch": 1.81, "grad_norm": 3.511898652789175, "learning_rate": 8.831183339296751e-08, "loss": 0.74, "step": 1811 }, { "epoch": 1.81, "grad_norm": 3.407883231848714, "learning_rate": 8.726355110832862e-08, "loss": 0.8093, "step": 1812 }, { "epoch": 1.81, "grad_norm": 3.3486631139510794, "learning_rate": 8.622141718456128e-08, "loss": 0.7474, "step": 1813 }, { "epoch": 1.81, "grad_norm": 3.4185955272488324, "learning_rate": 8.518543427732951e-08, "loss": 0.7735, "step": 1814 }, { "epoch": 1.81, "grad_norm": 4.032268887360276, "learning_rate": 8.415560502662152e-08, "loss": 0.7573, "step": 1815 }, { "epoch": 1.82, "grad_norm": 2.8813525615229802, "learning_rate": 8.313193205674391e-08, "loss": 0.7664, "step": 1816 }, { "epoch": 1.82, "grad_norm": 3.3951188618625587, "learning_rate": 8.211441797631753e-08, "loss": 0.7673, "step": 1817 }, { "epoch": 1.82, "grad_norm": 3.734588248321118, "learning_rate": 8.110306537826601e-08, "loss": 0.8027, "step": 1818 }, { "epoch": 1.82, "grad_norm": 3.1485524680422396, "learning_rate": 8.009787683981279e-08, "loss": 0.7993, "step": 1819 }, { "epoch": 1.82, "grad_norm": 3.3505148607359287, "learning_rate": 7.90988549224736e-08, "loss": 0.7766, "step": 1820 }, { "epoch": 1.82, "grad_norm": 3.028772808744086, "learning_rate": 7.81060021720495e-08, "loss": 0.7159, "step": 1821 }, { "epoch": 1.82, "grad_norm": 3.4345419750197532, "learning_rate": 7.711932111862024e-08, "loss": 0.7276, "step": 1822 }, { "epoch": 1.82, "grad_norm": 3.505240349744336, "learning_rate": 7.613881427654013e-08, "loss": 0.7905, "step": 1823 }, { "epoch": 1.82, "grad_norm": 3.429660566103403, "learning_rate": 7.51644841444274e-08, "loss": 0.8104, "step": 1824 }, { "epoch": 1.82, "grad_norm": 3.329014343734893, "learning_rate": 7.419633320516178e-08, "loss": 0.7955, "step": 1825 }, { "epoch": 1.83, "grad_norm": 3.16204906840086, "learning_rate": 7.323436392587591e-08, "loss": 0.7493, "step": 1826 }, { "epoch": 1.83, "grad_norm": 3.1351282660444486, "learning_rate": 7.227857875795024e-08, "loss": 0.7962, "step": 1827 }, { "epoch": 1.83, "grad_norm": 3.160679644072152, "learning_rate": 7.13289801370054e-08, "loss": 0.7651, "step": 1828 }, { "epoch": 1.83, "grad_norm": 3.832074131827653, "learning_rate": 7.038557048289818e-08, "loss": 0.8074, "step": 1829 }, { "epoch": 1.83, "grad_norm": 3.283494956152082, "learning_rate": 6.94483521997133e-08, "loss": 0.7529, "step": 1830 }, { "epoch": 1.83, "grad_norm": 7.290869517669055, "learning_rate": 6.851732767575752e-08, "loss": 0.7691, "step": 1831 }, { "epoch": 1.83, "grad_norm": 3.29750292787458, "learning_rate": 6.759249928355554e-08, "loss": 0.7621, "step": 1832 }, { "epoch": 1.83, "grad_norm": 3.7810322285900204, "learning_rate": 6.667386937984105e-08, "loss": 0.8088, "step": 1833 }, { "epoch": 1.83, "grad_norm": 3.170898616083802, "learning_rate": 6.576144030555259e-08, "loss": 0.7767, "step": 1834 }, { "epoch": 1.83, "grad_norm": 3.11329409919027, "learning_rate": 6.485521438582748e-08, "loss": 0.8128, "step": 1835 }, { "epoch": 1.84, "grad_norm": 3.3919917656310368, "learning_rate": 6.395519392999622e-08, "loss": 0.7492, "step": 1836 }, { "epoch": 1.84, "grad_norm": 3.3591959295112, "learning_rate": 6.30613812315739e-08, "loss": 0.8103, "step": 1837 }, { "epoch": 1.84, "grad_norm": 3.420990761337526, "learning_rate": 6.217377856825885e-08, "loss": 0.72, "step": 1838 }, { "epoch": 1.84, "grad_norm": 3.1297055647085283, "learning_rate": 6.129238820192284e-08, "loss": 0.7939, "step": 1839 }, { "epoch": 1.84, "grad_norm": 3.03968567640584, "learning_rate": 6.041721237860677e-08, "loss": 0.7543, "step": 1840 }, { "epoch": 1.84, "grad_norm": 3.732525128633042, "learning_rate": 5.9548253328516325e-08, "loss": 0.7241, "step": 1841 }, { "epoch": 1.84, "grad_norm": 3.0114242220216476, "learning_rate": 5.8685513266014125e-08, "loss": 0.7867, "step": 1842 }, { "epoch": 1.84, "grad_norm": 3.3849547694814826, "learning_rate": 5.7828994389614866e-08, "loss": 0.7239, "step": 1843 }, { "epoch": 1.84, "grad_norm": 3.119888413852809, "learning_rate": 5.697869888198065e-08, "loss": 0.7704, "step": 1844 }, { "epoch": 1.84, "grad_norm": 4.841012122130986, "learning_rate": 5.613462890991378e-08, "loss": 0.8012, "step": 1845 }, { "epoch": 1.85, "grad_norm": 3.325707190516144, "learning_rate": 5.529678662435228e-08, "loss": 0.7864, "step": 1846 }, { "epoch": 1.85, "grad_norm": 3.815849198305633, "learning_rate": 5.446517416036412e-08, "loss": 0.7967, "step": 1847 }, { "epoch": 1.85, "grad_norm": 3.20970092255699, "learning_rate": 5.3639793637142444e-08, "loss": 0.8, "step": 1848 }, { "epoch": 1.85, "grad_norm": 3.7739213780020924, "learning_rate": 5.282064715799895e-08, "loss": 0.7541, "step": 1849 }, { "epoch": 1.85, "grad_norm": 2.9529234011097363, "learning_rate": 5.2007736810359696e-08, "loss": 0.7686, "step": 1850 }, { "epoch": 1.85, "grad_norm": 3.409723892856057, "learning_rate": 5.120106466575875e-08, "loss": 0.8026, "step": 1851 }, { "epoch": 1.85, "grad_norm": 5.216458396500128, "learning_rate": 5.040063277983287e-08, "loss": 0.7952, "step": 1852 }, { "epoch": 1.85, "grad_norm": 3.1939652123410434, "learning_rate": 4.96064431923185e-08, "loss": 0.7826, "step": 1853 }, { "epoch": 1.85, "grad_norm": 3.394699587086276, "learning_rate": 4.881849792704368e-08, "loss": 0.7434, "step": 1854 }, { "epoch": 1.85, "grad_norm": 3.2722822816212003, "learning_rate": 4.8036798991923925e-08, "loss": 0.7726, "step": 1855 }, { "epoch": 1.86, "grad_norm": 3.486380698478845, "learning_rate": 4.7261348378958015e-08, "loss": 0.8109, "step": 1856 }, { "epoch": 1.86, "grad_norm": 3.1504822046939385, "learning_rate": 4.6492148064221646e-08, "loss": 0.794, "step": 1857 }, { "epoch": 1.86, "grad_norm": 3.1659637164728562, "learning_rate": 4.5729200007862686e-08, "loss": 0.7392, "step": 1858 }, { "epoch": 1.86, "grad_norm": 3.6802148910899084, "learning_rate": 4.497250615409732e-08, "loss": 0.9253, "step": 1859 }, { "epoch": 1.86, "grad_norm": 3.02270568540266, "learning_rate": 4.4222068431203634e-08, "loss": 0.8081, "step": 1860 }, { "epoch": 1.86, "grad_norm": 3.4414867265704565, "learning_rate": 4.34778887515172e-08, "loss": 0.7825, "step": 1861 }, { "epoch": 1.86, "grad_norm": 3.0772904517305704, "learning_rate": 4.2739969011426075e-08, "loss": 0.7433, "step": 1862 }, { "epoch": 1.86, "grad_norm": 2.9226473470834824, "learning_rate": 4.20083110913666e-08, "loss": 0.7501, "step": 1863 }, { "epoch": 1.86, "grad_norm": 4.223282908145395, "learning_rate": 4.128291685581792e-08, "loss": 0.8673, "step": 1864 }, { "epoch": 1.86, "grad_norm": 3.4940054443050363, "learning_rate": 4.0563788153297754e-08, "loss": 0.9213, "step": 1865 }, { "epoch": 1.87, "grad_norm": 3.3519572326966203, "learning_rate": 3.985092681635716e-08, "loss": 0.7648, "step": 1866 }, { "epoch": 1.87, "grad_norm": 3.294794217383537, "learning_rate": 3.914433466157608e-08, "loss": 0.8346, "step": 1867 }, { "epoch": 1.87, "grad_norm": 3.6989873248638774, "learning_rate": 3.844401348955834e-08, "loss": 0.8134, "step": 1868 }, { "epoch": 1.87, "grad_norm": 3.4825757951742617, "learning_rate": 3.7749965084928343e-08, "loss": 0.7579, "step": 1869 }, { "epoch": 1.87, "grad_norm": 3.2165836370120995, "learning_rate": 3.70621912163252e-08, "loss": 0.7566, "step": 1870 }, { "epoch": 1.87, "grad_norm": 3.111470767148195, "learning_rate": 3.6380693636398344e-08, "loss": 0.7684, "step": 1871 }, { "epoch": 1.87, "grad_norm": 3.183066596928477, "learning_rate": 3.570547408180442e-08, "loss": 0.8123, "step": 1872 }, { "epoch": 1.87, "grad_norm": 3.611941155804037, "learning_rate": 3.503653427320036e-08, "loss": 0.803, "step": 1873 }, { "epoch": 1.87, "grad_norm": 2.983125424995353, "learning_rate": 3.43738759152415e-08, "loss": 0.7564, "step": 1874 }, { "epoch": 1.87, "grad_norm": 3.211530878795052, "learning_rate": 3.3717500696575925e-08, "loss": 0.7712, "step": 1875 }, { "epoch": 1.88, "grad_norm": 3.2336625858563606, "learning_rate": 3.3067410289840115e-08, "loss": 0.7291, "step": 1876 }, { "epoch": 1.88, "grad_norm": 3.487794967952788, "learning_rate": 3.242360635165559e-08, "loss": 0.7766, "step": 1877 }, { "epoch": 1.88, "grad_norm": 4.6889663038700045, "learning_rate": 3.178609052262416e-08, "loss": 0.7749, "step": 1878 }, { "epoch": 1.88, "grad_norm": 3.257376560387505, "learning_rate": 3.115486442732268e-08, "loss": 0.7588, "step": 1879 }, { "epoch": 1.88, "grad_norm": 3.1183646738475046, "learning_rate": 3.052992967430085e-08, "loss": 0.7783, "step": 1880 }, { "epoch": 1.88, "grad_norm": 3.610808949460703, "learning_rate": 2.991128785607589e-08, "loss": 0.7642, "step": 1881 }, { "epoch": 1.88, "grad_norm": 5.289360971888713, "learning_rate": 2.9298940549128962e-08, "loss": 0.8222, "step": 1882 }, { "epoch": 1.88, "grad_norm": 3.2387446481069326, "learning_rate": 2.8692889313900185e-08, "loss": 0.7605, "step": 1883 }, { "epoch": 1.88, "grad_norm": 3.8815106435993294, "learning_rate": 2.8093135694786665e-08, "loss": 0.8041, "step": 1884 }, { "epoch": 1.89, "grad_norm": 4.004500860537093, "learning_rate": 2.7499681220136685e-08, "loss": 0.7681, "step": 1885 }, { "epoch": 1.89, "grad_norm": 3.2389631692834495, "learning_rate": 2.691252740224637e-08, "loss": 0.7973, "step": 1886 }, { "epoch": 1.89, "grad_norm": 3.502974765335723, "learning_rate": 2.633167573735579e-08, "loss": 0.7575, "step": 1887 }, { "epoch": 1.89, "grad_norm": 3.711666292718203, "learning_rate": 2.5757127705645924e-08, "loss": 0.7262, "step": 1888 }, { "epoch": 1.89, "grad_norm": 3.152696715676429, "learning_rate": 2.5188884771233657e-08, "loss": 0.8026, "step": 1889 }, { "epoch": 1.89, "grad_norm": 3.2757427323745514, "learning_rate": 2.4626948382168725e-08, "loss": 0.9247, "step": 1890 }, { "epoch": 1.89, "grad_norm": 3.5058517037876378, "learning_rate": 2.4071319970430385e-08, "loss": 0.756, "step": 1891 }, { "epoch": 1.89, "grad_norm": 5.019449958573384, "learning_rate": 2.352200095192242e-08, "loss": 0.7745, "step": 1892 }, { "epoch": 1.89, "grad_norm": 3.3388714116634115, "learning_rate": 2.2978992726471748e-08, "loss": 0.7987, "step": 1893 }, { "epoch": 1.89, "grad_norm": 2.8866415372773973, "learning_rate": 2.244229667782205e-08, "loss": 0.7704, "step": 1894 }, { "epoch": 1.9, "grad_norm": 3.2727749773182047, "learning_rate": 2.1911914173632642e-08, "loss": 0.7711, "step": 1895 }, { "epoch": 1.9, "grad_norm": 3.650767190321931, "learning_rate": 2.1387846565474047e-08, "loss": 0.7762, "step": 1896 }, { "epoch": 1.9, "grad_norm": 3.112961907966241, "learning_rate": 2.0870095188824103e-08, "loss": 0.7862, "step": 1897 }, { "epoch": 1.9, "grad_norm": 5.164969648329022, "learning_rate": 2.0358661363065746e-08, "loss": 0.7832, "step": 1898 }, { "epoch": 1.9, "grad_norm": 3.854270667774517, "learning_rate": 1.985354639148229e-08, "loss": 0.7705, "step": 1899 }, { "epoch": 1.9, "grad_norm": 4.262263260220136, "learning_rate": 1.9354751561254937e-08, "loss": 0.8125, "step": 1900 }, { "epoch": 1.9, "grad_norm": 3.275946878058362, "learning_rate": 1.8862278143459144e-08, "loss": 0.814, "step": 1901 }, { "epoch": 1.9, "grad_norm": 4.395005014767785, "learning_rate": 1.837612739306216e-08, "loss": 0.8078, "step": 1902 }, { "epoch": 1.9, "grad_norm": 7.644228688749442, "learning_rate": 1.789630054891883e-08, "loss": 0.7908, "step": 1903 }, { "epoch": 1.9, "grad_norm": 3.886400512022099, "learning_rate": 1.7422798833768572e-08, "loss": 0.9245, "step": 1904 }, { "epoch": 1.91, "grad_norm": 3.7528813262415373, "learning_rate": 1.6955623454233126e-08, "loss": 0.793, "step": 1905 }, { "epoch": 1.91, "grad_norm": 3.451703776509019, "learning_rate": 1.6494775600812418e-08, "loss": 0.7775, "step": 1906 }, { "epoch": 1.91, "grad_norm": 5.24226190201148, "learning_rate": 1.6040256447881762e-08, "loss": 0.8662, "step": 1907 }, { "epoch": 1.91, "grad_norm": 3.1920740668552257, "learning_rate": 1.559206715368966e-08, "loss": 0.7943, "step": 1908 }, { "epoch": 1.91, "grad_norm": 3.21792399839096, "learning_rate": 1.5150208860354175e-08, "loss": 0.7451, "step": 1909 }, { "epoch": 1.91, "grad_norm": 3.1707641635083497, "learning_rate": 1.4714682693859616e-08, "loss": 0.8093, "step": 1910 }, { "epoch": 1.91, "grad_norm": 4.016699382942119, "learning_rate": 1.428548976405486e-08, "loss": 0.7941, "step": 1911 }, { "epoch": 1.91, "grad_norm": 2.998601999306543, "learning_rate": 1.3862631164649477e-08, "loss": 0.7651, "step": 1912 }, { "epoch": 1.91, "grad_norm": 3.0577852946508552, "learning_rate": 1.344610797321122e-08, "loss": 0.9008, "step": 1913 }, { "epoch": 1.91, "grad_norm": 3.215420465717638, "learning_rate": 1.3035921251163263e-08, "loss": 0.7797, "step": 1914 }, { "epoch": 1.92, "grad_norm": 3.0738955094985845, "learning_rate": 1.2632072043782251e-08, "loss": 0.7963, "step": 1915 }, { "epoch": 1.92, "grad_norm": 3.1919963331144316, "learning_rate": 1.2234561380194132e-08, "loss": 0.7962, "step": 1916 }, { "epoch": 1.92, "grad_norm": 3.4072512125608525, "learning_rate": 1.1843390273373057e-08, "loss": 0.7763, "step": 1917 }, { "epoch": 1.92, "grad_norm": 6.136288886754454, "learning_rate": 1.1458559720137762e-08, "loss": 0.7945, "step": 1918 }, { "epoch": 1.92, "grad_norm": 2.967785582398141, "learning_rate": 1.108007070114936e-08, "loss": 0.7987, "step": 1919 }, { "epoch": 1.92, "grad_norm": 3.7713110999332975, "learning_rate": 1.070792418090938e-08, "loss": 0.7341, "step": 1920 }, { "epoch": 1.92, "grad_norm": 3.1290456318993978, "learning_rate": 1.0342121107755898e-08, "loss": 0.7416, "step": 1921 }, { "epoch": 1.92, "grad_norm": 2.8709100155475307, "learning_rate": 9.982662413862975e-09, "loss": 0.7411, "step": 1922 }, { "epoch": 1.92, "grad_norm": 3.440644563645032, "learning_rate": 9.62954901523705e-09, "loss": 0.8039, "step": 1923 }, { "epoch": 1.92, "grad_norm": 3.293664607433898, "learning_rate": 9.282781811714159e-09, "loss": 0.7512, "step": 1924 }, { "epoch": 1.93, "grad_norm": 3.5102912917143776, "learning_rate": 8.942361686959112e-09, "loss": 0.8055, "step": 1925 }, { "epoch": 1.93, "grad_norm": 3.076810107113977, "learning_rate": 8.608289508462708e-09, "loss": 0.8269, "step": 1926 }, { "epoch": 1.93, "grad_norm": 3.592767693073543, "learning_rate": 8.280566127538691e-09, "loss": 0.8256, "step": 1927 }, { "epoch": 1.93, "grad_norm": 3.7130084916245982, "learning_rate": 7.959192379322078e-09, "loss": 0.8043, "step": 1928 }, { "epoch": 1.93, "grad_norm": 3.745663007958313, "learning_rate": 7.644169082768327e-09, "loss": 0.7642, "step": 1929 }, { "epoch": 1.93, "grad_norm": 2.9818149028139387, "learning_rate": 7.335497040648898e-09, "loss": 0.8304, "step": 1930 }, { "epoch": 1.93, "grad_norm": 3.574641840905873, "learning_rate": 7.033177039550698e-09, "loss": 0.817, "step": 1931 }, { "epoch": 1.93, "grad_norm": 3.1328224981340935, "learning_rate": 6.73720984987386e-09, "loss": 0.827, "step": 1932 }, { "epoch": 1.93, "grad_norm": 2.961568026621921, "learning_rate": 6.4475962258297994e-09, "loss": 0.8012, "step": 1933 }, { "epoch": 1.93, "grad_norm": 2.9917340909933428, "learning_rate": 6.164336905438994e-09, "loss": 0.7851, "step": 1934 }, { "epoch": 1.94, "grad_norm": 7.596539024844979, "learning_rate": 5.88743261052932e-09, "loss": 0.7311, "step": 1935 }, { "epoch": 1.94, "grad_norm": 3.066116656958356, "learning_rate": 5.616884046734383e-09, "loss": 0.7181, "step": 1936 }, { "epoch": 1.94, "grad_norm": 3.0083960893722286, "learning_rate": 5.352691903491303e-09, "loss": 0.7731, "step": 1937 }, { "epoch": 1.94, "grad_norm": 3.4022412128960147, "learning_rate": 5.094856854039043e-09, "loss": 0.7899, "step": 1938 }, { "epoch": 1.94, "grad_norm": 3.2103108603991792, "learning_rate": 4.843379555417304e-09, "loss": 0.7363, "step": 1939 }, { "epoch": 1.94, "grad_norm": 3.765239107302818, "learning_rate": 4.598260648463748e-09, "loss": 0.7606, "step": 1940 }, { "epoch": 1.94, "grad_norm": 3.087778448876169, "learning_rate": 4.359500757813717e-09, "loss": 0.8729, "step": 1941 }, { "epoch": 1.94, "grad_norm": 2.9274838102617804, "learning_rate": 4.1271004918971845e-09, "loss": 0.738, "step": 1942 }, { "epoch": 1.94, "grad_norm": 3.3609036336599067, "learning_rate": 3.9010604429381985e-09, "loss": 0.8309, "step": 1943 }, { "epoch": 1.94, "grad_norm": 3.427665373219867, "learning_rate": 3.6813811869537698e-09, "loss": 0.775, "step": 1944 }, { "epoch": 1.95, "grad_norm": 3.146336635296433, "learning_rate": 3.468063283750267e-09, "loss": 0.7714, "step": 1945 }, { "epoch": 1.95, "grad_norm": 3.6012042097770864, "learning_rate": 3.2611072769250797e-09, "loss": 1.0305, "step": 1946 }, { "epoch": 1.95, "grad_norm": 3.2770839629650546, "learning_rate": 3.060513693862455e-09, "loss": 0.7582, "step": 1947 }, { "epoch": 1.95, "grad_norm": 3.2353210793243137, "learning_rate": 2.866283045734053e-09, "loss": 0.7801, "step": 1948 }, { "epoch": 1.95, "grad_norm": 3.2593351635542973, "learning_rate": 2.67841582749645e-09, "loss": 0.7996, "step": 1949 }, { "epoch": 1.95, "grad_norm": 3.522716363873415, "learning_rate": 2.4969125178903043e-09, "loss": 0.8641, "step": 1950 }, { "epoch": 1.95, "grad_norm": 3.6131553308723285, "learning_rate": 2.321773579439246e-09, "loss": 0.8131, "step": 1951 }, { "epoch": 1.95, "grad_norm": 3.389236555118622, "learning_rate": 2.152999458449323e-09, "loss": 0.7608, "step": 1952 }, { "epoch": 1.95, "grad_norm": 3.6213987478343173, "learning_rate": 1.990590585005947e-09, "loss": 0.8152, "step": 1953 }, { "epoch": 1.95, "grad_norm": 3.202690410119613, "learning_rate": 1.834547372975004e-09, "loss": 0.7977, "step": 1954 }, { "epoch": 1.96, "grad_norm": 3.288623021899843, "learning_rate": 1.6848702200000787e-09, "loss": 0.7834, "step": 1955 }, { "epoch": 1.96, "grad_norm": 2.991531777248238, "learning_rate": 1.5415595075027323e-09, "loss": 0.7423, "step": 1956 }, { "epoch": 1.96, "grad_norm": 3.7577918290988905, "learning_rate": 1.4046156006808365e-09, "loss": 0.9146, "step": 1957 }, { "epoch": 1.96, "grad_norm": 3.500747930732869, "learning_rate": 1.2740388485071864e-09, "loss": 0.7524, "step": 1958 }, { "epoch": 1.96, "grad_norm": 3.459655675099245, "learning_rate": 1.149829583730333e-09, "loss": 0.8232, "step": 1959 }, { "epoch": 1.96, "grad_norm": 3.0513435382323695, "learning_rate": 1.03198812287153e-09, "loss": 0.8061, "step": 1960 }, { "epoch": 1.96, "grad_norm": 3.548539947690588, "learning_rate": 9.205147662252889e-10, "loss": 0.7549, "step": 1961 }, { "epoch": 1.96, "grad_norm": 3.2934774940065066, "learning_rate": 8.154097978591014e-10, "loss": 0.8063, "step": 1962 }, { "epoch": 1.96, "grad_norm": 3.1609951740282485, "learning_rate": 7.166734856103863e-10, "loss": 0.7679, "step": 1963 }, { "epoch": 1.96, "grad_norm": 3.3709508885001056, "learning_rate": 6.243060810892653e-10, "loss": 0.7704, "step": 1964 }, { "epoch": 1.97, "grad_norm": 3.098953177233062, "learning_rate": 5.38307819674122e-10, "loss": 0.7744, "step": 1965 }, { "epoch": 1.97, "grad_norm": 3.149077360815716, "learning_rate": 4.586789205140996e-10, "loss": 0.8039, "step": 1966 }, { "epoch": 1.97, "grad_norm": 3.419136409086045, "learning_rate": 3.854195865271582e-10, "loss": 0.9447, "step": 1967 }, { "epoch": 1.97, "grad_norm": 4.800326597047391, "learning_rate": 3.1853000439951987e-10, "loss": 0.8073, "step": 1968 }, { "epoch": 1.97, "grad_norm": 2.974519221341341, "learning_rate": 2.58010344585391e-10, "loss": 0.7247, "step": 1969 }, { "epoch": 1.97, "grad_norm": 3.6992889534822098, "learning_rate": 2.0386076130668453e-10, "loss": 0.7731, "step": 1970 }, { "epoch": 1.97, "grad_norm": 3.531552897523052, "learning_rate": 1.5608139255246514e-10, "loss": 0.7221, "step": 1971 }, { "epoch": 1.97, "grad_norm": 3.0165130197219754, "learning_rate": 1.1467236007867144e-10, "loss": 0.7545, "step": 1972 }, { "epoch": 1.97, "grad_norm": 3.6280436074985842, "learning_rate": 7.963376940728352e-11, "loss": 0.7784, "step": 1973 }, { "epoch": 1.97, "grad_norm": 3.274422287510792, "learning_rate": 5.0965709827432984e-11, "loss": 0.7814, "step": 1974 }, { "epoch": 1.98, "grad_norm": 3.3675801249210964, "learning_rate": 2.8668254393460216e-11, "loss": 0.7672, "step": 1975 }, { "epoch": 1.98, "grad_norm": 3.261376186501408, "learning_rate": 1.2741459925746936e-11, "loss": 0.7935, "step": 1976 }, { "epoch": 1.98, "grad_norm": 3.355252994634647, "learning_rate": 3.185367010716256e-12, "loss": 0.817, "step": 1977 }, { "epoch": 1.98, "grad_norm": 3.5548050258769, "learning_rate": 0.0, "loss": 0.7676, "step": 1978 } ], "logging_steps": 1, "max_steps": 1978, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 989, "total_flos": 3171453260267520.0, "train_batch_size": 4, "trial_name": null, "trial_params": null }