|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.9981341625944022, |
|
"global_step": 4218, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 3.5294117647058825e-06, |
|
"loss": 3.8713, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 7.058823529411765e-06, |
|
"loss": 3.4478, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.0588235294117648e-05, |
|
"loss": 3.2108, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.411764705882353e-05, |
|
"loss": 2.9405, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.7647058823529414e-05, |
|
"loss": 2.6643, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.1176470588235296e-05, |
|
"loss": 2.4552, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.4705882352941178e-05, |
|
"loss": 2.4638, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 2.823529411764706e-05, |
|
"loss": 2.182, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.176470588235294e-05, |
|
"loss": 2.0387, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.529411764705883e-05, |
|
"loss": 1.9353, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.882352941176471e-05, |
|
"loss": 1.9432, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.235294117647059e-05, |
|
"loss": 1.8205, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.588235294117647e-05, |
|
"loss": 1.7445, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.9411764705882355e-05, |
|
"loss": 1.638, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 5.294117647058824e-05, |
|
"loss": 1.6144, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 5.647058823529412e-05, |
|
"loss": 1.597, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 6e-05, |
|
"loss": 1.5933, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 6.352941176470588e-05, |
|
"loss": 1.6111, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 6.705882352941176e-05, |
|
"loss": 1.5282, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 7.058823529411765e-05, |
|
"loss": 1.499, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 7.411764705882354e-05, |
|
"loss": 1.4422, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 7.764705882352942e-05, |
|
"loss": 1.4665, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.11764705882353e-05, |
|
"loss": 1.4113, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.470588235294118e-05, |
|
"loss": 1.3772, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.823529411764706e-05, |
|
"loss": 1.3554, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.176470588235295e-05, |
|
"loss": 1.3363, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.529411764705883e-05, |
|
"loss": 1.3221, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.882352941176471e-05, |
|
"loss": 1.231, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.999994222115565e-05, |
|
"loss": 1.2604, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.999963888258794e-05, |
|
"loss": 1.1821, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.999907554116106e-05, |
|
"loss": 1.2191, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.999825219980447e-05, |
|
"loss": 1.1625, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.999716886279962e-05, |
|
"loss": 1.1489, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.999582553577995e-05, |
|
"loss": 1.1297, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.999422222573088e-05, |
|
"loss": 1.1592, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.999235894098979e-05, |
|
"loss": 1.1427, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.999023569124595e-05, |
|
"loss": 1.0463, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.998785248754044e-05, |
|
"loss": 1.0977, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.998520934226615e-05, |
|
"loss": 1.1207, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.99823062691677e-05, |
|
"loss": 1.0715, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.99791432833413e-05, |
|
"loss": 1.0612, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.997572040123483e-05, |
|
"loss": 0.9976, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.997203764064758e-05, |
|
"loss": 0.9853, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.996809502073023e-05, |
|
"loss": 0.9514, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.996389256198482e-05, |
|
"loss": 0.9948, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.995943028626451e-05, |
|
"loss": 0.9322, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.995470821677356e-05, |
|
"loss": 0.9208, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.994972637806718e-05, |
|
"loss": 0.893, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.994448479605138e-05, |
|
"loss": 0.8938, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.993898349798294e-05, |
|
"loss": 0.869, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.993322251246907e-05, |
|
"loss": 0.8657, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.992720186946747e-05, |
|
"loss": 0.9, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.992092160028605e-05, |
|
"loss": 0.8594, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.991438173758282e-05, |
|
"loss": 0.9043, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.990758231536565e-05, |
|
"loss": 0.8273, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.990052336899223e-05, |
|
"loss": 0.8295, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.989320493516969e-05, |
|
"loss": 0.818, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.988562705195463e-05, |
|
"loss": 0.8361, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.987778975875273e-05, |
|
"loss": 0.7643, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.986969309631866e-05, |
|
"loss": 0.8122, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.986133710675585e-05, |
|
"loss": 0.7909, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.985272183351621e-05, |
|
"loss": 0.7521, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.984384732140001e-05, |
|
"loss": 0.7736, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.983471361655554e-05, |
|
"loss": 0.757, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.982532076647893e-05, |
|
"loss": 0.7525, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.981566882001388e-05, |
|
"loss": 0.7251, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.980575782735142e-05, |
|
"loss": 0.6862, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.979558784002968e-05, |
|
"loss": 0.7208, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.978515891093355e-05, |
|
"loss": 0.7133, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.977447109429443e-05, |
|
"loss": 0.7081, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.976352444568999e-05, |
|
"loss": 0.6588, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.975231902204385e-05, |
|
"loss": 0.6957, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.974085488162525e-05, |
|
"loss": 0.7015, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.972913208404884e-05, |
|
"loss": 0.6278, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.971715069027426e-05, |
|
"loss": 0.7023, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.970491076260587e-05, |
|
"loss": 0.6808, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.969241236469249e-05, |
|
"loss": 0.6916, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.967965556152696e-05, |
|
"loss": 0.6941, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.966664041944586e-05, |
|
"loss": 0.6286, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.965336700612917e-05, |
|
"loss": 0.7053, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.963983539059989e-05, |
|
"loss": 0.7111, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.962604564322371e-05, |
|
"loss": 0.6585, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.96119978357086e-05, |
|
"loss": 0.6242, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.959769204110453e-05, |
|
"loss": 0.6834, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.958312833380296e-05, |
|
"loss": 0.6638, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.956830678953654e-05, |
|
"loss": 0.6184, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.955322748537875e-05, |
|
"loss": 0.5992, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.953789049974334e-05, |
|
"loss": 0.5889, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.952229591238411e-05, |
|
"loss": 0.5633, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.950644380439442e-05, |
|
"loss": 0.6044, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.949033425820671e-05, |
|
"loss": 0.577, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.947396735759213e-05, |
|
"loss": 0.5375, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.945734318766014e-05, |
|
"loss": 0.5692, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.9440461834858e-05, |
|
"loss": 0.5442, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.942332338697032e-05, |
|
"loss": 0.5688, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.940592793311866e-05, |
|
"loss": 0.5424, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.938827556376106e-05, |
|
"loss": 0.5837, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.937036637069147e-05, |
|
"loss": 0.5909, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.935220044703941e-05, |
|
"loss": 0.5029, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.93337778872694e-05, |
|
"loss": 0.4939, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.931509878718051e-05, |
|
"loss": 0.5687, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.929616324390581e-05, |
|
"loss": 0.5074, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.927697135591194e-05, |
|
"loss": 0.556, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.92575232229985e-05, |
|
"loss": 0.4956, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.923781894629767e-05, |
|
"loss": 0.5471, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.921785862827353e-05, |
|
"loss": 0.5511, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.919764237272162e-05, |
|
"loss": 0.5238, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.917717028476839e-05, |
|
"loss": 0.4945, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.915644247087064e-05, |
|
"loss": 0.5521, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.913545903881497e-05, |
|
"loss": 0.5046, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.91142200977172e-05, |
|
"loss": 0.4685, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.909272575802187e-05, |
|
"loss": 0.517, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.907097613150153e-05, |
|
"loss": 0.4779, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.904897133125633e-05, |
|
"loss": 0.4686, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.902671147171331e-05, |
|
"loss": 0.471, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.900419666862587e-05, |
|
"loss": 0.4871, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.89814270390731e-05, |
|
"loss": 0.4345, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.895840270145921e-05, |
|
"loss": 0.4674, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.893512377551298e-05, |
|
"loss": 0.4669, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.891159038228699e-05, |
|
"loss": 0.4817, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.888780264415713e-05, |
|
"loss": 0.4559, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.88637606848219e-05, |
|
"loss": 0.4319, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.883946462930177e-05, |
|
"loss": 0.4799, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.881491460393853e-05, |
|
"loss": 0.4257, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.879011073639459e-05, |
|
"loss": 0.5319, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.876505315565248e-05, |
|
"loss": 0.4366, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.873974199201396e-05, |
|
"loss": 0.468, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.871417737709949e-05, |
|
"loss": 0.4477, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.868835944384746e-05, |
|
"loss": 0.4231, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.866228832651361e-05, |
|
"loss": 0.4449, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.86359641606702e-05, |
|
"loss": 0.406, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.860938708320539e-05, |
|
"loss": 0.399, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.858255723232249e-05, |
|
"loss": 0.4787, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.855547474753928e-05, |
|
"loss": 0.4314, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.852813976968723e-05, |
|
"loss": 0.3915, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.850055244091084e-05, |
|
"loss": 0.3818, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.847271290466682e-05, |
|
"loss": 0.4128, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.844462130572339e-05, |
|
"loss": 0.4149, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.841627779015953e-05, |
|
"loss": 0.4368, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.838768250536421e-05, |
|
"loss": 0.3788, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.835883560003561e-05, |
|
"loss": 0.4315, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.83297372241804e-05, |
|
"loss": 0.4171, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.830038752911283e-05, |
|
"loss": 0.4019, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.827078666745418e-05, |
|
"loss": 0.4521, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.824093479313165e-05, |
|
"loss": 0.343, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.821083206137788e-05, |
|
"loss": 0.4224, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.81804786287299e-05, |
|
"loss": 0.4192, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.814987465302843e-05, |
|
"loss": 0.4035, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.811902029341704e-05, |
|
"loss": 0.3777, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.808791571034135e-05, |
|
"loss": 0.4148, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.80565610655481e-05, |
|
"loss": 0.3854, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.802495652208445e-05, |
|
"loss": 0.3892, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.799310224429701e-05, |
|
"loss": 0.4206, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.796099839783103e-05, |
|
"loss": 0.4363, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.792864514962956e-05, |
|
"loss": 0.3719, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.789604266793257e-05, |
|
"loss": 0.3671, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.786319112227603e-05, |
|
"loss": 0.4127, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.783009068349109e-05, |
|
"loss": 0.4179, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.77967415237032e-05, |
|
"loss": 0.371, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.776314381633111e-05, |
|
"loss": 0.3564, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.77292977360861e-05, |
|
"loss": 0.3611, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.769520345897099e-05, |
|
"loss": 0.3295, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.766086116227926e-05, |
|
"loss": 0.3418, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.76262710245941e-05, |
|
"loss": 0.3541, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.759143322578751e-05, |
|
"loss": 0.3386, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.755634794701933e-05, |
|
"loss": 0.3182, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.752101537073634e-05, |
|
"loss": 0.3904, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.74854356806713e-05, |
|
"loss": 0.3687, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.744960906184194e-05, |
|
"loss": 0.3435, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.741353570055009e-05, |
|
"loss": 0.3185, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.737721578438064e-05, |
|
"loss": 0.3605, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.734064950220061e-05, |
|
"loss": 0.3285, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.730383704415811e-05, |
|
"loss": 0.3139, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.726677860168142e-05, |
|
"loss": 0.3675, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.722947436747796e-05, |
|
"loss": 0.3471, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.719192453553325e-05, |
|
"loss": 0.3423, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.715412930111002e-05, |
|
"loss": 0.3339, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.711608886074703e-05, |
|
"loss": 0.351, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.707780341225821e-05, |
|
"loss": 0.2691, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.703927315473149e-05, |
|
"loss": 0.3315, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.700049828852786e-05, |
|
"loss": 0.3261, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.696147901528032e-05, |
|
"loss": 0.2997, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.692221553789277e-05, |
|
"loss": 0.3094, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.6882708060539e-05, |
|
"loss": 0.3018, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.684295678866169e-05, |
|
"loss": 0.2936, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.680296192897114e-05, |
|
"loss": 0.3058, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.676272368944448e-05, |
|
"loss": 0.3237, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.672224227932433e-05, |
|
"loss": 0.272, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.668151790911786e-05, |
|
"loss": 0.2933, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 9.664055079059566e-05, |
|
"loss": 0.3095, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 9.659934113679064e-05, |
|
"loss": 0.2977, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 9.655788916199688e-05, |
|
"loss": 0.294, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 9.651619508176859e-05, |
|
"loss": 0.2843, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 9.647425911291894e-05, |
|
"loss": 0.28, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 9.643208147351893e-05, |
|
"loss": 0.2794, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 9.638966238289628e-05, |
|
"loss": 0.3104, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 9.63470020616343e-05, |
|
"loss": 0.2917, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 9.630410073157069e-05, |
|
"loss": 0.2701, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 9.626095861579643e-05, |
|
"loss": 0.2661, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 9.621757593865459e-05, |
|
"loss": 0.2869, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 9.617395292573923e-05, |
|
"loss": 0.2935, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 9.613008980389416e-05, |
|
"loss": 0.3391, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 9.608598680121172e-05, |
|
"loss": 0.2654, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.604164414703175e-05, |
|
"loss": 0.2974, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.599706207194022e-05, |
|
"loss": 0.3128, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.595224080776817e-05, |
|
"loss": 0.28, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.59071805875904e-05, |
|
"loss": 0.2658, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.586188164572435e-05, |
|
"loss": 0.244, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 9.581634421772878e-05, |
|
"loss": 0.2884, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 9.577056854040262e-05, |
|
"loss": 0.2768, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 9.572455485178377e-05, |
|
"loss": 0.2649, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 9.567830339114776e-05, |
|
"loss": 0.3047, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 9.563181439900655e-05, |
|
"loss": 0.296, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 9.558508811710731e-05, |
|
"loss": 0.2854, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 9.553812478843114e-05, |
|
"loss": 0.2982, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 9.549092465719178e-05, |
|
"loss": 0.2845, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 9.544348796883441e-05, |
|
"loss": 0.2385, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 9.539581497003429e-05, |
|
"loss": 0.2782, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 9.534790590869552e-05, |
|
"loss": 0.2555, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 9.529976103394978e-05, |
|
"loss": 0.2351, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 9.525138059615495e-05, |
|
"loss": 0.2713, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 9.520276484689392e-05, |
|
"loss": 0.2828, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 9.515391403897317e-05, |
|
"loss": 0.2697, |
|
"step": 669 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 9.510482842642157e-05, |
|
"loss": 0.2552, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 9.50555082644889e-05, |
|
"loss": 0.2809, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 9.50059538096447e-05, |
|
"loss": 0.2809, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 9.495616531957684e-05, |
|
"loss": 0.2342, |
|
"step": 681 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 9.490614305319017e-05, |
|
"loss": 0.2461, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 9.485588727060521e-05, |
|
"loss": 0.249, |
|
"step": 687 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 9.480539823315682e-05, |
|
"loss": 0.2609, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 9.475467620339275e-05, |
|
"loss": 0.2744, |
|
"step": 693 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 9.470372144507237e-05, |
|
"loss": 0.2688, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.465253422316524e-05, |
|
"loss": 0.2487, |
|
"step": 699 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.460111480384979e-05, |
|
"loss": 0.2488, |
|
"step": 702 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.454946345451183e-05, |
|
"loss": 0.2727, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.449758044374331e-05, |
|
"loss": 0.2446, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.44454660413408e-05, |
|
"loss": 0.2686, |
|
"step": 711 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.439312051830409e-05, |
|
"loss": 0.2424, |
|
"step": 714 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.434054414683489e-05, |
|
"loss": 0.2853, |
|
"step": 717 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.42877372003353e-05, |
|
"loss": 0.2317, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.423469995340648e-05, |
|
"loss": 0.245, |
|
"step": 723 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.41814326818471e-05, |
|
"loss": 0.2661, |
|
"step": 726 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.412793566265201e-05, |
|
"loss": 0.2433, |
|
"step": 729 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.407420917401082e-05, |
|
"loss": 0.2316, |
|
"step": 732 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.402025349530633e-05, |
|
"loss": 0.2172, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.39660689071132e-05, |
|
"loss": 0.2379, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.391165569119638e-05, |
|
"loss": 0.227, |
|
"step": 741 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.38570141305098e-05, |
|
"loss": 0.2239, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.380214450919474e-05, |
|
"loss": 0.2413, |
|
"step": 747 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.374704711257843e-05, |
|
"loss": 0.2193, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.369172222717251e-05, |
|
"loss": 0.2278, |
|
"step": 753 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.363617014067168e-05, |
|
"loss": 0.2092, |
|
"step": 756 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.358039114195203e-05, |
|
"loss": 0.2193, |
|
"step": 759 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.352438552106964e-05, |
|
"loss": 0.2251, |
|
"step": 762 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.346815356925904e-05, |
|
"loss": 0.1976, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.341169557893171e-05, |
|
"loss": 0.2125, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.335501184367455e-05, |
|
"loss": 0.2191, |
|
"step": 771 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.329810265824837e-05, |
|
"loss": 0.2488, |
|
"step": 774 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.324096831858628e-05, |
|
"loss": 0.2331, |
|
"step": 777 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.318360912179232e-05, |
|
"loss": 0.2159, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.312602536613968e-05, |
|
"loss": 0.2258, |
|
"step": 783 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.30682173510694e-05, |
|
"loss": 0.2157, |
|
"step": 786 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.301018537718862e-05, |
|
"loss": 0.2438, |
|
"step": 789 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.29519297462691e-05, |
|
"loss": 0.2276, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.289345076124562e-05, |
|
"loss": 0.2056, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.283474872621448e-05, |
|
"loss": 0.2274, |
|
"step": 798 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.277582394643179e-05, |
|
"loss": 0.2272, |
|
"step": 801 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.271667672831199e-05, |
|
"loss": 0.2033, |
|
"step": 804 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.265730737942625e-05, |
|
"loss": 0.2068, |
|
"step": 807 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.259771620850077e-05, |
|
"loss": 0.2137, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.253790352541532e-05, |
|
"loss": 0.2475, |
|
"step": 813 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.24778696412015e-05, |
|
"loss": 0.2361, |
|
"step": 816 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.241761486804119e-05, |
|
"loss": 0.2405, |
|
"step": 819 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.235713951926493e-05, |
|
"loss": 0.2116, |
|
"step": 822 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 9.229644390935024e-05, |
|
"loss": 0.2054, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 9.223552835392008e-05, |
|
"loss": 0.231, |
|
"step": 828 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 9.217439316974108e-05, |
|
"loss": 0.2106, |
|
"step": 831 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 9.2113038674722e-05, |
|
"loss": 0.2189, |
|
"step": 834 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 9.205146518791204e-05, |
|
"loss": 0.2116, |
|
"step": 837 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 9.198967302949912e-05, |
|
"loss": 0.2278, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 9.192766252080837e-05, |
|
"loss": 0.2017, |
|
"step": 843 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 9.18654339843003e-05, |
|
"loss": 0.2019, |
|
"step": 846 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 9.180298774356917e-05, |
|
"loss": 0.1702, |
|
"step": 849 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 9.174032412334135e-05, |
|
"loss": 0.2299, |
|
"step": 852 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 9.167744344947363e-05, |
|
"loss": 0.1961, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 9.161434604895146e-05, |
|
"loss": 0.1934, |
|
"step": 858 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 9.155103224988728e-05, |
|
"loss": 0.1995, |
|
"step": 861 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 9.148750238151887e-05, |
|
"loss": 0.2047, |
|
"step": 864 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 9.142375677420756e-05, |
|
"loss": 0.2101, |
|
"step": 867 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 9.135979575943653e-05, |
|
"loss": 0.2227, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 9.129561966980913e-05, |
|
"loss": 0.1714, |
|
"step": 873 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 9.123122883904711e-05, |
|
"loss": 0.2298, |
|
"step": 876 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 9.116662360198888e-05, |
|
"loss": 0.2121, |
|
"step": 879 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 9.11018042945878e-05, |
|
"loss": 0.2061, |
|
"step": 882 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 9.10367712539104e-05, |
|
"loss": 0.1741, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 9.097152481813466e-05, |
|
"loss": 0.2126, |
|
"step": 888 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 9.090606532654823e-05, |
|
"loss": 0.1853, |
|
"step": 891 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 9.084039311954667e-05, |
|
"loss": 0.2013, |
|
"step": 894 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 9.077450853863168e-05, |
|
"loss": 0.1789, |
|
"step": 897 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 9.070841192640931e-05, |
|
"loss": 0.2109, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 9.064210362658819e-05, |
|
"loss": 0.2062, |
|
"step": 903 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 9.057558398397781e-05, |
|
"loss": 0.2128, |
|
"step": 906 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 9.050885334448658e-05, |
|
"loss": 0.1825, |
|
"step": 909 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 9.044191205512013e-05, |
|
"loss": 0.2, |
|
"step": 912 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 9.037476046397957e-05, |
|
"loss": 0.2142, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 9.030739892025947e-05, |
|
"loss": 0.1759, |
|
"step": 918 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 9.023982777424627e-05, |
|
"loss": 0.205, |
|
"step": 921 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 9.017204737731632e-05, |
|
"loss": 0.1673, |
|
"step": 924 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 9.010405808193412e-05, |
|
"loss": 0.1765, |
|
"step": 927 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 9.003586024165044e-05, |
|
"loss": 0.1781, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 8.996745421110051e-05, |
|
"loss": 0.1831, |
|
"step": 933 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 8.989884034600217e-05, |
|
"loss": 0.1855, |
|
"step": 936 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 8.983001900315402e-05, |
|
"loss": 0.1725, |
|
"step": 939 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 8.976099054043357e-05, |
|
"loss": 0.1843, |
|
"step": 942 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 8.969175531679533e-05, |
|
"loss": 0.1838, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 8.962231369226906e-05, |
|
"loss": 0.1848, |
|
"step": 948 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 8.955266602795779e-05, |
|
"loss": 0.1707, |
|
"step": 951 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 8.948281268603595e-05, |
|
"loss": 0.1491, |
|
"step": 954 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 8.941275402974754e-05, |
|
"loss": 0.1794, |
|
"step": 957 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 8.934249042340421e-05, |
|
"loss": 0.19, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 8.92720222323834e-05, |
|
"loss": 0.1985, |
|
"step": 963 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 8.920134982312633e-05, |
|
"loss": 0.1989, |
|
"step": 966 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 8.913047356313625e-05, |
|
"loss": 0.1682, |
|
"step": 969 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 8.905939382097643e-05, |
|
"loss": 0.1681, |
|
"step": 972 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 8.898811096626826e-05, |
|
"loss": 0.1606, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 8.891662536968932e-05, |
|
"loss": 0.1593, |
|
"step": 978 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 8.884493740297152e-05, |
|
"loss": 0.193, |
|
"step": 981 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 8.877304743889905e-05, |
|
"loss": 0.1803, |
|
"step": 984 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 8.870095585130653e-05, |
|
"loss": 0.1911, |
|
"step": 987 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 8.862866301507709e-05, |
|
"loss": 0.181, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 8.855616930614027e-05, |
|
"loss": 0.1673, |
|
"step": 993 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 8.848347510147026e-05, |
|
"loss": 0.1811, |
|
"step": 996 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 8.841058077908379e-05, |
|
"loss": 0.1868, |
|
"step": 999 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 8.833748671803826e-05, |
|
"loss": 0.1542, |
|
"step": 1002 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 8.826419329842972e-05, |
|
"loss": 0.1497, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 8.819070090139086e-05, |
|
"loss": 0.1908, |
|
"step": 1008 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 8.811700990908914e-05, |
|
"loss": 0.182, |
|
"step": 1011 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 8.80431207047247e-05, |
|
"loss": 0.1592, |
|
"step": 1014 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 8.796903367252837e-05, |
|
"loss": 0.1718, |
|
"step": 1017 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 8.789474919775979e-05, |
|
"loss": 0.1649, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 8.782026766670522e-05, |
|
"loss": 0.1837, |
|
"step": 1023 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 8.774558946667569e-05, |
|
"loss": 0.1811, |
|
"step": 1026 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 8.767071498600492e-05, |
|
"loss": 0.1839, |
|
"step": 1029 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 8.759564461404731e-05, |
|
"loss": 0.185, |
|
"step": 1032 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 8.75203787411759e-05, |
|
"loss": 0.1817, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 8.744491775878036e-05, |
|
"loss": 0.1937, |
|
"step": 1038 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 8.736926205926495e-05, |
|
"loss": 0.1834, |
|
"step": 1041 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 8.729341203604646e-05, |
|
"loss": 0.1789, |
|
"step": 1044 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 8.72173680835522e-05, |
|
"loss": 0.1389, |
|
"step": 1047 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 8.714113059721796e-05, |
|
"loss": 0.1955, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 8.706469997348585e-05, |
|
"loss": 0.154, |
|
"step": 1053 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 8.698807660980237e-05, |
|
"loss": 0.1485, |
|
"step": 1056 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 8.691126090461628e-05, |
|
"loss": 0.1723, |
|
"step": 1059 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 8.683425325737651e-05, |
|
"loss": 0.185, |
|
"step": 1062 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 8.675705406853014e-05, |
|
"loss": 0.1656, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 8.667966373952023e-05, |
|
"loss": 0.1458, |
|
"step": 1068 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 8.660208267278387e-05, |
|
"loss": 0.1592, |
|
"step": 1071 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 8.652431127174994e-05, |
|
"loss": 0.177, |
|
"step": 1074 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 8.644634994083707e-05, |
|
"loss": 0.1663, |
|
"step": 1077 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 8.636819908545157e-05, |
|
"loss": 0.1685, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 8.628985911198533e-05, |
|
"loss": 0.1425, |
|
"step": 1083 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 8.62113304278136e-05, |
|
"loss": 0.1742, |
|
"step": 1086 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 8.6132613441293e-05, |
|
"loss": 0.1788, |
|
"step": 1089 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 8.60537085617593e-05, |
|
"loss": 0.1601, |
|
"step": 1092 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 8.597461619952534e-05, |
|
"loss": 0.1678, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 8.589533676587893e-05, |
|
"loss": 0.1641, |
|
"step": 1098 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 8.58158706730806e-05, |
|
"loss": 0.1818, |
|
"step": 1101 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 8.573621833436158e-05, |
|
"loss": 0.166, |
|
"step": 1104 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 8.565638016392158e-05, |
|
"loss": 0.156, |
|
"step": 1107 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 8.557635657692665e-05, |
|
"loss": 0.1745, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 8.5496147989507e-05, |
|
"loss": 0.1614, |
|
"step": 1113 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 8.541575481875493e-05, |
|
"loss": 0.1691, |
|
"step": 1116 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 8.53351774827225e-05, |
|
"loss": 0.1511, |
|
"step": 1119 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 8.525441640041951e-05, |
|
"loss": 0.1781, |
|
"step": 1122 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 8.517347199181124e-05, |
|
"loss": 0.1433, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 8.509234467781627e-05, |
|
"loss": 0.1607, |
|
"step": 1128 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 8.501103488030435e-05, |
|
"loss": 0.1533, |
|
"step": 1131 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 8.49295430220941e-05, |
|
"loss": 0.145, |
|
"step": 1134 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 8.484786952695091e-05, |
|
"loss": 0.1526, |
|
"step": 1137 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 8.476601481958471e-05, |
|
"loss": 0.1568, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 8.46839793256477e-05, |
|
"loss": 0.1585, |
|
"step": 1143 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 8.460176347173224e-05, |
|
"loss": 0.1598, |
|
"step": 1146 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 8.451936768536858e-05, |
|
"loss": 0.1586, |
|
"step": 1149 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 8.443679239502258e-05, |
|
"loss": 0.1358, |
|
"step": 1152 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 8.435403803009359e-05, |
|
"loss": 0.1371, |
|
"step": 1155 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 8.427110502091214e-05, |
|
"loss": 0.1559, |
|
"step": 1158 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 8.418799379873775e-05, |
|
"loss": 0.1457, |
|
"step": 1161 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 8.410470479575663e-05, |
|
"loss": 0.1481, |
|
"step": 1164 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 8.40212384450795e-05, |
|
"loss": 0.1593, |
|
"step": 1167 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 8.39375951807393e-05, |
|
"loss": 0.1423, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 8.385377543768888e-05, |
|
"loss": 0.1408, |
|
"step": 1173 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 8.376977965179891e-05, |
|
"loss": 0.1386, |
|
"step": 1176 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 8.36856082598554e-05, |
|
"loss": 0.1524, |
|
"step": 1179 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 8.360126169955756e-05, |
|
"loss": 0.1497, |
|
"step": 1182 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 8.351674040951551e-05, |
|
"loss": 0.1568, |
|
"step": 1185 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 8.343204482924796e-05, |
|
"loss": 0.1583, |
|
"step": 1188 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 8.334717539917994e-05, |
|
"loss": 0.1399, |
|
"step": 1191 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 8.326213256064053e-05, |
|
"loss": 0.1484, |
|
"step": 1194 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 8.317691675586057e-05, |
|
"loss": 0.1509, |
|
"step": 1197 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 8.30915284279703e-05, |
|
"loss": 0.1418, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 8.30059680209971e-05, |
|
"loss": 0.139, |
|
"step": 1203 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 8.292023597986323e-05, |
|
"loss": 0.1756, |
|
"step": 1206 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 8.283433275038339e-05, |
|
"loss": 0.136, |
|
"step": 1209 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 8.27482587792625e-05, |
|
"loss": 0.1163, |
|
"step": 1212 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 8.266201451409343e-05, |
|
"loss": 0.1489, |
|
"step": 1215 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 8.257560040335449e-05, |
|
"loss": 0.1536, |
|
"step": 1218 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 8.248901689640724e-05, |
|
"loss": 0.1259, |
|
"step": 1221 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 8.240226444349411e-05, |
|
"loss": 0.1449, |
|
"step": 1224 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 8.231534349573609e-05, |
|
"loss": 0.1492, |
|
"step": 1227 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 8.222825450513035e-05, |
|
"loss": 0.139, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 8.214099792454787e-05, |
|
"loss": 0.14, |
|
"step": 1233 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 8.205357420773115e-05, |
|
"loss": 0.1426, |
|
"step": 1236 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 8.19659838092918e-05, |
|
"loss": 0.1341, |
|
"step": 1239 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 8.187822718470818e-05, |
|
"loss": 0.141, |
|
"step": 1242 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 8.179030479032305e-05, |
|
"loss": 0.1552, |
|
"step": 1245 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 8.170221708334122e-05, |
|
"loss": 0.1429, |
|
"step": 1248 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 8.161396452182709e-05, |
|
"loss": 0.1202, |
|
"step": 1251 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 8.152554756470234e-05, |
|
"loss": 0.1411, |
|
"step": 1254 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 8.143696667174352e-05, |
|
"loss": 0.1476, |
|
"step": 1257 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 8.134822230357969e-05, |
|
"loss": 0.1551, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 8.125931492168992e-05, |
|
"loss": 0.1464, |
|
"step": 1263 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 8.117024498840108e-05, |
|
"loss": 0.1473, |
|
"step": 1266 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 8.108101296688523e-05, |
|
"loss": 0.1311, |
|
"step": 1269 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 8.099161932115733e-05, |
|
"loss": 0.1314, |
|
"step": 1272 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 8.090206451607283e-05, |
|
"loss": 0.1427, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 8.081234901732517e-05, |
|
"loss": 0.1288, |
|
"step": 1278 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 8.072247329144346e-05, |
|
"loss": 0.1377, |
|
"step": 1281 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 8.063243780579002e-05, |
|
"loss": 0.1492, |
|
"step": 1284 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 8.054224302855787e-05, |
|
"loss": 0.1519, |
|
"step": 1287 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 8.045188942876838e-05, |
|
"loss": 0.1336, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 8.036137747626886e-05, |
|
"loss": 0.1249, |
|
"step": 1293 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 8.027070764173002e-05, |
|
"loss": 0.1313, |
|
"step": 1296 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 8.017988039664361e-05, |
|
"loss": 0.146, |
|
"step": 1299 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 8.00888962133199e-05, |
|
"loss": 0.131, |
|
"step": 1302 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 7.999775556488526e-05, |
|
"loss": 0.1507, |
|
"step": 1305 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 7.99064589252797e-05, |
|
"loss": 0.1328, |
|
"step": 1308 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 7.981500676925438e-05, |
|
"loss": 0.1255, |
|
"step": 1311 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 7.97233995723692e-05, |
|
"loss": 0.1426, |
|
"step": 1314 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 7.963163781099026e-05, |
|
"loss": 0.1363, |
|
"step": 1317 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 7.953972196228738e-05, |
|
"loss": 0.1436, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 7.944765250423173e-05, |
|
"loss": 0.1356, |
|
"step": 1323 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 7.935542991559319e-05, |
|
"loss": 0.1378, |
|
"step": 1326 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 7.926305467593793e-05, |
|
"loss": 0.1228, |
|
"step": 1329 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 7.917052726562599e-05, |
|
"loss": 0.1234, |
|
"step": 1332 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 7.907784816580868e-05, |
|
"loss": 0.1369, |
|
"step": 1335 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 7.898501785842605e-05, |
|
"loss": 0.1354, |
|
"step": 1338 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 7.889203682620452e-05, |
|
"loss": 0.1371, |
|
"step": 1341 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 7.879890555265428e-05, |
|
"loss": 0.1358, |
|
"step": 1344 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 7.870562452206674e-05, |
|
"loss": 0.1268, |
|
"step": 1347 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 7.861219421951211e-05, |
|
"loss": 0.1345, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 7.851861513083685e-05, |
|
"loss": 0.1428, |
|
"step": 1353 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 7.842488774266106e-05, |
|
"loss": 0.1241, |
|
"step": 1356 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 7.833101254237601e-05, |
|
"loss": 0.1252, |
|
"step": 1359 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 7.823699001814167e-05, |
|
"loss": 0.1455, |
|
"step": 1362 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 7.814282065888408e-05, |
|
"loss": 0.1153, |
|
"step": 1365 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 7.804850495429281e-05, |
|
"loss": 0.136, |
|
"step": 1368 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 7.795404339481847e-05, |
|
"loss": 0.1347, |
|
"step": 1371 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 7.78594364716701e-05, |
|
"loss": 0.1175, |
|
"step": 1374 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 7.776468467681268e-05, |
|
"loss": 0.1349, |
|
"step": 1377 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 7.766978850296452e-05, |
|
"loss": 0.1374, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 7.75747484435947e-05, |
|
"loss": 0.1288, |
|
"step": 1383 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 7.747956499292057e-05, |
|
"loss": 0.1127, |
|
"step": 1386 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 7.738423864590506e-05, |
|
"loss": 0.1297, |
|
"step": 1389 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 7.728876989825418e-05, |
|
"loss": 0.1381, |
|
"step": 1392 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 7.719315924641452e-05, |
|
"loss": 0.1243, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 7.70974071875705e-05, |
|
"loss": 0.1247, |
|
"step": 1398 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 7.700151421964187e-05, |
|
"loss": 0.1423, |
|
"step": 1401 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 7.690548084128122e-05, |
|
"loss": 0.1316, |
|
"step": 1404 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 7.680930755187115e-05, |
|
"loss": 0.1207, |
|
"step": 1407 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 7.671299485152192e-05, |
|
"loss": 0.1035, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 7.661654324106865e-05, |
|
"loss": 0.1116, |
|
"step": 1413 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 7.651995322206888e-05, |
|
"loss": 0.1145, |
|
"step": 1416 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 7.642322529679988e-05, |
|
"loss": 0.1167, |
|
"step": 1419 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 7.632635996825598e-05, |
|
"loss": 0.1166, |
|
"step": 1422 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 7.622935774014609e-05, |
|
"loss": 0.1098, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 7.6132219116891e-05, |
|
"loss": 0.0963, |
|
"step": 1428 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 7.603494460362072e-05, |
|
"loss": 0.0987, |
|
"step": 1431 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 7.593753470617196e-05, |
|
"loss": 0.1057, |
|
"step": 1434 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 7.58399899310854e-05, |
|
"loss": 0.1, |
|
"step": 1437 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 7.574231078560314e-05, |
|
"loss": 0.1062, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 7.564449777766595e-05, |
|
"loss": 0.1115, |
|
"step": 1443 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 7.554655141591077e-05, |
|
"loss": 0.1204, |
|
"step": 1446 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 7.544847220966795e-05, |
|
"loss": 0.1065, |
|
"step": 1449 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 7.535026066895864e-05, |
|
"loss": 0.0947, |
|
"step": 1452 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 7.525191730449219e-05, |
|
"loss": 0.0989, |
|
"step": 1455 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 7.515344262766338e-05, |
|
"loss": 0.1073, |
|
"step": 1458 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 7.505483715054988e-05, |
|
"loss": 0.1014, |
|
"step": 1461 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 7.495610138590949e-05, |
|
"loss": 0.1014, |
|
"step": 1464 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 7.485723584717757e-05, |
|
"loss": 0.1033, |
|
"step": 1467 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 7.475824104846427e-05, |
|
"loss": 0.1299, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 7.465911750455192e-05, |
|
"loss": 0.1021, |
|
"step": 1473 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 7.455986573089234e-05, |
|
"loss": 0.1116, |
|
"step": 1476 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 7.446048624360417e-05, |
|
"loss": 0.1016, |
|
"step": 1479 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 7.436097955947014e-05, |
|
"loss": 0.1017, |
|
"step": 1482 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 7.426134619593443e-05, |
|
"loss": 0.1117, |
|
"step": 1485 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 7.416158667110002e-05, |
|
"loss": 0.1176, |
|
"step": 1488 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 7.406170150372583e-05, |
|
"loss": 0.0998, |
|
"step": 1491 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 7.396169121322422e-05, |
|
"loss": 0.0899, |
|
"step": 1494 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 7.386155631965818e-05, |
|
"loss": 0.1014, |
|
"step": 1497 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 7.376129734373862e-05, |
|
"loss": 0.1114, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 7.366091480682173e-05, |
|
"loss": 0.1049, |
|
"step": 1503 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 7.356040923090624e-05, |
|
"loss": 0.1066, |
|
"step": 1506 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 7.345978113863063e-05, |
|
"loss": 0.1046, |
|
"step": 1509 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 7.33590310532705e-05, |
|
"loss": 0.0984, |
|
"step": 1512 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 7.325815949873587e-05, |
|
"loss": 0.1061, |
|
"step": 1515 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 7.315716699956837e-05, |
|
"loss": 0.1059, |
|
"step": 1518 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 7.305605408093858e-05, |
|
"loss": 0.1102, |
|
"step": 1521 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 7.295482126864323e-05, |
|
"loss": 0.1044, |
|
"step": 1524 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 7.285346908910254e-05, |
|
"loss": 0.1026, |
|
"step": 1527 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 7.275199806935745e-05, |
|
"loss": 0.1199, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 7.265040873706687e-05, |
|
"loss": 0.1055, |
|
"step": 1533 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 7.254870162050498e-05, |
|
"loss": 0.106, |
|
"step": 1536 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 7.24468772485584e-05, |
|
"loss": 0.1055, |
|
"step": 1539 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 7.23449361507235e-05, |
|
"loss": 0.1, |
|
"step": 1542 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 7.224287885710366e-05, |
|
"loss": 0.1091, |
|
"step": 1545 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 7.214070589840648e-05, |
|
"loss": 0.1061, |
|
"step": 1548 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 7.203841780594101e-05, |
|
"loss": 0.0959, |
|
"step": 1551 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 7.193601511161505e-05, |
|
"loss": 0.1197, |
|
"step": 1554 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 7.183349834793229e-05, |
|
"loss": 0.0991, |
|
"step": 1557 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 7.173086804798961e-05, |
|
"loss": 0.1019, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 7.16281247454743e-05, |
|
"loss": 0.1148, |
|
"step": 1563 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 7.152526897466126e-05, |
|
"loss": 0.1151, |
|
"step": 1566 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 7.142230127041025e-05, |
|
"loss": 0.1098, |
|
"step": 1569 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 7.131922216816311e-05, |
|
"loss": 0.1049, |
|
"step": 1572 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 7.121603220394089e-05, |
|
"loss": 0.1026, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 7.111273191434122e-05, |
|
"loss": 0.1031, |
|
"step": 1578 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 7.100932183653537e-05, |
|
"loss": 0.0932, |
|
"step": 1581 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 7.090580250826554e-05, |
|
"loss": 0.0985, |
|
"step": 1584 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 7.080217446784206e-05, |
|
"loss": 0.0979, |
|
"step": 1587 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 7.069843825414056e-05, |
|
"loss": 0.1042, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 7.059459440659917e-05, |
|
"loss": 0.0973, |
|
"step": 1593 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 7.049064346521573e-05, |
|
"loss": 0.0944, |
|
"step": 1596 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 7.0386585970545e-05, |
|
"loss": 0.1044, |
|
"step": 1599 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 7.028242246369577e-05, |
|
"loss": 0.1143, |
|
"step": 1602 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 7.01781534863282e-05, |
|
"loss": 0.0982, |
|
"step": 1605 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 7.007377958065083e-05, |
|
"loss": 0.1065, |
|
"step": 1608 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 6.996930128941783e-05, |
|
"loss": 0.0932, |
|
"step": 1611 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 6.986471915592625e-05, |
|
"loss": 0.1032, |
|
"step": 1614 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 6.976003372401307e-05, |
|
"loss": 0.0992, |
|
"step": 1617 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 6.965524553805245e-05, |
|
"loss": 0.0939, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 6.955035514295292e-05, |
|
"loss": 0.0915, |
|
"step": 1623 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 6.944536308415444e-05, |
|
"loss": 0.1157, |
|
"step": 1626 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 6.934026990762568e-05, |
|
"loss": 0.1028, |
|
"step": 1629 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 6.923507615986109e-05, |
|
"loss": 0.12, |
|
"step": 1632 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 6.912978238787813e-05, |
|
"loss": 0.1102, |
|
"step": 1635 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 6.90243891392144e-05, |
|
"loss": 0.0964, |
|
"step": 1638 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 6.891889696192481e-05, |
|
"loss": 0.0948, |
|
"step": 1641 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 6.881330640457862e-05, |
|
"loss": 0.1052, |
|
"step": 1644 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 6.870761801625676e-05, |
|
"loss": 0.1002, |
|
"step": 1647 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 6.860183234654887e-05, |
|
"loss": 0.1059, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 6.84959499455504e-05, |
|
"loss": 0.0844, |
|
"step": 1653 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 6.838997136385996e-05, |
|
"loss": 0.0964, |
|
"step": 1656 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 6.828389715257614e-05, |
|
"loss": 0.0959, |
|
"step": 1659 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 6.81777278632949e-05, |
|
"loss": 0.0976, |
|
"step": 1662 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 6.807146404810664e-05, |
|
"loss": 0.0992, |
|
"step": 1665 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 6.796510625959321e-05, |
|
"loss": 0.097, |
|
"step": 1668 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 6.785865505082524e-05, |
|
"loss": 0.1005, |
|
"step": 1671 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 6.775211097535908e-05, |
|
"loss": 0.1216, |
|
"step": 1674 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 6.764547458723401e-05, |
|
"loss": 0.1074, |
|
"step": 1677 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 6.753874644096935e-05, |
|
"loss": 0.0992, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 6.743192709156159e-05, |
|
"loss": 0.1034, |
|
"step": 1683 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 6.732501709448144e-05, |
|
"loss": 0.0804, |
|
"step": 1686 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 6.721801700567103e-05, |
|
"loss": 0.1002, |
|
"step": 1689 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 6.711092738154097e-05, |
|
"loss": 0.0954, |
|
"step": 1692 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 6.700374877896743e-05, |
|
"loss": 0.0903, |
|
"step": 1695 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 6.689648175528931e-05, |
|
"loss": 0.0922, |
|
"step": 1698 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 6.678912686830528e-05, |
|
"loss": 0.0943, |
|
"step": 1701 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 6.668168467627094e-05, |
|
"loss": 0.0858, |
|
"step": 1704 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 6.657415573789588e-05, |
|
"loss": 0.1088, |
|
"step": 1707 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 6.646654061234072e-05, |
|
"loss": 0.0955, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 6.635883985921435e-05, |
|
"loss": 0.1152, |
|
"step": 1713 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 6.625105403857089e-05, |
|
"loss": 0.0967, |
|
"step": 1716 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 6.614318371090678e-05, |
|
"loss": 0.0926, |
|
"step": 1719 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 6.603522943715798e-05, |
|
"loss": 0.0912, |
|
"step": 1722 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 6.592719177869693e-05, |
|
"loss": 0.1041, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 6.581907129732971e-05, |
|
"loss": 0.1026, |
|
"step": 1728 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 6.571086855529303e-05, |
|
"loss": 0.0923, |
|
"step": 1731 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 6.560258411525139e-05, |
|
"loss": 0.09, |
|
"step": 1734 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 6.54942185402942e-05, |
|
"loss": 0.0954, |
|
"step": 1737 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 6.538577239393269e-05, |
|
"loss": 0.0785, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 6.527724624009706e-05, |
|
"loss": 0.0986, |
|
"step": 1743 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 6.516864064313362e-05, |
|
"loss": 0.0932, |
|
"step": 1746 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 6.505995616780176e-05, |
|
"loss": 0.0914, |
|
"step": 1749 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 6.495119337927103e-05, |
|
"loss": 0.0889, |
|
"step": 1752 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 6.484235284311828e-05, |
|
"loss": 0.0922, |
|
"step": 1755 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 6.47334351253246e-05, |
|
"loss": 0.0834, |
|
"step": 1758 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 6.462444079227239e-05, |
|
"loss": 0.0889, |
|
"step": 1761 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 6.451537041074255e-05, |
|
"loss": 0.0919, |
|
"step": 1764 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 6.440622454791139e-05, |
|
"loss": 0.0823, |
|
"step": 1767 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 6.429700377134773e-05, |
|
"loss": 0.0932, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 6.418770864900997e-05, |
|
"loss": 0.0819, |
|
"step": 1773 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 6.40783397492431e-05, |
|
"loss": 0.0768, |
|
"step": 1776 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 6.396889764077575e-05, |
|
"loss": 0.0883, |
|
"step": 1779 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 6.385938289271725e-05, |
|
"loss": 0.1014, |
|
"step": 1782 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 6.374979607455465e-05, |
|
"loss": 0.0918, |
|
"step": 1785 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 6.364013775614984e-05, |
|
"loss": 0.0711, |
|
"step": 1788 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 6.353040850773644e-05, |
|
"loss": 0.0918, |
|
"step": 1791 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 6.342060889991691e-05, |
|
"loss": 0.093, |
|
"step": 1794 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 6.331073950365963e-05, |
|
"loss": 0.1003, |
|
"step": 1797 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 6.320080089029589e-05, |
|
"loss": 0.0888, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 6.309079363151685e-05, |
|
"loss": 0.0968, |
|
"step": 1803 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 6.298071829937072e-05, |
|
"loss": 0.0952, |
|
"step": 1806 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 6.287057546625965e-05, |
|
"loss": 0.0822, |
|
"step": 1809 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 6.27603657049368e-05, |
|
"loss": 0.0797, |
|
"step": 1812 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 6.265008958850336e-05, |
|
"loss": 0.0934, |
|
"step": 1815 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 6.253974769040558e-05, |
|
"loss": 0.091, |
|
"step": 1818 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 6.242934058443181e-05, |
|
"loss": 0.1042, |
|
"step": 1821 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 6.231886884470944e-05, |
|
"loss": 0.0864, |
|
"step": 1824 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 6.2208333045702e-05, |
|
"loss": 0.0813, |
|
"step": 1827 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 6.209773376220612e-05, |
|
"loss": 0.0834, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 6.198707156934851e-05, |
|
"loss": 0.0844, |
|
"step": 1833 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 6.187634704258307e-05, |
|
"loss": 0.0975, |
|
"step": 1836 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 6.176556075768789e-05, |
|
"loss": 0.0865, |
|
"step": 1839 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 6.165471329076208e-05, |
|
"loss": 0.0873, |
|
"step": 1842 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 6.154380521822298e-05, |
|
"loss": 0.0867, |
|
"step": 1845 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 6.14328371168031e-05, |
|
"loss": 0.0894, |
|
"step": 1848 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 6.132180956354705e-05, |
|
"loss": 0.1123, |
|
"step": 1851 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 6.121072313580866e-05, |
|
"loss": 0.0859, |
|
"step": 1854 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 6.109957841124786e-05, |
|
"loss": 0.0843, |
|
"step": 1857 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 6.098837596782776e-05, |
|
"loss": 0.0851, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 6.087711638381157e-05, |
|
"loss": 0.0797, |
|
"step": 1863 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 6.0765800237759695e-05, |
|
"loss": 0.1038, |
|
"step": 1866 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 6.0654428108526615e-05, |
|
"loss": 0.0832, |
|
"step": 1869 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 6.0543000575257994e-05, |
|
"loss": 0.0878, |
|
"step": 1872 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 6.043151821738753e-05, |
|
"loss": 0.0967, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 6.0319981614634056e-05, |
|
"loss": 0.1, |
|
"step": 1878 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 6.020839134699847e-05, |
|
"loss": 0.1083, |
|
"step": 1881 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 6.009674799476073e-05, |
|
"loss": 0.1017, |
|
"step": 1884 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 5.998505213847686e-05, |
|
"loss": 0.0896, |
|
"step": 1887 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 5.987330435897589e-05, |
|
"loss": 0.1017, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 5.976150523735686e-05, |
|
"loss": 0.0933, |
|
"step": 1893 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 5.964965535498579e-05, |
|
"loss": 0.0818, |
|
"step": 1896 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 5.953775529349267e-05, |
|
"loss": 0.0822, |
|
"step": 1899 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 5.942580563476839e-05, |
|
"loss": 0.0895, |
|
"step": 1902 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 5.931380696096184e-05, |
|
"loss": 0.0741, |
|
"step": 1905 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 5.920175985447669e-05, |
|
"loss": 0.0825, |
|
"step": 1908 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 5.908966489796854e-05, |
|
"loss": 0.0982, |
|
"step": 1911 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 5.8977522674341756e-05, |
|
"loss": 0.0868, |
|
"step": 1914 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 5.8865333766746534e-05, |
|
"loss": 0.1042, |
|
"step": 1917 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 5.8753098758575805e-05, |
|
"loss": 0.0807, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 5.864081823346229e-05, |
|
"loss": 0.0764, |
|
"step": 1923 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 5.852849277527534e-05, |
|
"loss": 0.0831, |
|
"step": 1926 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 5.8416122968117945e-05, |
|
"loss": 0.0689, |
|
"step": 1929 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 5.830370939632378e-05, |
|
"loss": 0.083, |
|
"step": 1932 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 5.8191252644454056e-05, |
|
"loss": 0.0738, |
|
"step": 1935 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 5.807875329729454e-05, |
|
"loss": 0.073, |
|
"step": 1938 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 5.796621193985251e-05, |
|
"loss": 0.0803, |
|
"step": 1941 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 5.785362915735368e-05, |
|
"loss": 0.0753, |
|
"step": 1944 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 5.774100553523918e-05, |
|
"loss": 0.0974, |
|
"step": 1947 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 5.76283416591625e-05, |
|
"loss": 0.0806, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 5.7515638114986484e-05, |
|
"loss": 0.074, |
|
"step": 1953 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 5.740289548878024e-05, |
|
"loss": 0.0956, |
|
"step": 1956 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 5.72901143668161e-05, |
|
"loss": 0.0919, |
|
"step": 1959 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 5.7177295335566615e-05, |
|
"loss": 0.0884, |
|
"step": 1962 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 5.7064438981701393e-05, |
|
"loss": 0.0785, |
|
"step": 1965 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 5.695154589208417e-05, |
|
"loss": 0.0801, |
|
"step": 1968 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 5.683861665376973e-05, |
|
"loss": 0.0784, |
|
"step": 1971 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 5.6725651854000793e-05, |
|
"loss": 0.0827, |
|
"step": 1974 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 5.661265208020505e-05, |
|
"loss": 0.0795, |
|
"step": 1977 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 5.6499617919991985e-05, |
|
"loss": 0.0899, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 5.638654996114997e-05, |
|
"loss": 0.0842, |
|
"step": 1983 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 5.627344879164309e-05, |
|
"loss": 0.0855, |
|
"step": 1986 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 5.6160314999608155e-05, |
|
"loss": 0.0864, |
|
"step": 1989 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 5.604714917335159e-05, |
|
"loss": 0.0817, |
|
"step": 1992 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 5.5933951901346424e-05, |
|
"loss": 0.0811, |
|
"step": 1995 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 5.5820723772229186e-05, |
|
"loss": 0.0931, |
|
"step": 1998 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 5.570746537479687e-05, |
|
"loss": 0.0866, |
|
"step": 2001 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 5.559417729800389e-05, |
|
"loss": 0.0698, |
|
"step": 2004 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 5.548086013095897e-05, |
|
"loss": 0.0886, |
|
"step": 2007 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 5.53675144629221e-05, |
|
"loss": 0.0817, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 5.525414088330152e-05, |
|
"loss": 0.0779, |
|
"step": 2013 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 5.514073998165057e-05, |
|
"loss": 0.0874, |
|
"step": 2016 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 5.502731234766466e-05, |
|
"loss": 0.0869, |
|
"step": 2019 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 5.4913858571178265e-05, |
|
"loss": 0.0904, |
|
"step": 2022 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 5.480037924216177e-05, |
|
"loss": 0.0717, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 5.4686874950718435e-05, |
|
"loss": 0.0777, |
|
"step": 2028 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 5.457334628708132e-05, |
|
"loss": 0.077, |
|
"step": 2031 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 5.4459793841610206e-05, |
|
"loss": 0.0953, |
|
"step": 2034 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 5.434621820478863e-05, |
|
"loss": 0.0807, |
|
"step": 2037 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 5.42326199672206e-05, |
|
"loss": 0.0764, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 5.411899971962776e-05, |
|
"loss": 0.072, |
|
"step": 2043 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 5.400535805284614e-05, |
|
"loss": 0.0885, |
|
"step": 2046 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 5.389169555782319e-05, |
|
"loss": 0.091, |
|
"step": 2049 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 5.377801282561461e-05, |
|
"loss": 0.0772, |
|
"step": 2052 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 5.3664310447381425e-05, |
|
"loss": 0.0767, |
|
"step": 2055 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 5.3550589014386765e-05, |
|
"loss": 0.0851, |
|
"step": 2058 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 5.343684911799285e-05, |
|
"loss": 0.0778, |
|
"step": 2061 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 5.332309134965792e-05, |
|
"loss": 0.0865, |
|
"step": 2064 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 5.320931630093312e-05, |
|
"loss": 0.0766, |
|
"step": 2067 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 5.3095524563459495e-05, |
|
"loss": 0.076, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 5.2981716728964894e-05, |
|
"loss": 0.0844, |
|
"step": 2073 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 5.286789338926081e-05, |
|
"loss": 0.0756, |
|
"step": 2076 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 5.275405513623941e-05, |
|
"loss": 0.0732, |
|
"step": 2079 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 5.2640202561870364e-05, |
|
"loss": 0.0724, |
|
"step": 2082 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 5.252633625819786e-05, |
|
"loss": 0.0747, |
|
"step": 2085 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 5.241245681733748e-05, |
|
"loss": 0.0862, |
|
"step": 2088 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 5.22985648314731e-05, |
|
"loss": 0.0814, |
|
"step": 2091 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 5.2184660892853824e-05, |
|
"loss": 0.0741, |
|
"step": 2094 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 5.207074559379094e-05, |
|
"loss": 0.075, |
|
"step": 2097 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 5.195681952665478e-05, |
|
"loss": 0.0815, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 5.1842883283871666e-05, |
|
"loss": 0.0763, |
|
"step": 2103 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 5.172893745792091e-05, |
|
"loss": 0.0915, |
|
"step": 2106 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 5.161498264133157e-05, |
|
"loss": 0.0767, |
|
"step": 2109 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 5.1501019426679483e-05, |
|
"loss": 0.0777, |
|
"step": 2112 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 5.1387048406584196e-05, |
|
"loss": 0.083, |
|
"step": 2115 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 5.127307017370575e-05, |
|
"loss": 0.0833, |
|
"step": 2118 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 5.115908532074181e-05, |
|
"loss": 0.0807, |
|
"step": 2121 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 5.1045094440424377e-05, |
|
"loss": 0.0873, |
|
"step": 2124 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 5.093109812551685e-05, |
|
"loss": 0.0754, |
|
"step": 2127 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 5.0817096968810875e-05, |
|
"loss": 0.0729, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 5.070309156312324e-05, |
|
"loss": 0.0815, |
|
"step": 2133 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 5.058908250129285e-05, |
|
"loss": 0.0868, |
|
"step": 2136 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 5.0475070376177644e-05, |
|
"loss": 0.0708, |
|
"step": 2139 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 5.036105578065149e-05, |
|
"loss": 0.0761, |
|
"step": 2142 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 5.024703930760106e-05, |
|
"loss": 0.0923, |
|
"step": 2145 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 5.013302154992285e-05, |
|
"loss": 0.0694, |
|
"step": 2148 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 5.0019003100519946e-05, |
|
"loss": 0.0816, |
|
"step": 2151 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 4.990498455229913e-05, |
|
"loss": 0.0815, |
|
"step": 2154 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 4.979096649816766e-05, |
|
"loss": 0.0721, |
|
"step": 2157 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 4.9676949531030175e-05, |
|
"loss": 0.0711, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 4.9562934243785765e-05, |
|
"loss": 0.0874, |
|
"step": 2163 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 4.944892122932466e-05, |
|
"loss": 0.0818, |
|
"step": 2166 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 4.9334911080525394e-05, |
|
"loss": 0.0722, |
|
"step": 2169 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 4.9220904390251524e-05, |
|
"loss": 0.0642, |
|
"step": 2172 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 4.910690175134862e-05, |
|
"loss": 0.0793, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 4.899290375664124e-05, |
|
"loss": 0.0894, |
|
"step": 2178 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 4.887891099892972e-05, |
|
"loss": 0.0807, |
|
"step": 2181 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 4.8764924070987226e-05, |
|
"loss": 0.074, |
|
"step": 2184 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 4.865094356555657e-05, |
|
"loss": 0.0698, |
|
"step": 2187 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 4.8536970075347176e-05, |
|
"loss": 0.0706, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 4.842300419303202e-05, |
|
"loss": 0.0882, |
|
"step": 2193 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 4.8309046511244435e-05, |
|
"loss": 0.0876, |
|
"step": 2196 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 4.819509762257523e-05, |
|
"loss": 0.0851, |
|
"step": 2199 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 4.808115811956935e-05, |
|
"loss": 0.0804, |
|
"step": 2202 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 4.796722859472305e-05, |
|
"loss": 0.0854, |
|
"step": 2205 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 4.7853309640480645e-05, |
|
"loss": 0.0927, |
|
"step": 2208 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 4.773940184923148e-05, |
|
"loss": 0.073, |
|
"step": 2211 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 4.762550581330688e-05, |
|
"loss": 0.0812, |
|
"step": 2214 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 4.7511622124976995e-05, |
|
"loss": 0.0766, |
|
"step": 2217 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 4.7397751376447816e-05, |
|
"loss": 0.0779, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 4.728389415985803e-05, |
|
"loss": 0.0728, |
|
"step": 2223 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 4.7170051067275914e-05, |
|
"loss": 0.0634, |
|
"step": 2226 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 4.705622269069637e-05, |
|
"loss": 0.0787, |
|
"step": 2229 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 4.694240962203771e-05, |
|
"loss": 0.0746, |
|
"step": 2232 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 4.682861245313871e-05, |
|
"loss": 0.0768, |
|
"step": 2235 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 4.671483177575541e-05, |
|
"loss": 0.092, |
|
"step": 2238 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 4.6601068181558106e-05, |
|
"loss": 0.0776, |
|
"step": 2241 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 4.648732226212828e-05, |
|
"loss": 0.0801, |
|
"step": 2244 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 4.637359460895547e-05, |
|
"loss": 0.0731, |
|
"step": 2247 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 4.625988581343428e-05, |
|
"loss": 0.0752, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 4.614619646686117e-05, |
|
"loss": 0.0779, |
|
"step": 2253 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 4.6032527160431554e-05, |
|
"loss": 0.0833, |
|
"step": 2256 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 4.591887848523659e-05, |
|
"loss": 0.072, |
|
"step": 2259 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 4.580525103226013e-05, |
|
"loss": 0.078, |
|
"step": 2262 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 4.569164539237574e-05, |
|
"loss": 0.0832, |
|
"step": 2265 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 4.557806215634345e-05, |
|
"loss": 0.0768, |
|
"step": 2268 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 4.5464501914806894e-05, |
|
"loss": 0.0741, |
|
"step": 2271 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 4.535096525829009e-05, |
|
"loss": 0.0772, |
|
"step": 2274 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 4.5237452777194366e-05, |
|
"loss": 0.0717, |
|
"step": 2277 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 4.5123965061795425e-05, |
|
"loss": 0.0788, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 4.5010502702240096e-05, |
|
"loss": 0.0773, |
|
"step": 2283 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 4.4897066288543435e-05, |
|
"loss": 0.0784, |
|
"step": 2286 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 4.4783656410585515e-05, |
|
"loss": 0.071, |
|
"step": 2289 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 4.4670273658108473e-05, |
|
"loss": 0.0726, |
|
"step": 2292 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 4.455691862071335e-05, |
|
"loss": 0.0921, |
|
"step": 2295 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 4.444359188785705e-05, |
|
"loss": 0.0801, |
|
"step": 2298 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 4.433029404884936e-05, |
|
"loss": 0.0691, |
|
"step": 2301 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 4.4217025692849765e-05, |
|
"loss": 0.0839, |
|
"step": 2304 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 4.4103787408864444e-05, |
|
"loss": 0.0685, |
|
"step": 2307 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 4.3990579785743216e-05, |
|
"loss": 0.0752, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 4.387740341217641e-05, |
|
"loss": 0.0672, |
|
"step": 2313 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 4.376425887669194e-05, |
|
"loss": 0.0805, |
|
"step": 2316 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 4.365114676765206e-05, |
|
"loss": 0.0809, |
|
"step": 2319 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 4.353806767325049e-05, |
|
"loss": 0.0723, |
|
"step": 2322 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 4.342502218150923e-05, |
|
"loss": 0.0709, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 4.331201088027552e-05, |
|
"loss": 0.0737, |
|
"step": 2328 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 4.319903435721888e-05, |
|
"loss": 0.0633, |
|
"step": 2331 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 4.308609319982787e-05, |
|
"loss": 0.0686, |
|
"step": 2334 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 4.297318799540726e-05, |
|
"loss": 0.0725, |
|
"step": 2337 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 4.2860319331074786e-05, |
|
"loss": 0.0745, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 4.274748779375822e-05, |
|
"loss": 0.0718, |
|
"step": 2343 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 4.263469397019224e-05, |
|
"loss": 0.0594, |
|
"step": 2346 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 4.252193844691537e-05, |
|
"loss": 0.0794, |
|
"step": 2349 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 4.240922181026707e-05, |
|
"loss": 0.074, |
|
"step": 2352 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 4.229654464638449e-05, |
|
"loss": 0.0731, |
|
"step": 2355 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 4.218390754119958e-05, |
|
"loss": 0.0689, |
|
"step": 2358 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 4.207131108043596e-05, |
|
"loss": 0.0764, |
|
"step": 2361 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 4.1958755849605866e-05, |
|
"loss": 0.0688, |
|
"step": 2364 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 4.184624243400721e-05, |
|
"loss": 0.0722, |
|
"step": 2367 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 4.1733771418720354e-05, |
|
"loss": 0.0723, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 4.162134338860527e-05, |
|
"loss": 0.0775, |
|
"step": 2373 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 4.1508958928298376e-05, |
|
"loss": 0.0752, |
|
"step": 2376 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 4.139661862220947e-05, |
|
"loss": 0.0641, |
|
"step": 2379 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 4.128432305451884e-05, |
|
"loss": 0.072, |
|
"step": 2382 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 4.117207280917402e-05, |
|
"loss": 0.0611, |
|
"step": 2385 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 4.1059868469886956e-05, |
|
"loss": 0.0787, |
|
"step": 2388 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 4.094771062013083e-05, |
|
"loss": 0.0689, |
|
"step": 2391 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 4.083559984313712e-05, |
|
"loss": 0.072, |
|
"step": 2394 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 4.072353672189245e-05, |
|
"loss": 0.0723, |
|
"step": 2397 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 4.0611521839135666e-05, |
|
"loss": 0.0651, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 4.049955577735479e-05, |
|
"loss": 0.0654, |
|
"step": 2403 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 4.038763911878393e-05, |
|
"loss": 0.0749, |
|
"step": 2406 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 4.0275772445400344e-05, |
|
"loss": 0.0699, |
|
"step": 2409 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 4.016395633892132e-05, |
|
"loss": 0.0801, |
|
"step": 2412 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 4.0052191380801176e-05, |
|
"loss": 0.0713, |
|
"step": 2415 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 3.9940478152228315e-05, |
|
"loss": 0.0704, |
|
"step": 2418 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 3.982881723412211e-05, |
|
"loss": 0.0699, |
|
"step": 2421 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 3.971720920712989e-05, |
|
"loss": 0.0759, |
|
"step": 2424 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 3.9605654651624005e-05, |
|
"loss": 0.0691, |
|
"step": 2427 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 3.9494154147698666e-05, |
|
"loss": 0.0674, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 3.93827082751671e-05, |
|
"loss": 0.0632, |
|
"step": 2433 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 3.927131761355837e-05, |
|
"loss": 0.0917, |
|
"step": 2436 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 3.915998274211449e-05, |
|
"loss": 0.0634, |
|
"step": 2439 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 3.9048704239787345e-05, |
|
"loss": 0.0686, |
|
"step": 2442 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 3.8937482685235657e-05, |
|
"loss": 0.0693, |
|
"step": 2445 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 3.8826318656822084e-05, |
|
"loss": 0.0737, |
|
"step": 2448 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 3.8715212732610076e-05, |
|
"loss": 0.0655, |
|
"step": 2451 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 3.860416549036098e-05, |
|
"loss": 0.0714, |
|
"step": 2454 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 3.849317750753098e-05, |
|
"loss": 0.0705, |
|
"step": 2457 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 3.83822493612681e-05, |
|
"loss": 0.0624, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 3.8271381628409206e-05, |
|
"loss": 0.0787, |
|
"step": 2463 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 3.8160574885477e-05, |
|
"loss": 0.0835, |
|
"step": 2466 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 3.804982970867706e-05, |
|
"loss": 0.069, |
|
"step": 2469 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 3.79391466738948e-05, |
|
"loss": 0.0749, |
|
"step": 2472 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 3.7828526356692476e-05, |
|
"loss": 0.0711, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 3.7717969332306233e-05, |
|
"loss": 0.0636, |
|
"step": 2478 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 3.760747617564304e-05, |
|
"loss": 0.0573, |
|
"step": 2481 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 3.7497047461277825e-05, |
|
"loss": 0.0694, |
|
"step": 2484 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 3.738668376345032e-05, |
|
"loss": 0.0661, |
|
"step": 2487 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 3.727638565606223e-05, |
|
"loss": 0.086, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 3.716615371267417e-05, |
|
"loss": 0.0742, |
|
"step": 2493 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 3.705598850650267e-05, |
|
"loss": 0.0769, |
|
"step": 2496 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 3.694589061041727e-05, |
|
"loss": 0.075, |
|
"step": 2499 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 3.683586059693742e-05, |
|
"loss": 0.075, |
|
"step": 2502 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 3.672589903822964e-05, |
|
"loss": 0.0771, |
|
"step": 2505 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 3.661600650610444e-05, |
|
"loss": 0.0664, |
|
"step": 2508 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 3.650618357201343e-05, |
|
"loss": 0.0656, |
|
"step": 2511 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 3.6396430807046255e-05, |
|
"loss": 0.0615, |
|
"step": 2514 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 3.6286748781927654e-05, |
|
"loss": 0.0756, |
|
"step": 2517 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 3.617713806701459e-05, |
|
"loss": 0.0569, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 3.6067599232293145e-05, |
|
"loss": 0.0812, |
|
"step": 2523 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 3.595813284737567e-05, |
|
"loss": 0.0647, |
|
"step": 2526 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 3.5848739481497716e-05, |
|
"loss": 0.0693, |
|
"step": 2529 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 3.573941970351512e-05, |
|
"loss": 0.0758, |
|
"step": 2532 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 3.5630174081901125e-05, |
|
"loss": 0.0738, |
|
"step": 2535 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 3.552100318474331e-05, |
|
"loss": 0.0667, |
|
"step": 2538 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 3.5411907579740653e-05, |
|
"loss": 0.0589, |
|
"step": 2541 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 3.5302887834200684e-05, |
|
"loss": 0.0604, |
|
"step": 2544 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 3.5193944515036346e-05, |
|
"loss": 0.0773, |
|
"step": 2547 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 3.5085078188763285e-05, |
|
"loss": 0.0773, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 3.4976289421496655e-05, |
|
"loss": 0.0652, |
|
"step": 2553 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 3.486757877894838e-05, |
|
"loss": 0.0676, |
|
"step": 2556 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 3.47589468264241e-05, |
|
"loss": 0.066, |
|
"step": 2559 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 3.465039412882029e-05, |
|
"loss": 0.0632, |
|
"step": 2562 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 3.454192125062121e-05, |
|
"loss": 0.0564, |
|
"step": 2565 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 3.443352875589611e-05, |
|
"loss": 0.0745, |
|
"step": 2568 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 3.432521720829626e-05, |
|
"loss": 0.0489, |
|
"step": 2571 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 3.421698717105194e-05, |
|
"loss": 0.0684, |
|
"step": 2574 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 3.410883920696963e-05, |
|
"loss": 0.0657, |
|
"step": 2577 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 3.400077387842897e-05, |
|
"loss": 0.0581, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 3.389279174737989e-05, |
|
"loss": 0.0675, |
|
"step": 2583 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 3.378489337533973e-05, |
|
"loss": 0.0794, |
|
"step": 2586 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 3.367707932339024e-05, |
|
"loss": 0.0695, |
|
"step": 2589 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 3.3569350152174694e-05, |
|
"loss": 0.0609, |
|
"step": 2592 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 3.3461706421895003e-05, |
|
"loss": 0.0652, |
|
"step": 2595 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 3.3354148692308736e-05, |
|
"loss": 0.0675, |
|
"step": 2598 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 3.3246677522726324e-05, |
|
"loss": 0.0701, |
|
"step": 2601 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 3.313929347200797e-05, |
|
"loss": 0.0693, |
|
"step": 2604 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 3.303199709856096e-05, |
|
"loss": 0.0735, |
|
"step": 2607 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 3.2924788960336575e-05, |
|
"loss": 0.0652, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 3.281766961482732e-05, |
|
"loss": 0.0617, |
|
"step": 2613 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 3.271063961906392e-05, |
|
"loss": 0.0555, |
|
"step": 2616 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 3.260369952961248e-05, |
|
"loss": 0.0572, |
|
"step": 2619 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 3.24968499025716e-05, |
|
"loss": 0.0719, |
|
"step": 2622 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 3.239009129356948e-05, |
|
"loss": 0.0646, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 3.2283424257761005e-05, |
|
"loss": 0.0678, |
|
"step": 2628 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 3.2176849349824865e-05, |
|
"loss": 0.0711, |
|
"step": 2631 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 3.207036712396065e-05, |
|
"loss": 0.069, |
|
"step": 2634 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 3.196397813388605e-05, |
|
"loss": 0.0635, |
|
"step": 2637 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 3.1857682932833864e-05, |
|
"loss": 0.0713, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 3.175148207354927e-05, |
|
"loss": 0.0757, |
|
"step": 2643 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 3.164537610828674e-05, |
|
"loss": 0.063, |
|
"step": 2646 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 3.1539365588807346e-05, |
|
"loss": 0.07, |
|
"step": 2649 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 3.1433451066375856e-05, |
|
"loss": 0.0652, |
|
"step": 2652 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 3.132763309175782e-05, |
|
"loss": 0.0785, |
|
"step": 2655 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 3.122191221521672e-05, |
|
"loss": 0.0688, |
|
"step": 2658 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 3.111628898651113e-05, |
|
"loss": 0.0569, |
|
"step": 2661 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 3.101076395489187e-05, |
|
"loss": 0.0497, |
|
"step": 2664 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 3.090533766909908e-05, |
|
"loss": 0.0662, |
|
"step": 2667 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 3.080001067735942e-05, |
|
"loss": 0.0777, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 3.0694783527383255e-05, |
|
"loss": 0.0638, |
|
"step": 2673 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 3.058965676636171e-05, |
|
"loss": 0.0641, |
|
"step": 2676 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 3.0484630940963932e-05, |
|
"loss": 0.0578, |
|
"step": 2679 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 3.0379706597334146e-05, |
|
"loss": 0.0627, |
|
"step": 2682 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 3.0274884281088856e-05, |
|
"loss": 0.068, |
|
"step": 2685 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 3.0170164537314072e-05, |
|
"loss": 0.0632, |
|
"step": 2688 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 3.0065547910562364e-05, |
|
"loss": 0.0781, |
|
"step": 2691 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 2.996103494485013e-05, |
|
"loss": 0.0632, |
|
"step": 2694 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 2.985662618365467e-05, |
|
"loss": 0.0711, |
|
"step": 2697 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 2.975232216991141e-05, |
|
"loss": 0.0542, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 2.9648123446011146e-05, |
|
"loss": 0.0514, |
|
"step": 2703 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 2.9544030553797086e-05, |
|
"loss": 0.0641, |
|
"step": 2706 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 2.9440044034562163e-05, |
|
"loss": 0.0679, |
|
"step": 2709 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 2.9336164429046098e-05, |
|
"loss": 0.0663, |
|
"step": 2712 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 2.9232392277432695e-05, |
|
"loss": 0.0761, |
|
"step": 2715 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 2.912872811934698e-05, |
|
"loss": 0.07, |
|
"step": 2718 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 2.902517249385236e-05, |
|
"loss": 0.0603, |
|
"step": 2721 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 2.8921725939447965e-05, |
|
"loss": 0.0604, |
|
"step": 2724 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 2.8818388994065594e-05, |
|
"loss": 0.0606, |
|
"step": 2727 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 2.8715162195067214e-05, |
|
"loss": 0.064, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 2.861204607924197e-05, |
|
"loss": 0.0714, |
|
"step": 2733 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 2.8509041182803364e-05, |
|
"loss": 0.0553, |
|
"step": 2736 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 2.8406148041386698e-05, |
|
"loss": 0.0708, |
|
"step": 2739 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 2.830336719004598e-05, |
|
"loss": 0.0542, |
|
"step": 2742 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 2.820069916325144e-05, |
|
"loss": 0.0566, |
|
"step": 2745 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 2.8098144494886513e-05, |
|
"loss": 0.0507, |
|
"step": 2748 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 2.7995703718245157e-05, |
|
"loss": 0.0638, |
|
"step": 2751 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 2.78933773660292e-05, |
|
"loss": 0.067, |
|
"step": 2754 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 2.7791165970345278e-05, |
|
"loss": 0.0731, |
|
"step": 2757 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 2.7689070062702404e-05, |
|
"loss": 0.0752, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 2.758709017400891e-05, |
|
"loss": 0.0673, |
|
"step": 2763 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 2.748522683456992e-05, |
|
"loss": 0.0669, |
|
"step": 2766 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 2.7383480574084462e-05, |
|
"loss": 0.0704, |
|
"step": 2769 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 2.7281851921642672e-05, |
|
"loss": 0.0586, |
|
"step": 2772 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 2.7180341405723235e-05, |
|
"loss": 0.0663, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 2.7078949554190388e-05, |
|
"loss": 0.0686, |
|
"step": 2778 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 2.6977676894291414e-05, |
|
"loss": 0.0713, |
|
"step": 2781 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 2.6876523952653736e-05, |
|
"loss": 0.0606, |
|
"step": 2784 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 2.677549125528219e-05, |
|
"loss": 0.0624, |
|
"step": 2787 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 2.6674579327556436e-05, |
|
"loss": 0.0594, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 2.6573788694227997e-05, |
|
"loss": 0.0545, |
|
"step": 2793 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 2.6473119879417755e-05, |
|
"loss": 0.0596, |
|
"step": 2796 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 2.63725734066131e-05, |
|
"loss": 0.064, |
|
"step": 2799 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 2.6272149798665146e-05, |
|
"loss": 0.0585, |
|
"step": 2802 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 2.6171849577786262e-05, |
|
"loss": 0.0712, |
|
"step": 2805 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 2.6071673265547003e-05, |
|
"loss": 0.0606, |
|
"step": 2808 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 2.5971621382873774e-05, |
|
"loss": 0.0607, |
|
"step": 2811 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 2.5871694450045758e-05, |
|
"loss": 0.0614, |
|
"step": 2814 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 2.5771892986692537e-05, |
|
"loss": 0.0579, |
|
"step": 2817 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 2.5672217511791175e-05, |
|
"loss": 0.0475, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 2.5572668543663526e-05, |
|
"loss": 0.0618, |
|
"step": 2823 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 2.5473246599973722e-05, |
|
"loss": 0.0552, |
|
"step": 2826 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 2.5373952197725232e-05, |
|
"loss": 0.0527, |
|
"step": 2829 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 2.52747858532584e-05, |
|
"loss": 0.051, |
|
"step": 2832 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 2.517574808224761e-05, |
|
"loss": 0.053, |
|
"step": 2835 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 2.5076839399698605e-05, |
|
"loss": 0.0538, |
|
"step": 2838 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 2.497806031994598e-05, |
|
"loss": 0.0522, |
|
"step": 2841 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 2.4879411356650223e-05, |
|
"loss": 0.0565, |
|
"step": 2844 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 2.478089302279536e-05, |
|
"loss": 0.049, |
|
"step": 2847 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 2.4682505830686048e-05, |
|
"loss": 0.0673, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 2.458425029194496e-05, |
|
"loss": 0.0579, |
|
"step": 2853 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 2.448612691751027e-05, |
|
"loss": 0.0542, |
|
"step": 2856 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 2.4388136217632762e-05, |
|
"loss": 0.0554, |
|
"step": 2859 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 2.429027870187342e-05, |
|
"loss": 0.0464, |
|
"step": 2862 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 2.4192554879100548e-05, |
|
"loss": 0.0483, |
|
"step": 2865 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 2.4094965257487286e-05, |
|
"loss": 0.049, |
|
"step": 2868 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 2.3997510344508956e-05, |
|
"loss": 0.0604, |
|
"step": 2871 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 2.390019064694027e-05, |
|
"loss": 0.0524, |
|
"step": 2874 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 2.3803006670852956e-05, |
|
"loss": 0.051, |
|
"step": 2877 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 2.3705958921612814e-05, |
|
"loss": 0.0665, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 2.3609047903877392e-05, |
|
"loss": 0.0547, |
|
"step": 2883 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 2.3512274121593163e-05, |
|
"loss": 0.0584, |
|
"step": 2886 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 2.3415638077992914e-05, |
|
"loss": 0.0466, |
|
"step": 2889 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 2.3319140275593297e-05, |
|
"loss": 0.0603, |
|
"step": 2892 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 2.3222781216191948e-05, |
|
"loss": 0.0573, |
|
"step": 2895 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 2.3126561400865166e-05, |
|
"loss": 0.0534, |
|
"step": 2898 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 2.303048132996511e-05, |
|
"loss": 0.0608, |
|
"step": 2901 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 2.2934541503117185e-05, |
|
"loss": 0.0459, |
|
"step": 2904 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 2.2838742419217663e-05, |
|
"loss": 0.0732, |
|
"step": 2907 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 2.2743084576430775e-05, |
|
"loss": 0.0529, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 2.264756847218643e-05, |
|
"loss": 0.0563, |
|
"step": 2913 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 2.255219460317741e-05, |
|
"loss": 0.0591, |
|
"step": 2916 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 2.2456963465356822e-05, |
|
"loss": 0.0548, |
|
"step": 2919 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 2.236187555393568e-05, |
|
"loss": 0.0498, |
|
"step": 2922 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 2.226693136338005e-05, |
|
"loss": 0.0588, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 2.2172131387408796e-05, |
|
"loss": 0.0617, |
|
"step": 2928 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 2.2077476118990698e-05, |
|
"loss": 0.0503, |
|
"step": 2931 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 2.198296605034218e-05, |
|
"loss": 0.0523, |
|
"step": 2934 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 2.1888601672924548e-05, |
|
"loss": 0.0516, |
|
"step": 2937 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 2.1794383477441453e-05, |
|
"loss": 0.0518, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 2.170031195383651e-05, |
|
"loss": 0.0564, |
|
"step": 2943 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 2.1606387591290476e-05, |
|
"loss": 0.0551, |
|
"step": 2946 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 2.1512610878218998e-05, |
|
"loss": 0.061, |
|
"step": 2949 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 2.1418982302269864e-05, |
|
"loss": 0.052, |
|
"step": 2952 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 2.132550235032048e-05, |
|
"loss": 0.0624, |
|
"step": 2955 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 2.1232171508475535e-05, |
|
"loss": 0.0541, |
|
"step": 2958 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 2.1138990262064172e-05, |
|
"loss": 0.0599, |
|
"step": 2961 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 2.104595909563775e-05, |
|
"loss": 0.0465, |
|
"step": 2964 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 2.0953078492967155e-05, |
|
"loss": 0.0545, |
|
"step": 2967 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 2.086034893704026e-05, |
|
"loss": 0.0507, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 2.076777091005962e-05, |
|
"loss": 0.0491, |
|
"step": 2973 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 2.0675344893439662e-05, |
|
"loss": 0.0578, |
|
"step": 2976 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 2.05830713678045e-05, |
|
"loss": 0.0525, |
|
"step": 2979 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 2.049095081298518e-05, |
|
"loss": 0.0591, |
|
"step": 2982 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 2.0398983708017316e-05, |
|
"loss": 0.0604, |
|
"step": 2985 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 2.0307170531138585e-05, |
|
"loss": 0.0565, |
|
"step": 2988 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 2.0215511759786158e-05, |
|
"loss": 0.0577, |
|
"step": 2991 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 2.012400787059439e-05, |
|
"loss": 0.0541, |
|
"step": 2994 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 2.0032659339392112e-05, |
|
"loss": 0.0592, |
|
"step": 2997 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 1.9941466641200386e-05, |
|
"loss": 0.0619, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 1.9850430250229885e-05, |
|
"loss": 0.0459, |
|
"step": 3003 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 1.9759550639878422e-05, |
|
"loss": 0.0534, |
|
"step": 3006 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 1.9668828282728634e-05, |
|
"loss": 0.0554, |
|
"step": 3009 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 1.957826365054532e-05, |
|
"loss": 0.0581, |
|
"step": 3012 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 1.9487857214273192e-05, |
|
"loss": 0.0548, |
|
"step": 3015 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 1.9397609444034288e-05, |
|
"loss": 0.0584, |
|
"step": 3018 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 1.93075208091255e-05, |
|
"loss": 0.0557, |
|
"step": 3021 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 1.9217591778016346e-05, |
|
"loss": 0.0617, |
|
"step": 3024 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 1.9127822818346235e-05, |
|
"loss": 0.0611, |
|
"step": 3027 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 1.9038214396922317e-05, |
|
"loss": 0.056, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 1.8948766979716848e-05, |
|
"loss": 0.0518, |
|
"step": 3033 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 1.8859481031864874e-05, |
|
"loss": 0.0582, |
|
"step": 3036 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 1.87703570176618e-05, |
|
"loss": 0.0547, |
|
"step": 3039 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 1.8681395400560898e-05, |
|
"loss": 0.0552, |
|
"step": 3042 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 1.8592596643171068e-05, |
|
"loss": 0.0641, |
|
"step": 3045 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 1.8503961207254188e-05, |
|
"loss": 0.0654, |
|
"step": 3048 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 1.8415489553722988e-05, |
|
"loss": 0.0545, |
|
"step": 3051 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 1.832718214263843e-05, |
|
"loss": 0.0583, |
|
"step": 3054 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 1.8239039433207384e-05, |
|
"loss": 0.0453, |
|
"step": 3057 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 1.8151061883780347e-05, |
|
"loss": 0.0493, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 1.8063249951848844e-05, |
|
"loss": 0.0577, |
|
"step": 3063 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 1.7975604094043295e-05, |
|
"loss": 0.0542, |
|
"step": 3066 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 1.7888124766130464e-05, |
|
"loss": 0.0553, |
|
"step": 3069 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 1.7800812423011087e-05, |
|
"loss": 0.0614, |
|
"step": 3072 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 1.7713667518717697e-05, |
|
"loss": 0.0588, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 1.7626690506411974e-05, |
|
"loss": 0.0571, |
|
"step": 3078 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 1.7539881838382684e-05, |
|
"loss": 0.0635, |
|
"step": 3081 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 1.74532419660431e-05, |
|
"loss": 0.0498, |
|
"step": 3084 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 1.7366771339928774e-05, |
|
"loss": 0.0495, |
|
"step": 3087 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 1.728047040969516e-05, |
|
"loss": 0.0462, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 1.719433962411523e-05, |
|
"loss": 0.0467, |
|
"step": 3093 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 1.7108379431077282e-05, |
|
"loss": 0.0528, |
|
"step": 3096 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 1.7022590277582436e-05, |
|
"loss": 0.0472, |
|
"step": 3099 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 1.6936972609742417e-05, |
|
"loss": 0.0558, |
|
"step": 3102 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 1.685152687277724e-05, |
|
"loss": 0.0617, |
|
"step": 3105 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 1.6766253511012775e-05, |
|
"loss": 0.0541, |
|
"step": 3108 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 1.668115296787865e-05, |
|
"loss": 0.0418, |
|
"step": 3111 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 1.6596225685905675e-05, |
|
"loss": 0.0566, |
|
"step": 3114 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 1.651147210672383e-05, |
|
"loss": 0.0587, |
|
"step": 3117 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 1.642689267105974e-05, |
|
"loss": 0.0569, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 1.6342487818734443e-05, |
|
"loss": 0.0481, |
|
"step": 3123 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 1.625825798866123e-05, |
|
"loss": 0.0506, |
|
"step": 3126 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 1.6174203618843132e-05, |
|
"loss": 0.0571, |
|
"step": 3129 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 1.6090325146370887e-05, |
|
"loss": 0.0506, |
|
"step": 3132 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 1.600662300742048e-05, |
|
"loss": 0.0486, |
|
"step": 3135 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 1.5923097637250962e-05, |
|
"loss": 0.0493, |
|
"step": 3138 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 1.5839749470202193e-05, |
|
"loss": 0.0504, |
|
"step": 3141 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 1.5756578939692477e-05, |
|
"loss": 0.05, |
|
"step": 3144 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 1.5673586478216516e-05, |
|
"loss": 0.0497, |
|
"step": 3147 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 1.559077251734295e-05, |
|
"loss": 0.0615, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 1.5508137487712223e-05, |
|
"loss": 0.0523, |
|
"step": 3153 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 1.5425681819034326e-05, |
|
"loss": 0.0605, |
|
"step": 3156 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 1.534340594008651e-05, |
|
"loss": 0.0587, |
|
"step": 3159 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 1.52613102787112e-05, |
|
"loss": 0.053, |
|
"step": 3162 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 1.5179395261813544e-05, |
|
"loss": 0.0443, |
|
"step": 3165 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 1.5097661315359446e-05, |
|
"loss": 0.0536, |
|
"step": 3168 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 1.5016108864373169e-05, |
|
"loss": 0.0543, |
|
"step": 3171 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 1.4934738332935138e-05, |
|
"loss": 0.0538, |
|
"step": 3174 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 1.4853550144179884e-05, |
|
"loss": 0.0588, |
|
"step": 3177 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 1.4772544720293625e-05, |
|
"loss": 0.0529, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 1.4691722482512277e-05, |
|
"loss": 0.0465, |
|
"step": 3183 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 1.4611083851119122e-05, |
|
"loss": 0.0525, |
|
"step": 3186 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 1.4530629245442684e-05, |
|
"loss": 0.0469, |
|
"step": 3189 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 1.4450359083854553e-05, |
|
"loss": 0.0512, |
|
"step": 3192 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 1.4370273783767124e-05, |
|
"loss": 0.0578, |
|
"step": 3195 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 1.4290373761631593e-05, |
|
"loss": 0.047, |
|
"step": 3198 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 1.4210659432935625e-05, |
|
"loss": 0.0548, |
|
"step": 3201 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 1.4131131212201287e-05, |
|
"loss": 0.0474, |
|
"step": 3204 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 1.4051789512982877e-05, |
|
"loss": 0.0496, |
|
"step": 3207 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 1.3972634747864693e-05, |
|
"loss": 0.0558, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 1.389366732845907e-05, |
|
"loss": 0.0558, |
|
"step": 3213 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 1.3814887665404047e-05, |
|
"loss": 0.0479, |
|
"step": 3216 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 1.3736296168361334e-05, |
|
"loss": 0.0545, |
|
"step": 3219 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 1.3657893246014187e-05, |
|
"loss": 0.0522, |
|
"step": 3222 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 1.3579679306065173e-05, |
|
"loss": 0.049, |
|
"step": 3225 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 1.3501654755234261e-05, |
|
"loss": 0.0535, |
|
"step": 3228 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 1.3423819999256437e-05, |
|
"loss": 0.0588, |
|
"step": 3231 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 1.334617544287986e-05, |
|
"loss": 0.051, |
|
"step": 3234 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 1.3268721489863573e-05, |
|
"loss": 0.0434, |
|
"step": 3237 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 1.3191458542975427e-05, |
|
"loss": 0.0571, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 1.3114387003990125e-05, |
|
"loss": 0.0487, |
|
"step": 3243 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 1.303750727368691e-05, |
|
"loss": 0.0615, |
|
"step": 3246 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 1.2960819751847708e-05, |
|
"loss": 0.0435, |
|
"step": 3249 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 1.288432483725489e-05, |
|
"loss": 0.0459, |
|
"step": 3252 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 1.2808022927689251e-05, |
|
"loss": 0.0521, |
|
"step": 3255 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 1.273191441992797e-05, |
|
"loss": 0.0521, |
|
"step": 3258 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 1.265599970974245e-05, |
|
"loss": 0.0583, |
|
"step": 3261 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 1.2580279191896422e-05, |
|
"loss": 0.0551, |
|
"step": 3264 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 1.2504753260143736e-05, |
|
"loss": 0.0576, |
|
"step": 3267 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 1.242942230722639e-05, |
|
"loss": 0.0434, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 1.23542867248725e-05, |
|
"loss": 0.0538, |
|
"step": 3273 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 1.2279346903794159e-05, |
|
"loss": 0.0537, |
|
"step": 3276 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 1.2204603233685597e-05, |
|
"loss": 0.0527, |
|
"step": 3279 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 1.2130056103220926e-05, |
|
"loss": 0.0447, |
|
"step": 3282 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 1.2055705900052328e-05, |
|
"loss": 0.051, |
|
"step": 3285 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 1.1981553010807905e-05, |
|
"loss": 0.0587, |
|
"step": 3288 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 1.1907597821089656e-05, |
|
"loss": 0.0547, |
|
"step": 3291 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 1.1833840715471639e-05, |
|
"loss": 0.045, |
|
"step": 3294 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 1.1760282077497737e-05, |
|
"loss": 0.0562, |
|
"step": 3297 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 1.1686922289679864e-05, |
|
"loss": 0.0505, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 1.1613761733495865e-05, |
|
"loss": 0.0516, |
|
"step": 3303 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 1.1540800789387562e-05, |
|
"loss": 0.0519, |
|
"step": 3306 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 1.1468039836758793e-05, |
|
"loss": 0.0554, |
|
"step": 3309 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 1.1395479253973351e-05, |
|
"loss": 0.0517, |
|
"step": 3312 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 1.13231194183532e-05, |
|
"loss": 0.0565, |
|
"step": 3315 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 1.1250960706176312e-05, |
|
"loss": 0.057, |
|
"step": 3318 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 1.1179003492674834e-05, |
|
"loss": 0.0513, |
|
"step": 3321 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 1.1107248152033106e-05, |
|
"loss": 0.047, |
|
"step": 3324 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 1.1035695057385653e-05, |
|
"loss": 0.0454, |
|
"step": 3327 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 1.096434458081539e-05, |
|
"loss": 0.0505, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 1.0893197093351543e-05, |
|
"loss": 0.0476, |
|
"step": 3333 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 1.0822252964967782e-05, |
|
"loss": 0.0472, |
|
"step": 3336 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 1.0751512564580319e-05, |
|
"loss": 0.0493, |
|
"step": 3339 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 1.0680976260045889e-05, |
|
"loss": 0.0519, |
|
"step": 3342 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 1.0610644418160015e-05, |
|
"loss": 0.0449, |
|
"step": 3345 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 1.0540517404654881e-05, |
|
"loss": 0.0592, |
|
"step": 3348 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 1.0470595584197645e-05, |
|
"loss": 0.0554, |
|
"step": 3351 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 1.040087932038838e-05, |
|
"loss": 0.0555, |
|
"step": 3354 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 1.033136897575827e-05, |
|
"loss": 0.0435, |
|
"step": 3357 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 1.0262064911767706e-05, |
|
"loss": 0.044, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 1.0192967488804333e-05, |
|
"loss": 0.0496, |
|
"step": 3363 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 1.0124077066181353e-05, |
|
"loss": 0.045, |
|
"step": 3366 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 1.0055394002135466e-05, |
|
"loss": 0.0499, |
|
"step": 3369 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 9.986918653825117e-06, |
|
"loss": 0.0547, |
|
"step": 3372 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 9.918651377328619e-06, |
|
"loss": 0.0532, |
|
"step": 3375 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 9.850592527642232e-06, |
|
"loss": 0.0518, |
|
"step": 3378 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 9.782742458678473e-06, |
|
"loss": 0.0543, |
|
"step": 3381 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 9.715101523264102e-06, |
|
"loss": 0.0439, |
|
"step": 3384 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 9.647670073138399e-06, |
|
"loss": 0.0526, |
|
"step": 3387 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 9.580448458951303e-06, |
|
"loss": 0.0549, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 9.51343703026154e-06, |
|
"loss": 0.05, |
|
"step": 3393 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 9.446636135534932e-06, |
|
"loss": 0.056, |
|
"step": 3396 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 9.380046122142433e-06, |
|
"loss": 0.0545, |
|
"step": 3399 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 9.313667336358433e-06, |
|
"loss": 0.0603, |
|
"step": 3402 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 9.2475001233589e-06, |
|
"loss": 0.0486, |
|
"step": 3405 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 9.181544827219619e-06, |
|
"loss": 0.0524, |
|
"step": 3408 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 9.11580179091438e-06, |
|
"loss": 0.0466, |
|
"step": 3411 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 9.050271356313172e-06, |
|
"loss": 0.057, |
|
"step": 3414 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 8.984953864180495e-06, |
|
"loss": 0.0513, |
|
"step": 3417 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 8.91984965417349e-06, |
|
"loss": 0.0563, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 8.854959064840206e-06, |
|
"loss": 0.046, |
|
"step": 3423 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 8.79028243361788e-06, |
|
"loss": 0.0494, |
|
"step": 3426 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 8.725820096831077e-06, |
|
"loss": 0.0502, |
|
"step": 3429 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 8.661572389690098e-06, |
|
"loss": 0.048, |
|
"step": 3432 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 8.597539646289088e-06, |
|
"loss": 0.054, |
|
"step": 3435 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 8.533722199604388e-06, |
|
"loss": 0.0572, |
|
"step": 3438 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 8.470120381492769e-06, |
|
"loss": 0.0473, |
|
"step": 3441 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 8.40673452268968e-06, |
|
"loss": 0.0485, |
|
"step": 3444 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 8.343564952807626e-06, |
|
"loss": 0.0512, |
|
"step": 3447 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 8.280612000334342e-06, |
|
"loss": 0.0542, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 8.217875992631153e-06, |
|
"loss": 0.0622, |
|
"step": 3453 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 8.155357255931234e-06, |
|
"loss": 0.0494, |
|
"step": 3456 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 8.093056115337939e-06, |
|
"loss": 0.047, |
|
"step": 3459 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 8.030972894823119e-06, |
|
"loss": 0.0465, |
|
"step": 3462 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 7.969107917225365e-06, |
|
"loss": 0.0451, |
|
"step": 3465 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 7.907461504248453e-06, |
|
"loss": 0.0542, |
|
"step": 3468 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 7.846033976459566e-06, |
|
"loss": 0.0528, |
|
"step": 3471 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 7.784825653287664e-06, |
|
"loss": 0.0505, |
|
"step": 3474 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 7.723836853021832e-06, |
|
"loss": 0.0545, |
|
"step": 3477 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 7.663067892809589e-06, |
|
"loss": 0.056, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 7.602519088655307e-06, |
|
"loss": 0.0447, |
|
"step": 3483 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 7.542190755418499e-06, |
|
"loss": 0.0488, |
|
"step": 3486 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 7.4820832068122125e-06, |
|
"loss": 0.0547, |
|
"step": 3489 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 7.422196755401406e-06, |
|
"loss": 0.0498, |
|
"step": 3492 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 7.362531712601261e-06, |
|
"loss": 0.0405, |
|
"step": 3495 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 7.303088388675694e-06, |
|
"loss": 0.0459, |
|
"step": 3498 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 7.243867092735612e-06, |
|
"loss": 0.05, |
|
"step": 3501 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 7.1848681327373545e-06, |
|
"loss": 0.0503, |
|
"step": 3504 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 7.126091815481112e-06, |
|
"loss": 0.0528, |
|
"step": 3507 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 7.067538446609295e-06, |
|
"loss": 0.0487, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 7.009208330604971e-06, |
|
"loss": 0.0492, |
|
"step": 3513 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 6.9511017707902806e-06, |
|
"loss": 0.0525, |
|
"step": 3516 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 6.893219069324824e-06, |
|
"loss": 0.0384, |
|
"step": 3519 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 6.835560527204143e-06, |
|
"loss": 0.0474, |
|
"step": 3522 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 6.77812644425811e-06, |
|
"loss": 0.0514, |
|
"step": 3525 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 6.720917119149417e-06, |
|
"loss": 0.0552, |
|
"step": 3528 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 6.66393284937194e-06, |
|
"loss": 0.0527, |
|
"step": 3531 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 6.607173931249316e-06, |
|
"loss": 0.0505, |
|
"step": 3534 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 6.550640659933277e-06, |
|
"loss": 0.0474, |
|
"step": 3537 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 6.4943333294021915e-06, |
|
"loss": 0.0466, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 6.438252232459535e-06, |
|
"loss": 0.0558, |
|
"step": 3543 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 6.382397660732276e-06, |
|
"loss": 0.0514, |
|
"step": 3546 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 6.326769904669521e-06, |
|
"loss": 0.0425, |
|
"step": 3549 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 6.271369253540865e-06, |
|
"loss": 0.0463, |
|
"step": 3552 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 6.216195995434931e-06, |
|
"loss": 0.0487, |
|
"step": 3555 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 6.161250417257902e-06, |
|
"loss": 0.0459, |
|
"step": 3558 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 6.1065328047319736e-06, |
|
"loss": 0.0517, |
|
"step": 3561 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 6.052043442393928e-06, |
|
"loss": 0.0542, |
|
"step": 3564 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 5.997782613593606e-06, |
|
"loss": 0.0488, |
|
"step": 3567 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 5.943750600492459e-06, |
|
"loss": 0.0472, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 5.889947684062075e-06, |
|
"loss": 0.0559, |
|
"step": 3573 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 5.836374144082718e-06, |
|
"loss": 0.0466, |
|
"step": 3576 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 5.783030259141881e-06, |
|
"loss": 0.0558, |
|
"step": 3579 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 5.729916306632782e-06, |
|
"loss": 0.054, |
|
"step": 3582 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 5.677032562753049e-06, |
|
"loss": 0.0515, |
|
"step": 3585 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 5.6243793025031354e-06, |
|
"loss": 0.0518, |
|
"step": 3588 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 5.571956799684996e-06, |
|
"loss": 0.0498, |
|
"step": 3591 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 5.519765326900617e-06, |
|
"loss": 0.0542, |
|
"step": 3594 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 5.46780515555057e-06, |
|
"loss": 0.0532, |
|
"step": 3597 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 5.416076555832706e-06, |
|
"loss": 0.0524, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 5.3645797967406204e-06, |
|
"loss": 0.0427, |
|
"step": 3603 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 5.313315146062348e-06, |
|
"loss": 0.0436, |
|
"step": 3606 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 5.2622828703789315e-06, |
|
"loss": 0.0347, |
|
"step": 3609 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 5.2114832350630365e-06, |
|
"loss": 0.0506, |
|
"step": 3612 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 5.16091650427758e-06, |
|
"loss": 0.0509, |
|
"step": 3615 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 5.110582940974351e-06, |
|
"loss": 0.0543, |
|
"step": 3618 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 5.060482806892647e-06, |
|
"loss": 0.0509, |
|
"step": 3621 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 5.0106163625579124e-06, |
|
"loss": 0.0503, |
|
"step": 3624 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 4.960983867280384e-06, |
|
"loss": 0.0494, |
|
"step": 3627 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 4.911585579153732e-06, |
|
"loss": 0.0571, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 4.862421755053747e-06, |
|
"loss": 0.0474, |
|
"step": 3633 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 4.813492650636964e-06, |
|
"loss": 0.0461, |
|
"step": 3636 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 4.764798520339364e-06, |
|
"loss": 0.0474, |
|
"step": 3639 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 4.7163396173750375e-06, |
|
"loss": 0.0522, |
|
"step": 3642 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 4.668116193734878e-06, |
|
"loss": 0.0494, |
|
"step": 3645 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 4.620128500185233e-06, |
|
"loss": 0.0554, |
|
"step": 3648 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 4.57237678626668e-06, |
|
"loss": 0.0473, |
|
"step": 3651 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 4.5248613002926485e-06, |
|
"loss": 0.0449, |
|
"step": 3654 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 4.477582289348164e-06, |
|
"loss": 0.0494, |
|
"step": 3657 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 4.430539999288574e-06, |
|
"loss": 0.0439, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 4.383734674738216e-06, |
|
"loss": 0.0499, |
|
"step": 3663 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 4.337166559089239e-06, |
|
"loss": 0.05, |
|
"step": 3666 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 4.2908358945002505e-06, |
|
"loss": 0.0428, |
|
"step": 3669 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 4.24474292189509e-06, |
|
"loss": 0.0549, |
|
"step": 3672 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 4.198887880961583e-06, |
|
"loss": 0.0508, |
|
"step": 3675 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 4.153271010150289e-06, |
|
"loss": 0.0557, |
|
"step": 3678 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 4.107892546673259e-06, |
|
"loss": 0.0537, |
|
"step": 3681 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 4.0627527265028e-06, |
|
"loss": 0.0565, |
|
"step": 3684 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 4.017851784370252e-06, |
|
"loss": 0.0481, |
|
"step": 3687 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 3.973189953764766e-06, |
|
"loss": 0.047, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 3.928767466932104e-06, |
|
"loss": 0.0457, |
|
"step": 3693 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 3.884584554873405e-06, |
|
"loss": 0.0396, |
|
"step": 3696 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 3.840641447343968e-06, |
|
"loss": 0.05, |
|
"step": 3699 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 3.7969383728521544e-06, |
|
"loss": 0.0419, |
|
"step": 3702 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 3.7534755586580707e-06, |
|
"loss": 0.0404, |
|
"step": 3705 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 3.7102532307724757e-06, |
|
"loss": 0.0469, |
|
"step": 3708 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 3.667271613955575e-06, |
|
"loss": 0.0473, |
|
"step": 3711 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 3.6245309317158193e-06, |
|
"loss": 0.0619, |
|
"step": 3714 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 3.5820314063088388e-06, |
|
"loss": 0.0464, |
|
"step": 3717 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 3.5397732587361785e-06, |
|
"loss": 0.0498, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 3.4977567087442153e-06, |
|
"loss": 0.0479, |
|
"step": 3723 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 3.4559819748229807e-06, |
|
"loss": 0.0451, |
|
"step": 3726 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 3.4144492742050503e-06, |
|
"loss": 0.0503, |
|
"step": 3729 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 3.3731588228644086e-06, |
|
"loss": 0.051, |
|
"step": 3732 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 3.3321108355153063e-06, |
|
"loss": 0.054, |
|
"step": 3735 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 3.291305525611177e-06, |
|
"loss": 0.0451, |
|
"step": 3738 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 3.25074310534349e-06, |
|
"loss": 0.0402, |
|
"step": 3741 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 3.2104237856406703e-06, |
|
"loss": 0.0565, |
|
"step": 3744 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 3.1703477761670095e-06, |
|
"loss": 0.0441, |
|
"step": 3747 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 3.1305152853215504e-06, |
|
"loss": 0.0559, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 3.0909265202370197e-06, |
|
"loss": 0.0509, |
|
"step": 3753 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 3.0515816867787426e-06, |
|
"loss": 0.0435, |
|
"step": 3756 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 3.012480989543581e-06, |
|
"loss": 0.0566, |
|
"step": 3759 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 2.973624631858879e-06, |
|
"loss": 0.0496, |
|
"step": 3762 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 2.9350128157813462e-06, |
|
"loss": 0.0436, |
|
"step": 3765 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 2.896645742096116e-06, |
|
"loss": 0.0534, |
|
"step": 3768 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 2.858523610315589e-06, |
|
"loss": 0.0392, |
|
"step": 3771 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 2.8206466186784743e-06, |
|
"loss": 0.043, |
|
"step": 3774 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 2.783014964148706e-06, |
|
"loss": 0.0492, |
|
"step": 3777 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 2.7456288424144547e-06, |
|
"loss": 0.0528, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 2.7084884478870977e-06, |
|
"loss": 0.0489, |
|
"step": 3783 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 2.6715939737002095e-06, |
|
"loss": 0.0688, |
|
"step": 3786 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 2.6349456117085447e-06, |
|
"loss": 0.0478, |
|
"step": 3789 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 2.5985435524870594e-06, |
|
"loss": 0.0455, |
|
"step": 3792 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 2.5623879853299115e-06, |
|
"loss": 0.0557, |
|
"step": 3795 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 2.5264790982494733e-06, |
|
"loss": 0.0462, |
|
"step": 3798 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 2.490817077975366e-06, |
|
"loss": 0.0453, |
|
"step": 3801 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 2.455402109953464e-06, |
|
"loss": 0.0491, |
|
"step": 3804 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 2.42023437834496e-06, |
|
"loss": 0.0526, |
|
"step": 3807 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 2.385314066025396e-06, |
|
"loss": 0.0478, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 2.3506413545837047e-06, |
|
"loss": 0.0479, |
|
"step": 3813 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 2.3162164243212493e-06, |
|
"loss": 0.0605, |
|
"step": 3816 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 2.2820394542509506e-06, |
|
"loss": 0.0467, |
|
"step": 3819 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 2.248110622096289e-06, |
|
"loss": 0.0542, |
|
"step": 3822 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 2.214430104290399e-06, |
|
"loss": 0.0468, |
|
"step": 3825 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 2.1809980759751715e-06, |
|
"loss": 0.0534, |
|
"step": 3828 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 2.1478147110003122e-06, |
|
"loss": 0.0555, |
|
"step": 3831 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 2.1148801819224683e-06, |
|
"loss": 0.052, |
|
"step": 3834 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 2.0821946600043163e-06, |
|
"loss": 0.0468, |
|
"step": 3837 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 2.0497583152136513e-06, |
|
"loss": 0.0529, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 2.017571316222555e-06, |
|
"loss": 0.0443, |
|
"step": 3843 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 1.985633830406464e-06, |
|
"loss": 0.0535, |
|
"step": 3846 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 1.9539460238433283e-06, |
|
"loss": 0.0526, |
|
"step": 3849 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 1.92250806131275e-06, |
|
"loss": 0.0456, |
|
"step": 3852 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 1.891320106295108e-06, |
|
"loss": 0.0521, |
|
"step": 3855 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 1.8603823209707271e-06, |
|
"loss": 0.0411, |
|
"step": 3858 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 1.8296948662190226e-06, |
|
"loss": 0.0433, |
|
"step": 3861 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 1.7992579016176625e-06, |
|
"loss": 0.0475, |
|
"step": 3864 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 1.769071585441745e-06, |
|
"loss": 0.0495, |
|
"step": 3867 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 1.7391360746629782e-06, |
|
"loss": 0.0536, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 1.7094515249488518e-06, |
|
"loss": 0.0544, |
|
"step": 3873 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 1.680018090661839e-06, |
|
"loss": 0.0567, |
|
"step": 3876 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 1.6508359248585848e-06, |
|
"loss": 0.0468, |
|
"step": 3879 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 1.6219051792891183e-06, |
|
"loss": 0.0478, |
|
"step": 3882 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 1.5932260043960535e-06, |
|
"loss": 0.0486, |
|
"step": 3885 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 1.564798549313823e-06, |
|
"loss": 0.0353, |
|
"step": 3888 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 1.5366229618678784e-06, |
|
"loss": 0.0516, |
|
"step": 3891 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 1.5086993885739476e-06, |
|
"loss": 0.0485, |
|
"step": 3894 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 1.4810279746372614e-06, |
|
"loss": 0.0442, |
|
"step": 3897 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 1.4536088639517887e-06, |
|
"loss": 0.0409, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 1.4264421990995036e-06, |
|
"loss": 0.0557, |
|
"step": 3903 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 1.3995281213496359e-06, |
|
"loss": 0.0489, |
|
"step": 3906 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 1.3728667706579434e-06, |
|
"loss": 0.0544, |
|
"step": 3909 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 1.3464582856659747e-06, |
|
"loss": 0.0473, |
|
"step": 3912 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 1.3203028037003518e-06, |
|
"loss": 0.0514, |
|
"step": 3915 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 1.294400460772055e-06, |
|
"loss": 0.0439, |
|
"step": 3918 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 1.2687513915757342e-06, |
|
"loss": 0.0489, |
|
"step": 3921 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 1.2433557294889653e-06, |
|
"loss": 0.0537, |
|
"step": 3924 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 1.2182136065716054e-06, |
|
"loss": 0.049, |
|
"step": 3927 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 1.1933251535650726e-06, |
|
"loss": 0.0543, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 1.168690499891678e-06, |
|
"loss": 0.0482, |
|
"step": 3933 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 1.1443097736539621e-06, |
|
"loss": 0.0511, |
|
"step": 3936 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 1.1201831016340036e-06, |
|
"loss": 0.048, |
|
"step": 3939 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 1.0963106092927888e-06, |
|
"loss": 0.0499, |
|
"step": 3942 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 1.0726924207695387e-06, |
|
"loss": 0.0444, |
|
"step": 3945 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 1.0493286588810657e-06, |
|
"loss": 0.045, |
|
"step": 3948 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 1.026219445121146e-06, |
|
"loss": 0.0487, |
|
"step": 3951 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 1.0033648996598866e-06, |
|
"loss": 0.0489, |
|
"step": 3954 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 9.807651413430874e-07, |
|
"loss": 0.048, |
|
"step": 3957 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 9.584202876916304e-07, |
|
"loss": 0.0487, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 9.363304549008689e-07, |
|
"loss": 0.0487, |
|
"step": 3963 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 9.144957578400281e-07, |
|
"loss": 0.0463, |
|
"step": 3966 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 8.929163100516058e-07, |
|
"loss": 0.0458, |
|
"step": 3969 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 8.715922237507723e-07, |
|
"loss": 0.0544, |
|
"step": 3972 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 8.505236098247993e-07, |
|
"loss": 0.0497, |
|
"step": 3975 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 8.297105778324654e-07, |
|
"loss": 0.0465, |
|
"step": 3978 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 8.091532360035182e-07, |
|
"loss": 0.0538, |
|
"step": 3981 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 7.888516912380794e-07, |
|
"loss": 0.0511, |
|
"step": 3984 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 7.688060491061078e-07, |
|
"loss": 0.048, |
|
"step": 3987 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 7.490164138468369e-07, |
|
"loss": 0.0564, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 7.294828883682492e-07, |
|
"loss": 0.044, |
|
"step": 3993 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 7.102055742465252e-07, |
|
"loss": 0.044, |
|
"step": 3996 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 6.911845717255283e-07, |
|
"loss": 0.0476, |
|
"step": 3999 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 6.724199797162711e-07, |
|
"loss": 0.0581, |
|
"step": 4002 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 6.539118957964052e-07, |
|
"loss": 0.0562, |
|
"step": 4005 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 6.356604162097268e-07, |
|
"loss": 0.0509, |
|
"step": 4008 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 6.176656358656608e-07, |
|
"loss": 0.0568, |
|
"step": 4011 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 5.999276483387605e-07, |
|
"loss": 0.0457, |
|
"step": 4014 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 5.824465458682538e-07, |
|
"loss": 0.0387, |
|
"step": 4017 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 5.65222419357514e-07, |
|
"loss": 0.0464, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 5.482553583736449e-07, |
|
"loss": 0.0582, |
|
"step": 4023 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 5.315454511469698e-07, |
|
"loss": 0.0537, |
|
"step": 4026 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 5.150927845705811e-07, |
|
"loss": 0.0486, |
|
"step": 4029 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 4.988974441999084e-07, |
|
"loss": 0.0537, |
|
"step": 4032 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 4.82959514252257e-07, |
|
"loss": 0.0456, |
|
"step": 4035 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 4.672790776063696e-07, |
|
"loss": 0.0468, |
|
"step": 4038 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 4.518562158020045e-07, |
|
"loss": 0.0491, |
|
"step": 4041 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 4.366910090395027e-07, |
|
"loss": 0.043, |
|
"step": 4044 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 4.2178353617938207e-07, |
|
"loss": 0.0456, |
|
"step": 4047 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 4.0713387474190514e-07, |
|
"loss": 0.0522, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 3.927421009067012e-07, |
|
"loss": 0.0474, |
|
"step": 4053 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 3.7860828951236103e-07, |
|
"loss": 0.0476, |
|
"step": 4056 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 3.6473251405604317e-07, |
|
"loss": 0.0492, |
|
"step": 4059 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 3.5111484669309024e-07, |
|
"loss": 0.0601, |
|
"step": 4062 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 3.377553582366688e-07, |
|
"loss": 0.0495, |
|
"step": 4065 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 3.2465411815738036e-07, |
|
"loss": 0.0512, |
|
"step": 4068 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 3.1181119458291164e-07, |
|
"loss": 0.0375, |
|
"step": 4071 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 2.992266542976907e-07, |
|
"loss": 0.0502, |
|
"step": 4074 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 2.869005627425092e-07, |
|
"loss": 0.0515, |
|
"step": 4077 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 2.748329840142172e-07, |
|
"loss": 0.0606, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 2.6302398086536784e-07, |
|
"loss": 0.0459, |
|
"step": 4083 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 2.514736147038954e-07, |
|
"loss": 0.0495, |
|
"step": 4086 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 2.401819455927934e-07, |
|
"loss": 0.0461, |
|
"step": 4089 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 2.291490322498202e-07, |
|
"loss": 0.0386, |
|
"step": 4092 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 2.1837493204716064e-07, |
|
"loss": 0.0509, |
|
"step": 4095 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 2.0785970101115938e-07, |
|
"loss": 0.0541, |
|
"step": 4098 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 1.976033938220212e-07, |
|
"loss": 0.0521, |
|
"step": 4101 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 1.8760606381350575e-07, |
|
"loss": 0.0504, |
|
"step": 4104 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 1.7786776297268327e-07, |
|
"loss": 0.0496, |
|
"step": 4107 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 1.683885419396347e-07, |
|
"loss": 0.0411, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 1.591684500072077e-07, |
|
"loss": 0.0573, |
|
"step": 4113 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 1.5020753512076104e-07, |
|
"loss": 0.0547, |
|
"step": 4116 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 1.4150584387789267e-07, |
|
"loss": 0.049, |
|
"step": 4119 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 1.3306342152822893e-07, |
|
"loss": 0.0468, |
|
"step": 4122 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 1.248803119731634e-07, |
|
"loss": 0.0454, |
|
"step": 4125 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 1.1695655776563508e-07, |
|
"loss": 0.0436, |
|
"step": 4128 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 1.0929220010992281e-07, |
|
"loss": 0.0527, |
|
"step": 4131 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 1.0188727886141225e-07, |
|
"loss": 0.0495, |
|
"step": 4134 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 9.474183252639601e-08, |
|
"loss": 0.0489, |
|
"step": 4137 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 8.785589826186825e-08, |
|
"loss": 0.0474, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 8.122951187534144e-08, |
|
"loss": 0.0508, |
|
"step": 4143 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 7.486270782465222e-08, |
|
"loss": 0.048, |
|
"step": 4146 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 6.875551921778357e-08, |
|
"loss": 0.0614, |
|
"step": 4149 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 6.290797781269842e-08, |
|
"loss": 0.0474, |
|
"step": 4152 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 5.7320114017161976e-08, |
|
"loss": 0.0494, |
|
"step": 4155 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 5.199195688860292e-08, |
|
"loss": 0.0493, |
|
"step": 4158 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 4.692353413393025e-08, |
|
"loss": 0.0523, |
|
"step": 4161 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 4.2114872109438876e-08, |
|
"loss": 0.0512, |
|
"step": 4164 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 3.756599582060982e-08, |
|
"loss": 0.0444, |
|
"step": 4167 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 3.3276928922043595e-08, |
|
"loss": 0.0579, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 2.9247693717288083e-08, |
|
"loss": 0.0397, |
|
"step": 4173 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 2.5478311158749768e-08, |
|
"loss": 0.043, |
|
"step": 4176 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 2.196880084757158e-08, |
|
"loss": 0.038, |
|
"step": 4179 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 1.8719181033549637e-08, |
|
"loss": 0.0533, |
|
"step": 4182 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 1.572946861500557e-08, |
|
"loss": 0.0576, |
|
"step": 4185 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 1.2999679138725462e-08, |
|
"loss": 0.0454, |
|
"step": 4188 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 1.0529826799882125e-08, |
|
"loss": 0.0467, |
|
"step": 4191 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 8.31992444193519e-09, |
|
"loss": 0.0484, |
|
"step": 4194 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 6.3699835565922406e-09, |
|
"loss": 0.0481, |
|
"step": 4197 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 4.680014283725553e-09, |
|
"loss": 0.0476, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 3.250025411333235e-09, |
|
"loss": 0.0444, |
|
"step": 4203 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 2.0800243755003667e-09, |
|
"loss": 0.0439, |
|
"step": 4206 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 1.17001726033239e-09, |
|
"loss": 0.0498, |
|
"step": 4209 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 5.200087979662093e-10, |
|
"loss": 0.0476, |
|
"step": 4212 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 1.3000236849802783e-10, |
|
"loss": 0.0439, |
|
"step": 4215 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 0.0, |
|
"loss": 0.0418, |
|
"step": 4218 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"step": 4218, |
|
"total_flos": 3.2919915274960896e+17, |
|
"train_loss": 0.18810525719710525, |
|
"train_runtime": 27776.8918, |
|
"train_samples_per_second": 19.449, |
|
"train_steps_per_second": 0.152 |
|
} |
|
], |
|
"max_steps": 4218, |
|
"num_train_epochs": 3, |
|
"total_flos": 3.2919915274960896e+17, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|