|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.0, |
|
"eval_steps": 500, |
|
"global_step": 920, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.008695652173913044, |
|
"grad_norm": 14.8125, |
|
"learning_rate": 8.000000000000001e-07, |
|
"loss": 1.8918, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.017391304347826087, |
|
"grad_norm": 7.8125, |
|
"learning_rate": 1.6000000000000001e-06, |
|
"loss": 1.816, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.02608695652173913, |
|
"grad_norm": 3.171875, |
|
"learning_rate": 2.4000000000000003e-06, |
|
"loss": 1.7715, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.034782608695652174, |
|
"grad_norm": 4.0, |
|
"learning_rate": 3.2000000000000003e-06, |
|
"loss": 1.7775, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.043478260869565216, |
|
"grad_norm": 4.5625, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 1.7134, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.05217391304347826, |
|
"grad_norm": 5.84375, |
|
"learning_rate": 4.800000000000001e-06, |
|
"loss": 1.7756, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.06086956521739131, |
|
"grad_norm": 6.78125, |
|
"learning_rate": 5.600000000000001e-06, |
|
"loss": 1.7208, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.06956521739130435, |
|
"grad_norm": 6.65625, |
|
"learning_rate": 6.4000000000000006e-06, |
|
"loss": 1.6998, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.0782608695652174, |
|
"grad_norm": 8.75, |
|
"learning_rate": 7.2000000000000005e-06, |
|
"loss": 1.8301, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.08695652173913043, |
|
"grad_norm": 7.0625, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 1.7552, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.09565217391304348, |
|
"grad_norm": 4.5625, |
|
"learning_rate": 8.8e-06, |
|
"loss": 1.7062, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.10434782608695652, |
|
"grad_norm": 5.375, |
|
"learning_rate": 9.600000000000001e-06, |
|
"loss": 1.7292, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.11304347826086956, |
|
"grad_norm": 7.53125, |
|
"learning_rate": 9.999869605476246e-06, |
|
"loss": 1.8108, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.12173913043478261, |
|
"grad_norm": 9.125, |
|
"learning_rate": 9.99882649009242e-06, |
|
"loss": 1.8255, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.13043478260869565, |
|
"grad_norm": 6.78125, |
|
"learning_rate": 9.996740476948386e-06, |
|
"loss": 1.7476, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.1391304347826087, |
|
"grad_norm": 7.09375, |
|
"learning_rate": 9.99361200124597e-06, |
|
"loss": 1.7619, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.14782608695652175, |
|
"grad_norm": 4.75, |
|
"learning_rate": 9.989441715674422e-06, |
|
"loss": 1.7593, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.1565217391304348, |
|
"grad_norm": 6.21875, |
|
"learning_rate": 9.98423049027424e-06, |
|
"loss": 1.7353, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.16521739130434782, |
|
"grad_norm": 6.03125, |
|
"learning_rate": 9.977979412255651e-06, |
|
"loss": 1.7301, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.17391304347826086, |
|
"grad_norm": 6.78125, |
|
"learning_rate": 9.970689785771798e-06, |
|
"loss": 1.7558, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.1826086956521739, |
|
"grad_norm": 5.6875, |
|
"learning_rate": 9.962363131646649e-06, |
|
"loss": 1.8001, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.19130434782608696, |
|
"grad_norm": 8.875, |
|
"learning_rate": 9.953001187057705e-06, |
|
"loss": 1.751, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 6.3125, |
|
"learning_rate": 9.942605905173593e-06, |
|
"loss": 1.7262, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.20869565217391303, |
|
"grad_norm": 6.28125, |
|
"learning_rate": 9.931179454746556e-06, |
|
"loss": 1.8109, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.21739130434782608, |
|
"grad_norm": 6.46875, |
|
"learning_rate": 9.918724219660013e-06, |
|
"loss": 1.9357, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.22608695652173913, |
|
"grad_norm": 6.25, |
|
"learning_rate": 9.905242798431196e-06, |
|
"loss": 1.6968, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.23478260869565218, |
|
"grad_norm": 7.8125, |
|
"learning_rate": 9.890738003669029e-06, |
|
"loss": 1.8129, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.24347826086956523, |
|
"grad_norm": 5.6875, |
|
"learning_rate": 9.87521286148734e-06, |
|
"loss": 1.688, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.25217391304347825, |
|
"grad_norm": 6.34375, |
|
"learning_rate": 9.858670610873528e-06, |
|
"loss": 1.7948, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.2608695652173913, |
|
"grad_norm": 6.03125, |
|
"learning_rate": 9.841114703012817e-06, |
|
"loss": 1.7484, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.26956521739130435, |
|
"grad_norm": 5.84375, |
|
"learning_rate": 9.822548800568238e-06, |
|
"loss": 1.7572, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.2782608695652174, |
|
"grad_norm": 4.96875, |
|
"learning_rate": 9.802976776916493e-06, |
|
"loss": 1.6758, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.28695652173913044, |
|
"grad_norm": 6.0, |
|
"learning_rate": 9.782402715339866e-06, |
|
"loss": 1.7555, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.2956521739130435, |
|
"grad_norm": 5.75, |
|
"learning_rate": 9.760830908174323e-06, |
|
"loss": 1.7926, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.30434782608695654, |
|
"grad_norm": 5.0625, |
|
"learning_rate": 9.738265855914014e-06, |
|
"loss": 1.7392, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.3130434782608696, |
|
"grad_norm": 6.375, |
|
"learning_rate": 9.714712266272339e-06, |
|
"loss": 1.7365, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.3217391304347826, |
|
"grad_norm": 6.21875, |
|
"learning_rate": 9.690175053199789e-06, |
|
"loss": 1.8047, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.33043478260869563, |
|
"grad_norm": 7.03125, |
|
"learning_rate": 9.664659335858755e-06, |
|
"loss": 1.691, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.3391304347826087, |
|
"grad_norm": 5.8125, |
|
"learning_rate": 9.63817043755552e-06, |
|
"loss": 1.635, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.34782608695652173, |
|
"grad_norm": 12.1875, |
|
"learning_rate": 9.610713884629667e-06, |
|
"loss": 1.7294, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.3565217391304348, |
|
"grad_norm": 7.28125, |
|
"learning_rate": 9.582295405301131e-06, |
|
"loss": 1.8021, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.3652173913043478, |
|
"grad_norm": 6.46875, |
|
"learning_rate": 9.552920928475127e-06, |
|
"loss": 1.7255, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.3739130434782609, |
|
"grad_norm": 4.75, |
|
"learning_rate": 9.522596582505208e-06, |
|
"loss": 1.7484, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.3826086956521739, |
|
"grad_norm": 6.25, |
|
"learning_rate": 9.491328693914723e-06, |
|
"loss": 1.8143, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.391304347826087, |
|
"grad_norm": 3.921875, |
|
"learning_rate": 9.459123786076911e-06, |
|
"loss": 1.6716, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 6.71875, |
|
"learning_rate": 9.425988577853959e-06, |
|
"loss": 1.8105, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.40869565217391307, |
|
"grad_norm": 9.375, |
|
"learning_rate": 9.391929982195233e-06, |
|
"loss": 1.7907, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.41739130434782606, |
|
"grad_norm": 3.84375, |
|
"learning_rate": 9.356955104695057e-06, |
|
"loss": 1.6777, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.4260869565217391, |
|
"grad_norm": 5.71875, |
|
"learning_rate": 9.321071242110275e-06, |
|
"loss": 1.8068, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.43478260869565216, |
|
"grad_norm": 5.53125, |
|
"learning_rate": 9.284285880837947e-06, |
|
"loss": 1.718, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.4434782608695652, |
|
"grad_norm": 3.4375, |
|
"learning_rate": 9.24660669535346e-06, |
|
"loss": 1.6196, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.45217391304347826, |
|
"grad_norm": 2.828125, |
|
"learning_rate": 9.208041546609424e-06, |
|
"loss": 1.7354, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.4608695652173913, |
|
"grad_norm": 3.34375, |
|
"learning_rate": 9.168598480395653e-06, |
|
"loss": 1.7442, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.46956521739130436, |
|
"grad_norm": 3.6875, |
|
"learning_rate": 9.128285725660577e-06, |
|
"loss": 1.742, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.4782608695652174, |
|
"grad_norm": 4.125, |
|
"learning_rate": 9.08711169279446e-06, |
|
"loss": 1.8496, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.48695652173913045, |
|
"grad_norm": 3.5625, |
|
"learning_rate": 9.045084971874738e-06, |
|
"loss": 1.6437, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.4956521739130435, |
|
"grad_norm": 4.28125, |
|
"learning_rate": 9.002214330873895e-06, |
|
"loss": 1.7723, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.5043478260869565, |
|
"grad_norm": 2.921875, |
|
"learning_rate": 8.958508713830212e-06, |
|
"loss": 1.5517, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.5130434782608696, |
|
"grad_norm": 6.625, |
|
"learning_rate": 8.91397723898178e-06, |
|
"loss": 1.6924, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.5217391304347826, |
|
"grad_norm": 3.78125, |
|
"learning_rate": 8.868629196864182e-06, |
|
"loss": 1.6444, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.5304347826086957, |
|
"grad_norm": 3.9375, |
|
"learning_rate": 8.82247404837222e-06, |
|
"loss": 1.751, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.5391304347826087, |
|
"grad_norm": 4.28125, |
|
"learning_rate": 8.775521422786104e-06, |
|
"loss": 1.8369, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.5478260869565217, |
|
"grad_norm": 3.6875, |
|
"learning_rate": 8.727781115762503e-06, |
|
"loss": 1.7386, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.5565217391304348, |
|
"grad_norm": 4.46875, |
|
"learning_rate": 8.679263087290903e-06, |
|
"loss": 1.8209, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.5652173913043478, |
|
"grad_norm": 3.953125, |
|
"learning_rate": 8.629977459615655e-06, |
|
"loss": 1.7278, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.5739130434782609, |
|
"grad_norm": 4.53125, |
|
"learning_rate": 8.579934515124202e-06, |
|
"loss": 1.7605, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.5826086956521739, |
|
"grad_norm": 3.640625, |
|
"learning_rate": 8.529144694201866e-06, |
|
"loss": 1.6836, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.591304347826087, |
|
"grad_norm": 3.375, |
|
"learning_rate": 8.477618593053693e-06, |
|
"loss": 1.6158, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 4.40625, |
|
"learning_rate": 8.425366961493784e-06, |
|
"loss": 1.821, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.6086956521739131, |
|
"grad_norm": 3.71875, |
|
"learning_rate": 8.372400700702569e-06, |
|
"loss": 1.667, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.6173913043478261, |
|
"grad_norm": 3.734375, |
|
"learning_rate": 8.318730860952523e-06, |
|
"loss": 1.7708, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.6260869565217392, |
|
"grad_norm": 3.5625, |
|
"learning_rate": 8.264368639302746e-06, |
|
"loss": 1.6117, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.6347826086956522, |
|
"grad_norm": 3.609375, |
|
"learning_rate": 8.209325377262955e-06, |
|
"loss": 1.7218, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.6434782608695652, |
|
"grad_norm": 3.578125, |
|
"learning_rate": 8.153612558427311e-06, |
|
"loss": 1.775, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.6521739130434783, |
|
"grad_norm": 3.1875, |
|
"learning_rate": 8.097241806078616e-06, |
|
"loss": 1.5923, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.6608695652173913, |
|
"grad_norm": 3.78125, |
|
"learning_rate": 8.040224880763368e-06, |
|
"loss": 1.6577, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.6695652173913044, |
|
"grad_norm": 4.90625, |
|
"learning_rate": 7.982573677838172e-06, |
|
"loss": 1.7276, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.6782608695652174, |
|
"grad_norm": 3.390625, |
|
"learning_rate": 7.92430022498803e-06, |
|
"loss": 1.6829, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.6869565217391305, |
|
"grad_norm": 3.671875, |
|
"learning_rate": 7.865416679717025e-06, |
|
"loss": 1.7478, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.6956521739130435, |
|
"grad_norm": 3.46875, |
|
"learning_rate": 7.805935326811913e-06, |
|
"loss": 1.7896, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.7043478260869566, |
|
"grad_norm": 3.234375, |
|
"learning_rate": 7.745868575779176e-06, |
|
"loss": 1.6258, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.7130434782608696, |
|
"grad_norm": 3.296875, |
|
"learning_rate": 7.685228958256036e-06, |
|
"loss": 1.6437, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.7217391304347827, |
|
"grad_norm": 3.9375, |
|
"learning_rate": 7.624029125396004e-06, |
|
"loss": 1.7755, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.7304347826086957, |
|
"grad_norm": 3.171875, |
|
"learning_rate": 7.562281845229483e-06, |
|
"loss": 1.7433, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.7391304347826086, |
|
"grad_norm": 3.1875, |
|
"learning_rate": 7.500000000000001e-06, |
|
"loss": 1.7589, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.7478260869565218, |
|
"grad_norm": 3.265625, |
|
"learning_rate": 7.437196583476597e-06, |
|
"loss": 1.6831, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.7565217391304347, |
|
"grad_norm": 3.359375, |
|
"learning_rate": 7.373884698242959e-06, |
|
"loss": 1.7514, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.7652173913043478, |
|
"grad_norm": 3.359375, |
|
"learning_rate": 7.310077552963849e-06, |
|
"loss": 1.6971, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.7739130434782608, |
|
"grad_norm": 3.515625, |
|
"learning_rate": 7.245788459629397e-06, |
|
"loss": 1.7533, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.782608695652174, |
|
"grad_norm": 3.25, |
|
"learning_rate": 7.181030830777838e-06, |
|
"loss": 1.6642, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.7913043478260869, |
|
"grad_norm": 3.453125, |
|
"learning_rate": 7.115818176697285e-06, |
|
"loss": 1.7967, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 3.984375, |
|
"learning_rate": 7.050164102607081e-06, |
|
"loss": 1.8408, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.808695652173913, |
|
"grad_norm": 3.015625, |
|
"learning_rate": 6.984082305819379e-06, |
|
"loss": 1.5956, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.8173913043478261, |
|
"grad_norm": 3.078125, |
|
"learning_rate": 6.9175865728814806e-06, |
|
"loss": 1.7632, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.8260869565217391, |
|
"grad_norm": 3.9375, |
|
"learning_rate": 6.850690776699574e-06, |
|
"loss": 1.7298, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.8347826086956521, |
|
"grad_norm": 3.125, |
|
"learning_rate": 6.7834088736444435e-06, |
|
"loss": 1.7808, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.8434782608695652, |
|
"grad_norm": 2.703125, |
|
"learning_rate": 6.715754900639789e-06, |
|
"loss": 1.7084, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.8521739130434782, |
|
"grad_norm": 4.0, |
|
"learning_rate": 6.647742972233703e-06, |
|
"loss": 1.8133, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.8608695652173913, |
|
"grad_norm": 3.40625, |
|
"learning_rate": 6.579387277653986e-06, |
|
"loss": 1.7081, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.8695652173913043, |
|
"grad_norm": 2.953125, |
|
"learning_rate": 6.510702077847864e-06, |
|
"loss": 1.8246, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.8782608695652174, |
|
"grad_norm": 3.015625, |
|
"learning_rate": 6.441701702506755e-06, |
|
"loss": 1.6652, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.8869565217391304, |
|
"grad_norm": 2.78125, |
|
"learning_rate": 6.372400547076675e-06, |
|
"loss": 1.7344, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.8956521739130435, |
|
"grad_norm": 2.9375, |
|
"learning_rate": 6.302813069754949e-06, |
|
"loss": 1.7272, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 0.9043478260869565, |
|
"grad_norm": 2.9375, |
|
"learning_rate": 6.2329537884738115e-06, |
|
"loss": 1.7836, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 0.9130434782608695, |
|
"grad_norm": 3.09375, |
|
"learning_rate": 6.162837277871553e-06, |
|
"loss": 1.5905, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.9217391304347826, |
|
"grad_norm": 3.53125, |
|
"learning_rate": 6.092478166251839e-06, |
|
"loss": 1.7248, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 0.9304347826086956, |
|
"grad_norm": 3.5625, |
|
"learning_rate": 6.021891132531825e-06, |
|
"loss": 1.7304, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 0.9391304347826087, |
|
"grad_norm": 3.28125, |
|
"learning_rate": 5.9510909031797135e-06, |
|
"loss": 1.6762, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.9478260869565217, |
|
"grad_norm": 3.84375, |
|
"learning_rate": 5.880092249142401e-06, |
|
"loss": 1.6284, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 0.9565217391304348, |
|
"grad_norm": 2.828125, |
|
"learning_rate": 5.808909982763825e-06, |
|
"loss": 1.7868, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.9652173913043478, |
|
"grad_norm": 3.203125, |
|
"learning_rate": 5.737558954694698e-06, |
|
"loss": 1.6698, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 0.9739130434782609, |
|
"grad_norm": 3.015625, |
|
"learning_rate": 5.666054050794234e-06, |
|
"loss": 1.7537, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 0.9826086956521739, |
|
"grad_norm": 2.90625, |
|
"learning_rate": 5.594410189024533e-06, |
|
"loss": 1.6714, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 0.991304347826087, |
|
"grad_norm": 6.25, |
|
"learning_rate": 5.522642316338268e-06, |
|
"loss": 1.7342, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 2.53125, |
|
"learning_rate": 5.450765405560328e-06, |
|
"loss": 1.7088, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.008695652173913, |
|
"grad_norm": 2.875, |
|
"learning_rate": 5.378794452264053e-06, |
|
"loss": 1.6256, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 1.017391304347826, |
|
"grad_norm": 3.46875, |
|
"learning_rate": 5.30674447164274e-06, |
|
"loss": 1.7419, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 1.0260869565217392, |
|
"grad_norm": 3.109375, |
|
"learning_rate": 5.234630495377035e-06, |
|
"loss": 1.6345, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 1.0347826086956522, |
|
"grad_norm": 2.609375, |
|
"learning_rate": 5.1624675684989035e-06, |
|
"loss": 1.668, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 1.0434782608695652, |
|
"grad_norm": 3.03125, |
|
"learning_rate": 5.090270746252803e-06, |
|
"loss": 1.8276, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.0521739130434782, |
|
"grad_norm": 2.859375, |
|
"learning_rate": 5.018055090954733e-06, |
|
"loss": 1.6851, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 1.0608695652173914, |
|
"grad_norm": 2.875, |
|
"learning_rate": 4.945835668849801e-06, |
|
"loss": 1.7474, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 1.0695652173913044, |
|
"grad_norm": 2.96875, |
|
"learning_rate": 4.873627546968983e-06, |
|
"loss": 1.7416, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 1.0782608695652174, |
|
"grad_norm": 3.421875, |
|
"learning_rate": 4.801445789985703e-06, |
|
"loss": 1.6386, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 1.0869565217391304, |
|
"grad_norm": 2.84375, |
|
"learning_rate": 4.729305457072913e-06, |
|
"loss": 1.6194, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.0956521739130434, |
|
"grad_norm": 2.546875, |
|
"learning_rate": 4.657221598761329e-06, |
|
"loss": 1.6924, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 1.1043478260869566, |
|
"grad_norm": 3.109375, |
|
"learning_rate": 4.585209253799451e-06, |
|
"loss": 1.6767, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 1.1130434782608696, |
|
"grad_norm": 3.046875, |
|
"learning_rate": 4.513283446016052e-06, |
|
"loss": 1.7603, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 1.1217391304347826, |
|
"grad_norm": 2.46875, |
|
"learning_rate": 4.441459181185786e-06, |
|
"loss": 1.758, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 1.1304347826086956, |
|
"grad_norm": 2.859375, |
|
"learning_rate": 4.369751443898554e-06, |
|
"loss": 1.6899, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.1391304347826088, |
|
"grad_norm": 2.875, |
|
"learning_rate": 4.298175194433279e-06, |
|
"loss": 1.7596, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 1.1478260869565218, |
|
"grad_norm": 2.984375, |
|
"learning_rate": 4.226745365636783e-06, |
|
"loss": 1.7741, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 1.1565217391304348, |
|
"grad_norm": 2.71875, |
|
"learning_rate": 4.155476859808356e-06, |
|
"loss": 1.7412, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 1.1652173913043478, |
|
"grad_norm": 2.78125, |
|
"learning_rate": 4.0843845455907195e-06, |
|
"loss": 1.7079, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 1.1739130434782608, |
|
"grad_norm": 3.03125, |
|
"learning_rate": 4.013483254868001e-06, |
|
"loss": 1.6852, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.182608695652174, |
|
"grad_norm": 3.0, |
|
"learning_rate": 3.942787779671376e-06, |
|
"loss": 1.781, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 1.191304347826087, |
|
"grad_norm": 3.25, |
|
"learning_rate": 3.87231286909303e-06, |
|
"loss": 1.64, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 2.671875, |
|
"learning_rate": 3.8020732262090876e-06, |
|
"loss": 1.5995, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 1.208695652173913, |
|
"grad_norm": 2.875, |
|
"learning_rate": 3.732083505012118e-06, |
|
"loss": 1.5924, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 1.2173913043478262, |
|
"grad_norm": 2.671875, |
|
"learning_rate": 3.662358307353897e-06, |
|
"loss": 1.5588, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.2260869565217392, |
|
"grad_norm": 2.59375, |
|
"learning_rate": 3.5929121798990494e-06, |
|
"loss": 1.6359, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 1.2347826086956522, |
|
"grad_norm": 2.6875, |
|
"learning_rate": 3.5237596110902007e-06, |
|
"loss": 1.6286, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 1.2434782608695651, |
|
"grad_norm": 2.21875, |
|
"learning_rate": 3.4549150281252635e-06, |
|
"loss": 1.7253, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 1.2521739130434781, |
|
"grad_norm": 2.71875, |
|
"learning_rate": 3.3863927939475274e-06, |
|
"loss": 1.8196, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 1.2608695652173914, |
|
"grad_norm": 2.59375, |
|
"learning_rate": 3.3182072042491244e-06, |
|
"loss": 1.8091, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 1.2695652173913043, |
|
"grad_norm": 2.53125, |
|
"learning_rate": 3.250372484488558e-06, |
|
"loss": 1.772, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 1.2782608695652173, |
|
"grad_norm": 2.8125, |
|
"learning_rate": 3.1829027869228506e-06, |
|
"loss": 1.6914, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 1.2869565217391306, |
|
"grad_norm": 2.671875, |
|
"learning_rate": 3.1158121876549853e-06, |
|
"loss": 1.8261, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 1.2956521739130435, |
|
"grad_norm": 2.671875, |
|
"learning_rate": 3.0491146836972273e-06, |
|
"loss": 1.712, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 1.3043478260869565, |
|
"grad_norm": 2.609375, |
|
"learning_rate": 2.982824190050958e-06, |
|
"loss": 1.7076, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.3130434782608695, |
|
"grad_norm": 1.9609375, |
|
"learning_rate": 2.9169545368035933e-06, |
|
"loss": 1.6249, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 1.3217391304347825, |
|
"grad_norm": 2.265625, |
|
"learning_rate": 2.8515194662432423e-06, |
|
"loss": 1.6602, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 1.3304347826086955, |
|
"grad_norm": 2.3125, |
|
"learning_rate": 2.7865326299916736e-06, |
|
"loss": 1.72, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 1.3391304347826087, |
|
"grad_norm": 1.8828125, |
|
"learning_rate": 2.7220075861561956e-06, |
|
"loss": 1.6759, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 1.3478260869565217, |
|
"grad_norm": 2.046875, |
|
"learning_rate": 2.65795779650105e-06, |
|
"loss": 1.8399, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 1.3565217391304347, |
|
"grad_norm": 1.90625, |
|
"learning_rate": 2.5943966236389028e-06, |
|
"loss": 1.736, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 1.365217391304348, |
|
"grad_norm": 2.0625, |
|
"learning_rate": 2.5313373282430244e-06, |
|
"loss": 1.7337, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 1.373913043478261, |
|
"grad_norm": 2.15625, |
|
"learning_rate": 2.46879306628073e-06, |
|
"loss": 1.7536, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 1.382608695652174, |
|
"grad_norm": 2.015625, |
|
"learning_rate": 2.406776886268677e-06, |
|
"loss": 1.6754, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 1.391304347826087, |
|
"grad_norm": 2.015625, |
|
"learning_rate": 2.345301726550567e-06, |
|
"loss": 1.8216, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 1.8515625, |
|
"learning_rate": 2.2843804125978356e-06, |
|
"loss": 1.6752, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 1.4086956521739131, |
|
"grad_norm": 2.125, |
|
"learning_rate": 2.224025654333905e-06, |
|
"loss": 1.7638, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 1.4173913043478261, |
|
"grad_norm": 1.890625, |
|
"learning_rate": 2.1642500434825207e-06, |
|
"loss": 1.6406, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 1.4260869565217391, |
|
"grad_norm": 1.7578125, |
|
"learning_rate": 2.105066050940758e-06, |
|
"loss": 1.695, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 1.434782608695652, |
|
"grad_norm": 1.96875, |
|
"learning_rate": 2.0464860241772454e-06, |
|
"loss": 1.7199, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 1.4434782608695653, |
|
"grad_norm": 1.8984375, |
|
"learning_rate": 1.9885221846561267e-06, |
|
"loss": 1.7604, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 1.4521739130434783, |
|
"grad_norm": 1.8203125, |
|
"learning_rate": 1.931186625287313e-06, |
|
"loss": 1.592, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 1.4608695652173913, |
|
"grad_norm": 1.796875, |
|
"learning_rate": 1.8744913079035598e-06, |
|
"loss": 1.654, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 1.4695652173913043, |
|
"grad_norm": 1.9609375, |
|
"learning_rate": 1.8184480607648863e-06, |
|
"loss": 1.6098, |
|
"step": 676 |
|
}, |
|
{ |
|
"epoch": 1.4782608695652173, |
|
"grad_norm": 1.9140625, |
|
"learning_rate": 1.7630685760908623e-06, |
|
"loss": 1.6708, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 1.4869565217391305, |
|
"grad_norm": 2.59375, |
|
"learning_rate": 1.7083644076212785e-06, |
|
"loss": 1.6816, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 1.4956521739130435, |
|
"grad_norm": 2.46875, |
|
"learning_rate": 1.6543469682057105e-06, |
|
"loss": 1.6811, |
|
"step": 688 |
|
}, |
|
{ |
|
"epoch": 1.5043478260869565, |
|
"grad_norm": 2.015625, |
|
"learning_rate": 1.6010275274224607e-06, |
|
"loss": 1.7303, |
|
"step": 692 |
|
}, |
|
{ |
|
"epoch": 1.5130434782608697, |
|
"grad_norm": 1.8046875, |
|
"learning_rate": 1.548417209227422e-06, |
|
"loss": 1.7151, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 1.5217391304347827, |
|
"grad_norm": 1.9453125, |
|
"learning_rate": 1.4965269896332884e-06, |
|
"loss": 1.6849, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.5304347826086957, |
|
"grad_norm": 2.421875, |
|
"learning_rate": 1.4453676944196477e-06, |
|
"loss": 1.7114, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 1.5391304347826087, |
|
"grad_norm": 3.015625, |
|
"learning_rate": 1.3949499968744208e-06, |
|
"loss": 1.7589, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 1.5478260869565217, |
|
"grad_norm": 1.890625, |
|
"learning_rate": 1.3452844155671052e-06, |
|
"loss": 1.6126, |
|
"step": 712 |
|
}, |
|
{ |
|
"epoch": 1.5565217391304347, |
|
"grad_norm": 1.8203125, |
|
"learning_rate": 1.296381312154305e-06, |
|
"loss": 1.7336, |
|
"step": 716 |
|
}, |
|
{ |
|
"epoch": 1.5652173913043477, |
|
"grad_norm": 1.890625, |
|
"learning_rate": 1.2482508892179884e-06, |
|
"loss": 1.7618, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 1.5739130434782609, |
|
"grad_norm": 1.75, |
|
"learning_rate": 1.2009031881369431e-06, |
|
"loss": 1.7413, |
|
"step": 724 |
|
}, |
|
{ |
|
"epoch": 1.5826086956521739, |
|
"grad_norm": 1.8046875, |
|
"learning_rate": 1.1543480869918555e-06, |
|
"loss": 1.6844, |
|
"step": 728 |
|
}, |
|
{ |
|
"epoch": 1.591304347826087, |
|
"grad_norm": 1.921875, |
|
"learning_rate": 1.1085952985044636e-06, |
|
"loss": 1.7543, |
|
"step": 732 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 1.9453125, |
|
"learning_rate": 1.0636543680112043e-06, |
|
"loss": 1.7416, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 1.608695652173913, |
|
"grad_norm": 1.8203125, |
|
"learning_rate": 1.0195346714717813e-06, |
|
"loss": 1.6915, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 1.617391304347826, |
|
"grad_norm": 1.8828125, |
|
"learning_rate": 9.762454135130828e-07, |
|
"loss": 1.6498, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 1.626086956521739, |
|
"grad_norm": 2.421875, |
|
"learning_rate": 9.337956255088238e-07, |
|
"loss": 1.7677, |
|
"step": 748 |
|
}, |
|
{ |
|
"epoch": 1.634782608695652, |
|
"grad_norm": 1.984375, |
|
"learning_rate": 8.921941636953435e-07, |
|
"loss": 1.7559, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 1.643478260869565, |
|
"grad_norm": 1.7890625, |
|
"learning_rate": 8.514497073239491e-07, |
|
"loss": 1.8024, |
|
"step": 756 |
|
}, |
|
{ |
|
"epoch": 1.6521739130434783, |
|
"grad_norm": 1.8125, |
|
"learning_rate": 8.115707568501768e-07, |
|
"loss": 1.7869, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 1.6608695652173913, |
|
"grad_norm": 1.890625, |
|
"learning_rate": 7.725656321603414e-07, |
|
"loss": 1.6925, |
|
"step": 764 |
|
}, |
|
{ |
|
"epoch": 1.6695652173913045, |
|
"grad_norm": 2.859375, |
|
"learning_rate": 7.344424708357867e-07, |
|
"loss": 1.6703, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 1.6782608695652175, |
|
"grad_norm": 2.03125, |
|
"learning_rate": 6.972092264551439e-07, |
|
"loss": 1.745, |
|
"step": 772 |
|
}, |
|
{ |
|
"epoch": 1.6869565217391305, |
|
"grad_norm": 2.3125, |
|
"learning_rate": 6.60873666934993e-07, |
|
"loss": 1.714, |
|
"step": 776 |
|
}, |
|
{ |
|
"epoch": 1.6956521739130435, |
|
"grad_norm": 1.8828125, |
|
"learning_rate": 6.254433729092518e-07, |
|
"loss": 1.724, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 1.7043478260869565, |
|
"grad_norm": 1.9375, |
|
"learning_rate": 5.909257361476406e-07, |
|
"loss": 1.7012, |
|
"step": 784 |
|
}, |
|
{ |
|
"epoch": 1.7130434782608694, |
|
"grad_norm": 3.65625, |
|
"learning_rate": 5.573279580135438e-07, |
|
"loss": 1.5692, |
|
"step": 788 |
|
}, |
|
{ |
|
"epoch": 1.7217391304347827, |
|
"grad_norm": 1.890625, |
|
"learning_rate": 5.246570479616103e-07, |
|
"loss": 1.753, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 1.7304347826086957, |
|
"grad_norm": 1.7421875, |
|
"learning_rate": 4.929198220753723e-07, |
|
"loss": 1.6207, |
|
"step": 796 |
|
}, |
|
{ |
|
"epoch": 1.7391304347826086, |
|
"grad_norm": 2.171875, |
|
"learning_rate": 4.6212290164521554e-07, |
|
"loss": 1.7742, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.7478260869565219, |
|
"grad_norm": 1.7734375, |
|
"learning_rate": 4.322727117869951e-07, |
|
"loss": 1.832, |
|
"step": 804 |
|
}, |
|
{ |
|
"epoch": 1.7565217391304349, |
|
"grad_norm": 1.9609375, |
|
"learning_rate": 4.0337548010157324e-07, |
|
"loss": 1.7819, |
|
"step": 808 |
|
}, |
|
{ |
|
"epoch": 1.7652173913043478, |
|
"grad_norm": 2.5, |
|
"learning_rate": 3.754372353755559e-07, |
|
"loss": 1.6899, |
|
"step": 812 |
|
}, |
|
{ |
|
"epoch": 1.7739130434782608, |
|
"grad_norm": 2.0, |
|
"learning_rate": 3.4846380632352463e-07, |
|
"loss": 1.7287, |
|
"step": 816 |
|
}, |
|
{ |
|
"epoch": 1.7826086956521738, |
|
"grad_norm": 1.8359375, |
|
"learning_rate": 3.224608203719953e-07, |
|
"loss": 1.7088, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 1.7913043478260868, |
|
"grad_norm": 1.8984375, |
|
"learning_rate": 2.974337024853802e-07, |
|
"loss": 1.7639, |
|
"step": 824 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 1.90625, |
|
"learning_rate": 2.733876740341829e-07, |
|
"loss": 1.787, |
|
"step": 828 |
|
}, |
|
{ |
|
"epoch": 1.808695652173913, |
|
"grad_norm": 1.8125, |
|
"learning_rate": 2.5032775170567293e-07, |
|
"loss": 1.7659, |
|
"step": 832 |
|
}, |
|
{ |
|
"epoch": 1.8173913043478263, |
|
"grad_norm": 1.8125, |
|
"learning_rate": 2.2825874645725942e-07, |
|
"loss": 1.7699, |
|
"step": 836 |
|
}, |
|
{ |
|
"epoch": 1.8260869565217392, |
|
"grad_norm": 2.625, |
|
"learning_rate": 2.0718526251279346e-07, |
|
"loss": 1.7371, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 1.8347826086956522, |
|
"grad_norm": 2.046875, |
|
"learning_rate": 1.871116964019898e-07, |
|
"loss": 1.7021, |
|
"step": 844 |
|
}, |
|
{ |
|
"epoch": 1.8434782608695652, |
|
"grad_norm": 1.9765625, |
|
"learning_rate": 1.6804223604318825e-07, |
|
"loss": 1.69, |
|
"step": 848 |
|
}, |
|
{ |
|
"epoch": 1.8521739130434782, |
|
"grad_norm": 1.8671875, |
|
"learning_rate": 1.4998085986963283e-07, |
|
"loss": 1.639, |
|
"step": 852 |
|
}, |
|
{ |
|
"epoch": 1.8608695652173912, |
|
"grad_norm": 2.0, |
|
"learning_rate": 1.329313359994633e-07, |
|
"loss": 1.7022, |
|
"step": 856 |
|
}, |
|
{ |
|
"epoch": 1.8695652173913042, |
|
"grad_norm": 1.828125, |
|
"learning_rate": 1.1689722144956672e-07, |
|
"loss": 1.695, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 1.8782608695652174, |
|
"grad_norm": 1.859375, |
|
"learning_rate": 1.0188186139349355e-07, |
|
"loss": 1.7447, |
|
"step": 864 |
|
}, |
|
{ |
|
"epoch": 1.8869565217391304, |
|
"grad_norm": 1.78125, |
|
"learning_rate": 8.788838846355341e-08, |
|
"loss": 1.7429, |
|
"step": 868 |
|
}, |
|
{ |
|
"epoch": 1.8956521739130436, |
|
"grad_norm": 2.6875, |
|
"learning_rate": 7.491972209725807e-08, |
|
"loss": 1.7296, |
|
"step": 872 |
|
}, |
|
{ |
|
"epoch": 1.9043478260869566, |
|
"grad_norm": 1.75, |
|
"learning_rate": 6.297856792824741e-08, |
|
"loss": 1.6088, |
|
"step": 876 |
|
}, |
|
{ |
|
"epoch": 1.9130434782608696, |
|
"grad_norm": 1.78125, |
|
"learning_rate": 5.206741722181385e-08, |
|
"loss": 1.6972, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 1.9217391304347826, |
|
"grad_norm": 1.75, |
|
"learning_rate": 4.2188546355153016e-08, |
|
"loss": 1.7105, |
|
"step": 884 |
|
}, |
|
{ |
|
"epoch": 1.9304347826086956, |
|
"grad_norm": 1.8359375, |
|
"learning_rate": 3.334401634245033e-08, |
|
"loss": 1.6165, |
|
"step": 888 |
|
}, |
|
{ |
|
"epoch": 1.9391304347826086, |
|
"grad_norm": 1.828125, |
|
"learning_rate": 2.5535672404890523e-08, |
|
"loss": 1.7325, |
|
"step": 892 |
|
}, |
|
{ |
|
"epoch": 1.9478260869565216, |
|
"grad_norm": 2.140625, |
|
"learning_rate": 1.8765143585693924e-08, |
|
"loss": 1.6871, |
|
"step": 896 |
|
}, |
|
{ |
|
"epoch": 1.9565217391304348, |
|
"grad_norm": 1.828125, |
|
"learning_rate": 1.3033842410251074e-08, |
|
"loss": 1.7506, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 1.9652173913043478, |
|
"grad_norm": 1.75, |
|
"learning_rate": 8.342964591430137e-09, |
|
"loss": 1.6842, |
|
"step": 904 |
|
}, |
|
{ |
|
"epoch": 1.973913043478261, |
|
"grad_norm": 1.7578125, |
|
"learning_rate": 4.69348878011644e-09, |
|
"loss": 1.6753, |
|
"step": 908 |
|
}, |
|
{ |
|
"epoch": 1.982608695652174, |
|
"grad_norm": 2.140625, |
|
"learning_rate": 2.0861763610385834e-09, |
|
"loss": 1.6069, |
|
"step": 912 |
|
}, |
|
{ |
|
"epoch": 1.991304347826087, |
|
"grad_norm": 1.890625, |
|
"learning_rate": 5.215712939210527e-10, |
|
"loss": 1.7117, |
|
"step": 916 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 2.578125, |
|
"learning_rate": 0.0, |
|
"loss": 1.7504, |
|
"step": 920 |
|
} |
|
], |
|
"logging_steps": 4, |
|
"max_steps": 920, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 230, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 8.371395096477696e+18, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|