|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.9064422143088378, |
|
"eval_steps": 100, |
|
"global_step": 700, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": "2.5641e-07", |
|
"loss": 0.5987, |
|
"slid_loss": 0.5987, |
|
"step": 1, |
|
"time": 416.96 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": "5.1282e-07", |
|
"loss": 0.6236, |
|
"slid_loss": 0.6112, |
|
"step": 2, |
|
"time": 308.68 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": "7.6923e-07", |
|
"loss": 0.61, |
|
"slid_loss": 0.6108, |
|
"step": 3, |
|
"time": 303.89 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": "1.0256e-06", |
|
"loss": 0.6193, |
|
"slid_loss": 0.6129, |
|
"step": 4, |
|
"time": 300.55 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": "1.2821e-06", |
|
"loss": 0.62, |
|
"slid_loss": 0.6143, |
|
"step": 5, |
|
"time": 301.85 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": "1.5385e-06", |
|
"loss": 0.6036, |
|
"slid_loss": 0.6125, |
|
"step": 6, |
|
"time": 303.95 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": "1.7949e-06", |
|
"loss": 0.6116, |
|
"slid_loss": 0.6124, |
|
"step": 7, |
|
"time": 301.83 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": "2.0513e-06", |
|
"loss": 0.5796, |
|
"slid_loss": 0.6083, |
|
"step": 8, |
|
"time": 304.86 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": "2.3077e-06", |
|
"loss": 0.6255, |
|
"slid_loss": 0.6102, |
|
"step": 9, |
|
"time": 299.19 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": "2.5641e-06", |
|
"loss": 0.6125, |
|
"slid_loss": 0.6104, |
|
"step": 10, |
|
"time": 303.97 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": "2.8205e-06", |
|
"loss": 0.6187, |
|
"slid_loss": 0.6112, |
|
"step": 11, |
|
"time": 299.93 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": "3.0769e-06", |
|
"loss": 0.6229, |
|
"slid_loss": 0.6122, |
|
"step": 12, |
|
"time": 299.47 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": "3.3333e-06", |
|
"loss": 0.6, |
|
"slid_loss": 0.6112, |
|
"step": 13, |
|
"time": 297.05 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": "3.5897e-06", |
|
"loss": 0.6079, |
|
"slid_loss": 0.611, |
|
"step": 14, |
|
"time": 300.11 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": "3.8462e-06", |
|
"loss": 0.5676, |
|
"slid_loss": 0.6081, |
|
"step": 15, |
|
"time": 301.79 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": "4.1026e-06", |
|
"loss": 0.5754, |
|
"slid_loss": 0.606, |
|
"step": 16, |
|
"time": 294.41 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": "4.3590e-06", |
|
"loss": 0.5977, |
|
"slid_loss": 0.6056, |
|
"step": 17, |
|
"time": 299.96 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": "4.6154e-06", |
|
"loss": 0.5934, |
|
"slid_loss": 0.6049, |
|
"step": 18, |
|
"time": 300.94 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": "4.8718e-06", |
|
"loss": 0.6201, |
|
"slid_loss": 0.6057, |
|
"step": 19, |
|
"time": 299.48 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": "5.1282e-06", |
|
"loss": 0.6358, |
|
"slid_loss": 0.6072, |
|
"step": 20, |
|
"time": 303.69 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": "5.3846e-06", |
|
"loss": 0.6163, |
|
"slid_loss": 0.6076, |
|
"step": 21, |
|
"time": 302.21 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": "5.6410e-06", |
|
"loss": 0.5778, |
|
"slid_loss": 0.6063, |
|
"step": 22, |
|
"time": 299.39 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": "5.8974e-06", |
|
"loss": 0.6006, |
|
"slid_loss": 0.606, |
|
"step": 23, |
|
"time": 303.27 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": "6.1538e-06", |
|
"loss": 0.5979, |
|
"slid_loss": 0.6057, |
|
"step": 24, |
|
"time": 301.63 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": "6.4103e-06", |
|
"loss": 0.5862, |
|
"slid_loss": 0.6049, |
|
"step": 25, |
|
"time": 298.76 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": "6.6667e-06", |
|
"loss": 0.5932, |
|
"slid_loss": 0.6045, |
|
"step": 26, |
|
"time": 301.27 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": "6.9231e-06", |
|
"loss": 0.6079, |
|
"slid_loss": 0.6046, |
|
"step": 27, |
|
"time": 301.95 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": "7.1795e-06", |
|
"loss": 0.6066, |
|
"slid_loss": 0.6047, |
|
"step": 28, |
|
"time": 306.01 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": "7.4359e-06", |
|
"loss": 0.6065, |
|
"slid_loss": 0.6047, |
|
"step": 29, |
|
"time": 305.69 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": "7.6923e-06", |
|
"loss": 0.6172, |
|
"slid_loss": 0.6051, |
|
"step": 30, |
|
"time": 307.86 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": "7.9487e-06", |
|
"loss": 0.5885, |
|
"slid_loss": 0.6046, |
|
"step": 31, |
|
"time": 302.34 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": "8.2051e-06", |
|
"loss": 0.616, |
|
"slid_loss": 0.605, |
|
"step": 32, |
|
"time": 297.77 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": "8.4615e-06", |
|
"loss": 0.5881, |
|
"slid_loss": 0.6044, |
|
"step": 33, |
|
"time": 303.91 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": "8.7179e-06", |
|
"loss": 0.6276, |
|
"slid_loss": 0.6051, |
|
"step": 34, |
|
"time": 300.64 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": "8.9744e-06", |
|
"loss": 0.6095, |
|
"slid_loss": 0.6052, |
|
"step": 35, |
|
"time": 293.11 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": "9.2308e-06", |
|
"loss": 0.6124, |
|
"slid_loss": 0.6054, |
|
"step": 36, |
|
"time": 302.12 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": "9.4872e-06", |
|
"loss": 0.5959, |
|
"slid_loss": 0.6052, |
|
"step": 37, |
|
"time": 300.08 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": "9.7436e-06", |
|
"loss": 0.6022, |
|
"slid_loss": 0.6051, |
|
"step": 38, |
|
"time": 301.67 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": "1.0000e-05", |
|
"loss": 0.5919, |
|
"slid_loss": 0.6048, |
|
"step": 39, |
|
"time": 300.9 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": "1.0000e-05", |
|
"loss": 0.5992, |
|
"slid_loss": 0.6046, |
|
"step": 40, |
|
"time": 303.93 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": "1.0000e-05", |
|
"loss": 0.581, |
|
"slid_loss": 0.6041, |
|
"step": 41, |
|
"time": 303.43 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": "1.0000e-05", |
|
"loss": 0.5893, |
|
"slid_loss": 0.6037, |
|
"step": 42, |
|
"time": 296.02 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": "1.0000e-05", |
|
"loss": 0.6011, |
|
"slid_loss": 0.6036, |
|
"step": 43, |
|
"time": 296.37 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": "1.0000e-05", |
|
"loss": 0.619, |
|
"slid_loss": 0.604, |
|
"step": 44, |
|
"time": 332.28 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": "1.0000e-05", |
|
"loss": 0.5861, |
|
"slid_loss": 0.6036, |
|
"step": 45, |
|
"time": 299.29 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": "1.0000e-05", |
|
"loss": 0.5954, |
|
"slid_loss": 0.6034, |
|
"step": 46, |
|
"time": 303.55 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": "9.9999e-06", |
|
"loss": 0.6072, |
|
"slid_loss": 0.6035, |
|
"step": 47, |
|
"time": 304.38 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": "9.9999e-06", |
|
"loss": 0.5787, |
|
"slid_loss": 0.603, |
|
"step": 48, |
|
"time": 304.24 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": "9.9999e-06", |
|
"loss": 0.5957, |
|
"slid_loss": 0.6028, |
|
"step": 49, |
|
"time": 298.5 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": "9.9999e-06", |
|
"loss": 0.6328, |
|
"slid_loss": 0.6034, |
|
"step": 50, |
|
"time": 303.53 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": "9.9999e-06", |
|
"loss": 0.581, |
|
"slid_loss": 0.603, |
|
"step": 51, |
|
"time": 302.2 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": "9.9999e-06", |
|
"loss": 0.6062, |
|
"slid_loss": 0.6031, |
|
"step": 52, |
|
"time": 301.68 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": "9.9998e-06", |
|
"loss": 0.6003, |
|
"slid_loss": 0.603, |
|
"step": 53, |
|
"time": 296.0 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": "9.9998e-06", |
|
"loss": 0.6157, |
|
"slid_loss": 0.6032, |
|
"step": 54, |
|
"time": 299.54 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": "9.9998e-06", |
|
"loss": 0.604, |
|
"slid_loss": 0.6032, |
|
"step": 55, |
|
"time": 299.1 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": "9.9998e-06", |
|
"loss": 0.5752, |
|
"slid_loss": 0.6027, |
|
"step": 56, |
|
"time": 294.14 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": "9.9997e-06", |
|
"loss": 0.5931, |
|
"slid_loss": 0.6026, |
|
"step": 57, |
|
"time": 381.96 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": "9.9997e-06", |
|
"loss": 0.5937, |
|
"slid_loss": 0.6024, |
|
"step": 58, |
|
"time": 293.92 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": "9.9997e-06", |
|
"loss": 0.6232, |
|
"slid_loss": 0.6028, |
|
"step": 59, |
|
"time": 304.51 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": "9.9996e-06", |
|
"loss": 0.6031, |
|
"slid_loss": 0.6028, |
|
"step": 60, |
|
"time": 306.02 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": "9.9996e-06", |
|
"loss": 0.601, |
|
"slid_loss": 0.6028, |
|
"step": 61, |
|
"time": 301.96 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": "9.9996e-06", |
|
"loss": 0.6268, |
|
"slid_loss": 0.6031, |
|
"step": 62, |
|
"time": 302.5 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": "9.9995e-06", |
|
"loss": 0.6078, |
|
"slid_loss": 0.6032, |
|
"step": 63, |
|
"time": 302.51 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": "9.9995e-06", |
|
"loss": 0.6252, |
|
"slid_loss": 0.6036, |
|
"step": 64, |
|
"time": 297.93 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": "9.9994e-06", |
|
"loss": 0.6007, |
|
"slid_loss": 0.6035, |
|
"step": 65, |
|
"time": 302.92 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": "9.9994e-06", |
|
"loss": 0.5985, |
|
"slid_loss": 0.6034, |
|
"step": 66, |
|
"time": 298.35 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": "9.9993e-06", |
|
"loss": 0.5966, |
|
"slid_loss": 0.6033, |
|
"step": 67, |
|
"time": 300.99 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": "9.9993e-06", |
|
"loss": 0.5678, |
|
"slid_loss": 0.6028, |
|
"step": 68, |
|
"time": 296.29 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": "9.9992e-06", |
|
"loss": 0.5818, |
|
"slid_loss": 0.6025, |
|
"step": 69, |
|
"time": 294.95 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": "9.9992e-06", |
|
"loss": 0.6138, |
|
"slid_loss": 0.6027, |
|
"step": 70, |
|
"time": 298.16 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": "9.9991e-06", |
|
"loss": 0.6071, |
|
"slid_loss": 0.6027, |
|
"step": 71, |
|
"time": 302.87 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": "9.9991e-06", |
|
"loss": 0.6126, |
|
"slid_loss": 0.6029, |
|
"step": 72, |
|
"time": 302.98 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": "9.9990e-06", |
|
"loss": 0.5945, |
|
"slid_loss": 0.6028, |
|
"step": 73, |
|
"time": 429.67 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": "9.9990e-06", |
|
"loss": 0.5973, |
|
"slid_loss": 0.6027, |
|
"step": 74, |
|
"time": 303.2 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": "9.9989e-06", |
|
"loss": 0.6045, |
|
"slid_loss": 0.6027, |
|
"step": 75, |
|
"time": 298.66 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": "9.9988e-06", |
|
"loss": 0.5884, |
|
"slid_loss": 0.6025, |
|
"step": 76, |
|
"time": 305.59 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": "9.9988e-06", |
|
"loss": 0.6009, |
|
"slid_loss": 0.6025, |
|
"step": 77, |
|
"time": 300.33 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": "9.9987e-06", |
|
"loss": 0.6005, |
|
"slid_loss": 0.6025, |
|
"step": 78, |
|
"time": 300.6 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": "9.9986e-06", |
|
"loss": 0.5811, |
|
"slid_loss": 0.6022, |
|
"step": 79, |
|
"time": 304.71 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": "9.9986e-06", |
|
"loss": 0.5954, |
|
"slid_loss": 0.6021, |
|
"step": 80, |
|
"time": 295.84 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": "9.9985e-06", |
|
"loss": 0.5864, |
|
"slid_loss": 0.6019, |
|
"step": 81, |
|
"time": 307.04 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": "9.9984e-06", |
|
"loss": 0.6053, |
|
"slid_loss": 0.602, |
|
"step": 82, |
|
"time": 301.52 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": "9.9984e-06", |
|
"loss": 0.5725, |
|
"slid_loss": 0.6016, |
|
"step": 83, |
|
"time": 297.7 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": "9.9983e-06", |
|
"loss": 0.6035, |
|
"slid_loss": 0.6016, |
|
"step": 84, |
|
"time": 304.19 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": "9.9982e-06", |
|
"loss": 0.6073, |
|
"slid_loss": 0.6017, |
|
"step": 85, |
|
"time": 338.5 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": "9.9981e-06", |
|
"loss": 0.6044, |
|
"slid_loss": 0.6017, |
|
"step": 86, |
|
"time": 296.47 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": "9.9981e-06", |
|
"loss": 0.6035, |
|
"slid_loss": 0.6017, |
|
"step": 87, |
|
"time": 300.44 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": "9.9980e-06", |
|
"loss": 0.6143, |
|
"slid_loss": 0.6019, |
|
"step": 88, |
|
"time": 299.11 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": "9.9979e-06", |
|
"loss": 0.6097, |
|
"slid_loss": 0.602, |
|
"step": 89, |
|
"time": 299.75 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": "9.9978e-06", |
|
"loss": 0.5806, |
|
"slid_loss": 0.6017, |
|
"step": 90, |
|
"time": 301.43 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": "9.9977e-06", |
|
"loss": 0.5961, |
|
"slid_loss": 0.6017, |
|
"step": 91, |
|
"time": 304.29 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": "9.9976e-06", |
|
"loss": 0.6096, |
|
"slid_loss": 0.6018, |
|
"step": 92, |
|
"time": 303.86 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": "9.9975e-06", |
|
"loss": 0.625, |
|
"slid_loss": 0.602, |
|
"step": 93, |
|
"time": 302.51 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": "9.9974e-06", |
|
"loss": 0.5965, |
|
"slid_loss": 0.602, |
|
"step": 94, |
|
"time": 301.0 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": "9.9974e-06", |
|
"loss": 0.6001, |
|
"slid_loss": 0.6019, |
|
"step": 95, |
|
"time": 299.4 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": "9.9973e-06", |
|
"loss": 0.6154, |
|
"slid_loss": 0.6021, |
|
"step": 96, |
|
"time": 301.42 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": "9.9972e-06", |
|
"loss": 0.6055, |
|
"slid_loss": 0.6021, |
|
"step": 97, |
|
"time": 304.69 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": "9.9971e-06", |
|
"loss": 0.5986, |
|
"slid_loss": 0.6021, |
|
"step": 98, |
|
"time": 301.12 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": "9.9970e-06", |
|
"loss": 0.6033, |
|
"slid_loss": 0.6021, |
|
"step": 99, |
|
"time": 302.85 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": "9.9969e-06", |
|
"loss": 0.5926, |
|
"slid_loss": 0.602, |
|
"step": 100, |
|
"time": 298.85 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": "9.9968e-06", |
|
"loss": 0.5972, |
|
"slid_loss": 0.602, |
|
"step": 101, |
|
"time": 2115.7 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": "9.9966e-06", |
|
"loss": 0.6335, |
|
"slid_loss": 0.6021, |
|
"step": 102, |
|
"time": 297.71 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": "9.9965e-06", |
|
"loss": 0.6178, |
|
"slid_loss": 0.6022, |
|
"step": 103, |
|
"time": 296.19 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": "9.9964e-06", |
|
"loss": 0.5811, |
|
"slid_loss": 0.6018, |
|
"step": 104, |
|
"time": 297.39 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": "9.9963e-06", |
|
"loss": 0.6264, |
|
"slid_loss": 0.6018, |
|
"step": 105, |
|
"time": 302.43 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": "9.9962e-06", |
|
"loss": 0.5883, |
|
"slid_loss": 0.6017, |
|
"step": 106, |
|
"time": 297.74 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": "9.9961e-06", |
|
"loss": 0.5763, |
|
"slid_loss": 0.6013, |
|
"step": 107, |
|
"time": 295.11 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": "9.9960e-06", |
|
"loss": 0.5981, |
|
"slid_loss": 0.6015, |
|
"step": 108, |
|
"time": 297.78 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": "9.9959e-06", |
|
"loss": 0.5965, |
|
"slid_loss": 0.6012, |
|
"step": 109, |
|
"time": 303.32 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": "9.9957e-06", |
|
"loss": 0.5904, |
|
"slid_loss": 0.601, |
|
"step": 110, |
|
"time": 305.55 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": "9.9956e-06", |
|
"loss": 0.6257, |
|
"slid_loss": 0.6011, |
|
"step": 111, |
|
"time": 309.15 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": "9.9955e-06", |
|
"loss": 0.596, |
|
"slid_loss": 0.6008, |
|
"step": 112, |
|
"time": 295.49 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": "9.9954e-06", |
|
"loss": 0.6059, |
|
"slid_loss": 0.6009, |
|
"step": 113, |
|
"time": 297.05 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": "9.9952e-06", |
|
"loss": 0.5955, |
|
"slid_loss": 0.6007, |
|
"step": 114, |
|
"time": 298.43 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": "9.9951e-06", |
|
"loss": 0.5853, |
|
"slid_loss": 0.6009, |
|
"step": 115, |
|
"time": 298.88 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": "9.9950e-06", |
|
"loss": 0.5772, |
|
"slid_loss": 0.6009, |
|
"step": 116, |
|
"time": 303.47 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": "9.9949e-06", |
|
"loss": 0.6111, |
|
"slid_loss": 0.6011, |
|
"step": 117, |
|
"time": 297.58 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": "9.9947e-06", |
|
"loss": 0.5853, |
|
"slid_loss": 0.601, |
|
"step": 118, |
|
"time": 303.99 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": "9.9946e-06", |
|
"loss": 0.5996, |
|
"slid_loss": 0.6008, |
|
"step": 119, |
|
"time": 298.4 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": "9.9945e-06", |
|
"loss": 0.6205, |
|
"slid_loss": 0.6006, |
|
"step": 120, |
|
"time": 344.13 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": "9.9943e-06", |
|
"loss": 0.5966, |
|
"slid_loss": 0.6004, |
|
"step": 121, |
|
"time": 634.45 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": "9.9942e-06", |
|
"loss": 0.5996, |
|
"slid_loss": 0.6007, |
|
"step": 122, |
|
"time": 298.46 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": "9.9940e-06", |
|
"loss": 0.6048, |
|
"slid_loss": 0.6007, |
|
"step": 123, |
|
"time": 305.47 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": "9.9939e-06", |
|
"loss": 0.6087, |
|
"slid_loss": 0.6008, |
|
"step": 124, |
|
"time": 298.35 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": "9.9938e-06", |
|
"loss": 0.5717, |
|
"slid_loss": 0.6007, |
|
"step": 125, |
|
"time": 304.29 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": "9.9936e-06", |
|
"loss": 0.587, |
|
"slid_loss": 0.6006, |
|
"step": 126, |
|
"time": 299.88 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": "9.9935e-06", |
|
"loss": 0.5759, |
|
"slid_loss": 0.6003, |
|
"step": 127, |
|
"time": 298.54 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": "9.9933e-06", |
|
"loss": 0.5657, |
|
"slid_loss": 0.5999, |
|
"step": 128, |
|
"time": 297.77 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": "9.9932e-06", |
|
"loss": 0.6105, |
|
"slid_loss": 0.5999, |
|
"step": 129, |
|
"time": 295.44 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": "9.9930e-06", |
|
"loss": 0.5847, |
|
"slid_loss": 0.5996, |
|
"step": 130, |
|
"time": 304.95 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": "9.9929e-06", |
|
"loss": 0.58, |
|
"slid_loss": 0.5995, |
|
"step": 131, |
|
"time": 301.05 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": "9.9927e-06", |
|
"loss": 0.5847, |
|
"slid_loss": 0.5992, |
|
"step": 132, |
|
"time": 302.39 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": "9.9925e-06", |
|
"loss": 0.6222, |
|
"slid_loss": 0.5995, |
|
"step": 133, |
|
"time": 306.34 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": "9.9924e-06", |
|
"loss": 0.6114, |
|
"slid_loss": 0.5994, |
|
"step": 134, |
|
"time": 299.73 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": "9.9922e-06", |
|
"loss": 0.5823, |
|
"slid_loss": 0.5991, |
|
"step": 135, |
|
"time": 340.32 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": "9.9921e-06", |
|
"loss": 0.5989, |
|
"slid_loss": 0.599, |
|
"step": 136, |
|
"time": 302.77 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": "9.9919e-06", |
|
"loss": 0.5885, |
|
"slid_loss": 0.5989, |
|
"step": 137, |
|
"time": 300.49 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": "9.9917e-06", |
|
"loss": 0.6046, |
|
"slid_loss": 0.5989, |
|
"step": 138, |
|
"time": 300.88 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": "9.9916e-06", |
|
"loss": 0.6067, |
|
"slid_loss": 0.5991, |
|
"step": 139, |
|
"time": 302.46 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": "9.9914e-06", |
|
"loss": 0.5841, |
|
"slid_loss": 0.5989, |
|
"step": 140, |
|
"time": 299.23 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": "9.9912e-06", |
|
"loss": 0.5965, |
|
"slid_loss": 0.5991, |
|
"step": 141, |
|
"time": 304.79 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": "9.9910e-06", |
|
"loss": 0.6078, |
|
"slid_loss": 0.5992, |
|
"step": 142, |
|
"time": 303.42 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": "9.9909e-06", |
|
"loss": 0.5876, |
|
"slid_loss": 0.5991, |
|
"step": 143, |
|
"time": 300.34 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": "9.9907e-06", |
|
"loss": 0.5699, |
|
"slid_loss": 0.5986, |
|
"step": 144, |
|
"time": 301.76 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": "9.9905e-06", |
|
"loss": 0.594, |
|
"slid_loss": 0.5987, |
|
"step": 145, |
|
"time": 301.5 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": "9.9903e-06", |
|
"loss": 0.5813, |
|
"slid_loss": 0.5986, |
|
"step": 146, |
|
"time": 305.53 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": "9.9902e-06", |
|
"loss": 0.6037, |
|
"slid_loss": 0.5985, |
|
"step": 147, |
|
"time": 299.21 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": "9.9900e-06", |
|
"loss": 0.6152, |
|
"slid_loss": 0.5989, |
|
"step": 148, |
|
"time": 302.99 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": "9.9898e-06", |
|
"loss": 0.5807, |
|
"slid_loss": 0.5987, |
|
"step": 149, |
|
"time": 300.63 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": "9.9896e-06", |
|
"loss": 0.6118, |
|
"slid_loss": 0.5985, |
|
"step": 150, |
|
"time": 300.98 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": "9.9894e-06", |
|
"loss": 0.6146, |
|
"slid_loss": 0.5989, |
|
"step": 151, |
|
"time": 302.41 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": "9.9892e-06", |
|
"loss": 0.5905, |
|
"slid_loss": 0.5987, |
|
"step": 152, |
|
"time": 302.49 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": "9.9890e-06", |
|
"loss": 0.5948, |
|
"slid_loss": 0.5987, |
|
"step": 153, |
|
"time": 298.31 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": "9.9888e-06", |
|
"loss": 0.5909, |
|
"slid_loss": 0.5984, |
|
"step": 154, |
|
"time": 303.58 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": "9.9886e-06", |
|
"loss": 0.5935, |
|
"slid_loss": 0.5983, |
|
"step": 155, |
|
"time": 300.33 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": "9.9884e-06", |
|
"loss": 0.5986, |
|
"slid_loss": 0.5985, |
|
"step": 156, |
|
"time": 302.57 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": "9.9882e-06", |
|
"loss": 0.5854, |
|
"slid_loss": 0.5985, |
|
"step": 157, |
|
"time": 300.71 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": "9.9880e-06", |
|
"loss": 0.5969, |
|
"slid_loss": 0.5985, |
|
"step": 158, |
|
"time": 303.74 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": "9.9878e-06", |
|
"loss": 0.5823, |
|
"slid_loss": 0.5981, |
|
"step": 159, |
|
"time": 297.54 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": "9.9876e-06", |
|
"loss": 0.6209, |
|
"slid_loss": 0.5983, |
|
"step": 160, |
|
"time": 301.04 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": "9.9874e-06", |
|
"loss": 0.6104, |
|
"slid_loss": 0.5984, |
|
"step": 161, |
|
"time": 299.43 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": "9.9872e-06", |
|
"loss": 0.6082, |
|
"slid_loss": 0.5982, |
|
"step": 162, |
|
"time": 303.51 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": "9.9870e-06", |
|
"loss": 0.594, |
|
"slid_loss": 0.598, |
|
"step": 163, |
|
"time": 299.33 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": "9.9868e-06", |
|
"loss": 0.5984, |
|
"slid_loss": 0.5978, |
|
"step": 164, |
|
"time": 300.6 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": "9.9866e-06", |
|
"loss": 0.5662, |
|
"slid_loss": 0.5974, |
|
"step": 165, |
|
"time": 300.92 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": "9.9864e-06", |
|
"loss": 0.5915, |
|
"slid_loss": 0.5973, |
|
"step": 166, |
|
"time": 302.81 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": "9.9862e-06", |
|
"loss": 0.5758, |
|
"slid_loss": 0.5971, |
|
"step": 167, |
|
"time": 297.08 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": "9.9860e-06", |
|
"loss": 0.5816, |
|
"slid_loss": 0.5973, |
|
"step": 168, |
|
"time": 299.58 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": "9.9857e-06", |
|
"loss": 0.6093, |
|
"slid_loss": 0.5976, |
|
"step": 169, |
|
"time": 298.08 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": "9.9855e-06", |
|
"loss": 0.5916, |
|
"slid_loss": 0.5973, |
|
"step": 170, |
|
"time": 299.47 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": "9.9853e-06", |
|
"loss": 0.6068, |
|
"slid_loss": 0.5973, |
|
"step": 171, |
|
"time": 301.28 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": "9.9851e-06", |
|
"loss": 0.5746, |
|
"slid_loss": 0.5969, |
|
"step": 172, |
|
"time": 298.82 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": "9.9848e-06", |
|
"loss": 0.5839, |
|
"slid_loss": 0.5968, |
|
"step": 173, |
|
"time": 297.42 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": "9.9846e-06", |
|
"loss": 0.606, |
|
"slid_loss": 0.5969, |
|
"step": 174, |
|
"time": 300.27 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": "9.9844e-06", |
|
"loss": 0.5745, |
|
"slid_loss": 0.5966, |
|
"step": 175, |
|
"time": 299.0 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": "9.9842e-06", |
|
"loss": 0.5781, |
|
"slid_loss": 0.5965, |
|
"step": 176, |
|
"time": 301.89 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": "9.9839e-06", |
|
"loss": 0.6075, |
|
"slid_loss": 0.5966, |
|
"step": 177, |
|
"time": 302.45 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": "9.9837e-06", |
|
"loss": 0.5993, |
|
"slid_loss": 0.5966, |
|
"step": 178, |
|
"time": 303.55 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": "9.9835e-06", |
|
"loss": 0.5941, |
|
"slid_loss": 0.5967, |
|
"step": 179, |
|
"time": 300.89 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": "9.9832e-06", |
|
"loss": 0.5923, |
|
"slid_loss": 0.5967, |
|
"step": 180, |
|
"time": 296.88 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": "9.9830e-06", |
|
"loss": 0.6073, |
|
"slid_loss": 0.5969, |
|
"step": 181, |
|
"time": 297.73 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": "9.9827e-06", |
|
"loss": 0.5899, |
|
"slid_loss": 0.5967, |
|
"step": 182, |
|
"time": 304.98 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": "9.9825e-06", |
|
"loss": 0.6225, |
|
"slid_loss": 0.5972, |
|
"step": 183, |
|
"time": 299.1 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": "9.9823e-06", |
|
"loss": 0.5727, |
|
"slid_loss": 0.5969, |
|
"step": 184, |
|
"time": 299.82 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": "9.9820e-06", |
|
"loss": 0.5644, |
|
"slid_loss": 0.5965, |
|
"step": 185, |
|
"time": 302.65 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": "9.9818e-06", |
|
"loss": 0.6012, |
|
"slid_loss": 0.5965, |
|
"step": 186, |
|
"time": 294.28 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": "9.9815e-06", |
|
"loss": 0.583, |
|
"slid_loss": 0.5963, |
|
"step": 187, |
|
"time": 300.57 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": "9.9813e-06", |
|
"loss": 0.6199, |
|
"slid_loss": 0.5963, |
|
"step": 188, |
|
"time": 300.21 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": "9.9810e-06", |
|
"loss": 0.5737, |
|
"slid_loss": 0.596, |
|
"step": 189, |
|
"time": 298.53 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": "9.9808e-06", |
|
"loss": 0.6084, |
|
"slid_loss": 0.5962, |
|
"step": 190, |
|
"time": 303.74 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": "9.9805e-06", |
|
"loss": 0.587, |
|
"slid_loss": 0.5961, |
|
"step": 191, |
|
"time": 303.28 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": "9.9802e-06", |
|
"loss": 0.5778, |
|
"slid_loss": 0.5958, |
|
"step": 192, |
|
"time": 296.74 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": "9.9800e-06", |
|
"loss": 0.5752, |
|
"slid_loss": 0.5953, |
|
"step": 193, |
|
"time": 301.16 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": "9.9797e-06", |
|
"loss": 0.6127, |
|
"slid_loss": 0.5955, |
|
"step": 194, |
|
"time": 296.39 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": "9.9795e-06", |
|
"loss": 0.5864, |
|
"slid_loss": 0.5953, |
|
"step": 195, |
|
"time": 298.05 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": "9.9792e-06", |
|
"loss": 0.5788, |
|
"slid_loss": 0.595, |
|
"step": 196, |
|
"time": 297.51 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": "9.9789e-06", |
|
"loss": 0.5977, |
|
"slid_loss": 0.5949, |
|
"step": 197, |
|
"time": 299.48 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": "9.9787e-06", |
|
"loss": 0.568, |
|
"slid_loss": 0.5946, |
|
"step": 198, |
|
"time": 303.2 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": "9.9784e-06", |
|
"loss": 0.5954, |
|
"slid_loss": 0.5945, |
|
"step": 199, |
|
"time": 298.52 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": "9.9781e-06", |
|
"loss": 0.618, |
|
"slid_loss": 0.5948, |
|
"step": 200, |
|
"time": 297.2 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": "9.9779e-06", |
|
"loss": 0.568, |
|
"slid_loss": 0.5945, |
|
"step": 201, |
|
"time": 2159.56 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": "9.9776e-06", |
|
"loss": 0.5864, |
|
"slid_loss": 0.594, |
|
"step": 202, |
|
"time": 304.84 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": "9.9773e-06", |
|
"loss": 0.5866, |
|
"slid_loss": 0.5937, |
|
"step": 203, |
|
"time": 301.84 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": "9.9770e-06", |
|
"loss": 0.633, |
|
"slid_loss": 0.5942, |
|
"step": 204, |
|
"time": 300.37 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": "9.9768e-06", |
|
"loss": 0.5984, |
|
"slid_loss": 0.5939, |
|
"step": 205, |
|
"time": 298.58 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": "9.9765e-06", |
|
"loss": 0.5859, |
|
"slid_loss": 0.5939, |
|
"step": 206, |
|
"time": 298.88 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": "9.9762e-06", |
|
"loss": 0.5895, |
|
"slid_loss": 0.594, |
|
"step": 207, |
|
"time": 302.29 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": "9.9759e-06", |
|
"loss": 0.5931, |
|
"slid_loss": 0.594, |
|
"step": 208, |
|
"time": 295.71 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": "9.9756e-06", |
|
"loss": 0.586, |
|
"slid_loss": 0.5939, |
|
"step": 209, |
|
"time": 297.38 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": "9.9753e-06", |
|
"loss": 0.5702, |
|
"slid_loss": 0.5937, |
|
"step": 210, |
|
"time": 292.88 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": "9.9750e-06", |
|
"loss": 0.5841, |
|
"slid_loss": 0.5933, |
|
"step": 211, |
|
"time": 302.19 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": "9.9748e-06", |
|
"loss": 0.5839, |
|
"slid_loss": 0.5932, |
|
"step": 212, |
|
"time": 295.71 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": "9.9745e-06", |
|
"loss": 0.6085, |
|
"slid_loss": 0.5932, |
|
"step": 213, |
|
"time": 301.35 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": "9.9742e-06", |
|
"loss": 0.5931, |
|
"slid_loss": 0.5932, |
|
"step": 214, |
|
"time": 297.9 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": "9.9739e-06", |
|
"loss": 0.6003, |
|
"slid_loss": 0.5933, |
|
"step": 215, |
|
"time": 305.13 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": "9.9736e-06", |
|
"loss": 0.5904, |
|
"slid_loss": 0.5934, |
|
"step": 216, |
|
"time": 299.72 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": "9.9733e-06", |
|
"loss": 0.5789, |
|
"slid_loss": 0.5931, |
|
"step": 217, |
|
"time": 294.25 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": "9.9730e-06", |
|
"loss": 0.5975, |
|
"slid_loss": 0.5932, |
|
"step": 218, |
|
"time": 300.49 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": "9.9727e-06", |
|
"loss": 0.5814, |
|
"slid_loss": 0.5931, |
|
"step": 219, |
|
"time": 300.88 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": "9.9724e-06", |
|
"loss": 0.6156, |
|
"slid_loss": 0.593, |
|
"step": 220, |
|
"time": 298.38 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": "9.9721e-06", |
|
"loss": 0.5958, |
|
"slid_loss": 0.593, |
|
"step": 221, |
|
"time": 295.82 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": "9.9718e-06", |
|
"loss": 0.5821, |
|
"slid_loss": 0.5928, |
|
"step": 222, |
|
"time": 297.92 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": "9.9714e-06", |
|
"loss": 0.6157, |
|
"slid_loss": 0.5929, |
|
"step": 223, |
|
"time": 301.05 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": "9.9711e-06", |
|
"loss": 0.5596, |
|
"slid_loss": 0.5924, |
|
"step": 224, |
|
"time": 300.03 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": "9.9708e-06", |
|
"loss": 0.6077, |
|
"slid_loss": 0.5928, |
|
"step": 225, |
|
"time": 302.72 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": "9.9705e-06", |
|
"loss": 0.6011, |
|
"slid_loss": 0.5929, |
|
"step": 226, |
|
"time": 300.68 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": "9.9702e-06", |
|
"loss": 0.5792, |
|
"slid_loss": 0.593, |
|
"step": 227, |
|
"time": 301.7 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": "9.9699e-06", |
|
"loss": 0.5819, |
|
"slid_loss": 0.5931, |
|
"step": 228, |
|
"time": 307.53 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": "9.9696e-06", |
|
"loss": 0.5668, |
|
"slid_loss": 0.5927, |
|
"step": 229, |
|
"time": 298.86 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": "9.9692e-06", |
|
"loss": 0.5833, |
|
"slid_loss": 0.5927, |
|
"step": 230, |
|
"time": 301.51 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": "9.9689e-06", |
|
"loss": 0.5911, |
|
"slid_loss": 0.5928, |
|
"step": 231, |
|
"time": 301.86 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": "9.9686e-06", |
|
"loss": 0.6039, |
|
"slid_loss": 0.593, |
|
"step": 232, |
|
"time": 301.91 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": "9.9683e-06", |
|
"loss": 0.6113, |
|
"slid_loss": 0.5929, |
|
"step": 233, |
|
"time": 299.99 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": "9.9679e-06", |
|
"loss": 0.5821, |
|
"slid_loss": 0.5926, |
|
"step": 234, |
|
"time": 300.63 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": "9.9676e-06", |
|
"loss": 0.584, |
|
"slid_loss": 0.5926, |
|
"step": 235, |
|
"time": 294.85 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": "9.9673e-06", |
|
"loss": 0.593, |
|
"slid_loss": 0.5925, |
|
"step": 236, |
|
"time": 300.63 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": "9.9669e-06", |
|
"loss": 0.5875, |
|
"slid_loss": 0.5925, |
|
"step": 237, |
|
"time": 301.49 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": "9.9666e-06", |
|
"loss": 0.6002, |
|
"slid_loss": 0.5925, |
|
"step": 238, |
|
"time": 300.88 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": "9.9663e-06", |
|
"loss": 0.5786, |
|
"slid_loss": 0.5922, |
|
"step": 239, |
|
"time": 300.01 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": "9.9659e-06", |
|
"loss": 0.596, |
|
"slid_loss": 0.5923, |
|
"step": 240, |
|
"time": 299.06 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": "9.9656e-06", |
|
"loss": 0.5881, |
|
"slid_loss": 0.5922, |
|
"step": 241, |
|
"time": 398.92 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": "9.9653e-06", |
|
"loss": 0.5988, |
|
"slid_loss": 0.5921, |
|
"step": 242, |
|
"time": 512.79 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": "9.9649e-06", |
|
"loss": 0.5969, |
|
"slid_loss": 0.5922, |
|
"step": 243, |
|
"time": 333.47 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": "9.9646e-06", |
|
"loss": 0.58, |
|
"slid_loss": 0.5923, |
|
"step": 244, |
|
"time": 303.21 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": "9.9642e-06", |
|
"loss": 0.5971, |
|
"slid_loss": 0.5924, |
|
"step": 245, |
|
"time": 296.87 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": "9.9639e-06", |
|
"loss": 0.5809, |
|
"slid_loss": 0.5924, |
|
"step": 246, |
|
"time": 303.53 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": "9.9635e-06", |
|
"loss": 0.6039, |
|
"slid_loss": 0.5924, |
|
"step": 247, |
|
"time": 304.38 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": "9.9632e-06", |
|
"loss": 0.6062, |
|
"slid_loss": 0.5923, |
|
"step": 248, |
|
"time": 299.2 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": "9.9628e-06", |
|
"loss": 0.5842, |
|
"slid_loss": 0.5923, |
|
"step": 249, |
|
"time": 295.64 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": "9.9625e-06", |
|
"loss": 0.6137, |
|
"slid_loss": 0.5923, |
|
"step": 250, |
|
"time": 300.97 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": "9.9621e-06", |
|
"loss": 0.5899, |
|
"slid_loss": 0.5921, |
|
"step": 251, |
|
"time": 304.22 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": "9.9618e-06", |
|
"loss": 0.5839, |
|
"slid_loss": 0.592, |
|
"step": 252, |
|
"time": 300.56 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": "9.9614e-06", |
|
"loss": 0.6023, |
|
"slid_loss": 0.5921, |
|
"step": 253, |
|
"time": 302.46 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": "9.9610e-06", |
|
"loss": 0.6089, |
|
"slid_loss": 0.5923, |
|
"step": 254, |
|
"time": 300.0 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": "9.9607e-06", |
|
"loss": 0.6055, |
|
"slid_loss": 0.5924, |
|
"step": 255, |
|
"time": 300.65 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": "9.9603e-06", |
|
"loss": 0.5927, |
|
"slid_loss": 0.5923, |
|
"step": 256, |
|
"time": 296.14 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": "9.9599e-06", |
|
"loss": 0.5967, |
|
"slid_loss": 0.5925, |
|
"step": 257, |
|
"time": 294.8 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": "9.9596e-06", |
|
"loss": 0.5717, |
|
"slid_loss": 0.5922, |
|
"step": 258, |
|
"time": 303.38 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": "9.9592e-06", |
|
"loss": 0.5853, |
|
"slid_loss": 0.5922, |
|
"step": 259, |
|
"time": 301.95 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": "9.9588e-06", |
|
"loss": 0.5944, |
|
"slid_loss": 0.592, |
|
"step": 260, |
|
"time": 300.31 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": "9.9585e-06", |
|
"loss": 0.6069, |
|
"slid_loss": 0.5919, |
|
"step": 261, |
|
"time": 303.26 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": "9.9581e-06", |
|
"loss": 0.5901, |
|
"slid_loss": 0.5917, |
|
"step": 262, |
|
"time": 301.92 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": "9.9577e-06", |
|
"loss": 0.604, |
|
"slid_loss": 0.5918, |
|
"step": 263, |
|
"time": 302.96 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": "9.9573e-06", |
|
"loss": 0.5833, |
|
"slid_loss": 0.5917, |
|
"step": 264, |
|
"time": 301.58 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": "9.9570e-06", |
|
"loss": 0.5949, |
|
"slid_loss": 0.592, |
|
"step": 265, |
|
"time": 301.0 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": "9.9566e-06", |
|
"loss": 0.5695, |
|
"slid_loss": 0.5918, |
|
"step": 266, |
|
"time": 300.87 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": "9.9562e-06", |
|
"loss": 0.5978, |
|
"slid_loss": 0.592, |
|
"step": 267, |
|
"time": 299.95 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": "9.9558e-06", |
|
"loss": 0.5796, |
|
"slid_loss": 0.592, |
|
"step": 268, |
|
"time": 293.55 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": "9.9554e-06", |
|
"loss": 0.6032, |
|
"slid_loss": 0.5919, |
|
"step": 269, |
|
"time": 302.36 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": "9.9550e-06", |
|
"loss": 0.6002, |
|
"slid_loss": 0.592, |
|
"step": 270, |
|
"time": 303.01 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": "9.9547e-06", |
|
"loss": 0.6025, |
|
"slid_loss": 0.5919, |
|
"step": 271, |
|
"time": 303.25 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": "9.9543e-06", |
|
"loss": 0.6, |
|
"slid_loss": 0.5922, |
|
"step": 272, |
|
"time": 303.51 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": "9.9539e-06", |
|
"loss": 0.5961, |
|
"slid_loss": 0.5923, |
|
"step": 273, |
|
"time": 300.67 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": "9.9535e-06", |
|
"loss": 0.5879, |
|
"slid_loss": 0.5921, |
|
"step": 274, |
|
"time": 298.22 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": "9.9531e-06", |
|
"loss": 0.5915, |
|
"slid_loss": 0.5923, |
|
"step": 275, |
|
"time": 294.33 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": "9.9527e-06", |
|
"loss": 0.5785, |
|
"slid_loss": 0.5923, |
|
"step": 276, |
|
"time": 303.84 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": "9.9523e-06", |
|
"loss": 0.5918, |
|
"slid_loss": 0.5922, |
|
"step": 277, |
|
"time": 301.67 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": "9.9519e-06", |
|
"loss": 0.5767, |
|
"slid_loss": 0.5919, |
|
"step": 278, |
|
"time": 298.92 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": "9.9515e-06", |
|
"loss": 0.5884, |
|
"slid_loss": 0.5919, |
|
"step": 279, |
|
"time": 301.91 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": "9.9511e-06", |
|
"loss": 0.5535, |
|
"slid_loss": 0.5915, |
|
"step": 280, |
|
"time": 305.25 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": "9.9507e-06", |
|
"loss": 0.5793, |
|
"slid_loss": 0.5912, |
|
"step": 281, |
|
"time": 304.08 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": "9.9503e-06", |
|
"loss": 0.6179, |
|
"slid_loss": 0.5915, |
|
"step": 282, |
|
"time": 302.88 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": "9.9499e-06", |
|
"loss": 0.5902, |
|
"slid_loss": 0.5912, |
|
"step": 283, |
|
"time": 304.51 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": "9.9495e-06", |
|
"loss": 0.607, |
|
"slid_loss": 0.5915, |
|
"step": 284, |
|
"time": 303.7 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": "9.9490e-06", |
|
"loss": 0.5717, |
|
"slid_loss": 0.5916, |
|
"step": 285, |
|
"time": 293.89 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": "9.9486e-06", |
|
"loss": 0.5917, |
|
"slid_loss": 0.5915, |
|
"step": 286, |
|
"time": 294.93 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": "9.9482e-06", |
|
"loss": 0.6037, |
|
"slid_loss": 0.5917, |
|
"step": 287, |
|
"time": 292.8 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": "9.9478e-06", |
|
"loss": 0.5852, |
|
"slid_loss": 0.5913, |
|
"step": 288, |
|
"time": 305.96 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": "9.9474e-06", |
|
"loss": 0.6088, |
|
"slid_loss": 0.5917, |
|
"step": 289, |
|
"time": 300.0 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": "9.9470e-06", |
|
"loss": 0.589, |
|
"slid_loss": 0.5915, |
|
"step": 290, |
|
"time": 298.64 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": "9.9465e-06", |
|
"loss": 0.5874, |
|
"slid_loss": 0.5915, |
|
"step": 291, |
|
"time": 299.51 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": "9.9461e-06", |
|
"loss": 0.6055, |
|
"slid_loss": 0.5918, |
|
"step": 292, |
|
"time": 303.2 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": "9.9457e-06", |
|
"loss": 0.5701, |
|
"slid_loss": 0.5917, |
|
"step": 293, |
|
"time": 306.15 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": "9.9453e-06", |
|
"loss": 0.551, |
|
"slid_loss": 0.5911, |
|
"step": 294, |
|
"time": 296.03 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": "9.9448e-06", |
|
"loss": 0.5914, |
|
"slid_loss": 0.5912, |
|
"step": 295, |
|
"time": 299.9 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": "9.9444e-06", |
|
"loss": 0.6068, |
|
"slid_loss": 0.5914, |
|
"step": 296, |
|
"time": 303.06 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": "9.9440e-06", |
|
"loss": 0.5881, |
|
"slid_loss": 0.5914, |
|
"step": 297, |
|
"time": 302.97 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": "9.9435e-06", |
|
"loss": 0.5652, |
|
"slid_loss": 0.5913, |
|
"step": 298, |
|
"time": 299.7 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": "9.9431e-06", |
|
"loss": 0.5771, |
|
"slid_loss": 0.5911, |
|
"step": 299, |
|
"time": 297.81 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": "9.9427e-06", |
|
"loss": 0.5984, |
|
"slid_loss": 0.5909, |
|
"step": 300, |
|
"time": 309.93 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": "9.9422e-06", |
|
"loss": 0.5798, |
|
"slid_loss": 0.5911, |
|
"step": 301, |
|
"time": 2145.71 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": "9.9418e-06", |
|
"loss": 0.5467, |
|
"slid_loss": 0.5907, |
|
"step": 302, |
|
"time": 304.18 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": "9.9413e-06", |
|
"loss": 0.6062, |
|
"slid_loss": 0.5909, |
|
"step": 303, |
|
"time": 303.3 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": "9.9409e-06", |
|
"loss": 0.5981, |
|
"slid_loss": 0.5905, |
|
"step": 304, |
|
"time": 301.71 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": "9.9404e-06", |
|
"loss": 0.5806, |
|
"slid_loss": 0.5903, |
|
"step": 305, |
|
"time": 298.37 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": "9.9400e-06", |
|
"loss": 0.6079, |
|
"slid_loss": 0.5906, |
|
"step": 306, |
|
"time": 297.14 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": "9.9396e-06", |
|
"loss": 0.5935, |
|
"slid_loss": 0.5906, |
|
"step": 307, |
|
"time": 302.82 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": "9.9391e-06", |
|
"loss": 0.561, |
|
"slid_loss": 0.5903, |
|
"step": 308, |
|
"time": 298.85 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": "9.9387e-06", |
|
"loss": 0.603, |
|
"slid_loss": 0.5904, |
|
"step": 309, |
|
"time": 298.61 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": "9.9382e-06", |
|
"loss": 0.5999, |
|
"slid_loss": 0.5907, |
|
"step": 310, |
|
"time": 301.25 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": "9.9377e-06", |
|
"loss": 0.5995, |
|
"slid_loss": 0.5909, |
|
"step": 311, |
|
"time": 299.59 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": "9.9373e-06", |
|
"loss": 0.5712, |
|
"slid_loss": 0.5908, |
|
"step": 312, |
|
"time": 302.55 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": "9.9368e-06", |
|
"loss": 0.5979, |
|
"slid_loss": 0.5907, |
|
"step": 313, |
|
"time": 298.82 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": "9.9364e-06", |
|
"loss": 0.5781, |
|
"slid_loss": 0.5905, |
|
"step": 314, |
|
"time": 296.46 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": "9.9359e-06", |
|
"loss": 0.5776, |
|
"slid_loss": 0.5903, |
|
"step": 315, |
|
"time": 299.03 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": "9.9354e-06", |
|
"loss": 0.575, |
|
"slid_loss": 0.5901, |
|
"step": 316, |
|
"time": 298.66 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": "9.9350e-06", |
|
"loss": 0.5841, |
|
"slid_loss": 0.5902, |
|
"step": 317, |
|
"time": 299.26 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": "9.9345e-06", |
|
"loss": 0.5874, |
|
"slid_loss": 0.5901, |
|
"step": 318, |
|
"time": 298.97 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": "9.9340e-06", |
|
"loss": 0.5836, |
|
"slid_loss": 0.5901, |
|
"step": 319, |
|
"time": 301.89 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": "9.9336e-06", |
|
"loss": 0.5722, |
|
"slid_loss": 0.5897, |
|
"step": 320, |
|
"time": 301.23 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": "9.9331e-06", |
|
"loss": 0.5815, |
|
"slid_loss": 0.5895, |
|
"step": 321, |
|
"time": 300.73 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": "9.9326e-06", |
|
"loss": 0.582, |
|
"slid_loss": 0.5895, |
|
"step": 322, |
|
"time": 306.95 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": "9.9322e-06", |
|
"loss": 0.5925, |
|
"slid_loss": 0.5893, |
|
"step": 323, |
|
"time": 303.77 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": "9.9317e-06", |
|
"loss": 0.5637, |
|
"slid_loss": 0.5893, |
|
"step": 324, |
|
"time": 302.23 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": "9.9312e-06", |
|
"loss": 0.5697, |
|
"slid_loss": 0.589, |
|
"step": 325, |
|
"time": 300.31 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": "9.9307e-06", |
|
"loss": 0.5983, |
|
"slid_loss": 0.5889, |
|
"step": 326, |
|
"time": 304.98 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": "9.9302e-06", |
|
"loss": 0.5884, |
|
"slid_loss": 0.589, |
|
"step": 327, |
|
"time": 295.6 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": "9.9298e-06", |
|
"loss": 0.5902, |
|
"slid_loss": 0.5891, |
|
"step": 328, |
|
"time": 300.44 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": "9.9293e-06", |
|
"loss": 0.5706, |
|
"slid_loss": 0.5891, |
|
"step": 329, |
|
"time": 299.45 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": "9.9288e-06", |
|
"loss": 0.5751, |
|
"slid_loss": 0.5891, |
|
"step": 330, |
|
"time": 300.15 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": "9.9283e-06", |
|
"loss": 0.5948, |
|
"slid_loss": 0.5891, |
|
"step": 331, |
|
"time": 301.39 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": "9.9278e-06", |
|
"loss": 0.585, |
|
"slid_loss": 0.5889, |
|
"step": 332, |
|
"time": 300.47 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": "9.9273e-06", |
|
"loss": 0.598, |
|
"slid_loss": 0.5888, |
|
"step": 333, |
|
"time": 301.51 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": "9.9268e-06", |
|
"loss": 0.5929, |
|
"slid_loss": 0.5889, |
|
"step": 334, |
|
"time": 301.08 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": "9.9263e-06", |
|
"loss": 0.6013, |
|
"slid_loss": 0.5891, |
|
"step": 335, |
|
"time": 301.25 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": "9.9258e-06", |
|
"loss": 0.5635, |
|
"slid_loss": 0.5888, |
|
"step": 336, |
|
"time": 302.84 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": "9.9253e-06", |
|
"loss": 0.583, |
|
"slid_loss": 0.5887, |
|
"step": 337, |
|
"time": 302.95 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": "9.9248e-06", |
|
"loss": 0.5615, |
|
"slid_loss": 0.5883, |
|
"step": 338, |
|
"time": 304.5 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": "9.9243e-06", |
|
"loss": 0.5832, |
|
"slid_loss": 0.5884, |
|
"step": 339, |
|
"time": 301.31 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": "9.9238e-06", |
|
"loss": 0.5584, |
|
"slid_loss": 0.588, |
|
"step": 340, |
|
"time": 301.03 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": "9.9233e-06", |
|
"loss": 0.5838, |
|
"slid_loss": 0.588, |
|
"step": 341, |
|
"time": 300.61 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": "9.9228e-06", |
|
"loss": 0.5854, |
|
"slid_loss": 0.5878, |
|
"step": 342, |
|
"time": 304.07 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": "9.9223e-06", |
|
"loss": 0.5775, |
|
"slid_loss": 0.5876, |
|
"step": 343, |
|
"time": 304.08 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": "9.9218e-06", |
|
"loss": 0.587, |
|
"slid_loss": 0.5877, |
|
"step": 344, |
|
"time": 301.56 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": "9.9213e-06", |
|
"loss": 0.5812, |
|
"slid_loss": 0.5875, |
|
"step": 345, |
|
"time": 297.96 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": "9.9208e-06", |
|
"loss": 0.5783, |
|
"slid_loss": 0.5875, |
|
"step": 346, |
|
"time": 302.47 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": "9.9203e-06", |
|
"loss": 0.5731, |
|
"slid_loss": 0.5872, |
|
"step": 347, |
|
"time": 297.21 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": "9.9198e-06", |
|
"loss": 0.5725, |
|
"slid_loss": 0.5869, |
|
"step": 348, |
|
"time": 300.15 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": "9.9192e-06", |
|
"loss": 0.5565, |
|
"slid_loss": 0.5866, |
|
"step": 349, |
|
"time": 299.24 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": "9.9187e-06", |
|
"loss": 0.6033, |
|
"slid_loss": 0.5865, |
|
"step": 350, |
|
"time": 297.98 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": "9.9182e-06", |
|
"loss": 0.5588, |
|
"slid_loss": 0.5862, |
|
"step": 351, |
|
"time": 298.4 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": "9.9177e-06", |
|
"loss": 0.5595, |
|
"slid_loss": 0.5859, |
|
"step": 352, |
|
"time": 300.12 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": "9.9171e-06", |
|
"loss": 0.5777, |
|
"slid_loss": 0.5857, |
|
"step": 353, |
|
"time": 293.74 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": "9.9166e-06", |
|
"loss": 0.587, |
|
"slid_loss": 0.5855, |
|
"step": 354, |
|
"time": 300.09 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": "9.9161e-06", |
|
"loss": 0.5791, |
|
"slid_loss": 0.5852, |
|
"step": 355, |
|
"time": 299.74 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": "9.9156e-06", |
|
"loss": 0.5859, |
|
"slid_loss": 0.5851, |
|
"step": 356, |
|
"time": 299.89 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": "9.9150e-06", |
|
"loss": 0.5808, |
|
"slid_loss": 0.585, |
|
"step": 357, |
|
"time": 299.12 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": "9.9145e-06", |
|
"loss": 0.5901, |
|
"slid_loss": 0.5852, |
|
"step": 358, |
|
"time": 301.57 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": "9.9140e-06", |
|
"loss": 0.5809, |
|
"slid_loss": 0.5851, |
|
"step": 359, |
|
"time": 302.98 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": "9.9134e-06", |
|
"loss": 0.5859, |
|
"slid_loss": 0.585, |
|
"step": 360, |
|
"time": 296.56 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": "9.9129e-06", |
|
"loss": 0.592, |
|
"slid_loss": 0.5849, |
|
"step": 361, |
|
"time": 309.77 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": "9.9124e-06", |
|
"loss": 0.5672, |
|
"slid_loss": 0.5847, |
|
"step": 362, |
|
"time": 442.06 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": "9.9118e-06", |
|
"loss": 0.5837, |
|
"slid_loss": 0.5844, |
|
"step": 363, |
|
"time": 543.81 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": "9.9113e-06", |
|
"loss": 0.5883, |
|
"slid_loss": 0.5845, |
|
"step": 364, |
|
"time": 307.58 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": "9.9107e-06", |
|
"loss": 0.5914, |
|
"slid_loss": 0.5845, |
|
"step": 365, |
|
"time": 302.4 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": "9.9102e-06", |
|
"loss": 0.5901, |
|
"slid_loss": 0.5847, |
|
"step": 366, |
|
"time": 304.61 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": "9.9096e-06", |
|
"loss": 0.6143, |
|
"slid_loss": 0.5848, |
|
"step": 367, |
|
"time": 301.14 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": "9.9091e-06", |
|
"loss": 0.5647, |
|
"slid_loss": 0.5847, |
|
"step": 368, |
|
"time": 304.0 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": "9.9085e-06", |
|
"loss": 0.5874, |
|
"slid_loss": 0.5845, |
|
"step": 369, |
|
"time": 299.09 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": "9.9080e-06", |
|
"loss": 0.5769, |
|
"slid_loss": 0.5843, |
|
"step": 370, |
|
"time": 300.46 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": "9.9074e-06", |
|
"loss": 0.5568, |
|
"slid_loss": 0.5838, |
|
"step": 371, |
|
"time": 299.6 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": "9.9069e-06", |
|
"loss": 0.5953, |
|
"slid_loss": 0.5838, |
|
"step": 372, |
|
"time": 303.64 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": "9.9063e-06", |
|
"loss": 0.5935, |
|
"slid_loss": 0.5838, |
|
"step": 373, |
|
"time": 302.9 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": "9.9058e-06", |
|
"loss": 0.5781, |
|
"slid_loss": 0.5837, |
|
"step": 374, |
|
"time": 303.66 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": "9.9052e-06", |
|
"loss": 0.5954, |
|
"slid_loss": 0.5837, |
|
"step": 375, |
|
"time": 301.25 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": "9.9046e-06", |
|
"loss": 0.5562, |
|
"slid_loss": 0.5835, |
|
"step": 376, |
|
"time": 302.16 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": "9.9041e-06", |
|
"loss": 0.5978, |
|
"slid_loss": 0.5835, |
|
"step": 377, |
|
"time": 301.0 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": "9.9035e-06", |
|
"loss": 0.5904, |
|
"slid_loss": 0.5837, |
|
"step": 378, |
|
"time": 301.57 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": "9.9030e-06", |
|
"loss": 0.61, |
|
"slid_loss": 0.5839, |
|
"step": 379, |
|
"time": 298.34 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": "9.9024e-06", |
|
"loss": 0.5824, |
|
"slid_loss": 0.5842, |
|
"step": 380, |
|
"time": 298.45 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": "9.9018e-06", |
|
"loss": 0.5803, |
|
"slid_loss": 0.5842, |
|
"step": 381, |
|
"time": 299.83 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": "9.9012e-06", |
|
"loss": 0.6062, |
|
"slid_loss": 0.5841, |
|
"step": 382, |
|
"time": 302.79 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": "9.9007e-06", |
|
"loss": 0.5783, |
|
"slid_loss": 0.584, |
|
"step": 383, |
|
"time": 304.89 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": "9.9001e-06", |
|
"loss": 0.5873, |
|
"slid_loss": 0.5838, |
|
"step": 384, |
|
"time": 299.4 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": "9.8995e-06", |
|
"loss": 0.5714, |
|
"slid_loss": 0.5838, |
|
"step": 385, |
|
"time": 302.76 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": "9.8989e-06", |
|
"loss": 0.5849, |
|
"slid_loss": 0.5837, |
|
"step": 386, |
|
"time": 298.68 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": "9.8984e-06", |
|
"loss": 0.5512, |
|
"slid_loss": 0.5832, |
|
"step": 387, |
|
"time": 302.47 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": "9.8978e-06", |
|
"loss": 0.5851, |
|
"slid_loss": 0.5832, |
|
"step": 388, |
|
"time": 300.3 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": "9.8972e-06", |
|
"loss": 0.5749, |
|
"slid_loss": 0.5828, |
|
"step": 389, |
|
"time": 303.56 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": "9.8966e-06", |
|
"loss": 0.5559, |
|
"slid_loss": 0.5825, |
|
"step": 390, |
|
"time": 303.89 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": "9.8960e-06", |
|
"loss": 0.5666, |
|
"slid_loss": 0.5823, |
|
"step": 391, |
|
"time": 298.94 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": "9.8954e-06", |
|
"loss": 0.5939, |
|
"slid_loss": 0.5822, |
|
"step": 392, |
|
"time": 299.49 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": "9.8949e-06", |
|
"loss": 0.5914, |
|
"slid_loss": 0.5824, |
|
"step": 393, |
|
"time": 300.26 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": "9.8943e-06", |
|
"loss": 0.5602, |
|
"slid_loss": 0.5825, |
|
"step": 394, |
|
"time": 301.06 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": "9.8937e-06", |
|
"loss": 0.5878, |
|
"slid_loss": 0.5824, |
|
"step": 395, |
|
"time": 298.55 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": "9.8931e-06", |
|
"loss": 0.5964, |
|
"slid_loss": 0.5823, |
|
"step": 396, |
|
"time": 297.95 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": "9.8925e-06", |
|
"loss": 0.6124, |
|
"slid_loss": 0.5826, |
|
"step": 397, |
|
"time": 302.23 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": "9.8919e-06", |
|
"loss": 0.5672, |
|
"slid_loss": 0.5826, |
|
"step": 398, |
|
"time": 298.67 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": "9.8913e-06", |
|
"loss": 0.5678, |
|
"slid_loss": 0.5825, |
|
"step": 399, |
|
"time": 296.19 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": "9.8907e-06", |
|
"loss": 0.5952, |
|
"slid_loss": 0.5825, |
|
"step": 400, |
|
"time": 294.41 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": "9.8901e-06", |
|
"loss": 0.5817, |
|
"slid_loss": 0.5825, |
|
"step": 401, |
|
"time": 2131.8 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": "9.8895e-06", |
|
"loss": 0.5577, |
|
"slid_loss": 0.5826, |
|
"step": 402, |
|
"time": 303.45 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": "9.8889e-06", |
|
"loss": 0.5867, |
|
"slid_loss": 0.5824, |
|
"step": 403, |
|
"time": 303.95 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": "9.8883e-06", |
|
"loss": 0.5767, |
|
"slid_loss": 0.5822, |
|
"step": 404, |
|
"time": 299.6 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": "9.8877e-06", |
|
"loss": 0.5631, |
|
"slid_loss": 0.582, |
|
"step": 405, |
|
"time": 296.97 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": "9.8870e-06", |
|
"loss": 0.5671, |
|
"slid_loss": 0.5816, |
|
"step": 406, |
|
"time": 296.01 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": "9.8864e-06", |
|
"loss": 0.5607, |
|
"slid_loss": 0.5813, |
|
"step": 407, |
|
"time": 292.73 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": "9.8858e-06", |
|
"loss": 0.5612, |
|
"slid_loss": 0.5813, |
|
"step": 408, |
|
"time": 299.45 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": "9.8852e-06", |
|
"loss": 0.5591, |
|
"slid_loss": 0.5808, |
|
"step": 409, |
|
"time": 296.08 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": "9.8846e-06", |
|
"loss": 0.5762, |
|
"slid_loss": 0.5806, |
|
"step": 410, |
|
"time": 299.42 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": "9.8840e-06", |
|
"loss": 0.5762, |
|
"slid_loss": 0.5804, |
|
"step": 411, |
|
"time": 302.72 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": "9.8834e-06", |
|
"loss": 0.5675, |
|
"slid_loss": 0.5803, |
|
"step": 412, |
|
"time": 304.89 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": "9.8827e-06", |
|
"loss": 0.5656, |
|
"slid_loss": 0.58, |
|
"step": 413, |
|
"time": 300.11 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": "9.8821e-06", |
|
"loss": 0.599, |
|
"slid_loss": 0.5802, |
|
"step": 414, |
|
"time": 300.68 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": "9.8815e-06", |
|
"loss": 0.5735, |
|
"slid_loss": 0.5802, |
|
"step": 415, |
|
"time": 300.05 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": "9.8809e-06", |
|
"loss": 0.5825, |
|
"slid_loss": 0.5803, |
|
"step": 416, |
|
"time": 300.83 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": "9.8802e-06", |
|
"loss": 0.5733, |
|
"slid_loss": 0.5802, |
|
"step": 417, |
|
"time": 298.88 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": "9.8796e-06", |
|
"loss": 0.5703, |
|
"slid_loss": 0.58, |
|
"step": 418, |
|
"time": 305.66 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": "9.8790e-06", |
|
"loss": 0.5645, |
|
"slid_loss": 0.5798, |
|
"step": 419, |
|
"time": 300.94 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": "9.8783e-06", |
|
"loss": 0.5623, |
|
"slid_loss": 0.5797, |
|
"step": 420, |
|
"time": 299.63 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": "9.8777e-06", |
|
"loss": 0.5444, |
|
"slid_loss": 0.5793, |
|
"step": 421, |
|
"time": 298.09 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": "9.8771e-06", |
|
"loss": 0.5864, |
|
"slid_loss": 0.5794, |
|
"step": 422, |
|
"time": 297.47 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": "9.8764e-06", |
|
"loss": 0.6055, |
|
"slid_loss": 0.5795, |
|
"step": 423, |
|
"time": 304.79 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": "9.8758e-06", |
|
"loss": 0.57, |
|
"slid_loss": 0.5796, |
|
"step": 424, |
|
"time": 301.25 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": "9.8752e-06", |
|
"loss": 0.5871, |
|
"slid_loss": 0.5797, |
|
"step": 425, |
|
"time": 298.86 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": "9.8745e-06", |
|
"loss": 0.5809, |
|
"slid_loss": 0.5796, |
|
"step": 426, |
|
"time": 301.44 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": "9.8739e-06", |
|
"loss": 0.5766, |
|
"slid_loss": 0.5794, |
|
"step": 427, |
|
"time": 302.51 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": "9.8732e-06", |
|
"loss": 0.5997, |
|
"slid_loss": 0.5795, |
|
"step": 428, |
|
"time": 295.31 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": "9.8726e-06", |
|
"loss": 0.5724, |
|
"slid_loss": 0.5796, |
|
"step": 429, |
|
"time": 304.51 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": "9.8719e-06", |
|
"loss": 0.5924, |
|
"slid_loss": 0.5797, |
|
"step": 430, |
|
"time": 300.1 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": "9.8713e-06", |
|
"loss": 0.611, |
|
"slid_loss": 0.5799, |
|
"step": 431, |
|
"time": 302.85 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": "9.8706e-06", |
|
"loss": 0.5887, |
|
"slid_loss": 0.5799, |
|
"step": 432, |
|
"time": 300.26 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": "9.8700e-06", |
|
"loss": 0.5958, |
|
"slid_loss": 0.5799, |
|
"step": 433, |
|
"time": 297.76 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": "9.8693e-06", |
|
"loss": 0.5765, |
|
"slid_loss": 0.5797, |
|
"step": 434, |
|
"time": 301.87 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": "9.8687e-06", |
|
"loss": 0.5859, |
|
"slid_loss": 0.5796, |
|
"step": 435, |
|
"time": 303.1 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": "9.8680e-06", |
|
"loss": 0.5864, |
|
"slid_loss": 0.5798, |
|
"step": 436, |
|
"time": 301.61 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": "9.8673e-06", |
|
"loss": 0.5822, |
|
"slid_loss": 0.5798, |
|
"step": 437, |
|
"time": 302.85 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": "9.8667e-06", |
|
"loss": 0.5707, |
|
"slid_loss": 0.5799, |
|
"step": 438, |
|
"time": 301.6 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": "9.8660e-06", |
|
"loss": 0.5834, |
|
"slid_loss": 0.5799, |
|
"step": 439, |
|
"time": 303.64 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": "9.8653e-06", |
|
"loss": 0.575, |
|
"slid_loss": 0.5801, |
|
"step": 440, |
|
"time": 300.26 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": "9.8647e-06", |
|
"loss": 0.5684, |
|
"slid_loss": 0.5799, |
|
"step": 441, |
|
"time": 298.04 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": "9.8640e-06", |
|
"loss": 0.5908, |
|
"slid_loss": 0.58, |
|
"step": 442, |
|
"time": 299.82 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": "9.8633e-06", |
|
"loss": 0.6021, |
|
"slid_loss": 0.5802, |
|
"step": 443, |
|
"time": 299.4 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": "9.8627e-06", |
|
"loss": 0.5846, |
|
"slid_loss": 0.5802, |
|
"step": 444, |
|
"time": 302.43 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": "9.8620e-06", |
|
"loss": 0.5923, |
|
"slid_loss": 0.5803, |
|
"step": 445, |
|
"time": 303.88 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": "9.8613e-06", |
|
"loss": 0.5359, |
|
"slid_loss": 0.5799, |
|
"step": 446, |
|
"time": 298.65 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": "9.8607e-06", |
|
"loss": 0.5764, |
|
"slid_loss": 0.5799, |
|
"step": 447, |
|
"time": 300.04 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": "9.8600e-06", |
|
"loss": 0.5913, |
|
"slid_loss": 0.5801, |
|
"step": 448, |
|
"time": 303.88 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": "9.8593e-06", |
|
"loss": 0.5527, |
|
"slid_loss": 0.5801, |
|
"step": 449, |
|
"time": 301.12 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": "9.8586e-06", |
|
"loss": 0.563, |
|
"slid_loss": 0.5797, |
|
"step": 450, |
|
"time": 298.58 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": "9.8579e-06", |
|
"loss": 0.5561, |
|
"slid_loss": 0.5796, |
|
"step": 451, |
|
"time": 299.14 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": "9.8572e-06", |
|
"loss": 0.5594, |
|
"slid_loss": 0.5796, |
|
"step": 452, |
|
"time": 299.44 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": "9.8566e-06", |
|
"loss": 0.5778, |
|
"slid_loss": 0.5796, |
|
"step": 453, |
|
"time": 302.98 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": "9.8559e-06", |
|
"loss": 0.5815, |
|
"slid_loss": 0.5796, |
|
"step": 454, |
|
"time": 298.23 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": "9.8552e-06", |
|
"loss": 0.5589, |
|
"slid_loss": 0.5794, |
|
"step": 455, |
|
"time": 300.7 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": "9.8545e-06", |
|
"loss": 0.5716, |
|
"slid_loss": 0.5792, |
|
"step": 456, |
|
"time": 299.94 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": "9.8538e-06", |
|
"loss": 0.5927, |
|
"slid_loss": 0.5793, |
|
"step": 457, |
|
"time": 302.7 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": "9.8531e-06", |
|
"loss": 0.5675, |
|
"slid_loss": 0.5791, |
|
"step": 458, |
|
"time": 297.41 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": "9.8524e-06", |
|
"loss": 0.5934, |
|
"slid_loss": 0.5792, |
|
"step": 459, |
|
"time": 302.7 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": "9.8517e-06", |
|
"loss": 0.6052, |
|
"slid_loss": 0.5794, |
|
"step": 460, |
|
"time": 300.68 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": "9.8510e-06", |
|
"loss": 0.5644, |
|
"slid_loss": 0.5792, |
|
"step": 461, |
|
"time": 295.54 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": "9.8503e-06", |
|
"loss": 0.562, |
|
"slid_loss": 0.5791, |
|
"step": 462, |
|
"time": 298.32 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": "9.8496e-06", |
|
"loss": 0.5411, |
|
"slid_loss": 0.5787, |
|
"step": 463, |
|
"time": 303.31 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": "9.8489e-06", |
|
"loss": 0.549, |
|
"slid_loss": 0.5783, |
|
"step": 464, |
|
"time": 299.02 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": "9.8482e-06", |
|
"loss": 0.5705, |
|
"slid_loss": 0.5781, |
|
"step": 465, |
|
"time": 302.64 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": "9.8475e-06", |
|
"loss": 0.5853, |
|
"slid_loss": 0.578, |
|
"step": 466, |
|
"time": 301.21 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": "9.8468e-06", |
|
"loss": 0.5954, |
|
"slid_loss": 0.5778, |
|
"step": 467, |
|
"time": 299.8 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": "9.8461e-06", |
|
"loss": 0.6089, |
|
"slid_loss": 0.5783, |
|
"step": 468, |
|
"time": 299.48 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": "9.8454e-06", |
|
"loss": 0.5906, |
|
"slid_loss": 0.5783, |
|
"step": 469, |
|
"time": 295.69 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": "9.8447e-06", |
|
"loss": 0.5781, |
|
"slid_loss": 0.5783, |
|
"step": 470, |
|
"time": 297.51 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": "9.8440e-06", |
|
"loss": 0.5976, |
|
"slid_loss": 0.5787, |
|
"step": 471, |
|
"time": 302.98 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": "9.8432e-06", |
|
"loss": 0.5703, |
|
"slid_loss": 0.5785, |
|
"step": 472, |
|
"time": 300.12 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": "9.8425e-06", |
|
"loss": 0.5825, |
|
"slid_loss": 0.5784, |
|
"step": 473, |
|
"time": 304.01 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": "9.8418e-06", |
|
"loss": 0.5771, |
|
"slid_loss": 0.5784, |
|
"step": 474, |
|
"time": 296.5 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": "9.8411e-06", |
|
"loss": 0.5687, |
|
"slid_loss": 0.5781, |
|
"step": 475, |
|
"time": 300.4 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": "9.8404e-06", |
|
"loss": 0.6119, |
|
"slid_loss": 0.5787, |
|
"step": 476, |
|
"time": 305.33 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": "9.8396e-06", |
|
"loss": 0.5623, |
|
"slid_loss": 0.5783, |
|
"step": 477, |
|
"time": 298.81 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": "9.8389e-06", |
|
"loss": 0.5784, |
|
"slid_loss": 0.5782, |
|
"step": 478, |
|
"time": 299.61 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": "9.8382e-06", |
|
"loss": 0.5713, |
|
"slid_loss": 0.5778, |
|
"step": 479, |
|
"time": 299.37 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": "9.8375e-06", |
|
"loss": 0.5819, |
|
"slid_loss": 0.5778, |
|
"step": 480, |
|
"time": 303.75 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": "9.8367e-06", |
|
"loss": 0.6179, |
|
"slid_loss": 0.5782, |
|
"step": 481, |
|
"time": 301.54 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": "9.8360e-06", |
|
"loss": 0.5672, |
|
"slid_loss": 0.5778, |
|
"step": 482, |
|
"time": 320.09 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": "9.8353e-06", |
|
"loss": 0.5843, |
|
"slid_loss": 0.5778, |
|
"step": 483, |
|
"time": 457.75 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": "9.8345e-06", |
|
"loss": 0.5752, |
|
"slid_loss": 0.5777, |
|
"step": 484, |
|
"time": 431.51 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": "9.8338e-06", |
|
"loss": 0.5523, |
|
"slid_loss": 0.5775, |
|
"step": 485, |
|
"time": 321.57 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": "9.8331e-06", |
|
"loss": 0.5891, |
|
"slid_loss": 0.5776, |
|
"step": 486, |
|
"time": 300.63 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": "9.8323e-06", |
|
"loss": 0.5725, |
|
"slid_loss": 0.5778, |
|
"step": 487, |
|
"time": 303.72 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": "9.8316e-06", |
|
"loss": 0.5792, |
|
"slid_loss": 0.5777, |
|
"step": 488, |
|
"time": 298.59 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": "9.8308e-06", |
|
"loss": 0.5648, |
|
"slid_loss": 0.5776, |
|
"step": 489, |
|
"time": 298.27 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": "9.8301e-06", |
|
"loss": 0.5668, |
|
"slid_loss": 0.5777, |
|
"step": 490, |
|
"time": 305.54 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": "9.8293e-06", |
|
"loss": 0.568, |
|
"slid_loss": 0.5777, |
|
"step": 491, |
|
"time": 303.7 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": "9.8286e-06", |
|
"loss": 0.5719, |
|
"slid_loss": 0.5775, |
|
"step": 492, |
|
"time": 300.89 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": "9.8278e-06", |
|
"loss": 0.6006, |
|
"slid_loss": 0.5776, |
|
"step": 493, |
|
"time": 297.17 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": "9.8271e-06", |
|
"loss": 0.6049, |
|
"slid_loss": 0.5781, |
|
"step": 494, |
|
"time": 302.1 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": "9.8263e-06", |
|
"loss": 0.5823, |
|
"slid_loss": 0.578, |
|
"step": 495, |
|
"time": 304.26 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": "9.8256e-06", |
|
"loss": 0.5863, |
|
"slid_loss": 0.5779, |
|
"step": 496, |
|
"time": 300.98 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": "9.8248e-06", |
|
"loss": 0.5983, |
|
"slid_loss": 0.5778, |
|
"step": 497, |
|
"time": 295.65 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": "9.8241e-06", |
|
"loss": 0.5871, |
|
"slid_loss": 0.578, |
|
"step": 498, |
|
"time": 301.29 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": "9.8233e-06", |
|
"loss": 0.5856, |
|
"slid_loss": 0.5781, |
|
"step": 499, |
|
"time": 298.92 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": "9.8226e-06", |
|
"loss": 0.5778, |
|
"slid_loss": 0.578, |
|
"step": 500, |
|
"time": 303.74 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": "9.8218e-06", |
|
"loss": 0.5678, |
|
"slid_loss": 0.5778, |
|
"step": 501, |
|
"time": 2049.88 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": "9.8210e-06", |
|
"loss": 0.591, |
|
"slid_loss": 0.5782, |
|
"step": 502, |
|
"time": 297.95 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": "9.8203e-06", |
|
"loss": 0.5714, |
|
"slid_loss": 0.578, |
|
"step": 503, |
|
"time": 302.64 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": "9.8195e-06", |
|
"loss": 0.5987, |
|
"slid_loss": 0.5782, |
|
"step": 504, |
|
"time": 301.99 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": "9.8187e-06", |
|
"loss": 0.575, |
|
"slid_loss": 0.5783, |
|
"step": 505, |
|
"time": 297.94 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": "9.8180e-06", |
|
"loss": 0.5822, |
|
"slid_loss": 0.5785, |
|
"step": 506, |
|
"time": 293.53 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": "9.8172e-06", |
|
"loss": 0.5876, |
|
"slid_loss": 0.5788, |
|
"step": 507, |
|
"time": 300.75 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": "9.8164e-06", |
|
"loss": 0.5721, |
|
"slid_loss": 0.5789, |
|
"step": 508, |
|
"time": 303.66 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": "9.8157e-06", |
|
"loss": 0.5543, |
|
"slid_loss": 0.5788, |
|
"step": 509, |
|
"time": 302.79 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": "9.8149e-06", |
|
"loss": 0.5652, |
|
"slid_loss": 0.5787, |
|
"step": 510, |
|
"time": 300.58 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": "9.8141e-06", |
|
"loss": 0.6082, |
|
"slid_loss": 0.579, |
|
"step": 511, |
|
"time": 305.49 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": "9.8133e-06", |
|
"loss": 0.5794, |
|
"slid_loss": 0.5792, |
|
"step": 512, |
|
"time": 301.02 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": "9.8125e-06", |
|
"loss": 0.5701, |
|
"slid_loss": 0.5792, |
|
"step": 513, |
|
"time": 303.23 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": "9.8118e-06", |
|
"loss": 0.5764, |
|
"slid_loss": 0.579, |
|
"step": 514, |
|
"time": 300.22 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": "9.8110e-06", |
|
"loss": 0.5837, |
|
"slid_loss": 0.5791, |
|
"step": 515, |
|
"time": 305.91 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": "9.8102e-06", |
|
"loss": 0.5901, |
|
"slid_loss": 0.5792, |
|
"step": 516, |
|
"time": 301.4 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": "9.8094e-06", |
|
"loss": 0.5621, |
|
"slid_loss": 0.579, |
|
"step": 517, |
|
"time": 296.36 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": "9.8086e-06", |
|
"loss": 0.5756, |
|
"slid_loss": 0.5791, |
|
"step": 518, |
|
"time": 301.91 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": "9.8078e-06", |
|
"loss": 0.5763, |
|
"slid_loss": 0.5792, |
|
"step": 519, |
|
"time": 301.14 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": "9.8070e-06", |
|
"loss": 0.5869, |
|
"slid_loss": 0.5795, |
|
"step": 520, |
|
"time": 296.65 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": "9.8062e-06", |
|
"loss": 0.5763, |
|
"slid_loss": 0.5798, |
|
"step": 521, |
|
"time": 299.31 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": "9.8054e-06", |
|
"loss": 0.5921, |
|
"slid_loss": 0.5798, |
|
"step": 522, |
|
"time": 300.99 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": "9.8047e-06", |
|
"loss": 0.5499, |
|
"slid_loss": 0.5793, |
|
"step": 523, |
|
"time": 300.13 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": "9.8039e-06", |
|
"loss": 0.5583, |
|
"slid_loss": 0.5792, |
|
"step": 524, |
|
"time": 298.62 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": "9.8031e-06", |
|
"loss": 0.5705, |
|
"slid_loss": 0.579, |
|
"step": 525, |
|
"time": 298.48 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": "9.8023e-06", |
|
"loss": 0.618, |
|
"slid_loss": 0.5794, |
|
"step": 526, |
|
"time": 296.88 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": "9.8015e-06", |
|
"loss": 0.5895, |
|
"slid_loss": 0.5795, |
|
"step": 527, |
|
"time": 296.39 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": "9.8007e-06", |
|
"loss": 0.5516, |
|
"slid_loss": 0.579, |
|
"step": 528, |
|
"time": 299.57 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": "9.7998e-06", |
|
"loss": 0.593, |
|
"slid_loss": 0.5792, |
|
"step": 529, |
|
"time": 302.3 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": "9.7990e-06", |
|
"loss": 0.5789, |
|
"slid_loss": 0.5791, |
|
"step": 530, |
|
"time": 300.1 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": "9.7982e-06", |
|
"loss": 0.5832, |
|
"slid_loss": 0.5788, |
|
"step": 531, |
|
"time": 300.39 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": "9.7974e-06", |
|
"loss": 0.5875, |
|
"slid_loss": 0.5788, |
|
"step": 532, |
|
"time": 302.23 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": "9.7966e-06", |
|
"loss": 0.5849, |
|
"slid_loss": 0.5787, |
|
"step": 533, |
|
"time": 304.22 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": "9.7958e-06", |
|
"loss": 0.574, |
|
"slid_loss": 0.5787, |
|
"step": 534, |
|
"time": 296.74 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": "9.7950e-06", |
|
"loss": 0.5671, |
|
"slid_loss": 0.5785, |
|
"step": 535, |
|
"time": 298.14 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": "9.7942e-06", |
|
"loss": 0.5839, |
|
"slid_loss": 0.5785, |
|
"step": 536, |
|
"time": 294.14 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": "9.7933e-06", |
|
"loss": 0.5725, |
|
"slid_loss": 0.5784, |
|
"step": 537, |
|
"time": 299.7 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": "9.7925e-06", |
|
"loss": 0.5664, |
|
"slid_loss": 0.5783, |
|
"step": 538, |
|
"time": 293.94 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": "9.7917e-06", |
|
"loss": 0.581, |
|
"slid_loss": 0.5783, |
|
"step": 539, |
|
"time": 299.01 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": "9.7909e-06", |
|
"loss": 0.6031, |
|
"slid_loss": 0.5786, |
|
"step": 540, |
|
"time": 299.36 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": "9.7901e-06", |
|
"loss": 0.5569, |
|
"slid_loss": 0.5785, |
|
"step": 541, |
|
"time": 302.66 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": "9.7892e-06", |
|
"loss": 0.5837, |
|
"slid_loss": 0.5784, |
|
"step": 542, |
|
"time": 297.96 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": "9.7884e-06", |
|
"loss": 0.5852, |
|
"slid_loss": 0.5782, |
|
"step": 543, |
|
"time": 302.07 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": "9.7876e-06", |
|
"loss": 0.5566, |
|
"slid_loss": 0.5779, |
|
"step": 544, |
|
"time": 299.15 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": "9.7868e-06", |
|
"loss": 0.5821, |
|
"slid_loss": 0.5778, |
|
"step": 545, |
|
"time": 295.31 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": "9.7859e-06", |
|
"loss": 0.5775, |
|
"slid_loss": 0.5783, |
|
"step": 546, |
|
"time": 302.12 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": "9.7851e-06", |
|
"loss": 0.6021, |
|
"slid_loss": 0.5785, |
|
"step": 547, |
|
"time": 298.46 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": "9.7843e-06", |
|
"loss": 0.6014, |
|
"slid_loss": 0.5786, |
|
"step": 548, |
|
"time": 297.01 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": "9.7834e-06", |
|
"loss": 0.5815, |
|
"slid_loss": 0.5789, |
|
"step": 549, |
|
"time": 303.58 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": "9.7826e-06", |
|
"loss": 0.5721, |
|
"slid_loss": 0.579, |
|
"step": 550, |
|
"time": 293.17 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": "9.7817e-06", |
|
"loss": 0.5768, |
|
"slid_loss": 0.5792, |
|
"step": 551, |
|
"time": 297.87 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": "9.7809e-06", |
|
"loss": 0.5622, |
|
"slid_loss": 0.5792, |
|
"step": 552, |
|
"time": 303.11 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": "9.7801e-06", |
|
"loss": 0.5484, |
|
"slid_loss": 0.5789, |
|
"step": 553, |
|
"time": 298.86 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": "9.7792e-06", |
|
"loss": 0.5793, |
|
"slid_loss": 0.5789, |
|
"step": 554, |
|
"time": 300.88 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": "9.7784e-06", |
|
"loss": 0.5794, |
|
"slid_loss": 0.5791, |
|
"step": 555, |
|
"time": 297.44 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": "9.7775e-06", |
|
"loss": 0.5711, |
|
"slid_loss": 0.5791, |
|
"step": 556, |
|
"time": 297.39 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": "9.7767e-06", |
|
"loss": 0.5841, |
|
"slid_loss": 0.579, |
|
"step": 557, |
|
"time": 299.79 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": "9.7758e-06", |
|
"loss": 0.5912, |
|
"slid_loss": 0.5793, |
|
"step": 558, |
|
"time": 298.87 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": "9.7750e-06", |
|
"loss": 0.5774, |
|
"slid_loss": 0.5791, |
|
"step": 559, |
|
"time": 301.18 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": "9.7741e-06", |
|
"loss": 0.5456, |
|
"slid_loss": 0.5785, |
|
"step": 560, |
|
"time": 298.72 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": "9.7733e-06", |
|
"loss": 0.5367, |
|
"slid_loss": 0.5782, |
|
"step": 561, |
|
"time": 299.38 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": "9.7724e-06", |
|
"loss": 0.5657, |
|
"slid_loss": 0.5783, |
|
"step": 562, |
|
"time": 298.86 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": "9.7716e-06", |
|
"loss": 0.5877, |
|
"slid_loss": 0.5787, |
|
"step": 563, |
|
"time": 303.56 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": "9.7707e-06", |
|
"loss": 0.5731, |
|
"slid_loss": 0.579, |
|
"step": 564, |
|
"time": 302.98 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": "9.7698e-06", |
|
"loss": 0.5662, |
|
"slid_loss": 0.5789, |
|
"step": 565, |
|
"time": 299.55 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": "9.7690e-06", |
|
"loss": 0.5598, |
|
"slid_loss": 0.5787, |
|
"step": 566, |
|
"time": 296.33 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": "9.7681e-06", |
|
"loss": 0.5672, |
|
"slid_loss": 0.5784, |
|
"step": 567, |
|
"time": 299.91 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": "9.7672e-06", |
|
"loss": 0.5767, |
|
"slid_loss": 0.5781, |
|
"step": 568, |
|
"time": 300.64 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": "9.7664e-06", |
|
"loss": 0.5659, |
|
"slid_loss": 0.5778, |
|
"step": 569, |
|
"time": 302.66 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": "9.7655e-06", |
|
"loss": 0.5682, |
|
"slid_loss": 0.5777, |
|
"step": 570, |
|
"time": 298.52 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": "9.7646e-06", |
|
"loss": 0.5822, |
|
"slid_loss": 0.5776, |
|
"step": 571, |
|
"time": 302.79 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": "9.7638e-06", |
|
"loss": 0.587, |
|
"slid_loss": 0.5777, |
|
"step": 572, |
|
"time": 303.17 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": "9.7629e-06", |
|
"loss": 0.5786, |
|
"slid_loss": 0.5777, |
|
"step": 573, |
|
"time": 295.3 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": "9.7620e-06", |
|
"loss": 0.6122, |
|
"slid_loss": 0.578, |
|
"step": 574, |
|
"time": 303.38 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": "9.7611e-06", |
|
"loss": 0.5696, |
|
"slid_loss": 0.5781, |
|
"step": 575, |
|
"time": 300.1 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": "9.7603e-06", |
|
"loss": 0.5628, |
|
"slid_loss": 0.5776, |
|
"step": 576, |
|
"time": 300.2 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": "9.7594e-06", |
|
"loss": 0.5719, |
|
"slid_loss": 0.5777, |
|
"step": 577, |
|
"time": 300.67 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": "9.7585e-06", |
|
"loss": 0.592, |
|
"slid_loss": 0.5778, |
|
"step": 578, |
|
"time": 302.63 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": "9.7576e-06", |
|
"loss": 0.5533, |
|
"slid_loss": 0.5776, |
|
"step": 579, |
|
"time": 303.71 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": "9.7567e-06", |
|
"loss": 0.5822, |
|
"slid_loss": 0.5776, |
|
"step": 580, |
|
"time": 301.66 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": "9.7559e-06", |
|
"loss": 0.5546, |
|
"slid_loss": 0.577, |
|
"step": 581, |
|
"time": 299.5 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": "9.7550e-06", |
|
"loss": 0.5803, |
|
"slid_loss": 0.5771, |
|
"step": 582, |
|
"time": 299.14 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": "9.7541e-06", |
|
"loss": 0.5686, |
|
"slid_loss": 0.577, |
|
"step": 583, |
|
"time": 304.85 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": "9.7532e-06", |
|
"loss": 0.5827, |
|
"slid_loss": 0.577, |
|
"step": 584, |
|
"time": 304.42 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": "9.7523e-06", |
|
"loss": 0.5796, |
|
"slid_loss": 0.5773, |
|
"step": 585, |
|
"time": 304.0 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": "9.7514e-06", |
|
"loss": 0.5625, |
|
"slid_loss": 0.577, |
|
"step": 586, |
|
"time": 297.89 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": "9.7505e-06", |
|
"loss": 0.5733, |
|
"slid_loss": 0.5771, |
|
"step": 587, |
|
"time": 298.03 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": "9.7496e-06", |
|
"loss": 0.5815, |
|
"slid_loss": 0.5771, |
|
"step": 588, |
|
"time": 303.5 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": "9.7487e-06", |
|
"loss": 0.5898, |
|
"slid_loss": 0.5773, |
|
"step": 589, |
|
"time": 301.59 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": "9.7478e-06", |
|
"loss": 0.5994, |
|
"slid_loss": 0.5777, |
|
"step": 590, |
|
"time": 303.28 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": "9.7469e-06", |
|
"loss": 0.5832, |
|
"slid_loss": 0.5778, |
|
"step": 591, |
|
"time": 294.78 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": "9.7460e-06", |
|
"loss": 0.5676, |
|
"slid_loss": 0.5778, |
|
"step": 592, |
|
"time": 304.99 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": "9.7451e-06", |
|
"loss": 0.5801, |
|
"slid_loss": 0.5776, |
|
"step": 593, |
|
"time": 298.88 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": "9.7442e-06", |
|
"loss": 0.5912, |
|
"slid_loss": 0.5774, |
|
"step": 594, |
|
"time": 295.82 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": "9.7433e-06", |
|
"loss": 0.5739, |
|
"slid_loss": 0.5773, |
|
"step": 595, |
|
"time": 300.49 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": "9.7424e-06", |
|
"loss": 0.5688, |
|
"slid_loss": 0.5772, |
|
"step": 596, |
|
"time": 294.8 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": "9.7415e-06", |
|
"loss": 0.5834, |
|
"slid_loss": 0.577, |
|
"step": 597, |
|
"time": 303.61 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": "9.7406e-06", |
|
"loss": 0.5988, |
|
"slid_loss": 0.5771, |
|
"step": 598, |
|
"time": 296.42 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": "9.7397e-06", |
|
"loss": 0.578, |
|
"slid_loss": 0.5771, |
|
"step": 599, |
|
"time": 293.08 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": "9.7387e-06", |
|
"loss": 0.5955, |
|
"slid_loss": 0.5772, |
|
"step": 600, |
|
"time": 299.55 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": "9.7378e-06", |
|
"loss": 0.5693, |
|
"slid_loss": 0.5772, |
|
"step": 601, |
|
"time": 2053.43 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": "9.7369e-06", |
|
"loss": 0.5697, |
|
"slid_loss": 0.577, |
|
"step": 602, |
|
"time": 297.5 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": "9.7360e-06", |
|
"loss": 0.5897, |
|
"slid_loss": 0.5772, |
|
"step": 603, |
|
"time": 355.72 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": "9.7351e-06", |
|
"loss": 0.5582, |
|
"slid_loss": 0.5768, |
|
"step": 604, |
|
"time": 543.71 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": "9.7341e-06", |
|
"loss": 0.5588, |
|
"slid_loss": 0.5766, |
|
"step": 605, |
|
"time": 450.24 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": "9.7332e-06", |
|
"loss": 0.5721, |
|
"slid_loss": 0.5765, |
|
"step": 606, |
|
"time": 309.18 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": "9.7323e-06", |
|
"loss": 0.5639, |
|
"slid_loss": 0.5763, |
|
"step": 607, |
|
"time": 300.7 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": "9.7314e-06", |
|
"loss": 0.567, |
|
"slid_loss": 0.5763, |
|
"step": 608, |
|
"time": 301.61 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": "9.7304e-06", |
|
"loss": 0.5601, |
|
"slid_loss": 0.5763, |
|
"step": 609, |
|
"time": 300.66 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": "9.7295e-06", |
|
"loss": 0.5744, |
|
"slid_loss": 0.5764, |
|
"step": 610, |
|
"time": 296.57 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": "9.7286e-06", |
|
"loss": 0.5749, |
|
"slid_loss": 0.5761, |
|
"step": 611, |
|
"time": 294.55 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": "9.7277e-06", |
|
"loss": 0.5752, |
|
"slid_loss": 0.576, |
|
"step": 612, |
|
"time": 296.3 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": "9.7267e-06", |
|
"loss": 0.5614, |
|
"slid_loss": 0.5759, |
|
"step": 613, |
|
"time": 298.5 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": "9.7258e-06", |
|
"loss": 0.5752, |
|
"slid_loss": 0.5759, |
|
"step": 614, |
|
"time": 292.6 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": "9.7249e-06", |
|
"loss": 0.587, |
|
"slid_loss": 0.576, |
|
"step": 615, |
|
"time": 297.26 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": "9.7239e-06", |
|
"loss": 0.5631, |
|
"slid_loss": 0.5757, |
|
"step": 616, |
|
"time": 307.76 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": "9.7230e-06", |
|
"loss": 0.6094, |
|
"slid_loss": 0.5762, |
|
"step": 617, |
|
"time": 300.4 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": "9.7220e-06", |
|
"loss": 0.5581, |
|
"slid_loss": 0.576, |
|
"step": 618, |
|
"time": 298.17 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": "9.7211e-06", |
|
"loss": 0.5693, |
|
"slid_loss": 0.5759, |
|
"step": 619, |
|
"time": 299.92 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": "9.7201e-06", |
|
"loss": 0.5682, |
|
"slid_loss": 0.5757, |
|
"step": 620, |
|
"time": 302.01 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": "9.7192e-06", |
|
"loss": 0.5668, |
|
"slid_loss": 0.5756, |
|
"step": 621, |
|
"time": 302.89 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": "9.7183e-06", |
|
"loss": 0.5831, |
|
"slid_loss": 0.5755, |
|
"step": 622, |
|
"time": 300.0 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": "9.7173e-06", |
|
"loss": 0.5892, |
|
"slid_loss": 0.5759, |
|
"step": 623, |
|
"time": 297.76 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": "9.7164e-06", |
|
"loss": 0.5668, |
|
"slid_loss": 0.576, |
|
"step": 624, |
|
"time": 302.83 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": "9.7154e-06", |
|
"loss": 0.5879, |
|
"slid_loss": 0.5762, |
|
"step": 625, |
|
"time": 308.26 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": "9.7144e-06", |
|
"loss": 0.5769, |
|
"slid_loss": 0.5758, |
|
"step": 626, |
|
"time": 298.8 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": "9.7135e-06", |
|
"loss": 0.5729, |
|
"slid_loss": 0.5756, |
|
"step": 627, |
|
"time": 302.73 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": "9.7125e-06", |
|
"loss": 0.5622, |
|
"slid_loss": 0.5757, |
|
"step": 628, |
|
"time": 304.61 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": "9.7116e-06", |
|
"loss": 0.5882, |
|
"slid_loss": 0.5757, |
|
"step": 629, |
|
"time": 302.35 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": "9.7106e-06", |
|
"loss": 0.5668, |
|
"slid_loss": 0.5756, |
|
"step": 630, |
|
"time": 301.02 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": "9.7097e-06", |
|
"loss": 0.5877, |
|
"slid_loss": 0.5756, |
|
"step": 631, |
|
"time": 300.32 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": "9.7087e-06", |
|
"loss": 0.5718, |
|
"slid_loss": 0.5754, |
|
"step": 632, |
|
"time": 293.92 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": "9.7077e-06", |
|
"loss": 0.5553, |
|
"slid_loss": 0.5752, |
|
"step": 633, |
|
"time": 296.84 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": "9.7068e-06", |
|
"loss": 0.5692, |
|
"slid_loss": 0.5751, |
|
"step": 634, |
|
"time": 303.08 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": "9.7058e-06", |
|
"loss": 0.5597, |
|
"slid_loss": 0.575, |
|
"step": 635, |
|
"time": 304.45 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": "9.7048e-06", |
|
"loss": 0.5771, |
|
"slid_loss": 0.575, |
|
"step": 636, |
|
"time": 300.98 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": "9.7039e-06", |
|
"loss": 0.5856, |
|
"slid_loss": 0.5751, |
|
"step": 637, |
|
"time": 303.39 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": "9.7029e-06", |
|
"loss": 0.5705, |
|
"slid_loss": 0.5751, |
|
"step": 638, |
|
"time": 300.56 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": "9.7019e-06", |
|
"loss": 0.5753, |
|
"slid_loss": 0.5751, |
|
"step": 639, |
|
"time": 302.55 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": "9.7009e-06", |
|
"loss": 0.5651, |
|
"slid_loss": 0.5747, |
|
"step": 640, |
|
"time": 300.37 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": "9.7000e-06", |
|
"loss": 0.5477, |
|
"slid_loss": 0.5746, |
|
"step": 641, |
|
"time": 299.12 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": "9.6990e-06", |
|
"loss": 0.5573, |
|
"slid_loss": 0.5743, |
|
"step": 642, |
|
"time": 302.26 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": "9.6980e-06", |
|
"loss": 0.5694, |
|
"slid_loss": 0.5742, |
|
"step": 643, |
|
"time": 300.38 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": "9.6970e-06", |
|
"loss": 0.596, |
|
"slid_loss": 0.5746, |
|
"step": 644, |
|
"time": 303.72 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": "9.6961e-06", |
|
"loss": 0.568, |
|
"slid_loss": 0.5744, |
|
"step": 645, |
|
"time": 299.1 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": "9.6951e-06", |
|
"loss": 0.5573, |
|
"slid_loss": 0.5742, |
|
"step": 646, |
|
"time": 296.37 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": "9.6941e-06", |
|
"loss": 0.572, |
|
"slid_loss": 0.5739, |
|
"step": 647, |
|
"time": 302.76 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": "9.6931e-06", |
|
"loss": 0.6008, |
|
"slid_loss": 0.5739, |
|
"step": 648, |
|
"time": 304.74 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": "9.6921e-06", |
|
"loss": 0.5735, |
|
"slid_loss": 0.5738, |
|
"step": 649, |
|
"time": 303.11 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": "9.6911e-06", |
|
"loss": 0.5629, |
|
"slid_loss": 0.5738, |
|
"step": 650, |
|
"time": 301.38 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": "9.6901e-06", |
|
"loss": 0.5738, |
|
"slid_loss": 0.5737, |
|
"step": 651, |
|
"time": 303.03 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": "9.6891e-06", |
|
"loss": 0.5789, |
|
"slid_loss": 0.5739, |
|
"step": 652, |
|
"time": 299.84 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": "9.6881e-06", |
|
"loss": 0.5989, |
|
"slid_loss": 0.5744, |
|
"step": 653, |
|
"time": 304.95 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": "9.6872e-06", |
|
"loss": 0.5602, |
|
"slid_loss": 0.5742, |
|
"step": 654, |
|
"time": 305.98 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": "9.6862e-06", |
|
"loss": 0.5695, |
|
"slid_loss": 0.5741, |
|
"step": 655, |
|
"time": 299.64 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": "9.6852e-06", |
|
"loss": 0.5742, |
|
"slid_loss": 0.5741, |
|
"step": 656, |
|
"time": 300.47 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": "9.6842e-06", |
|
"loss": 0.5649, |
|
"slid_loss": 0.5739, |
|
"step": 657, |
|
"time": 299.06 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": "9.6832e-06", |
|
"loss": 0.5553, |
|
"slid_loss": 0.5736, |
|
"step": 658, |
|
"time": 302.71 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": "9.6822e-06", |
|
"loss": 0.5709, |
|
"slid_loss": 0.5735, |
|
"step": 659, |
|
"time": 302.34 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": "9.6812e-06", |
|
"loss": 0.5565, |
|
"slid_loss": 0.5736, |
|
"step": 660, |
|
"time": 298.23 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": "9.6801e-06", |
|
"loss": 0.573, |
|
"slid_loss": 0.574, |
|
"step": 661, |
|
"time": 295.97 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": "9.6791e-06", |
|
"loss": 0.5797, |
|
"slid_loss": 0.5741, |
|
"step": 662, |
|
"time": 299.7 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": "9.6781e-06", |
|
"loss": 0.5464, |
|
"slid_loss": 0.5737, |
|
"step": 663, |
|
"time": 299.82 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": "9.6771e-06", |
|
"loss": 0.5701, |
|
"slid_loss": 0.5737, |
|
"step": 664, |
|
"time": 305.05 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": "9.6761e-06", |
|
"loss": 0.5567, |
|
"slid_loss": 0.5736, |
|
"step": 665, |
|
"time": 299.79 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": "9.6751e-06", |
|
"loss": 0.5588, |
|
"slid_loss": 0.5736, |
|
"step": 666, |
|
"time": 300.74 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": "9.6741e-06", |
|
"loss": 0.5927, |
|
"slid_loss": 0.5738, |
|
"step": 667, |
|
"time": 300.0 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": "9.6731e-06", |
|
"loss": 0.5572, |
|
"slid_loss": 0.5736, |
|
"step": 668, |
|
"time": 301.77 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": "9.6721e-06", |
|
"loss": 0.5461, |
|
"slid_loss": 0.5734, |
|
"step": 669, |
|
"time": 297.96 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": "9.6710e-06", |
|
"loss": 0.564, |
|
"slid_loss": 0.5734, |
|
"step": 670, |
|
"time": 301.97 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": "9.6700e-06", |
|
"loss": 0.5656, |
|
"slid_loss": 0.5732, |
|
"step": 671, |
|
"time": 299.58 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": "9.6690e-06", |
|
"loss": 0.5534, |
|
"slid_loss": 0.5729, |
|
"step": 672, |
|
"time": 296.02 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": "9.6680e-06", |
|
"loss": 0.5292, |
|
"slid_loss": 0.5724, |
|
"step": 673, |
|
"time": 302.8 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": "9.6669e-06", |
|
"loss": 0.5549, |
|
"slid_loss": 0.5718, |
|
"step": 674, |
|
"time": 301.94 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": "9.6659e-06", |
|
"loss": 0.5647, |
|
"slid_loss": 0.5718, |
|
"step": 675, |
|
"time": 300.67 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": "9.6649e-06", |
|
"loss": 0.555, |
|
"slid_loss": 0.5717, |
|
"step": 676, |
|
"time": 296.84 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": "9.6639e-06", |
|
"loss": 0.5872, |
|
"slid_loss": 0.5719, |
|
"step": 677, |
|
"time": 303.18 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": "9.6628e-06", |
|
"loss": 0.5538, |
|
"slid_loss": 0.5715, |
|
"step": 678, |
|
"time": 303.01 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": "9.6618e-06", |
|
"loss": 0.5584, |
|
"slid_loss": 0.5715, |
|
"step": 679, |
|
"time": 302.31 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": "9.6608e-06", |
|
"loss": 0.5835, |
|
"slid_loss": 0.5715, |
|
"step": 680, |
|
"time": 302.48 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": "9.6597e-06", |
|
"loss": 0.5938, |
|
"slid_loss": 0.5719, |
|
"step": 681, |
|
"time": 298.4 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": "9.6587e-06", |
|
"loss": 0.5878, |
|
"slid_loss": 0.572, |
|
"step": 682, |
|
"time": 300.58 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": "9.6577e-06", |
|
"loss": 0.5725, |
|
"slid_loss": 0.572, |
|
"step": 683, |
|
"time": 302.24 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": "9.6566e-06", |
|
"loss": 0.5648, |
|
"slid_loss": 0.5719, |
|
"step": 684, |
|
"time": 301.51 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": "9.6556e-06", |
|
"loss": 0.5772, |
|
"slid_loss": 0.5718, |
|
"step": 685, |
|
"time": 298.21 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": "9.6545e-06", |
|
"loss": 0.5852, |
|
"slid_loss": 0.5721, |
|
"step": 686, |
|
"time": 302.79 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": "9.6535e-06", |
|
"loss": 0.5793, |
|
"slid_loss": 0.5721, |
|
"step": 687, |
|
"time": 299.38 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": "9.6525e-06", |
|
"loss": 0.597, |
|
"slid_loss": 0.5723, |
|
"step": 688, |
|
"time": 304.05 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": "9.6514e-06", |
|
"loss": 0.5599, |
|
"slid_loss": 0.572, |
|
"step": 689, |
|
"time": 297.95 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": "9.6504e-06", |
|
"loss": 0.5822, |
|
"slid_loss": 0.5718, |
|
"step": 690, |
|
"time": 304.74 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": "9.6493e-06", |
|
"loss": 0.5838, |
|
"slid_loss": 0.5718, |
|
"step": 691, |
|
"time": 302.75 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": "9.6483e-06", |
|
"loss": 0.5442, |
|
"slid_loss": 0.5716, |
|
"step": 692, |
|
"time": 300.4 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": "9.6472e-06", |
|
"loss": 0.575, |
|
"slid_loss": 0.5715, |
|
"step": 693, |
|
"time": 302.57 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": "9.6462e-06", |
|
"loss": 0.5481, |
|
"slid_loss": 0.5711, |
|
"step": 694, |
|
"time": 302.08 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": "9.6451e-06", |
|
"loss": 0.5555, |
|
"slid_loss": 0.5709, |
|
"step": 695, |
|
"time": 296.83 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": "9.6440e-06", |
|
"loss": 0.5521, |
|
"slid_loss": 0.5708, |
|
"step": 696, |
|
"time": 305.62 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": "9.6430e-06", |
|
"loss": 0.5705, |
|
"slid_loss": 0.5706, |
|
"step": 697, |
|
"time": 297.63 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": "9.6419e-06", |
|
"loss": 0.5724, |
|
"slid_loss": 0.5704, |
|
"step": 698, |
|
"time": 296.55 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": "9.6409e-06", |
|
"loss": 0.5564, |
|
"slid_loss": 0.5701, |
|
"step": 699, |
|
"time": 301.49 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": "9.6398e-06", |
|
"loss": 0.5864, |
|
"slid_loss": 0.5701, |
|
"step": 700, |
|
"time": 306.47 |
|
} |
|
], |
|
"logging_steps": 1.0, |
|
"max_steps": 3860, |
|
"num_train_epochs": 5, |
|
"save_steps": 50000.0, |
|
"total_flos": 0.0, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|