|
{ |
|
"best_metric": 1.2486178874969482, |
|
"best_model_checkpoint": "data/qwen_1_8B_llamafied/checkpoint-2000", |
|
"epoch": 3.0, |
|
"eval_steps": 100, |
|
"global_step": 2997, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.6666666666666668e-07, |
|
"loss": 2.8016, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 8.333333333333333e-07, |
|
"loss": 2.7718, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.6666666666666667e-06, |
|
"loss": 2.7343, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 2.5e-06, |
|
"loss": 2.5458, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.3333333333333333e-06, |
|
"loss": 2.0935, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.166666666666667e-06, |
|
"loss": 1.8378, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 5e-06, |
|
"loss": 1.7056, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 5.833333333333334e-06, |
|
"loss": 1.6188, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 6.666666666666667e-06, |
|
"loss": 1.5864, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 7.500000000000001e-06, |
|
"loss": 1.5367, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.333333333333334e-06, |
|
"loss": 1.5154, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.166666666666666e-06, |
|
"loss": 1.5062, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1e-05, |
|
"loss": 1.4631, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.0833333333333334e-05, |
|
"loss": 1.4677, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.1666666666666668e-05, |
|
"loss": 1.4489, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.25e-05, |
|
"loss": 1.4412, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.3333333333333333e-05, |
|
"loss": 1.4387, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.416666666666667e-05, |
|
"loss": 1.3953, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.5000000000000002e-05, |
|
"loss": 1.3908, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.5833333333333333e-05, |
|
"loss": 1.4252, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.6666666666666667e-05, |
|
"loss": 1.3881, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"eval_loss": 1.3936688899993896, |
|
"eval_runtime": 187.5396, |
|
"eval_samples_per_second": 75.451, |
|
"eval_steps_per_second": 2.362, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.7500000000000002e-05, |
|
"loss": 1.3971, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.8333333333333333e-05, |
|
"loss": 1.3798, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.916666666666667e-05, |
|
"loss": 1.3629, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 2e-05, |
|
"loss": 1.3734, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.9999850951038044e-05, |
|
"loss": 1.3569, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.999940380859529e-05, |
|
"loss": 1.3637, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.9998658586000968e-05, |
|
"loss": 1.3416, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.9997615305470002e-05, |
|
"loss": 1.3236, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.9996273998102368e-05, |
|
"loss": 1.3579, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.9994634703882166e-05, |
|
"loss": 1.3346, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.9992697471676413e-05, |
|
"loss": 1.3171, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.9990462359233595e-05, |
|
"loss": 1.3493, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.998792943318196e-05, |
|
"loss": 1.33, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.99850987690275e-05, |
|
"loss": 1.3414, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.9981970451151724e-05, |
|
"loss": 1.3217, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.9978544572809143e-05, |
|
"loss": 1.3345, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.997482123612448e-05, |
|
"loss": 1.337, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.9970800552089623e-05, |
|
"loss": 1.3474, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.9966482640560335e-05, |
|
"loss": 1.329, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.996186763025266e-05, |
|
"loss": 1.3499, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"eval_loss": 1.337212324142456, |
|
"eval_runtime": 187.5232, |
|
"eval_samples_per_second": 75.457, |
|
"eval_steps_per_second": 2.362, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.9956955658739098e-05, |
|
"loss": 1.3359, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.995174687244449e-05, |
|
"loss": 1.3417, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.994624142664169e-05, |
|
"loss": 1.3162, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.9940439485446884e-05, |
|
"loss": 1.3135, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.993434122181474e-05, |
|
"loss": 1.3419, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.9927946817533224e-05, |
|
"loss": 1.324, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.9921256463218207e-05, |
|
"loss": 1.3066, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.9914270358307762e-05, |
|
"loss": 1.321, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.9906988711056215e-05, |
|
"loss": 1.3189, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.9899411738527977e-05, |
|
"loss": 1.3164, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.989153966659101e-05, |
|
"loss": 1.3075, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.9883372729910154e-05, |
|
"loss": 1.3072, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.9874911171940087e-05, |
|
"loss": 1.33, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.9866155244918106e-05, |
|
"loss": 1.3254, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.985710520985657e-05, |
|
"loss": 1.3208, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.9847761336535148e-05, |
|
"loss": 1.3101, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.9838123903492764e-05, |
|
"loss": 1.2976, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.9828193198019297e-05, |
|
"loss": 1.3126, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.981796951614701e-05, |
|
"loss": 1.3094, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.9807453162641745e-05, |
|
"loss": 1.3138, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"eval_loss": 1.3167798519134521, |
|
"eval_runtime": 187.2287, |
|
"eval_samples_per_second": 75.576, |
|
"eval_steps_per_second": 2.366, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.9796644450993812e-05, |
|
"loss": 1.2995, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.978554370340866e-05, |
|
"loss": 1.314, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.9774151250797278e-05, |
|
"loss": 1.3188, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.9762467432766303e-05, |
|
"loss": 1.314, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.975049259760793e-05, |
|
"loss": 1.2998, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.973822710228951e-05, |
|
"loss": 1.2831, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.972567131244291e-05, |
|
"loss": 1.3072, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.9712825602353617e-05, |
|
"loss": 1.3056, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.969969035494958e-05, |
|
"loss": 1.2919, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.9686265961789807e-05, |
|
"loss": 1.293, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.9672552823052663e-05, |
|
"loss": 1.3041, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.965855134752397e-05, |
|
"loss": 1.2889, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.9644261952584804e-05, |
|
"loss": 1.3081, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.962968506419906e-05, |
|
"loss": 1.3317, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.9614821116900763e-05, |
|
"loss": 1.2802, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.9599670553781087e-05, |
|
"loss": 1.2958, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.9584233826475178e-05, |
|
"loss": 1.2925, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.9568511395148674e-05, |
|
"loss": 1.2985, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.955250372848398e-05, |
|
"loss": 1.2998, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.9536211303666324e-05, |
|
"loss": 1.3152, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"eval_loss": 1.3044512271881104, |
|
"eval_runtime": 187.4307, |
|
"eval_samples_per_second": 75.495, |
|
"eval_steps_per_second": 2.364, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.9519634606369506e-05, |
|
"loss": 1.289, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.9502774130741427e-05, |
|
"loss": 1.3092, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.948563037938937e-05, |
|
"loss": 1.2971, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.9468203863365002e-05, |
|
"loss": 1.2988, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.9450495102149144e-05, |
|
"loss": 1.2971, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.94325046236363e-05, |
|
"loss": 1.287, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.9414232964118893e-05, |
|
"loss": 1.3097, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.93956806682713e-05, |
|
"loss": 1.3012, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.9376848289133607e-05, |
|
"loss": 1.2917, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.9357736388095132e-05, |
|
"loss": 1.3051, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.9338345534877675e-05, |
|
"loss": 1.2817, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.9318676307518543e-05, |
|
"loss": 1.2784, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.9298729292353318e-05, |
|
"loss": 1.3045, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.9278505083998393e-05, |
|
"loss": 1.2936, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.9258004285333204e-05, |
|
"loss": 1.2808, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.923722750748231e-05, |
|
"loss": 1.2869, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.9216175369797145e-05, |
|
"loss": 1.2843, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.9194848499837567e-05, |
|
"loss": 1.2776, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.9173247533353137e-05, |
|
"loss": 1.2914, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.9151373114264183e-05, |
|
"loss": 1.2897, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"eval_loss": 1.2954026460647583, |
|
"eval_runtime": 187.4576, |
|
"eval_samples_per_second": 75.484, |
|
"eval_steps_per_second": 2.363, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.9129225894642594e-05, |
|
"loss": 1.2949, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.9106806534692392e-05, |
|
"loss": 1.3079, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.9084115702730043e-05, |
|
"loss": 1.2716, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.9061154075164535e-05, |
|
"loss": 1.2734, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 1.903792233647722e-05, |
|
"loss": 1.2873, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 1.9014421179201407e-05, |
|
"loss": 1.2895, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 1.899065130390171e-05, |
|
"loss": 1.2875, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 1.896661341915318e-05, |
|
"loss": 1.2865, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.8942308241520174e-05, |
|
"loss": 1.2938, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.8917736495534988e-05, |
|
"loss": 1.2705, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.8892898913676264e-05, |
|
"loss": 1.2814, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.8867796236347168e-05, |
|
"loss": 1.2905, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 1.8842429211853298e-05, |
|
"loss": 1.2844, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 1.8816798596380388e-05, |
|
"loss": 1.2799, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.879090515397176e-05, |
|
"loss": 1.2958, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.876474965650556e-05, |
|
"loss": 1.2946, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.8738332883671738e-05, |
|
"loss": 1.2864, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.8711655622948805e-05, |
|
"loss": 1.2983, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 1.8684718669580368e-05, |
|
"loss": 1.284, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 1.8657522826551412e-05, |
|
"loss": 1.28, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"eval_loss": 1.2882355451583862, |
|
"eval_runtime": 187.4557, |
|
"eval_samples_per_second": 75.485, |
|
"eval_steps_per_second": 2.363, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.8630068904564373e-05, |
|
"loss": 1.2817, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.8602357722014966e-05, |
|
"loss": 1.3014, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.8574390104967784e-05, |
|
"loss": 1.2887, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.85461668871317e-05, |
|
"loss": 1.2768, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 1.851768890983496e-05, |
|
"loss": 1.2828, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 1.8488957022000164e-05, |
|
"loss": 1.2506, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.8459972080118925e-05, |
|
"loss": 1.2773, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.8430734948226336e-05, |
|
"loss": 1.2714, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.8401246497875238e-05, |
|
"loss": 1.2715, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.8371507608110206e-05, |
|
"loss": 1.2931, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 1.8341519165441373e-05, |
|
"loss": 1.2845, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 1.831128206381799e-05, |
|
"loss": 1.2625, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.8280797204601782e-05, |
|
"loss": 1.2622, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.8250065496540062e-05, |
|
"loss": 1.2719, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.8219087855738678e-05, |
|
"loss": 1.2722, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.818786520563467e-05, |
|
"loss": 1.2766, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.815639847696875e-05, |
|
"loss": 1.2765, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.812468860775757e-05, |
|
"loss": 1.2888, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.809273654326575e-05, |
|
"loss": 1.2772, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.8060543235977696e-05, |
|
"loss": 1.2669, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"eval_loss": 1.2820061445236206, |
|
"eval_runtime": 187.8058, |
|
"eval_samples_per_second": 75.344, |
|
"eval_steps_per_second": 2.359, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.8028109645569212e-05, |
|
"loss": 1.2602, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.79954367388789e-05, |
|
"loss": 1.2583, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.7962525489879324e-05, |
|
"loss": 1.2693, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.7929376879647984e-05, |
|
"loss": 1.2679, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.7895991896338066e-05, |
|
"loss": 1.2588, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.7862371535149e-05, |
|
"loss": 1.2836, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.782851679829676e-05, |
|
"loss": 1.2691, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.779442869498403e-05, |
|
"loss": 1.2778, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.7760108241370097e-05, |
|
"loss": 1.2715, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.7725556460540553e-05, |
|
"loss": 1.2563, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.7690774382476808e-05, |
|
"loss": 1.2825, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.7655763044025395e-05, |
|
"loss": 1.2882, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.7620523488867042e-05, |
|
"loss": 1.2704, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.7585056767485574e-05, |
|
"loss": 1.2585, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.7549363937136586e-05, |
|
"loss": 1.272, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.7513446061815946e-05, |
|
"loss": 1.2668, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.747730421222806e-05, |
|
"loss": 1.2811, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.744093946575396e-05, |
|
"loss": 1.2629, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.7404352906419187e-05, |
|
"loss": 1.2508, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.7367545624861483e-05, |
|
"loss": 1.2591, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"eval_loss": 1.2767587900161743, |
|
"eval_runtime": 187.4947, |
|
"eval_samples_per_second": 75.469, |
|
"eval_steps_per_second": 2.363, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.7330518718298263e-05, |
|
"loss": 1.276, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.7293273290493932e-05, |
|
"loss": 1.2746, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.725581045172696e-05, |
|
"loss": 1.2693, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.721813131875679e-05, |
|
"loss": 1.2668, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.718023701479055e-05, |
|
"loss": 1.269, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.714212866944958e-05, |
|
"loss": 1.2699, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.710380741873574e-05, |
|
"loss": 1.2695, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.706527440499756e-05, |
|
"loss": 1.2731, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.702653077689618e-05, |
|
"loss": 1.2701, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.6987577689371107e-05, |
|
"loss": 1.2877, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.6948416303605796e-05, |
|
"loss": 1.2306, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.690904778699302e-05, |
|
"loss": 1.2618, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.68694733131001e-05, |
|
"loss": 1.2432, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.6829694061633874e-05, |
|
"loss": 1.2709, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.6789711218405575e-05, |
|
"loss": 1.264, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.6749525975295455e-05, |
|
"loss": 1.2534, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.670913953021727e-05, |
|
"loss": 1.2451, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.666855308708257e-05, |
|
"loss": 1.2583, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.6627767855764786e-05, |
|
"loss": 1.2787, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.6586785052063206e-05, |
|
"loss": 1.2447, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"eval_loss": 1.2721099853515625, |
|
"eval_runtime": 187.5451, |
|
"eval_samples_per_second": 75.449, |
|
"eval_steps_per_second": 2.362, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.65456058976667e-05, |
|
"loss": 1.2618, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.6504231620117305e-05, |
|
"loss": 1.2708, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.646266345277365e-05, |
|
"loss": 1.238, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.642090263477417e-05, |
|
"loss": 1.2572, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.6378950411000183e-05, |
|
"loss": 1.2712, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.6336808032038768e-05, |
|
"loss": 1.2915, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.6294476754145497e-05, |
|
"loss": 1.2624, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.625195783920697e-05, |
|
"loss": 1.2527, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.6209252554703215e-05, |
|
"loss": 1.273, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.61663621736699e-05, |
|
"loss": 1.264, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.6123287974660378e-05, |
|
"loss": 1.2745, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.608003124170758e-05, |
|
"loss": 1.2594, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.603659326428573e-05, |
|
"loss": 1.2692, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.599297533727192e-05, |
|
"loss": 1.2635, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.59491787609075e-05, |
|
"loss": 1.2618, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.590520484075932e-05, |
|
"loss": 1.2594, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.5861054887680805e-05, |
|
"loss": 1.2728, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.581673021777289e-05, |
|
"loss": 1.2531, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 1.5772232152344797e-05, |
|
"loss": 1.2677, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 1.5727562017874602e-05, |
|
"loss": 1.2867, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 1.2679548263549805, |
|
"eval_runtime": 187.1108, |
|
"eval_samples_per_second": 75.624, |
|
"eval_steps_per_second": 2.368, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 1.5682721145969742e-05, |
|
"loss": 1.1971, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 1.5637710873327305e-05, |
|
"loss": 1.2015, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 1.5592532541694173e-05, |
|
"loss": 1.1858, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 1.554718749782703e-05, |
|
"loss": 1.1876, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 1.5501677093452218e-05, |
|
"loss": 1.2026, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 1.5456002685225447e-05, |
|
"loss": 1.2048, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 1.5410165634691343e-05, |
|
"loss": 1.1949, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 1.5364167308242866e-05, |
|
"loss": 1.19, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 1.5318009077080577e-05, |
|
"loss": 1.1918, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 1.5271692317171765e-05, |
|
"loss": 1.2024, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 1.5225218409209424e-05, |
|
"loss": 1.2118, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 1.5178588738571108e-05, |
|
"loss": 1.1784, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.5131804695277612e-05, |
|
"loss": 1.1648, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.5084867673951558e-05, |
|
"loss": 1.1834, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.5037779073775805e-05, |
|
"loss": 1.197, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.4990540298451748e-05, |
|
"loss": 1.2048, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.4943152756157476e-05, |
|
"loss": 1.1843, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.4895617859505782e-05, |
|
"loss": 1.1959, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.4847937025502067e-05, |
|
"loss": 1.2008, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.4800111675502095e-05, |
|
"loss": 1.1918, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"eval_loss": 1.2683547735214233, |
|
"eval_runtime": 187.3811, |
|
"eval_samples_per_second": 75.515, |
|
"eval_steps_per_second": 2.364, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.4752143235169623e-05, |
|
"loss": 1.2103, |
|
"step": 1105 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.4704033134433895e-05, |
|
"loss": 1.1977, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.4655782807447025e-05, |
|
"loss": 1.2042, |
|
"step": 1115 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.460739369254125e-05, |
|
"loss": 1.2113, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.4558867232186034e-05, |
|
"loss": 1.2063, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.4510204872945084e-05, |
|
"loss": 1.209, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.446140806543323e-05, |
|
"loss": 1.1723, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.4412478264273173e-05, |
|
"loss": 1.1861, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.4363416928052126e-05, |
|
"loss": 1.1916, |
|
"step": 1145 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.4314225519278333e-05, |
|
"loss": 1.1912, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.4264905504337483e-05, |
|
"loss": 1.1996, |
|
"step": 1155 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.4215458353448981e-05, |
|
"loss": 1.1973, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 1.4165885540622128e-05, |
|
"loss": 1.1967, |
|
"step": 1165 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 1.4116188543612182e-05, |
|
"loss": 1.2018, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 1.4066368843876305e-05, |
|
"loss": 1.1946, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 1.4016427926529406e-05, |
|
"loss": 1.1964, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 1.3966367280299865e-05, |
|
"loss": 1.2016, |
|
"step": 1185 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 1.3916188397485147e-05, |
|
"loss": 1.198, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 1.3865892773907335e-05, |
|
"loss": 1.2003, |
|
"step": 1195 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 1.3815481908868524e-05, |
|
"loss": 1.2002, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"eval_loss": 1.2659939527511597, |
|
"eval_runtime": 187.4768, |
|
"eval_samples_per_second": 75.476, |
|
"eval_steps_per_second": 2.363, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 1.3764957305106142e-05, |
|
"loss": 1.2018, |
|
"step": 1205 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 1.3714320468748126e-05, |
|
"loss": 1.1792, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 1.3663572909268064e-05, |
|
"loss": 1.2, |
|
"step": 1215 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 1.3612716139440164e-05, |
|
"loss": 1.1935, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 1.3561751675294182e-05, |
|
"loss": 1.1894, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 1.3510681036070206e-05, |
|
"loss": 1.2167, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.3459505744173392e-05, |
|
"loss": 1.2214, |
|
"step": 1235 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.3408227325128569e-05, |
|
"loss": 1.1916, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.3356847307534759e-05, |
|
"loss": 1.1956, |
|
"step": 1245 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.3305367223019622e-05, |
|
"loss": 1.1871, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 1.3253788606193791e-05, |
|
"loss": 1.1828, |
|
"step": 1255 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 1.3202112994605128e-05, |
|
"loss": 1.1989, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 1.3150341928692875e-05, |
|
"loss": 1.2084, |
|
"step": 1265 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 1.309847695174177e-05, |
|
"loss": 1.2011, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 1.3046519609836002e-05, |
|
"loss": 1.1951, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 1.2994471451813149e-05, |
|
"loss": 1.2113, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 1.2942334029217993e-05, |
|
"loss": 1.1971, |
|
"step": 1285 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 1.2890108896256282e-05, |
|
"loss": 1.1994, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 1.2837797609748384e-05, |
|
"loss": 1.2019, |
|
"step": 1295 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 1.2785401729082893e-05, |
|
"loss": 1.1943, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"eval_loss": 1.2633405923843384, |
|
"eval_runtime": 187.4908, |
|
"eval_samples_per_second": 75.47, |
|
"eval_steps_per_second": 2.363, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 1.273292281617013e-05, |
|
"loss": 1.2003, |
|
"step": 1305 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 1.2680362435395595e-05, |
|
"loss": 1.1832, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 1.2627722153573326e-05, |
|
"loss": 1.2033, |
|
"step": 1315 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 1.2575003539899198e-05, |
|
"loss": 1.1892, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 1.2522208165904137e-05, |
|
"loss": 1.1815, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 1.2469337605407284e-05, |
|
"loss": 1.1735, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 1.2416393434469068e-05, |
|
"loss": 1.1798, |
|
"step": 1335 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 1.2363377231344239e-05, |
|
"loss": 1.2016, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 1.2310290576434795e-05, |
|
"loss": 1.1974, |
|
"step": 1345 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 1.2257135052242906e-05, |
|
"loss": 1.1884, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 1.220391224332371e-05, |
|
"loss": 1.1969, |
|
"step": 1355 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 1.215062373623809e-05, |
|
"loss": 1.2057, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 1.2097271119505382e-05, |
|
"loss": 1.1895, |
|
"step": 1365 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 1.2043855983556014e-05, |
|
"loss": 1.1997, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 1.1990379920684097e-05, |
|
"loss": 1.1946, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 1.1936844524999966e-05, |
|
"loss": 1.1937, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 1.1883251392382649e-05, |
|
"loss": 1.2074, |
|
"step": 1385 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 1.1829602120432305e-05, |
|
"loss": 1.183, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 1.177589830842259e-05, |
|
"loss": 1.1866, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 1.1722141557252998e-05, |
|
"loss": 1.199, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"eval_loss": 1.2606513500213623, |
|
"eval_runtime": 187.2567, |
|
"eval_samples_per_second": 75.565, |
|
"eval_steps_per_second": 2.366, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 1.1668333469401109e-05, |
|
"loss": 1.1867, |
|
"step": 1405 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 1.1614475648874861e-05, |
|
"loss": 1.1821, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 1.1560569701164696e-05, |
|
"loss": 1.1928, |
|
"step": 1415 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 1.1506617233195732e-05, |
|
"loss": 1.1916, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 1.145261985327983e-05, |
|
"loss": 1.2152, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 1.1398579171067679e-05, |
|
"loss": 1.2126, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 1.1344496797500795e-05, |
|
"loss": 1.2264, |
|
"step": 1435 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 1.1290374344763506e-05, |
|
"loss": 1.1955, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 1.1236213426234894e-05, |
|
"loss": 1.1955, |
|
"step": 1445 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 1.1182015656440692e-05, |
|
"loss": 1.1947, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 1.112778265100517e-05, |
|
"loss": 1.1963, |
|
"step": 1455 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 1.1073516026602959e-05, |
|
"loss": 1.1984, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 1.1019217400910864e-05, |
|
"loss": 1.1858, |
|
"step": 1465 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 1.096488839255965e-05, |
|
"loss": 1.1998, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 1.0910530621085769e-05, |
|
"loss": 1.2096, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 1.085614570688311e-05, |
|
"loss": 1.1876, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 1.080173527115467e-05, |
|
"loss": 1.1893, |
|
"step": 1485 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 1.0747300935864245e-05, |
|
"loss": 1.1856, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 1.0692844323688065e-05, |
|
"loss": 1.1836, |
|
"step": 1495 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 1.0638367057966442e-05, |
|
"loss": 1.1887, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"eval_loss": 1.2581257820129395, |
|
"eval_runtime": 187.5926, |
|
"eval_samples_per_second": 75.429, |
|
"eval_steps_per_second": 2.362, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 1.0583870762655347e-05, |
|
"loss": 1.1821, |
|
"step": 1505 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 1.0529357062278033e-05, |
|
"loss": 1.1622, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 1.0474827581876594e-05, |
|
"loss": 1.1928, |
|
"step": 1515 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 1.042028394696352e-05, |
|
"loss": 1.1972, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 1.0365727783473238e-05, |
|
"loss": 1.196, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 1.031116071771366e-05, |
|
"loss": 1.1925, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 1.0256584376317686e-05, |
|
"loss": 1.2037, |
|
"step": 1535 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 1.0202000386194726e-05, |
|
"loss": 1.1895, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 1.0147410374482188e-05, |
|
"loss": 1.2098, |
|
"step": 1545 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 1.009281596849699e-05, |
|
"loss": 1.1774, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 1.003821879568704e-05, |
|
"loss": 1.1939, |
|
"step": 1555 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 9.983620483582726e-06, |
|
"loss": 1.2046, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 9.929022659748395e-06, |
|
"loss": 1.201, |
|
"step": 1565 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 9.874426951733844e-06, |
|
"loss": 1.1794, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 9.81983498702579e-06, |
|
"loss": 1.1879, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 9.76524839299937e-06, |
|
"loss": 1.1795, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 9.710668796869616e-06, |
|
"loss": 1.2116, |
|
"step": 1585 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 9.65609782564296e-06, |
|
"loss": 1.1764, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 9.601537106068727e-06, |
|
"loss": 1.167, |
|
"step": 1595 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 9.546988264590637e-06, |
|
"loss": 1.1987, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"eval_loss": 1.255649447441101, |
|
"eval_runtime": 187.4393, |
|
"eval_samples_per_second": 75.491, |
|
"eval_steps_per_second": 2.363, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 9.49245292729833e-06, |
|
"loss": 1.1838, |
|
"step": 1605 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 9.437932719878888e-06, |
|
"loss": 1.2057, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 9.383429267568379e-06, |
|
"loss": 1.193, |
|
"step": 1615 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 9.328944195103397e-06, |
|
"loss": 1.1937, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 9.27447912667264e-06, |
|
"loss": 1.1815, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 9.220035685868495e-06, |
|
"loss": 1.1862, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 9.165615495638627e-06, |
|
"loss": 1.1945, |
|
"step": 1635 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 9.111220178237607e-06, |
|
"loss": 1.1938, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 9.056851355178556e-06, |
|
"loss": 1.2155, |
|
"step": 1645 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 9.002510647184803e-06, |
|
"loss": 1.1971, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 8.94819967414157e-06, |
|
"loss": 1.1858, |
|
"step": 1655 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 8.893920055047686e-06, |
|
"loss": 1.2008, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 8.83967340796733e-06, |
|
"loss": 1.1987, |
|
"step": 1665 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 8.785461349981789e-06, |
|
"loss": 1.1958, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 8.731285497141258e-06, |
|
"loss": 1.1798, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 8.67714746441666e-06, |
|
"loss": 1.1942, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 8.62304886565151e-06, |
|
"loss": 1.1841, |
|
"step": 1685 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 8.56899131351381e-06, |
|
"loss": 1.1846, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 8.514976419447963e-06, |
|
"loss": 1.1902, |
|
"step": 1695 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 8.461005793626752e-06, |
|
"loss": 1.1954, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"eval_loss": 1.2533518075942993, |
|
"eval_runtime": 187.5718, |
|
"eval_samples_per_second": 75.438, |
|
"eval_steps_per_second": 2.362, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 8.407081044903323e-06, |
|
"loss": 1.1936, |
|
"step": 1705 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 8.353203780763239e-06, |
|
"loss": 1.1766, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 8.299375607276564e-06, |
|
"loss": 1.1717, |
|
"step": 1715 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 8.245598129049973e-06, |
|
"loss": 1.1983, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 8.191872949178926e-06, |
|
"loss": 1.176, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 8.138201669199878e-06, |
|
"loss": 1.1887, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 8.084585889042547e-06, |
|
"loss": 1.1819, |
|
"step": 1735 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 8.031027206982211e-06, |
|
"loss": 1.1958, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 7.977527219592057e-06, |
|
"loss": 1.1914, |
|
"step": 1745 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 7.924087521695602e-06, |
|
"loss": 1.1871, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 7.87070970631915e-06, |
|
"loss": 1.1762, |
|
"step": 1755 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 7.817395364644294e-06, |
|
"loss": 1.1991, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 7.76414608596049e-06, |
|
"loss": 1.1903, |
|
"step": 1765 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 7.710963457617681e-06, |
|
"loss": 1.1848, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 7.657849064978976e-06, |
|
"loss": 1.1776, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 7.604804491373396e-06, |
|
"loss": 1.1702, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 7.551831318048663e-06, |
|
"loss": 1.2081, |
|
"step": 1785 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 7.498931124124077e-06, |
|
"loss": 1.1857, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 7.446105486543438e-06, |
|
"loss": 1.1817, |
|
"step": 1795 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 7.393355980028039e-06, |
|
"loss": 1.1869, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"eval_loss": 1.251145362854004, |
|
"eval_runtime": 187.49, |
|
"eval_samples_per_second": 75.471, |
|
"eval_steps_per_second": 2.363, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 7.340684177029709e-06, |
|
"loss": 1.2018, |
|
"step": 1805 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 7.288091647683966e-06, |
|
"loss": 1.187, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 7.235579959763189e-06, |
|
"loss": 1.1738, |
|
"step": 1815 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 7.183150678629898e-06, |
|
"loss": 1.1771, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 7.1308053671900675e-06, |
|
"loss": 1.1917, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 7.078545585846578e-06, |
|
"loss": 1.1803, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 7.026372892452653e-06, |
|
"loss": 1.1866, |
|
"step": 1835 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 6.974288842265455e-06, |
|
"loss": 1.2103, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 6.922294987899705e-06, |
|
"loss": 1.1954, |
|
"step": 1845 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 6.870392879281407e-06, |
|
"loss": 1.1907, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 6.818584063601646e-06, |
|
"loss": 1.1741, |
|
"step": 1855 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 6.766870085270459e-06, |
|
"loss": 1.1833, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 6.715252485870804e-06, |
|
"loss": 1.166, |
|
"step": 1865 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 6.663732804112604e-06, |
|
"loss": 1.1901, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 6.6123125757868764e-06, |
|
"loss": 1.1776, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 6.5609933337199515e-06, |
|
"loss": 1.1661, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 6.50977660772778e-06, |
|
"loss": 1.1917, |
|
"step": 1885 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 6.458663924570332e-06, |
|
"loss": 1.1842, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 6.4076568079060815e-06, |
|
"loss": 1.1824, |
|
"step": 1895 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 6.356756778246587e-06, |
|
"loss": 1.1744, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"eval_loss": 1.2491811513900757, |
|
"eval_runtime": 187.4757, |
|
"eval_samples_per_second": 75.476, |
|
"eval_steps_per_second": 2.363, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 6.305965352911162e-06, |
|
"loss": 1.1877, |
|
"step": 1905 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 6.255284045981653e-06, |
|
"loss": 1.1934, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 6.204714368257303e-06, |
|
"loss": 1.1995, |
|
"step": 1915 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 6.1542578272096975e-06, |
|
"loss": 1.1825, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 6.103915926937853e-06, |
|
"loss": 1.1885, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 6.053690168123367e-06, |
|
"loss": 1.2088, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 6.003582047985684e-06, |
|
"loss": 1.2033, |
|
"step": 1935 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 5.953593060237457e-06, |
|
"loss": 1.1946, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 5.903724695040036e-06, |
|
"loss": 1.1907, |
|
"step": 1945 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 5.8539784389590345e-06, |
|
"loss": 1.1708, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 5.804355774920017e-06, |
|
"loss": 1.1791, |
|
"step": 1955 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 5.754858182164298e-06, |
|
"loss": 1.1778, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 5.705487136204846e-06, |
|
"loss": 1.1918, |
|
"step": 1965 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 5.65624410878228e-06, |
|
"loss": 1.179, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 5.60713056782103e-06, |
|
"loss": 1.1841, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 5.558147977385558e-06, |
|
"loss": 1.1646, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 5.509297797636712e-06, |
|
"loss": 1.178, |
|
"step": 1985 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 5.46058148478821e-06, |
|
"loss": 1.1854, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 5.412000491063225e-06, |
|
"loss": 1.1738, |
|
"step": 1995 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 5.363556264651093e-06, |
|
"loss": 1.1718, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 1.2486178874969482, |
|
"eval_runtime": 187.198, |
|
"eval_samples_per_second": 75.588, |
|
"eval_steps_per_second": 2.366, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 5.315250249664151e-06, |
|
"loss": 1.1109, |
|
"step": 2005 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 5.267083886094668e-06, |
|
"loss": 1.1148, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 5.219058609771945e-06, |
|
"loss": 1.1205, |
|
"step": 2015 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 5.171175852319503e-06, |
|
"loss": 1.1116, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 5.123437041112395e-06, |
|
"loss": 1.1223, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 5.075843599234678e-06, |
|
"loss": 1.1306, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 5.0283969454369685e-06, |
|
"loss": 1.1276, |
|
"step": 2035 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 4.98109849409417e-06, |
|
"loss": 1.1394, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 4.933949655163295e-06, |
|
"loss": 1.1337, |
|
"step": 2045 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 4.886951834141446e-06, |
|
"loss": 1.1308, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 4.840106432023908e-06, |
|
"loss": 1.1257, |
|
"step": 2055 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 4.7934148452623906e-06, |
|
"loss": 1.1331, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 4.7468784657234055e-06, |
|
"loss": 1.1367, |
|
"step": 2065 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 4.7004986806467635e-06, |
|
"loss": 1.137, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 4.65427687260423e-06, |
|
"loss": 1.1224, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 4.608214419458306e-06, |
|
"loss": 1.1308, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 4.562312694321156e-06, |
|
"loss": 1.1196, |
|
"step": 2085 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 4.51657306551368e-06, |
|
"loss": 1.1362, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 4.470996896524713e-06, |
|
"loss": 1.1177, |
|
"step": 2095 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 4.425585545970392e-06, |
|
"loss": 1.1456, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"eval_loss": 1.253213882446289, |
|
"eval_runtime": 187.4923, |
|
"eval_samples_per_second": 75.47, |
|
"eval_steps_per_second": 2.363, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 4.380340367553649e-06, |
|
"loss": 1.1093, |
|
"step": 2105 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 4.335262710023861e-06, |
|
"loss": 1.1492, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 4.290353917136639e-06, |
|
"loss": 1.1202, |
|
"step": 2115 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 4.245615327613779e-06, |
|
"loss": 1.124, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 4.201048275103339e-06, |
|
"loss": 1.1398, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 4.156654088139907e-06, |
|
"loss": 1.1274, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 4.112434090104978e-06, |
|
"loss": 1.1421, |
|
"step": 2135 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 4.068389599187514e-06, |
|
"loss": 1.1198, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 4.024521928344646e-06, |
|
"loss": 1.1235, |
|
"step": 2145 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 3.980832385262532e-06, |
|
"loss": 1.1294, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 3.937322272317383e-06, |
|
"loss": 1.1484, |
|
"step": 2155 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 3.893992886536637e-06, |
|
"loss": 1.1328, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 3.850845519560279e-06, |
|
"loss": 1.1334, |
|
"step": 2165 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 3.807881457602367e-06, |
|
"loss": 1.1413, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 3.7651019814126656e-06, |
|
"loss": 1.1253, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 3.722508366238481e-06, |
|
"loss": 1.1141, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 3.6801018817866375e-06, |
|
"loss": 1.1358, |
|
"step": 2185 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 3.6378837921856324e-06, |
|
"loss": 1.1257, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 3.5958553559479524e-06, |
|
"loss": 1.1285, |
|
"step": 2195 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 3.5540178259325565e-06, |
|
"loss": 1.1204, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"eval_loss": 1.2528735399246216, |
|
"eval_runtime": 187.555, |
|
"eval_samples_per_second": 75.445, |
|
"eval_steps_per_second": 2.362, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 3.5123724493075318e-06, |
|
"loss": 1.111, |
|
"step": 2205 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 3.4709204675129025e-06, |
|
"loss": 1.1238, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 3.429663116223644e-06, |
|
"loss": 1.1265, |
|
"step": 2215 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 3.388601625312833e-06, |
|
"loss": 1.1217, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 3.347737218814987e-06, |
|
"loss": 1.1045, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 3.307071114889583e-06, |
|
"loss": 1.1302, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 3.2666045257847332e-06, |
|
"loss": 1.1359, |
|
"step": 2235 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 3.22633865780106e-06, |
|
"loss": 1.1121, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 3.1862747112557313e-06, |
|
"loss": 1.1278, |
|
"step": 2245 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 3.146413880446668e-06, |
|
"loss": 1.1233, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 3.106757353616966e-06, |
|
"loss": 1.1281, |
|
"step": 2255 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 3.067306312919456e-06, |
|
"loss": 1.1142, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 3.028061934381471e-06, |
|
"loss": 1.1173, |
|
"step": 2265 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 2.9890253878697885e-06, |
|
"loss": 1.1209, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 2.950197837055755e-06, |
|
"loss": 1.1336, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 2.911580439380599e-06, |
|
"loss": 1.1295, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 2.8731743460209325e-06, |
|
"loss": 1.1264, |
|
"step": 2285 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 2.834980701854417e-06, |
|
"loss": 1.1306, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 2.7970006454256605e-06, |
|
"loss": 1.126, |
|
"step": 2295 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 2.7592353089122582e-06, |
|
"loss": 1.1347, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"eval_loss": 1.2519265413284302, |
|
"eval_runtime": 187.5322, |
|
"eval_samples_per_second": 75.454, |
|
"eval_steps_per_second": 2.362, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 2.7216858180910532e-06, |
|
"loss": 1.1285, |
|
"step": 2305 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 2.6843532923045702e-06, |
|
"loss": 1.0964, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 2.647238844427653e-06, |
|
"loss": 1.1213, |
|
"step": 2315 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 2.610343580834287e-06, |
|
"loss": 1.1104, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 2.5736686013646226e-06, |
|
"loss": 1.1245, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 2.53721499929218e-06, |
|
"loss": 1.1252, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 2.5009838612912697e-06, |
|
"loss": 1.1356, |
|
"step": 2335 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 2.464976267404593e-06, |
|
"loss": 1.1232, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 2.4291932910110473e-06, |
|
"loss": 1.1347, |
|
"step": 2345 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 2.3936359987937353e-06, |
|
"loss": 1.1233, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 2.3583054507081526e-06, |
|
"loss": 1.1441, |
|
"step": 2355 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 2.3232026999506062e-06, |
|
"loss": 1.1234, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 2.2883287929268027e-06, |
|
"loss": 1.1313, |
|
"step": 2365 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 2.2536847692206722e-06, |
|
"loss": 1.1192, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 2.21927166156337e-06, |
|
"loss": 1.1196, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 2.185090495802491e-06, |
|
"loss": 1.1275, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 2.1511422908714897e-06, |
|
"loss": 1.1366, |
|
"step": 2385 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 2.1174280587593076e-06, |
|
"loss": 1.1447, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 2.083948804480204e-06, |
|
"loss": 1.1341, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 2.0507055260438014e-06, |
|
"loss": 1.1312, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"eval_loss": 1.251314640045166, |
|
"eval_runtime": 187.6132, |
|
"eval_samples_per_second": 75.421, |
|
"eval_steps_per_second": 2.361, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 2.017699214425323e-06, |
|
"loss": 1.1394, |
|
"step": 2405 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 1.9849308535360688e-06, |
|
"loss": 1.132, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 1.9524014201940732e-06, |
|
"loss": 1.1198, |
|
"step": 2415 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 1.9201118840949905e-06, |
|
"loss": 1.133, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 1.8880632077831906e-06, |
|
"loss": 1.1245, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 1.8562563466230577e-06, |
|
"loss": 1.1158, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 1.824692248770521e-06, |
|
"loss": 1.1329, |
|
"step": 2435 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 1.7933718551447877e-06, |
|
"loss": 1.1271, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 1.7622960994002836e-06, |
|
"loss": 1.1118, |
|
"step": 2445 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 1.7314659078988383e-06, |
|
"loss": 1.1345, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 1.700882199682059e-06, |
|
"loss": 1.1148, |
|
"step": 2455 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 1.6705458864439405e-06, |
|
"loss": 1.1091, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 1.6404578725036802e-06, |
|
"loss": 1.1218, |
|
"step": 2465 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 1.6106190547787282e-06, |
|
"loss": 1.1482, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 1.5810303227580448e-06, |
|
"loss": 1.1214, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 1.5516925584755938e-06, |
|
"loss": 1.1365, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 1.5226066364840298e-06, |
|
"loss": 1.1382, |
|
"step": 2485 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 1.4937734238286539e-06, |
|
"loss": 1.1355, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 1.4651937800215465e-06, |
|
"loss": 1.1417, |
|
"step": 2495 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 1.436868557015959e-06, |
|
"loss": 1.1229, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"eval_loss": 1.2507939338684082, |
|
"eval_runtime": 187.2741, |
|
"eval_samples_per_second": 75.558, |
|
"eval_steps_per_second": 2.366, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 1.4087985991809072e-06, |
|
"loss": 1.1236, |
|
"step": 2505 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 1.380984743276006e-06, |
|
"loss": 1.1418, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 1.353427818426527e-06, |
|
"loss": 1.1285, |
|
"step": 2515 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 1.32612864609868e-06, |
|
"loss": 1.129, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 1.2990880400751204e-06, |
|
"loss": 1.1223, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 1.272306806430702e-06, |
|
"loss": 1.1346, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 1.245785743508441e-06, |
|
"loss": 1.1287, |
|
"step": 2535 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 1.2195256418957146e-06, |
|
"loss": 1.133, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 1.1935272844007029e-06, |
|
"loss": 1.1458, |
|
"step": 2545 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 1.167791446029043e-06, |
|
"loss": 1.1291, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 1.1423188939607356e-06, |
|
"loss": 1.1175, |
|
"step": 2555 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 1.1171103875272693e-06, |
|
"loss": 1.1275, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 1.0921666781889861e-06, |
|
"loss": 1.1362, |
|
"step": 2565 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 1.067488509512683e-06, |
|
"loss": 1.1202, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 1.0430766171494444e-06, |
|
"loss": 1.1385, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 1.0189317288127154e-06, |
|
"loss": 1.1232, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 9.950545642566045e-07, |
|
"loss": 1.1201, |
|
"step": 2585 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 9.714458352544276e-07, |
|
"loss": 1.1086, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 9.481062455774981e-07, |
|
"loss": 1.1154, |
|
"step": 2595 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 9.25036490974136e-07, |
|
"loss": 1.1287, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"eval_loss": 1.2499793767929077, |
|
"eval_runtime": 187.3438, |
|
"eval_samples_per_second": 75.53, |
|
"eval_steps_per_second": 2.365, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 9.022372591489381e-07, |
|
"loss": 1.1222, |
|
"step": 2605 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 8.797092297422715e-07, |
|
"loss": 1.1174, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 8.574530743100151e-07, |
|
"loss": 1.1299, |
|
"step": 2615 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 8.354694563035426e-07, |
|
"loss": 1.1145, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 8.137590310499433e-07, |
|
"loss": 1.1362, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 7.923224457324852e-07, |
|
"loss": 1.1245, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 7.711603393713297e-07, |
|
"loss": 1.1288, |
|
"step": 2635 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 7.502733428044684e-07, |
|
"loss": 1.145, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 7.296620786689356e-07, |
|
"loss": 1.1319, |
|
"step": 2645 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 7.093271613822373e-07, |
|
"loss": 1.1267, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 6.89269197124034e-07, |
|
"loss": 1.1307, |
|
"step": 2655 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 6.694887838180764e-07, |
|
"loss": 1.1417, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 6.499865111143788e-07, |
|
"loss": 1.141, |
|
"step": 2665 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 6.307629603716425e-07, |
|
"loss": 1.1078, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 6.11818704639926e-07, |
|
"loss": 1.1363, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 5.931543086435543e-07, |
|
"loss": 1.1374, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 5.747703287643003e-07, |
|
"loss": 1.1353, |
|
"step": 2685 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 5.566673130247879e-07, |
|
"loss": 1.131, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 5.388458010721575e-07, |
|
"loss": 1.1376, |
|
"step": 2695 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 5.213063241619798e-07, |
|
"loss": 1.1252, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"eval_loss": 1.2499687671661377, |
|
"eval_runtime": 196.4522, |
|
"eval_samples_per_second": 72.028, |
|
"eval_steps_per_second": 4.505, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 5.040494051424205e-07, |
|
"loss": 1.1257, |
|
"step": 2705 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 4.870755584386544e-07, |
|
"loss": 1.114, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 4.703852900375272e-07, |
|
"loss": 1.1383, |
|
"step": 2715 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 4.539790974724723e-07, |
|
"loss": 1.1017, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 4.378574698086868e-07, |
|
"loss": 1.1043, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 4.2202088762854255e-07, |
|
"loss": 1.1242, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 4.064698230172681e-07, |
|
"loss": 1.1048, |
|
"step": 2735 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 3.912047395488705e-07, |
|
"loss": 1.12, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 3.762260922723182e-07, |
|
"loss": 1.1194, |
|
"step": 2745 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 3.615343276979777e-07, |
|
"loss": 1.1323, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 3.471298837843007e-07, |
|
"loss": 1.1461, |
|
"step": 2755 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 3.3301318992476904e-07, |
|
"loss": 1.1443, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 3.191846669350973e-07, |
|
"loss": 1.1128, |
|
"step": 2765 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 3.056447270406826e-07, |
|
"loss": 1.1371, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 2.9239377386432345e-07, |
|
"loss": 1.1182, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 2.7943220241418376e-07, |
|
"loss": 1.12, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 2.667603990720169e-07, |
|
"loss": 1.1394, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 2.5437874158164897e-07, |
|
"loss": 1.1155, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 2.422875990377205e-07, |
|
"loss": 1.1358, |
|
"step": 2795 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 2.3048733187468032e-07, |
|
"loss": 1.139, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"eval_loss": 1.2498047351837158, |
|
"eval_runtime": 196.456, |
|
"eval_samples_per_second": 72.026, |
|
"eval_steps_per_second": 4.505, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 2.1897829185604235e-07, |
|
"loss": 1.1373, |
|
"step": 2805 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 2.0776082206389936e-07, |
|
"loss": 1.127, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 1.9683525688869776e-07, |
|
"loss": 1.1277, |
|
"step": 2815 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 1.862019220192679e-07, |
|
"loss": 1.1192, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 1.7586113443311404e-07, |
|
"loss": 1.1621, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 1.658132023869663e-07, |
|
"loss": 1.1335, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 1.5605842540759476e-07, |
|
"loss": 1.1206, |
|
"step": 2835 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 1.4659709428287538e-07, |
|
"loss": 1.1355, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 1.3742949105312375e-07, |
|
"loss": 1.1225, |
|
"step": 2845 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 1.2855588900269057e-07, |
|
"loss": 1.1141, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 1.199765526518082e-07, |
|
"loss": 1.1299, |
|
"step": 2855 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 1.1169173774871478e-07, |
|
"loss": 1.1216, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 1.0370169126202145e-07, |
|
"loss": 1.1237, |
|
"step": 2865 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 9.600665137335374e-08, |
|
"loss": 1.121, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 8.860684747025727e-08, |
|
"loss": 1.1194, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 8.150250013934547e-08, |
|
"loss": 1.1187, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 7.469382115974034e-08, |
|
"loss": 1.1365, |
|
"step": 2885 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 6.818101349674756e-08, |
|
"loss": 1.1118, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 6.196427129581129e-08, |
|
"loss": 1.1273, |
|
"step": 2895 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 5.60437798767266e-08, |
|
"loss": 1.1282, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"eval_loss": 1.2497347593307495, |
|
"eval_runtime": 196.0724, |
|
"eval_samples_per_second": 72.167, |
|
"eval_steps_per_second": 4.514, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 5.041971572811277e-08, |
|
"loss": 1.137, |
|
"step": 2905 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 4.509224650215527e-08, |
|
"loss": 1.1452, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 4.0061531009605305e-08, |
|
"loss": 1.125, |
|
"step": 2915 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 3.5327719215046965e-08, |
|
"loss": 1.1291, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 3.089095223242744e-08, |
|
"loss": 1.112, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 2.6751362320849293e-08, |
|
"loss": 1.1385, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 2.2909072880629156e-08, |
|
"loss": 1.1345, |
|
"step": 2935 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 1.9364198449616233e-08, |
|
"loss": 1.1271, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 1.6116844699781697e-08, |
|
"loss": 1.1138, |
|
"step": 2945 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 1.3167108434066766e-08, |
|
"loss": 1.1118, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 1.0515077583498346e-08, |
|
"loss": 1.1192, |
|
"step": 2955 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 8.160831204563347e-09, |
|
"loss": 1.1291, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 6.1044394768594585e-09, |
|
"loss": 1.1321, |
|
"step": 2965 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 4.345963700995714e-09, |
|
"loss": 1.1255, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 2.8854562967706167e-09, |
|
"loss": 1.1247, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 1.7229608016067213e-09, |
|
"loss": 1.1246, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 8.585118692538974e-10, |
|
"loss": 1.1297, |
|
"step": 2985 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 2.9213526875349063e-10, |
|
"loss": 1.1431, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 2.384788367670865e-11, |
|
"loss": 1.1115, |
|
"step": 2995 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"step": 2997, |
|
"total_flos": 1399131230699520.0, |
|
"train_loss": 0.14928224097103288, |
|
"train_runtime": 3033.3773, |
|
"train_samples_per_second": 126.446, |
|
"train_steps_per_second": 0.988 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 2997, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 100, |
|
"total_flos": 1399131230699520.0, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|