|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.0, |
|
"eval_steps": 1000, |
|
"global_step": 676, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 7.142857142857143e-07, |
|
"loss": 1.7121, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.4285714285714286e-06, |
|
"loss": 1.6351, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.1428571428571427e-06, |
|
"loss": 0.9576, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.8571428571428573e-06, |
|
"loss": 0.6691, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 3.5714285714285718e-06, |
|
"loss": 1.155, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.2857142857142855e-06, |
|
"loss": 0.5538, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 5e-06, |
|
"loss": 1.5242, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 5.7142857142857145e-06, |
|
"loss": 1.0917, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 6.4285714285714295e-06, |
|
"loss": 0.9385, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 7.1428571428571436e-06, |
|
"loss": 0.745, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 7.857142857142858e-06, |
|
"loss": 1.0332, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.571428571428571e-06, |
|
"loss": 0.8973, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.285714285714288e-06, |
|
"loss": 0.8048, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1e-05, |
|
"loss": 0.9936, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.0714285714285714e-05, |
|
"loss": 0.8611, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.1428571428571429e-05, |
|
"loss": 0.951, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.2142857142857142e-05, |
|
"loss": 0.636, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1.2857142857142859e-05, |
|
"loss": 0.2592, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.3571428571428574e-05, |
|
"loss": 0.7124, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.4285714285714287e-05, |
|
"loss": 0.6169, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.5000000000000002e-05, |
|
"loss": 0.6877, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.5714285714285715e-05, |
|
"loss": 0.5239, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.642857142857143e-05, |
|
"loss": 0.7335, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.7142857142857142e-05, |
|
"loss": 0.7291, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 1.785714285714286e-05, |
|
"loss": 0.6331, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.8571428571428575e-05, |
|
"loss": 0.3293, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.928571428571429e-05, |
|
"loss": 0.5166, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 2e-05, |
|
"loss": 0.3995, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.999252615844544e-05, |
|
"loss": 0.4698, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.9985052316890884e-05, |
|
"loss": 0.599, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.9977578475336326e-05, |
|
"loss": 0.3244, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 1.9970104633781765e-05, |
|
"loss": 0.4321, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.9962630792227208e-05, |
|
"loss": 0.2851, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.9955156950672647e-05, |
|
"loss": 0.5091, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.9947683109118087e-05, |
|
"loss": 0.4354, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.994020926756353e-05, |
|
"loss": 0.4124, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.9932735426008972e-05, |
|
"loss": 0.4631, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 1.992526158445441e-05, |
|
"loss": 0.2247, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.9917787742899854e-05, |
|
"loss": 0.3505, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.9910313901345293e-05, |
|
"loss": 0.2736, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.9902840059790732e-05, |
|
"loss": 0.3274, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.9895366218236175e-05, |
|
"loss": 0.4126, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.9887892376681618e-05, |
|
"loss": 0.4884, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.9880418535127057e-05, |
|
"loss": 0.3895, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.98729446935725e-05, |
|
"loss": 0.3829, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.986547085201794e-05, |
|
"loss": 0.2123, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.9857997010463378e-05, |
|
"loss": 0.3708, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.985052316890882e-05, |
|
"loss": 0.3577, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.9843049327354263e-05, |
|
"loss": 0.4285, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.9835575485799703e-05, |
|
"loss": 0.495, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.9828101644245142e-05, |
|
"loss": 0.3795, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 1.9820627802690585e-05, |
|
"loss": 0.254, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.9813153961136024e-05, |
|
"loss": 0.2853, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.9805680119581466e-05, |
|
"loss": 0.2242, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.979820627802691e-05, |
|
"loss": 0.3997, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.979073243647235e-05, |
|
"loss": 0.3478, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.9783258594917788e-05, |
|
"loss": 0.3803, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.977578475336323e-05, |
|
"loss": 0.4557, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.976831091180867e-05, |
|
"loss": 0.4193, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.9760837070254112e-05, |
|
"loss": 0.2113, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.9753363228699555e-05, |
|
"loss": 0.3107, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.9745889387144994e-05, |
|
"loss": 0.337, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.9738415545590433e-05, |
|
"loss": 0.375, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.9730941704035876e-05, |
|
"loss": 0.3449, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.972346786248132e-05, |
|
"loss": 0.2696, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.9715994020926758e-05, |
|
"loss": 0.3429, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.97085201793722e-05, |
|
"loss": 0.3534, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.970104633781764e-05, |
|
"loss": 0.2249, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.969357249626308e-05, |
|
"loss": 0.2489, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.9686098654708522e-05, |
|
"loss": 0.4954, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.9678624813153964e-05, |
|
"loss": 0.3308, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.9671150971599404e-05, |
|
"loss": 0.2431, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.9663677130044846e-05, |
|
"loss": 0.2689, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.9656203288490286e-05, |
|
"loss": 0.2281, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.9648729446935725e-05, |
|
"loss": 0.3062, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.9641255605381167e-05, |
|
"loss": 0.339, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.963378176382661e-05, |
|
"loss": 0.2751, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.962630792227205e-05, |
|
"loss": 0.3185, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.9618834080717492e-05, |
|
"loss": 0.3796, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.961136023916293e-05, |
|
"loss": 0.3768, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.960388639760837e-05, |
|
"loss": 0.2139, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.9596412556053813e-05, |
|
"loss": 0.1631, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.9588938714499256e-05, |
|
"loss": 0.2343, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.9581464872944695e-05, |
|
"loss": 0.3333, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.9573991031390134e-05, |
|
"loss": 0.2805, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.9566517189835577e-05, |
|
"loss": 0.3305, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.9559043348281016e-05, |
|
"loss": 0.2159, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.955156950672646e-05, |
|
"loss": 0.3456, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.95440956651719e-05, |
|
"loss": 0.2844, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.953662182361734e-05, |
|
"loss": 0.2367, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.952914798206278e-05, |
|
"loss": 0.2048, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.9521674140508223e-05, |
|
"loss": 0.2589, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.9514200298953662e-05, |
|
"loss": 0.2653, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.9506726457399105e-05, |
|
"loss": 0.2004, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.9499252615844547e-05, |
|
"loss": 0.2007, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.9491778774289987e-05, |
|
"loss": 0.1923, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.9484304932735426e-05, |
|
"loss": 0.1986, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.947683109118087e-05, |
|
"loss": 0.4276, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.946935724962631e-05, |
|
"loss": 0.2475, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.946188340807175e-05, |
|
"loss": 0.1941, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.9454409566517193e-05, |
|
"loss": 0.2069, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.9446935724962632e-05, |
|
"loss": 0.2331, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.943946188340807e-05, |
|
"loss": 0.2208, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.9431988041853514e-05, |
|
"loss": 0.2585, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.9424514200298957e-05, |
|
"loss": 0.2988, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.9417040358744396e-05, |
|
"loss": 0.2965, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.940956651718984e-05, |
|
"loss": 0.2611, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.9402092675635278e-05, |
|
"loss": 0.2909, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.9394618834080717e-05, |
|
"loss": 0.1858, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.938714499252616e-05, |
|
"loss": 0.1401, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.9379671150971603e-05, |
|
"loss": 0.2057, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.9372197309417042e-05, |
|
"loss": 0.128, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.9364723467862484e-05, |
|
"loss": 0.2158, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.9357249626307924e-05, |
|
"loss": 0.2921, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.9349775784753363e-05, |
|
"loss": 0.2217, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.9342301943198806e-05, |
|
"loss": 0.1605, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.9334828101644248e-05, |
|
"loss": 0.2123, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.9327354260089688e-05, |
|
"loss": 0.1571, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.931988041853513e-05, |
|
"loss": 0.3153, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.931240657698057e-05, |
|
"loss": 0.1478, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.930493273542601e-05, |
|
"loss": 0.2461, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.929745889387145e-05, |
|
"loss": 0.1576, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.9289985052316894e-05, |
|
"loss": 0.2529, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.9282511210762333e-05, |
|
"loss": 0.1751, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.9275037369207773e-05, |
|
"loss": 0.2999, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.9267563527653215e-05, |
|
"loss": 0.3322, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.9260089686098654e-05, |
|
"loss": 0.2252, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.9252615844544097e-05, |
|
"loss": 0.1505, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.924514200298954e-05, |
|
"loss": 0.1555, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.923766816143498e-05, |
|
"loss": 0.1885, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.9230194319880418e-05, |
|
"loss": 0.2109, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.922272047832586e-05, |
|
"loss": 0.1625, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.9215246636771304e-05, |
|
"loss": 0.1517, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.9207772795216743e-05, |
|
"loss": 0.2655, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.9200298953662185e-05, |
|
"loss": 0.1839, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.9192825112107625e-05, |
|
"loss": 0.258, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.9185351270553064e-05, |
|
"loss": 0.1707, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.9177877428998507e-05, |
|
"loss": 0.0959, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.917040358744395e-05, |
|
"loss": 0.1505, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.916292974588939e-05, |
|
"loss": 0.1793, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.915545590433483e-05, |
|
"loss": 0.1991, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.914798206278027e-05, |
|
"loss": 0.1646, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.914050822122571e-05, |
|
"loss": 0.1877, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.9133034379671152e-05, |
|
"loss": 0.2178, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.9125560538116595e-05, |
|
"loss": 0.1381, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.9118086696562034e-05, |
|
"loss": 0.1962, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.9110612855007477e-05, |
|
"loss": 0.2094, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.9103139013452916e-05, |
|
"loss": 0.1103, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.9095665171898355e-05, |
|
"loss": 0.1808, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.9088191330343798e-05, |
|
"loss": 0.3039, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.908071748878924e-05, |
|
"loss": 0.1886, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.907324364723468e-05, |
|
"loss": 0.1659, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.9065769805680123e-05, |
|
"loss": 0.1135, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.9058295964125562e-05, |
|
"loss": 0.1233, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.9050822122571e-05, |
|
"loss": 0.1556, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.9043348281016444e-05, |
|
"loss": 0.1469, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.9035874439461886e-05, |
|
"loss": 0.126, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.9028400597907326e-05, |
|
"loss": 0.1184, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.9020926756352765e-05, |
|
"loss": 0.1788, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.9013452914798208e-05, |
|
"loss": 0.1624, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.9005979073243647e-05, |
|
"loss": 0.3195, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.899850523168909e-05, |
|
"loss": 0.1057, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.8991031390134532e-05, |
|
"loss": 0.1392, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.898355754857997e-05, |
|
"loss": 0.1107, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.897608370702541e-05, |
|
"loss": 0.217, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.8968609865470853e-05, |
|
"loss": 0.161, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.8961136023916296e-05, |
|
"loss": 0.1969, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.8953662182361735e-05, |
|
"loss": 0.1386, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.8946188340807178e-05, |
|
"loss": 0.1529, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 1.8938714499252617e-05, |
|
"loss": 0.109, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.8931240657698056e-05, |
|
"loss": 0.1202, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.89237668161435e-05, |
|
"loss": 0.0875, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.8916292974588942e-05, |
|
"loss": 0.1248, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.890881913303438e-05, |
|
"loss": 0.0989, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.8901345291479824e-05, |
|
"loss": 0.1993, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.8893871449925263e-05, |
|
"loss": 0.094, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 1.8886397608370702e-05, |
|
"loss": 0.2574, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 1.8878923766816145e-05, |
|
"loss": 0.1246, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 1.8871449925261588e-05, |
|
"loss": 0.1293, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 1.8863976083707027e-05, |
|
"loss": 0.2807, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 1.885650224215247e-05, |
|
"loss": 0.0824, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 1.884902840059791e-05, |
|
"loss": 0.1072, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 1.8841554559043348e-05, |
|
"loss": 0.1424, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 1.883408071748879e-05, |
|
"loss": 0.136, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.8826606875934233e-05, |
|
"loss": 0.1441, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.8819133034379672e-05, |
|
"loss": 0.1195, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.8811659192825115e-05, |
|
"loss": 0.1971, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.8804185351270554e-05, |
|
"loss": 0.1315, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.8796711509715994e-05, |
|
"loss": 0.2408, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 1.8789237668161436e-05, |
|
"loss": 0.1223, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 1.878176382660688e-05, |
|
"loss": 0.13, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 1.8774289985052318e-05, |
|
"loss": 0.1561, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 1.8766816143497757e-05, |
|
"loss": 0.174, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 1.87593423019432e-05, |
|
"loss": 0.1024, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.875186846038864e-05, |
|
"loss": 0.1497, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.8744394618834082e-05, |
|
"loss": 0.1025, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.8736920777279525e-05, |
|
"loss": 0.1955, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.8729446935724964e-05, |
|
"loss": 0.3608, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.8721973094170403e-05, |
|
"loss": 0.1002, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.8714499252615846e-05, |
|
"loss": 0.0897, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 1.870702541106129e-05, |
|
"loss": 0.1505, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 1.8699551569506728e-05, |
|
"loss": 0.1407, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 1.869207772795217e-05, |
|
"loss": 0.1, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 1.868460388639761e-05, |
|
"loss": 0.1155, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.867713004484305e-05, |
|
"loss": 0.1011, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.866965620328849e-05, |
|
"loss": 0.1389, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.8662182361733934e-05, |
|
"loss": 0.1635, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.8654708520179373e-05, |
|
"loss": 0.0981, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.8647234678624816e-05, |
|
"loss": 0.0854, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.8639760837070255e-05, |
|
"loss": 0.0481, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.8632286995515695e-05, |
|
"loss": 0.1173, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 1.8624813153961137e-05, |
|
"loss": 0.1232, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 1.861733931240658e-05, |
|
"loss": 0.1024, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 1.860986547085202e-05, |
|
"loss": 0.0921, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.8602391629297462e-05, |
|
"loss": 0.0635, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.85949177877429e-05, |
|
"loss": 0.0682, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.858744394618834e-05, |
|
"loss": 0.1079, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.8579970104633783e-05, |
|
"loss": 0.0951, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.8572496263079226e-05, |
|
"loss": 0.0614, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.8565022421524665e-05, |
|
"loss": 0.1315, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 1.8557548579970108e-05, |
|
"loss": 0.0428, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 1.8550074738415547e-05, |
|
"loss": 0.0627, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 1.8542600896860986e-05, |
|
"loss": 0.1626, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 1.853512705530643e-05, |
|
"loss": 0.2031, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.852765321375187e-05, |
|
"loss": 0.2312, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.852017937219731e-05, |
|
"loss": 0.077, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.8512705530642753e-05, |
|
"loss": 0.0495, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 1.8505231689088193e-05, |
|
"loss": 0.118, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.8497757847533632e-05, |
|
"loss": 0.1213, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.8490284005979075e-05, |
|
"loss": 0.2259, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.8482810164424517e-05, |
|
"loss": 0.1215, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.8475336322869956e-05, |
|
"loss": 0.0947, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.8467862481315396e-05, |
|
"loss": 0.1236, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 1.846038863976084e-05, |
|
"loss": 0.1157, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.845291479820628e-05, |
|
"loss": 0.1132, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.844544095665172e-05, |
|
"loss": 0.0576, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.8437967115097163e-05, |
|
"loss": 0.262, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.8430493273542602e-05, |
|
"loss": 0.0824, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.842301943198804e-05, |
|
"loss": 0.1578, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.8415545590433484e-05, |
|
"loss": 0.0914, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.8408071748878927e-05, |
|
"loss": 0.1206, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.8400597907324366e-05, |
|
"loss": 0.1859, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.839312406576981e-05, |
|
"loss": 0.1447, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.8385650224215248e-05, |
|
"loss": 0.0675, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.8378176382660687e-05, |
|
"loss": 0.0735, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.837070254110613e-05, |
|
"loss": 0.0905, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.8363228699551572e-05, |
|
"loss": 0.0429, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 1.835575485799701e-05, |
|
"loss": 0.0729, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.8348281016442454e-05, |
|
"loss": 0.0806, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.8340807174887894e-05, |
|
"loss": 0.0601, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.8333333333333333e-05, |
|
"loss": 0.062, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.8325859491778776e-05, |
|
"loss": 0.1074, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.8318385650224218e-05, |
|
"loss": 0.0427, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.8310911808669657e-05, |
|
"loss": 0.1015, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 1.83034379671151e-05, |
|
"loss": 0.0521, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.829596412556054e-05, |
|
"loss": 0.0448, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.828849028400598e-05, |
|
"loss": 0.0585, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.828101644245142e-05, |
|
"loss": 0.0849, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.8273542600896864e-05, |
|
"loss": 0.0897, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.8266068759342303e-05, |
|
"loss": 0.0221, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.8258594917787746e-05, |
|
"loss": 0.0745, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.8251121076233185e-05, |
|
"loss": 0.0403, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.8243647234678624e-05, |
|
"loss": 0.0425, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.8236173393124067e-05, |
|
"loss": 0.1864, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.822869955156951e-05, |
|
"loss": 0.1687, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.822122571001495e-05, |
|
"loss": 0.1226, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.8213751868460388e-05, |
|
"loss": 0.0665, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 1.820627802690583e-05, |
|
"loss": 0.0278, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.8198804185351273e-05, |
|
"loss": 0.0579, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.8191330343796713e-05, |
|
"loss": 0.076, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.8183856502242155e-05, |
|
"loss": 0.1066, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.8176382660687595e-05, |
|
"loss": 0.0512, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.8168908819133034e-05, |
|
"loss": 0.0652, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.8161434977578477e-05, |
|
"loss": 0.0533, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.815396113602392e-05, |
|
"loss": 0.0637, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.814648729446936e-05, |
|
"loss": 0.042, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.81390134529148e-05, |
|
"loss": 0.031, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.813153961136024e-05, |
|
"loss": 0.0618, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.812406576980568e-05, |
|
"loss": 0.1812, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.8116591928251122e-05, |
|
"loss": 0.0516, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.8109118086696565e-05, |
|
"loss": 0.0574, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.8101644245142004e-05, |
|
"loss": 0.0619, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.8094170403587447e-05, |
|
"loss": 0.0776, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.8086696562032886e-05, |
|
"loss": 0.1249, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.8079222720478325e-05, |
|
"loss": 0.1777, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.8071748878923768e-05, |
|
"loss": 0.0577, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.806427503736921e-05, |
|
"loss": 0.0734, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.805680119581465e-05, |
|
"loss": 0.1135, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.8049327354260093e-05, |
|
"loss": 0.032, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.8041853512705532e-05, |
|
"loss": 0.1834, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.803437967115097e-05, |
|
"loss": 0.0563, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.8026905829596414e-05, |
|
"loss": 0.0525, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.8019431988041856e-05, |
|
"loss": 0.0602, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.8011958146487296e-05, |
|
"loss": 0.1962, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.8004484304932738e-05, |
|
"loss": 0.0732, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.7997010463378178e-05, |
|
"loss": 0.0252, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.7989536621823617e-05, |
|
"loss": 0.1243, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.798206278026906e-05, |
|
"loss": 0.0363, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 1.7974588938714502e-05, |
|
"loss": 0.0544, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.796711509715994e-05, |
|
"loss": 0.0315, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.7959641255605384e-05, |
|
"loss": 0.1768, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.7952167414050823e-05, |
|
"loss": 0.0601, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.7944693572496266e-05, |
|
"loss": 0.0393, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.7937219730941705e-05, |
|
"loss": 0.0289, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.7929745889387148e-05, |
|
"loss": 0.1323, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.7922272047832587e-05, |
|
"loss": 0.1477, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.7914798206278026e-05, |
|
"loss": 0.0374, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.790732436472347e-05, |
|
"loss": 0.1036, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.789985052316891e-05, |
|
"loss": 0.0417, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.789237668161435e-05, |
|
"loss": 0.0736, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.7884902840059794e-05, |
|
"loss": 0.0307, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.7877428998505233e-05, |
|
"loss": 0.0684, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.7869955156950672e-05, |
|
"loss": 0.0391, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.7862481315396115e-05, |
|
"loss": 0.0781, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.7855007473841557e-05, |
|
"loss": 0.2455, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.7847533632286997e-05, |
|
"loss": 0.1698, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.784005979073244e-05, |
|
"loss": 0.0173, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.783258594917788e-05, |
|
"loss": 0.1204, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.7825112107623318e-05, |
|
"loss": 0.0667, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.781763826606876e-05, |
|
"loss": 0.0214, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.7810164424514203e-05, |
|
"loss": 0.0351, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.7802690582959642e-05, |
|
"loss": 0.2676, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.7795216741405085e-05, |
|
"loss": 0.0633, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.7787742899850524e-05, |
|
"loss": 0.0249, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.7780269058295964e-05, |
|
"loss": 0.1116, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.7772795216741406e-05, |
|
"loss": 0.0439, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.776532137518685e-05, |
|
"loss": 0.0798, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.7757847533632288e-05, |
|
"loss": 0.0249, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.775037369207773e-05, |
|
"loss": 0.0285, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.774289985052317e-05, |
|
"loss": 0.0945, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.773542600896861e-05, |
|
"loss": 0.1555, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.7727952167414052e-05, |
|
"loss": 0.0378, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.7720478325859495e-05, |
|
"loss": 0.0234, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.7713004484304934e-05, |
|
"loss": 0.084, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.7705530642750376e-05, |
|
"loss": 0.0862, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.7698056801195816e-05, |
|
"loss": 0.0364, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 1.769058295964126e-05, |
|
"loss": 0.0281, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 1.7683109118086698e-05, |
|
"loss": 0.0277, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 1.767563527653214e-05, |
|
"loss": 0.014, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 1.766816143497758e-05, |
|
"loss": 0.0129, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 1.766068759342302e-05, |
|
"loss": 0.0411, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 1.765321375186846e-05, |
|
"loss": 0.0182, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 1.7645739910313904e-05, |
|
"loss": 0.0105, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 1.7638266068759343e-05, |
|
"loss": 0.0083, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 1.7630792227204786e-05, |
|
"loss": 0.0113, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 1.7623318385650225e-05, |
|
"loss": 0.0051, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 1.7615844544095665e-05, |
|
"loss": 0.01, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 1.7608370702541107e-05, |
|
"loss": 0.0148, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 1.760089686098655e-05, |
|
"loss": 0.0527, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 1.759342301943199e-05, |
|
"loss": 0.0112, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 1.7585949177877432e-05, |
|
"loss": 0.0759, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 1.757847533632287e-05, |
|
"loss": 0.0126, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 1.757100149476831e-05, |
|
"loss": 0.0165, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 1.7563527653213753e-05, |
|
"loss": 0.0113, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 1.7556053811659196e-05, |
|
"loss": 0.0394, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 1.7548579970104635e-05, |
|
"loss": 0.009, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 1.7541106128550077e-05, |
|
"loss": 0.0126, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 1.7533632286995517e-05, |
|
"loss": 0.0162, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 1.7526158445440956e-05, |
|
"loss": 0.0052, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.75186846038864e-05, |
|
"loss": 0.059, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.751121076233184e-05, |
|
"loss": 0.0444, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.750373692077728e-05, |
|
"loss": 0.006, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.7496263079222723e-05, |
|
"loss": 0.0109, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.7488789237668162e-05, |
|
"loss": 0.0062, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.7481315396113602e-05, |
|
"loss": 0.0102, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 1.7473841554559044e-05, |
|
"loss": 0.01, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.7466367713004487e-05, |
|
"loss": 0.0094, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.7458893871449926e-05, |
|
"loss": 0.0457, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.745142002989537e-05, |
|
"loss": 0.009, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.7443946188340808e-05, |
|
"loss": 0.0058, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.7436472346786247e-05, |
|
"loss": 0.0078, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.742899850523169e-05, |
|
"loss": 0.011, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 1.7421524663677133e-05, |
|
"loss": 0.0078, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.7414050822122572e-05, |
|
"loss": 0.0081, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.740657698056801e-05, |
|
"loss": 0.0059, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.7399103139013454e-05, |
|
"loss": 0.0101, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.7391629297458897e-05, |
|
"loss": 0.0474, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.7384155455904336e-05, |
|
"loss": 0.0085, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.737668161434978e-05, |
|
"loss": 0.0097, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 1.7369207772795218e-05, |
|
"loss": 0.0073, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.7361733931240657e-05, |
|
"loss": 0.0101, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.73542600896861e-05, |
|
"loss": 0.0569, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.7346786248131542e-05, |
|
"loss": 0.0099, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.733931240657698e-05, |
|
"loss": 0.0083, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.7331838565022424e-05, |
|
"loss": 0.0178, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.7324364723467863e-05, |
|
"loss": 0.0059, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 1.7316890881913303e-05, |
|
"loss": 0.0073, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.7309417040358745e-05, |
|
"loss": 0.0295, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.7301943198804188e-05, |
|
"loss": 0.0382, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 1.7294469357249627e-05, |
|
"loss": 0.0195, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.728699551569507e-05, |
|
"loss": 0.0097, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.727952167414051e-05, |
|
"loss": 0.0111, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.727204783258595e-05, |
|
"loss": 0.0214, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 1.726457399103139e-05, |
|
"loss": 0.0106, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 1.7257100149476834e-05, |
|
"loss": 0.0035, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 1.7249626307922273e-05, |
|
"loss": 0.0935, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 1.7242152466367716e-05, |
|
"loss": 0.0049, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 1.7234678624813155e-05, |
|
"loss": 0.0112, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 1.7227204783258594e-05, |
|
"loss": 0.0231, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 1.7219730941704037e-05, |
|
"loss": 0.016, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 1.721225710014948e-05, |
|
"loss": 0.0096, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 1.720478325859492e-05, |
|
"loss": 0.0199, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 1.719730941704036e-05, |
|
"loss": 0.0155, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 1.71898355754858e-05, |
|
"loss": 0.0723, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 1.718236173393124e-05, |
|
"loss": 0.0126, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 1.7174887892376683e-05, |
|
"loss": 0.0065, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 1.7167414050822125e-05, |
|
"loss": 0.0155, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 1.7159940209267564e-05, |
|
"loss": 0.0112, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 1.7152466367713007e-05, |
|
"loss": 0.0134, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 1.7144992526158446e-05, |
|
"loss": 0.0664, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 1.713751868460389e-05, |
|
"loss": 0.0233, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 1.7130044843049328e-05, |
|
"loss": 0.0259, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 1.712257100149477e-05, |
|
"loss": 0.1194, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 1.711509715994021e-05, |
|
"loss": 0.0067, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 1.710762331838565e-05, |
|
"loss": 0.0146, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 1.7100149476831092e-05, |
|
"loss": 0.0164, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 1.7092675635276535e-05, |
|
"loss": 0.0465, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.7085201793721974e-05, |
|
"loss": 0.0421, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.7077727952167417e-05, |
|
"loss": 0.0049, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.7070254110612856e-05, |
|
"loss": 0.028, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.7062780269058295e-05, |
|
"loss": 0.0055, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.7055306427503738e-05, |
|
"loss": 0.0451, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.704783258594918e-05, |
|
"loss": 0.0283, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.704035874439462e-05, |
|
"loss": 0.0129, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 1.7032884902840062e-05, |
|
"loss": 0.0117, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 1.70254110612855e-05, |
|
"loss": 0.0129, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 1.701793721973094e-05, |
|
"loss": 0.0109, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 1.7010463378176384e-05, |
|
"loss": 0.0095, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 1.7002989536621826e-05, |
|
"loss": 0.0711, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 1.6995515695067265e-05, |
|
"loss": 0.0716, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 1.6988041853512708e-05, |
|
"loss": 0.0669, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 1.6980568011958147e-05, |
|
"loss": 0.0067, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 1.6973094170403587e-05, |
|
"loss": 0.0241, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 1.696562032884903e-05, |
|
"loss": 0.0068, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 1.6958146487294472e-05, |
|
"loss": 0.0297, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 1.695067264573991e-05, |
|
"loss": 0.0581, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 1.6943198804185354e-05, |
|
"loss": 0.0109, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 1.6935724962630793e-05, |
|
"loss": 0.0059, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 1.6928251121076232e-05, |
|
"loss": 0.0265, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 1.6920777279521675e-05, |
|
"loss": 0.0597, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 1.6913303437967118e-05, |
|
"loss": 0.0061, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 1.6905829596412557e-05, |
|
"loss": 0.0114, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 1.6898355754858e-05, |
|
"loss": 0.0123, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 1.689088191330344e-05, |
|
"loss": 0.0064, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 1.688340807174888e-05, |
|
"loss": 0.0107, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 1.687593423019432e-05, |
|
"loss": 0.0153, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 1.6868460388639763e-05, |
|
"loss": 0.0122, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 1.6860986547085203e-05, |
|
"loss": 0.0212, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 1.6853512705530642e-05, |
|
"loss": 0.0412, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 1.6846038863976085e-05, |
|
"loss": 0.187, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 1.6838565022421527e-05, |
|
"loss": 0.0641, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 1.6831091180866966e-05, |
|
"loss": 0.0902, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 1.682361733931241e-05, |
|
"loss": 0.0073, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 1.681614349775785e-05, |
|
"loss": 0.033, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 1.6808669656203288e-05, |
|
"loss": 0.0104, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 1.680119581464873e-05, |
|
"loss": 0.0122, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 1.6793721973094173e-05, |
|
"loss": 0.0131, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 1.6786248131539612e-05, |
|
"loss": 0.0673, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 1.6778774289985055e-05, |
|
"loss": 0.0271, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 1.6771300448430494e-05, |
|
"loss": 0.0143, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 1.6763826606875933e-05, |
|
"loss": 0.01, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 1.6756352765321376e-05, |
|
"loss": 0.0201, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 1.674887892376682e-05, |
|
"loss": 0.0056, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 1.6741405082212258e-05, |
|
"loss": 0.0618, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 1.67339312406577e-05, |
|
"loss": 0.0141, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 1.672645739910314e-05, |
|
"loss": 0.0118, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 1.671898355754858e-05, |
|
"loss": 0.0105, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 1.6711509715994022e-05, |
|
"loss": 0.0082, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 1.6704035874439464e-05, |
|
"loss": 0.0104, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 1.6696562032884904e-05, |
|
"loss": 0.03, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 1.6689088191330346e-05, |
|
"loss": 0.0099, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 1.6681614349775786e-05, |
|
"loss": 0.0157, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 1.6674140508221225e-05, |
|
"loss": 0.0101, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 1.6666666666666667e-05, |
|
"loss": 0.0048, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 1.665919282511211e-05, |
|
"loss": 0.0157, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 1.665171898355755e-05, |
|
"loss": 0.0271, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 1.6644245142002992e-05, |
|
"loss": 0.0232, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 1.663677130044843e-05, |
|
"loss": 0.0088, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 1.6629297458893874e-05, |
|
"loss": 0.0147, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 1.6621823617339313e-05, |
|
"loss": 0.0054, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 1.6614349775784756e-05, |
|
"loss": 0.0254, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 1.6606875934230195e-05, |
|
"loss": 0.0089, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 1.6599402092675638e-05, |
|
"loss": 0.0425, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 1.6591928251121077e-05, |
|
"loss": 0.0813, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 1.658445440956652e-05, |
|
"loss": 0.029, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 1.657698056801196e-05, |
|
"loss": 0.0117, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 1.65695067264574e-05, |
|
"loss": 0.0259, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 1.656203288490284e-05, |
|
"loss": 0.011, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 1.655455904334828e-05, |
|
"loss": 0.012, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 1.6547085201793723e-05, |
|
"loss": 0.0072, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 1.6539611360239165e-05, |
|
"loss": 0.0224, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 1.6532137518684605e-05, |
|
"loss": 0.0073, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 1.6524663677130047e-05, |
|
"loss": 0.0607, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 1.6517189835575487e-05, |
|
"loss": 0.0037, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 1.6509715994020926e-05, |
|
"loss": 0.0111, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 1.650224215246637e-05, |
|
"loss": 0.0187, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 1.649476831091181e-05, |
|
"loss": 0.0161, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 1.648729446935725e-05, |
|
"loss": 0.0065, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 1.6479820627802693e-05, |
|
"loss": 0.0091, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 1.6472346786248132e-05, |
|
"loss": 0.009, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 1.646487294469357e-05, |
|
"loss": 0.0109, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 1.6457399103139014e-05, |
|
"loss": 0.0117, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 1.6449925261584457e-05, |
|
"loss": 0.0344, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 1.6442451420029896e-05, |
|
"loss": 0.0193, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 1.643497757847534e-05, |
|
"loss": 0.0116, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 1.6427503736920778e-05, |
|
"loss": 0.0151, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 1.6420029895366217e-05, |
|
"loss": 0.0116, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 1.641255605381166e-05, |
|
"loss": 0.0198, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 1.6405082212257103e-05, |
|
"loss": 0.0079, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 1.6397608370702542e-05, |
|
"loss": 0.0096, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 1.6390134529147984e-05, |
|
"loss": 0.0225, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 1.6382660687593424e-05, |
|
"loss": 0.0403, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 1.6375186846038866e-05, |
|
"loss": 0.0078, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 1.6367713004484306e-05, |
|
"loss": 0.0018, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 1.6360239162929748e-05, |
|
"loss": 0.0154, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 1.6352765321375188e-05, |
|
"loss": 0.0192, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 1.634529147982063e-05, |
|
"loss": 0.0578, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 1.633781763826607e-05, |
|
"loss": 0.0641, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 1.6330343796711512e-05, |
|
"loss": 0.0141, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 1.632286995515695e-05, |
|
"loss": 0.0174, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 1.6315396113602394e-05, |
|
"loss": 0.0095, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 1.6307922272047833e-05, |
|
"loss": 0.0068, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 1.6300448430493273e-05, |
|
"loss": 0.0154, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 1.6292974588938715e-05, |
|
"loss": 0.0706, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 1.6285500747384158e-05, |
|
"loss": 0.0065, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 1.6278026905829597e-05, |
|
"loss": 0.0174, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 1.627055306427504e-05, |
|
"loss": 0.0121, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 1.626307922272048e-05, |
|
"loss": 0.0078, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 1.6255605381165918e-05, |
|
"loss": 0.0217, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 1.624813153961136e-05, |
|
"loss": 0.049, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 1.6240657698056804e-05, |
|
"loss": 0.0124, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 1.6233183856502243e-05, |
|
"loss": 0.0085, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 1.6225710014947685e-05, |
|
"loss": 0.0038, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 1.6218236173393125e-05, |
|
"loss": 0.0382, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 1.6210762331838564e-05, |
|
"loss": 0.0235, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 1.6203288490284007e-05, |
|
"loss": 0.0067, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 1.619581464872945e-05, |
|
"loss": 0.0496, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 1.618834080717489e-05, |
|
"loss": 0.0433, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 1.618086696562033e-05, |
|
"loss": 0.0158, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 1.617339312406577e-05, |
|
"loss": 0.0294, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 1.616591928251121e-05, |
|
"loss": 0.0128, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 1.6158445440956652e-05, |
|
"loss": 0.0215, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 1.6150971599402095e-05, |
|
"loss": 0.0075, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 1.6143497757847534e-05, |
|
"loss": 0.0061, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 1.6136023916292977e-05, |
|
"loss": 0.0123, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 1.6128550074738416e-05, |
|
"loss": 0.0613, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 1.612107623318386e-05, |
|
"loss": 0.0182, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 1.6113602391629298e-05, |
|
"loss": 0.0121, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 1.610612855007474e-05, |
|
"loss": 0.0114, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 1.609865470852018e-05, |
|
"loss": 0.0084, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 1.6091180866965623e-05, |
|
"loss": 0.0035, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 1.6083707025411062e-05, |
|
"loss": 0.0063, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 1.6076233183856505e-05, |
|
"loss": 0.0732, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 1.6068759342301944e-05, |
|
"loss": 0.0442, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 1.6061285500747386e-05, |
|
"loss": 0.0308, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 1.6053811659192826e-05, |
|
"loss": 0.008, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 1.6046337817638265e-05, |
|
"loss": 0.0068, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 1.6038863976083708e-05, |
|
"loss": 0.0074, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 1.603139013452915e-05, |
|
"loss": 0.0077, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 1.602391629297459e-05, |
|
"loss": 0.0037, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 1.6016442451420032e-05, |
|
"loss": 0.0081, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 1.600896860986547e-05, |
|
"loss": 0.0736, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 1.600149476831091e-05, |
|
"loss": 0.0096, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 1.5994020926756357e-05, |
|
"loss": 0.004, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 1.5986547085201796e-05, |
|
"loss": 0.0095, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 1.5979073243647235e-05, |
|
"loss": 0.0211, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 1.5971599402092678e-05, |
|
"loss": 0.0105, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 1.5964125560538117e-05, |
|
"loss": 0.0087, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 1.5956651718983556e-05, |
|
"loss": 0.0046, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 1.5949177877429e-05, |
|
"loss": 0.0112, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 1.5941704035874442e-05, |
|
"loss": 0.0809, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 1.593423019431988e-05, |
|
"loss": 0.0059, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 1.5926756352765324e-05, |
|
"loss": 0.0195, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 1.5919282511210763e-05, |
|
"loss": 0.0064, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 1.5911808669656202e-05, |
|
"loss": 0.0082, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 1.5904334828101645e-05, |
|
"loss": 0.0197, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 1.5896860986547088e-05, |
|
"loss": 0.034, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 1.5889387144992527e-05, |
|
"loss": 0.0088, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 1.588191330343797e-05, |
|
"loss": 0.0076, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 1.587443946188341e-05, |
|
"loss": 0.0169, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 1.586696562032885e-05, |
|
"loss": 0.0064, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 1.585949177877429e-05, |
|
"loss": 0.0474, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 1.5852017937219733e-05, |
|
"loss": 0.0137, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 1.5844544095665172e-05, |
|
"loss": 0.0142, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 1.5837070254110615e-05, |
|
"loss": 0.0393, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 1.5829596412556054e-05, |
|
"loss": 0.008, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 1.5822122571001497e-05, |
|
"loss": 0.0087, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 1.5814648729446936e-05, |
|
"loss": 0.0078, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 1.580717488789238e-05, |
|
"loss": 0.0067, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 1.5799701046337818e-05, |
|
"loss": 0.0254, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 1.579222720478326e-05, |
|
"loss": 0.0087, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 1.57847533632287e-05, |
|
"loss": 0.1282, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 1.5777279521674143e-05, |
|
"loss": 0.0153, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 1.5769805680119582e-05, |
|
"loss": 0.0102, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 1.5762331838565025e-05, |
|
"loss": 0.0258, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 1.5754857997010464e-05, |
|
"loss": 0.0055, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 1.5747384155455903e-05, |
|
"loss": 0.008, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 1.573991031390135e-05, |
|
"loss": 0.0115, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 1.573243647234679e-05, |
|
"loss": 0.0089, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 1.5724962630792228e-05, |
|
"loss": 0.0061, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 1.571748878923767e-05, |
|
"loss": 0.0377, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 1.571001494768311e-05, |
|
"loss": 0.0102, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 1.570254110612855e-05, |
|
"loss": 0.0156, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 1.569506726457399e-05, |
|
"loss": 0.0353, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 1.5687593423019434e-05, |
|
"loss": 0.0088, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 1.5680119581464873e-05, |
|
"loss": 0.006, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 1.5672645739910316e-05, |
|
"loss": 0.0198, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 1.5665171898355755e-05, |
|
"loss": 0.0117, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 1.5657698056801195e-05, |
|
"loss": 0.0069, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 1.5650224215246637e-05, |
|
"loss": 0.0097, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 1.564275037369208e-05, |
|
"loss": 0.0097, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 1.563527653213752e-05, |
|
"loss": 0.0127, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 1.5627802690582962e-05, |
|
"loss": 0.0243, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 1.56203288490284e-05, |
|
"loss": 0.0062, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 1.5612855007473844e-05, |
|
"loss": 0.0056, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 1.5605381165919283e-05, |
|
"loss": 0.008, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 1.5597907324364726e-05, |
|
"loss": 0.0081, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 1.5590433482810165e-05, |
|
"loss": 0.0115, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 1.5582959641255608e-05, |
|
"loss": 0.0615, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 1.5575485799701047e-05, |
|
"loss": 0.0088, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 1.556801195814649e-05, |
|
"loss": 0.0067, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 1.556053811659193e-05, |
|
"loss": 0.0072, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 1.555306427503737e-05, |
|
"loss": 0.0066, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 1.554559043348281e-05, |
|
"loss": 0.0052, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 1.5538116591928253e-05, |
|
"loss": 0.0301, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 1.5530642750373693e-05, |
|
"loss": 0.0171, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 1.5523168908819135e-05, |
|
"loss": 0.0083, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 1.5515695067264575e-05, |
|
"loss": 0.0138, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 1.5508221225710017e-05, |
|
"loss": 0.0092, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 1.5500747384155456e-05, |
|
"loss": 0.0047, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.5493273542600896e-05, |
|
"loss": 0.0235, |
|
"step": 631 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.5485799701046342e-05, |
|
"loss": 0.0481, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.547832585949178e-05, |
|
"loss": 0.0129, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.547085201793722e-05, |
|
"loss": 0.013, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.5463378176382663e-05, |
|
"loss": 0.0256, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.5455904334828102e-05, |
|
"loss": 0.0046, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.544843049327354e-05, |
|
"loss": 0.0236, |
|
"step": 637 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.5440956651718984e-05, |
|
"loss": 0.0058, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.5433482810164427e-05, |
|
"loss": 0.0123, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.5426008968609866e-05, |
|
"loss": 0.0049, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 1.541853512705531e-05, |
|
"loss": 0.009, |
|
"step": 641 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 1.5411061285500748e-05, |
|
"loss": 0.0156, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 1.5403587443946187e-05, |
|
"loss": 0.0026, |
|
"step": 643 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 1.539611360239163e-05, |
|
"loss": 0.0942, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 1.5388639760837072e-05, |
|
"loss": 0.0079, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 1.538116591928251e-05, |
|
"loss": 0.0091, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 1.5373692077727954e-05, |
|
"loss": 0.009, |
|
"step": 647 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 1.5366218236173394e-05, |
|
"loss": 0.0098, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 1.5358744394618836e-05, |
|
"loss": 0.0077, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 1.5351270553064276e-05, |
|
"loss": 0.0083, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 1.5343796711509718e-05, |
|
"loss": 0.0068, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 1.5336322869955157e-05, |
|
"loss": 0.0117, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 1.53288490284006e-05, |
|
"loss": 0.008, |
|
"step": 653 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 1.532137518684604e-05, |
|
"loss": 0.0177, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 1.5313901345291482e-05, |
|
"loss": 0.0041, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 1.530642750373692e-05, |
|
"loss": 0.0104, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 1.5298953662182364e-05, |
|
"loss": 0.0213, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 1.5291479820627803e-05, |
|
"loss": 0.0491, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 1.5284005979073246e-05, |
|
"loss": 0.0739, |
|
"step": 659 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 1.5276532137518685e-05, |
|
"loss": 0.0083, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 1.5269058295964128e-05, |
|
"loss": 0.0078, |
|
"step": 661 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 1.5261584454409567e-05, |
|
"loss": 0.0103, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 1.5254110612855008e-05, |
|
"loss": 0.0163, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 1.5246636771300449e-05, |
|
"loss": 0.0066, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 1.523916292974589e-05, |
|
"loss": 0.0083, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 1.5231689088191332e-05, |
|
"loss": 0.0195, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 1.5224215246636773e-05, |
|
"loss": 0.0068, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 1.5216741405082214e-05, |
|
"loss": 0.008, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 1.5209267563527654e-05, |
|
"loss": 0.1062, |
|
"step": 669 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 1.5201793721973095e-05, |
|
"loss": 0.0096, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 1.5194319880418536e-05, |
|
"loss": 0.0138, |
|
"step": 671 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 1.5186846038863978e-05, |
|
"loss": 0.15, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 1.517937219730942e-05, |
|
"loss": 0.0048, |
|
"step": 673 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 1.517189835575486e-05, |
|
"loss": 0.0037, |
|
"step": 674 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 1.51644245142003e-05, |
|
"loss": 0.0078, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 1.515695067264574e-05, |
|
"loss": 0.0067, |
|
"step": 676 |
|
} |
|
], |
|
"logging_steps": 1.0, |
|
"max_steps": 2704, |
|
"num_train_epochs": 8, |
|
"save_steps": 500, |
|
"total_flos": 8.565880284061368e+17, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|