|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.0, |
|
"global_step": 134942, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.981473521957582e-05, |
|
"loss": 1.4713, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.962947043915164e-05, |
|
"loss": 1.48, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.944420565872746e-05, |
|
"loss": 1.5275, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 4.9258940878303276e-05, |
|
"loss": 1.5373, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.907367609787909e-05, |
|
"loss": 1.5363, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 4.8888411317454905e-05, |
|
"loss": 1.5386, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 4.8703146537030724e-05, |
|
"loss": 1.5377, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 4.851788175660654e-05, |
|
"loss": 1.5505, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 4.833261697618236e-05, |
|
"loss": 1.5443, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 4.814735219575818e-05, |
|
"loss": 1.5775, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 4.7962087415334e-05, |
|
"loss": 1.5587, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 4.7776822634909815e-05, |
|
"loss": 1.566, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.7591557854485634e-05, |
|
"loss": 1.5715, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.740629307406145e-05, |
|
"loss": 1.5798, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 4.722102829363727e-05, |
|
"loss": 1.5844, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 4.703576351321309e-05, |
|
"loss": 1.5983, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 4.685049873278891e-05, |
|
"loss": 1.5914, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 4.6665233952364725e-05, |
|
"loss": 1.5988, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 4.6479969171940543e-05, |
|
"loss": 1.5802, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 4.6294704391516355e-05, |
|
"loss": 1.5803, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 4.610943961109217e-05, |
|
"loss": 1.567, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 4.592417483066799e-05, |
|
"loss": 1.612, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 4.573891005024381e-05, |
|
"loss": 1.6182, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 4.555364526981963e-05, |
|
"loss": 1.5818, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 4.5368380489395447e-05, |
|
"loss": 1.602, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 4.5183115708971265e-05, |
|
"loss": 1.5877, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 4.499785092854708e-05, |
|
"loss": 1.5995, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 4.48125861481229e-05, |
|
"loss": 1.6053, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 4.462732136769871e-05, |
|
"loss": 1.5905, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 4.444205658727453e-05, |
|
"loss": 1.5788, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 4.425679180685035e-05, |
|
"loss": 1.6116, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 4.407152702642617e-05, |
|
"loss": 1.602, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 4.3886262246001986e-05, |
|
"loss": 1.6282, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 4.3700997465577804e-05, |
|
"loss": 1.6232, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 4.351573268515363e-05, |
|
"loss": 1.6178, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 4.333046790472944e-05, |
|
"loss": 1.6305, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 4.314520312430526e-05, |
|
"loss": 1.614, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 4.295993834388108e-05, |
|
"loss": 1.6242, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 4.2774673563456896e-05, |
|
"loss": 1.6125, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 4.2589408783032714e-05, |
|
"loss": 1.6274, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 4.240414400260853e-05, |
|
"loss": 1.6265, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 4.221887922218435e-05, |
|
"loss": 1.6125, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 4.203361444176017e-05, |
|
"loss": 1.6225, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.184834966133598e-05, |
|
"loss": 1.631, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 4.16630848809118e-05, |
|
"loss": 1.6501, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.147782010048762e-05, |
|
"loss": 1.6484, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 4.1292555320063436e-05, |
|
"loss": 1.6428, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.1107290539639254e-05, |
|
"loss": 1.6506, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 4.092202575921507e-05, |
|
"loss": 1.6386, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 4.073676097879089e-05, |
|
"loss": 1.6137, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 4.055149619836671e-05, |
|
"loss": 1.6623, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 4.036623141794252e-05, |
|
"loss": 1.6357, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 4.018096663751834e-05, |
|
"loss": 1.6643, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 3.999570185709416e-05, |
|
"loss": 1.6583, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 3.9810437076669975e-05, |
|
"loss": 1.6552, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 3.9625172296245794e-05, |
|
"loss": 1.6485, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 3.943990751582161e-05, |
|
"loss": 1.6431, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 3.925464273539744e-05, |
|
"loss": 1.6368, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 3.906937795497325e-05, |
|
"loss": 1.6616, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 3.888411317454907e-05, |
|
"loss": 1.6341, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 3.8698848394124885e-05, |
|
"loss": 1.6563, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 3.8513583613700704e-05, |
|
"loss": 1.6536, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 3.832831883327652e-05, |
|
"loss": 1.6615, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 3.814305405285234e-05, |
|
"loss": 1.6548, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 3.795778927242816e-05, |
|
"loss": 1.6732, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 3.777252449200398e-05, |
|
"loss": 1.6839, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 3.7587259711579795e-05, |
|
"loss": 1.6735, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 3.7401994931155607e-05, |
|
"loss": 1.672, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 3.7216730150731425e-05, |
|
"loss": 1.6569, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 3.703146537030724e-05, |
|
"loss": 1.6462, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 3.684620058988306e-05, |
|
"loss": 1.6683, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 3.666093580945888e-05, |
|
"loss": 1.6682, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 3.64756710290347e-05, |
|
"loss": 1.6598, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 3.6290406248610516e-05, |
|
"loss": 1.6505, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 3.6105141468186335e-05, |
|
"loss": 1.6535, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 3.5919876687762146e-05, |
|
"loss": 1.6637, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 3.5734611907337965e-05, |
|
"loss": 1.6534, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 3.554934712691378e-05, |
|
"loss": 1.6512, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.53640823464896e-05, |
|
"loss": 1.6602, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 3.5178817566065426e-05, |
|
"loss": 1.6802, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 3.4993552785641245e-05, |
|
"loss": 1.6637, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 3.480828800521706e-05, |
|
"loss": 1.6669, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.4623023224792874e-05, |
|
"loss": 1.6595, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 3.443775844436869e-05, |
|
"loss": 1.6746, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.425249366394451e-05, |
|
"loss": 1.6679, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.406722888352033e-05, |
|
"loss": 1.6596, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.388196410309615e-05, |
|
"loss": 1.6818, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 3.3696699322671966e-05, |
|
"loss": 1.6862, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 3.3511434542247784e-05, |
|
"loss": 1.6578, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.33261697618236e-05, |
|
"loss": 1.6567, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 3.314090498139942e-05, |
|
"loss": 1.6701, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 3.295564020097523e-05, |
|
"loss": 1.6861, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 3.277037542055105e-05, |
|
"loss": 1.6905, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.258511064012687e-05, |
|
"loss": 1.6935, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 3.239984585970269e-05, |
|
"loss": 1.6464, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 3.2214581079278506e-05, |
|
"loss": 1.6827, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.2029316298854324e-05, |
|
"loss": 1.6565, |
|
"step": 48500 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.184405151843014e-05, |
|
"loss": 1.6695, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.165878673800596e-05, |
|
"loss": 1.6949, |
|
"step": 49500 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.147352195758177e-05, |
|
"loss": 1.6847, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.128825717715759e-05, |
|
"loss": 1.6726, |
|
"step": 50500 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.110299239673341e-05, |
|
"loss": 1.6729, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 3.0917727616309234e-05, |
|
"loss": 1.7076, |
|
"step": 51500 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 3.073246283588505e-05, |
|
"loss": 1.7006, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 3.054719805546087e-05, |
|
"loss": 1.6812, |
|
"step": 52500 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 3.0361933275036685e-05, |
|
"loss": 1.6976, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 3.0176668494612504e-05, |
|
"loss": 1.6668, |
|
"step": 53500 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.9991403714188322e-05, |
|
"loss": 1.7091, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 2.9806138933764137e-05, |
|
"loss": 1.6627, |
|
"step": 54500 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 2.9620874153339955e-05, |
|
"loss": 1.7057, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 2.9435609372915773e-05, |
|
"loss": 1.6779, |
|
"step": 55500 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 2.9250344592491592e-05, |
|
"loss": 1.6818, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 2.9065079812067407e-05, |
|
"loss": 1.692, |
|
"step": 56500 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 2.8879815031643225e-05, |
|
"loss": 1.6898, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 2.8694550251219043e-05, |
|
"loss": 1.6998, |
|
"step": 57500 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 2.850928547079486e-05, |
|
"loss": 1.691, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 2.832402069037068e-05, |
|
"loss": 1.6854, |
|
"step": 58500 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 2.8138755909946495e-05, |
|
"loss": 1.6901, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.7953491129522313e-05, |
|
"loss": 1.6782, |
|
"step": 59500 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.776822634909813e-05, |
|
"loss": 1.6921, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.758296156867395e-05, |
|
"loss": 1.6973, |
|
"step": 60500 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.7397696788249765e-05, |
|
"loss": 1.6879, |
|
"step": 61000 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 2.7212432007825583e-05, |
|
"loss": 1.6839, |
|
"step": 61500 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 2.70271672274014e-05, |
|
"loss": 1.7067, |
|
"step": 62000 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.684190244697722e-05, |
|
"loss": 1.6797, |
|
"step": 62500 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.665663766655304e-05, |
|
"loss": 1.7041, |
|
"step": 63000 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.647137288612886e-05, |
|
"loss": 1.6987, |
|
"step": 63500 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 2.6286108105704678e-05, |
|
"loss": 1.7034, |
|
"step": 64000 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 2.6100843325280493e-05, |
|
"loss": 1.694, |
|
"step": 64500 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 2.591557854485631e-05, |
|
"loss": 1.7114, |
|
"step": 65000 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 2.573031376443213e-05, |
|
"loss": 1.7016, |
|
"step": 65500 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.5545048984007948e-05, |
|
"loss": 1.6879, |
|
"step": 66000 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.5359784203583763e-05, |
|
"loss": 1.6962, |
|
"step": 66500 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.517451942315958e-05, |
|
"loss": 1.7212, |
|
"step": 67000 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 2.49892546427354e-05, |
|
"loss": 1.6942, |
|
"step": 67500 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 2.4803989862311218e-05, |
|
"loss": 1.4384, |
|
"step": 68000 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 2.4618725081887032e-05, |
|
"loss": 1.458, |
|
"step": 68500 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 2.443346030146285e-05, |
|
"loss": 1.4683, |
|
"step": 69000 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 2.424819552103867e-05, |
|
"loss": 1.4516, |
|
"step": 69500 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 2.4062930740614487e-05, |
|
"loss": 1.4581, |
|
"step": 70000 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 2.3877665960190306e-05, |
|
"loss": 1.4695, |
|
"step": 70500 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 2.369240117976612e-05, |
|
"loss": 1.4513, |
|
"step": 71000 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 2.3507136399341942e-05, |
|
"loss": 1.4649, |
|
"step": 71500 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 2.332187161891776e-05, |
|
"loss": 1.4755, |
|
"step": 72000 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 2.3136606838493576e-05, |
|
"loss": 1.4694, |
|
"step": 72500 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 2.2951342058069394e-05, |
|
"loss": 1.4587, |
|
"step": 73000 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 2.2766077277645212e-05, |
|
"loss": 1.4701, |
|
"step": 73500 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 2.258081249722103e-05, |
|
"loss": 1.4596, |
|
"step": 74000 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 2.2395547716796845e-05, |
|
"loss": 1.4757, |
|
"step": 74500 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 2.2210282936372664e-05, |
|
"loss": 1.4876, |
|
"step": 75000 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 2.2025018155948482e-05, |
|
"loss": 1.4819, |
|
"step": 75500 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 2.18397533755243e-05, |
|
"loss": 1.4758, |
|
"step": 76000 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 2.165448859510012e-05, |
|
"loss": 1.4917, |
|
"step": 76500 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 2.1469223814675937e-05, |
|
"loss": 1.4939, |
|
"step": 77000 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 2.1283959034251755e-05, |
|
"loss": 1.509, |
|
"step": 77500 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 2.1098694253827574e-05, |
|
"loss": 1.4854, |
|
"step": 78000 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 2.091342947340339e-05, |
|
"loss": 1.5043, |
|
"step": 78500 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 2.0728164692979207e-05, |
|
"loss": 1.5026, |
|
"step": 79000 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 2.0542899912555025e-05, |
|
"loss": 1.4936, |
|
"step": 79500 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 2.0357635132130843e-05, |
|
"loss": 1.5001, |
|
"step": 80000 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 2.0172370351706658e-05, |
|
"loss": 1.4855, |
|
"step": 80500 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 1.9987105571282477e-05, |
|
"loss": 1.4758, |
|
"step": 81000 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 1.9801840790858295e-05, |
|
"loss": 1.496, |
|
"step": 81500 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 1.9616576010434113e-05, |
|
"loss": 1.4905, |
|
"step": 82000 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 1.9431311230009928e-05, |
|
"loss": 1.4971, |
|
"step": 82500 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 1.924604644958575e-05, |
|
"loss": 1.4487, |
|
"step": 83000 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.9060781669161568e-05, |
|
"loss": 1.4366, |
|
"step": 83500 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.8875516888737386e-05, |
|
"loss": 1.4482, |
|
"step": 84000 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.86902521083132e-05, |
|
"loss": 1.4364, |
|
"step": 84500 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 1.850498732788902e-05, |
|
"loss": 1.4334, |
|
"step": 85000 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 1.8319722547464838e-05, |
|
"loss": 1.45, |
|
"step": 85500 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 1.8134457767040656e-05, |
|
"loss": 1.4507, |
|
"step": 86000 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 1.794919298661647e-05, |
|
"loss": 1.4476, |
|
"step": 86500 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 1.776392820619229e-05, |
|
"loss": 1.4402, |
|
"step": 87000 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 1.7578663425768108e-05, |
|
"loss": 1.4357, |
|
"step": 87500 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 1.7393398645343926e-05, |
|
"loss": 1.4292, |
|
"step": 88000 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 1.7208133864919744e-05, |
|
"loss": 1.4501, |
|
"step": 88500 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 1.7022869084495563e-05, |
|
"loss": 1.4635, |
|
"step": 89000 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 1.683760430407138e-05, |
|
"loss": 1.4629, |
|
"step": 89500 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 1.66523395236472e-05, |
|
"loss": 1.4496, |
|
"step": 90000 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 1.6467074743223014e-05, |
|
"loss": 1.5029, |
|
"step": 90500 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 1.6281809962798833e-05, |
|
"loss": 1.4813, |
|
"step": 91000 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 1.609654518237465e-05, |
|
"loss": 1.4797, |
|
"step": 91500 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 1.591128040195047e-05, |
|
"loss": 1.4808, |
|
"step": 92000 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 1.5726015621526284e-05, |
|
"loss": 1.4953, |
|
"step": 92500 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 1.5540750841102102e-05, |
|
"loss": 1.4928, |
|
"step": 93000 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 1.535548606067792e-05, |
|
"loss": 1.4918, |
|
"step": 93500 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 1.5170221280253737e-05, |
|
"loss": 1.4859, |
|
"step": 94000 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 1.4984956499829559e-05, |
|
"loss": 1.5046, |
|
"step": 94500 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 1.4799691719405376e-05, |
|
"loss": 1.4953, |
|
"step": 95000 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 1.4614426938981194e-05, |
|
"loss": 1.5017, |
|
"step": 95500 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 1.442916215855701e-05, |
|
"loss": 1.4902, |
|
"step": 96000 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 1.4243897378132829e-05, |
|
"loss": 1.5008, |
|
"step": 96500 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 1.4058632597708645e-05, |
|
"loss": 1.5013, |
|
"step": 97000 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 1.3873367817284464e-05, |
|
"loss": 1.4889, |
|
"step": 97500 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 1.368810303686028e-05, |
|
"loss": 1.4879, |
|
"step": 98000 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 1.3502838256436099e-05, |
|
"loss": 1.5108, |
|
"step": 98500 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 1.3317573476011915e-05, |
|
"loss": 1.5042, |
|
"step": 99000 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 1.3132308695587734e-05, |
|
"loss": 1.5153, |
|
"step": 99500 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 1.2947043915163554e-05, |
|
"loss": 1.5044, |
|
"step": 100000 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 1.2761779134739372e-05, |
|
"loss": 1.5163, |
|
"step": 100500 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 1.2576514354315189e-05, |
|
"loss": 1.5069, |
|
"step": 101000 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 1.2391249573891007e-05, |
|
"loss": 1.4993, |
|
"step": 101500 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 1.2205984793466823e-05, |
|
"loss": 1.5073, |
|
"step": 102000 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 1.2020720013042642e-05, |
|
"loss": 1.5076, |
|
"step": 102500 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 1.1835455232618458e-05, |
|
"loss": 1.5086, |
|
"step": 103000 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 1.1650190452194277e-05, |
|
"loss": 1.535, |
|
"step": 103500 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 1.1464925671770093e-05, |
|
"loss": 1.5124, |
|
"step": 104000 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 1.1279660891345913e-05, |
|
"loss": 1.5062, |
|
"step": 104500 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 1.109439611092173e-05, |
|
"loss": 1.5339, |
|
"step": 105000 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 1.0909131330497548e-05, |
|
"loss": 1.5385, |
|
"step": 105500 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 1.0723866550073365e-05, |
|
"loss": 1.5343, |
|
"step": 106000 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 1.0538601769649183e-05, |
|
"loss": 1.5351, |
|
"step": 106500 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 1.0353336989225001e-05, |
|
"loss": 1.5182, |
|
"step": 107000 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 1.016807220880082e-05, |
|
"loss": 1.5242, |
|
"step": 107500 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 9.982807428376636e-06, |
|
"loss": 1.5444, |
|
"step": 108000 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 9.797542647952455e-06, |
|
"loss": 1.533, |
|
"step": 108500 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 9.612277867528271e-06, |
|
"loss": 1.5202, |
|
"step": 109000 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 9.42701308710409e-06, |
|
"loss": 1.5468, |
|
"step": 109500 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 9.241748306679908e-06, |
|
"loss": 1.5181, |
|
"step": 110000 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 9.056483526255726e-06, |
|
"loss": 1.5231, |
|
"step": 110500 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 8.871218745831543e-06, |
|
"loss": 1.5209, |
|
"step": 111000 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 8.685953965407361e-06, |
|
"loss": 1.52, |
|
"step": 111500 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 8.500689184983178e-06, |
|
"loss": 1.5131, |
|
"step": 112000 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 8.315424404558996e-06, |
|
"loss": 1.5404, |
|
"step": 112500 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 8.130159624134814e-06, |
|
"loss": 1.5523, |
|
"step": 113000 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 7.944894843710633e-06, |
|
"loss": 1.5608, |
|
"step": 113500 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 7.75963006328645e-06, |
|
"loss": 1.5172, |
|
"step": 114000 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 7.574365282862267e-06, |
|
"loss": 1.5192, |
|
"step": 114500 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 7.389100502438084e-06, |
|
"loss": 1.5329, |
|
"step": 115000 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 7.203835722013903e-06, |
|
"loss": 1.5195, |
|
"step": 115500 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 7.018570941589721e-06, |
|
"loss": 1.5377, |
|
"step": 116000 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 6.833306161165538e-06, |
|
"loss": 1.5332, |
|
"step": 116500 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 6.648041380741356e-06, |
|
"loss": 1.546, |
|
"step": 117000 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 6.462776600317173e-06, |
|
"loss": 1.5475, |
|
"step": 117500 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 6.277511819892991e-06, |
|
"loss": 1.5593, |
|
"step": 118000 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 6.092247039468809e-06, |
|
"loss": 1.5384, |
|
"step": 118500 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 5.906982259044627e-06, |
|
"loss": 1.5448, |
|
"step": 119000 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 5.721717478620445e-06, |
|
"loss": 1.5546, |
|
"step": 119500 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 5.536452698196262e-06, |
|
"loss": 1.5641, |
|
"step": 120000 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 5.3511879177720805e-06, |
|
"loss": 1.5636, |
|
"step": 120500 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 5.165923137347898e-06, |
|
"loss": 1.5474, |
|
"step": 121000 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 4.980658356923715e-06, |
|
"loss": 1.5317, |
|
"step": 121500 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 4.795393576499534e-06, |
|
"loss": 1.5344, |
|
"step": 122000 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 4.610128796075351e-06, |
|
"loss": 1.5568, |
|
"step": 122500 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 4.424864015651169e-06, |
|
"loss": 1.5598, |
|
"step": 123000 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 4.239599235226987e-06, |
|
"loss": 1.559, |
|
"step": 123500 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 4.054334454802804e-06, |
|
"loss": 1.5571, |
|
"step": 124000 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 3.869069674378622e-06, |
|
"loss": 1.5597, |
|
"step": 124500 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 3.68380489395444e-06, |
|
"loss": 1.5629, |
|
"step": 125000 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 3.4985401135302576e-06, |
|
"loss": 1.5573, |
|
"step": 125500 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 3.313275333106075e-06, |
|
"loss": 1.5486, |
|
"step": 126000 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 3.1280105526818934e-06, |
|
"loss": 1.5755, |
|
"step": 126500 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 2.942745772257711e-06, |
|
"loss": 1.5626, |
|
"step": 127000 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 2.7574809918335287e-06, |
|
"loss": 1.5385, |
|
"step": 127500 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 2.5722162114093466e-06, |
|
"loss": 1.5594, |
|
"step": 128000 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 2.386951430985164e-06, |
|
"loss": 1.5513, |
|
"step": 128500 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 2.201686650560982e-06, |
|
"loss": 1.5725, |
|
"step": 129000 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 2.0164218701368e-06, |
|
"loss": 1.5727, |
|
"step": 129500 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 1.8311570897126173e-06, |
|
"loss": 1.6023, |
|
"step": 130000 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 1.645892309288435e-06, |
|
"loss": 1.5643, |
|
"step": 130500 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 1.4606275288642529e-06, |
|
"loss": 1.5606, |
|
"step": 131000 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 1.2753627484400705e-06, |
|
"loss": 1.5998, |
|
"step": 131500 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 1.0900979680158882e-06, |
|
"loss": 1.5882, |
|
"step": 132000 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 9.048331875917061e-07, |
|
"loss": 1.59, |
|
"step": 132500 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 7.195684071675239e-07, |
|
"loss": 1.5851, |
|
"step": 133000 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 5.343036267433415e-07, |
|
"loss": 1.6012, |
|
"step": 133500 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 3.490388463191593e-07, |
|
"loss": 1.5951, |
|
"step": 134000 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 1.637740658949771e-07, |
|
"loss": 1.6155, |
|
"step": 134500 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"step": 134942, |
|
"total_flos": 2.5064145824199475e+17, |
|
"train_loss": 1.578441581144295, |
|
"train_runtime": 31772.5798, |
|
"train_samples_per_second": 8.494, |
|
"train_steps_per_second": 4.247 |
|
} |
|
], |
|
"max_steps": 134942, |
|
"num_train_epochs": 2, |
|
"total_flos": 2.5064145824199475e+17, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|