|
{ |
|
"best_metric": 0.9828028941855711, |
|
"best_model_checkpoint": "../results/SersegformerForSemanticSegmentation_decoder/checkpoint-27500", |
|
"epoch": 19.642857142857142, |
|
"eval_steps": 500, |
|
"global_step": 27500, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 4e-06, |
|
"loss": 0.0169, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 8e-06, |
|
"loss": 0.0171, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.2e-05, |
|
"loss": 0.0182, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.6e-05, |
|
"loss": 0.0173, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.9999999999999998e-05, |
|
"loss": 0.017, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"eval_iou": 0.9819581976123885, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 2.4e-05, |
|
"loss": 0.0169, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 2.8e-05, |
|
"loss": 0.0163, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 3.2e-05, |
|
"loss": 0.018, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 3.6e-05, |
|
"loss": 0.0175, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 3.9999999999999996e-05, |
|
"loss": 0.0167, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"eval_iou": 0.9816833290399289, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 4.4e-05, |
|
"loss": 0.0174, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 4.8e-05, |
|
"loss": 0.0171, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 5.2000000000000004e-05, |
|
"loss": 0.0169, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 5.6e-05, |
|
"loss": 0.0171, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 6e-05, |
|
"loss": 0.0175, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"eval_iou": 0.9816984183648053, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 5.98520987654321e-05, |
|
"loss": 0.0178, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 5.970419753086419e-05, |
|
"loss": 0.0184, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 5.95562962962963e-05, |
|
"loss": 0.018, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 5.94083950617284e-05, |
|
"loss": 0.0186, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 5.926049382716049e-05, |
|
"loss": 0.0205, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"eval_iou": 0.9816799380031287, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 5.911259259259259e-05, |
|
"loss": 0.0194, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 5.8964691358024696e-05, |
|
"loss": 0.018, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 5.8816790123456795e-05, |
|
"loss": 0.0192, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 5.866888888888889e-05, |
|
"loss": 0.0186, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 5.852098765432099e-05, |
|
"loss": 0.0179, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"eval_iou": 0.9817095761205635, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 5.837308641975309e-05, |
|
"loss": 0.0195, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 5.8225185185185185e-05, |
|
"loss": 0.0189, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 5.8077283950617285e-05, |
|
"loss": 0.0187, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 5.7929382716049384e-05, |
|
"loss": 0.0179, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 5.778148148148149e-05, |
|
"loss": 0.018, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"eval_iou": 0.9814854831087823, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 5.763358024691358e-05, |
|
"loss": 0.0192, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 5.748567901234568e-05, |
|
"loss": 0.0168, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 5.733777777777778e-05, |
|
"loss": 0.0179, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 5.718987654320987e-05, |
|
"loss": 0.0178, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 5.704197530864198e-05, |
|
"loss": 0.0182, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"eval_iou": 0.9810651301168928, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 5.689407407407408e-05, |
|
"loss": 0.0181, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 5.674617283950617e-05, |
|
"loss": 0.019, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 5.659827160493827e-05, |
|
"loss": 0.0181, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 5.6450370370370376e-05, |
|
"loss": 0.0179, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 5.630246913580247e-05, |
|
"loss": 0.0181, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"eval_iou": 0.9813077029059915, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 5.615456790123457e-05, |
|
"loss": 0.0176, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 5.600666666666667e-05, |
|
"loss": 0.0178, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"learning_rate": 5.585876543209877e-05, |
|
"loss": 0.0185, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 5.5710864197530865e-05, |
|
"loss": 0.0172, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 5.5562962962962965e-05, |
|
"loss": 0.0169, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"eval_iou": 0.981582362277835, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"learning_rate": 5.5415061728395064e-05, |
|
"loss": 0.0174, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 5.526716049382716e-05, |
|
"loss": 0.0188, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"learning_rate": 5.511925925925926e-05, |
|
"loss": 0.0175, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 5.497135802469136e-05, |
|
"loss": 0.0177, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"learning_rate": 5.482345679012346e-05, |
|
"loss": 0.0179, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"eval_iou": 0.981942740725536, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 3.64, |
|
"learning_rate": 5.467555555555555e-05, |
|
"loss": 0.0195, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"learning_rate": 5.452765432098766e-05, |
|
"loss": 0.0202, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 3.79, |
|
"learning_rate": 5.437975308641975e-05, |
|
"loss": 0.0198, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 3.86, |
|
"learning_rate": 5.423185185185185e-05, |
|
"loss": 0.0198, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 3.93, |
|
"learning_rate": 5.408395061728395e-05, |
|
"loss": 0.0177, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 3.93, |
|
"eval_iou": 0.9819124785124087, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 5.3936049382716056e-05, |
|
"loss": 0.0197, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 4.07, |
|
"learning_rate": 5.378814814814815e-05, |
|
"loss": 0.0186, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 4.14, |
|
"learning_rate": 5.364024691358025e-05, |
|
"loss": 0.018, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 4.21, |
|
"learning_rate": 5.349234567901235e-05, |
|
"loss": 0.0188, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 4.29, |
|
"learning_rate": 5.3344444444444446e-05, |
|
"loss": 0.0187, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 4.29, |
|
"eval_iou": 0.9817866309016359, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 4.36, |
|
"learning_rate": 5.3196543209876545e-05, |
|
"loss": 0.019, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 4.43, |
|
"learning_rate": 5.3048641975308645e-05, |
|
"loss": 0.018, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 4.5, |
|
"learning_rate": 5.2900740740740744e-05, |
|
"loss": 0.0193, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 4.57, |
|
"learning_rate": 5.2752839506172836e-05, |
|
"loss": 0.0187, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 4.64, |
|
"learning_rate": 5.260493827160494e-05, |
|
"loss": 0.0182, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 4.64, |
|
"eval_iou": 0.9818342994241154, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 4.71, |
|
"learning_rate": 5.245703703703704e-05, |
|
"loss": 0.0184, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 4.79, |
|
"learning_rate": 5.2309135802469134e-05, |
|
"loss": 0.0182, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 4.86, |
|
"learning_rate": 5.216123456790123e-05, |
|
"loss": 0.0175, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 4.93, |
|
"learning_rate": 5.201333333333334e-05, |
|
"loss": 0.0178, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 5.186543209876544e-05, |
|
"loss": 0.0178, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_iou": 0.9820612844298867, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_iou": 0.9820612844298867, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 5.07, |
|
"learning_rate": 5.171753086419753e-05, |
|
"loss": 0.018, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 5.14, |
|
"learning_rate": 5.156962962962963e-05, |
|
"loss": 0.0196, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 5.21, |
|
"learning_rate": 5.142172839506173e-05, |
|
"loss": 0.0198, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 5.29, |
|
"learning_rate": 5.127382716049383e-05, |
|
"loss": 0.018, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 5.36, |
|
"learning_rate": 5.112592592592593e-05, |
|
"loss": 0.0191, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 5.36, |
|
"eval_iou": 0.9817404121744866, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 5.43, |
|
"learning_rate": 5.097802469135803e-05, |
|
"loss": 0.0174, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 5.5, |
|
"learning_rate": 5.0830123456790126e-05, |
|
"loss": 0.0174, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 5.57, |
|
"learning_rate": 5.0682222222222225e-05, |
|
"loss": 0.019, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 5.64, |
|
"learning_rate": 5.0534320987654324e-05, |
|
"loss": 0.0186, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 5.71, |
|
"learning_rate": 5.038641975308642e-05, |
|
"loss": 0.018, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 5.71, |
|
"eval_iou": 0.981840242357815, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 5.79, |
|
"learning_rate": 5.0238518518518516e-05, |
|
"loss": 0.0174, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 5.86, |
|
"learning_rate": 5.009061728395062e-05, |
|
"loss": 0.0179, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 5.93, |
|
"learning_rate": 4.994271604938272e-05, |
|
"loss": 0.0182, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"learning_rate": 4.9794814814814814e-05, |
|
"loss": 0.0181, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 6.07, |
|
"learning_rate": 4.964691358024691e-05, |
|
"loss": 0.0167, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 6.07, |
|
"eval_iou": 0.9818143762794196, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 6.14, |
|
"learning_rate": 4.949901234567901e-05, |
|
"loss": 0.0173, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 6.21, |
|
"learning_rate": 4.935111111111111e-05, |
|
"loss": 0.0184, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 6.29, |
|
"learning_rate": 4.920320987654321e-05, |
|
"loss": 0.0188, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 6.36, |
|
"learning_rate": 4.905530864197531e-05, |
|
"loss": 0.0183, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 6.43, |
|
"learning_rate": 4.8907407407407416e-05, |
|
"loss": 0.0176, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 6.43, |
|
"eval_iou": 0.9819974823325288, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 6.5, |
|
"learning_rate": 4.875950617283951e-05, |
|
"loss": 0.0195, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 6.57, |
|
"learning_rate": 4.861160493827161e-05, |
|
"loss": 0.0176, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 6.64, |
|
"learning_rate": 4.84637037037037e-05, |
|
"loss": 0.0186, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 6.71, |
|
"learning_rate": 4.8315802469135806e-05, |
|
"loss": 0.0194, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 6.79, |
|
"learning_rate": 4.8167901234567905e-05, |
|
"loss": 0.0179, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 6.79, |
|
"eval_iou": 0.9821633836802315, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 6.86, |
|
"learning_rate": 4.8020000000000004e-05, |
|
"loss": 0.0167, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 6.93, |
|
"learning_rate": 4.78720987654321e-05, |
|
"loss": 0.0175, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"learning_rate": 4.7724197530864196e-05, |
|
"loss": 0.0171, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 7.07, |
|
"learning_rate": 4.75762962962963e-05, |
|
"loss": 0.0191, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 7.14, |
|
"learning_rate": 4.7428395061728395e-05, |
|
"loss": 0.0188, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 7.14, |
|
"eval_iou": 0.9819790539788265, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 7.21, |
|
"learning_rate": 4.7280493827160494e-05, |
|
"loss": 0.0179, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 7.29, |
|
"learning_rate": 4.713259259259259e-05, |
|
"loss": 0.0171, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 7.36, |
|
"learning_rate": 4.69846913580247e-05, |
|
"loss": 0.0178, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 7.43, |
|
"learning_rate": 4.683679012345679e-05, |
|
"loss": 0.0179, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"learning_rate": 4.668888888888889e-05, |
|
"loss": 0.0183, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 7.5, |
|
"eval_iou": 0.9820494610889285, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 7.57, |
|
"learning_rate": 4.654098765432098e-05, |
|
"loss": 0.0184, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 7.64, |
|
"learning_rate": 4.639308641975309e-05, |
|
"loss": 0.0186, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 7.71, |
|
"learning_rate": 4.624518518518519e-05, |
|
"loss": 0.0178, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 7.79, |
|
"learning_rate": 4.609728395061729e-05, |
|
"loss": 0.0176, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 7.86, |
|
"learning_rate": 4.594938271604939e-05, |
|
"loss": 0.0175, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 7.86, |
|
"eval_iou": 0.9821178334600785, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 7.93, |
|
"learning_rate": 4.5801481481481486e-05, |
|
"loss": 0.019, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"learning_rate": 4.5653580246913585e-05, |
|
"loss": 0.0181, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 8.07, |
|
"learning_rate": 4.550567901234568e-05, |
|
"loss": 0.0176, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 8.14, |
|
"learning_rate": 4.535777777777778e-05, |
|
"loss": 0.0171, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 8.21, |
|
"learning_rate": 4.5209876543209876e-05, |
|
"loss": 0.0178, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 8.21, |
|
"eval_iou": 0.9818687801267586, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 8.29, |
|
"learning_rate": 4.506197530864198e-05, |
|
"loss": 0.0173, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 8.36, |
|
"learning_rate": 4.491407407407408e-05, |
|
"loss": 0.0179, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 8.43, |
|
"learning_rate": 4.4766172839506174e-05, |
|
"loss": 0.0182, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 8.5, |
|
"learning_rate": 4.4618271604938266e-05, |
|
"loss": 0.0185, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 8.57, |
|
"learning_rate": 4.447037037037037e-05, |
|
"loss": 0.0175, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 8.57, |
|
"eval_iou": 0.9821109941192702, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 8.64, |
|
"learning_rate": 4.432246913580247e-05, |
|
"loss": 0.0171, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 8.71, |
|
"learning_rate": 4.417456790123457e-05, |
|
"loss": 0.0178, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 8.79, |
|
"learning_rate": 4.402666666666667e-05, |
|
"loss": 0.0179, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 8.86, |
|
"learning_rate": 4.387876543209877e-05, |
|
"loss": 0.0178, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 8.93, |
|
"learning_rate": 4.373086419753087e-05, |
|
"loss": 0.0177, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 8.93, |
|
"eval_iou": 0.9821157449600744, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"learning_rate": 4.358296296296296e-05, |
|
"loss": 0.0171, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 9.07, |
|
"learning_rate": 4.343506172839506e-05, |
|
"loss": 0.0172, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 9.14, |
|
"learning_rate": 4.3287160493827166e-05, |
|
"loss": 0.0183, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 9.21, |
|
"learning_rate": 4.3139259259259265e-05, |
|
"loss": 0.0184, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 9.29, |
|
"learning_rate": 4.2991358024691364e-05, |
|
"loss": 0.0168, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 9.29, |
|
"eval_iou": 0.982087265627057, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 9.36, |
|
"learning_rate": 4.284345679012346e-05, |
|
"loss": 0.0186, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 9.43, |
|
"learning_rate": 4.2695555555555556e-05, |
|
"loss": 0.0167, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 9.5, |
|
"learning_rate": 4.2547654320987655e-05, |
|
"loss": 0.0192, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 9.57, |
|
"learning_rate": 4.2399753086419754e-05, |
|
"loss": 0.017, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 9.64, |
|
"learning_rate": 4.2251851851851854e-05, |
|
"loss": 0.0174, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 9.64, |
|
"eval_iou": 0.9819462275972554, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 9.71, |
|
"learning_rate": 4.210395061728395e-05, |
|
"loss": 0.017, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 9.79, |
|
"learning_rate": 4.195604938271605e-05, |
|
"loss": 0.0182, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 9.86, |
|
"learning_rate": 4.180814814814815e-05, |
|
"loss": 0.0179, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 9.93, |
|
"learning_rate": 4.1660246913580244e-05, |
|
"loss": 0.018, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 4.151234567901234e-05, |
|
"loss": 0.0179, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_iou": 0.9823968813699835, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_iou": 0.9823968813699835, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 10.07, |
|
"learning_rate": 4.136444444444445e-05, |
|
"loss": 0.0176, |
|
"step": 14100 |
|
}, |
|
{ |
|
"epoch": 10.14, |
|
"learning_rate": 4.121654320987655e-05, |
|
"loss": 0.0171, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 10.21, |
|
"learning_rate": 4.106864197530865e-05, |
|
"loss": 0.0178, |
|
"step": 14300 |
|
}, |
|
{ |
|
"epoch": 10.29, |
|
"learning_rate": 4.092074074074074e-05, |
|
"loss": 0.0181, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 10.36, |
|
"learning_rate": 4.077283950617284e-05, |
|
"loss": 0.0187, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 10.36, |
|
"eval_iou": 0.9823005129309261, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 10.43, |
|
"learning_rate": 4.062493827160494e-05, |
|
"loss": 0.0189, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 10.5, |
|
"learning_rate": 4.047703703703704e-05, |
|
"loss": 0.0172, |
|
"step": 14700 |
|
}, |
|
{ |
|
"epoch": 10.57, |
|
"learning_rate": 4.032913580246914e-05, |
|
"loss": 0.0175, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 10.64, |
|
"learning_rate": 4.0181234567901236e-05, |
|
"loss": 0.0179, |
|
"step": 14900 |
|
}, |
|
{ |
|
"epoch": 10.71, |
|
"learning_rate": 4.003333333333334e-05, |
|
"loss": 0.0167, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 10.71, |
|
"eval_iou": 0.9822724671924815, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 10.79, |
|
"learning_rate": 3.9885432098765434e-05, |
|
"loss": 0.0175, |
|
"step": 15100 |
|
}, |
|
{ |
|
"epoch": 10.86, |
|
"learning_rate": 3.973753086419753e-05, |
|
"loss": 0.0177, |
|
"step": 15200 |
|
}, |
|
{ |
|
"epoch": 10.93, |
|
"learning_rate": 3.9589629629629626e-05, |
|
"loss": 0.0184, |
|
"step": 15300 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"learning_rate": 3.944172839506173e-05, |
|
"loss": 0.0174, |
|
"step": 15400 |
|
}, |
|
{ |
|
"epoch": 11.07, |
|
"learning_rate": 3.929382716049383e-05, |
|
"loss": 0.0174, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 11.07, |
|
"eval_iou": 0.982211804857441, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 11.14, |
|
"learning_rate": 3.914592592592593e-05, |
|
"loss": 0.0168, |
|
"step": 15600 |
|
}, |
|
{ |
|
"epoch": 11.21, |
|
"learning_rate": 3.899802469135802e-05, |
|
"loss": 0.0168, |
|
"step": 15700 |
|
}, |
|
{ |
|
"epoch": 11.29, |
|
"learning_rate": 3.885012345679013e-05, |
|
"loss": 0.0173, |
|
"step": 15800 |
|
}, |
|
{ |
|
"epoch": 11.36, |
|
"learning_rate": 3.870222222222222e-05, |
|
"loss": 0.0174, |
|
"step": 15900 |
|
}, |
|
{ |
|
"epoch": 11.43, |
|
"learning_rate": 3.855432098765432e-05, |
|
"loss": 0.0179, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 11.43, |
|
"eval_iou": 0.9822788746917949, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 11.5, |
|
"learning_rate": 3.840641975308642e-05, |
|
"loss": 0.0178, |
|
"step": 16100 |
|
}, |
|
{ |
|
"epoch": 11.57, |
|
"learning_rate": 3.825851851851852e-05, |
|
"loss": 0.0176, |
|
"step": 16200 |
|
}, |
|
{ |
|
"epoch": 11.64, |
|
"learning_rate": 3.8110617283950625e-05, |
|
"loss": 0.0167, |
|
"step": 16300 |
|
}, |
|
{ |
|
"epoch": 11.71, |
|
"learning_rate": 3.796271604938272e-05, |
|
"loss": 0.0175, |
|
"step": 16400 |
|
}, |
|
{ |
|
"epoch": 11.79, |
|
"learning_rate": 3.781481481481482e-05, |
|
"loss": 0.0187, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 11.79, |
|
"eval_iou": 0.9822942889204213, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 11.86, |
|
"learning_rate": 3.766691358024691e-05, |
|
"loss": 0.0189, |
|
"step": 16600 |
|
}, |
|
{ |
|
"epoch": 11.93, |
|
"learning_rate": 3.7519012345679015e-05, |
|
"loss": 0.0174, |
|
"step": 16700 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"learning_rate": 3.7371111111111114e-05, |
|
"loss": 0.0188, |
|
"step": 16800 |
|
}, |
|
{ |
|
"epoch": 12.07, |
|
"learning_rate": 3.7223209876543214e-05, |
|
"loss": 0.0186, |
|
"step": 16900 |
|
}, |
|
{ |
|
"epoch": 12.14, |
|
"learning_rate": 3.707530864197531e-05, |
|
"loss": 0.0184, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 12.14, |
|
"eval_iou": 0.9820192654927572, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 12.21, |
|
"learning_rate": 3.692740740740741e-05, |
|
"loss": 0.0167, |
|
"step": 17100 |
|
}, |
|
{ |
|
"epoch": 12.29, |
|
"learning_rate": 3.6779506172839504e-05, |
|
"loss": 0.017, |
|
"step": 17200 |
|
}, |
|
{ |
|
"epoch": 12.36, |
|
"learning_rate": 3.6631604938271604e-05, |
|
"loss": 0.0175, |
|
"step": 17300 |
|
}, |
|
{ |
|
"epoch": 12.43, |
|
"learning_rate": 3.64837037037037e-05, |
|
"loss": 0.0179, |
|
"step": 17400 |
|
}, |
|
{ |
|
"epoch": 12.5, |
|
"learning_rate": 3.633580246913581e-05, |
|
"loss": 0.0178, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 12.5, |
|
"eval_iou": 0.981830325196771, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 12.57, |
|
"learning_rate": 3.618790123456791e-05, |
|
"loss": 0.0179, |
|
"step": 17600 |
|
}, |
|
{ |
|
"epoch": 12.64, |
|
"learning_rate": 3.604e-05, |
|
"loss": 0.0179, |
|
"step": 17700 |
|
}, |
|
{ |
|
"epoch": 12.71, |
|
"learning_rate": 3.58920987654321e-05, |
|
"loss": 0.0172, |
|
"step": 17800 |
|
}, |
|
{ |
|
"epoch": 12.79, |
|
"learning_rate": 3.574419753086419e-05, |
|
"loss": 0.017, |
|
"step": 17900 |
|
}, |
|
{ |
|
"epoch": 12.86, |
|
"learning_rate": 3.55962962962963e-05, |
|
"loss": 0.0174, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 12.86, |
|
"eval_iou": 0.9824123674747991, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 12.93, |
|
"learning_rate": 3.54483950617284e-05, |
|
"loss": 0.0184, |
|
"step": 18100 |
|
}, |
|
{ |
|
"epoch": 13.0, |
|
"learning_rate": 3.53004938271605e-05, |
|
"loss": 0.0172, |
|
"step": 18200 |
|
}, |
|
{ |
|
"epoch": 13.07, |
|
"learning_rate": 3.5152592592592596e-05, |
|
"loss": 0.0171, |
|
"step": 18300 |
|
}, |
|
{ |
|
"epoch": 13.14, |
|
"learning_rate": 3.5004691358024695e-05, |
|
"loss": 0.0172, |
|
"step": 18400 |
|
}, |
|
{ |
|
"epoch": 13.21, |
|
"learning_rate": 3.4856790123456794e-05, |
|
"loss": 0.0173, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 13.21, |
|
"eval_iou": 0.9818773731297138, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 13.29, |
|
"learning_rate": 3.470888888888889e-05, |
|
"loss": 0.0167, |
|
"step": 18600 |
|
}, |
|
{ |
|
"epoch": 13.36, |
|
"learning_rate": 3.4560987654320986e-05, |
|
"loss": 0.0168, |
|
"step": 18700 |
|
}, |
|
{ |
|
"epoch": 13.43, |
|
"learning_rate": 3.441308641975309e-05, |
|
"loss": 0.0171, |
|
"step": 18800 |
|
}, |
|
{ |
|
"epoch": 13.5, |
|
"learning_rate": 3.426518518518519e-05, |
|
"loss": 0.0165, |
|
"step": 18900 |
|
}, |
|
{ |
|
"epoch": 13.57, |
|
"learning_rate": 3.411728395061729e-05, |
|
"loss": 0.0172, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 13.57, |
|
"eval_iou": 0.9824096741629582, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 13.64, |
|
"learning_rate": 3.396938271604938e-05, |
|
"loss": 0.0168, |
|
"step": 19100 |
|
}, |
|
{ |
|
"epoch": 13.71, |
|
"learning_rate": 3.382148148148148e-05, |
|
"loss": 0.0172, |
|
"step": 19200 |
|
}, |
|
{ |
|
"epoch": 13.79, |
|
"learning_rate": 3.367358024691358e-05, |
|
"loss": 0.017, |
|
"step": 19300 |
|
}, |
|
{ |
|
"epoch": 13.86, |
|
"learning_rate": 3.352567901234568e-05, |
|
"loss": 0.0172, |
|
"step": 19400 |
|
}, |
|
{ |
|
"epoch": 13.93, |
|
"learning_rate": 3.337777777777778e-05, |
|
"loss": 0.018, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 13.93, |
|
"eval_iou": 0.9822233032946494, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"learning_rate": 3.322987654320988e-05, |
|
"loss": 0.017, |
|
"step": 19600 |
|
}, |
|
{ |
|
"epoch": 14.07, |
|
"learning_rate": 3.308197530864198e-05, |
|
"loss": 0.017, |
|
"step": 19700 |
|
}, |
|
{ |
|
"epoch": 14.14, |
|
"learning_rate": 3.293407407407408e-05, |
|
"loss": 0.0176, |
|
"step": 19800 |
|
}, |
|
{ |
|
"epoch": 14.21, |
|
"learning_rate": 3.278617283950617e-05, |
|
"loss": 0.0173, |
|
"step": 19900 |
|
}, |
|
{ |
|
"epoch": 14.29, |
|
"learning_rate": 3.263827160493827e-05, |
|
"loss": 0.0172, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 14.29, |
|
"eval_iou": 0.9821930264725405, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 14.36, |
|
"learning_rate": 3.2490370370370375e-05, |
|
"loss": 0.0174, |
|
"step": 20100 |
|
}, |
|
{ |
|
"epoch": 14.43, |
|
"learning_rate": 3.2342469135802474e-05, |
|
"loss": 0.017, |
|
"step": 20200 |
|
}, |
|
{ |
|
"epoch": 14.5, |
|
"learning_rate": 3.2194567901234573e-05, |
|
"loss": 0.0175, |
|
"step": 20300 |
|
}, |
|
{ |
|
"epoch": 14.57, |
|
"learning_rate": 3.2046666666666666e-05, |
|
"loss": 0.0179, |
|
"step": 20400 |
|
}, |
|
{ |
|
"epoch": 14.64, |
|
"learning_rate": 3.189876543209877e-05, |
|
"loss": 0.0173, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 14.64, |
|
"eval_iou": 0.9825124956813513, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 14.71, |
|
"learning_rate": 3.1750864197530864e-05, |
|
"loss": 0.0168, |
|
"step": 20600 |
|
}, |
|
{ |
|
"epoch": 14.79, |
|
"learning_rate": 3.1602962962962964e-05, |
|
"loss": 0.018, |
|
"step": 20700 |
|
}, |
|
{ |
|
"epoch": 14.86, |
|
"learning_rate": 3.145506172839506e-05, |
|
"loss": 0.0158, |
|
"step": 20800 |
|
}, |
|
{ |
|
"epoch": 14.93, |
|
"learning_rate": 3.130716049382716e-05, |
|
"loss": 0.0174, |
|
"step": 20900 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"learning_rate": 3.115925925925927e-05, |
|
"loss": 0.017, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"eval_iou": 0.9824787465965047, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"eval_iou": 0.9824787465965047, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 15.07, |
|
"learning_rate": 3.101135802469136e-05, |
|
"loss": 0.0164, |
|
"step": 21100 |
|
}, |
|
{ |
|
"epoch": 15.14, |
|
"learning_rate": 3.086345679012345e-05, |
|
"loss": 0.0169, |
|
"step": 21200 |
|
}, |
|
{ |
|
"epoch": 15.21, |
|
"learning_rate": 3.071555555555555e-05, |
|
"loss": 0.0172, |
|
"step": 21300 |
|
}, |
|
{ |
|
"epoch": 15.29, |
|
"learning_rate": 3.056765432098765e-05, |
|
"loss": 0.0171, |
|
"step": 21400 |
|
}, |
|
{ |
|
"epoch": 15.36, |
|
"learning_rate": 3.0419753086419754e-05, |
|
"loss": 0.0174, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 15.36, |
|
"eval_iou": 0.9827606730601367, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 15.43, |
|
"learning_rate": 3.0271851851851857e-05, |
|
"loss": 0.0169, |
|
"step": 21600 |
|
}, |
|
{ |
|
"epoch": 15.5, |
|
"learning_rate": 3.0123950617283952e-05, |
|
"loss": 0.0175, |
|
"step": 21700 |
|
}, |
|
{ |
|
"epoch": 15.57, |
|
"learning_rate": 2.9976049382716048e-05, |
|
"loss": 0.0178, |
|
"step": 21800 |
|
}, |
|
{ |
|
"epoch": 15.64, |
|
"learning_rate": 2.982814814814815e-05, |
|
"loss": 0.0184, |
|
"step": 21900 |
|
}, |
|
{ |
|
"epoch": 15.71, |
|
"learning_rate": 2.9680246913580243e-05, |
|
"loss": 0.0167, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 15.71, |
|
"eval_iou": 0.9823599522020302, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 15.79, |
|
"learning_rate": 2.9532345679012346e-05, |
|
"loss": 0.0171, |
|
"step": 22100 |
|
}, |
|
{ |
|
"epoch": 15.86, |
|
"learning_rate": 2.9384444444444445e-05, |
|
"loss": 0.0171, |
|
"step": 22200 |
|
}, |
|
{ |
|
"epoch": 15.93, |
|
"learning_rate": 2.923654320987654e-05, |
|
"loss": 0.0162, |
|
"step": 22300 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"learning_rate": 2.908864197530864e-05, |
|
"loss": 0.0172, |
|
"step": 22400 |
|
}, |
|
{ |
|
"epoch": 16.07, |
|
"learning_rate": 2.8940740740740743e-05, |
|
"loss": 0.0168, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 16.07, |
|
"eval_iou": 0.9826575839052013, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 16.14, |
|
"learning_rate": 2.8792839506172835e-05, |
|
"loss": 0.0183, |
|
"step": 22600 |
|
}, |
|
{ |
|
"epoch": 16.21, |
|
"learning_rate": 2.8644938271604934e-05, |
|
"loss": 0.0174, |
|
"step": 22700 |
|
}, |
|
{ |
|
"epoch": 16.29, |
|
"learning_rate": 2.8497037037037037e-05, |
|
"loss": 0.0166, |
|
"step": 22800 |
|
}, |
|
{ |
|
"epoch": 16.36, |
|
"learning_rate": 2.8349135802469136e-05, |
|
"loss": 0.0166, |
|
"step": 22900 |
|
}, |
|
{ |
|
"epoch": 16.43, |
|
"learning_rate": 2.8201234567901232e-05, |
|
"loss": 0.0184, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 16.43, |
|
"eval_iou": 0.9823403113028583, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 16.5, |
|
"learning_rate": 2.805333333333333e-05, |
|
"loss": 0.0172, |
|
"step": 23100 |
|
}, |
|
{ |
|
"epoch": 16.57, |
|
"learning_rate": 2.7905432098765434e-05, |
|
"loss": 0.018, |
|
"step": 23200 |
|
}, |
|
{ |
|
"epoch": 16.64, |
|
"learning_rate": 2.7757530864197526e-05, |
|
"loss": 0.0177, |
|
"step": 23300 |
|
}, |
|
{ |
|
"epoch": 16.71, |
|
"learning_rate": 2.760962962962963e-05, |
|
"loss": 0.0174, |
|
"step": 23400 |
|
}, |
|
{ |
|
"epoch": 16.79, |
|
"learning_rate": 2.7461728395061728e-05, |
|
"loss": 0.0176, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 16.79, |
|
"eval_iou": 0.982518440952488, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 16.86, |
|
"learning_rate": 2.7313827160493824e-05, |
|
"loss": 0.0173, |
|
"step": 23600 |
|
}, |
|
{ |
|
"epoch": 16.93, |
|
"learning_rate": 2.7165925925925923e-05, |
|
"loss": 0.0173, |
|
"step": 23700 |
|
}, |
|
{ |
|
"epoch": 17.0, |
|
"learning_rate": 2.7018024691358026e-05, |
|
"loss": 0.0173, |
|
"step": 23800 |
|
}, |
|
{ |
|
"epoch": 17.07, |
|
"learning_rate": 2.6870123456790125e-05, |
|
"loss": 0.0163, |
|
"step": 23900 |
|
}, |
|
{ |
|
"epoch": 17.14, |
|
"learning_rate": 2.672222222222222e-05, |
|
"loss": 0.0174, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 17.14, |
|
"eval_iou": 0.9825776663481021, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 17.21, |
|
"learning_rate": 2.657432098765432e-05, |
|
"loss": 0.0167, |
|
"step": 24100 |
|
}, |
|
{ |
|
"epoch": 17.29, |
|
"learning_rate": 2.642641975308642e-05, |
|
"loss": 0.0172, |
|
"step": 24200 |
|
}, |
|
{ |
|
"epoch": 17.36, |
|
"learning_rate": 2.6278518518518515e-05, |
|
"loss": 0.0167, |
|
"step": 24300 |
|
}, |
|
{ |
|
"epoch": 17.43, |
|
"learning_rate": 2.6130617283950614e-05, |
|
"loss": 0.0173, |
|
"step": 24400 |
|
}, |
|
{ |
|
"epoch": 17.5, |
|
"learning_rate": 2.5982716049382717e-05, |
|
"loss": 0.0167, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 17.5, |
|
"eval_iou": 0.9825901704676011, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 17.57, |
|
"learning_rate": 2.583481481481481e-05, |
|
"loss": 0.0169, |
|
"step": 24600 |
|
}, |
|
{ |
|
"epoch": 17.64, |
|
"learning_rate": 2.5686913580246912e-05, |
|
"loss": 0.0168, |
|
"step": 24700 |
|
}, |
|
{ |
|
"epoch": 17.71, |
|
"learning_rate": 2.553901234567901e-05, |
|
"loss": 0.0175, |
|
"step": 24800 |
|
}, |
|
{ |
|
"epoch": 17.79, |
|
"learning_rate": 2.5391111111111114e-05, |
|
"loss": 0.0183, |
|
"step": 24900 |
|
}, |
|
{ |
|
"epoch": 17.86, |
|
"learning_rate": 2.5243209876543206e-05, |
|
"loss": 0.0169, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 17.86, |
|
"eval_iou": 0.9825106870894339, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 17.93, |
|
"learning_rate": 2.509530864197531e-05, |
|
"loss": 0.0172, |
|
"step": 25100 |
|
}, |
|
{ |
|
"epoch": 18.0, |
|
"learning_rate": 2.4947407407407408e-05, |
|
"loss": 0.017, |
|
"step": 25200 |
|
}, |
|
{ |
|
"epoch": 18.07, |
|
"learning_rate": 2.4799506172839504e-05, |
|
"loss": 0.0175, |
|
"step": 25300 |
|
}, |
|
{ |
|
"epoch": 18.14, |
|
"learning_rate": 2.4651604938271603e-05, |
|
"loss": 0.0162, |
|
"step": 25400 |
|
}, |
|
{ |
|
"epoch": 18.21, |
|
"learning_rate": 2.4503703703703706e-05, |
|
"loss": 0.0161, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 18.21, |
|
"eval_iou": 0.9824175115893868, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 18.29, |
|
"learning_rate": 2.4355802469135798e-05, |
|
"loss": 0.0166, |
|
"step": 25600 |
|
}, |
|
{ |
|
"epoch": 18.36, |
|
"learning_rate": 2.42079012345679e-05, |
|
"loss": 0.0171, |
|
"step": 25700 |
|
}, |
|
{ |
|
"epoch": 18.43, |
|
"learning_rate": 2.406e-05, |
|
"loss": 0.0176, |
|
"step": 25800 |
|
}, |
|
{ |
|
"epoch": 18.5, |
|
"learning_rate": 2.3912098765432092e-05, |
|
"loss": 0.0173, |
|
"step": 25900 |
|
}, |
|
{ |
|
"epoch": 18.57, |
|
"learning_rate": 2.3764197530864195e-05, |
|
"loss": 0.0171, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 18.57, |
|
"eval_iou": 0.9823989424051023, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 18.64, |
|
"learning_rate": 2.3616296296296294e-05, |
|
"loss": 0.0163, |
|
"step": 26100 |
|
}, |
|
{ |
|
"epoch": 18.71, |
|
"learning_rate": 2.3468395061728397e-05, |
|
"loss": 0.0185, |
|
"step": 26200 |
|
}, |
|
{ |
|
"epoch": 18.79, |
|
"learning_rate": 2.332049382716049e-05, |
|
"loss": 0.0164, |
|
"step": 26300 |
|
}, |
|
{ |
|
"epoch": 18.86, |
|
"learning_rate": 2.3172592592592592e-05, |
|
"loss": 0.0169, |
|
"step": 26400 |
|
}, |
|
{ |
|
"epoch": 18.93, |
|
"learning_rate": 2.302469135802469e-05, |
|
"loss": 0.0173, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 18.93, |
|
"eval_iou": 0.9824401590169645, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 19.0, |
|
"learning_rate": 2.2876790123456787e-05, |
|
"loss": 0.0173, |
|
"step": 26600 |
|
}, |
|
{ |
|
"epoch": 19.07, |
|
"learning_rate": 2.2728888888888886e-05, |
|
"loss": 0.017, |
|
"step": 26700 |
|
}, |
|
{ |
|
"epoch": 19.14, |
|
"learning_rate": 2.258098765432099e-05, |
|
"loss": 0.0171, |
|
"step": 26800 |
|
}, |
|
{ |
|
"epoch": 19.21, |
|
"learning_rate": 2.243308641975308e-05, |
|
"loss": 0.0169, |
|
"step": 26900 |
|
}, |
|
{ |
|
"epoch": 19.29, |
|
"learning_rate": 2.2285185185185184e-05, |
|
"loss": 0.0169, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 19.29, |
|
"eval_iou": 0.9826161189406526, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 19.36, |
|
"learning_rate": 2.2137283950617283e-05, |
|
"loss": 0.0176, |
|
"step": 27100 |
|
}, |
|
{ |
|
"epoch": 19.43, |
|
"learning_rate": 2.1989382716049386e-05, |
|
"loss": 0.0163, |
|
"step": 27200 |
|
}, |
|
{ |
|
"epoch": 19.5, |
|
"learning_rate": 2.1841481481481478e-05, |
|
"loss": 0.0168, |
|
"step": 27300 |
|
}, |
|
{ |
|
"epoch": 19.57, |
|
"learning_rate": 2.169358024691358e-05, |
|
"loss": 0.0154, |
|
"step": 27400 |
|
}, |
|
{ |
|
"epoch": 19.64, |
|
"learning_rate": 2.154567901234568e-05, |
|
"loss": 0.0167, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 19.64, |
|
"eval_iou": 0.9828028941855711, |
|
"step": 27500 |
|
} |
|
], |
|
"logging_steps": 100, |
|
"max_steps": 42000, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 30, |
|
"save_steps": 500, |
|
"total_flos": 0.0, |
|
"train_batch_size": 10, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|