|
{ |
|
"best_metric": 0.5357989568515884, |
|
"best_model_checkpoint": "./indobertweet-review-rating/checkpoint-93240", |
|
"epoch": 6.0, |
|
"eval_steps": 500, |
|
"global_step": 93240, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 1.996782496782497e-05, |
|
"loss": 0.4164, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.9935649935649937e-05, |
|
"loss": 0.3806, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.9903474903474905e-05, |
|
"loss": 0.3775, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 1.9871299871299873e-05, |
|
"loss": 0.373, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.983912483912484e-05, |
|
"loss": 0.3697, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.980694980694981e-05, |
|
"loss": 0.3687, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.9774774774774777e-05, |
|
"loss": 0.3683, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.9742599742599745e-05, |
|
"loss": 0.3591, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.9710424710424713e-05, |
|
"loss": 0.37, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.967824967824968e-05, |
|
"loss": 0.3643, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.964607464607465e-05, |
|
"loss": 0.3664, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.9613899613899616e-05, |
|
"loss": 0.3644, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.9581724581724584e-05, |
|
"loss": 0.3633, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.9549549549549552e-05, |
|
"loss": 0.3616, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 1.951737451737452e-05, |
|
"loss": 0.3659, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.9485199485199485e-05, |
|
"loss": 0.3613, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 1.9453024453024453e-05, |
|
"loss": 0.3661, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 1.942084942084942e-05, |
|
"loss": 0.3605, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 1.9388674388674392e-05, |
|
"loss": 0.3544, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.935649935649936e-05, |
|
"loss": 0.3599, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.9324324324324328e-05, |
|
"loss": 0.3583, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 1.9292149292149292e-05, |
|
"loss": 0.3571, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.925997425997426e-05, |
|
"loss": 0.3598, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 1.9227799227799228e-05, |
|
"loss": 0.3567, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.9195624195624196e-05, |
|
"loss": 0.3572, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.9163449163449167e-05, |
|
"loss": 0.3512, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.9131274131274135e-05, |
|
"loss": 0.3565, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 1.90990990990991e-05, |
|
"loss": 0.3452, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.9066924066924067e-05, |
|
"loss": 0.3531, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.9034749034749035e-05, |
|
"loss": 0.3475, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 1.9002574002574003e-05, |
|
"loss": 0.3579, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_akurasi": 0.38799603174603176, |
|
"eval_f1": 0.4929409550103533, |
|
"eval_loss": 0.3472347557544708, |
|
"eval_roc_auc": 0.6709201388888889, |
|
"eval_runtime": 170.7708, |
|
"eval_samples_per_second": 59.026, |
|
"eval_steps_per_second": 7.378, |
|
"step": 15540 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 1.897039897039897e-05, |
|
"loss": 0.3402, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 1.8938223938223942e-05, |
|
"loss": 0.3325, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.8906048906048907e-05, |
|
"loss": 0.3306, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.8873873873873875e-05, |
|
"loss": 0.3326, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.8841698841698843e-05, |
|
"loss": 0.3387, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 1.880952380952381e-05, |
|
"loss": 0.3336, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 1.877734877734878e-05, |
|
"loss": 0.3367, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 1.8745173745173746e-05, |
|
"loss": 0.3332, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 1.8712998712998714e-05, |
|
"loss": 0.3305, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 1.8680823680823682e-05, |
|
"loss": 0.3335, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 1.864864864864865e-05, |
|
"loss": 0.3349, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 1.8616473616473618e-05, |
|
"loss": 0.3327, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 1.8584298584298586e-05, |
|
"loss": 0.3268, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 1.8552123552123554e-05, |
|
"loss": 0.3264, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 1.8519948519948522e-05, |
|
"loss": 0.3375, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 1.848777348777349e-05, |
|
"loss": 0.333, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 1.8455598455598458e-05, |
|
"loss": 0.3373, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 1.8423423423423425e-05, |
|
"loss": 0.3327, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 1.8391248391248393e-05, |
|
"loss": 0.3297, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 1.835907335907336e-05, |
|
"loss": 0.3413, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 1.832689832689833e-05, |
|
"loss": 0.3303, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 1.8294723294723297e-05, |
|
"loss": 0.3305, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 1.826254826254826e-05, |
|
"loss": 0.3334, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 1.8230373230373233e-05, |
|
"loss": 0.331, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 1.81981981981982e-05, |
|
"loss": 0.3382, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 1.816602316602317e-05, |
|
"loss": 0.3361, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.8133848133848137e-05, |
|
"loss": 0.34, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 1.8101673101673104e-05, |
|
"loss": 0.3382, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 1.806949806949807e-05, |
|
"loss": 0.3293, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 1.8037323037323037e-05, |
|
"loss": 0.3291, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 1.8005148005148008e-05, |
|
"loss": 0.3286, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_akurasi": 0.43134920634920637, |
|
"eval_f1": 0.5178539599548453, |
|
"eval_loss": 0.34680771827697754, |
|
"eval_roc_auc": 0.6864955357142858, |
|
"eval_runtime": 170.8274, |
|
"eval_samples_per_second": 59.007, |
|
"eval_steps_per_second": 7.376, |
|
"step": 31080 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 1.7972972972972976e-05, |
|
"loss": 0.3128, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 1.7940797940797944e-05, |
|
"loss": 0.301, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 1.7908622908622912e-05, |
|
"loss": 0.3053, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 1.7876447876447876e-05, |
|
"loss": 0.3079, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 1.7844272844272844e-05, |
|
"loss": 0.3001, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 1.7812097812097812e-05, |
|
"loss": 0.3017, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 1.7779922779922783e-05, |
|
"loss": 0.2996, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 1.774774774774775e-05, |
|
"loss": 0.306, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 1.7715572715572716e-05, |
|
"loss": 0.3084, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 1.7683397683397684e-05, |
|
"loss": 0.3139, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 1.7651222651222652e-05, |
|
"loss": 0.3106, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 1.761904761904762e-05, |
|
"loss": 0.306, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 1.7586872586872588e-05, |
|
"loss": 0.3108, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 1.7554697554697555e-05, |
|
"loss": 0.3155, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 1.7522522522522523e-05, |
|
"loss": 0.3021, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 1.749034749034749e-05, |
|
"loss": 0.31, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 1.745817245817246e-05, |
|
"loss": 0.3061, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 1.7425997425997427e-05, |
|
"loss": 0.3108, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 1.7393822393822395e-05, |
|
"loss": 0.3028, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 1.7361647361647363e-05, |
|
"loss": 0.3081, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 1.732947232947233e-05, |
|
"loss": 0.3091, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 1.72972972972973e-05, |
|
"loss": 0.3063, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 1.7265122265122267e-05, |
|
"loss": 0.3085, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 1.7232947232947234e-05, |
|
"loss": 0.3079, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 1.7200772200772202e-05, |
|
"loss": 0.3109, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 1.716859716859717e-05, |
|
"loss": 0.3135, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 1.7136422136422138e-05, |
|
"loss": 0.3063, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 1.7104247104247106e-05, |
|
"loss": 0.3173, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 1.7072072072072074e-05, |
|
"loss": 0.3082, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 1.7039897039897042e-05, |
|
"loss": 0.3047, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 1.700772200772201e-05, |
|
"loss": 0.3124, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_akurasi": 0.4599206349206349, |
|
"eval_f1": 0.5274112829845313, |
|
"eval_loss": 0.3588548004627228, |
|
"eval_roc_auc": 0.6944568452380953, |
|
"eval_runtime": 171.0225, |
|
"eval_samples_per_second": 58.94, |
|
"eval_steps_per_second": 7.367, |
|
"step": 46620 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 1.6975546975546978e-05, |
|
"loss": 0.2838, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 1.6943371943371946e-05, |
|
"loss": 0.2749, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 1.6911196911196913e-05, |
|
"loss": 0.2808, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"learning_rate": 1.6879021879021878e-05, |
|
"loss": 0.2751, |
|
"step": 48500 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"learning_rate": 1.6846846846846846e-05, |
|
"loss": 0.279, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 1.6814671814671817e-05, |
|
"loss": 0.2743, |
|
"step": 49500 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 1.6782496782496785e-05, |
|
"loss": 0.2787, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"learning_rate": 1.6750321750321753e-05, |
|
"loss": 0.2796, |
|
"step": 50500 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 1.671814671814672e-05, |
|
"loss": 0.2723, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"learning_rate": 1.6685971685971685e-05, |
|
"loss": 0.2751, |
|
"step": 51500 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"learning_rate": 1.6653796653796653e-05, |
|
"loss": 0.2682, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"learning_rate": 1.662162162162162e-05, |
|
"loss": 0.2722, |
|
"step": 52500 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 1.6589446589446592e-05, |
|
"loss": 0.2841, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 1.655727155727156e-05, |
|
"loss": 0.2762, |
|
"step": 53500 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 1.6525096525096528e-05, |
|
"loss": 0.2762, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 3.51, |
|
"learning_rate": 1.6492921492921493e-05, |
|
"loss": 0.2785, |
|
"step": 54500 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"learning_rate": 1.646074646074646e-05, |
|
"loss": 0.2753, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"learning_rate": 1.642857142857143e-05, |
|
"loss": 0.2742, |
|
"step": 55500 |
|
}, |
|
{ |
|
"epoch": 3.6, |
|
"learning_rate": 1.6396396396396396e-05, |
|
"loss": 0.282, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 3.64, |
|
"learning_rate": 1.6364221364221368e-05, |
|
"loss": 0.2767, |
|
"step": 56500 |
|
}, |
|
{ |
|
"epoch": 3.67, |
|
"learning_rate": 1.6332046332046336e-05, |
|
"loss": 0.2851, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 3.7, |
|
"learning_rate": 1.62998712998713e-05, |
|
"loss": 0.2818, |
|
"step": 57500 |
|
}, |
|
{ |
|
"epoch": 3.73, |
|
"learning_rate": 1.6267696267696268e-05, |
|
"loss": 0.278, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 3.76, |
|
"learning_rate": 1.6235521235521236e-05, |
|
"loss": 0.2695, |
|
"step": 58500 |
|
}, |
|
{ |
|
"epoch": 3.8, |
|
"learning_rate": 1.6203346203346204e-05, |
|
"loss": 0.2758, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 3.83, |
|
"learning_rate": 1.6171171171171172e-05, |
|
"loss": 0.2805, |
|
"step": 59500 |
|
}, |
|
{ |
|
"epoch": 3.86, |
|
"learning_rate": 1.613899613899614e-05, |
|
"loss": 0.2846, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 3.89, |
|
"learning_rate": 1.6106821106821108e-05, |
|
"loss": 0.2774, |
|
"step": 60500 |
|
}, |
|
{ |
|
"epoch": 3.93, |
|
"learning_rate": 1.6074646074646076e-05, |
|
"loss": 0.2808, |
|
"step": 61000 |
|
}, |
|
{ |
|
"epoch": 3.96, |
|
"learning_rate": 1.6042471042471043e-05, |
|
"loss": 0.2833, |
|
"step": 61500 |
|
}, |
|
{ |
|
"epoch": 3.99, |
|
"learning_rate": 1.601029601029601e-05, |
|
"loss": 0.2709, |
|
"step": 62000 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_akurasi": 0.491765873015873, |
|
"eval_f1": 0.5355716290618591, |
|
"eval_loss": 0.37542611360549927, |
|
"eval_roc_auc": 0.7028645833333333, |
|
"eval_runtime": 171.2358, |
|
"eval_samples_per_second": 58.866, |
|
"eval_steps_per_second": 7.358, |
|
"step": 62160 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 1.597812097812098e-05, |
|
"loss": 0.2595, |
|
"step": 62500 |
|
}, |
|
{ |
|
"epoch": 4.05, |
|
"learning_rate": 1.5945945945945947e-05, |
|
"loss": 0.2407, |
|
"step": 63000 |
|
}, |
|
{ |
|
"epoch": 4.09, |
|
"learning_rate": 1.5913770913770915e-05, |
|
"loss": 0.234, |
|
"step": 63500 |
|
}, |
|
{ |
|
"epoch": 4.12, |
|
"learning_rate": 1.5881595881595883e-05, |
|
"loss": 0.2333, |
|
"step": 64000 |
|
}, |
|
{ |
|
"epoch": 4.15, |
|
"learning_rate": 1.584942084942085e-05, |
|
"loss": 0.2502, |
|
"step": 64500 |
|
}, |
|
{ |
|
"epoch": 4.18, |
|
"learning_rate": 1.581724581724582e-05, |
|
"loss": 0.2441, |
|
"step": 65000 |
|
}, |
|
{ |
|
"epoch": 4.21, |
|
"learning_rate": 1.5785070785070787e-05, |
|
"loss": 0.2452, |
|
"step": 65500 |
|
}, |
|
{ |
|
"epoch": 4.25, |
|
"learning_rate": 1.5752895752895755e-05, |
|
"loss": 0.2418, |
|
"step": 66000 |
|
}, |
|
{ |
|
"epoch": 4.28, |
|
"learning_rate": 1.5720720720720722e-05, |
|
"loss": 0.2442, |
|
"step": 66500 |
|
}, |
|
{ |
|
"epoch": 4.31, |
|
"learning_rate": 1.568854568854569e-05, |
|
"loss": 0.2421, |
|
"step": 67000 |
|
}, |
|
{ |
|
"epoch": 4.34, |
|
"learning_rate": 1.5656370656370658e-05, |
|
"loss": 0.2406, |
|
"step": 67500 |
|
}, |
|
{ |
|
"epoch": 4.38, |
|
"learning_rate": 1.5624195624195626e-05, |
|
"loss": 0.2551, |
|
"step": 68000 |
|
}, |
|
{ |
|
"epoch": 4.41, |
|
"learning_rate": 1.5592020592020594e-05, |
|
"loss": 0.2465, |
|
"step": 68500 |
|
}, |
|
{ |
|
"epoch": 4.44, |
|
"learning_rate": 1.5559845559845562e-05, |
|
"loss": 0.2489, |
|
"step": 69000 |
|
}, |
|
{ |
|
"epoch": 4.47, |
|
"learning_rate": 1.552767052767053e-05, |
|
"loss": 0.2486, |
|
"step": 69500 |
|
}, |
|
{ |
|
"epoch": 4.5, |
|
"learning_rate": 1.5495495495495498e-05, |
|
"loss": 0.2535, |
|
"step": 70000 |
|
}, |
|
{ |
|
"epoch": 4.54, |
|
"learning_rate": 1.5463320463320462e-05, |
|
"loss": 0.2515, |
|
"step": 70500 |
|
}, |
|
{ |
|
"epoch": 4.57, |
|
"learning_rate": 1.5431145431145434e-05, |
|
"loss": 0.2585, |
|
"step": 71000 |
|
}, |
|
{ |
|
"epoch": 4.6, |
|
"learning_rate": 1.53989703989704e-05, |
|
"loss": 0.2451, |
|
"step": 71500 |
|
}, |
|
{ |
|
"epoch": 4.63, |
|
"learning_rate": 1.536679536679537e-05, |
|
"loss": 0.2493, |
|
"step": 72000 |
|
}, |
|
{ |
|
"epoch": 4.67, |
|
"learning_rate": 1.5334620334620337e-05, |
|
"loss": 0.2522, |
|
"step": 72500 |
|
}, |
|
{ |
|
"epoch": 4.7, |
|
"learning_rate": 1.5302445302445305e-05, |
|
"loss": 0.2517, |
|
"step": 73000 |
|
}, |
|
{ |
|
"epoch": 4.73, |
|
"learning_rate": 1.527027027027027e-05, |
|
"loss": 0.2511, |
|
"step": 73500 |
|
}, |
|
{ |
|
"epoch": 4.76, |
|
"learning_rate": 1.523809523809524e-05, |
|
"loss": 0.2542, |
|
"step": 74000 |
|
}, |
|
{ |
|
"epoch": 4.79, |
|
"learning_rate": 1.5205920205920209e-05, |
|
"loss": 0.2568, |
|
"step": 74500 |
|
}, |
|
{ |
|
"epoch": 4.83, |
|
"learning_rate": 1.5173745173745175e-05, |
|
"loss": 0.2528, |
|
"step": 75000 |
|
}, |
|
{ |
|
"epoch": 4.86, |
|
"learning_rate": 1.5141570141570143e-05, |
|
"loss": 0.2463, |
|
"step": 75500 |
|
}, |
|
{ |
|
"epoch": 4.89, |
|
"learning_rate": 1.5109395109395111e-05, |
|
"loss": 0.2491, |
|
"step": 76000 |
|
}, |
|
{ |
|
"epoch": 4.92, |
|
"learning_rate": 1.5077220077220079e-05, |
|
"loss": 0.2519, |
|
"step": 76500 |
|
}, |
|
{ |
|
"epoch": 4.95, |
|
"learning_rate": 1.5045045045045045e-05, |
|
"loss": 0.2594, |
|
"step": 77000 |
|
}, |
|
{ |
|
"epoch": 4.99, |
|
"learning_rate": 1.5012870012870013e-05, |
|
"loss": 0.2539, |
|
"step": 77500 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_akurasi": 0.489781746031746, |
|
"eval_f1": 0.5299071938200739, |
|
"eval_loss": 0.404236763715744, |
|
"eval_roc_auc": 0.7000744047619047, |
|
"eval_runtime": 171.3763, |
|
"eval_samples_per_second": 58.818, |
|
"eval_steps_per_second": 7.352, |
|
"step": 77700 |
|
}, |
|
{ |
|
"epoch": 5.02, |
|
"learning_rate": 1.498069498069498e-05, |
|
"loss": 0.2286, |
|
"step": 78000 |
|
}, |
|
{ |
|
"epoch": 5.05, |
|
"learning_rate": 1.494851994851995e-05, |
|
"loss": 0.2122, |
|
"step": 78500 |
|
}, |
|
{ |
|
"epoch": 5.08, |
|
"learning_rate": 1.4916344916344918e-05, |
|
"loss": 0.2104, |
|
"step": 79000 |
|
}, |
|
{ |
|
"epoch": 5.12, |
|
"learning_rate": 1.4884169884169886e-05, |
|
"loss": 0.218, |
|
"step": 79500 |
|
}, |
|
{ |
|
"epoch": 5.15, |
|
"learning_rate": 1.4851994851994852e-05, |
|
"loss": 0.2155, |
|
"step": 80000 |
|
}, |
|
{ |
|
"epoch": 5.18, |
|
"learning_rate": 1.481981981981982e-05, |
|
"loss": 0.2241, |
|
"step": 80500 |
|
}, |
|
{ |
|
"epoch": 5.21, |
|
"learning_rate": 1.4787644787644788e-05, |
|
"loss": 0.2214, |
|
"step": 81000 |
|
}, |
|
{ |
|
"epoch": 5.24, |
|
"learning_rate": 1.4755469755469756e-05, |
|
"loss": 0.217, |
|
"step": 81500 |
|
}, |
|
{ |
|
"epoch": 5.28, |
|
"learning_rate": 1.4723294723294726e-05, |
|
"loss": 0.2222, |
|
"step": 82000 |
|
}, |
|
{ |
|
"epoch": 5.31, |
|
"learning_rate": 1.4691119691119694e-05, |
|
"loss": 0.2207, |
|
"step": 82500 |
|
}, |
|
{ |
|
"epoch": 5.34, |
|
"learning_rate": 1.465894465894466e-05, |
|
"loss": 0.2277, |
|
"step": 83000 |
|
}, |
|
{ |
|
"epoch": 5.37, |
|
"learning_rate": 1.4626769626769628e-05, |
|
"loss": 0.2265, |
|
"step": 83500 |
|
}, |
|
{ |
|
"epoch": 5.41, |
|
"learning_rate": 1.4594594594594596e-05, |
|
"loss": 0.224, |
|
"step": 84000 |
|
}, |
|
{ |
|
"epoch": 5.44, |
|
"learning_rate": 1.4562419562419564e-05, |
|
"loss": 0.2177, |
|
"step": 84500 |
|
}, |
|
{ |
|
"epoch": 5.47, |
|
"learning_rate": 1.453024453024453e-05, |
|
"loss": 0.2192, |
|
"step": 85000 |
|
}, |
|
{ |
|
"epoch": 5.5, |
|
"learning_rate": 1.4498069498069501e-05, |
|
"loss": 0.2245, |
|
"step": 85500 |
|
}, |
|
{ |
|
"epoch": 5.53, |
|
"learning_rate": 1.4465894465894467e-05, |
|
"loss": 0.2212, |
|
"step": 86000 |
|
}, |
|
{ |
|
"epoch": 5.57, |
|
"learning_rate": 1.4433719433719435e-05, |
|
"loss": 0.2297, |
|
"step": 86500 |
|
}, |
|
{ |
|
"epoch": 5.6, |
|
"learning_rate": 1.4401544401544403e-05, |
|
"loss": 0.2309, |
|
"step": 87000 |
|
}, |
|
{ |
|
"epoch": 5.63, |
|
"learning_rate": 1.4369369369369371e-05, |
|
"loss": 0.2293, |
|
"step": 87500 |
|
}, |
|
{ |
|
"epoch": 5.66, |
|
"learning_rate": 1.4337194337194337e-05, |
|
"loss": 0.2236, |
|
"step": 88000 |
|
}, |
|
{ |
|
"epoch": 5.69, |
|
"learning_rate": 1.4305019305019305e-05, |
|
"loss": 0.227, |
|
"step": 88500 |
|
}, |
|
{ |
|
"epoch": 5.73, |
|
"learning_rate": 1.4272844272844275e-05, |
|
"loss": 0.2221, |
|
"step": 89000 |
|
}, |
|
{ |
|
"epoch": 5.76, |
|
"learning_rate": 1.4240669240669243e-05, |
|
"loss": 0.2315, |
|
"step": 89500 |
|
}, |
|
{ |
|
"epoch": 5.79, |
|
"learning_rate": 1.420849420849421e-05, |
|
"loss": 0.2289, |
|
"step": 90000 |
|
}, |
|
{ |
|
"epoch": 5.82, |
|
"learning_rate": 1.4176319176319178e-05, |
|
"loss": 0.2285, |
|
"step": 90500 |
|
}, |
|
{ |
|
"epoch": 5.86, |
|
"learning_rate": 1.4144144144144145e-05, |
|
"loss": 0.2291, |
|
"step": 91000 |
|
}, |
|
{ |
|
"epoch": 5.89, |
|
"learning_rate": 1.4111969111969112e-05, |
|
"loss": 0.2201, |
|
"step": 91500 |
|
}, |
|
{ |
|
"epoch": 5.92, |
|
"learning_rate": 1.407979407979408e-05, |
|
"loss": 0.2308, |
|
"step": 92000 |
|
}, |
|
{ |
|
"epoch": 5.95, |
|
"learning_rate": 1.4047619047619048e-05, |
|
"loss": 0.2301, |
|
"step": 92500 |
|
}, |
|
{ |
|
"epoch": 5.98, |
|
"learning_rate": 1.4015444015444018e-05, |
|
"loss": 0.2273, |
|
"step": 93000 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_akurasi": 0.5041666666666667, |
|
"eval_f1": 0.5357989568515884, |
|
"eval_loss": 0.4383452534675598, |
|
"eval_roc_auc": 0.7049107142857142, |
|
"eval_runtime": 171.6368, |
|
"eval_samples_per_second": 58.729, |
|
"eval_steps_per_second": 7.341, |
|
"step": 93240 |
|
} |
|
], |
|
"logging_steps": 500, |
|
"max_steps": 310800, |
|
"num_train_epochs": 20, |
|
"save_steps": 500, |
|
"total_flos": 1.471940774934958e+17, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|