|
{ |
|
"best_metric": 13.085066864784547, |
|
"best_model_checkpoint": "./whisper-medium-el/checkpoint-1000", |
|
"epoch": 4.0392, |
|
"global_step": 1000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.0000000000000003e-07, |
|
"loss": 8.8645, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.000000000000001e-07, |
|
"loss": 5.8694, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.4000000000000001e-06, |
|
"loss": 3.1568, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.8800000000000002e-06, |
|
"loss": 1.9833, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 2.38e-06, |
|
"loss": 1.4934, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 2.88e-06, |
|
"loss": 1.1115, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 3.3800000000000007e-06, |
|
"loss": 0.9083, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 3.88e-06, |
|
"loss": 0.6967, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 4.38e-06, |
|
"loss": 0.5981, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 4.880000000000001e-06, |
|
"loss": 0.5506, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 5.380000000000001e-06, |
|
"loss": 0.453, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 5.8800000000000005e-06, |
|
"loss": 0.3967, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 6.380000000000001e-06, |
|
"loss": 0.4095, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 6.88e-06, |
|
"loss": 0.3197, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 7.3800000000000005e-06, |
|
"loss": 0.3176, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 7.88e-06, |
|
"loss": 0.278, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 8.380000000000001e-06, |
|
"loss": 0.2536, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 8.880000000000001e-06, |
|
"loss": 0.2529, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 9.38e-06, |
|
"loss": 0.2158, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 9.88e-06, |
|
"loss": 0.204, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 9.957777777777779e-06, |
|
"loss": 0.21, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 9.902222222222223e-06, |
|
"loss": 0.1758, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 9.846666666666668e-06, |
|
"loss": 0.1668, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 9.791111111111112e-06, |
|
"loss": 0.1475, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 9.735555555555556e-06, |
|
"loss": 0.1416, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 9.68e-06, |
|
"loss": 0.138, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 9.624444444444445e-06, |
|
"loss": 0.1225, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 9.56888888888889e-06, |
|
"loss": 0.1156, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 9.513333333333334e-06, |
|
"loss": 0.1093, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 9.457777777777778e-06, |
|
"loss": 0.1041, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 9.402222222222222e-06, |
|
"loss": 0.0971, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 9.346666666666666e-06, |
|
"loss": 0.0875, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 9.291111111111112e-06, |
|
"loss": 0.0837, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 9.235555555555556e-06, |
|
"loss": 0.0826, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 9.180000000000002e-06, |
|
"loss": 0.0737, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 9.124444444444444e-06, |
|
"loss": 0.0717, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 9.06888888888889e-06, |
|
"loss": 0.0707, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 9.013333333333334e-06, |
|
"loss": 0.0647, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 8.957777777777778e-06, |
|
"loss": 0.0593, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 8.902222222222224e-06, |
|
"loss": 0.0582, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"eval_loss": 0.3256979286670685, |
|
"eval_runtime": 1207.5418, |
|
"eval_samples_per_second": 1.405, |
|
"eval_steps_per_second": 0.088, |
|
"eval_wer": 13.085066864784547, |
|
"step": 1000 |
|
} |
|
], |
|
"max_steps": 5000, |
|
"num_train_epochs": 9223372036854775807, |
|
"total_flos": 3.264302736211968e+19, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|