|
{ |
|
"best_metric": 0.7447552447552448, |
|
"best_model_checkpoint": "wav2vec2-5Class-train-test-finetune-Medium/checkpoint-2640", |
|
"epoch": 237.28813559322035, |
|
"eval_steps": 500, |
|
"global_step": 3500, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.95, |
|
"grad_norm": 0.8927629590034485, |
|
"learning_rate": 1.2000000000000002e-06, |
|
"loss": 1.6852, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"eval_accuracy": 0.34265734265734266, |
|
"eval_loss": 1.5986738204956055, |
|
"eval_runtime": 4.6543, |
|
"eval_samples_per_second": 61.448, |
|
"eval_steps_per_second": 3.867, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"grad_norm": 0.6501776576042175, |
|
"learning_rate": 2.4857142857142858e-06, |
|
"loss": 1.5721, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"eval_accuracy": 0.34265734265734266, |
|
"eval_loss": 1.5975812673568726, |
|
"eval_runtime": 4.4155, |
|
"eval_samples_per_second": 64.772, |
|
"eval_steps_per_second": 4.077, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"grad_norm": 0.6221457719802856, |
|
"learning_rate": 3.771428571428572e-06, |
|
"loss": 1.5696, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"eval_accuracy": 0.34265734265734266, |
|
"eval_loss": 1.5957201719284058, |
|
"eval_runtime": 4.5928, |
|
"eval_samples_per_second": 62.272, |
|
"eval_steps_per_second": 3.919, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"grad_norm": 1.563844919204712, |
|
"learning_rate": 5.057142857142857e-06, |
|
"loss": 1.5671, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_accuracy": 0.3356643356643357, |
|
"eval_loss": 1.5932137966156006, |
|
"eval_runtime": 4.5483, |
|
"eval_samples_per_second": 62.881, |
|
"eval_steps_per_second": 3.958, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 4.95, |
|
"grad_norm": 0.7324579358100891, |
|
"learning_rate": 6.257142857142858e-06, |
|
"loss": 1.6757, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 4.95, |
|
"eval_accuracy": 0.32517482517482516, |
|
"eval_loss": 1.5901715755462646, |
|
"eval_runtime": 4.9986, |
|
"eval_samples_per_second": 57.216, |
|
"eval_steps_per_second": 3.601, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 5.97, |
|
"grad_norm": 0.878238320350647, |
|
"learning_rate": 7.542857142857144e-06, |
|
"loss": 1.5595, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 5.97, |
|
"eval_accuracy": 0.32167832167832167, |
|
"eval_loss": 1.5863642692565918, |
|
"eval_runtime": 4.9711, |
|
"eval_samples_per_second": 57.532, |
|
"eval_steps_per_second": 3.621, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 6.98, |
|
"grad_norm": 1.4210667610168457, |
|
"learning_rate": 8.828571428571429e-06, |
|
"loss": 1.5536, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 6.98, |
|
"eval_accuracy": 0.3181818181818182, |
|
"eval_loss": 1.5817956924438477, |
|
"eval_runtime": 5.5676, |
|
"eval_samples_per_second": 51.368, |
|
"eval_steps_per_second": 3.233, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"grad_norm": 0.49261531233787537, |
|
"learning_rate": 1.0114285714285715e-05, |
|
"loss": 1.5484, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_accuracy": 0.3111888111888112, |
|
"eval_loss": 1.576475739479065, |
|
"eval_runtime": 5.5008, |
|
"eval_samples_per_second": 51.992, |
|
"eval_steps_per_second": 3.272, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 8.95, |
|
"grad_norm": 0.9810589551925659, |
|
"learning_rate": 1.1314285714285715e-05, |
|
"loss": 1.6506, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 8.95, |
|
"eval_accuracy": 0.3076923076923077, |
|
"eval_loss": 1.570804238319397, |
|
"eval_runtime": 4.0121, |
|
"eval_samples_per_second": 71.285, |
|
"eval_steps_per_second": 4.486, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 9.97, |
|
"grad_norm": 0.6308433413505554, |
|
"learning_rate": 1.26e-05, |
|
"loss": 1.5317, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 9.97, |
|
"eval_accuracy": 0.3006993006993007, |
|
"eval_loss": 1.5640877485275269, |
|
"eval_runtime": 5.5443, |
|
"eval_samples_per_second": 51.585, |
|
"eval_steps_per_second": 3.247, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 10.98, |
|
"grad_norm": 1.381785273551941, |
|
"learning_rate": 1.3885714285714286e-05, |
|
"loss": 1.5226, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 10.98, |
|
"eval_accuracy": 0.2867132867132867, |
|
"eval_loss": 1.5564289093017578, |
|
"eval_runtime": 5.0151, |
|
"eval_samples_per_second": 57.027, |
|
"eval_steps_per_second": 3.589, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"grad_norm": 0.552884578704834, |
|
"learning_rate": 1.517142857142857e-05, |
|
"loss": 1.5116, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_accuracy": 0.2692307692307692, |
|
"eval_loss": 1.5478310585021973, |
|
"eval_runtime": 5.4183, |
|
"eval_samples_per_second": 52.784, |
|
"eval_steps_per_second": 3.322, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 12.95, |
|
"grad_norm": 0.8657445907592773, |
|
"learning_rate": 1.6371428571428572e-05, |
|
"loss": 1.6046, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 12.95, |
|
"eval_accuracy": 0.26223776223776224, |
|
"eval_loss": 1.5384888648986816, |
|
"eval_runtime": 5.1185, |
|
"eval_samples_per_second": 55.875, |
|
"eval_steps_per_second": 3.517, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 13.97, |
|
"grad_norm": 0.8375154733657837, |
|
"learning_rate": 1.7657142857142857e-05, |
|
"loss": 1.4822, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 13.97, |
|
"eval_accuracy": 0.25524475524475526, |
|
"eval_loss": 1.5284570455551147, |
|
"eval_runtime": 5.5939, |
|
"eval_samples_per_second": 51.127, |
|
"eval_steps_per_second": 3.218, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 14.98, |
|
"grad_norm": 0.7776028513908386, |
|
"learning_rate": 1.8942857142857145e-05, |
|
"loss": 1.4614, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 14.98, |
|
"eval_accuracy": 0.25524475524475526, |
|
"eval_loss": 1.5183566808700562, |
|
"eval_runtime": 5.2849, |
|
"eval_samples_per_second": 54.116, |
|
"eval_steps_per_second": 3.406, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"grad_norm": 0.3529152572154999, |
|
"learning_rate": 2.022857142857143e-05, |
|
"loss": 1.4396, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 16.0, |
|
"eval_accuracy": 0.2517482517482518, |
|
"eval_loss": 1.510148048400879, |
|
"eval_runtime": 6.3848, |
|
"eval_samples_per_second": 44.794, |
|
"eval_steps_per_second": 2.819, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 16.95, |
|
"grad_norm": 0.5623555779457092, |
|
"learning_rate": 2.1428571428571428e-05, |
|
"loss": 1.5047, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 16.95, |
|
"eval_accuracy": 0.24475524475524477, |
|
"eval_loss": 1.510407567024231, |
|
"eval_runtime": 5.762, |
|
"eval_samples_per_second": 49.635, |
|
"eval_steps_per_second": 3.124, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 17.97, |
|
"grad_norm": 0.889561116695404, |
|
"learning_rate": 2.2714285714285713e-05, |
|
"loss": 1.3741, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 17.97, |
|
"eval_accuracy": 0.2517482517482518, |
|
"eval_loss": 1.524996280670166, |
|
"eval_runtime": 5.3825, |
|
"eval_samples_per_second": 53.135, |
|
"eval_steps_per_second": 3.344, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 18.98, |
|
"grad_norm": 0.837506890296936, |
|
"learning_rate": 2.4e-05, |
|
"loss": 1.3512, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 18.98, |
|
"eval_accuracy": 0.26573426573426573, |
|
"eval_loss": 1.5328779220581055, |
|
"eval_runtime": 4.8496, |
|
"eval_samples_per_second": 58.974, |
|
"eval_steps_per_second": 3.712, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"grad_norm": 0.6774041056632996, |
|
"learning_rate": 2.5285714285714285e-05, |
|
"loss": 1.3286, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"eval_accuracy": 0.32517482517482516, |
|
"eval_loss": 1.509504795074463, |
|
"eval_runtime": 4.7142, |
|
"eval_samples_per_second": 60.667, |
|
"eval_steps_per_second": 3.818, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 20.95, |
|
"grad_norm": 0.5629450678825378, |
|
"learning_rate": 2.6485714285714287e-05, |
|
"loss": 1.3967, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 20.95, |
|
"eval_accuracy": 0.34965034965034963, |
|
"eval_loss": 1.482858419418335, |
|
"eval_runtime": 5.0574, |
|
"eval_samples_per_second": 56.551, |
|
"eval_steps_per_second": 3.559, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 21.97, |
|
"grad_norm": 0.5265232920646667, |
|
"learning_rate": 2.7771428571428572e-05, |
|
"loss": 1.2779, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 21.97, |
|
"eval_accuracy": 0.38461538461538464, |
|
"eval_loss": 1.4598273038864136, |
|
"eval_runtime": 5.2206, |
|
"eval_samples_per_second": 54.783, |
|
"eval_steps_per_second": 3.448, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 22.98, |
|
"grad_norm": 0.4762378931045532, |
|
"learning_rate": 2.9057142857142856e-05, |
|
"loss": 1.2449, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 22.98, |
|
"eval_accuracy": 0.4160839160839161, |
|
"eval_loss": 1.4284018278121948, |
|
"eval_runtime": 5.3164, |
|
"eval_samples_per_second": 53.796, |
|
"eval_steps_per_second": 3.386, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"grad_norm": 0.598441481590271, |
|
"learning_rate": 2.996190476190476e-05, |
|
"loss": 1.2118, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 24.0, |
|
"eval_accuracy": 0.4230769230769231, |
|
"eval_loss": 1.4161577224731445, |
|
"eval_runtime": 4.9587, |
|
"eval_samples_per_second": 57.677, |
|
"eval_steps_per_second": 3.63, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 24.95, |
|
"grad_norm": 1.1748360395431519, |
|
"learning_rate": 2.982857142857143e-05, |
|
"loss": 1.2521, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 24.95, |
|
"eval_accuracy": 0.44755244755244755, |
|
"eval_loss": 1.3797944784164429, |
|
"eval_runtime": 7.057, |
|
"eval_samples_per_second": 40.527, |
|
"eval_steps_per_second": 2.551, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 25.97, |
|
"grad_norm": 0.7063133716583252, |
|
"learning_rate": 2.9685714285714284e-05, |
|
"loss": 1.1183, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 25.97, |
|
"eval_accuracy": 0.479020979020979, |
|
"eval_loss": 1.324568271636963, |
|
"eval_runtime": 5.5598, |
|
"eval_samples_per_second": 51.44, |
|
"eval_steps_per_second": 3.238, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 26.98, |
|
"grad_norm": 0.91399085521698, |
|
"learning_rate": 2.9542857142857142e-05, |
|
"loss": 1.0778, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 26.98, |
|
"eval_accuracy": 0.493006993006993, |
|
"eval_loss": 1.276139259338379, |
|
"eval_runtime": 4.8059, |
|
"eval_samples_per_second": 59.51, |
|
"eval_steps_per_second": 3.745, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 28.0, |
|
"grad_norm": 1.1466203927993774, |
|
"learning_rate": 2.94e-05, |
|
"loss": 1.0306, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 28.0, |
|
"eval_accuracy": 0.5104895104895105, |
|
"eval_loss": 1.2299922704696655, |
|
"eval_runtime": 6.0199, |
|
"eval_samples_per_second": 47.509, |
|
"eval_steps_per_second": 2.99, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 28.95, |
|
"grad_norm": 1.1782015562057495, |
|
"learning_rate": 2.9266666666666665e-05, |
|
"loss": 1.0808, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 28.95, |
|
"eval_accuracy": 0.513986013986014, |
|
"eval_loss": 1.2064085006713867, |
|
"eval_runtime": 6.0946, |
|
"eval_samples_per_second": 46.927, |
|
"eval_steps_per_second": 2.953, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 29.97, |
|
"grad_norm": 0.9410634636878967, |
|
"learning_rate": 2.9123809523809523e-05, |
|
"loss": 0.9767, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 29.97, |
|
"eval_accuracy": 0.5524475524475524, |
|
"eval_loss": 1.1669567823410034, |
|
"eval_runtime": 4.8236, |
|
"eval_samples_per_second": 59.292, |
|
"eval_steps_per_second": 3.732, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 30.98, |
|
"grad_norm": 1.2911593914031982, |
|
"learning_rate": 2.898095238095238e-05, |
|
"loss": 0.9589, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 30.98, |
|
"eval_accuracy": 0.5734265734265734, |
|
"eval_loss": 1.126379132270813, |
|
"eval_runtime": 5.0702, |
|
"eval_samples_per_second": 56.408, |
|
"eval_steps_per_second": 3.55, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 32.0, |
|
"grad_norm": 1.2354660034179688, |
|
"learning_rate": 2.883809523809524e-05, |
|
"loss": 0.9193, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 32.0, |
|
"eval_accuracy": 0.5874125874125874, |
|
"eval_loss": 1.1012728214263916, |
|
"eval_runtime": 5.9275, |
|
"eval_samples_per_second": 48.249, |
|
"eval_steps_per_second": 3.037, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 32.95, |
|
"grad_norm": 0.8961493968963623, |
|
"learning_rate": 2.8704761904761905e-05, |
|
"loss": 0.9462, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 32.95, |
|
"eval_accuracy": 0.6048951048951049, |
|
"eval_loss": 1.0736197233200073, |
|
"eval_runtime": 5.7691, |
|
"eval_samples_per_second": 49.575, |
|
"eval_steps_per_second": 3.12, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 33.97, |
|
"grad_norm": 0.9503061175346375, |
|
"learning_rate": 2.8561904761904763e-05, |
|
"loss": 0.85, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 33.97, |
|
"eval_accuracy": 0.6048951048951049, |
|
"eval_loss": 1.0628284215927124, |
|
"eval_runtime": 6.1054, |
|
"eval_samples_per_second": 46.844, |
|
"eval_steps_per_second": 2.948, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 34.98, |
|
"grad_norm": 1.0706520080566406, |
|
"learning_rate": 2.841904761904762e-05, |
|
"loss": 0.8294, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 34.98, |
|
"eval_accuracy": 0.6188811188811189, |
|
"eval_loss": 1.0472813844680786, |
|
"eval_runtime": 4.7734, |
|
"eval_samples_per_second": 59.916, |
|
"eval_steps_per_second": 3.771, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 36.0, |
|
"grad_norm": 0.9784532189369202, |
|
"learning_rate": 2.827619047619048e-05, |
|
"loss": 0.8025, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 36.0, |
|
"eval_accuracy": 0.6328671328671329, |
|
"eval_loss": 1.0030184984207153, |
|
"eval_runtime": 5.5291, |
|
"eval_samples_per_second": 51.726, |
|
"eval_steps_per_second": 3.255, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 36.95, |
|
"grad_norm": 0.8138810396194458, |
|
"learning_rate": 2.8142857142857145e-05, |
|
"loss": 0.8206, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 36.95, |
|
"eval_accuracy": 0.6398601398601399, |
|
"eval_loss": 0.9963561296463013, |
|
"eval_runtime": 5.8221, |
|
"eval_samples_per_second": 49.123, |
|
"eval_steps_per_second": 3.092, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 37.97, |
|
"grad_norm": 0.8114917278289795, |
|
"learning_rate": 2.8e-05, |
|
"loss": 0.7541, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 37.97, |
|
"eval_accuracy": 0.6608391608391608, |
|
"eval_loss": 0.9604987502098083, |
|
"eval_runtime": 4.9473, |
|
"eval_samples_per_second": 57.809, |
|
"eval_steps_per_second": 3.638, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 38.98, |
|
"grad_norm": 0.9221667051315308, |
|
"learning_rate": 2.7857142857142858e-05, |
|
"loss": 0.7413, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 38.98, |
|
"eval_accuracy": 0.6643356643356644, |
|
"eval_loss": 0.946722686290741, |
|
"eval_runtime": 4.7955, |
|
"eval_samples_per_second": 59.64, |
|
"eval_steps_per_second": 3.754, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"grad_norm": 1.2665307521820068, |
|
"learning_rate": 2.7714285714285716e-05, |
|
"loss": 0.709, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"eval_accuracy": 0.6678321678321678, |
|
"eval_loss": 0.9348079562187195, |
|
"eval_runtime": 5.3252, |
|
"eval_samples_per_second": 53.707, |
|
"eval_steps_per_second": 3.38, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 40.95, |
|
"grad_norm": 1.0893254280090332, |
|
"learning_rate": 2.758095238095238e-05, |
|
"loss": 0.7817, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 40.95, |
|
"eval_accuracy": 0.6678321678321678, |
|
"eval_loss": 0.9366168975830078, |
|
"eval_runtime": 6.4862, |
|
"eval_samples_per_second": 44.094, |
|
"eval_steps_per_second": 2.775, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 41.97, |
|
"grad_norm": 0.7136903405189514, |
|
"learning_rate": 2.743809523809524e-05, |
|
"loss": 0.7034, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 41.97, |
|
"eval_accuracy": 0.6818181818181818, |
|
"eval_loss": 0.9109411835670471, |
|
"eval_runtime": 6.1724, |
|
"eval_samples_per_second": 46.335, |
|
"eval_steps_per_second": 2.916, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 42.98, |
|
"grad_norm": 1.142288327217102, |
|
"learning_rate": 2.7295238095238097e-05, |
|
"loss": 0.6856, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 42.98, |
|
"eval_accuracy": 0.6573426573426573, |
|
"eval_loss": 0.9276965856552124, |
|
"eval_runtime": 4.6805, |
|
"eval_samples_per_second": 61.105, |
|
"eval_steps_per_second": 3.846, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 44.0, |
|
"grad_norm": 1.3661341667175293, |
|
"learning_rate": 2.7152380952380952e-05, |
|
"loss": 0.6625, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 44.0, |
|
"eval_accuracy": 0.6783216783216783, |
|
"eval_loss": 0.8979520201683044, |
|
"eval_runtime": 5.8084, |
|
"eval_samples_per_second": 49.239, |
|
"eval_steps_per_second": 3.099, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 44.95, |
|
"grad_norm": 0.7071289420127869, |
|
"learning_rate": 2.701904761904762e-05, |
|
"loss": 0.7207, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 44.95, |
|
"eval_accuracy": 0.6713286713286714, |
|
"eval_loss": 0.9049975275993347, |
|
"eval_runtime": 5.5071, |
|
"eval_samples_per_second": 51.933, |
|
"eval_steps_per_second": 3.268, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 45.97, |
|
"grad_norm": 0.9930222034454346, |
|
"learning_rate": 2.6876190476190476e-05, |
|
"loss": 0.6684, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 45.97, |
|
"eval_accuracy": 0.6748251748251748, |
|
"eval_loss": 0.8973459005355835, |
|
"eval_runtime": 4.8744, |
|
"eval_samples_per_second": 58.674, |
|
"eval_steps_per_second": 3.693, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 46.98, |
|
"grad_norm": 1.2905802726745605, |
|
"learning_rate": 2.6733333333333334e-05, |
|
"loss": 0.6651, |
|
"step": 693 |
|
}, |
|
{ |
|
"epoch": 46.98, |
|
"eval_accuracy": 0.6783216783216783, |
|
"eval_loss": 0.8934686779975891, |
|
"eval_runtime": 4.9699, |
|
"eval_samples_per_second": 57.546, |
|
"eval_steps_per_second": 3.622, |
|
"step": 693 |
|
}, |
|
{ |
|
"epoch": 48.0, |
|
"grad_norm": 1.5243291854858398, |
|
"learning_rate": 2.6590476190476192e-05, |
|
"loss": 0.6451, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 48.0, |
|
"eval_accuracy": 0.6748251748251748, |
|
"eval_loss": 0.8944538831710815, |
|
"eval_runtime": 4.9181, |
|
"eval_samples_per_second": 58.153, |
|
"eval_steps_per_second": 3.66, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 48.95, |
|
"grad_norm": 1.4421076774597168, |
|
"learning_rate": 2.6457142857142857e-05, |
|
"loss": 0.6774, |
|
"step": 722 |
|
}, |
|
{ |
|
"epoch": 48.95, |
|
"eval_accuracy": 0.6818181818181818, |
|
"eval_loss": 0.8879284858703613, |
|
"eval_runtime": 5.1997, |
|
"eval_samples_per_second": 55.003, |
|
"eval_steps_per_second": 3.462, |
|
"step": 722 |
|
}, |
|
{ |
|
"epoch": 49.97, |
|
"grad_norm": 1.3786752223968506, |
|
"learning_rate": 2.6314285714285715e-05, |
|
"loss": 0.6308, |
|
"step": 737 |
|
}, |
|
{ |
|
"epoch": 49.97, |
|
"eval_accuracy": 0.6818181818181818, |
|
"eval_loss": 0.8882840275764465, |
|
"eval_runtime": 5.1707, |
|
"eval_samples_per_second": 55.312, |
|
"eval_steps_per_second": 3.481, |
|
"step": 737 |
|
}, |
|
{ |
|
"epoch": 50.98, |
|
"grad_norm": 1.9291437864303589, |
|
"learning_rate": 2.617142857142857e-05, |
|
"loss": 0.6199, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 50.98, |
|
"eval_accuracy": 0.6818181818181818, |
|
"eval_loss": 0.8825680017471313, |
|
"eval_runtime": 4.9861, |
|
"eval_samples_per_second": 57.359, |
|
"eval_steps_per_second": 3.61, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 52.0, |
|
"grad_norm": 1.2054837942123413, |
|
"learning_rate": 2.6028571428571428e-05, |
|
"loss": 0.6379, |
|
"step": 767 |
|
}, |
|
{ |
|
"epoch": 52.0, |
|
"eval_accuracy": 0.6923076923076923, |
|
"eval_loss": 0.8582118153572083, |
|
"eval_runtime": 5.287, |
|
"eval_samples_per_second": 54.095, |
|
"eval_steps_per_second": 3.405, |
|
"step": 767 |
|
}, |
|
{ |
|
"epoch": 52.95, |
|
"grad_norm": 1.3063265085220337, |
|
"learning_rate": 2.5895238095238094e-05, |
|
"loss": 0.6588, |
|
"step": 781 |
|
}, |
|
{ |
|
"epoch": 52.95, |
|
"eval_accuracy": 0.6818181818181818, |
|
"eval_loss": 0.8825483918190002, |
|
"eval_runtime": 4.6024, |
|
"eval_samples_per_second": 62.141, |
|
"eval_steps_per_second": 3.911, |
|
"step": 781 |
|
}, |
|
{ |
|
"epoch": 53.97, |
|
"grad_norm": 1.2957918643951416, |
|
"learning_rate": 2.5752380952380952e-05, |
|
"loss": 0.5857, |
|
"step": 796 |
|
}, |
|
{ |
|
"epoch": 53.97, |
|
"eval_accuracy": 0.6748251748251748, |
|
"eval_loss": 0.8808070421218872, |
|
"eval_runtime": 5.2003, |
|
"eval_samples_per_second": 54.997, |
|
"eval_steps_per_second": 3.461, |
|
"step": 796 |
|
}, |
|
{ |
|
"epoch": 54.98, |
|
"grad_norm": 0.9337932467460632, |
|
"learning_rate": 2.560952380952381e-05, |
|
"loss": 0.6076, |
|
"step": 811 |
|
}, |
|
{ |
|
"epoch": 54.98, |
|
"eval_accuracy": 0.6958041958041958, |
|
"eval_loss": 0.8554547429084778, |
|
"eval_runtime": 5.108, |
|
"eval_samples_per_second": 55.991, |
|
"eval_steps_per_second": 3.524, |
|
"step": 811 |
|
}, |
|
{ |
|
"epoch": 56.0, |
|
"grad_norm": 1.4797887802124023, |
|
"learning_rate": 2.5466666666666668e-05, |
|
"loss": 0.5934, |
|
"step": 826 |
|
}, |
|
{ |
|
"epoch": 56.0, |
|
"eval_accuracy": 0.6888111888111889, |
|
"eval_loss": 0.8653700351715088, |
|
"eval_runtime": 4.9201, |
|
"eval_samples_per_second": 58.128, |
|
"eval_steps_per_second": 3.658, |
|
"step": 826 |
|
}, |
|
{ |
|
"epoch": 56.95, |
|
"grad_norm": 0.9749770164489746, |
|
"learning_rate": 2.5333333333333334e-05, |
|
"loss": 0.6427, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 56.95, |
|
"eval_accuracy": 0.6853146853146853, |
|
"eval_loss": 0.861566424369812, |
|
"eval_runtime": 5.5361, |
|
"eval_samples_per_second": 51.661, |
|
"eval_steps_per_second": 3.251, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 57.97, |
|
"grad_norm": 1.1588115692138672, |
|
"learning_rate": 2.519047619047619e-05, |
|
"loss": 0.5782, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 57.97, |
|
"eval_accuracy": 0.6678321678321678, |
|
"eval_loss": 0.8711130619049072, |
|
"eval_runtime": 5.1513, |
|
"eval_samples_per_second": 55.52, |
|
"eval_steps_per_second": 3.494, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 58.98, |
|
"grad_norm": 2.290128707885742, |
|
"learning_rate": 2.504761904761905e-05, |
|
"loss": 0.5819, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 58.98, |
|
"eval_accuracy": 0.6748251748251748, |
|
"eval_loss": 0.8689377903938293, |
|
"eval_runtime": 5.3476, |
|
"eval_samples_per_second": 53.482, |
|
"eval_steps_per_second": 3.366, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 60.0, |
|
"grad_norm": 1.8434489965438843, |
|
"learning_rate": 2.4904761904761908e-05, |
|
"loss": 0.5918, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 60.0, |
|
"eval_accuracy": 0.6923076923076923, |
|
"eval_loss": 0.8602246642112732, |
|
"eval_runtime": 5.1182, |
|
"eval_samples_per_second": 55.879, |
|
"eval_steps_per_second": 3.517, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 60.95, |
|
"grad_norm": 1.1310744285583496, |
|
"learning_rate": 2.4771428571428573e-05, |
|
"loss": 0.5845, |
|
"step": 899 |
|
}, |
|
{ |
|
"epoch": 60.95, |
|
"eval_accuracy": 0.6993006993006993, |
|
"eval_loss": 0.8458691835403442, |
|
"eval_runtime": 4.8028, |
|
"eval_samples_per_second": 59.549, |
|
"eval_steps_per_second": 3.748, |
|
"step": 899 |
|
}, |
|
{ |
|
"epoch": 61.97, |
|
"grad_norm": 1.319384217262268, |
|
"learning_rate": 2.462857142857143e-05, |
|
"loss": 0.5667, |
|
"step": 914 |
|
}, |
|
{ |
|
"epoch": 61.97, |
|
"eval_accuracy": 0.7027972027972028, |
|
"eval_loss": 0.8466892838478088, |
|
"eval_runtime": 4.9074, |
|
"eval_samples_per_second": 58.279, |
|
"eval_steps_per_second": 3.668, |
|
"step": 914 |
|
}, |
|
{ |
|
"epoch": 62.98, |
|
"grad_norm": 1.2371602058410645, |
|
"learning_rate": 2.448571428571429e-05, |
|
"loss": 0.5327, |
|
"step": 929 |
|
}, |
|
{ |
|
"epoch": 62.98, |
|
"eval_accuracy": 0.6923076923076923, |
|
"eval_loss": 0.8539786338806152, |
|
"eval_runtime": 5.32, |
|
"eval_samples_per_second": 53.759, |
|
"eval_steps_per_second": 3.383, |
|
"step": 929 |
|
}, |
|
{ |
|
"epoch": 64.0, |
|
"grad_norm": 2.1778481006622314, |
|
"learning_rate": 2.434285714285714e-05, |
|
"loss": 0.523, |
|
"step": 944 |
|
}, |
|
{ |
|
"epoch": 64.0, |
|
"eval_accuracy": 0.7062937062937062, |
|
"eval_loss": 0.8323072791099548, |
|
"eval_runtime": 5.1152, |
|
"eval_samples_per_second": 55.912, |
|
"eval_steps_per_second": 3.519, |
|
"step": 944 |
|
}, |
|
{ |
|
"epoch": 64.95, |
|
"grad_norm": 0.8219490051269531, |
|
"learning_rate": 2.420952380952381e-05, |
|
"loss": 0.548, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 64.95, |
|
"eval_accuracy": 0.6993006993006993, |
|
"eval_loss": 0.8406782746315002, |
|
"eval_runtime": 5.325, |
|
"eval_samples_per_second": 53.709, |
|
"eval_steps_per_second": 3.38, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 65.97, |
|
"grad_norm": 2.0205609798431396, |
|
"learning_rate": 2.4066666666666664e-05, |
|
"loss": 0.5399, |
|
"step": 973 |
|
}, |
|
{ |
|
"epoch": 65.97, |
|
"eval_accuracy": 0.6993006993006993, |
|
"eval_loss": 0.837882936000824, |
|
"eval_runtime": 5.6808, |
|
"eval_samples_per_second": 50.345, |
|
"eval_steps_per_second": 3.169, |
|
"step": 973 |
|
}, |
|
{ |
|
"epoch": 66.98, |
|
"grad_norm": 1.527024507522583, |
|
"learning_rate": 2.3923809523809522e-05, |
|
"loss": 0.5324, |
|
"step": 988 |
|
}, |
|
{ |
|
"epoch": 66.98, |
|
"eval_accuracy": 0.7027972027972028, |
|
"eval_loss": 0.8118743896484375, |
|
"eval_runtime": 5.543, |
|
"eval_samples_per_second": 51.596, |
|
"eval_steps_per_second": 3.247, |
|
"step": 988 |
|
}, |
|
{ |
|
"epoch": 68.0, |
|
"grad_norm": 1.3294800519943237, |
|
"learning_rate": 2.378095238095238e-05, |
|
"loss": 0.5171, |
|
"step": 1003 |
|
}, |
|
{ |
|
"epoch": 68.0, |
|
"eval_accuracy": 0.6923076923076923, |
|
"eval_loss": 0.8444966673851013, |
|
"eval_runtime": 5.5571, |
|
"eval_samples_per_second": 51.465, |
|
"eval_steps_per_second": 3.239, |
|
"step": 1003 |
|
}, |
|
{ |
|
"epoch": 68.95, |
|
"grad_norm": 1.6269547939300537, |
|
"learning_rate": 2.3647619047619046e-05, |
|
"loss": 0.538, |
|
"step": 1017 |
|
}, |
|
{ |
|
"epoch": 68.95, |
|
"eval_accuracy": 0.7097902097902098, |
|
"eval_loss": 0.8196296095848083, |
|
"eval_runtime": 4.8636, |
|
"eval_samples_per_second": 58.805, |
|
"eval_steps_per_second": 3.701, |
|
"step": 1017 |
|
}, |
|
{ |
|
"epoch": 69.97, |
|
"grad_norm": 1.0726577043533325, |
|
"learning_rate": 2.3504761904761904e-05, |
|
"loss": 0.5312, |
|
"step": 1032 |
|
}, |
|
{ |
|
"epoch": 69.97, |
|
"eval_accuracy": 0.6853146853146853, |
|
"eval_loss": 0.8414965271949768, |
|
"eval_runtime": 5.2322, |
|
"eval_samples_per_second": 54.661, |
|
"eval_steps_per_second": 3.44, |
|
"step": 1032 |
|
}, |
|
{ |
|
"epoch": 70.98, |
|
"grad_norm": 1.0595276355743408, |
|
"learning_rate": 2.3361904761904762e-05, |
|
"loss": 0.4914, |
|
"step": 1047 |
|
}, |
|
{ |
|
"epoch": 70.98, |
|
"eval_accuracy": 0.6958041958041958, |
|
"eval_loss": 0.8184240460395813, |
|
"eval_runtime": 4.6109, |
|
"eval_samples_per_second": 62.027, |
|
"eval_steps_per_second": 3.904, |
|
"step": 1047 |
|
}, |
|
{ |
|
"epoch": 72.0, |
|
"grad_norm": 1.3313194513320923, |
|
"learning_rate": 2.321904761904762e-05, |
|
"loss": 0.5055, |
|
"step": 1062 |
|
}, |
|
{ |
|
"epoch": 72.0, |
|
"eval_accuracy": 0.6923076923076923, |
|
"eval_loss": 0.8218348622322083, |
|
"eval_runtime": 5.5535, |
|
"eval_samples_per_second": 51.499, |
|
"eval_steps_per_second": 3.241, |
|
"step": 1062 |
|
}, |
|
{ |
|
"epoch": 72.95, |
|
"grad_norm": 1.9882231950759888, |
|
"learning_rate": 2.3085714285714286e-05, |
|
"loss": 0.5401, |
|
"step": 1076 |
|
}, |
|
{ |
|
"epoch": 72.95, |
|
"eval_accuracy": 0.7027972027972028, |
|
"eval_loss": 0.815979540348053, |
|
"eval_runtime": 6.0512, |
|
"eval_samples_per_second": 47.264, |
|
"eval_steps_per_second": 2.975, |
|
"step": 1076 |
|
}, |
|
{ |
|
"epoch": 73.97, |
|
"grad_norm": 1.8936933279037476, |
|
"learning_rate": 2.2942857142857144e-05, |
|
"loss": 0.4966, |
|
"step": 1091 |
|
}, |
|
{ |
|
"epoch": 73.97, |
|
"eval_accuracy": 0.6888111888111889, |
|
"eval_loss": 0.8237566351890564, |
|
"eval_runtime": 4.6622, |
|
"eval_samples_per_second": 61.344, |
|
"eval_steps_per_second": 3.861, |
|
"step": 1091 |
|
}, |
|
{ |
|
"epoch": 74.98, |
|
"grad_norm": 2.0124337673187256, |
|
"learning_rate": 2.2800000000000002e-05, |
|
"loss": 0.4768, |
|
"step": 1106 |
|
}, |
|
{ |
|
"epoch": 74.98, |
|
"eval_accuracy": 0.6993006993006993, |
|
"eval_loss": 0.8185241222381592, |
|
"eval_runtime": 4.2442, |
|
"eval_samples_per_second": 67.386, |
|
"eval_steps_per_second": 4.241, |
|
"step": 1106 |
|
}, |
|
{ |
|
"epoch": 76.0, |
|
"grad_norm": 1.7522495985031128, |
|
"learning_rate": 2.265714285714286e-05, |
|
"loss": 0.4789, |
|
"step": 1121 |
|
}, |
|
{ |
|
"epoch": 76.0, |
|
"eval_accuracy": 0.7027972027972028, |
|
"eval_loss": 0.8260769844055176, |
|
"eval_runtime": 6.128, |
|
"eval_samples_per_second": 46.671, |
|
"eval_steps_per_second": 2.937, |
|
"step": 1121 |
|
}, |
|
{ |
|
"epoch": 76.95, |
|
"grad_norm": 1.8254278898239136, |
|
"learning_rate": 2.2523809523809526e-05, |
|
"loss": 0.5176, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 76.95, |
|
"eval_accuracy": 0.7097902097902098, |
|
"eval_loss": 0.8109525442123413, |
|
"eval_runtime": 6.0147, |
|
"eval_samples_per_second": 47.55, |
|
"eval_steps_per_second": 2.993, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 77.97, |
|
"grad_norm": 1.4586073160171509, |
|
"learning_rate": 2.238095238095238e-05, |
|
"loss": 0.466, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 77.97, |
|
"eval_accuracy": 0.6993006993006993, |
|
"eval_loss": 0.814057469367981, |
|
"eval_runtime": 4.782, |
|
"eval_samples_per_second": 59.808, |
|
"eval_steps_per_second": 3.764, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 78.98, |
|
"grad_norm": 1.7543085813522339, |
|
"learning_rate": 2.223809523809524e-05, |
|
"loss": 0.4736, |
|
"step": 1165 |
|
}, |
|
{ |
|
"epoch": 78.98, |
|
"eval_accuracy": 0.7167832167832168, |
|
"eval_loss": 0.7970029711723328, |
|
"eval_runtime": 5.8124, |
|
"eval_samples_per_second": 49.205, |
|
"eval_steps_per_second": 3.097, |
|
"step": 1165 |
|
}, |
|
{ |
|
"epoch": 80.0, |
|
"grad_norm": 1.6676491498947144, |
|
"learning_rate": 2.2095238095238096e-05, |
|
"loss": 0.4785, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 80.0, |
|
"eval_accuracy": 0.7097902097902098, |
|
"eval_loss": 0.8062326312065125, |
|
"eval_runtime": 4.5174, |
|
"eval_samples_per_second": 63.31, |
|
"eval_steps_per_second": 3.985, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 80.95, |
|
"grad_norm": 1.827528476715088, |
|
"learning_rate": 2.1961904761904762e-05, |
|
"loss": 0.5309, |
|
"step": 1194 |
|
}, |
|
{ |
|
"epoch": 80.95, |
|
"eval_accuracy": 0.6958041958041958, |
|
"eval_loss": 0.8051398396492004, |
|
"eval_runtime": 4.9397, |
|
"eval_samples_per_second": 57.899, |
|
"eval_steps_per_second": 3.644, |
|
"step": 1194 |
|
}, |
|
{ |
|
"epoch": 81.97, |
|
"grad_norm": 2.3700757026672363, |
|
"learning_rate": 2.181904761904762e-05, |
|
"loss": 0.4571, |
|
"step": 1209 |
|
}, |
|
{ |
|
"epoch": 81.97, |
|
"eval_accuracy": 0.7097902097902098, |
|
"eval_loss": 0.8024002909660339, |
|
"eval_runtime": 6.3277, |
|
"eval_samples_per_second": 45.198, |
|
"eval_steps_per_second": 2.845, |
|
"step": 1209 |
|
}, |
|
{ |
|
"epoch": 82.98, |
|
"grad_norm": 1.5626410245895386, |
|
"learning_rate": 2.1676190476190478e-05, |
|
"loss": 0.47, |
|
"step": 1224 |
|
}, |
|
{ |
|
"epoch": 82.98, |
|
"eval_accuracy": 0.7167832167832168, |
|
"eval_loss": 0.8030509948730469, |
|
"eval_runtime": 5.4575, |
|
"eval_samples_per_second": 52.405, |
|
"eval_steps_per_second": 3.298, |
|
"step": 1224 |
|
}, |
|
{ |
|
"epoch": 84.0, |
|
"grad_norm": 1.4918785095214844, |
|
"learning_rate": 2.1533333333333333e-05, |
|
"loss": 0.4525, |
|
"step": 1239 |
|
}, |
|
{ |
|
"epoch": 84.0, |
|
"eval_accuracy": 0.7132867132867133, |
|
"eval_loss": 0.7910680174827576, |
|
"eval_runtime": 4.6183, |
|
"eval_samples_per_second": 61.928, |
|
"eval_steps_per_second": 3.898, |
|
"step": 1239 |
|
}, |
|
{ |
|
"epoch": 84.95, |
|
"grad_norm": 1.4631191492080688, |
|
"learning_rate": 2.1400000000000002e-05, |
|
"loss": 0.5058, |
|
"step": 1253 |
|
}, |
|
{ |
|
"epoch": 84.95, |
|
"eval_accuracy": 0.7132867132867133, |
|
"eval_loss": 0.7877373695373535, |
|
"eval_runtime": 4.6086, |
|
"eval_samples_per_second": 62.058, |
|
"eval_steps_per_second": 3.906, |
|
"step": 1253 |
|
}, |
|
{ |
|
"epoch": 85.97, |
|
"grad_norm": 1.463442087173462, |
|
"learning_rate": 2.1257142857142856e-05, |
|
"loss": 0.4627, |
|
"step": 1268 |
|
}, |
|
{ |
|
"epoch": 85.97, |
|
"eval_accuracy": 0.7062937062937062, |
|
"eval_loss": 0.7917687296867371, |
|
"eval_runtime": 6.4645, |
|
"eval_samples_per_second": 44.241, |
|
"eval_steps_per_second": 2.784, |
|
"step": 1268 |
|
}, |
|
{ |
|
"epoch": 86.98, |
|
"grad_norm": 1.5172946453094482, |
|
"learning_rate": 2.1114285714285714e-05, |
|
"loss": 0.4343, |
|
"step": 1283 |
|
}, |
|
{ |
|
"epoch": 86.98, |
|
"eval_accuracy": 0.7167832167832168, |
|
"eval_loss": 0.78819739818573, |
|
"eval_runtime": 5.3121, |
|
"eval_samples_per_second": 53.839, |
|
"eval_steps_per_second": 3.388, |
|
"step": 1283 |
|
}, |
|
{ |
|
"epoch": 88.0, |
|
"grad_norm": 1.8876270055770874, |
|
"learning_rate": 2.0971428571428572e-05, |
|
"loss": 0.4442, |
|
"step": 1298 |
|
}, |
|
{ |
|
"epoch": 88.0, |
|
"eval_accuracy": 0.7132867132867133, |
|
"eval_loss": 0.805809736251831, |
|
"eval_runtime": 5.4992, |
|
"eval_samples_per_second": 52.007, |
|
"eval_steps_per_second": 3.273, |
|
"step": 1298 |
|
}, |
|
{ |
|
"epoch": 88.95, |
|
"grad_norm": 1.4128847122192383, |
|
"learning_rate": 2.0838095238095238e-05, |
|
"loss": 0.4745, |
|
"step": 1312 |
|
}, |
|
{ |
|
"epoch": 88.95, |
|
"eval_accuracy": 0.7237762237762237, |
|
"eval_loss": 0.7810028791427612, |
|
"eval_runtime": 5.1967, |
|
"eval_samples_per_second": 55.035, |
|
"eval_steps_per_second": 3.464, |
|
"step": 1312 |
|
}, |
|
{ |
|
"epoch": 89.97, |
|
"grad_norm": 1.3974703550338745, |
|
"learning_rate": 2.0695238095238096e-05, |
|
"loss": 0.4282, |
|
"step": 1327 |
|
}, |
|
{ |
|
"epoch": 89.97, |
|
"eval_accuracy": 0.7097902097902098, |
|
"eval_loss": 0.7951435446739197, |
|
"eval_runtime": 5.9497, |
|
"eval_samples_per_second": 48.07, |
|
"eval_steps_per_second": 3.025, |
|
"step": 1327 |
|
}, |
|
{ |
|
"epoch": 90.98, |
|
"grad_norm": 1.8035740852355957, |
|
"learning_rate": 2.055238095238095e-05, |
|
"loss": 0.4307, |
|
"step": 1342 |
|
}, |
|
{ |
|
"epoch": 90.98, |
|
"eval_accuracy": 0.7167832167832168, |
|
"eval_loss": 0.7739275097846985, |
|
"eval_runtime": 5.0539, |
|
"eval_samples_per_second": 56.59, |
|
"eval_steps_per_second": 3.562, |
|
"step": 1342 |
|
}, |
|
{ |
|
"epoch": 92.0, |
|
"grad_norm": 2.104257822036743, |
|
"learning_rate": 2.040952380952381e-05, |
|
"loss": 0.4403, |
|
"step": 1357 |
|
}, |
|
{ |
|
"epoch": 92.0, |
|
"eval_accuracy": 0.7202797202797203, |
|
"eval_loss": 0.7788061499595642, |
|
"eval_runtime": 4.9104, |
|
"eval_samples_per_second": 58.244, |
|
"eval_steps_per_second": 3.666, |
|
"step": 1357 |
|
}, |
|
{ |
|
"epoch": 92.95, |
|
"grad_norm": 1.6670126914978027, |
|
"learning_rate": 2.0276190476190475e-05, |
|
"loss": 0.4567, |
|
"step": 1371 |
|
}, |
|
{ |
|
"epoch": 92.95, |
|
"eval_accuracy": 0.7167832167832168, |
|
"eval_loss": 0.7926862835884094, |
|
"eval_runtime": 5.6039, |
|
"eval_samples_per_second": 51.036, |
|
"eval_steps_per_second": 3.212, |
|
"step": 1371 |
|
}, |
|
{ |
|
"epoch": 93.97, |
|
"grad_norm": 0.9627218842506409, |
|
"learning_rate": 2.0133333333333333e-05, |
|
"loss": 0.4233, |
|
"step": 1386 |
|
}, |
|
{ |
|
"epoch": 93.97, |
|
"eval_accuracy": 0.7202797202797203, |
|
"eval_loss": 0.7885140180587769, |
|
"eval_runtime": 4.3051, |
|
"eval_samples_per_second": 66.433, |
|
"eval_steps_per_second": 4.181, |
|
"step": 1386 |
|
}, |
|
{ |
|
"epoch": 94.98, |
|
"grad_norm": 1.5669583082199097, |
|
"learning_rate": 1.999047619047619e-05, |
|
"loss": 0.4347, |
|
"step": 1401 |
|
}, |
|
{ |
|
"epoch": 94.98, |
|
"eval_accuracy": 0.7202797202797203, |
|
"eval_loss": 0.7849389314651489, |
|
"eval_runtime": 4.7453, |
|
"eval_samples_per_second": 60.27, |
|
"eval_steps_per_second": 3.793, |
|
"step": 1401 |
|
}, |
|
{ |
|
"epoch": 96.0, |
|
"grad_norm": 1.5593161582946777, |
|
"learning_rate": 1.984761904761905e-05, |
|
"loss": 0.4167, |
|
"step": 1416 |
|
}, |
|
{ |
|
"epoch": 96.0, |
|
"eval_accuracy": 0.7237762237762237, |
|
"eval_loss": 0.7880135774612427, |
|
"eval_runtime": 5.155, |
|
"eval_samples_per_second": 55.48, |
|
"eval_steps_per_second": 3.492, |
|
"step": 1416 |
|
}, |
|
{ |
|
"epoch": 96.95, |
|
"grad_norm": 1.337389349937439, |
|
"learning_rate": 1.9714285714285714e-05, |
|
"loss": 0.4394, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 96.95, |
|
"eval_accuracy": 0.7202797202797203, |
|
"eval_loss": 0.7888504862785339, |
|
"eval_runtime": 4.6698, |
|
"eval_samples_per_second": 61.244, |
|
"eval_steps_per_second": 3.855, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 97.97, |
|
"grad_norm": 1.8546737432479858, |
|
"learning_rate": 1.9571428571428572e-05, |
|
"loss": 0.4359, |
|
"step": 1445 |
|
}, |
|
{ |
|
"epoch": 97.97, |
|
"eval_accuracy": 0.7202797202797203, |
|
"eval_loss": 0.778459370136261, |
|
"eval_runtime": 4.7578, |
|
"eval_samples_per_second": 60.112, |
|
"eval_steps_per_second": 3.783, |
|
"step": 1445 |
|
}, |
|
{ |
|
"epoch": 98.98, |
|
"grad_norm": 1.3580291271209717, |
|
"learning_rate": 1.942857142857143e-05, |
|
"loss": 0.4085, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 98.98, |
|
"eval_accuracy": 0.7132867132867133, |
|
"eval_loss": 0.7851797342300415, |
|
"eval_runtime": 5.8892, |
|
"eval_samples_per_second": 48.564, |
|
"eval_steps_per_second": 3.056, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 100.0, |
|
"grad_norm": 1.4461547136306763, |
|
"learning_rate": 1.928571428571429e-05, |
|
"loss": 0.3965, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 100.0, |
|
"eval_accuracy": 0.7272727272727273, |
|
"eval_loss": 0.7784654498100281, |
|
"eval_runtime": 5.3557, |
|
"eval_samples_per_second": 53.401, |
|
"eval_steps_per_second": 3.361, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 100.95, |
|
"grad_norm": 1.765655755996704, |
|
"learning_rate": 1.9152380952380954e-05, |
|
"loss": 0.445, |
|
"step": 1489 |
|
}, |
|
{ |
|
"epoch": 100.95, |
|
"eval_accuracy": 0.7202797202797203, |
|
"eval_loss": 0.7826104760169983, |
|
"eval_runtime": 5.628, |
|
"eval_samples_per_second": 50.818, |
|
"eval_steps_per_second": 3.198, |
|
"step": 1489 |
|
}, |
|
{ |
|
"epoch": 101.97, |
|
"grad_norm": 1.7069743871688843, |
|
"learning_rate": 1.9009523809523812e-05, |
|
"loss": 0.3988, |
|
"step": 1504 |
|
}, |
|
{ |
|
"epoch": 101.97, |
|
"eval_accuracy": 0.7097902097902098, |
|
"eval_loss": 0.8045200109481812, |
|
"eval_runtime": 5.2204, |
|
"eval_samples_per_second": 54.785, |
|
"eval_steps_per_second": 3.448, |
|
"step": 1504 |
|
}, |
|
{ |
|
"epoch": 102.98, |
|
"grad_norm": 1.2846873998641968, |
|
"learning_rate": 1.886666666666667e-05, |
|
"loss": 0.4129, |
|
"step": 1519 |
|
}, |
|
{ |
|
"epoch": 102.98, |
|
"eval_accuracy": 0.7272727272727273, |
|
"eval_loss": 0.7685949206352234, |
|
"eval_runtime": 5.3375, |
|
"eval_samples_per_second": 53.583, |
|
"eval_steps_per_second": 3.372, |
|
"step": 1519 |
|
}, |
|
{ |
|
"epoch": 104.0, |
|
"grad_norm": 1.9040664434432983, |
|
"learning_rate": 1.872380952380952e-05, |
|
"loss": 0.3937, |
|
"step": 1534 |
|
}, |
|
{ |
|
"epoch": 104.0, |
|
"eval_accuracy": 0.7132867132867133, |
|
"eval_loss": 0.7911521196365356, |
|
"eval_runtime": 5.5547, |
|
"eval_samples_per_second": 51.488, |
|
"eval_steps_per_second": 3.24, |
|
"step": 1534 |
|
}, |
|
{ |
|
"epoch": 104.95, |
|
"grad_norm": 1.998403549194336, |
|
"learning_rate": 1.8590476190476194e-05, |
|
"loss": 0.4356, |
|
"step": 1548 |
|
}, |
|
{ |
|
"epoch": 104.95, |
|
"eval_accuracy": 0.7132867132867133, |
|
"eval_loss": 0.7922284007072449, |
|
"eval_runtime": 5.247, |
|
"eval_samples_per_second": 54.507, |
|
"eval_steps_per_second": 3.431, |
|
"step": 1548 |
|
}, |
|
{ |
|
"epoch": 105.97, |
|
"grad_norm": 1.3670754432678223, |
|
"learning_rate": 1.8447619047619045e-05, |
|
"loss": 0.3969, |
|
"step": 1563 |
|
}, |
|
{ |
|
"epoch": 105.97, |
|
"eval_accuracy": 0.7202797202797203, |
|
"eval_loss": 0.7751501202583313, |
|
"eval_runtime": 4.5814, |
|
"eval_samples_per_second": 62.427, |
|
"eval_steps_per_second": 3.929, |
|
"step": 1563 |
|
}, |
|
{ |
|
"epoch": 106.98, |
|
"grad_norm": 1.0057995319366455, |
|
"learning_rate": 1.8304761904761903e-05, |
|
"loss": 0.4051, |
|
"step": 1578 |
|
}, |
|
{ |
|
"epoch": 106.98, |
|
"eval_accuracy": 0.7132867132867133, |
|
"eval_loss": 0.7917311787605286, |
|
"eval_runtime": 5.1261, |
|
"eval_samples_per_second": 55.793, |
|
"eval_steps_per_second": 3.511, |
|
"step": 1578 |
|
}, |
|
{ |
|
"epoch": 108.0, |
|
"grad_norm": 1.7176306247711182, |
|
"learning_rate": 1.816190476190476e-05, |
|
"loss": 0.3982, |
|
"step": 1593 |
|
}, |
|
{ |
|
"epoch": 108.0, |
|
"eval_accuracy": 0.7097902097902098, |
|
"eval_loss": 0.7916986346244812, |
|
"eval_runtime": 6.2925, |
|
"eval_samples_per_second": 45.451, |
|
"eval_steps_per_second": 2.861, |
|
"step": 1593 |
|
}, |
|
{ |
|
"epoch": 108.95, |
|
"grad_norm": 1.6695342063903809, |
|
"learning_rate": 1.8028571428571427e-05, |
|
"loss": 0.4117, |
|
"step": 1607 |
|
}, |
|
{ |
|
"epoch": 108.95, |
|
"eval_accuracy": 0.7062937062937062, |
|
"eval_loss": 0.8070589303970337, |
|
"eval_runtime": 4.8482, |
|
"eval_samples_per_second": 58.99, |
|
"eval_steps_per_second": 3.713, |
|
"step": 1607 |
|
}, |
|
{ |
|
"epoch": 109.97, |
|
"grad_norm": 1.1871509552001953, |
|
"learning_rate": 1.7885714285714285e-05, |
|
"loss": 0.3666, |
|
"step": 1622 |
|
}, |
|
{ |
|
"epoch": 109.97, |
|
"eval_accuracy": 0.7202797202797203, |
|
"eval_loss": 0.7839590907096863, |
|
"eval_runtime": 5.4971, |
|
"eval_samples_per_second": 52.027, |
|
"eval_steps_per_second": 3.274, |
|
"step": 1622 |
|
}, |
|
{ |
|
"epoch": 110.98, |
|
"grad_norm": 2.196869134902954, |
|
"learning_rate": 1.7742857142857143e-05, |
|
"loss": 0.3894, |
|
"step": 1637 |
|
}, |
|
{ |
|
"epoch": 110.98, |
|
"eval_accuracy": 0.7237762237762237, |
|
"eval_loss": 0.7790002226829529, |
|
"eval_runtime": 6.5509, |
|
"eval_samples_per_second": 43.658, |
|
"eval_steps_per_second": 2.748, |
|
"step": 1637 |
|
}, |
|
{ |
|
"epoch": 112.0, |
|
"grad_norm": 2.341435432434082, |
|
"learning_rate": 1.76e-05, |
|
"loss": 0.3858, |
|
"step": 1652 |
|
}, |
|
{ |
|
"epoch": 112.0, |
|
"eval_accuracy": 0.7097902097902098, |
|
"eval_loss": 0.7961041927337646, |
|
"eval_runtime": 4.8263, |
|
"eval_samples_per_second": 59.258, |
|
"eval_steps_per_second": 3.73, |
|
"step": 1652 |
|
}, |
|
{ |
|
"epoch": 112.95, |
|
"grad_norm": 1.5795267820358276, |
|
"learning_rate": 1.7466666666666667e-05, |
|
"loss": 0.4037, |
|
"step": 1666 |
|
}, |
|
{ |
|
"epoch": 112.95, |
|
"eval_accuracy": 0.7202797202797203, |
|
"eval_loss": 0.7822238802909851, |
|
"eval_runtime": 5.6008, |
|
"eval_samples_per_second": 51.064, |
|
"eval_steps_per_second": 3.214, |
|
"step": 1666 |
|
}, |
|
{ |
|
"epoch": 113.97, |
|
"grad_norm": 1.5341583490371704, |
|
"learning_rate": 1.7333333333333332e-05, |
|
"loss": 0.3886, |
|
"step": 1681 |
|
}, |
|
{ |
|
"epoch": 113.97, |
|
"eval_accuracy": 0.7237762237762237, |
|
"eval_loss": 0.7748440504074097, |
|
"eval_runtime": 4.7211, |
|
"eval_samples_per_second": 60.579, |
|
"eval_steps_per_second": 3.813, |
|
"step": 1681 |
|
}, |
|
{ |
|
"epoch": 114.98, |
|
"grad_norm": 1.056776523590088, |
|
"learning_rate": 1.719047619047619e-05, |
|
"loss": 0.3762, |
|
"step": 1696 |
|
}, |
|
{ |
|
"epoch": 114.98, |
|
"eval_accuracy": 0.7167832167832168, |
|
"eval_loss": 0.7781729102134705, |
|
"eval_runtime": 6.5039, |
|
"eval_samples_per_second": 43.974, |
|
"eval_steps_per_second": 2.768, |
|
"step": 1696 |
|
}, |
|
{ |
|
"epoch": 116.0, |
|
"grad_norm": 2.7518413066864014, |
|
"learning_rate": 1.704761904761905e-05, |
|
"loss": 0.3444, |
|
"step": 1711 |
|
}, |
|
{ |
|
"epoch": 116.0, |
|
"eval_accuracy": 0.7307692307692307, |
|
"eval_loss": 0.7745847702026367, |
|
"eval_runtime": 6.1348, |
|
"eval_samples_per_second": 46.619, |
|
"eval_steps_per_second": 2.934, |
|
"step": 1711 |
|
}, |
|
{ |
|
"epoch": 116.95, |
|
"grad_norm": 1.8518308401107788, |
|
"learning_rate": 1.6914285714285714e-05, |
|
"loss": 0.3961, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 116.95, |
|
"eval_accuracy": 0.7202797202797203, |
|
"eval_loss": 0.7842003703117371, |
|
"eval_runtime": 4.8863, |
|
"eval_samples_per_second": 58.531, |
|
"eval_steps_per_second": 3.684, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 117.97, |
|
"grad_norm": 1.487108588218689, |
|
"learning_rate": 1.6771428571428572e-05, |
|
"loss": 0.3578, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 117.97, |
|
"eval_accuracy": 0.7202797202797203, |
|
"eval_loss": 0.7819164991378784, |
|
"eval_runtime": 5.6099, |
|
"eval_samples_per_second": 50.981, |
|
"eval_steps_per_second": 3.209, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 118.98, |
|
"grad_norm": 1.7816615104675293, |
|
"learning_rate": 1.662857142857143e-05, |
|
"loss": 0.3578, |
|
"step": 1755 |
|
}, |
|
{ |
|
"epoch": 118.98, |
|
"eval_accuracy": 0.7202797202797203, |
|
"eval_loss": 0.7805832624435425, |
|
"eval_runtime": 5.5595, |
|
"eval_samples_per_second": 51.443, |
|
"eval_steps_per_second": 3.238, |
|
"step": 1755 |
|
}, |
|
{ |
|
"epoch": 120.0, |
|
"grad_norm": 2.4715042114257812, |
|
"learning_rate": 1.6485714285714285e-05, |
|
"loss": 0.3489, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 120.0, |
|
"eval_accuracy": 0.7237762237762237, |
|
"eval_loss": 0.7809211015701294, |
|
"eval_runtime": 6.1205, |
|
"eval_samples_per_second": 46.728, |
|
"eval_steps_per_second": 2.941, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 120.95, |
|
"grad_norm": 1.1066619157791138, |
|
"learning_rate": 1.635238095238095e-05, |
|
"loss": 0.3622, |
|
"step": 1784 |
|
}, |
|
{ |
|
"epoch": 120.95, |
|
"eval_accuracy": 0.7097902097902098, |
|
"eval_loss": 0.7947035431861877, |
|
"eval_runtime": 5.0722, |
|
"eval_samples_per_second": 56.386, |
|
"eval_steps_per_second": 3.549, |
|
"step": 1784 |
|
}, |
|
{ |
|
"epoch": 121.97, |
|
"grad_norm": 1.3460161685943604, |
|
"learning_rate": 1.620952380952381e-05, |
|
"loss": 0.3545, |
|
"step": 1799 |
|
}, |
|
{ |
|
"epoch": 121.97, |
|
"eval_accuracy": 0.7167832167832168, |
|
"eval_loss": 0.7877638339996338, |
|
"eval_runtime": 5.4395, |
|
"eval_samples_per_second": 52.578, |
|
"eval_steps_per_second": 3.309, |
|
"step": 1799 |
|
}, |
|
{ |
|
"epoch": 122.98, |
|
"grad_norm": 1.7303364276885986, |
|
"learning_rate": 1.6066666666666666e-05, |
|
"loss": 0.3361, |
|
"step": 1814 |
|
}, |
|
{ |
|
"epoch": 122.98, |
|
"eval_accuracy": 0.7167832167832168, |
|
"eval_loss": 0.7854802012443542, |
|
"eval_runtime": 4.9544, |
|
"eval_samples_per_second": 57.726, |
|
"eval_steps_per_second": 3.633, |
|
"step": 1814 |
|
}, |
|
{ |
|
"epoch": 124.0, |
|
"grad_norm": 1.5440623760223389, |
|
"learning_rate": 1.5923809523809524e-05, |
|
"loss": 0.3618, |
|
"step": 1829 |
|
}, |
|
{ |
|
"epoch": 124.0, |
|
"eval_accuracy": 0.7132867132867133, |
|
"eval_loss": 0.7890083193778992, |
|
"eval_runtime": 4.6202, |
|
"eval_samples_per_second": 61.902, |
|
"eval_steps_per_second": 3.896, |
|
"step": 1829 |
|
}, |
|
{ |
|
"epoch": 124.95, |
|
"grad_norm": 1.2985795736312866, |
|
"learning_rate": 1.579047619047619e-05, |
|
"loss": 0.3472, |
|
"step": 1843 |
|
}, |
|
{ |
|
"epoch": 124.95, |
|
"eval_accuracy": 0.7167832167832168, |
|
"eval_loss": 0.7809839844703674, |
|
"eval_runtime": 4.5623, |
|
"eval_samples_per_second": 62.687, |
|
"eval_steps_per_second": 3.945, |
|
"step": 1843 |
|
}, |
|
{ |
|
"epoch": 125.97, |
|
"grad_norm": 1.3930552005767822, |
|
"learning_rate": 1.5647619047619048e-05, |
|
"loss": 0.3511, |
|
"step": 1858 |
|
}, |
|
{ |
|
"epoch": 125.97, |
|
"eval_accuracy": 0.7132867132867133, |
|
"eval_loss": 0.7897189259529114, |
|
"eval_runtime": 5.1961, |
|
"eval_samples_per_second": 55.041, |
|
"eval_steps_per_second": 3.464, |
|
"step": 1858 |
|
}, |
|
{ |
|
"epoch": 126.98, |
|
"grad_norm": 2.0258781909942627, |
|
"learning_rate": 1.5504761904761906e-05, |
|
"loss": 0.3389, |
|
"step": 1873 |
|
}, |
|
{ |
|
"epoch": 126.98, |
|
"eval_accuracy": 0.7132867132867133, |
|
"eval_loss": 0.7922906875610352, |
|
"eval_runtime": 5.0606, |
|
"eval_samples_per_second": 56.515, |
|
"eval_steps_per_second": 3.557, |
|
"step": 1873 |
|
}, |
|
{ |
|
"epoch": 128.0, |
|
"grad_norm": 1.0623548030853271, |
|
"learning_rate": 1.5361904761904764e-05, |
|
"loss": 0.3391, |
|
"step": 1888 |
|
}, |
|
{ |
|
"epoch": 128.0, |
|
"eval_accuracy": 0.7272727272727273, |
|
"eval_loss": 0.7781790494918823, |
|
"eval_runtime": 4.5848, |
|
"eval_samples_per_second": 62.38, |
|
"eval_steps_per_second": 3.926, |
|
"step": 1888 |
|
}, |
|
{ |
|
"epoch": 128.95, |
|
"grad_norm": 1.2839761972427368, |
|
"learning_rate": 1.522857142857143e-05, |
|
"loss": 0.3746, |
|
"step": 1902 |
|
}, |
|
{ |
|
"epoch": 128.95, |
|
"eval_accuracy": 0.7202797202797203, |
|
"eval_loss": 0.7838397026062012, |
|
"eval_runtime": 6.371, |
|
"eval_samples_per_second": 44.891, |
|
"eval_steps_per_second": 2.825, |
|
"step": 1902 |
|
}, |
|
{ |
|
"epoch": 129.97, |
|
"grad_norm": 1.7424538135528564, |
|
"learning_rate": 1.5085714285714288e-05, |
|
"loss": 0.3238, |
|
"step": 1917 |
|
}, |
|
{ |
|
"epoch": 129.97, |
|
"eval_accuracy": 0.7167832167832168, |
|
"eval_loss": 0.7943305969238281, |
|
"eval_runtime": 5.0078, |
|
"eval_samples_per_second": 57.111, |
|
"eval_steps_per_second": 3.594, |
|
"step": 1917 |
|
}, |
|
{ |
|
"epoch": 130.98, |
|
"grad_norm": 1.9641584157943726, |
|
"learning_rate": 1.4942857142857143e-05, |
|
"loss": 0.3601, |
|
"step": 1932 |
|
}, |
|
{ |
|
"epoch": 130.98, |
|
"eval_accuracy": 0.7167832167832168, |
|
"eval_loss": 0.786338746547699, |
|
"eval_runtime": 5.7221, |
|
"eval_samples_per_second": 49.982, |
|
"eval_steps_per_second": 3.146, |
|
"step": 1932 |
|
}, |
|
{ |
|
"epoch": 132.0, |
|
"grad_norm": 1.6700012683868408, |
|
"learning_rate": 1.48e-05, |
|
"loss": 0.3339, |
|
"step": 1947 |
|
}, |
|
{ |
|
"epoch": 132.0, |
|
"eval_accuracy": 0.7132867132867133, |
|
"eval_loss": 0.7948570847511292, |
|
"eval_runtime": 4.5956, |
|
"eval_samples_per_second": 62.234, |
|
"eval_steps_per_second": 3.917, |
|
"step": 1947 |
|
}, |
|
{ |
|
"epoch": 132.95, |
|
"grad_norm": 1.4314067363739014, |
|
"learning_rate": 1.4666666666666666e-05, |
|
"loss": 0.3805, |
|
"step": 1961 |
|
}, |
|
{ |
|
"epoch": 132.95, |
|
"eval_accuracy": 0.7237762237762237, |
|
"eval_loss": 0.7823219299316406, |
|
"eval_runtime": 5.3022, |
|
"eval_samples_per_second": 53.94, |
|
"eval_steps_per_second": 3.395, |
|
"step": 1961 |
|
}, |
|
{ |
|
"epoch": 133.97, |
|
"grad_norm": 1.9961072206497192, |
|
"learning_rate": 1.4523809523809524e-05, |
|
"loss": 0.3524, |
|
"step": 1976 |
|
}, |
|
{ |
|
"epoch": 133.97, |
|
"eval_accuracy": 0.7097902097902098, |
|
"eval_loss": 0.8052372932434082, |
|
"eval_runtime": 4.8993, |
|
"eval_samples_per_second": 58.375, |
|
"eval_steps_per_second": 3.674, |
|
"step": 1976 |
|
}, |
|
{ |
|
"epoch": 134.98, |
|
"grad_norm": 1.1320672035217285, |
|
"learning_rate": 1.4380952380952382e-05, |
|
"loss": 0.3103, |
|
"step": 1991 |
|
}, |
|
{ |
|
"epoch": 134.98, |
|
"eval_accuracy": 0.7237762237762237, |
|
"eval_loss": 0.7809023261070251, |
|
"eval_runtime": 5.8892, |
|
"eval_samples_per_second": 48.563, |
|
"eval_steps_per_second": 3.056, |
|
"step": 1991 |
|
}, |
|
{ |
|
"epoch": 136.0, |
|
"grad_norm": 2.298990249633789, |
|
"learning_rate": 1.4238095238095239e-05, |
|
"loss": 0.3484, |
|
"step": 2006 |
|
}, |
|
{ |
|
"epoch": 136.0, |
|
"eval_accuracy": 0.7202797202797203, |
|
"eval_loss": 0.787878692150116, |
|
"eval_runtime": 5.2761, |
|
"eval_samples_per_second": 54.207, |
|
"eval_steps_per_second": 3.412, |
|
"step": 2006 |
|
}, |
|
{ |
|
"epoch": 136.95, |
|
"grad_norm": 2.3017499446868896, |
|
"learning_rate": 1.4104761904761906e-05, |
|
"loss": 0.3424, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 136.95, |
|
"eval_accuracy": 0.7272727272727273, |
|
"eval_loss": 0.7875267267227173, |
|
"eval_runtime": 4.7804, |
|
"eval_samples_per_second": 59.828, |
|
"eval_steps_per_second": 3.765, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 137.97, |
|
"grad_norm": 1.4992213249206543, |
|
"learning_rate": 1.3961904761904762e-05, |
|
"loss": 0.316, |
|
"step": 2035 |
|
}, |
|
{ |
|
"epoch": 137.97, |
|
"eval_accuracy": 0.7272727272727273, |
|
"eval_loss": 0.7829337120056152, |
|
"eval_runtime": 4.9522, |
|
"eval_samples_per_second": 57.752, |
|
"eval_steps_per_second": 3.635, |
|
"step": 2035 |
|
}, |
|
{ |
|
"epoch": 138.98, |
|
"grad_norm": 1.634419560432434, |
|
"learning_rate": 1.3819047619047619e-05, |
|
"loss": 0.3171, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 138.98, |
|
"eval_accuracy": 0.7202797202797203, |
|
"eval_loss": 0.7882408499717712, |
|
"eval_runtime": 4.8523, |
|
"eval_samples_per_second": 58.941, |
|
"eval_steps_per_second": 3.71, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 140.0, |
|
"grad_norm": 1.2980549335479736, |
|
"learning_rate": 1.3676190476190477e-05, |
|
"loss": 0.3155, |
|
"step": 2065 |
|
}, |
|
{ |
|
"epoch": 140.0, |
|
"eval_accuracy": 0.7167832167832168, |
|
"eval_loss": 0.7830457091331482, |
|
"eval_runtime": 4.9611, |
|
"eval_samples_per_second": 57.649, |
|
"eval_steps_per_second": 3.628, |
|
"step": 2065 |
|
}, |
|
{ |
|
"epoch": 140.95, |
|
"grad_norm": 1.5925828218460083, |
|
"learning_rate": 1.3542857142857142e-05, |
|
"loss": 0.3382, |
|
"step": 2079 |
|
}, |
|
{ |
|
"epoch": 140.95, |
|
"eval_accuracy": 0.7272727272727273, |
|
"eval_loss": 0.7825762033462524, |
|
"eval_runtime": 4.6558, |
|
"eval_samples_per_second": 61.428, |
|
"eval_steps_per_second": 3.866, |
|
"step": 2079 |
|
}, |
|
{ |
|
"epoch": 141.97, |
|
"grad_norm": 1.9007666110992432, |
|
"learning_rate": 1.34e-05, |
|
"loss": 0.3175, |
|
"step": 2094 |
|
}, |
|
{ |
|
"epoch": 141.97, |
|
"eval_accuracy": 0.7202797202797203, |
|
"eval_loss": 0.7963916659355164, |
|
"eval_runtime": 5.2803, |
|
"eval_samples_per_second": 54.164, |
|
"eval_steps_per_second": 3.409, |
|
"step": 2094 |
|
}, |
|
{ |
|
"epoch": 142.98, |
|
"grad_norm": 1.8862413167953491, |
|
"learning_rate": 1.3257142857142858e-05, |
|
"loss": 0.3444, |
|
"step": 2109 |
|
}, |
|
{ |
|
"epoch": 142.98, |
|
"eval_accuracy": 0.7237762237762237, |
|
"eval_loss": 0.7858501672744751, |
|
"eval_runtime": 5.4501, |
|
"eval_samples_per_second": 52.476, |
|
"eval_steps_per_second": 3.303, |
|
"step": 2109 |
|
}, |
|
{ |
|
"epoch": 144.0, |
|
"grad_norm": 2.4451475143432617, |
|
"learning_rate": 1.3114285714285715e-05, |
|
"loss": 0.3208, |
|
"step": 2124 |
|
}, |
|
{ |
|
"epoch": 144.0, |
|
"eval_accuracy": 0.7272727272727273, |
|
"eval_loss": 0.7859659194946289, |
|
"eval_runtime": 4.5157, |
|
"eval_samples_per_second": 63.334, |
|
"eval_steps_per_second": 3.986, |
|
"step": 2124 |
|
}, |
|
{ |
|
"epoch": 144.95, |
|
"grad_norm": 1.6911462545394897, |
|
"learning_rate": 1.298095238095238e-05, |
|
"loss": 0.3286, |
|
"step": 2138 |
|
}, |
|
{ |
|
"epoch": 144.95, |
|
"eval_accuracy": 0.7272727272727273, |
|
"eval_loss": 0.7869133353233337, |
|
"eval_runtime": 5.6207, |
|
"eval_samples_per_second": 50.883, |
|
"eval_steps_per_second": 3.202, |
|
"step": 2138 |
|
}, |
|
{ |
|
"epoch": 145.97, |
|
"grad_norm": 1.9792907238006592, |
|
"learning_rate": 1.2838095238095239e-05, |
|
"loss": 0.3319, |
|
"step": 2153 |
|
}, |
|
{ |
|
"epoch": 145.97, |
|
"eval_accuracy": 0.7167832167832168, |
|
"eval_loss": 0.7916193604469299, |
|
"eval_runtime": 5.434, |
|
"eval_samples_per_second": 52.631, |
|
"eval_steps_per_second": 3.312, |
|
"step": 2153 |
|
}, |
|
{ |
|
"epoch": 146.98, |
|
"grad_norm": 1.3403879404067993, |
|
"learning_rate": 1.2695238095238095e-05, |
|
"loss": 0.2954, |
|
"step": 2168 |
|
}, |
|
{ |
|
"epoch": 146.98, |
|
"eval_accuracy": 0.7237762237762237, |
|
"eval_loss": 0.7937904000282288, |
|
"eval_runtime": 4.5934, |
|
"eval_samples_per_second": 62.264, |
|
"eval_steps_per_second": 3.919, |
|
"step": 2168 |
|
}, |
|
{ |
|
"epoch": 148.0, |
|
"grad_norm": 2.6414599418640137, |
|
"learning_rate": 1.2552380952380953e-05, |
|
"loss": 0.3283, |
|
"step": 2183 |
|
}, |
|
{ |
|
"epoch": 148.0, |
|
"eval_accuracy": 0.7167832167832168, |
|
"eval_loss": 0.7974053025245667, |
|
"eval_runtime": 4.6622, |
|
"eval_samples_per_second": 61.344, |
|
"eval_steps_per_second": 3.861, |
|
"step": 2183 |
|
}, |
|
{ |
|
"epoch": 148.95, |
|
"grad_norm": 1.7404941320419312, |
|
"learning_rate": 1.241904761904762e-05, |
|
"loss": 0.3306, |
|
"step": 2197 |
|
}, |
|
{ |
|
"epoch": 148.95, |
|
"eval_accuracy": 0.7307692307692307, |
|
"eval_loss": 0.7795438766479492, |
|
"eval_runtime": 4.8202, |
|
"eval_samples_per_second": 59.334, |
|
"eval_steps_per_second": 3.734, |
|
"step": 2197 |
|
}, |
|
{ |
|
"epoch": 149.97, |
|
"grad_norm": 1.3558377027511597, |
|
"learning_rate": 1.2276190476190477e-05, |
|
"loss": 0.3073, |
|
"step": 2212 |
|
}, |
|
{ |
|
"epoch": 149.97, |
|
"eval_accuracy": 0.7202797202797203, |
|
"eval_loss": 0.7910019755363464, |
|
"eval_runtime": 5.1476, |
|
"eval_samples_per_second": 55.56, |
|
"eval_steps_per_second": 3.497, |
|
"step": 2212 |
|
}, |
|
{ |
|
"epoch": 150.98, |
|
"grad_norm": 1.2379992008209229, |
|
"learning_rate": 1.2133333333333333e-05, |
|
"loss": 0.3089, |
|
"step": 2227 |
|
}, |
|
{ |
|
"epoch": 150.98, |
|
"eval_accuracy": 0.7202797202797203, |
|
"eval_loss": 0.7942177653312683, |
|
"eval_runtime": 5.4133, |
|
"eval_samples_per_second": 52.833, |
|
"eval_steps_per_second": 3.325, |
|
"step": 2227 |
|
}, |
|
{ |
|
"epoch": 152.0, |
|
"grad_norm": 1.3452534675598145, |
|
"learning_rate": 1.1990476190476191e-05, |
|
"loss": 0.2915, |
|
"step": 2242 |
|
}, |
|
{ |
|
"epoch": 152.0, |
|
"eval_accuracy": 0.7167832167832168, |
|
"eval_loss": 0.7933531403541565, |
|
"eval_runtime": 6.3553, |
|
"eval_samples_per_second": 45.002, |
|
"eval_steps_per_second": 2.832, |
|
"step": 2242 |
|
}, |
|
{ |
|
"epoch": 152.95, |
|
"grad_norm": 0.9324106574058533, |
|
"learning_rate": 1.1857142857142857e-05, |
|
"loss": 0.3286, |
|
"step": 2256 |
|
}, |
|
{ |
|
"epoch": 152.95, |
|
"eval_accuracy": 0.7307692307692307, |
|
"eval_loss": 0.7807846069335938, |
|
"eval_runtime": 5.094, |
|
"eval_samples_per_second": 56.144, |
|
"eval_steps_per_second": 3.534, |
|
"step": 2256 |
|
}, |
|
{ |
|
"epoch": 153.97, |
|
"grad_norm": 1.5133529901504517, |
|
"learning_rate": 1.1714285714285715e-05, |
|
"loss": 0.2817, |
|
"step": 2271 |
|
}, |
|
{ |
|
"epoch": 153.97, |
|
"eval_accuracy": 0.7307692307692307, |
|
"eval_loss": 0.7788205742835999, |
|
"eval_runtime": 5.0531, |
|
"eval_samples_per_second": 56.599, |
|
"eval_steps_per_second": 3.562, |
|
"step": 2271 |
|
}, |
|
{ |
|
"epoch": 154.98, |
|
"grad_norm": 2.8459701538085938, |
|
"learning_rate": 1.1571428571428573e-05, |
|
"loss": 0.3118, |
|
"step": 2286 |
|
}, |
|
{ |
|
"epoch": 154.98, |
|
"eval_accuracy": 0.7272727272727273, |
|
"eval_loss": 0.7898407578468323, |
|
"eval_runtime": 4.8393, |
|
"eval_samples_per_second": 59.1, |
|
"eval_steps_per_second": 3.72, |
|
"step": 2286 |
|
}, |
|
{ |
|
"epoch": 156.0, |
|
"grad_norm": 1.8526194095611572, |
|
"learning_rate": 1.1428571428571429e-05, |
|
"loss": 0.3155, |
|
"step": 2301 |
|
}, |
|
{ |
|
"epoch": 156.0, |
|
"eval_accuracy": 0.7202797202797203, |
|
"eval_loss": 0.796588122844696, |
|
"eval_runtime": 5.8687, |
|
"eval_samples_per_second": 48.733, |
|
"eval_steps_per_second": 3.067, |
|
"step": 2301 |
|
}, |
|
{ |
|
"epoch": 156.95, |
|
"grad_norm": 1.130321741104126, |
|
"learning_rate": 1.1295238095238096e-05, |
|
"loss": 0.3156, |
|
"step": 2315 |
|
}, |
|
{ |
|
"epoch": 156.95, |
|
"eval_accuracy": 0.7202797202797203, |
|
"eval_loss": 0.7947112321853638, |
|
"eval_runtime": 4.5714, |
|
"eval_samples_per_second": 62.564, |
|
"eval_steps_per_second": 3.938, |
|
"step": 2315 |
|
}, |
|
{ |
|
"epoch": 157.97, |
|
"grad_norm": 1.3919442892074585, |
|
"learning_rate": 1.1152380952380953e-05, |
|
"loss": 0.2936, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 157.97, |
|
"eval_accuracy": 0.7167832167832168, |
|
"eval_loss": 0.7916660904884338, |
|
"eval_runtime": 4.554, |
|
"eval_samples_per_second": 62.802, |
|
"eval_steps_per_second": 3.953, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 158.98, |
|
"grad_norm": 2.1628525257110596, |
|
"learning_rate": 1.1009523809523809e-05, |
|
"loss": 0.3049, |
|
"step": 2345 |
|
}, |
|
{ |
|
"epoch": 158.98, |
|
"eval_accuracy": 0.7307692307692307, |
|
"eval_loss": 0.7779849171638489, |
|
"eval_runtime": 5.5631, |
|
"eval_samples_per_second": 51.41, |
|
"eval_steps_per_second": 3.236, |
|
"step": 2345 |
|
}, |
|
{ |
|
"epoch": 160.0, |
|
"grad_norm": 1.0748810768127441, |
|
"learning_rate": 1.0866666666666667e-05, |
|
"loss": 0.2896, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 160.0, |
|
"eval_accuracy": 0.7272727272727273, |
|
"eval_loss": 0.792616605758667, |
|
"eval_runtime": 5.3985, |
|
"eval_samples_per_second": 52.978, |
|
"eval_steps_per_second": 3.334, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 160.95, |
|
"grad_norm": 1.9327268600463867, |
|
"learning_rate": 1.0733333333333333e-05, |
|
"loss": 0.3194, |
|
"step": 2374 |
|
}, |
|
{ |
|
"epoch": 160.95, |
|
"eval_accuracy": 0.7202797202797203, |
|
"eval_loss": 0.8022683262825012, |
|
"eval_runtime": 5.5802, |
|
"eval_samples_per_second": 51.253, |
|
"eval_steps_per_second": 3.226, |
|
"step": 2374 |
|
}, |
|
{ |
|
"epoch": 161.97, |
|
"grad_norm": 1.610187292098999, |
|
"learning_rate": 1.059047619047619e-05, |
|
"loss": 0.2918, |
|
"step": 2389 |
|
}, |
|
{ |
|
"epoch": 161.97, |
|
"eval_accuracy": 0.7272727272727273, |
|
"eval_loss": 0.7933264374732971, |
|
"eval_runtime": 5.0333, |
|
"eval_samples_per_second": 56.821, |
|
"eval_steps_per_second": 3.576, |
|
"step": 2389 |
|
}, |
|
{ |
|
"epoch": 162.98, |
|
"grad_norm": 1.961053729057312, |
|
"learning_rate": 1.0447619047619049e-05, |
|
"loss": 0.2992, |
|
"step": 2404 |
|
}, |
|
{ |
|
"epoch": 162.98, |
|
"eval_accuracy": 0.7412587412587412, |
|
"eval_loss": 0.7828559875488281, |
|
"eval_runtime": 5.4812, |
|
"eval_samples_per_second": 52.178, |
|
"eval_steps_per_second": 3.284, |
|
"step": 2404 |
|
}, |
|
{ |
|
"epoch": 164.0, |
|
"grad_norm": 1.2120810747146606, |
|
"learning_rate": 1.0304761904761905e-05, |
|
"loss": 0.3, |
|
"step": 2419 |
|
}, |
|
{ |
|
"epoch": 164.0, |
|
"eval_accuracy": 0.7202797202797203, |
|
"eval_loss": 0.7946493029594421, |
|
"eval_runtime": 5.351, |
|
"eval_samples_per_second": 53.448, |
|
"eval_steps_per_second": 3.364, |
|
"step": 2419 |
|
}, |
|
{ |
|
"epoch": 164.95, |
|
"grad_norm": 1.112051010131836, |
|
"learning_rate": 1.0171428571428571e-05, |
|
"loss": 0.322, |
|
"step": 2433 |
|
}, |
|
{ |
|
"epoch": 164.95, |
|
"eval_accuracy": 0.7272727272727273, |
|
"eval_loss": 0.79693204164505, |
|
"eval_runtime": 6.2682, |
|
"eval_samples_per_second": 45.627, |
|
"eval_steps_per_second": 2.872, |
|
"step": 2433 |
|
}, |
|
{ |
|
"epoch": 165.97, |
|
"grad_norm": 1.2547581195831299, |
|
"learning_rate": 1.0028571428571429e-05, |
|
"loss": 0.2994, |
|
"step": 2448 |
|
}, |
|
{ |
|
"epoch": 165.97, |
|
"eval_accuracy": 0.7237762237762237, |
|
"eval_loss": 0.8075766563415527, |
|
"eval_runtime": 5.4676, |
|
"eval_samples_per_second": 52.308, |
|
"eval_steps_per_second": 3.292, |
|
"step": 2448 |
|
}, |
|
{ |
|
"epoch": 166.98, |
|
"grad_norm": 3.3027942180633545, |
|
"learning_rate": 9.885714285714285e-06, |
|
"loss": 0.2849, |
|
"step": 2463 |
|
}, |
|
{ |
|
"epoch": 166.98, |
|
"eval_accuracy": 0.7307692307692307, |
|
"eval_loss": 0.7950677275657654, |
|
"eval_runtime": 5.6445, |
|
"eval_samples_per_second": 50.668, |
|
"eval_steps_per_second": 3.189, |
|
"step": 2463 |
|
}, |
|
{ |
|
"epoch": 168.0, |
|
"grad_norm": 3.2351794242858887, |
|
"learning_rate": 9.742857142857143e-06, |
|
"loss": 0.2745, |
|
"step": 2478 |
|
}, |
|
{ |
|
"epoch": 168.0, |
|
"eval_accuracy": 0.7342657342657343, |
|
"eval_loss": 0.7892395853996277, |
|
"eval_runtime": 6.6609, |
|
"eval_samples_per_second": 42.937, |
|
"eval_steps_per_second": 2.702, |
|
"step": 2478 |
|
}, |
|
{ |
|
"epoch": 168.95, |
|
"grad_norm": 2.2650949954986572, |
|
"learning_rate": 9.60952380952381e-06, |
|
"loss": 0.2974, |
|
"step": 2492 |
|
}, |
|
{ |
|
"epoch": 168.95, |
|
"eval_accuracy": 0.7307692307692307, |
|
"eval_loss": 0.791623055934906, |
|
"eval_runtime": 5.8468, |
|
"eval_samples_per_second": 48.916, |
|
"eval_steps_per_second": 3.079, |
|
"step": 2492 |
|
}, |
|
{ |
|
"epoch": 169.97, |
|
"grad_norm": 1.3676173686981201, |
|
"learning_rate": 9.466666666666667e-06, |
|
"loss": 0.2656, |
|
"step": 2507 |
|
}, |
|
{ |
|
"epoch": 169.97, |
|
"eval_accuracy": 0.7342657342657343, |
|
"eval_loss": 0.7994617223739624, |
|
"eval_runtime": 4.7953, |
|
"eval_samples_per_second": 59.642, |
|
"eval_steps_per_second": 3.754, |
|
"step": 2507 |
|
}, |
|
{ |
|
"epoch": 170.98, |
|
"grad_norm": 1.5123904943466187, |
|
"learning_rate": 9.323809523809523e-06, |
|
"loss": 0.295, |
|
"step": 2522 |
|
}, |
|
{ |
|
"epoch": 170.98, |
|
"eval_accuracy": 0.7237762237762237, |
|
"eval_loss": 0.802618682384491, |
|
"eval_runtime": 6.3216, |
|
"eval_samples_per_second": 45.241, |
|
"eval_steps_per_second": 2.847, |
|
"step": 2522 |
|
}, |
|
{ |
|
"epoch": 172.0, |
|
"grad_norm": 1.5210902690887451, |
|
"learning_rate": 9.180952380952381e-06, |
|
"loss": 0.2791, |
|
"step": 2537 |
|
}, |
|
{ |
|
"epoch": 172.0, |
|
"eval_accuracy": 0.7342657342657343, |
|
"eval_loss": 0.797250509262085, |
|
"eval_runtime": 5.4447, |
|
"eval_samples_per_second": 52.528, |
|
"eval_steps_per_second": 3.306, |
|
"step": 2537 |
|
}, |
|
{ |
|
"epoch": 172.95, |
|
"grad_norm": 1.0823432207107544, |
|
"learning_rate": 9.047619047619047e-06, |
|
"loss": 0.2836, |
|
"step": 2551 |
|
}, |
|
{ |
|
"epoch": 172.95, |
|
"eval_accuracy": 0.7307692307692307, |
|
"eval_loss": 0.8022569417953491, |
|
"eval_runtime": 5.0556, |
|
"eval_samples_per_second": 56.571, |
|
"eval_steps_per_second": 3.56, |
|
"step": 2551 |
|
}, |
|
{ |
|
"epoch": 173.97, |
|
"grad_norm": 1.4100669622421265, |
|
"learning_rate": 8.904761904761905e-06, |
|
"loss": 0.2806, |
|
"step": 2566 |
|
}, |
|
{ |
|
"epoch": 173.97, |
|
"eval_accuracy": 0.7307692307692307, |
|
"eval_loss": 0.8013490438461304, |
|
"eval_runtime": 5.503, |
|
"eval_samples_per_second": 51.972, |
|
"eval_steps_per_second": 3.271, |
|
"step": 2566 |
|
}, |
|
{ |
|
"epoch": 174.98, |
|
"grad_norm": 1.389672875404358, |
|
"learning_rate": 8.761904761904763e-06, |
|
"loss": 0.2661, |
|
"step": 2581 |
|
}, |
|
{ |
|
"epoch": 174.98, |
|
"eval_accuracy": 0.7307692307692307, |
|
"eval_loss": 0.796493649482727, |
|
"eval_runtime": 5.7105, |
|
"eval_samples_per_second": 50.083, |
|
"eval_steps_per_second": 3.152, |
|
"step": 2581 |
|
}, |
|
{ |
|
"epoch": 176.0, |
|
"grad_norm": 1.8702772855758667, |
|
"learning_rate": 8.61904761904762e-06, |
|
"loss": 0.2695, |
|
"step": 2596 |
|
}, |
|
{ |
|
"epoch": 176.0, |
|
"eval_accuracy": 0.7272727272727273, |
|
"eval_loss": 0.8062567114830017, |
|
"eval_runtime": 7.451, |
|
"eval_samples_per_second": 38.384, |
|
"eval_steps_per_second": 2.416, |
|
"step": 2596 |
|
}, |
|
{ |
|
"epoch": 176.95, |
|
"grad_norm": 3.131314992904663, |
|
"learning_rate": 8.485714285714287e-06, |
|
"loss": 0.286, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 176.95, |
|
"eval_accuracy": 0.7237762237762237, |
|
"eval_loss": 0.7963144779205322, |
|
"eval_runtime": 4.431, |
|
"eval_samples_per_second": 64.546, |
|
"eval_steps_per_second": 4.062, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 177.97, |
|
"grad_norm": 1.7147862911224365, |
|
"learning_rate": 8.342857142857143e-06, |
|
"loss": 0.2743, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 177.97, |
|
"eval_accuracy": 0.7412587412587412, |
|
"eval_loss": 0.7928534150123596, |
|
"eval_runtime": 5.7357, |
|
"eval_samples_per_second": 49.863, |
|
"eval_steps_per_second": 3.138, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 178.98, |
|
"grad_norm": 1.669243574142456, |
|
"learning_rate": 8.2e-06, |
|
"loss": 0.2775, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 178.98, |
|
"eval_accuracy": 0.7447552447552448, |
|
"eval_loss": 0.7855107188224792, |
|
"eval_runtime": 5.7848, |
|
"eval_samples_per_second": 49.44, |
|
"eval_steps_per_second": 3.112, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 180.0, |
|
"grad_norm": 2.0756781101226807, |
|
"learning_rate": 8.057142857142857e-06, |
|
"loss": 0.2878, |
|
"step": 2655 |
|
}, |
|
{ |
|
"epoch": 180.0, |
|
"eval_accuracy": 0.7377622377622378, |
|
"eval_loss": 0.7894182801246643, |
|
"eval_runtime": 6.034, |
|
"eval_samples_per_second": 47.398, |
|
"eval_steps_per_second": 2.983, |
|
"step": 2655 |
|
}, |
|
{ |
|
"epoch": 180.95, |
|
"grad_norm": 1.7884511947631836, |
|
"learning_rate": 7.923809523809525e-06, |
|
"loss": 0.2757, |
|
"step": 2669 |
|
}, |
|
{ |
|
"epoch": 180.95, |
|
"eval_accuracy": 0.7272727272727273, |
|
"eval_loss": 0.8012902736663818, |
|
"eval_runtime": 5.2896, |
|
"eval_samples_per_second": 54.068, |
|
"eval_steps_per_second": 3.403, |
|
"step": 2669 |
|
}, |
|
{ |
|
"epoch": 181.97, |
|
"grad_norm": 1.1550542116165161, |
|
"learning_rate": 7.780952380952381e-06, |
|
"loss": 0.3067, |
|
"step": 2684 |
|
}, |
|
{ |
|
"epoch": 181.97, |
|
"eval_accuracy": 0.7307692307692307, |
|
"eval_loss": 0.8015040755271912, |
|
"eval_runtime": 5.5658, |
|
"eval_samples_per_second": 51.385, |
|
"eval_steps_per_second": 3.234, |
|
"step": 2684 |
|
}, |
|
{ |
|
"epoch": 182.98, |
|
"grad_norm": 1.497018814086914, |
|
"learning_rate": 7.63809523809524e-06, |
|
"loss": 0.2412, |
|
"step": 2699 |
|
}, |
|
{ |
|
"epoch": 182.98, |
|
"eval_accuracy": 0.7272727272727273, |
|
"eval_loss": 0.7975053191184998, |
|
"eval_runtime": 4.7894, |
|
"eval_samples_per_second": 59.715, |
|
"eval_steps_per_second": 3.758, |
|
"step": 2699 |
|
}, |
|
{ |
|
"epoch": 184.0, |
|
"grad_norm": 1.293284296989441, |
|
"learning_rate": 7.4952380952380955e-06, |
|
"loss": 0.2686, |
|
"step": 2714 |
|
}, |
|
{ |
|
"epoch": 184.0, |
|
"eval_accuracy": 0.7237762237762237, |
|
"eval_loss": 0.8036520481109619, |
|
"eval_runtime": 5.5902, |
|
"eval_samples_per_second": 51.161, |
|
"eval_steps_per_second": 3.22, |
|
"step": 2714 |
|
}, |
|
{ |
|
"epoch": 184.95, |
|
"grad_norm": 1.8103567361831665, |
|
"learning_rate": 7.361904761904762e-06, |
|
"loss": 0.3176, |
|
"step": 2728 |
|
}, |
|
{ |
|
"epoch": 184.95, |
|
"eval_accuracy": 0.7307692307692307, |
|
"eval_loss": 0.8016763925552368, |
|
"eval_runtime": 4.3848, |
|
"eval_samples_per_second": 65.225, |
|
"eval_steps_per_second": 4.105, |
|
"step": 2728 |
|
}, |
|
{ |
|
"epoch": 185.97, |
|
"grad_norm": 1.6021767854690552, |
|
"learning_rate": 7.219047619047619e-06, |
|
"loss": 0.269, |
|
"step": 2743 |
|
}, |
|
{ |
|
"epoch": 185.97, |
|
"eval_accuracy": 0.7307692307692307, |
|
"eval_loss": 0.8060269355773926, |
|
"eval_runtime": 4.679, |
|
"eval_samples_per_second": 61.125, |
|
"eval_steps_per_second": 3.847, |
|
"step": 2743 |
|
}, |
|
{ |
|
"epoch": 186.98, |
|
"grad_norm": 1.6407794952392578, |
|
"learning_rate": 7.076190476190476e-06, |
|
"loss": 0.2668, |
|
"step": 2758 |
|
}, |
|
{ |
|
"epoch": 186.98, |
|
"eval_accuracy": 0.7307692307692307, |
|
"eval_loss": 0.8030104041099548, |
|
"eval_runtime": 5.0657, |
|
"eval_samples_per_second": 56.458, |
|
"eval_steps_per_second": 3.553, |
|
"step": 2758 |
|
}, |
|
{ |
|
"epoch": 188.0, |
|
"grad_norm": 1.3405295610427856, |
|
"learning_rate": 6.933333333333334e-06, |
|
"loss": 0.2761, |
|
"step": 2773 |
|
}, |
|
{ |
|
"epoch": 188.0, |
|
"eval_accuracy": 0.7412587412587412, |
|
"eval_loss": 0.800729513168335, |
|
"eval_runtime": 5.2351, |
|
"eval_samples_per_second": 54.631, |
|
"eval_steps_per_second": 3.438, |
|
"step": 2773 |
|
}, |
|
{ |
|
"epoch": 188.95, |
|
"grad_norm": 1.5641790628433228, |
|
"learning_rate": 6.8e-06, |
|
"loss": 0.2731, |
|
"step": 2787 |
|
}, |
|
{ |
|
"epoch": 188.95, |
|
"eval_accuracy": 0.7342657342657343, |
|
"eval_loss": 0.8038657903671265, |
|
"eval_runtime": 5.309, |
|
"eval_samples_per_second": 53.871, |
|
"eval_steps_per_second": 3.39, |
|
"step": 2787 |
|
}, |
|
{ |
|
"epoch": 189.97, |
|
"grad_norm": 1.6037400960922241, |
|
"learning_rate": 6.657142857142857e-06, |
|
"loss": 0.2678, |
|
"step": 2802 |
|
}, |
|
{ |
|
"epoch": 189.97, |
|
"eval_accuracy": 0.7307692307692307, |
|
"eval_loss": 0.8054195642471313, |
|
"eval_runtime": 5.3283, |
|
"eval_samples_per_second": 53.675, |
|
"eval_steps_per_second": 3.378, |
|
"step": 2802 |
|
}, |
|
{ |
|
"epoch": 190.98, |
|
"grad_norm": 1.289398193359375, |
|
"learning_rate": 6.5142857142857145e-06, |
|
"loss": 0.2686, |
|
"step": 2817 |
|
}, |
|
{ |
|
"epoch": 190.98, |
|
"eval_accuracy": 0.7342657342657343, |
|
"eval_loss": 0.8041976094245911, |
|
"eval_runtime": 4.8716, |
|
"eval_samples_per_second": 58.708, |
|
"eval_steps_per_second": 3.695, |
|
"step": 2817 |
|
}, |
|
{ |
|
"epoch": 192.0, |
|
"grad_norm": 2.173532009124756, |
|
"learning_rate": 6.371428571428572e-06, |
|
"loss": 0.2721, |
|
"step": 2832 |
|
}, |
|
{ |
|
"epoch": 192.0, |
|
"eval_accuracy": 0.7307692307692307, |
|
"eval_loss": 0.806255042552948, |
|
"eval_runtime": 4.423, |
|
"eval_samples_per_second": 64.662, |
|
"eval_steps_per_second": 4.07, |
|
"step": 2832 |
|
}, |
|
{ |
|
"epoch": 192.95, |
|
"grad_norm": 1.6040128469467163, |
|
"learning_rate": 6.238095238095238e-06, |
|
"loss": 0.3111, |
|
"step": 2846 |
|
}, |
|
{ |
|
"epoch": 192.95, |
|
"eval_accuracy": 0.7377622377622378, |
|
"eval_loss": 0.8075191378593445, |
|
"eval_runtime": 5.3815, |
|
"eval_samples_per_second": 53.145, |
|
"eval_steps_per_second": 3.345, |
|
"step": 2846 |
|
}, |
|
{ |
|
"epoch": 193.97, |
|
"grad_norm": 1.7221040725708008, |
|
"learning_rate": 6.095238095238095e-06, |
|
"loss": 0.2525, |
|
"step": 2861 |
|
}, |
|
{ |
|
"epoch": 193.97, |
|
"eval_accuracy": 0.7272727272727273, |
|
"eval_loss": 0.8114144802093506, |
|
"eval_runtime": 5.0641, |
|
"eval_samples_per_second": 56.476, |
|
"eval_steps_per_second": 3.554, |
|
"step": 2861 |
|
}, |
|
{ |
|
"epoch": 194.98, |
|
"grad_norm": 1.8386750221252441, |
|
"learning_rate": 5.9523809523809525e-06, |
|
"loss": 0.2589, |
|
"step": 2876 |
|
}, |
|
{ |
|
"epoch": 194.98, |
|
"eval_accuracy": 0.7412587412587412, |
|
"eval_loss": 0.807061493396759, |
|
"eval_runtime": 4.8818, |
|
"eval_samples_per_second": 58.585, |
|
"eval_steps_per_second": 3.687, |
|
"step": 2876 |
|
}, |
|
{ |
|
"epoch": 196.0, |
|
"grad_norm": 0.7993654012680054, |
|
"learning_rate": 5.80952380952381e-06, |
|
"loss": 0.2597, |
|
"step": 2891 |
|
}, |
|
{ |
|
"epoch": 196.0, |
|
"eval_accuracy": 0.7377622377622378, |
|
"eval_loss": 0.8161465525627136, |
|
"eval_runtime": 5.7644, |
|
"eval_samples_per_second": 49.615, |
|
"eval_steps_per_second": 3.123, |
|
"step": 2891 |
|
}, |
|
{ |
|
"epoch": 196.95, |
|
"grad_norm": 1.822001338005066, |
|
"learning_rate": 5.676190476190476e-06, |
|
"loss": 0.3034, |
|
"step": 2905 |
|
}, |
|
{ |
|
"epoch": 196.95, |
|
"eval_accuracy": 0.7377622377622378, |
|
"eval_loss": 0.8163018822669983, |
|
"eval_runtime": 5.4614, |
|
"eval_samples_per_second": 52.367, |
|
"eval_steps_per_second": 3.296, |
|
"step": 2905 |
|
}, |
|
{ |
|
"epoch": 197.97, |
|
"grad_norm": 2.4189586639404297, |
|
"learning_rate": 5.5333333333333334e-06, |
|
"loss": 0.2721, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 197.97, |
|
"eval_accuracy": 0.7307692307692307, |
|
"eval_loss": 0.8150739073753357, |
|
"eval_runtime": 4.8549, |
|
"eval_samples_per_second": 58.91, |
|
"eval_steps_per_second": 3.708, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 198.98, |
|
"grad_norm": 1.6175144910812378, |
|
"learning_rate": 5.390476190476191e-06, |
|
"loss": 0.2783, |
|
"step": 2935 |
|
}, |
|
{ |
|
"epoch": 198.98, |
|
"eval_accuracy": 0.7447552447552448, |
|
"eval_loss": 0.8093083500862122, |
|
"eval_runtime": 4.4141, |
|
"eval_samples_per_second": 64.792, |
|
"eval_steps_per_second": 4.078, |
|
"step": 2935 |
|
}, |
|
{ |
|
"epoch": 200.0, |
|
"grad_norm": 1.7226580381393433, |
|
"learning_rate": 5.247619047619048e-06, |
|
"loss": 0.259, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 200.0, |
|
"eval_accuracy": 0.7412587412587412, |
|
"eval_loss": 0.8025929927825928, |
|
"eval_runtime": 5.4584, |
|
"eval_samples_per_second": 52.397, |
|
"eval_steps_per_second": 3.298, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 200.95, |
|
"grad_norm": 2.5513391494750977, |
|
"learning_rate": 5.114285714285714e-06, |
|
"loss": 0.2896, |
|
"step": 2964 |
|
}, |
|
{ |
|
"epoch": 200.95, |
|
"eval_accuracy": 0.7307692307692307, |
|
"eval_loss": 0.8070325255393982, |
|
"eval_runtime": 4.9068, |
|
"eval_samples_per_second": 58.286, |
|
"eval_steps_per_second": 3.668, |
|
"step": 2964 |
|
}, |
|
{ |
|
"epoch": 201.97, |
|
"grad_norm": 1.116995930671692, |
|
"learning_rate": 4.9714285714285715e-06, |
|
"loss": 0.2584, |
|
"step": 2979 |
|
}, |
|
{ |
|
"epoch": 201.97, |
|
"eval_accuracy": 0.7307692307692307, |
|
"eval_loss": 0.8112642765045166, |
|
"eval_runtime": 5.2035, |
|
"eval_samples_per_second": 54.963, |
|
"eval_steps_per_second": 3.459, |
|
"step": 2979 |
|
}, |
|
{ |
|
"epoch": 202.98, |
|
"grad_norm": 1.8753693103790283, |
|
"learning_rate": 4.828571428571429e-06, |
|
"loss": 0.2672, |
|
"step": 2994 |
|
}, |
|
{ |
|
"epoch": 202.98, |
|
"eval_accuracy": 0.7342657342657343, |
|
"eval_loss": 0.8096128106117249, |
|
"eval_runtime": 4.9714, |
|
"eval_samples_per_second": 57.529, |
|
"eval_steps_per_second": 3.621, |
|
"step": 2994 |
|
}, |
|
{ |
|
"epoch": 204.0, |
|
"grad_norm": 1.2377465963363647, |
|
"learning_rate": 4.685714285714286e-06, |
|
"loss": 0.2622, |
|
"step": 3009 |
|
}, |
|
{ |
|
"epoch": 204.0, |
|
"eval_accuracy": 0.7412587412587412, |
|
"eval_loss": 0.8059563040733337, |
|
"eval_runtime": 5.4896, |
|
"eval_samples_per_second": 52.098, |
|
"eval_steps_per_second": 3.279, |
|
"step": 3009 |
|
}, |
|
{ |
|
"epoch": 204.95, |
|
"grad_norm": 1.5672334432601929, |
|
"learning_rate": 4.552380952380952e-06, |
|
"loss": 0.2677, |
|
"step": 3023 |
|
}, |
|
{ |
|
"epoch": 204.95, |
|
"eval_accuracy": 0.7342657342657343, |
|
"eval_loss": 0.8084142804145813, |
|
"eval_runtime": 5.7936, |
|
"eval_samples_per_second": 49.365, |
|
"eval_steps_per_second": 3.107, |
|
"step": 3023 |
|
}, |
|
{ |
|
"epoch": 205.97, |
|
"grad_norm": 0.9782791137695312, |
|
"learning_rate": 4.4095238095238096e-06, |
|
"loss": 0.263, |
|
"step": 3038 |
|
}, |
|
{ |
|
"epoch": 205.97, |
|
"eval_accuracy": 0.7377622377622378, |
|
"eval_loss": 0.8009874820709229, |
|
"eval_runtime": 4.9002, |
|
"eval_samples_per_second": 58.364, |
|
"eval_steps_per_second": 3.673, |
|
"step": 3038 |
|
}, |
|
{ |
|
"epoch": 206.98, |
|
"grad_norm": 2.033815622329712, |
|
"learning_rate": 4.266666666666667e-06, |
|
"loss": 0.2608, |
|
"step": 3053 |
|
}, |
|
{ |
|
"epoch": 206.98, |
|
"eval_accuracy": 0.7447552447552448, |
|
"eval_loss": 0.7988951802253723, |
|
"eval_runtime": 5.2669, |
|
"eval_samples_per_second": 54.302, |
|
"eval_steps_per_second": 3.418, |
|
"step": 3053 |
|
}, |
|
{ |
|
"epoch": 208.0, |
|
"grad_norm": 2.0094175338745117, |
|
"learning_rate": 4.123809523809524e-06, |
|
"loss": 0.2528, |
|
"step": 3068 |
|
}, |
|
{ |
|
"epoch": 208.0, |
|
"eval_accuracy": 0.7447552447552448, |
|
"eval_loss": 0.7953740954399109, |
|
"eval_runtime": 6.154, |
|
"eval_samples_per_second": 46.474, |
|
"eval_steps_per_second": 2.925, |
|
"step": 3068 |
|
}, |
|
{ |
|
"epoch": 208.95, |
|
"grad_norm": 1.639021635055542, |
|
"learning_rate": 3.9904761904761905e-06, |
|
"loss": 0.2553, |
|
"step": 3082 |
|
}, |
|
{ |
|
"epoch": 208.95, |
|
"eval_accuracy": 0.7412587412587412, |
|
"eval_loss": 0.7965402603149414, |
|
"eval_runtime": 5.6365, |
|
"eval_samples_per_second": 50.741, |
|
"eval_steps_per_second": 3.193, |
|
"step": 3082 |
|
}, |
|
{ |
|
"epoch": 209.97, |
|
"grad_norm": 2.0743513107299805, |
|
"learning_rate": 3.847619047619048e-06, |
|
"loss": 0.2652, |
|
"step": 3097 |
|
}, |
|
{ |
|
"epoch": 209.97, |
|
"eval_accuracy": 0.7412587412587412, |
|
"eval_loss": 0.799461841583252, |
|
"eval_runtime": 4.1526, |
|
"eval_samples_per_second": 68.873, |
|
"eval_steps_per_second": 4.335, |
|
"step": 3097 |
|
}, |
|
{ |
|
"epoch": 210.98, |
|
"grad_norm": 1.6969355344772339, |
|
"learning_rate": 3.704761904761905e-06, |
|
"loss": 0.246, |
|
"step": 3112 |
|
}, |
|
{ |
|
"epoch": 210.98, |
|
"eval_accuracy": 0.7377622377622378, |
|
"eval_loss": 0.8026475310325623, |
|
"eval_runtime": 4.4732, |
|
"eval_samples_per_second": 63.936, |
|
"eval_steps_per_second": 4.024, |
|
"step": 3112 |
|
}, |
|
{ |
|
"epoch": 212.0, |
|
"grad_norm": 2.7497739791870117, |
|
"learning_rate": 3.561904761904762e-06, |
|
"loss": 0.2665, |
|
"step": 3127 |
|
}, |
|
{ |
|
"epoch": 212.0, |
|
"eval_accuracy": 0.7377622377622378, |
|
"eval_loss": 0.8049183487892151, |
|
"eval_runtime": 5.7553, |
|
"eval_samples_per_second": 49.693, |
|
"eval_steps_per_second": 3.128, |
|
"step": 3127 |
|
}, |
|
{ |
|
"epoch": 212.95, |
|
"grad_norm": 1.5636447668075562, |
|
"learning_rate": 3.4285714285714285e-06, |
|
"loss": 0.2731, |
|
"step": 3141 |
|
}, |
|
{ |
|
"epoch": 212.95, |
|
"eval_accuracy": 0.7377622377622378, |
|
"eval_loss": 0.8051545023918152, |
|
"eval_runtime": 5.9082, |
|
"eval_samples_per_second": 48.407, |
|
"eval_steps_per_second": 3.047, |
|
"step": 3141 |
|
}, |
|
{ |
|
"epoch": 213.97, |
|
"grad_norm": 0.8648618459701538, |
|
"learning_rate": 3.2857142857142857e-06, |
|
"loss": 0.2539, |
|
"step": 3156 |
|
}, |
|
{ |
|
"epoch": 213.97, |
|
"eval_accuracy": 0.7377622377622378, |
|
"eval_loss": 0.8088578581809998, |
|
"eval_runtime": 5.3417, |
|
"eval_samples_per_second": 53.541, |
|
"eval_steps_per_second": 3.37, |
|
"step": 3156 |
|
}, |
|
{ |
|
"epoch": 214.98, |
|
"grad_norm": 1.2608678340911865, |
|
"learning_rate": 3.142857142857143e-06, |
|
"loss": 0.2376, |
|
"step": 3171 |
|
}, |
|
{ |
|
"epoch": 214.98, |
|
"eval_accuracy": 0.7377622377622378, |
|
"eval_loss": 0.8115506768226624, |
|
"eval_runtime": 5.2232, |
|
"eval_samples_per_second": 54.756, |
|
"eval_steps_per_second": 3.446, |
|
"step": 3171 |
|
}, |
|
{ |
|
"epoch": 216.0, |
|
"grad_norm": 1.721834421157837, |
|
"learning_rate": 3e-06, |
|
"loss": 0.2667, |
|
"step": 3186 |
|
}, |
|
{ |
|
"epoch": 216.0, |
|
"eval_accuracy": 0.7412587412587412, |
|
"eval_loss": 0.80994713306427, |
|
"eval_runtime": 6.1297, |
|
"eval_samples_per_second": 46.658, |
|
"eval_steps_per_second": 2.937, |
|
"step": 3186 |
|
}, |
|
{ |
|
"epoch": 216.95, |
|
"grad_norm": 2.3429391384124756, |
|
"learning_rate": 2.866666666666667e-06, |
|
"loss": 0.2768, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 216.95, |
|
"eval_accuracy": 0.7412587412587412, |
|
"eval_loss": 0.807944655418396, |
|
"eval_runtime": 4.2958, |
|
"eval_samples_per_second": 66.576, |
|
"eval_steps_per_second": 4.19, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 217.97, |
|
"grad_norm": 2.041189193725586, |
|
"learning_rate": 2.7238095238095238e-06, |
|
"loss": 0.2551, |
|
"step": 3215 |
|
}, |
|
{ |
|
"epoch": 217.97, |
|
"eval_accuracy": 0.7447552447552448, |
|
"eval_loss": 0.8064839243888855, |
|
"eval_runtime": 5.6836, |
|
"eval_samples_per_second": 50.32, |
|
"eval_steps_per_second": 3.167, |
|
"step": 3215 |
|
}, |
|
{ |
|
"epoch": 218.98, |
|
"grad_norm": 1.3890482187271118, |
|
"learning_rate": 2.580952380952381e-06, |
|
"loss": 0.256, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 218.98, |
|
"eval_accuracy": 0.7412587412587412, |
|
"eval_loss": 0.8079765439033508, |
|
"eval_runtime": 5.1325, |
|
"eval_samples_per_second": 55.724, |
|
"eval_steps_per_second": 3.507, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 220.0, |
|
"grad_norm": 1.2208611965179443, |
|
"learning_rate": 2.438095238095238e-06, |
|
"loss": 0.2387, |
|
"step": 3245 |
|
}, |
|
{ |
|
"epoch": 220.0, |
|
"eval_accuracy": 0.7447552447552448, |
|
"eval_loss": 0.8076029419898987, |
|
"eval_runtime": 5.8067, |
|
"eval_samples_per_second": 49.254, |
|
"eval_steps_per_second": 3.1, |
|
"step": 3245 |
|
}, |
|
{ |
|
"epoch": 220.95, |
|
"grad_norm": 1.1974842548370361, |
|
"learning_rate": 2.3047619047619047e-06, |
|
"loss": 0.2736, |
|
"step": 3259 |
|
}, |
|
{ |
|
"epoch": 220.95, |
|
"eval_accuracy": 0.7447552447552448, |
|
"eval_loss": 0.8080862760543823, |
|
"eval_runtime": 4.7889, |
|
"eval_samples_per_second": 59.722, |
|
"eval_steps_per_second": 3.759, |
|
"step": 3259 |
|
}, |
|
{ |
|
"epoch": 221.97, |
|
"grad_norm": 1.3414784669876099, |
|
"learning_rate": 2.161904761904762e-06, |
|
"loss": 0.2551, |
|
"step": 3274 |
|
}, |
|
{ |
|
"epoch": 221.97, |
|
"eval_accuracy": 0.7447552447552448, |
|
"eval_loss": 0.809248149394989, |
|
"eval_runtime": 4.967, |
|
"eval_samples_per_second": 57.58, |
|
"eval_steps_per_second": 3.624, |
|
"step": 3274 |
|
}, |
|
{ |
|
"epoch": 222.98, |
|
"grad_norm": 1.6082030534744263, |
|
"learning_rate": 2.0190476190476195e-06, |
|
"loss": 0.2482, |
|
"step": 3289 |
|
}, |
|
{ |
|
"epoch": 222.98, |
|
"eval_accuracy": 0.7447552447552448, |
|
"eval_loss": 0.8084933757781982, |
|
"eval_runtime": 4.7114, |
|
"eval_samples_per_second": 60.704, |
|
"eval_steps_per_second": 3.821, |
|
"step": 3289 |
|
}, |
|
{ |
|
"epoch": 224.0, |
|
"grad_norm": 1.8340036869049072, |
|
"learning_rate": 1.876190476190476e-06, |
|
"loss": 0.2432, |
|
"step": 3304 |
|
}, |
|
{ |
|
"epoch": 224.0, |
|
"eval_accuracy": 0.7447552447552448, |
|
"eval_loss": 0.8085303902626038, |
|
"eval_runtime": 5.6786, |
|
"eval_samples_per_second": 50.365, |
|
"eval_steps_per_second": 3.17, |
|
"step": 3304 |
|
}, |
|
{ |
|
"epoch": 224.95, |
|
"grad_norm": 1.9474186897277832, |
|
"learning_rate": 1.742857142857143e-06, |
|
"loss": 0.2857, |
|
"step": 3318 |
|
}, |
|
{ |
|
"epoch": 224.95, |
|
"eval_accuracy": 0.7447552447552448, |
|
"eval_loss": 0.8078930974006653, |
|
"eval_runtime": 7.5562, |
|
"eval_samples_per_second": 37.85, |
|
"eval_steps_per_second": 2.382, |
|
"step": 3318 |
|
}, |
|
{ |
|
"epoch": 225.97, |
|
"grad_norm": 1.7995171546936035, |
|
"learning_rate": 1.6000000000000001e-06, |
|
"loss": 0.2434, |
|
"step": 3333 |
|
}, |
|
{ |
|
"epoch": 225.97, |
|
"eval_accuracy": 0.7447552447552448, |
|
"eval_loss": 0.8078944087028503, |
|
"eval_runtime": 5.9694, |
|
"eval_samples_per_second": 47.911, |
|
"eval_steps_per_second": 3.015, |
|
"step": 3333 |
|
}, |
|
{ |
|
"epoch": 226.98, |
|
"grad_norm": 1.4122081995010376, |
|
"learning_rate": 1.4571428571428571e-06, |
|
"loss": 0.2531, |
|
"step": 3348 |
|
}, |
|
{ |
|
"epoch": 226.98, |
|
"eval_accuracy": 0.7447552447552448, |
|
"eval_loss": 0.8078155517578125, |
|
"eval_runtime": 4.7783, |
|
"eval_samples_per_second": 59.853, |
|
"eval_steps_per_second": 3.767, |
|
"step": 3348 |
|
}, |
|
{ |
|
"epoch": 228.0, |
|
"grad_norm": 1.6360557079315186, |
|
"learning_rate": 1.3142857142857145e-06, |
|
"loss": 0.222, |
|
"step": 3363 |
|
}, |
|
{ |
|
"epoch": 228.0, |
|
"eval_accuracy": 0.7447552447552448, |
|
"eval_loss": 0.8092585206031799, |
|
"eval_runtime": 5.1759, |
|
"eval_samples_per_second": 55.256, |
|
"eval_steps_per_second": 3.478, |
|
"step": 3363 |
|
}, |
|
{ |
|
"epoch": 228.95, |
|
"grad_norm": 1.262143611907959, |
|
"learning_rate": 1.180952380952381e-06, |
|
"loss": 0.2651, |
|
"step": 3377 |
|
}, |
|
{ |
|
"epoch": 228.95, |
|
"eval_accuracy": 0.7447552447552448, |
|
"eval_loss": 0.8105089664459229, |
|
"eval_runtime": 6.304, |
|
"eval_samples_per_second": 45.368, |
|
"eval_steps_per_second": 2.855, |
|
"step": 3377 |
|
}, |
|
{ |
|
"epoch": 229.97, |
|
"grad_norm": 1.6864830255508423, |
|
"learning_rate": 1.038095238095238e-06, |
|
"loss": 0.2886, |
|
"step": 3392 |
|
}, |
|
{ |
|
"epoch": 229.97, |
|
"eval_accuracy": 0.7447552447552448, |
|
"eval_loss": 0.8115819096565247, |
|
"eval_runtime": 4.9539, |
|
"eval_samples_per_second": 57.732, |
|
"eval_steps_per_second": 3.634, |
|
"step": 3392 |
|
}, |
|
{ |
|
"epoch": 230.98, |
|
"grad_norm": 2.204496383666992, |
|
"learning_rate": 8.952380952380953e-07, |
|
"loss": 0.2534, |
|
"step": 3407 |
|
}, |
|
{ |
|
"epoch": 230.98, |
|
"eval_accuracy": 0.7447552447552448, |
|
"eval_loss": 0.8115825653076172, |
|
"eval_runtime": 5.2788, |
|
"eval_samples_per_second": 54.179, |
|
"eval_steps_per_second": 3.41, |
|
"step": 3407 |
|
}, |
|
{ |
|
"epoch": 232.0, |
|
"grad_norm": 1.418483853340149, |
|
"learning_rate": 7.523809523809525e-07, |
|
"loss": 0.2483, |
|
"step": 3422 |
|
}, |
|
{ |
|
"epoch": 232.0, |
|
"eval_accuracy": 0.7447552447552448, |
|
"eval_loss": 0.8116432428359985, |
|
"eval_runtime": 6.7125, |
|
"eval_samples_per_second": 42.607, |
|
"eval_steps_per_second": 2.682, |
|
"step": 3422 |
|
}, |
|
{ |
|
"epoch": 232.95, |
|
"grad_norm": 1.5767290592193604, |
|
"learning_rate": 6.190476190476191e-07, |
|
"loss": 0.272, |
|
"step": 3436 |
|
}, |
|
{ |
|
"epoch": 232.95, |
|
"eval_accuracy": 0.7447552447552448, |
|
"eval_loss": 0.8118025660514832, |
|
"eval_runtime": 6.2184, |
|
"eval_samples_per_second": 45.993, |
|
"eval_steps_per_second": 2.895, |
|
"step": 3436 |
|
}, |
|
{ |
|
"epoch": 233.97, |
|
"grad_norm": 1.1176037788391113, |
|
"learning_rate": 4.761904761904762e-07, |
|
"loss": 0.2562, |
|
"step": 3451 |
|
}, |
|
{ |
|
"epoch": 233.97, |
|
"eval_accuracy": 0.7447552447552448, |
|
"eval_loss": 0.8119560480117798, |
|
"eval_runtime": 5.396, |
|
"eval_samples_per_second": 53.002, |
|
"eval_steps_per_second": 3.336, |
|
"step": 3451 |
|
}, |
|
{ |
|
"epoch": 234.98, |
|
"grad_norm": 2.2027063369750977, |
|
"learning_rate": 3.3333333333333335e-07, |
|
"loss": 0.2531, |
|
"step": 3466 |
|
}, |
|
{ |
|
"epoch": 234.98, |
|
"eval_accuracy": 0.7447552447552448, |
|
"eval_loss": 0.8122441172599792, |
|
"eval_runtime": 4.8695, |
|
"eval_samples_per_second": 58.733, |
|
"eval_steps_per_second": 3.697, |
|
"step": 3466 |
|
}, |
|
{ |
|
"epoch": 236.0, |
|
"grad_norm": 1.984872579574585, |
|
"learning_rate": 1.9047619047619048e-07, |
|
"loss": 0.2397, |
|
"step": 3481 |
|
}, |
|
{ |
|
"epoch": 236.0, |
|
"eval_accuracy": 0.7447552447552448, |
|
"eval_loss": 0.8123078346252441, |
|
"eval_runtime": 5.7128, |
|
"eval_samples_per_second": 50.063, |
|
"eval_steps_per_second": 3.151, |
|
"step": 3481 |
|
}, |
|
{ |
|
"epoch": 236.95, |
|
"grad_norm": 1.5013829469680786, |
|
"learning_rate": 5.7142857142857144e-08, |
|
"loss": 0.2573, |
|
"step": 3495 |
|
}, |
|
{ |
|
"epoch": 236.95, |
|
"eval_accuracy": 0.7447552447552448, |
|
"eval_loss": 0.8123247027397156, |
|
"eval_runtime": 5.9714, |
|
"eval_samples_per_second": 47.895, |
|
"eval_steps_per_second": 3.014, |
|
"step": 3495 |
|
}, |
|
{ |
|
"epoch": 237.29, |
|
"grad_norm": 1.3738845586776733, |
|
"learning_rate": 9.523809523809524e-09, |
|
"loss": 0.2023, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 237.29, |
|
"eval_accuracy": 0.7447552447552448, |
|
"eval_loss": 0.8122022151947021, |
|
"eval_runtime": 5.4038, |
|
"eval_samples_per_second": 52.926, |
|
"eval_steps_per_second": 3.331, |
|
"step": 3500 |
|
} |
|
], |
|
"logging_steps": 100.0, |
|
"max_steps": 3500, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 250, |
|
"save_steps": 500, |
|
"total_flos": 1.939107912046368e+18, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|