|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 3.063725490196078, |
|
"global_step": 50000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 2.3999999999999997e-05, |
|
"loss": 0.8848, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 4.7999999999999994e-05, |
|
"loss": 0.7411, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"eval_loss": 0.9070383906364441, |
|
"eval_runtime": 1.3223, |
|
"eval_samples_per_second": 756.258, |
|
"eval_steps_per_second": 12.1, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 7.199999999999999e-05, |
|
"loss": 0.7398, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.599999999999999e-05, |
|
"loss": 0.7395, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"eval_loss": 0.9064154028892517, |
|
"eval_runtime": 1.194, |
|
"eval_samples_per_second": 837.495, |
|
"eval_steps_per_second": 13.4, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00011999999999999999, |
|
"loss": 0.7392, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00014399999999999998, |
|
"loss": 0.7387, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"eval_loss": 0.9047173857688904, |
|
"eval_runtime": 1.3077, |
|
"eval_samples_per_second": 764.675, |
|
"eval_steps_per_second": 12.235, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.000168, |
|
"loss": 0.7384, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00019199999999999998, |
|
"loss": 0.7382, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"eval_loss": 0.9014638662338257, |
|
"eval_runtime": 1.2645, |
|
"eval_samples_per_second": 790.849, |
|
"eval_steps_per_second": 12.654, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00021599999999999996, |
|
"loss": 0.7382, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00023999999999999998, |
|
"loss": 0.7381, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"eval_loss": 0.9044438600540161, |
|
"eval_runtime": 1.2394, |
|
"eval_samples_per_second": 806.828, |
|
"eval_steps_per_second": 12.909, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00026399999999999997, |
|
"loss": 0.7379, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00028799999999999995, |
|
"loss": 0.7379, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"eval_loss": 0.9041813015937805, |
|
"eval_runtime": 1.2368, |
|
"eval_samples_per_second": 808.566, |
|
"eval_steps_per_second": 12.937, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.000312, |
|
"loss": 0.7381, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.000336, |
|
"loss": 0.7379, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"eval_loss": 0.9053561687469482, |
|
"eval_runtime": 1.2638, |
|
"eval_samples_per_second": 791.247, |
|
"eval_steps_per_second": 12.66, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00035999999999999997, |
|
"loss": 0.7378, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00038399999999999996, |
|
"loss": 0.7378, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"eval_loss": 0.9035025835037231, |
|
"eval_runtime": 1.2079, |
|
"eval_samples_per_second": 827.911, |
|
"eval_steps_per_second": 13.247, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.000408, |
|
"loss": 0.7378, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00043199999999999993, |
|
"loss": 0.7378, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"eval_loss": 0.9025757312774658, |
|
"eval_runtime": 1.2745, |
|
"eval_samples_per_second": 784.598, |
|
"eval_steps_per_second": 12.554, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 0.00045599999999999997, |
|
"loss": 0.7375, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 0.00047999999999999996, |
|
"loss": 0.7371, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"eval_loss": 0.903812050819397, |
|
"eval_runtime": 1.2376, |
|
"eval_samples_per_second": 807.983, |
|
"eval_steps_per_second": 12.928, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 0.0005039999999999999, |
|
"loss": 0.7369, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 0.0005279999999999999, |
|
"loss": 0.7369, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"eval_loss": 0.9027426838874817, |
|
"eval_runtime": 1.2459, |
|
"eval_samples_per_second": 802.657, |
|
"eval_steps_per_second": 12.843, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 0.000552, |
|
"loss": 0.7369, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 0.0005759999999999999, |
|
"loss": 0.7368, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"eval_loss": 0.9022310972213745, |
|
"eval_runtime": 1.2295, |
|
"eval_samples_per_second": 813.368, |
|
"eval_steps_per_second": 13.014, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 0.0006, |
|
"loss": 0.7367, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 0.0005999935478721662, |
|
"loss": 0.7368, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"eval_loss": 0.8987133502960205, |
|
"eval_runtime": 1.2643, |
|
"eval_samples_per_second": 790.939, |
|
"eval_steps_per_second": 12.655, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 0.000599974191770902, |
|
"loss": 0.7375, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 0.0005999419325429058, |
|
"loss": 0.7374, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"eval_loss": 0.9013972282409668, |
|
"eval_runtime": 1.236, |
|
"eval_samples_per_second": 809.042, |
|
"eval_steps_per_second": 12.945, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 0.0005998967715993009, |
|
"loss": 0.7369, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 0.0005998387109155732, |
|
"loss": 0.7369, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"eval_loss": 0.9001522660255432, |
|
"eval_runtime": 1.235, |
|
"eval_samples_per_second": 809.697, |
|
"eval_steps_per_second": 12.955, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 0.000599767753031485, |
|
"loss": 0.737, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 0.0005996839010509641, |
|
"loss": 0.7369, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"eval_loss": 0.9001864790916443, |
|
"eval_runtime": 1.1685, |
|
"eval_samples_per_second": 855.784, |
|
"eval_steps_per_second": 13.693, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 0.0005995871586419678, |
|
"loss": 0.7369, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 0.0005994775300363225, |
|
"loss": 0.7372, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"eval_loss": 0.9018534421920776, |
|
"eval_runtime": 1.1865, |
|
"eval_samples_per_second": 842.819, |
|
"eval_steps_per_second": 13.485, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 0.0005993550200295384, |
|
"loss": 0.7371, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 0.0005992196339806002, |
|
"loss": 0.737, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"eval_loss": 0.9001176953315735, |
|
"eval_runtime": 1.2219, |
|
"eval_samples_per_second": 818.375, |
|
"eval_steps_per_second": 13.094, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 0.0005990713778117324, |
|
"loss": 0.7369, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 0.0005989102580081398, |
|
"loss": 0.737, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"eval_loss": 0.9006143808364868, |
|
"eval_runtime": 1.2544, |
|
"eval_samples_per_second": 797.163, |
|
"eval_steps_per_second": 12.755, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 0.0005987362816177249, |
|
"loss": 0.7369, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 0.0005985494562507783, |
|
"loss": 0.7369, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"eval_loss": 0.9006676077842712, |
|
"eval_runtime": 1.2815, |
|
"eval_samples_per_second": 780.312, |
|
"eval_steps_per_second": 12.485, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 0.000598349790079647, |
|
"loss": 0.737, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 0.000598137291838376, |
|
"loss": 0.7365, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"eval_loss": 0.8698467016220093, |
|
"eval_runtime": 1.2584, |
|
"eval_samples_per_second": 794.654, |
|
"eval_steps_per_second": 12.714, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 0.000597911970822327, |
|
"loss": 0.7364, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 0.000597673836887771, |
|
"loss": 0.7363, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"eval_loss": 0.870046079158783, |
|
"eval_runtime": 1.2567, |
|
"eval_samples_per_second": 795.734, |
|
"eval_steps_per_second": 12.732, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 0.0005974229004514577, |
|
"loss": 0.7363, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 0.0005971591724901598, |
|
"loss": 0.7366, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"eval_loss": 0.9021453857421875, |
|
"eval_runtime": 1.277, |
|
"eval_samples_per_second": 783.107, |
|
"eval_steps_per_second": 12.53, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 0.0005968826645401927, |
|
"loss": 0.7368, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 0.00059659338869691, |
|
"loss": 0.7362, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"eval_loss": 0.8762597441673279, |
|
"eval_runtime": 1.2918, |
|
"eval_samples_per_second": 774.091, |
|
"eval_steps_per_second": 12.385, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 0.0005962913576141742, |
|
"loss": 0.7354, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 0.000595976584503803, |
|
"loss": 0.7082, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"eval_loss": 0.8719052672386169, |
|
"eval_runtime": 1.2648, |
|
"eval_samples_per_second": 790.61, |
|
"eval_steps_per_second": 12.65, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 0.0005956490831349923, |
|
"loss": 0.6914, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 0.0005953088678337129, |
|
"loss": 0.6774, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"eval_loss": 0.8876385688781738, |
|
"eval_runtime": 1.3046, |
|
"eval_samples_per_second": 766.511, |
|
"eval_steps_per_second": 12.264, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 0.0005949559534820841, |
|
"loss": 0.6675, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 0.0005945903555177229, |
|
"loss": 0.6525, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"eval_loss": 0.8904880285263062, |
|
"eval_runtime": 1.319, |
|
"eval_samples_per_second": 758.176, |
|
"eval_steps_per_second": 12.131, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 0.0005942120899330687, |
|
"loss": 0.6186, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 0.0005938211732746836, |
|
"loss": 0.6022, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"eval_loss": 0.8856265544891357, |
|
"eval_runtime": 1.2877, |
|
"eval_samples_per_second": 776.587, |
|
"eval_steps_per_second": 12.425, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 0.0005934176226425286, |
|
"loss": 0.593, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 0.0005930014556892158, |
|
"loss": 0.5874, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"eval_loss": 0.8793675303459167, |
|
"eval_runtime": 1.2489, |
|
"eval_samples_per_second": 800.707, |
|
"eval_steps_per_second": 12.811, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 0.0005925726906192357, |
|
"loss": 0.5837, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 0.0005921313461881617, |
|
"loss": 0.5765, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"eval_loss": 0.8805580735206604, |
|
"eval_runtime": 1.2758, |
|
"eval_samples_per_second": 783.816, |
|
"eval_steps_per_second": 12.541, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 0.0005916774417018287, |
|
"loss": 0.5714, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 0.0005912109970154897, |
|
"loss": 0.5685, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"eval_loss": 0.8747313022613525, |
|
"eval_runtime": 1.3003, |
|
"eval_samples_per_second": 769.027, |
|
"eval_steps_per_second": 12.304, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 0.0005907320325329461, |
|
"loss": 0.566, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 0.0005902405692056561, |
|
"loss": 0.564, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"eval_loss": 0.8779122233390808, |
|
"eval_runtime": 1.293, |
|
"eval_samples_per_second": 773.405, |
|
"eval_steps_per_second": 12.374, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 0.0005897366285318178, |
|
"loss": 0.5617, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 0.0005892202325554288, |
|
"loss": 0.5606, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"eval_loss": 0.8761873245239258, |
|
"eval_runtime": 1.3816, |
|
"eval_samples_per_second": 723.824, |
|
"eval_steps_per_second": 11.581, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 0.0005886914038653217, |
|
"loss": 0.5583, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 0.0005881501655941771, |
|
"loss": 0.5574, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"eval_loss": 0.8702684044837952, |
|
"eval_runtime": 1.3086, |
|
"eval_samples_per_second": 764.158, |
|
"eval_steps_per_second": 12.227, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 0.00058759654141751, |
|
"loss": 0.5548, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 0.0005870305555526355, |
|
"loss": 0.5528, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"eval_loss": 0.8663867115974426, |
|
"eval_runtime": 1.2619, |
|
"eval_samples_per_second": 792.469, |
|
"eval_steps_per_second": 12.679, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 0.0005864522327576088, |
|
"loss": 0.5509, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 0.0005858615983301424, |
|
"loss": 0.5494, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"eval_loss": 0.8717171549797058, |
|
"eval_runtime": 1.2691, |
|
"eval_samples_per_second": 787.953, |
|
"eval_steps_per_second": 12.607, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 0.0005852586781064997, |
|
"loss": 0.5471, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 0.0005846434984603645, |
|
"loss": 0.5448, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"eval_loss": 0.8672583103179932, |
|
"eval_runtime": 1.268, |
|
"eval_samples_per_second": 788.629, |
|
"eval_steps_per_second": 12.618, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 0.0005840160863016872, |
|
"loss": 0.5433, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 0.0005833764690755083, |
|
"loss": 0.5419, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"eval_loss": 0.8636866211891174, |
|
"eval_runtime": 1.3526, |
|
"eval_samples_per_second": 739.295, |
|
"eval_steps_per_second": 11.829, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 0.0005827246747607574, |
|
"loss": 0.5398, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 0.0005820607318690293, |
|
"loss": 0.5385, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"eval_loss": 0.8634054064750671, |
|
"eval_runtime": 1.2705, |
|
"eval_samples_per_second": 787.122, |
|
"eval_steps_per_second": 12.594, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 0.0005813846694433368, |
|
"loss": 0.5374, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 0.0005806965170568409, |
|
"loss": 0.536, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"eval_loss": 0.8661102652549744, |
|
"eval_runtime": 1.3401, |
|
"eval_samples_per_second": 746.204, |
|
"eval_steps_per_second": 11.939, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 0.0005799963048115559, |
|
"loss": 0.5353, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 0.0005792840633370341, |
|
"loss": 0.5336, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"eval_loss": 0.8631040453910828, |
|
"eval_runtime": 1.3088, |
|
"eval_samples_per_second": 764.037, |
|
"eval_steps_per_second": 12.225, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 0.0005785598237890247, |
|
"loss": 0.5327, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 0.0005778236178481119, |
|
"loss": 0.5316, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"eval_loss": 0.8605585098266602, |
|
"eval_runtime": 1.2836, |
|
"eval_samples_per_second": 779.077, |
|
"eval_steps_per_second": 12.465, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 0.0005770754777183285, |
|
"loss": 0.5306, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 0.0005763154361257473, |
|
"loss": 0.5297, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"eval_loss": 0.8589205145835876, |
|
"eval_runtime": 1.2763, |
|
"eval_samples_per_second": 783.485, |
|
"eval_steps_per_second": 12.536, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 0.0005755435263170498, |
|
"loss": 0.5287, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 0.0005747597820580717, |
|
"loss": 0.5305, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"eval_loss": 0.8569635152816772, |
|
"eval_runtime": 1.2713, |
|
"eval_samples_per_second": 786.624, |
|
"eval_steps_per_second": 12.586, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 0.000573964237632326, |
|
"loss": 0.527, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 0.0005731569278395029, |
|
"loss": 0.5262, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"eval_loss": 0.8558768033981323, |
|
"eval_runtime": 1.2738, |
|
"eval_samples_per_second": 785.051, |
|
"eval_steps_per_second": 12.561, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 0.0005723378879939481, |
|
"loss": 0.5254, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 0.0005715071539231178, |
|
"loss": 0.5247, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"eval_loss": 0.8633874654769897, |
|
"eval_runtime": 1.2747, |
|
"eval_samples_per_second": 784.504, |
|
"eval_steps_per_second": 12.552, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 0.0005706647619660116, |
|
"loss": 0.5243, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 0.0005698107489715823, |
|
"loss": 0.5235, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"eval_loss": 0.8606237769126892, |
|
"eval_runtime": 1.3283, |
|
"eval_samples_per_second": 752.838, |
|
"eval_steps_per_second": 12.045, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 0.0005689451522971252, |
|
"loss": 0.5228, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 0.0005680680098066429, |
|
"loss": 0.5227, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"eval_loss": 0.8610469698905945, |
|
"eval_runtime": 1.2783, |
|
"eval_samples_per_second": 782.274, |
|
"eval_steps_per_second": 12.516, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 0.0005671793598691895, |
|
"loss": 0.5215, |
|
"step": 48500 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 0.0005662792413571921, |
|
"loss": 0.5206, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 0.8610268235206604, |
|
"eval_runtime": 1.3555, |
|
"eval_samples_per_second": 737.72, |
|
"eval_steps_per_second": 11.804, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 0.0005653676936447504, |
|
"loss": 0.5201, |
|
"step": 49500 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 0.0005644447566059142, |
|
"loss": 0.5194, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"eval_loss": 0.8611247539520264, |
|
"eval_runtime": 1.2902, |
|
"eval_samples_per_second": 775.099, |
|
"eval_steps_per_second": 12.402, |
|
"step": 50000 |
|
} |
|
], |
|
"max_steps": 250000, |
|
"num_train_epochs": 16, |
|
"total_flos": 8.008099828181752e+20, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|