|
{ |
|
"best_metric": 0.8431240916252136, |
|
"best_model_checkpoint": "./output/training_results/C019_random_sample_llama3-8b-base_instruct_20240504_182259/checkpoint-20", |
|
"epoch": 4.0, |
|
"eval_steps": 20, |
|
"global_step": 192, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.020833333333333332, |
|
"grad_norm": 0.0, |
|
"learning_rate": 0.0, |
|
"loss": 0.9787, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.10416666666666667, |
|
"grad_norm": 12.192626949795681, |
|
"learning_rate": 2.25e-06, |
|
"loss": 0.9575, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.20833333333333334, |
|
"grad_norm": 4.52126946078145, |
|
"learning_rate": 5.25e-06, |
|
"loss": 0.8647, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.3125, |
|
"grad_norm": 4.831880513298797, |
|
"learning_rate": 9e-06, |
|
"loss": 0.8335, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.4166666666666667, |
|
"grad_norm": 4.892917165423346, |
|
"learning_rate": 1.275e-05, |
|
"loss": 0.8164, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.4166666666666667, |
|
"eval_loss": 0.8431240916252136, |
|
"eval_runtime": 1.9991, |
|
"eval_samples_per_second": 170.072, |
|
"eval_steps_per_second": 1.501, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.5208333333333334, |
|
"grad_norm": 4.415048447330865, |
|
"learning_rate": 1.3195176200175283e-05, |
|
"loss": 0.8222, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.625, |
|
"grad_norm": 4.0612946994016434, |
|
"learning_rate": 9.515676612044427e-06, |
|
"loss": 0.8995, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.7291666666666666, |
|
"grad_norm": 3.749755477511079, |
|
"learning_rate": 6.797580677308734e-06, |
|
"loss": 0.8188, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.8333333333333334, |
|
"grad_norm": 4.222046474278687, |
|
"learning_rate": 4.808575415542887e-06, |
|
"loss": 0.7962, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.8333333333333334, |
|
"eval_loss": 0.8461718559265137, |
|
"eval_runtime": 1.9687, |
|
"eval_samples_per_second": 172.702, |
|
"eval_steps_per_second": 1.524, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.9375, |
|
"grad_norm": 3.9382293828818082, |
|
"learning_rate": 3.3676619069852654e-06, |
|
"loss": 0.8768, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 1.0416666666666667, |
|
"grad_norm": 3.5738168321090593, |
|
"learning_rate": 2.334947896124909e-06, |
|
"loss": 0.721, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 1.1458333333333333, |
|
"grad_norm": 3.06331450723627, |
|
"learning_rate": 1.603233215095547e-06, |
|
"loss": 0.4634, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 3.6197487832931996, |
|
"learning_rate": 1.0911174606561334e-06, |
|
"loss": 0.4335, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"eval_loss": 0.8650219440460205, |
|
"eval_runtime": 1.9704, |
|
"eval_samples_per_second": 172.55, |
|
"eval_steps_per_second": 1.523, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 1.3541666666666667, |
|
"grad_norm": 3.7661003279348537, |
|
"learning_rate": 7.373930741131784e-07, |
|
"loss": 0.3981, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 1.4583333333333333, |
|
"grad_norm": 4.160189703076145, |
|
"learning_rate": 4.965174334325768e-07, |
|
"loss": 0.4411, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 1.5625, |
|
"grad_norm": 4.5757896800217885, |
|
"learning_rate": 3.349849877937343e-07, |
|
"loss": 0.4387, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 1.6666666666666665, |
|
"grad_norm": 3.832613476181348, |
|
"learning_rate": 2.2844505627726646e-07, |
|
"loss": 0.4578, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 1.6666666666666665, |
|
"eval_loss": 0.8532621264457703, |
|
"eval_runtime": 1.9686, |
|
"eval_samples_per_second": 172.711, |
|
"eval_steps_per_second": 1.524, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 1.7708333333333335, |
|
"grad_norm": 3.3221659720718155, |
|
"learning_rate": 1.594328760942437e-07, |
|
"loss": 0.4262, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 1.875, |
|
"grad_norm": 3.4150483727266727, |
|
"learning_rate": 1.156010161291434e-07, |
|
"loss": 0.4247, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 1.9791666666666665, |
|
"grad_norm": 3.7582137935644253, |
|
"learning_rate": 8.835555547373544e-08, |
|
"loss": 0.4835, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 2.0833333333333335, |
|
"grad_norm": 3.4652201585485707, |
|
"learning_rate": 7.181664349277562e-08, |
|
"loss": 0.3944, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 2.0833333333333335, |
|
"eval_loss": 0.848356306552887, |
|
"eval_runtime": 1.9641, |
|
"eval_samples_per_second": 173.109, |
|
"eval_steps_per_second": 1.527, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 2.1875, |
|
"grad_norm": 3.1934106643776103, |
|
"learning_rate": 6.203637972657601e-08, |
|
"loss": 0.3698, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 2.2916666666666665, |
|
"grad_norm": 3.1313040719559755, |
|
"learning_rate": 5.6418543066491835e-08, |
|
"loss": 0.3999, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 2.3958333333333335, |
|
"grad_norm": 3.455998545535417, |
|
"learning_rate": 5.329471712759216e-08, |
|
"loss": 0.3575, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"grad_norm": 4.316899239508259, |
|
"learning_rate": 5.161995210302015e-08, |
|
"loss": 0.3997, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"eval_loss": 0.8528212904930115, |
|
"eval_runtime": 1.9732, |
|
"eval_samples_per_second": 172.312, |
|
"eval_steps_per_second": 1.52, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 2.6041666666666665, |
|
"grad_norm": 3.301008618352482, |
|
"learning_rate": 5.075841465580837e-08, |
|
"loss": 0.3796, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 2.7083333333333335, |
|
"grad_norm": 3.631871358764962, |
|
"learning_rate": 5.033564114946932e-08, |
|
"loss": 0.3768, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 2.8125, |
|
"grad_norm": 3.5129100349792806, |
|
"learning_rate": 5.013915282607116e-08, |
|
"loss": 0.3762, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 2.9166666666666665, |
|
"grad_norm": 3.374052247691012, |
|
"learning_rate": 5.005343402153039e-08, |
|
"loss": 0.3752, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 2.9166666666666665, |
|
"eval_loss": 0.8572859764099121, |
|
"eval_runtime": 1.9683, |
|
"eval_samples_per_second": 172.738, |
|
"eval_steps_per_second": 1.524, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 3.0208333333333335, |
|
"grad_norm": 2.957006248440721, |
|
"learning_rate": 5.001872829857116e-08, |
|
"loss": 0.3768, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 3.125, |
|
"grad_norm": 3.215743121265744, |
|
"learning_rate": 5.000587713853837e-08, |
|
"loss": 0.3765, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 3.2291666666666665, |
|
"grad_norm": 3.202764952665669, |
|
"learning_rate": 5.0001608748597456e-08, |
|
"loss": 0.3702, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 3.3333333333333335, |
|
"grad_norm": 4.415632412034816, |
|
"learning_rate": 5.0000370319656156e-08, |
|
"loss": 0.3697, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 3.3333333333333335, |
|
"eval_loss": 0.8608274459838867, |
|
"eval_runtime": 1.9662, |
|
"eval_samples_per_second": 172.924, |
|
"eval_steps_per_second": 1.526, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 3.4375, |
|
"grad_norm": 3.3333903437391057, |
|
"learning_rate": 5.0000067945715855e-08, |
|
"loss": 0.3654, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 3.5416666666666665, |
|
"grad_norm": 3.495136224595858, |
|
"learning_rate": 5.0000009144677036e-08, |
|
"loss": 0.3523, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 3.6458333333333335, |
|
"grad_norm": 3.291266939205713, |
|
"learning_rate": 5.0000000785521776e-08, |
|
"loss": 0.3668, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"grad_norm": 3.4466892020449005, |
|
"learning_rate": 5.000000003317662e-08, |
|
"loss": 0.3636, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"eval_loss": 0.8633963465690613, |
|
"eval_runtime": 1.9654, |
|
"eval_samples_per_second": 172.993, |
|
"eval_steps_per_second": 1.526, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 3.8541666666666665, |
|
"grad_norm": 3.7460542764286364, |
|
"learning_rate": 5.000000000038355e-08, |
|
"loss": 0.3717, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 3.9583333333333335, |
|
"grad_norm": 3.076543710036104, |
|
"learning_rate": 5.000000000000018e-08, |
|
"loss": 0.3687, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"step": 192, |
|
"total_flos": 5362900008960.0, |
|
"train_loss": 0.5116761994237701, |
|
"train_runtime": 1046.9703, |
|
"train_samples_per_second": 11.668, |
|
"train_steps_per_second": 0.183 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 192, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 4, |
|
"save_steps": 20, |
|
"total_flos": 5362900008960.0, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|