|
{ |
|
"best_metric": 0.1676628440618515, |
|
"best_model_checkpoint": "saves/Qwen1.5-1.8B/WordProblem/checkpoint-9000", |
|
"epoch": 0.9999725884707108, |
|
"eval_steps": 1500, |
|
"global_step": 9120, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.03289383514706285, |
|
"grad_norm": 4.8125, |
|
"learning_rate": 3e-05, |
|
"loss": 0.7143, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.0657876702941257, |
|
"grad_norm": 3.5, |
|
"learning_rate": 4.998339850669331e-05, |
|
"loss": 0.2219, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.09868150544118856, |
|
"grad_norm": 4.40625, |
|
"learning_rate": 4.9734816848192624e-05, |
|
"loss": 0.2074, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.1315753405882514, |
|
"grad_norm": 4.03125, |
|
"learning_rate": 4.9190839785031474e-05, |
|
"loss": 0.1906, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.16446917573531428, |
|
"grad_norm": 3.78125, |
|
"learning_rate": 4.835796376008569e-05, |
|
"loss": 0.1923, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.16446917573531428, |
|
"eval_loss": 0.1849033087491989, |
|
"eval_runtime": 79.6037, |
|
"eval_samples_per_second": 37.423, |
|
"eval_steps_per_second": 9.359, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.19736301088237712, |
|
"grad_norm": 4.15625, |
|
"learning_rate": 4.7246135390382216e-05, |
|
"loss": 0.1839, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.23025684602944, |
|
"grad_norm": 4.5, |
|
"learning_rate": 4.586863267968384e-05, |
|
"loss": 0.1938, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.2631506811765028, |
|
"grad_norm": 3.1875, |
|
"learning_rate": 4.4241906446007296e-05, |
|
"loss": 0.1863, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.2960445163235657, |
|
"grad_norm": 2.71875, |
|
"learning_rate": 4.238538385782601e-05, |
|
"loss": 0.1797, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.32893835147062855, |
|
"grad_norm": 3.65625, |
|
"learning_rate": 4.032123642522486e-05, |
|
"loss": 0.176, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.32893835147062855, |
|
"eval_loss": 0.1760552078485489, |
|
"eval_runtime": 79.6333, |
|
"eval_samples_per_second": 37.409, |
|
"eval_steps_per_second": 9.355, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.3618321866176914, |
|
"grad_norm": 4.0, |
|
"learning_rate": 3.8074115216771435e-05, |
|
"loss": 0.1791, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.39472602176475424, |
|
"grad_norm": 3.90625, |
|
"learning_rate": 3.567085646427478e-05, |
|
"loss": 0.1808, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.4276198569118171, |
|
"grad_norm": 3.421875, |
|
"learning_rate": 3.3140161071244915e-05, |
|
"loss": 0.1805, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.46051369205888, |
|
"grad_norm": 2.640625, |
|
"learning_rate": 3.05122518525215e-05, |
|
"loss": 0.1738, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.49340752720594283, |
|
"grad_norm": 4.5, |
|
"learning_rate": 2.781851259848554e-05, |
|
"loss": 0.1736, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.49340752720594283, |
|
"eval_loss": 0.17090687155723572, |
|
"eval_runtime": 79.6329, |
|
"eval_samples_per_second": 37.409, |
|
"eval_steps_per_second": 9.355, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.5263013623530056, |
|
"grad_norm": 3.578125, |
|
"learning_rate": 2.509111327432736e-05, |
|
"loss": 0.1709, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.5591951975000685, |
|
"grad_norm": 3.515625, |
|
"learning_rate": 2.236262583042668e-05, |
|
"loss": 0.1775, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.5920890326471314, |
|
"grad_norm": 4.9375, |
|
"learning_rate": 1.966563521202681e-05, |
|
"loss": 0.1759, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.6249828677941942, |
|
"grad_norm": 3.640625, |
|
"learning_rate": 1.7032350213717874e-05, |
|
"loss": 0.1754, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.6578767029412571, |
|
"grad_norm": 3.578125, |
|
"learning_rate": 1.4494218826096939e-05, |
|
"loss": 0.1688, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.6578767029412571, |
|
"eval_loss": 0.16823573410511017, |
|
"eval_runtime": 79.6163, |
|
"eval_samples_per_second": 37.417, |
|
"eval_steps_per_second": 9.357, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.6907705380883199, |
|
"grad_norm": 3.6875, |
|
"learning_rate": 1.2081552668325321e-05, |
|
"loss": 0.1707, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.7236643732353828, |
|
"grad_norm": 4.0625, |
|
"learning_rate": 9.82316499179518e-06, |
|
"loss": 0.171, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.7565582083824457, |
|
"grad_norm": 3.984375, |
|
"learning_rate": 7.74602657804425e-06, |
|
"loss": 0.1702, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.7894520435295085, |
|
"grad_norm": 3.421875, |
|
"learning_rate": 5.874943640356082e-06, |
|
"loss": 0.1718, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.8223458786765714, |
|
"grad_norm": 4.1875, |
|
"learning_rate": 4.232261575703861e-06, |
|
"loss": 0.1689, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.8223458786765714, |
|
"eval_loss": 0.16773280501365662, |
|
"eval_runtime": 79.6198, |
|
"eval_samples_per_second": 37.415, |
|
"eval_steps_per_second": 9.357, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.8552397138236342, |
|
"grad_norm": 2.703125, |
|
"learning_rate": 2.83759810497852e-06, |
|
"loss": 0.1692, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 0.8881335489706971, |
|
"grad_norm": 4.0625, |
|
"learning_rate": 1.70760898847247e-06, |
|
"loss": 0.1787, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 0.92102738411776, |
|
"grad_norm": 4.71875, |
|
"learning_rate": 8.557891145603042e-07, |
|
"loss": 0.1733, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 0.9539212192648228, |
|
"grad_norm": 3.6875, |
|
"learning_rate": 2.923113370737779e-07, |
|
"loss": 0.1741, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 0.9868150544118857, |
|
"grad_norm": 4.3125, |
|
"learning_rate": 2.3904986054812396e-08, |
|
"loss": 0.168, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.9868150544118857, |
|
"eval_loss": 0.1676628440618515, |
|
"eval_runtime": 79.7085, |
|
"eval_samples_per_second": 37.374, |
|
"eval_steps_per_second": 9.347, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.9999725884707108, |
|
"step": 9120, |
|
"total_flos": 3.6976201313039155e+17, |
|
"train_loss": 0.19690959160788019, |
|
"train_runtime": 20589.0374, |
|
"train_samples_per_second": 7.087, |
|
"train_steps_per_second": 0.443 |
|
} |
|
], |
|
"logging_steps": 300, |
|
"max_steps": 9120, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 3000, |
|
"total_flos": 3.6976201313039155e+17, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|