|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 10.0, |
|
"eval_steps": 1000.0, |
|
"global_step": 25000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 3.1640737056732178, |
|
"learning_rate": 4.9e-05, |
|
"loss": 2.212, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 23.492929458618164, |
|
"learning_rate": 4.8e-05, |
|
"loss": 2.0008, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"grad_norm": 2.86879563331604, |
|
"learning_rate": 4.7e-05, |
|
"loss": 1.9454, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 2.5470571517944336, |
|
"learning_rate": 4.600000000000001e-05, |
|
"loss": 1.9098, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 3.299417018890381, |
|
"learning_rate": 4.5e-05, |
|
"loss": 1.8775, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 3.0483977794647217, |
|
"learning_rate": 4.4000000000000006e-05, |
|
"loss": 1.8398, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 3.0028903484344482, |
|
"learning_rate": 4.3e-05, |
|
"loss": 1.8147, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 2.972844123840332, |
|
"learning_rate": 4.2e-05, |
|
"loss": 1.8035, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 2.82574200630188, |
|
"learning_rate": 4.1e-05, |
|
"loss": 1.7974, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 2.6531100273132324, |
|
"learning_rate": 4e-05, |
|
"loss": 1.7818, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"grad_norm": 3.6078109741210938, |
|
"learning_rate": 3.9000000000000006e-05, |
|
"loss": 1.7492, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"grad_norm": 2.573765516281128, |
|
"learning_rate": 3.8e-05, |
|
"loss": 1.7536, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"grad_norm": 3.07545804977417, |
|
"learning_rate": 3.7e-05, |
|
"loss": 1.7341, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"grad_norm": 2.4871368408203125, |
|
"learning_rate": 3.6e-05, |
|
"loss": 1.7362, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"grad_norm": 2.881721019744873, |
|
"learning_rate": 3.5e-05, |
|
"loss": 1.7426, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"grad_norm": 3.6232352256774902, |
|
"learning_rate": 3.4000000000000007e-05, |
|
"loss": 1.6958, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"grad_norm": 2.4172329902648926, |
|
"learning_rate": 3.3e-05, |
|
"loss": 1.6961, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 3.6, |
|
"grad_norm": 2.5114290714263916, |
|
"learning_rate": 3.2000000000000005e-05, |
|
"loss": 1.6961, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 3.8, |
|
"grad_norm": 2.7974891662597656, |
|
"learning_rate": 3.1e-05, |
|
"loss": 1.6831, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"grad_norm": 2.511326551437378, |
|
"learning_rate": 3e-05, |
|
"loss": 1.7117, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 4.2, |
|
"grad_norm": 2.3756415843963623, |
|
"learning_rate": 2.9e-05, |
|
"loss": 1.6684, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 4.4, |
|
"grad_norm": 2.2897322177886963, |
|
"learning_rate": 2.8000000000000003e-05, |
|
"loss": 1.6826, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 4.6, |
|
"grad_norm": 2.7029635906219482, |
|
"learning_rate": 2.7000000000000002e-05, |
|
"loss": 1.6649, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 4.8, |
|
"grad_norm": 2.453059673309326, |
|
"learning_rate": 2.6000000000000002e-05, |
|
"loss": 1.6488, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"grad_norm": 2.173687219619751, |
|
"learning_rate": 2.5e-05, |
|
"loss": 1.652, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 5.2, |
|
"grad_norm": 2.9122655391693115, |
|
"learning_rate": 2.4e-05, |
|
"loss": 1.6354, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 5.4, |
|
"grad_norm": 2.8792638778686523, |
|
"learning_rate": 2.3000000000000003e-05, |
|
"loss": 1.6423, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 5.6, |
|
"grad_norm": 3.0389626026153564, |
|
"learning_rate": 2.2000000000000003e-05, |
|
"loss": 1.6591, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 5.8, |
|
"grad_norm": 2.6412346363067627, |
|
"learning_rate": 2.1e-05, |
|
"loss": 1.6365, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"grad_norm": 2.6655163764953613, |
|
"learning_rate": 2e-05, |
|
"loss": 1.6133, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 6.2, |
|
"grad_norm": 2.6776795387268066, |
|
"learning_rate": 1.9e-05, |
|
"loss": 1.603, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 6.4, |
|
"grad_norm": 3.4206550121307373, |
|
"learning_rate": 1.8e-05, |
|
"loss": 1.6142, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 6.6, |
|
"grad_norm": 3.59487247467041, |
|
"learning_rate": 1.7000000000000003e-05, |
|
"loss": 1.6193, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 6.8, |
|
"grad_norm": 2.707260847091675, |
|
"learning_rate": 1.6000000000000003e-05, |
|
"loss": 1.618, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"grad_norm": 2.4932329654693604, |
|
"learning_rate": 1.5e-05, |
|
"loss": 1.6179, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 7.2, |
|
"grad_norm": 2.7114312648773193, |
|
"learning_rate": 1.4000000000000001e-05, |
|
"loss": 1.6064, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 7.4, |
|
"grad_norm": 2.9900901317596436, |
|
"learning_rate": 1.3000000000000001e-05, |
|
"loss": 1.6079, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 7.6, |
|
"grad_norm": 2.5069007873535156, |
|
"learning_rate": 1.2e-05, |
|
"loss": 1.5917, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 7.8, |
|
"grad_norm": 2.816830635070801, |
|
"learning_rate": 1.1000000000000001e-05, |
|
"loss": 1.6159, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"grad_norm": 2.500258445739746, |
|
"learning_rate": 1e-05, |
|
"loss": 1.586, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 8.2, |
|
"grad_norm": 2.392134189605713, |
|
"learning_rate": 9e-06, |
|
"loss": 1.5948, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 8.4, |
|
"grad_norm": 2.19275164604187, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 1.5904, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 8.6, |
|
"grad_norm": 2.1356213092803955, |
|
"learning_rate": 7.000000000000001e-06, |
|
"loss": 1.5772, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 8.8, |
|
"grad_norm": 2.0939230918884277, |
|
"learning_rate": 6e-06, |
|
"loss": 1.604, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"grad_norm": 2.8503177165985107, |
|
"learning_rate": 5e-06, |
|
"loss": 1.5878, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 9.2, |
|
"grad_norm": 2.6212220191955566, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 1.5893, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 9.4, |
|
"grad_norm": 2.9165279865264893, |
|
"learning_rate": 3e-06, |
|
"loss": 1.5796, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 9.6, |
|
"grad_norm": 2.8491413593292236, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 1.5832, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 9.8, |
|
"grad_norm": 3.690595865249634, |
|
"learning_rate": 1.0000000000000002e-06, |
|
"loss": 1.5872, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"grad_norm": 3.193439245223999, |
|
"learning_rate": 0.0, |
|
"loss": 1.5781, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"step": 25000, |
|
"total_flos": 3.2973681787109376e+16, |
|
"train_loss": 1.6927287036132812, |
|
"train_runtime": 2178.3738, |
|
"train_samples_per_second": 91.812, |
|
"train_steps_per_second": 11.476 |
|
} |
|
], |
|
"logging_steps": 500, |
|
"max_steps": 25000, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 10, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 3.2973681787109376e+16, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|