|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.20976087260523005, |
|
"eval_steps": 5000, |
|
"global_step": 3000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.006992029086841001, |
|
"grad_norm": 75.90129852294922, |
|
"learning_rate": 6.780845858091577e-07, |
|
"loss": 16.865, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.013984058173682002, |
|
"grad_norm": 71.16612243652344, |
|
"learning_rate": 1.3771408598392172e-06, |
|
"loss": 16.1556, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.020976087260523003, |
|
"grad_norm": 130.67613220214844, |
|
"learning_rate": 2.0761971338692767e-06, |
|
"loss": 14.8008, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.027968116347364003, |
|
"grad_norm": 36.4179801940918, |
|
"learning_rate": 2.7682628451590356e-06, |
|
"loss": 12.4025, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.034960145434205, |
|
"grad_norm": 34.748844146728516, |
|
"learning_rate": 3.467319119189095e-06, |
|
"loss": 9.7465, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.04195217452104601, |
|
"grad_norm": 8.039146423339844, |
|
"learning_rate": 4.166375393219155e-06, |
|
"loss": 8.448, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.04894420360788701, |
|
"grad_norm": 10.288451194763184, |
|
"learning_rate": 4.865431667249214e-06, |
|
"loss": 8.1951, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.05593623269472801, |
|
"grad_norm": 6.3236308097839355, |
|
"learning_rate": 5.564487941279273e-06, |
|
"loss": 8.1093, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.06292826178156902, |
|
"grad_norm": 15.230368614196777, |
|
"learning_rate": 6.263544215309333e-06, |
|
"loss": 8.0567, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.06992029086841, |
|
"grad_norm": 5.573168754577637, |
|
"learning_rate": 6.962600489339393e-06, |
|
"loss": 8.0401, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.07691231995525101, |
|
"grad_norm": 6.7740325927734375, |
|
"learning_rate": 7.661656763369451e-06, |
|
"loss": 7.9491, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.08390434904209201, |
|
"grad_norm": 6.167610168457031, |
|
"learning_rate": 8.360713037399512e-06, |
|
"loss": 7.9494, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.09089637812893302, |
|
"grad_norm": 5.53561544418335, |
|
"learning_rate": 9.05976931142957e-06, |
|
"loss": 7.9386, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.09788840721577402, |
|
"grad_norm": 6.056953430175781, |
|
"learning_rate": 9.758825585459631e-06, |
|
"loss": 7.9033, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.10488043630261502, |
|
"grad_norm": 5.160706043243408, |
|
"learning_rate": 1.045788185948969e-05, |
|
"loss": 7.9055, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.11187246538945601, |
|
"grad_norm": 5.525616645812988, |
|
"learning_rate": 1.115693813351975e-05, |
|
"loss": 7.9203, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.11886449447629702, |
|
"grad_norm": 5.553597927093506, |
|
"learning_rate": 1.185599440754981e-05, |
|
"loss": 7.8381, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.12585652356313803, |
|
"grad_norm": 4.8832478523254395, |
|
"learning_rate": 1.2555050681579868e-05, |
|
"loss": 7.8679, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.13284855264997902, |
|
"grad_norm": 5.308164596557617, |
|
"learning_rate": 1.3254106955609927e-05, |
|
"loss": 7.8686, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.13984058173682, |
|
"grad_norm": 6.790646553039551, |
|
"learning_rate": 1.3953163229639988e-05, |
|
"loss": 7.8252, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.14683261082366103, |
|
"grad_norm": 5.003779888153076, |
|
"learning_rate": 1.4652219503670046e-05, |
|
"loss": 7.856, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.15382463991050202, |
|
"grad_norm": 5.0926642417907715, |
|
"learning_rate": 1.5351275777700107e-05, |
|
"loss": 7.8301, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.16081666899734304, |
|
"grad_norm": 4.154193878173828, |
|
"learning_rate": 1.6050332051730168e-05, |
|
"loss": 7.8595, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.16780869808418403, |
|
"grad_norm": 4.168135166168213, |
|
"learning_rate": 1.6749388325760225e-05, |
|
"loss": 7.8138, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.17480072717102504, |
|
"grad_norm": 4.69076681137085, |
|
"learning_rate": 1.7448444599790285e-05, |
|
"loss": 7.812, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.18179275625786603, |
|
"grad_norm": 4.386991500854492, |
|
"learning_rate": 1.8147500873820346e-05, |
|
"loss": 7.8261, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.18878478534470702, |
|
"grad_norm": 5.777139186859131, |
|
"learning_rate": 1.8846557147850403e-05, |
|
"loss": 7.7988, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.19577681443154804, |
|
"grad_norm": 18.3277645111084, |
|
"learning_rate": 1.9545613421880464e-05, |
|
"loss": 7.7965, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.20276884351838903, |
|
"grad_norm": 4.5356669425964355, |
|
"learning_rate": 1.997280814201919e-05, |
|
"loss": 7.783, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.20976087260523005, |
|
"grad_norm": 5.4392476081848145, |
|
"learning_rate": 1.9895117119216877e-05, |
|
"loss": 7.7752, |
|
"step": 3000 |
|
} |
|
], |
|
"logging_steps": 100, |
|
"max_steps": 28604, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 0.0, |
|
"train_batch_size": 128, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|