|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 3.0, |
|
"global_step": 18927, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 4.867913562635389e-05, |
|
"loss": 1.2978, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 4.735827125270778e-05, |
|
"loss": 1.2005, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 4.603740687906166e-05, |
|
"loss": 1.1692, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 4.471654250541555e-05, |
|
"loss": 1.1418, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 4.339567813176943e-05, |
|
"loss": 1.1284, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.2074813758123315e-05, |
|
"loss": 1.105, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 4.07539493844772e-05, |
|
"loss": 1.1058, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 3.943308501083109e-05, |
|
"loss": 1.0819, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 3.811222063718498e-05, |
|
"loss": 1.0774, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 3.6793997992286154e-05, |
|
"loss": 1.0645, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 3.547313361864004e-05, |
|
"loss": 1.059, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 3.415226924499392e-05, |
|
"loss": 1.0522, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 3.283140487134781e-05, |
|
"loss": 1.0476, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 3.15105404977017e-05, |
|
"loss": 1.0402, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 3.0189676124055582e-05, |
|
"loss": 1.0404, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 2.886881175040947e-05, |
|
"loss": 1.0339, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 2.7547947376763355e-05, |
|
"loss": 1.0283, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 2.6229724731864537e-05, |
|
"loss": 1.0299, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 2.4908860358218418e-05, |
|
"loss": 1.019, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 2.3587995984572305e-05, |
|
"loss": 1.0144, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 2.2269773339673484e-05, |
|
"loss": 1.0099, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 2.094890896602737e-05, |
|
"loss": 1.0064, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 1.9628044592381256e-05, |
|
"loss": 0.9982, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 1.830718021873514e-05, |
|
"loss": 0.9962, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 1.6986315845089025e-05, |
|
"loss": 1.0024, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 1.5665451471442913e-05, |
|
"loss": 1.0024, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 1.4344587097796799e-05, |
|
"loss": 0.9936, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 1.3026364452897977e-05, |
|
"loss": 0.9988, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 1.1705500079251864e-05, |
|
"loss": 0.9902, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 1.0384635705605748e-05, |
|
"loss": 0.9885, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 9.066413060706927e-06, |
|
"loss": 0.9836, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 7.745548687060814e-06, |
|
"loss": 0.9866, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 6.424684313414699e-06, |
|
"loss": 0.9871, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 5.103819939768585e-06, |
|
"loss": 0.992, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 3.782955566122471e-06, |
|
"loss": 0.9804, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 2.462091192476357e-06, |
|
"loss": 0.9887, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 1.1412268188302427e-06, |
|
"loss": 0.9785, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"step": 18927, |
|
"total_flos": 9.490460801340211e+16, |
|
"train_loss": 1.0424538015431204, |
|
"train_runtime": 40144.3005, |
|
"train_samples_per_second": 30.173, |
|
"train_steps_per_second": 0.471 |
|
} |
|
], |
|
"max_steps": 18927, |
|
"num_train_epochs": 3, |
|
"total_flos": 9.490460801340211e+16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|