|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 20.0, |
|
"global_step": 620, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.0009993582535855264, |
|
"loss": 0.7008, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 0.0009974346616959476, |
|
"loss": 0.4261, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 0.0009942341621640557, |
|
"loss": 0.3999, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 0.0009897649706262473, |
|
"loss": 0.403, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 0.000984038559433102, |
|
"loss": 0.3025, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 0.0009770696282000244, |
|
"loss": 0.3243, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 0.0009688760660735403, |
|
"loss": 0.1971, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 0.0009594789058101153, |
|
"loss": 0.3054, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 0.0009489022697853709, |
|
"loss": 0.3146, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 0.0009371733080722911, |
|
"loss": 0.1914, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"learning_rate": 0.0009243221287473755, |
|
"loss": 0.2848, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 3.87, |
|
"learning_rate": 0.0009103817206036382, |
|
"loss": 0.1316, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 4.19, |
|
"learning_rate": 0.0008953878684688492, |
|
"loss": 0.3139, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 4.52, |
|
"learning_rate": 0.0008793790613463954, |
|
"loss": 0.1033, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 4.84, |
|
"learning_rate": 0.00086239639361456, |
|
"loss": 0.1794, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 5.16, |
|
"learning_rate": 0.0008444834595378434, |
|
"loss": 0.1794, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 5.48, |
|
"learning_rate": 0.0008256862413611113, |
|
"loss": 0.1136, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 5.81, |
|
"learning_rate": 0.0008060529912738315, |
|
"loss": 0.1584, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 6.13, |
|
"learning_rate": 0.0007856341075473961, |
|
"loss": 0.1127, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 6.45, |
|
"learning_rate": 0.0007644820051634812, |
|
"loss": 0.1422, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 6.77, |
|
"learning_rate": 0.0007426509812655407, |
|
"loss": 0.1001, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 7.1, |
|
"learning_rate": 0.0007201970757788173, |
|
"loss": 0.0521, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 7.42, |
|
"learning_rate": 0.0006971779275566593, |
|
"loss": 0.0499, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 7.74, |
|
"learning_rate": 0.0006736526264224101, |
|
"loss": 0.1083, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 8.06, |
|
"learning_rate": 0.0006496815614866791, |
|
"loss": 0.0762, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 8.39, |
|
"learning_rate": 0.0006253262661293602, |
|
"loss": 0.0363, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 8.71, |
|
"learning_rate": 0.0006006492600443301, |
|
"loss": 0.0455, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 9.03, |
|
"learning_rate": 0.0005757138887522884, |
|
"loss": 0.0614, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 9.35, |
|
"learning_rate": 0.0005505841609937161, |
|
"loss": 0.0279, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 9.68, |
|
"learning_rate": 0.0005253245844193564, |
|
"loss": 0.0383, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 0.0005, |
|
"loss": 0.0219, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 10.32, |
|
"learning_rate": 0.0004746754155806437, |
|
"loss": 0.0069, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 10.65, |
|
"learning_rate": 0.000449415839006284, |
|
"loss": 0.0399, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 10.97, |
|
"learning_rate": 0.00042428611124771184, |
|
"loss": 0.0208, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 11.29, |
|
"learning_rate": 0.0003993507399556699, |
|
"loss": 0.0109, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 11.61, |
|
"learning_rate": 0.00037467373387063967, |
|
"loss": 0.0249, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 11.94, |
|
"learning_rate": 0.000350318438513321, |
|
"loss": 0.0041, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 12.26, |
|
"learning_rate": 0.0003263473735775899, |
|
"loss": 0.009, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 12.58, |
|
"learning_rate": 0.00030282207244334083, |
|
"loss": 0.007, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 12.9, |
|
"learning_rate": 0.0002798029242211828, |
|
"loss": 0.0089, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 13.23, |
|
"learning_rate": 0.0002573490187344596, |
|
"loss": 0.0103, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 13.55, |
|
"learning_rate": 0.0002355179948365189, |
|
"loss": 0.0054, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 13.87, |
|
"learning_rate": 0.00021436589245260373, |
|
"loss": 0.0037, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 14.19, |
|
"learning_rate": 0.00019394700872616856, |
|
"loss": 0.0047, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 14.52, |
|
"learning_rate": 0.00017431375863888898, |
|
"loss": 0.0061, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 14.84, |
|
"learning_rate": 0.00015551654046215668, |
|
"loss": 0.0043, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 15.16, |
|
"learning_rate": 0.00013760360638544013, |
|
"loss": 0.0058, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 15.48, |
|
"learning_rate": 0.00012062093865360457, |
|
"loss": 0.0029, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 15.81, |
|
"learning_rate": 0.00010461213153115079, |
|
"loss": 0.0037, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 16.13, |
|
"learning_rate": 8.961827939636197e-05, |
|
"loss": 0.004, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 16.45, |
|
"learning_rate": 7.567787125262449e-05, |
|
"loss": 0.0048, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 16.77, |
|
"learning_rate": 6.282669192770896e-05, |
|
"loss": 0.0036, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 17.1, |
|
"learning_rate": 5.109773021462921e-05, |
|
"loss": 0.0018, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 17.42, |
|
"learning_rate": 4.05210941898847e-05, |
|
"loss": 0.0039, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 17.74, |
|
"learning_rate": 3.112393392645985e-05, |
|
"loss": 0.004, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 18.06, |
|
"learning_rate": 2.2930371799975592e-05, |
|
"loss": 0.0038, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 18.39, |
|
"learning_rate": 1.596144056689791e-05, |
|
"loss": 0.0027, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 18.71, |
|
"learning_rate": 1.0235029373752758e-05, |
|
"loss": 0.0043, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 19.03, |
|
"learning_rate": 5.76583783594431e-06, |
|
"loss": 0.0034, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 19.35, |
|
"learning_rate": 2.5653383040524227e-06, |
|
"loss": 0.0034, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 19.68, |
|
"learning_rate": 6.417464144736207e-07, |
|
"loss": 0.0037, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"learning_rate": 0.0, |
|
"loss": 0.0038, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"step": 620, |
|
"total_flos": 7.220647174835405e+16, |
|
"train_loss": 0.09885127827404969, |
|
"train_runtime": 1180.114, |
|
"train_samples_per_second": 2.085, |
|
"train_steps_per_second": 0.525 |
|
} |
|
], |
|
"max_steps": 620, |
|
"num_train_epochs": 20, |
|
"total_flos": 7.220647174835405e+16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|