gpt-2-small-testing / checkpoint-100 /trainer_state.json
svwingerden's picture
Duplicate from georgeyw/gpt-2-small-testing
5326b39 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.09995002498750624,
"eval_steps": 500,
"global_step": 100,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0,
"grad_norm": 3.3340563149001086,
"learning_rate": 0.0,
"loss": 11.0,
"step": 1
},
{
"epoch": 0.0,
"grad_norm": 2.398812329952019,
"learning_rate": 5.9999999999999995e-05,
"loss": 10.125,
"step": 2
},
{
"epoch": 0.0,
"grad_norm": 2.394322446895115,
"learning_rate": 0.00011999999999999999,
"loss": 10.1172,
"step": 3
},
{
"epoch": 0.0,
"grad_norm": 1.9958816684399585,
"learning_rate": 0.00017999999999999998,
"loss": 9.875,
"step": 4
},
{
"epoch": 0.0,
"grad_norm": 1.8270465897882062,
"learning_rate": 0.00023999999999999998,
"loss": 9.6641,
"step": 5
},
{
"epoch": 0.01,
"grad_norm": 1.7854046471397795,
"learning_rate": 0.0003,
"loss": 9.4844,
"step": 6
},
{
"epoch": 0.01,
"grad_norm": 1.719416749115252,
"learning_rate": 0.00035999999999999997,
"loss": 9.3281,
"step": 7
},
{
"epoch": 0.01,
"grad_norm": 1.4637825746112274,
"learning_rate": 0.00041999999999999996,
"loss": 9.2109,
"step": 8
},
{
"epoch": 0.01,
"grad_norm": 1.4393631015406718,
"learning_rate": 0.00047999999999999996,
"loss": 8.9453,
"step": 9
},
{
"epoch": 0.01,
"grad_norm": 1.2936734586915988,
"learning_rate": 0.00054,
"loss": 8.7109,
"step": 10
},
{
"epoch": 0.01,
"grad_norm": 1.0756922378227356,
"learning_rate": 0.0005999986405514987,
"loss": 8.4609,
"step": 11
},
{
"epoch": 0.01,
"grad_norm": 0.9277829127413892,
"learning_rate": 0.0005999945622196846,
"loss": 8.2344,
"step": 12
},
{
"epoch": 0.01,
"grad_norm": 0.8084581786682467,
"learning_rate": 0.0005999877650456265,
"loss": 8.125,
"step": 13
},
{
"epoch": 0.01,
"grad_norm": 0.7635084596900947,
"learning_rate": 0.000599978249097772,
"loss": 7.9766,
"step": 14
},
{
"epoch": 0.01,
"grad_norm": 0.9186699644247788,
"learning_rate": 0.0005999660144719463,
"loss": 7.8555,
"step": 15
},
{
"epoch": 0.02,
"grad_norm": 0.6609504256551479,
"learning_rate": 0.0005999510612913519,
"loss": 7.7734,
"step": 16
},
{
"epoch": 0.02,
"grad_norm": 0.7086232844782971,
"learning_rate": 0.0005999333897065673,
"loss": 7.7148,
"step": 17
},
{
"epoch": 0.02,
"grad_norm": 16.38048851691348,
"learning_rate": 0.0005999129998955453,
"loss": 8.4844,
"step": 18
},
{
"epoch": 0.02,
"grad_norm": 1.3057527590449889,
"learning_rate": 0.0005998898920636111,
"loss": 7.7539,
"step": 19
},
{
"epoch": 0.02,
"grad_norm": 0.6966048242948986,
"learning_rate": 0.00059986406644346,
"loss": 7.75,
"step": 20
},
{
"epoch": 0.02,
"grad_norm": 0.6348089115348993,
"learning_rate": 0.0005998355232951559,
"loss": 7.7031,
"step": 21
},
{
"epoch": 0.02,
"grad_norm": 0.7829163518610293,
"learning_rate": 0.0005998042629061279,
"loss": 7.6992,
"step": 22
},
{
"epoch": 0.02,
"grad_norm": 0.5900591778980369,
"learning_rate": 0.0005997702855911678,
"loss": 7.6016,
"step": 23
},
{
"epoch": 0.02,
"grad_norm": 0.4655170213064256,
"learning_rate": 0.0005997335916924268,
"loss": 7.5977,
"step": 24
},
{
"epoch": 0.02,
"grad_norm": 0.6287348258915756,
"learning_rate": 0.0005996941815794121,
"loss": 7.5586,
"step": 25
},
{
"epoch": 0.03,
"grad_norm": 0.6137321903884564,
"learning_rate": 0.0005996520556489831,
"loss": 7.5898,
"step": 26
},
{
"epoch": 0.03,
"grad_norm": 0.44962562710631065,
"learning_rate": 0.0005996072143253473,
"loss": 7.4336,
"step": 27
},
{
"epoch": 0.03,
"grad_norm": 0.46130046454703316,
"learning_rate": 0.0005995596580600566,
"loss": 7.4023,
"step": 28
},
{
"epoch": 0.03,
"grad_norm": 0.4686712675731326,
"learning_rate": 0.0005995093873320018,
"loss": 7.3789,
"step": 29
},
{
"epoch": 0.03,
"grad_norm": 0.4672147564288997,
"learning_rate": 0.0005994564026474087,
"loss": 7.3711,
"step": 30
},
{
"epoch": 0.03,
"grad_norm": 0.40408354581233474,
"learning_rate": 0.0005994007045398324,
"loss": 7.3672,
"step": 31
},
{
"epoch": 0.03,
"grad_norm": 0.46032146732584733,
"learning_rate": 0.0005993422935701524,
"loss": 7.3477,
"step": 32
},
{
"epoch": 0.03,
"grad_norm": 0.4765534634593268,
"learning_rate": 0.0005992811703265664,
"loss": 7.3555,
"step": 33
},
{
"epoch": 0.03,
"grad_norm": 0.46208489386235113,
"learning_rate": 0.0005992173354245849,
"loss": 7.3047,
"step": 34
},
{
"epoch": 0.03,
"grad_norm": 0.2956144524964961,
"learning_rate": 0.0005991507895070244,
"loss": 7.3125,
"step": 35
},
{
"epoch": 0.04,
"grad_norm": 0.4834645389868856,
"learning_rate": 0.0005990815332440017,
"loss": 7.207,
"step": 36
},
{
"epoch": 0.04,
"grad_norm": 0.4411831350968505,
"learning_rate": 0.0005990095673329266,
"loss": 7.1758,
"step": 37
},
{
"epoch": 0.04,
"grad_norm": 0.24809297748968667,
"learning_rate": 0.0005989348924984951,
"loss": 7.2188,
"step": 38
},
{
"epoch": 0.04,
"grad_norm": 0.39402988416840584,
"learning_rate": 0.0005988575094926817,
"loss": 7.1953,
"step": 39
},
{
"epoch": 0.04,
"grad_norm": 0.3868345222189167,
"learning_rate": 0.0005987774190947328,
"loss": 7.1641,
"step": 40
},
{
"epoch": 0.04,
"grad_norm": 0.3777261230135448,
"learning_rate": 0.0005986946221111575,
"loss": 7.1328,
"step": 41
},
{
"epoch": 0.04,
"grad_norm": 0.4687511444077827,
"learning_rate": 0.0005986091193757206,
"loss": 7.0898,
"step": 42
},
{
"epoch": 0.04,
"grad_norm": 0.34935796211612463,
"learning_rate": 0.0005985209117494337,
"loss": 7.1367,
"step": 43
},
{
"epoch": 0.04,
"grad_norm": 0.38764476686849886,
"learning_rate": 0.0005984300001205466,
"loss": 7.125,
"step": 44
},
{
"epoch": 0.04,
"grad_norm": 0.3956487898882936,
"learning_rate": 0.0005983363854045386,
"loss": 7.1094,
"step": 45
},
{
"epoch": 0.05,
"grad_norm": 0.31140257544677513,
"learning_rate": 0.0005982400685441084,
"loss": 7.0898,
"step": 46
},
{
"epoch": 0.05,
"grad_norm": 0.3664476570531787,
"learning_rate": 0.0005981410505091662,
"loss": 7.0664,
"step": 47
},
{
"epoch": 0.05,
"grad_norm": 0.31891741142945207,
"learning_rate": 0.0005980393322968223,
"loss": 7.0273,
"step": 48
},
{
"epoch": 0.05,
"grad_norm": 0.4533529037337155,
"learning_rate": 0.0005979349149313778,
"loss": 7.0586,
"step": 49
},
{
"epoch": 0.05,
"grad_norm": 0.30532331638835586,
"learning_rate": 0.0005978277994643147,
"loss": 7.0195,
"step": 50
},
{
"epoch": 0.05,
"grad_norm": 0.6501991746260075,
"learning_rate": 0.0005977179869742844,
"loss": 6.9648,
"step": 51
},
{
"epoch": 0.05,
"grad_norm": 0.43904455901717926,
"learning_rate": 0.0005976054785670975,
"loss": 6.9805,
"step": 52
},
{
"epoch": 0.05,
"grad_norm": 0.4826001598483571,
"learning_rate": 0.0005974902753757124,
"loss": 6.9297,
"step": 53
},
{
"epoch": 0.05,
"grad_norm": 0.2924998027034648,
"learning_rate": 0.000597372378560224,
"loss": 6.8984,
"step": 54
},
{
"epoch": 0.05,
"grad_norm": 0.4439033666380787,
"learning_rate": 0.0005972517893078517,
"loss": 6.8945,
"step": 55
},
{
"epoch": 0.06,
"grad_norm": 0.6135914255073411,
"learning_rate": 0.0005971285088329284,
"loss": 6.9727,
"step": 56
},
{
"epoch": 0.06,
"grad_norm": 0.5575686565598483,
"learning_rate": 0.0005970025383768866,
"loss": 6.9219,
"step": 57
},
{
"epoch": 0.06,
"grad_norm": 0.4820951675994578,
"learning_rate": 0.0005968738792082478,
"loss": 6.8516,
"step": 58
},
{
"epoch": 0.06,
"grad_norm": 0.40164190019465584,
"learning_rate": 0.0005967425326226082,
"loss": 6.7734,
"step": 59
},
{
"epoch": 0.06,
"grad_norm": 0.46129863945181293,
"learning_rate": 0.0005966084999426265,
"loss": 6.8125,
"step": 60
},
{
"epoch": 0.06,
"grad_norm": 0.33322355827118677,
"learning_rate": 0.0005964717825180101,
"loss": 6.7891,
"step": 61
},
{
"epoch": 0.06,
"grad_norm": 0.3847525153855558,
"learning_rate": 0.0005963323817255024,
"loss": 6.8242,
"step": 62
},
{
"epoch": 0.06,
"grad_norm": 0.3384433591375982,
"learning_rate": 0.0005961902989688674,
"loss": 6.707,
"step": 63
},
{
"epoch": 0.06,
"grad_norm": 0.3937003195165685,
"learning_rate": 0.000596045535678877,
"loss": 6.8203,
"step": 64
},
{
"epoch": 0.06,
"grad_norm": 0.35423488053528107,
"learning_rate": 0.0005958980933132962,
"loss": 6.7383,
"step": 65
},
{
"epoch": 0.07,
"grad_norm": 0.36005939745315396,
"learning_rate": 0.0005957479733568675,
"loss": 6.7109,
"step": 66
},
{
"epoch": 0.07,
"grad_norm": 0.3499278317706933,
"learning_rate": 0.0005955951773212976,
"loss": 6.7266,
"step": 67
},
{
"epoch": 0.07,
"grad_norm": 0.3708385192137018,
"learning_rate": 0.0005954397067452407,
"loss": 6.7617,
"step": 68
},
{
"epoch": 0.07,
"grad_norm": 0.3775657656205869,
"learning_rate": 0.0005952815631942839,
"loss": 6.7148,
"step": 69
},
{
"epoch": 0.07,
"grad_norm": 0.3040083750375816,
"learning_rate": 0.0005951207482609307,
"loss": 6.5938,
"step": 70
},
{
"epoch": 0.07,
"grad_norm": 0.3443020808841468,
"learning_rate": 0.0005949572635645861,
"loss": 6.6523,
"step": 71
},
{
"epoch": 0.07,
"grad_norm": 0.3520066316939,
"learning_rate": 0.0005947911107515389,
"loss": 6.6211,
"step": 72
},
{
"epoch": 0.07,
"grad_norm": 0.3739040572679613,
"learning_rate": 0.0005946222914949462,
"loss": 6.5547,
"step": 73
},
{
"epoch": 0.07,
"grad_norm": 0.34890731989025553,
"learning_rate": 0.000594450807494816,
"loss": 6.5859,
"step": 74
},
{
"epoch": 0.07,
"grad_norm": 0.40910932350136514,
"learning_rate": 0.0005942766604779903,
"loss": 6.5547,
"step": 75
},
{
"epoch": 0.08,
"grad_norm": 0.5698342865852906,
"learning_rate": 0.0005940998521981274,
"loss": 6.457,
"step": 76
},
{
"epoch": 0.08,
"grad_norm": 0.5179452709555474,
"learning_rate": 0.0005939203844356852,
"loss": 6.5547,
"step": 77
},
{
"epoch": 0.08,
"grad_norm": 0.5222512938673792,
"learning_rate": 0.0005937382589979016,
"loss": 6.5039,
"step": 78
},
{
"epoch": 0.08,
"grad_norm": 0.5682332793686307,
"learning_rate": 0.0005935534777187781,
"loss": 6.5547,
"step": 79
},
{
"epoch": 0.08,
"grad_norm": 0.3869287710460676,
"learning_rate": 0.0005933660424590598,
"loss": 6.5156,
"step": 80
},
{
"epoch": 0.08,
"grad_norm": 0.3078211032807607,
"learning_rate": 0.000593175955106218,
"loss": 6.4258,
"step": 81
},
{
"epoch": 0.08,
"grad_norm": 0.3611357511872241,
"learning_rate": 0.00059298321757443,
"loss": 6.4727,
"step": 82
},
{
"epoch": 0.08,
"grad_norm": 0.29633467844266953,
"learning_rate": 0.0005927878318045608,
"loss": 6.3281,
"step": 83
},
{
"epoch": 0.08,
"grad_norm": 0.3257574200776832,
"learning_rate": 0.0005925897997641426,
"loss": 6.3203,
"step": 84
},
{
"epoch": 0.08,
"grad_norm": 0.2824054533852328,
"learning_rate": 0.0005923891234473562,
"loss": 6.4062,
"step": 85
},
{
"epoch": 0.09,
"grad_norm": 0.3056199770204573,
"learning_rate": 0.0005921858048750097,
"loss": 6.3984,
"step": 86
},
{
"epoch": 0.09,
"grad_norm": 0.2966438824341908,
"learning_rate": 0.000591979846094519,
"loss": 6.3555,
"step": 87
},
{
"epoch": 0.09,
"grad_norm": 0.32782438676663733,
"learning_rate": 0.0005917712491798866,
"loss": 6.4023,
"step": 88
},
{
"epoch": 0.09,
"grad_norm": 0.3538316399620157,
"learning_rate": 0.0005915600162316811,
"loss": 6.2812,
"step": 89
},
{
"epoch": 0.09,
"grad_norm": 0.375858298192913,
"learning_rate": 0.0005913461493770162,
"loss": 6.3086,
"step": 90
},
{
"epoch": 0.09,
"grad_norm": 0.5189251339815161,
"learning_rate": 0.0005911296507695284,
"loss": 6.2812,
"step": 91
},
{
"epoch": 0.09,
"grad_norm": 0.6304909542669104,
"learning_rate": 0.0005909105225893564,
"loss": 6.2969,
"step": 92
},
{
"epoch": 0.09,
"grad_norm": 0.4655662819622591,
"learning_rate": 0.0005906887670431187,
"loss": 6.1953,
"step": 93
},
{
"epoch": 0.09,
"grad_norm": 0.39035390983920965,
"learning_rate": 0.000590464386363891,
"loss": 6.2617,
"step": 94
},
{
"epoch": 0.09,
"grad_norm": 0.4918417851770978,
"learning_rate": 0.0005902373828111843,
"loss": 6.2148,
"step": 95
},
{
"epoch": 0.1,
"grad_norm": 0.35670770889552555,
"learning_rate": 0.0005900077586709219,
"loss": 6.2461,
"step": 96
},
{
"epoch": 0.1,
"grad_norm": 0.4177985869939347,
"learning_rate": 0.0005897755162554163,
"loss": 6.1797,
"step": 97
},
{
"epoch": 0.1,
"grad_norm": 0.3742471130708234,
"learning_rate": 0.000589540657903346,
"loss": 6.1406,
"step": 98
},
{
"epoch": 0.1,
"grad_norm": 0.28627666723978284,
"learning_rate": 0.0005893031859797322,
"loss": 6.2031,
"step": 99
},
{
"epoch": 0.1,
"grad_norm": 0.32238563846046103,
"learning_rate": 0.0005890631028759143,
"loss": 6.0625,
"step": 100
}
],
"logging_steps": 1,
"max_steps": 1000,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 100,
"total_flos": 0.0,
"train_batch_size": 32,
"trial_name": null,
"trial_params": null
}