AdminBERT-4GB / trainer_state.json
TSebbag's picture
Model files AdinBERT 4 GB
66903c9 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 6.0,
"eval_steps": 500,
"global_step": 94368,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0,
"learning_rate": 0.0,
"loss": 8.4152,
"step": 1
},
{
"epoch": 0.03,
"learning_rate": 4.950000000000001e-06,
"loss": 3.4041,
"step": 500
},
{
"epoch": 0.06,
"learning_rate": 9.950000000000001e-06,
"loss": 1.6651,
"step": 1000
},
{
"epoch": 0.1,
"learning_rate": 1.4950000000000001e-05,
"loss": 1.5062,
"step": 1500
},
{
"epoch": 0.13,
"learning_rate": 1.995e-05,
"loss": 1.4128,
"step": 2000
},
{
"epoch": 0.16,
"learning_rate": 2.495e-05,
"loss": 1.347,
"step": 2500
},
{
"epoch": 0.19,
"learning_rate": 2.995e-05,
"loss": 1.2902,
"step": 3000
},
{
"epoch": 0.22,
"learning_rate": 3.495e-05,
"loss": 1.2457,
"step": 3500
},
{
"epoch": 0.25,
"learning_rate": 3.995e-05,
"loss": 1.2037,
"step": 4000
},
{
"epoch": 0.29,
"learning_rate": 4.495e-05,
"loss": 1.1728,
"step": 4500
},
{
"epoch": 0.32,
"learning_rate": 4.995e-05,
"loss": 1.1493,
"step": 5000
},
{
"epoch": 0.35,
"learning_rate": 5.495e-05,
"loss": 1.1211,
"step": 5500
},
{
"epoch": 0.38,
"learning_rate": 5.995000000000001e-05,
"loss": 1.0958,
"step": 6000
},
{
"epoch": 0.41,
"learning_rate": 6.494999999999999e-05,
"loss": 1.0736,
"step": 6500
},
{
"epoch": 0.45,
"learning_rate": 6.995e-05,
"loss": 1.0591,
"step": 7000
},
{
"epoch": 0.48,
"learning_rate": 7.494e-05,
"loss": 1.0399,
"step": 7500
},
{
"epoch": 0.51,
"learning_rate": 7.994000000000001e-05,
"loss": 1.0244,
"step": 8000
},
{
"epoch": 0.54,
"learning_rate": 8.494000000000001e-05,
"loss": 1.0128,
"step": 8500
},
{
"epoch": 0.57,
"learning_rate": 8.994e-05,
"loss": 1.0026,
"step": 9000
},
{
"epoch": 0.6,
"learning_rate": 9.494000000000001e-05,
"loss": 0.9945,
"step": 9500
},
{
"epoch": 0.64,
"learning_rate": 9.994e-05,
"loss": 0.9798,
"step": 10000
},
{
"epoch": 0.67,
"learning_rate": 9.941446994120994e-05,
"loss": 0.9709,
"step": 10500
},
{
"epoch": 0.7,
"learning_rate": 9.882182818130097e-05,
"loss": 0.9623,
"step": 11000
},
{
"epoch": 0.73,
"learning_rate": 9.8229186421392e-05,
"loss": 0.9505,
"step": 11500
},
{
"epoch": 0.76,
"learning_rate": 9.763654466148303e-05,
"loss": 0.9428,
"step": 12000
},
{
"epoch": 0.79,
"learning_rate": 9.704390290157406e-05,
"loss": 0.9323,
"step": 12500
},
{
"epoch": 0.83,
"learning_rate": 9.645244642518491e-05,
"loss": 0.9265,
"step": 13000
},
{
"epoch": 0.86,
"learning_rate": 9.585980466527594e-05,
"loss": 0.9197,
"step": 13500
},
{
"epoch": 0.89,
"learning_rate": 9.526716290536697e-05,
"loss": 0.9122,
"step": 14000
},
{
"epoch": 0.92,
"learning_rate": 9.4674521145458e-05,
"loss": 0.9065,
"step": 14500
},
{
"epoch": 0.95,
"learning_rate": 9.408187938554904e-05,
"loss": 0.9023,
"step": 15000
},
{
"epoch": 0.99,
"learning_rate": 9.348923762564006e-05,
"loss": 0.8965,
"step": 15500
},
{
"epoch": 1.02,
"learning_rate": 9.289659586573108e-05,
"loss": 0.8901,
"step": 16000
},
{
"epoch": 1.05,
"learning_rate": 9.230395410582213e-05,
"loss": 0.8822,
"step": 16500
},
{
"epoch": 1.08,
"learning_rate": 9.171131234591315e-05,
"loss": 0.8767,
"step": 17000
},
{
"epoch": 1.11,
"learning_rate": 9.111867058600417e-05,
"loss": 0.8711,
"step": 17500
},
{
"epoch": 1.14,
"learning_rate": 9.052721410961502e-05,
"loss": 0.8692,
"step": 18000
},
{
"epoch": 1.18,
"learning_rate": 8.993457234970606e-05,
"loss": 0.8655,
"step": 18500
},
{
"epoch": 1.21,
"learning_rate": 8.934193058979709e-05,
"loss": 0.861,
"step": 19000
},
{
"epoch": 1.24,
"learning_rate": 8.874928882988811e-05,
"loss": 0.8566,
"step": 19500
},
{
"epoch": 1.27,
"learning_rate": 8.815664706997914e-05,
"loss": 0.8509,
"step": 20000
},
{
"epoch": 1.3,
"learning_rate": 8.756519059359e-05,
"loss": 0.8492,
"step": 20500
},
{
"epoch": 1.34,
"learning_rate": 8.697373411720084e-05,
"loss": 0.8475,
"step": 21000
},
{
"epoch": 1.37,
"learning_rate": 8.638109235729187e-05,
"loss": 0.8471,
"step": 21500
},
{
"epoch": 1.4,
"learning_rate": 8.578845059738289e-05,
"loss": 0.8393,
"step": 22000
},
{
"epoch": 1.43,
"learning_rate": 8.519580883747393e-05,
"loss": 0.8341,
"step": 22500
},
{
"epoch": 1.46,
"learning_rate": 8.460316707756496e-05,
"loss": 0.8332,
"step": 23000
},
{
"epoch": 1.49,
"learning_rate": 8.401052531765598e-05,
"loss": 0.8295,
"step": 23500
},
{
"epoch": 1.53,
"learning_rate": 8.341788355774702e-05,
"loss": 0.8295,
"step": 24000
},
{
"epoch": 1.56,
"learning_rate": 8.282524179783805e-05,
"loss": 0.8237,
"step": 24500
},
{
"epoch": 1.59,
"learning_rate": 8.22337853214489e-05,
"loss": 0.8213,
"step": 25000
},
{
"epoch": 1.62,
"learning_rate": 8.164232884505975e-05,
"loss": 0.8187,
"step": 25500
},
{
"epoch": 1.65,
"learning_rate": 8.104968708515077e-05,
"loss": 0.8154,
"step": 26000
},
{
"epoch": 1.68,
"learning_rate": 8.04570453252418e-05,
"loss": 0.8133,
"step": 26500
},
{
"epoch": 1.72,
"learning_rate": 7.986440356533284e-05,
"loss": 0.8064,
"step": 27000
},
{
"epoch": 1.75,
"learning_rate": 7.927176180542386e-05,
"loss": 0.8087,
"step": 27500
},
{
"epoch": 1.78,
"learning_rate": 7.86803053290347e-05,
"loss": 0.809,
"step": 28000
},
{
"epoch": 1.81,
"learning_rate": 7.808766356912574e-05,
"loss": 0.8087,
"step": 28500
},
{
"epoch": 1.84,
"learning_rate": 7.749502180921678e-05,
"loss": 0.8018,
"step": 29000
},
{
"epoch": 1.88,
"learning_rate": 7.690356533282762e-05,
"loss": 0.7998,
"step": 29500
},
{
"epoch": 1.91,
"learning_rate": 7.631092357291864e-05,
"loss": 0.7984,
"step": 30000
},
{
"epoch": 1.94,
"learning_rate": 7.571828181300968e-05,
"loss": 0.7941,
"step": 30500
},
{
"epoch": 1.97,
"learning_rate": 7.512564005310071e-05,
"loss": 0.7965,
"step": 31000
},
{
"epoch": 2.0,
"learning_rate": 7.453299829319173e-05,
"loss": 0.7941,
"step": 31500
},
{
"epoch": 2.03,
"learning_rate": 7.394035653328276e-05,
"loss": 0.7901,
"step": 32000
},
{
"epoch": 2.07,
"learning_rate": 7.33477147733738e-05,
"loss": 0.7855,
"step": 32500
},
{
"epoch": 2.1,
"learning_rate": 7.275507301346483e-05,
"loss": 0.7825,
"step": 33000
},
{
"epoch": 2.13,
"learning_rate": 7.216243125355585e-05,
"loss": 0.7848,
"step": 33500
},
{
"epoch": 2.16,
"learning_rate": 7.156978949364688e-05,
"loss": 0.779,
"step": 34000
},
{
"epoch": 2.19,
"learning_rate": 7.097714773373792e-05,
"loss": 0.7778,
"step": 34500
},
{
"epoch": 2.23,
"learning_rate": 7.038450597382894e-05,
"loss": 0.7789,
"step": 35000
},
{
"epoch": 2.26,
"learning_rate": 6.979186421391997e-05,
"loss": 0.7749,
"step": 35500
},
{
"epoch": 2.29,
"learning_rate": 6.920040773753082e-05,
"loss": 0.7719,
"step": 36000
},
{
"epoch": 2.32,
"learning_rate": 6.860776597762185e-05,
"loss": 0.7755,
"step": 36500
},
{
"epoch": 2.35,
"learning_rate": 6.801512421771288e-05,
"loss": 0.7708,
"step": 37000
},
{
"epoch": 2.38,
"learning_rate": 6.742248245780391e-05,
"loss": 0.7677,
"step": 37500
},
{
"epoch": 2.42,
"learning_rate": 6.682984069789495e-05,
"loss": 0.766,
"step": 38000
},
{
"epoch": 2.45,
"learning_rate": 6.623719893798597e-05,
"loss": 0.767,
"step": 38500
},
{
"epoch": 2.48,
"learning_rate": 6.5644557178077e-05,
"loss": 0.7628,
"step": 39000
},
{
"epoch": 2.51,
"learning_rate": 6.505191541816802e-05,
"loss": 0.7638,
"step": 39500
},
{
"epoch": 2.54,
"learning_rate": 6.445927365825906e-05,
"loss": 0.7631,
"step": 40000
},
{
"epoch": 2.58,
"learning_rate": 6.38666318983501e-05,
"loss": 0.7587,
"step": 40500
},
{
"epoch": 2.61,
"learning_rate": 6.327636070548075e-05,
"loss": 0.7595,
"step": 41000
},
{
"epoch": 2.64,
"learning_rate": 6.268371894557177e-05,
"loss": 0.7599,
"step": 41500
},
{
"epoch": 2.67,
"learning_rate": 6.209107718566282e-05,
"loss": 0.7583,
"step": 42000
},
{
"epoch": 2.7,
"learning_rate": 6.149843542575384e-05,
"loss": 0.7534,
"step": 42500
},
{
"epoch": 2.73,
"learning_rate": 6.090579366584487e-05,
"loss": 0.7564,
"step": 43000
},
{
"epoch": 2.77,
"learning_rate": 6.0313151905935906e-05,
"loss": 0.7527,
"step": 43500
},
{
"epoch": 2.8,
"learning_rate": 5.9720510146026934e-05,
"loss": 0.7526,
"step": 44000
},
{
"epoch": 2.83,
"learning_rate": 5.9127868386117956e-05,
"loss": 0.7519,
"step": 44500
},
{
"epoch": 2.86,
"learning_rate": 5.85364119097288e-05,
"loss": 0.749,
"step": 45000
},
{
"epoch": 2.89,
"learning_rate": 5.7943770149819844e-05,
"loss": 0.7492,
"step": 45500
},
{
"epoch": 2.92,
"learning_rate": 5.7351128389910865e-05,
"loss": 0.7468,
"step": 46000
},
{
"epoch": 2.96,
"learning_rate": 5.6758486630001894e-05,
"loss": 0.745,
"step": 46500
},
{
"epoch": 2.99,
"learning_rate": 5.616584487009293e-05,
"loss": 0.7479,
"step": 47000
},
{
"epoch": 3.02,
"learning_rate": 5.557320311018396e-05,
"loss": 0.7431,
"step": 47500
},
{
"epoch": 3.05,
"learning_rate": 5.4980561350274985e-05,
"loss": 0.7406,
"step": 48000
},
{
"epoch": 3.08,
"learning_rate": 5.438791959036602e-05,
"loss": 0.7401,
"step": 48500
},
{
"epoch": 3.12,
"learning_rate": 5.379527783045705e-05,
"loss": 0.7379,
"step": 49000
},
{
"epoch": 3.15,
"learning_rate": 5.3203821354067895e-05,
"loss": 0.7375,
"step": 49500
},
{
"epoch": 3.18,
"learning_rate": 5.261117959415892e-05,
"loss": 0.7339,
"step": 50000
},
{
"epoch": 3.21,
"learning_rate": 5.201853783424996e-05,
"loss": 0.7349,
"step": 50500
},
{
"epoch": 3.24,
"learning_rate": 5.1425896074340986e-05,
"loss": 0.7348,
"step": 51000
},
{
"epoch": 3.27,
"learning_rate": 5.083443959795183e-05,
"loss": 0.7319,
"step": 51500
},
{
"epoch": 3.31,
"learning_rate": 5.024179783804286e-05,
"loss": 0.7271,
"step": 52000
},
{
"epoch": 3.34,
"learning_rate": 4.964915607813389e-05,
"loss": 0.7309,
"step": 52500
},
{
"epoch": 3.37,
"learning_rate": 4.905769960174474e-05,
"loss": 0.733,
"step": 53000
},
{
"epoch": 3.4,
"learning_rate": 4.846505784183577e-05,
"loss": 0.7277,
"step": 53500
},
{
"epoch": 3.43,
"learning_rate": 4.7872416081926805e-05,
"loss": 0.728,
"step": 54000
},
{
"epoch": 3.47,
"learning_rate": 4.7279774322017827e-05,
"loss": 0.7268,
"step": 54500
},
{
"epoch": 3.5,
"learning_rate": 4.668713256210886e-05,
"loss": 0.7261,
"step": 55000
},
{
"epoch": 3.53,
"learning_rate": 4.609567608571971e-05,
"loss": 0.7251,
"step": 55500
},
{
"epoch": 3.56,
"learning_rate": 4.5503034325810736e-05,
"loss": 0.7261,
"step": 56000
},
{
"epoch": 3.59,
"learning_rate": 4.4910392565901764e-05,
"loss": 0.7231,
"step": 56500
},
{
"epoch": 3.62,
"learning_rate": 4.43177508059928e-05,
"loss": 0.7197,
"step": 57000
},
{
"epoch": 3.66,
"learning_rate": 4.3726294329603646e-05,
"loss": 0.7249,
"step": 57500
},
{
"epoch": 3.69,
"learning_rate": 4.313483785321449e-05,
"loss": 0.7204,
"step": 58000
},
{
"epoch": 3.72,
"learning_rate": 4.254219609330552e-05,
"loss": 0.718,
"step": 58500
},
{
"epoch": 3.75,
"learning_rate": 4.194955433339655e-05,
"loss": 0.7205,
"step": 59000
},
{
"epoch": 3.78,
"learning_rate": 4.1356912573487584e-05,
"loss": 0.7164,
"step": 59500
},
{
"epoch": 3.81,
"learning_rate": 4.076427081357861e-05,
"loss": 0.7194,
"step": 60000
},
{
"epoch": 3.85,
"learning_rate": 4.017162905366964e-05,
"loss": 0.7166,
"step": 60500
},
{
"epoch": 3.88,
"learning_rate": 3.9578987293760675e-05,
"loss": 0.7167,
"step": 61000
},
{
"epoch": 3.91,
"learning_rate": 3.8986345533851696e-05,
"loss": 0.7175,
"step": 61500
},
{
"epoch": 3.94,
"learning_rate": 3.839370377394273e-05,
"loss": 0.7104,
"step": 62000
},
{
"epoch": 3.97,
"learning_rate": 3.780106201403376e-05,
"loss": 0.7116,
"step": 62500
},
{
"epoch": 4.01,
"learning_rate": 3.720842025412479e-05,
"loss": 0.7125,
"step": 63000
},
{
"epoch": 4.04,
"learning_rate": 3.6615778494215816e-05,
"loss": 0.7098,
"step": 63500
},
{
"epoch": 4.07,
"learning_rate": 3.602432201782667e-05,
"loss": 0.708,
"step": 64000
},
{
"epoch": 4.1,
"learning_rate": 3.543168025791769e-05,
"loss": 0.7081,
"step": 64500
},
{
"epoch": 4.13,
"learning_rate": 3.4839038498008725e-05,
"loss": 0.7058,
"step": 65000
},
{
"epoch": 4.16,
"learning_rate": 3.4246396738099754e-05,
"loss": 0.7072,
"step": 65500
},
{
"epoch": 4.2,
"learning_rate": 3.365375497819078e-05,
"loss": 0.7068,
"step": 66000
},
{
"epoch": 4.23,
"learning_rate": 3.306111321828182e-05,
"loss": 0.706,
"step": 66500
},
{
"epoch": 4.26,
"learning_rate": 3.246965674189266e-05,
"loss": 0.705,
"step": 67000
},
{
"epoch": 4.29,
"learning_rate": 3.187701498198369e-05,
"loss": 0.7043,
"step": 67500
},
{
"epoch": 4.32,
"learning_rate": 3.128437322207472e-05,
"loss": 0.7025,
"step": 68000
},
{
"epoch": 4.36,
"learning_rate": 3.0691731462165755e-05,
"loss": 0.7019,
"step": 68500
},
{
"epoch": 4.39,
"learning_rate": 3.0099089702256783e-05,
"loss": 0.703,
"step": 69000
},
{
"epoch": 4.42,
"learning_rate": 2.9506447942347808e-05,
"loss": 0.7023,
"step": 69500
},
{
"epoch": 4.45,
"learning_rate": 2.891380618243884e-05,
"loss": 0.7023,
"step": 70000
},
{
"epoch": 4.48,
"learning_rate": 2.832234970604969e-05,
"loss": 0.7,
"step": 70500
},
{
"epoch": 4.51,
"learning_rate": 2.772970794614072e-05,
"loss": 0.6976,
"step": 71000
},
{
"epoch": 4.55,
"learning_rate": 2.7137066186231745e-05,
"loss": 0.7009,
"step": 71500
},
{
"epoch": 4.58,
"learning_rate": 2.6544424426322777e-05,
"loss": 0.6985,
"step": 72000
},
{
"epoch": 4.61,
"learning_rate": 2.595178266641381e-05,
"loss": 0.6976,
"step": 72500
},
{
"epoch": 4.64,
"learning_rate": 2.5359140906504837e-05,
"loss": 0.6988,
"step": 73000
},
{
"epoch": 4.67,
"learning_rate": 2.4766499146595868e-05,
"loss": 0.6988,
"step": 73500
},
{
"epoch": 4.7,
"learning_rate": 2.4173857386686896e-05,
"loss": 0.6964,
"step": 74000
},
{
"epoch": 4.74,
"learning_rate": 2.3582400910297743e-05,
"loss": 0.6972,
"step": 74500
},
{
"epoch": 4.77,
"learning_rate": 2.2989759150388775e-05,
"loss": 0.6948,
"step": 75000
},
{
"epoch": 4.8,
"learning_rate": 2.2397117390479806e-05,
"loss": 0.6946,
"step": 75500
},
{
"epoch": 4.83,
"learning_rate": 2.1804475630570834e-05,
"loss": 0.6927,
"step": 76000
},
{
"epoch": 4.86,
"learning_rate": 2.1211833870661862e-05,
"loss": 0.6909,
"step": 76500
},
{
"epoch": 4.9,
"learning_rate": 2.0620377394272712e-05,
"loss": 0.6906,
"step": 77000
},
{
"epoch": 4.93,
"learning_rate": 2.002892091788356e-05,
"loss": 0.6929,
"step": 77500
},
{
"epoch": 4.96,
"learning_rate": 1.9436279157974587e-05,
"loss": 0.6917,
"step": 78000
},
{
"epoch": 4.99,
"learning_rate": 1.884363739806562e-05,
"loss": 0.6925,
"step": 78500
},
{
"epoch": 5.02,
"learning_rate": 1.8250995638156647e-05,
"loss": 0.6895,
"step": 79000
},
{
"epoch": 5.05,
"learning_rate": 1.765835387824768e-05,
"loss": 0.686,
"step": 79500
},
{
"epoch": 5.09,
"learning_rate": 1.706571211833871e-05,
"loss": 0.6882,
"step": 80000
},
{
"epoch": 5.12,
"learning_rate": 1.6473070358429738e-05,
"loss": 0.6877,
"step": 80500
},
{
"epoch": 5.15,
"learning_rate": 1.5880428598520766e-05,
"loss": 0.688,
"step": 81000
},
{
"epoch": 5.18,
"learning_rate": 1.5287786838611798e-05,
"loss": 0.6863,
"step": 81500
},
{
"epoch": 5.21,
"learning_rate": 1.4696330362222643e-05,
"loss": 0.6882,
"step": 82000
},
{
"epoch": 5.25,
"learning_rate": 1.4104873885833491e-05,
"loss": 0.6892,
"step": 82500
},
{
"epoch": 5.28,
"learning_rate": 1.351223212592452e-05,
"loss": 0.6878,
"step": 83000
},
{
"epoch": 5.31,
"learning_rate": 1.2919590366015552e-05,
"loss": 0.6841,
"step": 83500
},
{
"epoch": 5.34,
"learning_rate": 1.232694860610658e-05,
"loss": 0.6841,
"step": 84000
},
{
"epoch": 5.37,
"learning_rate": 1.173430684619761e-05,
"loss": 0.6855,
"step": 84500
},
{
"epoch": 5.4,
"learning_rate": 1.1141665086288642e-05,
"loss": 0.6833,
"step": 85000
},
{
"epoch": 5.44,
"learning_rate": 1.054902332637967e-05,
"loss": 0.683,
"step": 85500
},
{
"epoch": 5.47,
"learning_rate": 9.9563815664707e-06,
"loss": 0.6847,
"step": 86000
},
{
"epoch": 5.5,
"learning_rate": 9.36373980656173e-06,
"loss": 0.6845,
"step": 86500
},
{
"epoch": 5.53,
"learning_rate": 8.77109804665276e-06,
"loss": 0.6795,
"step": 87000
},
{
"epoch": 5.56,
"learning_rate": 8.17845628674379e-06,
"loss": 0.6817,
"step": 87500
},
{
"epoch": 5.6,
"learning_rate": 7.585814526834819e-06,
"loss": 0.6796,
"step": 88000
},
{
"epoch": 5.63,
"learning_rate": 6.994358050445667e-06,
"loss": 0.684,
"step": 88500
},
{
"epoch": 5.66,
"learning_rate": 6.401716290536697e-06,
"loss": 0.6833,
"step": 89000
},
{
"epoch": 5.69,
"learning_rate": 5.810259814147544e-06,
"loss": 0.6805,
"step": 89500
},
{
"epoch": 5.72,
"learning_rate": 5.217618054238574e-06,
"loss": 0.6823,
"step": 90000
},
{
"epoch": 5.75,
"learning_rate": 4.626161577849421e-06,
"loss": 0.6786,
"step": 90500
},
{
"epoch": 5.79,
"learning_rate": 4.033519817940452e-06,
"loss": 0.676,
"step": 91000
},
{
"epoch": 5.82,
"learning_rate": 3.440878058031481e-06,
"loss": 0.6811,
"step": 91500
},
{
"epoch": 5.85,
"learning_rate": 2.848236298122511e-06,
"loss": 0.6805,
"step": 92000
},
{
"epoch": 5.88,
"learning_rate": 2.2567798217333587e-06,
"loss": 0.6805,
"step": 92500
},
{
"epoch": 5.91,
"learning_rate": 1.6641380618243886e-06,
"loss": 0.6783,
"step": 93000
},
{
"epoch": 5.94,
"learning_rate": 1.0714963019154182e-06,
"loss": 0.6806,
"step": 93500
},
{
"epoch": 5.98,
"learning_rate": 4.78854542006448e-07,
"loss": 0.6792,
"step": 94000
},
{
"before_init_mem_cpu": 1314717696,
"before_init_mem_gpu": 0,
"epoch": 6.0,
"init_mem_cpu_alloc_delta": 880553984,
"init_mem_cpu_peaked_delta": 93917184,
"init_mem_gpu_alloc_delta": 443396608,
"init_mem_gpu_peaked_delta": 0,
"step": 94368,
"total_flos": 1.9076000524147884e+19,
"train_loss": 0.8106447110522921,
"train_mem_cpu_alloc_delta": 31978672128,
"train_mem_cpu_peaked_delta": 500772864,
"train_mem_gpu_alloc_delta": 1348273664,
"train_mem_gpu_peaked_delta": 64082841600,
"train_runtime": 71211.051,
"train_samples_per_second": 1017.691,
"train_steps_per_second": 1.325
}
],
"logging_steps": 500,
"max_steps": 94368,
"num_train_epochs": 6,
"save_steps": 500,
"total_flos": 1.9076000524147884e+19,
"trial_name": null,
"trial_params": null
}