imdatta0's picture
End of training
8b1b653 verified
raw
history blame
27.8 kB
{
"best_metric": 2.2854506969451904,
"best_model_checkpoint": "/home/datta0/models/lora_final/Mistral-7B-v0.3_pct_ortho/checkpoint-8",
"epoch": 0.9980657640232108,
"eval_steps": 8,
"global_step": 387,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0025789813023855577,
"grad_norm": 10.031487464904785,
"learning_rate": 3.75e-05,
"loss": 2.128,
"step": 1
},
{
"epoch": 0.010315925209542231,
"grad_norm": 7.364378929138184,
"learning_rate": 0.00015,
"loss": 2.0612,
"step": 4
},
{
"epoch": 0.020631850419084462,
"grad_norm": 23.759828567504883,
"learning_rate": 0.0003,
"loss": 2.1243,
"step": 8
},
{
"epoch": 0.020631850419084462,
"eval_loss": 2.2854506969451904,
"eval_runtime": 11.1093,
"eval_samples_per_second": 22.054,
"eval_steps_per_second": 2.79,
"step": 8
},
{
"epoch": 0.030947775628626693,
"grad_norm": 133.17942810058594,
"learning_rate": 0.00029991755529206284,
"loss": 2.465,
"step": 12
},
{
"epoch": 0.041263700838168924,
"grad_norm": 254.61338806152344,
"learning_rate": 0.0002996703117966496,
"loss": 9.1803,
"step": 16
},
{
"epoch": 0.041263700838168924,
"eval_loss": 12.813275337219238,
"eval_runtime": 10.7499,
"eval_samples_per_second": 22.791,
"eval_steps_per_second": 2.884,
"step": 16
},
{
"epoch": 0.05157962604771115,
"grad_norm": 87.40137481689453,
"learning_rate": 0.00029925854129933066,
"loss": 12.8875,
"step": 20
},
{
"epoch": 0.061895551257253385,
"grad_norm": 22.197757720947266,
"learning_rate": 0.0002986826964440844,
"loss": 9.137,
"step": 24
},
{
"epoch": 0.061895551257253385,
"eval_loss": 8.456737518310547,
"eval_runtime": 10.6838,
"eval_samples_per_second": 22.932,
"eval_steps_per_second": 2.902,
"step": 24
},
{
"epoch": 0.07221147646679561,
"grad_norm": 11.230401992797852,
"learning_rate": 0.00029794341023572295,
"loss": 7.9575,
"step": 28
},
{
"epoch": 0.08252740167633785,
"grad_norm": 19.19782066345215,
"learning_rate": 0.0002970414953440533,
"loss": 8.3645,
"step": 32
},
{
"epoch": 0.08252740167633785,
"eval_loss": 8.26063346862793,
"eval_runtime": 10.699,
"eval_samples_per_second": 22.899,
"eval_steps_per_second": 2.897,
"step": 32
},
{
"epoch": 0.09284332688588008,
"grad_norm": 13.250898361206055,
"learning_rate": 0.00029597794321054006,
"loss": 8.0337,
"step": 36
},
{
"epoch": 0.1031592520954223,
"grad_norm": 15.708863258361816,
"learning_rate": 0.00029475392295845,
"loss": 8.5251,
"step": 40
},
{
"epoch": 0.1031592520954223,
"eval_loss": 7.756398677825928,
"eval_runtime": 10.6841,
"eval_samples_per_second": 22.931,
"eval_steps_per_second": 2.901,
"step": 40
},
{
"epoch": 0.11347517730496454,
"grad_norm": 8.025888442993164,
"learning_rate": 0.0002933707801076791,
"loss": 7.708,
"step": 44
},
{
"epoch": 0.12379110251450677,
"grad_norm": 72.3178482055664,
"learning_rate": 0.00029183003509567217,
"loss": 9.4881,
"step": 48
},
{
"epoch": 0.12379110251450677,
"eval_loss": 9.306961059570312,
"eval_runtime": 10.6939,
"eval_samples_per_second": 22.91,
"eval_steps_per_second": 2.899,
"step": 48
},
{
"epoch": 0.134107027724049,
"grad_norm": 13.741963386535645,
"learning_rate": 0.000290133381606063,
"loss": 8.36,
"step": 52
},
{
"epoch": 0.14442295293359123,
"grad_norm": 18.59068489074707,
"learning_rate": 0.0002882826847068703,
"loss": 7.7111,
"step": 56
},
{
"epoch": 0.14442295293359123,
"eval_loss": 7.73629903793335,
"eval_runtime": 10.7184,
"eval_samples_per_second": 22.858,
"eval_steps_per_second": 2.892,
"step": 56
},
{
"epoch": 0.15473887814313347,
"grad_norm": 20.837697982788086,
"learning_rate": 0.00028627997880029875,
"loss": 7.6612,
"step": 60
},
{
"epoch": 0.1650548033526757,
"grad_norm": 8.735236167907715,
"learning_rate": 0.0002841274653863955,
"loss": 7.6126,
"step": 64
},
{
"epoch": 0.1650548033526757,
"eval_loss": 7.594820022583008,
"eval_runtime": 10.664,
"eval_samples_per_second": 22.974,
"eval_steps_per_second": 2.907,
"step": 64
},
{
"epoch": 0.17537072856221791,
"grad_norm": 12.856707572937012,
"learning_rate": 0.00028182751064302397,
"loss": 7.5332,
"step": 68
},
{
"epoch": 0.18568665377176016,
"grad_norm": 13.578782081604004,
"learning_rate": 0.0002793826428248118,
"loss": 7.6789,
"step": 72
},
{
"epoch": 0.18568665377176016,
"eval_loss": 7.6147379875183105,
"eval_runtime": 10.6601,
"eval_samples_per_second": 22.983,
"eval_steps_per_second": 2.908,
"step": 72
},
{
"epoch": 0.19600257898130238,
"grad_norm": 7.203093528747559,
"learning_rate": 0.0002767955494839353,
"loss": 7.551,
"step": 76
},
{
"epoch": 0.2063185041908446,
"grad_norm": 5.764775276184082,
"learning_rate": 0.00027406907451579294,
"loss": 7.7404,
"step": 80
},
{
"epoch": 0.2063185041908446,
"eval_loss": 7.632232666015625,
"eval_runtime": 10.6044,
"eval_samples_per_second": 23.104,
"eval_steps_per_second": 2.923,
"step": 80
},
{
"epoch": 0.21663442940038685,
"grad_norm": 15.057692527770996,
"learning_rate": 0.0002712062150328175,
"loss": 7.7502,
"step": 84
},
{
"epoch": 0.22695035460992907,
"grad_norm": 7.761707305908203,
"learning_rate": 0.0002682101180698615,
"loss": 7.7173,
"step": 88
},
{
"epoch": 0.22695035460992907,
"eval_loss": 7.6740193367004395,
"eval_runtime": 10.575,
"eval_samples_per_second": 23.168,
"eval_steps_per_second": 2.931,
"step": 88
},
{
"epoch": 0.23726627981947132,
"grad_norm": 10.52561092376709,
"learning_rate": 0.000265084077124779,
"loss": 7.593,
"step": 92
},
{
"epoch": 0.24758220502901354,
"grad_norm": 13.96662425994873,
"learning_rate": 0.0002618315285380063,
"loss": 7.7113,
"step": 96
},
{
"epoch": 0.24758220502901354,
"eval_loss": 7.674198627471924,
"eval_runtime": 10.533,
"eval_samples_per_second": 23.26,
"eval_steps_per_second": 2.943,
"step": 96
},
{
"epoch": 0.2578981302385558,
"grad_norm": 5.200709819793701,
"learning_rate": 0.00025845604771512044,
"loss": 7.698,
"step": 100
},
{
"epoch": 0.268214055448098,
"grad_norm": 5.996535301208496,
"learning_rate": 0.00025496134519652946,
"loss": 7.6961,
"step": 104
},
{
"epoch": 0.268214055448098,
"eval_loss": 7.642160415649414,
"eval_runtime": 10.4668,
"eval_samples_per_second": 23.407,
"eval_steps_per_second": 2.962,
"step": 104
},
{
"epoch": 0.27852998065764023,
"grad_norm": 5.231250286102295,
"learning_rate": 0.00025135126257861296,
"loss": 7.7243,
"step": 108
},
{
"epoch": 0.28884590586718245,
"grad_norm": 4.705758094787598,
"learning_rate": 0.00024762976829079836,
"loss": 7.6729,
"step": 112
},
{
"epoch": 0.28884590586718245,
"eval_loss": 7.607565402984619,
"eval_runtime": 96.7081,
"eval_samples_per_second": 2.533,
"eval_steps_per_second": 0.321,
"step": 112
},
{
"epoch": 0.29916183107672467,
"grad_norm": 5.811936855316162,
"learning_rate": 0.00024380095323321433,
"loss": 7.6601,
"step": 116
},
{
"epoch": 0.30947775628626695,
"grad_norm": 6.612433433532715,
"learning_rate": 0.00023986902627971652,
"loss": 7.7225,
"step": 120
},
{
"epoch": 0.30947775628626695,
"eval_loss": 7.7170515060424805,
"eval_runtime": 102.1676,
"eval_samples_per_second": 2.398,
"eval_steps_per_second": 0.303,
"step": 120
},
{
"epoch": 0.31979368149580917,
"grad_norm": 22.143421173095703,
"learning_rate": 0.00023583830965122902,
"loss": 7.7705,
"step": 124
},
{
"epoch": 0.3301096067053514,
"grad_norm": 14.632792472839355,
"learning_rate": 0.000231713234164488,
"loss": 7.8259,
"step": 128
},
{
"epoch": 0.3301096067053514,
"eval_loss": 7.772371292114258,
"eval_runtime": 92.9761,
"eval_samples_per_second": 2.635,
"eval_steps_per_second": 0.333,
"step": 128
},
{
"epoch": 0.3404255319148936,
"grad_norm": 12.084867477416992,
"learning_rate": 0.00022749833436140976,
"loss": 7.753,
"step": 132
},
{
"epoch": 0.35074145712443583,
"grad_norm": 9.957462310791016,
"learning_rate": 0.00022319824352443768,
"loss": 7.6611,
"step": 136
},
{
"epoch": 0.35074145712443583,
"eval_loss": 7.5973920822143555,
"eval_runtime": 108.0726,
"eval_samples_per_second": 2.267,
"eval_steps_per_second": 0.287,
"step": 136
},
{
"epoch": 0.3610573823339781,
"grad_norm": 4.924723148345947,
"learning_rate": 0.0002188176885833471,
"loss": 7.6173,
"step": 140
},
{
"epoch": 0.3713733075435203,
"grad_norm": 9.477831840515137,
"learning_rate": 0.0002143614849191077,
"loss": 7.5696,
"step": 144
},
{
"epoch": 0.3713733075435203,
"eval_loss": 7.603182792663574,
"eval_runtime": 91.6779,
"eval_samples_per_second": 2.672,
"eval_steps_per_second": 0.338,
"step": 144
},
{
"epoch": 0.38168923275306255,
"grad_norm": 7.13740348815918,
"learning_rate": 0.00020983453107051425,
"loss": 7.6439,
"step": 148
},
{
"epoch": 0.39200515796260477,
"grad_norm": 10.324462890625,
"learning_rate": 0.00020524180334940528,
"loss": 7.6786,
"step": 152
},
{
"epoch": 0.39200515796260477,
"eval_loss": 7.616292476654053,
"eval_runtime": 106.0137,
"eval_samples_per_second": 2.311,
"eval_steps_per_second": 0.292,
"step": 152
},
{
"epoch": 0.402321083172147,
"grad_norm": 6.7738938331604,
"learning_rate": 0.00020058835037038873,
"loss": 7.6126,
"step": 156
},
{
"epoch": 0.4126370083816892,
"grad_norm": 4.637758255004883,
"learning_rate": 0.00019587928750108816,
"loss": 7.4746,
"step": 160
},
{
"epoch": 0.4126370083816892,
"eval_loss": 7.426816463470459,
"eval_runtime": 104.9303,
"eval_samples_per_second": 2.335,
"eval_steps_per_second": 0.295,
"step": 160
},
{
"epoch": 0.4229529335912315,
"grad_norm": 10.853010177612305,
"learning_rate": 0.000191119791239009,
"loss": 7.4631,
"step": 164
},
{
"epoch": 0.4332688588007737,
"grad_norm": 17.25995635986328,
"learning_rate": 0.000186315093521208,
"loss": 7.4383,
"step": 168
},
{
"epoch": 0.4332688588007737,
"eval_loss": 7.406857013702393,
"eval_runtime": 10.9396,
"eval_samples_per_second": 22.396,
"eval_steps_per_second": 2.834,
"step": 168
},
{
"epoch": 0.4435847840103159,
"grad_norm": 48.24932098388672,
"learning_rate": 0.00018147047597301952,
"loss": 7.5079,
"step": 172
},
{
"epoch": 0.45390070921985815,
"grad_norm": 10.03960132598877,
"learning_rate": 0.00017659126410216118,
"loss": 7.4469,
"step": 176
},
{
"epoch": 0.45390070921985815,
"eval_loss": 7.522530555725098,
"eval_runtime": 11.0052,
"eval_samples_per_second": 22.262,
"eval_steps_per_second": 2.817,
"step": 176
},
{
"epoch": 0.46421663442940037,
"grad_norm": 37.15389633178711,
"learning_rate": 0.00017168282144460167,
"loss": 7.4621,
"step": 180
},
{
"epoch": 0.47453255963894264,
"grad_norm": 17.32857894897461,
"learning_rate": 0.00016675054366862551,
"loss": 7.6465,
"step": 184
},
{
"epoch": 0.47453255963894264,
"eval_loss": 7.484851837158203,
"eval_runtime": 11.0169,
"eval_samples_per_second": 22.239,
"eval_steps_per_second": 2.814,
"step": 184
},
{
"epoch": 0.48484848484848486,
"grad_norm": 8.746517181396484,
"learning_rate": 0.0001617998526435754,
"loss": 7.5273,
"step": 188
},
{
"epoch": 0.4951644100580271,
"grad_norm": 10.883841514587402,
"learning_rate": 0.00015683619047979322,
"loss": 7.4025,
"step": 192
},
{
"epoch": 0.4951644100580271,
"eval_loss": 7.30989933013916,
"eval_runtime": 10.9893,
"eval_samples_per_second": 22.294,
"eval_steps_per_second": 2.821,
"step": 192
},
{
"epoch": 0.5054803352675693,
"grad_norm": 33.5476188659668,
"learning_rate": 0.000151865013546311,
"loss": 7.4748,
"step": 196
},
{
"epoch": 0.5157962604771116,
"grad_norm": 9.489357948303223,
"learning_rate": 0.00014689178647286702,
"loss": 7.3473,
"step": 200
},
{
"epoch": 0.5157962604771116,
"eval_loss": 7.262345314025879,
"eval_runtime": 11.0031,
"eval_samples_per_second": 22.267,
"eval_steps_per_second": 2.817,
"step": 200
},
{
"epoch": 0.5261121856866537,
"grad_norm": 17.06060791015625,
"learning_rate": 0.00014192197614284246,
"loss": 7.3059,
"step": 204
},
{
"epoch": 0.536428110896196,
"grad_norm": 6.841980934143066,
"learning_rate": 0.00013696104568371937,
"loss": 7.2821,
"step": 208
},
{
"epoch": 0.536428110896196,
"eval_loss": 7.248399257659912,
"eval_runtime": 10.9335,
"eval_samples_per_second": 22.408,
"eval_steps_per_second": 2.835,
"step": 208
},
{
"epoch": 0.5467440361057382,
"grad_norm": 30.315967559814453,
"learning_rate": 0.00013201444846166842,
"loss": 7.2134,
"step": 212
},
{
"epoch": 0.5570599613152805,
"grad_norm": 58.75193786621094,
"learning_rate": 0.00012708762208686638,
"loss": 7.389,
"step": 216
},
{
"epoch": 0.5570599613152805,
"eval_loss": 7.7176690101623535,
"eval_runtime": 11.0125,
"eval_samples_per_second": 22.247,
"eval_steps_per_second": 2.815,
"step": 216
},
{
"epoch": 0.5673758865248227,
"grad_norm": 12.3959379196167,
"learning_rate": 0.00012218598243613358,
"loss": 7.7131,
"step": 220
},
{
"epoch": 0.5776918117343649,
"grad_norm": 6.3016510009765625,
"learning_rate": 0.00011731491769946225,
"loss": 7.2912,
"step": 224
},
{
"epoch": 0.5776918117343649,
"eval_loss": 7.1140642166137695,
"eval_runtime": 10.9151,
"eval_samples_per_second": 22.446,
"eval_steps_per_second": 2.84,
"step": 224
},
{
"epoch": 0.5880077369439072,
"grad_norm": 12.621047019958496,
"learning_rate": 0.00011247978245697986,
"loss": 7.1747,
"step": 228
},
{
"epoch": 0.5983236621534493,
"grad_norm": 11.38538646697998,
"learning_rate": 0.00010768589179285843,
"loss": 7.1847,
"step": 232
},
{
"epoch": 0.5983236621534493,
"eval_loss": 7.114475250244141,
"eval_runtime": 10.877,
"eval_samples_per_second": 22.525,
"eval_steps_per_second": 2.85,
"step": 232
},
{
"epoch": 0.6086395873629916,
"grad_norm": 14.617621421813965,
"learning_rate": 0.0001029385154526404,
"loss": 7.1773,
"step": 236
},
{
"epoch": 0.6189555125725339,
"grad_norm": 8.394628524780273,
"learning_rate": 9.824287205040372e-05,
"loss": 7.2121,
"step": 240
},
{
"epoch": 0.6189555125725339,
"eval_loss": 7.146487236022949,
"eval_runtime": 10.8983,
"eval_samples_per_second": 22.481,
"eval_steps_per_second": 2.844,
"step": 240
},
{
"epoch": 0.6292714377820761,
"grad_norm": 9.852251052856445,
"learning_rate": 9.360412333213324e-05,
"loss": 7.2931,
"step": 244
},
{
"epoch": 0.6395873629916183,
"grad_norm": 9.686322212219238,
"learning_rate": 8.902736850160597e-05,
"loss": 7.1216,
"step": 248
},
{
"epoch": 0.6395873629916183,
"eval_loss": 7.147933006286621,
"eval_runtime": 10.9167,
"eval_samples_per_second": 22.443,
"eval_steps_per_second": 2.84,
"step": 248
},
{
"epoch": 0.6499032882011605,
"grad_norm": 13.205273628234863,
"learning_rate": 8.451763861502603e-05,
"loss": 7.4023,
"step": 252
},
{
"epoch": 0.6602192134107028,
"grad_norm": 7.406120300292969,
"learning_rate": 8.007989105057155e-05,
"loss": 7.2503,
"step": 256
},
{
"epoch": 0.6602192134107028,
"eval_loss": 7.110508441925049,
"eval_runtime": 10.8603,
"eval_samples_per_second": 22.559,
"eval_steps_per_second": 2.854,
"step": 256
},
{
"epoch": 0.670535138620245,
"grad_norm": 7.229670524597168,
"learning_rate": 7.571900405893403e-05,
"loss": 7.2731,
"step": 260
},
{
"epoch": 0.6808510638297872,
"grad_norm": 13.080286979675293,
"learning_rate": 7.143977140083847e-05,
"loss": 7.1416,
"step": 264
},
{
"epoch": 0.6808510638297872,
"eval_loss": 7.173031806945801,
"eval_runtime": 10.8292,
"eval_samples_per_second": 22.624,
"eval_steps_per_second": 2.863,
"step": 264
},
{
"epoch": 0.6911669890393295,
"grad_norm": 7.729543685913086,
"learning_rate": 6.724689707744056e-05,
"loss": 7.1927,
"step": 268
},
{
"epoch": 0.7014829142488717,
"grad_norm": 6.9558587074279785,
"learning_rate": 6.314499015939392e-05,
"loss": 7.2288,
"step": 272
},
{
"epoch": 0.7014829142488717,
"eval_loss": 7.149050235748291,
"eval_runtime": 101.2615,
"eval_samples_per_second": 2.419,
"eval_steps_per_second": 0.306,
"step": 272
},
{
"epoch": 0.7117988394584139,
"grad_norm": 37.099365234375,
"learning_rate": 5.913855972026981e-05,
"loss": 7.3082,
"step": 276
},
{
"epoch": 0.7221147646679562,
"grad_norm": 20.466567993164062,
"learning_rate": 5.52320098799004e-05,
"loss": 7.3502,
"step": 280
},
{
"epoch": 0.7221147646679562,
"eval_loss": 7.199110984802246,
"eval_runtime": 101.1627,
"eval_samples_per_second": 2.422,
"eval_steps_per_second": 0.306,
"step": 280
},
{
"epoch": 0.7324306898774984,
"grad_norm": 9.72053337097168,
"learning_rate": 5.1429634963094275e-05,
"loss": 7.1518,
"step": 284
},
{
"epoch": 0.7427466150870407,
"grad_norm": 8.377337455749512,
"learning_rate": 4.773561477904451e-05,
"loss": 7.2648,
"step": 288
},
{
"epoch": 0.7427466150870407,
"eval_loss": 7.140583038330078,
"eval_runtime": 102.1903,
"eval_samples_per_second": 2.397,
"eval_steps_per_second": 0.303,
"step": 288
},
{
"epoch": 0.7530625402965828,
"grad_norm": 11.30972671508789,
"learning_rate": 4.415401002662016e-05,
"loss": 7.2351,
"step": 292
},
{
"epoch": 0.7633784655061251,
"grad_norm": 11.135313987731934,
"learning_rate": 4.068875783059153e-05,
"loss": 7.1647,
"step": 296
},
{
"epoch": 0.7633784655061251,
"eval_loss": 7.122550964355469,
"eval_runtime": 94.8476,
"eval_samples_per_second": 2.583,
"eval_steps_per_second": 0.327,
"step": 296
},
{
"epoch": 0.7736943907156673,
"grad_norm": 3.4541268348693848,
"learning_rate": 3.734366741369488e-05,
"loss": 7.0893,
"step": 300
},
{
"epoch": 0.7840103159252095,
"grad_norm": 9.31264591217041,
"learning_rate": 3.4122415909296155e-05,
"loss": 7.1678,
"step": 304
},
{
"epoch": 0.7840103159252095,
"eval_loss": 7.084330081939697,
"eval_runtime": 97.4225,
"eval_samples_per_second": 2.515,
"eval_steps_per_second": 0.318,
"step": 304
},
{
"epoch": 0.7943262411347518,
"grad_norm": 4.439526081085205,
"learning_rate": 3.1028544319255097e-05,
"loss": 7.1907,
"step": 308
},
{
"epoch": 0.804642166344294,
"grad_norm": 7.1955766677856445,
"learning_rate": 2.8065453621433405e-05,
"loss": 7.1879,
"step": 312
},
{
"epoch": 0.804642166344294,
"eval_loss": 7.1045451164245605,
"eval_runtime": 98.3747,
"eval_samples_per_second": 2.49,
"eval_steps_per_second": 0.315,
"step": 312
},
{
"epoch": 0.8149580915538363,
"grad_norm": 7.633655548095703,
"learning_rate": 2.5236401031126713e-05,
"loss": 7.1379,
"step": 316
},
{
"epoch": 0.8252740167633784,
"grad_norm": 6.028517246246338,
"learning_rate": 2.2544496420529107e-05,
"loss": 7.2384,
"step": 320
},
{
"epoch": 0.8252740167633784,
"eval_loss": 7.113685607910156,
"eval_runtime": 104.0901,
"eval_samples_per_second": 2.354,
"eval_steps_per_second": 0.298,
"step": 320
},
{
"epoch": 0.8355899419729207,
"grad_norm": 13.628016471862793,
"learning_rate": 1.9992698900165983e-05,
"loss": 7.0224,
"step": 324
},
{
"epoch": 0.845905867182463,
"grad_norm": 9.757091522216797,
"learning_rate": 1.7583813566054454e-05,
"loss": 7.2301,
"step": 328
},
{
"epoch": 0.845905867182463,
"eval_loss": 7.094879150390625,
"eval_runtime": 11.0422,
"eval_samples_per_second": 22.188,
"eval_steps_per_second": 2.807,
"step": 328
},
{
"epoch": 0.8562217923920051,
"grad_norm": 12.490655899047852,
"learning_rate": 1.5320488416165604e-05,
"loss": 7.2313,
"step": 332
},
{
"epoch": 0.8665377176015474,
"grad_norm": 17.20357894897461,
"learning_rate": 1.3205211439578656e-05,
"loss": 7.2897,
"step": 336
},
{
"epoch": 0.8665377176015474,
"eval_loss": 7.127319812774658,
"eval_runtime": 11.0618,
"eval_samples_per_second": 22.148,
"eval_steps_per_second": 2.802,
"step": 336
},
{
"epoch": 0.8768536428110896,
"grad_norm": 9.043377876281738,
"learning_rate": 1.1240307881527516e-05,
"loss": 7.0591,
"step": 340
},
{
"epoch": 0.8871695680206318,
"grad_norm": 6.038938522338867,
"learning_rate": 9.427937687345144e-06,
"loss": 7.1483,
"step": 344
},
{
"epoch": 0.8871695680206318,
"eval_loss": 7.108406066894531,
"eval_runtime": 11.0489,
"eval_samples_per_second": 22.174,
"eval_steps_per_second": 2.806,
"step": 344
},
{
"epoch": 0.8974854932301741,
"grad_norm": 6.139174461364746,
"learning_rate": 7.770093128115911e-06,
"loss": 7.1885,
"step": 348
},
{
"epoch": 0.9078014184397163,
"grad_norm": 9.17184829711914,
"learning_rate": 6.268596610646382e-06,
"loss": 7.1119,
"step": 352
},
{
"epoch": 0.9078014184397163,
"eval_loss": 7.098378658294678,
"eval_runtime": 11.0353,
"eval_samples_per_second": 22.201,
"eval_steps_per_second": 2.809,
"step": 352
},
{
"epoch": 0.9181173436492586,
"grad_norm": 9.084051132202148,
"learning_rate": 4.9250986741612985e-06,
"loss": 7.0772,
"step": 356
},
{
"epoch": 0.9284332688588007,
"grad_norm": 4.8953070640563965,
"learning_rate": 3.7410761759270934e-06,
"loss": 7.2202,
"step": 360
},
{
"epoch": 0.9284332688588007,
"eval_loss": 7.076570510864258,
"eval_runtime": 11.0328,
"eval_samples_per_second": 22.206,
"eval_steps_per_second": 2.81,
"step": 360
},
{
"epoch": 0.938749194068343,
"grad_norm": 3.9921817779541016,
"learning_rate": 2.7178306677978767e-06,
"loss": 7.1844,
"step": 364
},
{
"epoch": 0.9490651192778853,
"grad_norm": 6.253467559814453,
"learning_rate": 1.8564869654679181e-06,
"loss": 7.1149,
"step": 368
},
{
"epoch": 0.9490651192778853,
"eval_loss": 7.073785781860352,
"eval_runtime": 10.9871,
"eval_samples_per_second": 22.299,
"eval_steps_per_second": 2.821,
"step": 368
},
{
"epoch": 0.9593810444874274,
"grad_norm": 5.679617881774902,
"learning_rate": 1.157991912003453e-06,
"loss": 7.204,
"step": 372
},
{
"epoch": 0.9696969696969697,
"grad_norm": 10.825676918029785,
"learning_rate": 6.231133370135422e-07,
"loss": 7.1986,
"step": 376
},
{
"epoch": 0.9696969696969697,
"eval_loss": 7.074854850769043,
"eval_runtime": 11.0251,
"eval_samples_per_second": 22.222,
"eval_steps_per_second": 2.812,
"step": 376
},
{
"epoch": 0.9800128949065119,
"grad_norm": 3.1997151374816895,
"learning_rate": 2.524392126034891e-07,
"loss": 7.289,
"step": 380
},
{
"epoch": 0.9903288201160542,
"grad_norm": 5.230665683746338,
"learning_rate": 4.63770070389724e-08,
"loss": 7.155,
"step": 384
},
{
"epoch": 0.9903288201160542,
"eval_loss": 7.072939872741699,
"eval_runtime": 11.0357,
"eval_samples_per_second": 22.201,
"eval_steps_per_second": 2.809,
"step": 384
}
],
"logging_steps": 4,
"max_steps": 387,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 8,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 9.288233513590456e+17,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}