chansung's picture
Model save
5292c8c verified
raw
history blame
85.4 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 10.0,
"eval_steps": 500,
"global_step": 2370,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.004219409282700422,
"grad_norm": 3.1438205242156982,
"learning_rate": 8.438818565400843e-07,
"loss": 2.5933,
"step": 1
},
{
"epoch": 0.02109704641350211,
"grad_norm": 3.45337176322937,
"learning_rate": 4.219409282700422e-06,
"loss": 2.5683,
"step": 5
},
{
"epoch": 0.04219409282700422,
"grad_norm": 3.8354620933532715,
"learning_rate": 8.438818565400844e-06,
"loss": 2.567,
"step": 10
},
{
"epoch": 0.06329113924050633,
"grad_norm": 3.2700271606445312,
"learning_rate": 1.2658227848101267e-05,
"loss": 2.4327,
"step": 15
},
{
"epoch": 0.08438818565400844,
"grad_norm": 3.2498395442962646,
"learning_rate": 1.6877637130801688e-05,
"loss": 2.2197,
"step": 20
},
{
"epoch": 0.10548523206751055,
"grad_norm": 2.3556034564971924,
"learning_rate": 2.1097046413502112e-05,
"loss": 1.8787,
"step": 25
},
{
"epoch": 0.12658227848101267,
"grad_norm": 1.5979266166687012,
"learning_rate": 2.5316455696202533e-05,
"loss": 1.5467,
"step": 30
},
{
"epoch": 0.14767932489451477,
"grad_norm": 1.260302186012268,
"learning_rate": 2.9535864978902954e-05,
"loss": 1.4303,
"step": 35
},
{
"epoch": 0.16877637130801687,
"grad_norm": 0.5591890811920166,
"learning_rate": 3.3755274261603375e-05,
"loss": 1.3705,
"step": 40
},
{
"epoch": 0.189873417721519,
"grad_norm": 0.5186863541603088,
"learning_rate": 3.79746835443038e-05,
"loss": 1.328,
"step": 45
},
{
"epoch": 0.2109704641350211,
"grad_norm": 0.5181670188903809,
"learning_rate": 4.2194092827004224e-05,
"loss": 1.2614,
"step": 50
},
{
"epoch": 0.2320675105485232,
"grad_norm": 0.5109447240829468,
"learning_rate": 4.641350210970464e-05,
"loss": 1.2328,
"step": 55
},
{
"epoch": 0.25316455696202533,
"grad_norm": 0.4200008809566498,
"learning_rate": 5.0632911392405066e-05,
"loss": 1.1882,
"step": 60
},
{
"epoch": 0.2742616033755274,
"grad_norm": 0.37015053629875183,
"learning_rate": 5.4852320675105484e-05,
"loss": 1.1461,
"step": 65
},
{
"epoch": 0.29535864978902954,
"grad_norm": 0.39964228868484497,
"learning_rate": 5.907172995780591e-05,
"loss": 1.1336,
"step": 70
},
{
"epoch": 0.31645569620253167,
"grad_norm": 0.3632591962814331,
"learning_rate": 6.329113924050633e-05,
"loss": 1.1081,
"step": 75
},
{
"epoch": 0.33755274261603374,
"grad_norm": 0.363908588886261,
"learning_rate": 6.751054852320675e-05,
"loss": 1.1042,
"step": 80
},
{
"epoch": 0.35864978902953587,
"grad_norm": 0.373738557100296,
"learning_rate": 7.172995780590718e-05,
"loss": 1.0802,
"step": 85
},
{
"epoch": 0.379746835443038,
"grad_norm": 0.3337308168411255,
"learning_rate": 7.59493670886076e-05,
"loss": 1.0781,
"step": 90
},
{
"epoch": 0.4008438818565401,
"grad_norm": 0.36707767844200134,
"learning_rate": 8.016877637130802e-05,
"loss": 1.07,
"step": 95
},
{
"epoch": 0.4219409282700422,
"grad_norm": 0.40698128938674927,
"learning_rate": 8.438818565400845e-05,
"loss": 1.0672,
"step": 100
},
{
"epoch": 0.4430379746835443,
"grad_norm": 0.4015476107597351,
"learning_rate": 8.860759493670887e-05,
"loss": 1.0508,
"step": 105
},
{
"epoch": 0.4641350210970464,
"grad_norm": 0.3830510675907135,
"learning_rate": 9.282700421940928e-05,
"loss": 1.0396,
"step": 110
},
{
"epoch": 0.48523206751054854,
"grad_norm": 0.540158748626709,
"learning_rate": 9.704641350210972e-05,
"loss": 1.0356,
"step": 115
},
{
"epoch": 0.5063291139240507,
"grad_norm": 0.5870048999786377,
"learning_rate": 0.00010126582278481013,
"loss": 1.0363,
"step": 120
},
{
"epoch": 0.5274261603375527,
"grad_norm": 0.6282536387443542,
"learning_rate": 0.00010548523206751055,
"loss": 1.0205,
"step": 125
},
{
"epoch": 0.5485232067510548,
"grad_norm": 0.5962668061256409,
"learning_rate": 0.00010970464135021097,
"loss": 1.0327,
"step": 130
},
{
"epoch": 0.569620253164557,
"grad_norm": 0.44145339727401733,
"learning_rate": 0.0001139240506329114,
"loss": 1.0036,
"step": 135
},
{
"epoch": 0.5907172995780591,
"grad_norm": 0.3850124776363373,
"learning_rate": 0.00011814345991561182,
"loss": 0.9939,
"step": 140
},
{
"epoch": 0.6118143459915611,
"grad_norm": 0.45476189255714417,
"learning_rate": 0.00012236286919831225,
"loss": 1.01,
"step": 145
},
{
"epoch": 0.6329113924050633,
"grad_norm": 0.4156922399997711,
"learning_rate": 0.00012658227848101267,
"loss": 1.0046,
"step": 150
},
{
"epoch": 0.6540084388185654,
"grad_norm": 0.5297821760177612,
"learning_rate": 0.00013080168776371308,
"loss": 0.9885,
"step": 155
},
{
"epoch": 0.6751054852320675,
"grad_norm": 0.4995609521865845,
"learning_rate": 0.0001350210970464135,
"loss": 0.9885,
"step": 160
},
{
"epoch": 0.6962025316455697,
"grad_norm": 0.43320751190185547,
"learning_rate": 0.00013924050632911395,
"loss": 0.9818,
"step": 165
},
{
"epoch": 0.7172995780590717,
"grad_norm": 0.3719841539859772,
"learning_rate": 0.00014345991561181436,
"loss": 0.9739,
"step": 170
},
{
"epoch": 0.7383966244725738,
"grad_norm": 0.40309059619903564,
"learning_rate": 0.00014767932489451478,
"loss": 0.9646,
"step": 175
},
{
"epoch": 0.759493670886076,
"grad_norm": 0.41251224279403687,
"learning_rate": 0.0001518987341772152,
"loss": 0.9606,
"step": 180
},
{
"epoch": 0.7805907172995781,
"grad_norm": 0.3959939181804657,
"learning_rate": 0.00015611814345991562,
"loss": 0.9585,
"step": 185
},
{
"epoch": 0.8016877637130801,
"grad_norm": 0.5289701819419861,
"learning_rate": 0.00016033755274261603,
"loss": 0.9709,
"step": 190
},
{
"epoch": 0.8227848101265823,
"grad_norm": 0.4239669740200043,
"learning_rate": 0.00016455696202531648,
"loss": 0.9577,
"step": 195
},
{
"epoch": 0.8438818565400844,
"grad_norm": 0.5463127493858337,
"learning_rate": 0.0001687763713080169,
"loss": 0.9703,
"step": 200
},
{
"epoch": 0.8649789029535865,
"grad_norm": 0.4942500591278076,
"learning_rate": 0.00017299578059071731,
"loss": 0.956,
"step": 205
},
{
"epoch": 0.8860759493670886,
"grad_norm": 0.361708402633667,
"learning_rate": 0.00017721518987341773,
"loss": 0.9598,
"step": 210
},
{
"epoch": 0.9071729957805907,
"grad_norm": 0.5146432518959045,
"learning_rate": 0.00018143459915611815,
"loss": 0.9606,
"step": 215
},
{
"epoch": 0.9282700421940928,
"grad_norm": 0.4746183156967163,
"learning_rate": 0.00018565400843881857,
"loss": 0.9255,
"step": 220
},
{
"epoch": 0.9493670886075949,
"grad_norm": 0.35139352083206177,
"learning_rate": 0.00018987341772151899,
"loss": 0.9452,
"step": 225
},
{
"epoch": 0.9704641350210971,
"grad_norm": 0.3744509816169739,
"learning_rate": 0.00019409282700421943,
"loss": 0.9388,
"step": 230
},
{
"epoch": 0.9915611814345991,
"grad_norm": 0.409365177154541,
"learning_rate": 0.00019831223628691985,
"loss": 0.9183,
"step": 235
},
{
"epoch": 1.0,
"eval_loss": 1.6516629457473755,
"eval_runtime": 0.5551,
"eval_samples_per_second": 3.603,
"eval_steps_per_second": 1.802,
"step": 237
},
{
"epoch": 1.0126582278481013,
"grad_norm": 0.3671579658985138,
"learning_rate": 0.00019999902382014363,
"loss": 0.9181,
"step": 240
},
{
"epoch": 1.0337552742616034,
"grad_norm": 0.37430304288864136,
"learning_rate": 0.0001999930583455953,
"loss": 0.9099,
"step": 245
},
{
"epoch": 1.0548523206751055,
"grad_norm": 0.3971017599105835,
"learning_rate": 0.00019998167004176888,
"loss": 0.9187,
"step": 250
},
{
"epoch": 1.0759493670886076,
"grad_norm": 0.35043320059776306,
"learning_rate": 0.00019996485952627552,
"loss": 0.9063,
"step": 255
},
{
"epoch": 1.0970464135021096,
"grad_norm": 0.34455016255378723,
"learning_rate": 0.00019994262771078406,
"loss": 0.9072,
"step": 260
},
{
"epoch": 1.1181434599156117,
"grad_norm": 0.358531653881073,
"learning_rate": 0.0001999149758009716,
"loss": 0.916,
"step": 265
},
{
"epoch": 1.139240506329114,
"grad_norm": 0.3874802589416504,
"learning_rate": 0.00019988190529645808,
"loss": 0.8913,
"step": 270
},
{
"epoch": 1.160337552742616,
"grad_norm": 0.3963850736618042,
"learning_rate": 0.00019984341799072504,
"loss": 0.9033,
"step": 275
},
{
"epoch": 1.1814345991561181,
"grad_norm": 0.485408753156662,
"learning_rate": 0.0001997995159710182,
"loss": 0.8965,
"step": 280
},
{
"epoch": 1.2025316455696202,
"grad_norm": 0.383575975894928,
"learning_rate": 0.00019975020161823445,
"loss": 0.8919,
"step": 285
},
{
"epoch": 1.2236286919831223,
"grad_norm": 0.3458747863769531,
"learning_rate": 0.00019969547760679258,
"loss": 0.8827,
"step": 290
},
{
"epoch": 1.2447257383966246,
"grad_norm": 0.36945974826812744,
"learning_rate": 0.00019963534690448835,
"loss": 0.8957,
"step": 295
},
{
"epoch": 1.2658227848101267,
"grad_norm": 0.39198634028434753,
"learning_rate": 0.0001995698127723334,
"loss": 0.879,
"step": 300
},
{
"epoch": 1.2869198312236287,
"grad_norm": 0.36909279227256775,
"learning_rate": 0.0001994988787643786,
"loss": 0.9014,
"step": 305
},
{
"epoch": 1.3080168776371308,
"grad_norm": 0.3728243112564087,
"learning_rate": 0.00019942254872752112,
"loss": 0.891,
"step": 310
},
{
"epoch": 1.3291139240506329,
"grad_norm": 0.34412381052970886,
"learning_rate": 0.00019934082680129586,
"loss": 0.8744,
"step": 315
},
{
"epoch": 1.350210970464135,
"grad_norm": 0.3310409486293793,
"learning_rate": 0.00019925371741765107,
"loss": 0.8788,
"step": 320
},
{
"epoch": 1.371308016877637,
"grad_norm": 0.3466126620769501,
"learning_rate": 0.00019916122530070783,
"loss": 0.8953,
"step": 325
},
{
"epoch": 1.3924050632911391,
"grad_norm": 0.3331877291202545,
"learning_rate": 0.00019906335546650392,
"loss": 0.8703,
"step": 330
},
{
"epoch": 1.4135021097046414,
"grad_norm": 0.37082529067993164,
"learning_rate": 0.0001989601132227218,
"loss": 0.8951,
"step": 335
},
{
"epoch": 1.4345991561181435,
"grad_norm": 0.3338633179664612,
"learning_rate": 0.00019885150416840082,
"loss": 0.8826,
"step": 340
},
{
"epoch": 1.4556962025316456,
"grad_norm": 0.32301968336105347,
"learning_rate": 0.00019873753419363336,
"loss": 0.8821,
"step": 345
},
{
"epoch": 1.4767932489451476,
"grad_norm": 0.3255464732646942,
"learning_rate": 0.00019861820947924565,
"loss": 0.87,
"step": 350
},
{
"epoch": 1.49789029535865,
"grad_norm": 0.33072352409362793,
"learning_rate": 0.0001984935364964625,
"loss": 0.8755,
"step": 355
},
{
"epoch": 1.518987341772152,
"grad_norm": 0.39160993695259094,
"learning_rate": 0.0001983635220065562,
"loss": 0.861,
"step": 360
},
{
"epoch": 1.540084388185654,
"grad_norm": 0.35338205099105835,
"learning_rate": 0.00019822817306048006,
"loss": 0.864,
"step": 365
},
{
"epoch": 1.5611814345991561,
"grad_norm": 0.30854344367980957,
"learning_rate": 0.00019808749699848593,
"loss": 0.8521,
"step": 370
},
{
"epoch": 1.5822784810126582,
"grad_norm": 0.3593827188014984,
"learning_rate": 0.00019794150144972602,
"loss": 0.8738,
"step": 375
},
{
"epoch": 1.6033755274261603,
"grad_norm": 0.30474522709846497,
"learning_rate": 0.0001977901943318393,
"loss": 0.8612,
"step": 380
},
{
"epoch": 1.6244725738396624,
"grad_norm": 0.3270438611507416,
"learning_rate": 0.0001976335838505221,
"loss": 0.8838,
"step": 385
},
{
"epoch": 1.6455696202531644,
"grad_norm": 0.3059784471988678,
"learning_rate": 0.00019747167849908304,
"loss": 0.8687,
"step": 390
},
{
"epoch": 1.6666666666666665,
"grad_norm": 0.3195934295654297,
"learning_rate": 0.00019730448705798239,
"loss": 0.8639,
"step": 395
},
{
"epoch": 1.6877637130801688,
"grad_norm": 0.31650060415267944,
"learning_rate": 0.00019713201859435602,
"loss": 0.8825,
"step": 400
},
{
"epoch": 1.7088607594936709,
"grad_norm": 0.33506250381469727,
"learning_rate": 0.0001969542824615235,
"loss": 0.8663,
"step": 405
},
{
"epoch": 1.729957805907173,
"grad_norm": 0.3536715805530548,
"learning_rate": 0.00019677128829848103,
"loss": 0.8498,
"step": 410
},
{
"epoch": 1.7510548523206753,
"grad_norm": 0.35253262519836426,
"learning_rate": 0.00019658304602937856,
"loss": 0.8614,
"step": 415
},
{
"epoch": 1.7721518987341773,
"grad_norm": 0.3264296054840088,
"learning_rate": 0.0001963895658629816,
"loss": 0.8456,
"step": 420
},
{
"epoch": 1.7932489451476794,
"grad_norm": 0.32499557733535767,
"learning_rate": 0.00019619085829211764,
"loss": 0.8435,
"step": 425
},
{
"epoch": 1.8143459915611815,
"grad_norm": 0.31721100211143494,
"learning_rate": 0.00019598693409310708,
"loss": 0.8716,
"step": 430
},
{
"epoch": 1.8354430379746836,
"grad_norm": 0.31418412923812866,
"learning_rate": 0.00019577780432517879,
"loss": 0.859,
"step": 435
},
{
"epoch": 1.8565400843881856,
"grad_norm": 0.3075924515724182,
"learning_rate": 0.0001955634803298703,
"loss": 0.8573,
"step": 440
},
{
"epoch": 1.8776371308016877,
"grad_norm": 0.30187729001045227,
"learning_rate": 0.00019534397373041285,
"loss": 0.8381,
"step": 445
},
{
"epoch": 1.8987341772151898,
"grad_norm": 0.3083683252334595,
"learning_rate": 0.00019511929643110097,
"loss": 0.8536,
"step": 450
},
{
"epoch": 1.9198312236286919,
"grad_norm": 0.2972455620765686,
"learning_rate": 0.0001948894606166468,
"loss": 0.8487,
"step": 455
},
{
"epoch": 1.9409282700421941,
"grad_norm": 0.3294317424297333,
"learning_rate": 0.00019465447875151946,
"loss": 0.8485,
"step": 460
},
{
"epoch": 1.9620253164556962,
"grad_norm": 0.28597962856292725,
"learning_rate": 0.00019441436357926892,
"loss": 0.8608,
"step": 465
},
{
"epoch": 1.9831223628691983,
"grad_norm": 0.30924198031425476,
"learning_rate": 0.00019416912812183498,
"loss": 0.8583,
"step": 470
},
{
"epoch": 2.0,
"eval_loss": 1.629499912261963,
"eval_runtime": 0.5557,
"eval_samples_per_second": 3.599,
"eval_steps_per_second": 1.8,
"step": 474
},
{
"epoch": 2.0042194092827006,
"grad_norm": 0.3041239380836487,
"learning_rate": 0.000193918785678841,
"loss": 0.8475,
"step": 475
},
{
"epoch": 2.0253164556962027,
"grad_norm": 0.31540679931640625,
"learning_rate": 0.0001936633498268728,
"loss": 0.8119,
"step": 480
},
{
"epoch": 2.0464135021097047,
"grad_norm": 0.3115026652812958,
"learning_rate": 0.0001934028344187421,
"loss": 0.8259,
"step": 485
},
{
"epoch": 2.067510548523207,
"grad_norm": 0.39703112840652466,
"learning_rate": 0.00019313725358273548,
"loss": 0.8041,
"step": 490
},
{
"epoch": 2.088607594936709,
"grad_norm": 0.3294001817703247,
"learning_rate": 0.00019286662172184808,
"loss": 0.8003,
"step": 495
},
{
"epoch": 2.109704641350211,
"grad_norm": 0.3266647160053253,
"learning_rate": 0.00019259095351300252,
"loss": 0.8109,
"step": 500
},
{
"epoch": 2.130801687763713,
"grad_norm": 0.31092244386672974,
"learning_rate": 0.0001923102639062529,
"loss": 0.8212,
"step": 505
},
{
"epoch": 2.151898734177215,
"grad_norm": 0.3094409704208374,
"learning_rate": 0.00019202456812397406,
"loss": 0.8187,
"step": 510
},
{
"epoch": 2.172995780590717,
"grad_norm": 0.32525700330734253,
"learning_rate": 0.00019173388166003613,
"loss": 0.8058,
"step": 515
},
{
"epoch": 2.1940928270042193,
"grad_norm": 0.296203076839447,
"learning_rate": 0.00019143822027896406,
"loss": 0.8037,
"step": 520
},
{
"epoch": 2.2151898734177213,
"grad_norm": 0.3076232373714447,
"learning_rate": 0.0001911376000150828,
"loss": 0.8208,
"step": 525
},
{
"epoch": 2.2362869198312234,
"grad_norm": 0.2956830859184265,
"learning_rate": 0.0001908320371716478,
"loss": 0.788,
"step": 530
},
{
"epoch": 2.257383966244726,
"grad_norm": 0.3284301161766052,
"learning_rate": 0.00019052154831996073,
"loss": 0.7986,
"step": 535
},
{
"epoch": 2.278481012658228,
"grad_norm": 0.31620386242866516,
"learning_rate": 0.00019020615029847072,
"loss": 0.8049,
"step": 540
},
{
"epoch": 2.29957805907173,
"grad_norm": 0.3297165632247925,
"learning_rate": 0.00018988586021186147,
"loss": 0.8309,
"step": 545
},
{
"epoch": 2.320675105485232,
"grad_norm": 0.2959255874156952,
"learning_rate": 0.0001895606954301233,
"loss": 0.7965,
"step": 550
},
{
"epoch": 2.3417721518987342,
"grad_norm": 0.3089437782764435,
"learning_rate": 0.00018923067358761136,
"loss": 0.7946,
"step": 555
},
{
"epoch": 2.3628691983122363,
"grad_norm": 0.30109426379203796,
"learning_rate": 0.00018889581258208903,
"loss": 0.8123,
"step": 560
},
{
"epoch": 2.3839662447257384,
"grad_norm": 0.32586753368377686,
"learning_rate": 0.0001885561305737577,
"loss": 0.8162,
"step": 565
},
{
"epoch": 2.4050632911392404,
"grad_norm": 0.3282499611377716,
"learning_rate": 0.00018821164598427145,
"loss": 0.8196,
"step": 570
},
{
"epoch": 2.4261603375527425,
"grad_norm": 0.3379076421260834,
"learning_rate": 0.00018786237749573837,
"loss": 0.816,
"step": 575
},
{
"epoch": 2.4472573839662446,
"grad_norm": 0.30591997504234314,
"learning_rate": 0.00018750834404970718,
"loss": 0.8015,
"step": 580
},
{
"epoch": 2.4683544303797467,
"grad_norm": 0.3031338155269623,
"learning_rate": 0.00018714956484613995,
"loss": 0.817,
"step": 585
},
{
"epoch": 2.489451476793249,
"grad_norm": 0.3209945261478424,
"learning_rate": 0.0001867860593423711,
"loss": 0.8134,
"step": 590
},
{
"epoch": 2.510548523206751,
"grad_norm": 0.31199783086776733,
"learning_rate": 0.000186417847252052,
"loss": 0.8053,
"step": 595
},
{
"epoch": 2.5316455696202533,
"grad_norm": 0.29871612787246704,
"learning_rate": 0.00018604494854408178,
"loss": 0.804,
"step": 600
},
{
"epoch": 2.5527426160337554,
"grad_norm": 0.31065070629119873,
"learning_rate": 0.0001856673834415246,
"loss": 0.8033,
"step": 605
},
{
"epoch": 2.5738396624472575,
"grad_norm": 0.2946309447288513,
"learning_rate": 0.00018528517242051283,
"loss": 0.8006,
"step": 610
},
{
"epoch": 2.5949367088607596,
"grad_norm": 0.2882119119167328,
"learning_rate": 0.00018489833620913642,
"loss": 0.8059,
"step": 615
},
{
"epoch": 2.6160337552742616,
"grad_norm": 0.31968575716018677,
"learning_rate": 0.00018450689578631898,
"loss": 0.8045,
"step": 620
},
{
"epoch": 2.6371308016877637,
"grad_norm": 0.3142683804035187,
"learning_rate": 0.00018411087238068003,
"loss": 0.8045,
"step": 625
},
{
"epoch": 2.6582278481012658,
"grad_norm": 0.3006449043750763,
"learning_rate": 0.0001837102874693836,
"loss": 0.8056,
"step": 630
},
{
"epoch": 2.679324894514768,
"grad_norm": 0.3158734440803528,
"learning_rate": 0.0001833051627769736,
"loss": 0.8275,
"step": 635
},
{
"epoch": 2.70042194092827,
"grad_norm": 0.30407387018203735,
"learning_rate": 0.00018289552027419558,
"loss": 0.8133,
"step": 640
},
{
"epoch": 2.721518987341772,
"grad_norm": 0.3022385835647583,
"learning_rate": 0.0001824813821768053,
"loss": 0.8026,
"step": 645
},
{
"epoch": 2.742616033755274,
"grad_norm": 0.29126378893852234,
"learning_rate": 0.00018206277094436377,
"loss": 0.8075,
"step": 650
},
{
"epoch": 2.7637130801687766,
"grad_norm": 0.29224568605422974,
"learning_rate": 0.00018163970927901937,
"loss": 0.811,
"step": 655
},
{
"epoch": 2.7848101265822782,
"grad_norm": 0.3017180860042572,
"learning_rate": 0.00018121222012427665,
"loss": 0.7945,
"step": 660
},
{
"epoch": 2.8059071729957807,
"grad_norm": 0.2974776029586792,
"learning_rate": 0.00018078032666375194,
"loss": 0.8078,
"step": 665
},
{
"epoch": 2.827004219409283,
"grad_norm": 0.2910807430744171,
"learning_rate": 0.0001803440523199162,
"loss": 0.7887,
"step": 670
},
{
"epoch": 2.848101265822785,
"grad_norm": 0.3062914311885834,
"learning_rate": 0.0001799034207528247,
"loss": 0.7928,
"step": 675
},
{
"epoch": 2.869198312236287,
"grad_norm": 0.29467758536338806,
"learning_rate": 0.0001794584558588338,
"loss": 0.8047,
"step": 680
},
{
"epoch": 2.890295358649789,
"grad_norm": 0.32635724544525146,
"learning_rate": 0.00017900918176930522,
"loss": 0.8144,
"step": 685
},
{
"epoch": 2.911392405063291,
"grad_norm": 0.31900787353515625,
"learning_rate": 0.00017855562284929718,
"loss": 0.8089,
"step": 690
},
{
"epoch": 2.932489451476793,
"grad_norm": 0.3085595667362213,
"learning_rate": 0.00017809780369624302,
"loss": 0.8048,
"step": 695
},
{
"epoch": 2.9535864978902953,
"grad_norm": 0.2946968078613281,
"learning_rate": 0.00017763574913861734,
"loss": 0.8157,
"step": 700
},
{
"epoch": 2.9746835443037973,
"grad_norm": 0.29723235964775085,
"learning_rate": 0.00017716948423458938,
"loss": 0.796,
"step": 705
},
{
"epoch": 2.9957805907173,
"grad_norm": 0.2712932229042053,
"learning_rate": 0.00017669903427066424,
"loss": 0.8179,
"step": 710
},
{
"epoch": 3.0,
"eval_loss": 1.6558986902236938,
"eval_runtime": 0.5507,
"eval_samples_per_second": 3.632,
"eval_steps_per_second": 1.816,
"step": 711
},
{
"epoch": 3.0168776371308015,
"grad_norm": 0.3356448709964752,
"learning_rate": 0.0001762244247603113,
"loss": 0.7628,
"step": 715
},
{
"epoch": 3.037974683544304,
"grad_norm": 0.3006523847579956,
"learning_rate": 0.00017574568144258077,
"loss": 0.7558,
"step": 720
},
{
"epoch": 3.059071729957806,
"grad_norm": 0.30827251076698303,
"learning_rate": 0.00017526283028070777,
"loss": 0.7567,
"step": 725
},
{
"epoch": 3.080168776371308,
"grad_norm": 0.3096933662891388,
"learning_rate": 0.00017477589746070417,
"loss": 0.7581,
"step": 730
},
{
"epoch": 3.1012658227848102,
"grad_norm": 0.32005831599235535,
"learning_rate": 0.00017428490938993862,
"loss": 0.7549,
"step": 735
},
{
"epoch": 3.1223628691983123,
"grad_norm": 0.30930569767951965,
"learning_rate": 0.00017378989269570437,
"loss": 0.7702,
"step": 740
},
{
"epoch": 3.1434599156118144,
"grad_norm": 0.32762596011161804,
"learning_rate": 0.0001732908742237752,
"loss": 0.7471,
"step": 745
},
{
"epoch": 3.1645569620253164,
"grad_norm": 0.32086798548698425,
"learning_rate": 0.00017278788103694943,
"loss": 0.7618,
"step": 750
},
{
"epoch": 3.1856540084388185,
"grad_norm": 0.3558262586593628,
"learning_rate": 0.00017228094041358248,
"loss": 0.7764,
"step": 755
},
{
"epoch": 3.2067510548523206,
"grad_norm": 0.3397001326084137,
"learning_rate": 0.0001717700798461074,
"loss": 0.753,
"step": 760
},
{
"epoch": 3.2278481012658227,
"grad_norm": 0.30650395154953003,
"learning_rate": 0.00017125532703954365,
"loss": 0.7595,
"step": 765
},
{
"epoch": 3.2489451476793247,
"grad_norm": 0.317777156829834,
"learning_rate": 0.0001707367099099951,
"loss": 0.7546,
"step": 770
},
{
"epoch": 3.270042194092827,
"grad_norm": 0.3183245062828064,
"learning_rate": 0.00017021425658313565,
"loss": 0.7633,
"step": 775
},
{
"epoch": 3.291139240506329,
"grad_norm": 0.3169344365596771,
"learning_rate": 0.00016968799539268407,
"loss": 0.7759,
"step": 780
},
{
"epoch": 3.3122362869198314,
"grad_norm": 0.30704858899116516,
"learning_rate": 0.00016915795487886746,
"loss": 0.7565,
"step": 785
},
{
"epoch": 3.3333333333333335,
"grad_norm": 0.3002530038356781,
"learning_rate": 0.0001686241637868734,
"loss": 0.7509,
"step": 790
},
{
"epoch": 3.3544303797468356,
"grad_norm": 0.3143273591995239,
"learning_rate": 0.00016808665106529094,
"loss": 0.7482,
"step": 795
},
{
"epoch": 3.3755274261603376,
"grad_norm": 0.30195352435112,
"learning_rate": 0.00016754544586454094,
"loss": 0.762,
"step": 800
},
{
"epoch": 3.3966244725738397,
"grad_norm": 0.32630935311317444,
"learning_rate": 0.00016700057753529484,
"loss": 0.7637,
"step": 805
},
{
"epoch": 3.4177215189873418,
"grad_norm": 0.31649506092071533,
"learning_rate": 0.0001664520756268832,
"loss": 0.7577,
"step": 810
},
{
"epoch": 3.438818565400844,
"grad_norm": 0.3301686644554138,
"learning_rate": 0.0001658999698856929,
"loss": 0.7534,
"step": 815
},
{
"epoch": 3.459915611814346,
"grad_norm": 0.3230050802230835,
"learning_rate": 0.00016534429025355426,
"loss": 0.7567,
"step": 820
},
{
"epoch": 3.481012658227848,
"grad_norm": 0.30652645230293274,
"learning_rate": 0.00016478506686611697,
"loss": 0.757,
"step": 825
},
{
"epoch": 3.50210970464135,
"grad_norm": 0.32210221886634827,
"learning_rate": 0.0001642223300512158,
"loss": 0.7734,
"step": 830
},
{
"epoch": 3.523206751054852,
"grad_norm": 0.3032419681549072,
"learning_rate": 0.00016365611032722604,
"loss": 0.7519,
"step": 835
},
{
"epoch": 3.5443037974683547,
"grad_norm": 0.2990473508834839,
"learning_rate": 0.00016308643840140828,
"loss": 0.7579,
"step": 840
},
{
"epoch": 3.5654008438818563,
"grad_norm": 0.32090187072753906,
"learning_rate": 0.000162513345168243,
"loss": 0.7569,
"step": 845
},
{
"epoch": 3.586497890295359,
"grad_norm": 0.3112528920173645,
"learning_rate": 0.00016193686170775537,
"loss": 0.7752,
"step": 850
},
{
"epoch": 3.607594936708861,
"grad_norm": 0.311675488948822,
"learning_rate": 0.00016135701928382952,
"loss": 0.7523,
"step": 855
},
{
"epoch": 3.628691983122363,
"grad_norm": 0.316641628742218,
"learning_rate": 0.000160773849342513,
"loss": 0.7651,
"step": 860
},
{
"epoch": 3.649789029535865,
"grad_norm": 0.32175716757774353,
"learning_rate": 0.00016018738351031156,
"loss": 0.7646,
"step": 865
},
{
"epoch": 3.670886075949367,
"grad_norm": 0.30499377846717834,
"learning_rate": 0.00015959765359247388,
"loss": 0.7654,
"step": 870
},
{
"epoch": 3.691983122362869,
"grad_norm": 0.3078381419181824,
"learning_rate": 0.0001590046915712667,
"loss": 0.7682,
"step": 875
},
{
"epoch": 3.7130801687763713,
"grad_norm": 0.3422172963619232,
"learning_rate": 0.00015840852960424036,
"loss": 0.7504,
"step": 880
},
{
"epoch": 3.7341772151898733,
"grad_norm": 0.30137816071510315,
"learning_rate": 0.00015780920002248484,
"loss": 0.75,
"step": 885
},
{
"epoch": 3.7552742616033754,
"grad_norm": 0.31054186820983887,
"learning_rate": 0.00015720673532887647,
"loss": 0.7511,
"step": 890
},
{
"epoch": 3.7763713080168775,
"grad_norm": 0.3199822008609772,
"learning_rate": 0.00015660116819631506,
"loss": 0.7659,
"step": 895
},
{
"epoch": 3.7974683544303796,
"grad_norm": 0.30703869462013245,
"learning_rate": 0.0001559925314659521,
"loss": 0.7641,
"step": 900
},
{
"epoch": 3.818565400843882,
"grad_norm": 0.3145774006843567,
"learning_rate": 0.00015538085814540962,
"loss": 0.7589,
"step": 905
},
{
"epoch": 3.8396624472573837,
"grad_norm": 0.3188943862915039,
"learning_rate": 0.00015476618140699034,
"loss": 0.7615,
"step": 910
},
{
"epoch": 3.8607594936708862,
"grad_norm": 0.32847416400909424,
"learning_rate": 0.00015414853458587833,
"loss": 0.7569,
"step": 915
},
{
"epoch": 3.8818565400843883,
"grad_norm": 0.33269551396369934,
"learning_rate": 0.00015352795117833145,
"loss": 0.7539,
"step": 920
},
{
"epoch": 3.9029535864978904,
"grad_norm": 0.30027341842651367,
"learning_rate": 0.00015290446483986472,
"loss": 0.76,
"step": 925
},
{
"epoch": 3.9240506329113924,
"grad_norm": 0.3010607063770294,
"learning_rate": 0.00015227810938342492,
"loss": 0.7574,
"step": 930
},
{
"epoch": 3.9451476793248945,
"grad_norm": 0.30083900690078735,
"learning_rate": 0.0001516489187775572,
"loss": 0.7556,
"step": 935
},
{
"epoch": 3.9662447257383966,
"grad_norm": 0.30435124039649963,
"learning_rate": 0.00015101692714456259,
"loss": 0.7612,
"step": 940
},
{
"epoch": 3.9873417721518987,
"grad_norm": 0.31260964274406433,
"learning_rate": 0.00015038216875864756,
"loss": 0.7533,
"step": 945
},
{
"epoch": 4.0,
"eval_loss": 1.6894222497940063,
"eval_runtime": 0.5552,
"eval_samples_per_second": 3.602,
"eval_steps_per_second": 1.801,
"step": 948
},
{
"epoch": 4.008438818565401,
"grad_norm": 0.3140685558319092,
"learning_rate": 0.00014974467804406533,
"loss": 0.749,
"step": 950
},
{
"epoch": 4.029535864978903,
"grad_norm": 0.3326011896133423,
"learning_rate": 0.00014910448957324897,
"loss": 0.7177,
"step": 955
},
{
"epoch": 4.050632911392405,
"grad_norm": 0.32034996151924133,
"learning_rate": 0.00014846163806493627,
"loss": 0.7061,
"step": 960
},
{
"epoch": 4.071729957805907,
"grad_norm": 0.31769704818725586,
"learning_rate": 0.00014781615838228715,
"loss": 0.6986,
"step": 965
},
{
"epoch": 4.0928270042194095,
"grad_norm": 0.35571393370628357,
"learning_rate": 0.00014716808553099286,
"loss": 0.7042,
"step": 970
},
{
"epoch": 4.113924050632911,
"grad_norm": 0.33056944608688354,
"learning_rate": 0.00014651745465737737,
"loss": 0.7195,
"step": 975
},
{
"epoch": 4.135021097046414,
"grad_norm": 0.35726672410964966,
"learning_rate": 0.00014586430104649163,
"loss": 0.7245,
"step": 980
},
{
"epoch": 4.156118143459915,
"grad_norm": 0.3273336887359619,
"learning_rate": 0.0001452086601201997,
"loss": 0.709,
"step": 985
},
{
"epoch": 4.177215189873418,
"grad_norm": 0.33940553665161133,
"learning_rate": 0.00014455056743525792,
"loss": 0.7115,
"step": 990
},
{
"epoch": 4.198312236286919,
"grad_norm": 0.34996211528778076,
"learning_rate": 0.00014389005868138658,
"loss": 0.7078,
"step": 995
},
{
"epoch": 4.219409282700422,
"grad_norm": 0.33837664127349854,
"learning_rate": 0.00014322716967933428,
"loss": 0.7042,
"step": 1000
},
{
"epoch": 4.2405063291139244,
"grad_norm": 0.3329886198043823,
"learning_rate": 0.0001425619363789354,
"loss": 0.7212,
"step": 1005
},
{
"epoch": 4.261603375527426,
"grad_norm": 0.35570377111434937,
"learning_rate": 0.00014189439485716053,
"loss": 0.7088,
"step": 1010
},
{
"epoch": 4.282700421940929,
"grad_norm": 0.3659791648387909,
"learning_rate": 0.00014122458131615975,
"loss": 0.7023,
"step": 1015
},
{
"epoch": 4.30379746835443,
"grad_norm": 0.3362638056278229,
"learning_rate": 0.00014055253208129938,
"loss": 0.7138,
"step": 1020
},
{
"epoch": 4.324894514767933,
"grad_norm": 0.3303203284740448,
"learning_rate": 0.00013987828359919222,
"loss": 0.7085,
"step": 1025
},
{
"epoch": 4.345991561181434,
"grad_norm": 0.32455962896347046,
"learning_rate": 0.00013920187243572057,
"loss": 0.7142,
"step": 1030
},
{
"epoch": 4.367088607594937,
"grad_norm": 0.33820950984954834,
"learning_rate": 0.00013852333527405346,
"loss": 0.7198,
"step": 1035
},
{
"epoch": 4.3881856540084385,
"grad_norm": 0.3443733751773834,
"learning_rate": 0.00013784270891265717,
"loss": 0.7281,
"step": 1040
},
{
"epoch": 4.409282700421941,
"grad_norm": 0.3376203179359436,
"learning_rate": 0.00013716003026329965,
"loss": 0.7157,
"step": 1045
},
{
"epoch": 4.430379746835443,
"grad_norm": 0.3343973159790039,
"learning_rate": 0.0001364753363490485,
"loss": 0.7157,
"step": 1050
},
{
"epoch": 4.451476793248945,
"grad_norm": 0.32973435521125793,
"learning_rate": 0.00013578866430226342,
"loss": 0.7183,
"step": 1055
},
{
"epoch": 4.472573839662447,
"grad_norm": 0.3444620370864868,
"learning_rate": 0.00013510005136258227,
"loss": 0.7196,
"step": 1060
},
{
"epoch": 4.493670886075949,
"grad_norm": 0.33004656434059143,
"learning_rate": 0.00013440953487490144,
"loss": 0.7139,
"step": 1065
},
{
"epoch": 4.514767932489452,
"grad_norm": 0.3244040608406067,
"learning_rate": 0.00013371715228735077,
"loss": 0.7144,
"step": 1070
},
{
"epoch": 4.5358649789029535,
"grad_norm": 0.3370364308357239,
"learning_rate": 0.0001330229411492625,
"loss": 0.7014,
"step": 1075
},
{
"epoch": 4.556962025316456,
"grad_norm": 0.3164542317390442,
"learning_rate": 0.00013232693910913485,
"loss": 0.7124,
"step": 1080
},
{
"epoch": 4.578059071729958,
"grad_norm": 0.3478745222091675,
"learning_rate": 0.0001316291839125904,
"loss": 0.7253,
"step": 1085
},
{
"epoch": 4.59915611814346,
"grad_norm": 0.33551761507987976,
"learning_rate": 0.00013092971340032905,
"loss": 0.7237,
"step": 1090
},
{
"epoch": 4.620253164556962,
"grad_norm": 0.3593490421772003,
"learning_rate": 0.00013022856550607572,
"loss": 0.7187,
"step": 1095
},
{
"epoch": 4.641350210970464,
"grad_norm": 0.33983170986175537,
"learning_rate": 0.0001295257782545233,
"loss": 0.715,
"step": 1100
},
{
"epoch": 4.662447257383966,
"grad_norm": 0.3238469064235687,
"learning_rate": 0.00012882138975927026,
"loss": 0.7024,
"step": 1105
},
{
"epoch": 4.6835443037974684,
"grad_norm": 0.3401734232902527,
"learning_rate": 0.00012811543822075397,
"loss": 0.7175,
"step": 1110
},
{
"epoch": 4.70464135021097,
"grad_norm": 0.35343295335769653,
"learning_rate": 0.00012740796192417875,
"loss": 0.7445,
"step": 1115
},
{
"epoch": 4.725738396624473,
"grad_norm": 0.3328774869441986,
"learning_rate": 0.00012669899923743968,
"loss": 0.7007,
"step": 1120
},
{
"epoch": 4.746835443037975,
"grad_norm": 0.341886967420578,
"learning_rate": 0.00012598858860904193,
"loss": 0.7275,
"step": 1125
},
{
"epoch": 4.767932489451477,
"grad_norm": 0.33224210143089294,
"learning_rate": 0.00012527676856601542,
"loss": 0.7093,
"step": 1130
},
{
"epoch": 4.789029535864979,
"grad_norm": 0.3608289062976837,
"learning_rate": 0.0001245635777118256,
"loss": 0.7237,
"step": 1135
},
{
"epoch": 4.810126582278481,
"grad_norm": 0.3258775472640991,
"learning_rate": 0.00012384905472427975,
"loss": 0.7068,
"step": 1140
},
{
"epoch": 4.831223628691983,
"grad_norm": 0.3356561064720154,
"learning_rate": 0.0001231332383534296,
"loss": 0.7208,
"step": 1145
},
{
"epoch": 4.852320675105485,
"grad_norm": 0.32746249437332153,
"learning_rate": 0.00012241616741946962,
"loss": 0.7143,
"step": 1150
},
{
"epoch": 4.8734177215189876,
"grad_norm": 0.33153674006462097,
"learning_rate": 0.0001216978808106318,
"loss": 0.726,
"step": 1155
},
{
"epoch": 4.894514767932489,
"grad_norm": 0.33083775639533997,
"learning_rate": 0.00012097841748107681,
"loss": 0.7015,
"step": 1160
},
{
"epoch": 4.915611814345992,
"grad_norm": 0.3352629542350769,
"learning_rate": 0.00012025781644878118,
"loss": 0.7234,
"step": 1165
},
{
"epoch": 4.936708860759493,
"grad_norm": 0.3423599898815155,
"learning_rate": 0.00011953611679342143,
"loss": 0.733,
"step": 1170
},
{
"epoch": 4.957805907172996,
"grad_norm": 0.3402250409126282,
"learning_rate": 0.00011881335765425473,
"loss": 0.7187,
"step": 1175
},
{
"epoch": 4.978902953586498,
"grad_norm": 0.345759779214859,
"learning_rate": 0.00011808957822799614,
"loss": 0.7119,
"step": 1180
},
{
"epoch": 5.0,
"grad_norm": 0.3274582326412201,
"learning_rate": 0.00011736481776669306,
"loss": 0.716,
"step": 1185
},
{
"epoch": 5.0,
"eval_loss": 1.7251324653625488,
"eval_runtime": 0.554,
"eval_samples_per_second": 3.61,
"eval_steps_per_second": 1.805,
"step": 1185
},
{
"epoch": 5.0210970464135025,
"grad_norm": 0.36755794286727905,
"learning_rate": 0.0001166391155755964,
"loss": 0.6589,
"step": 1190
},
{
"epoch": 5.042194092827004,
"grad_norm": 0.3618139624595642,
"learning_rate": 0.00011591251101102906,
"loss": 0.6697,
"step": 1195
},
{
"epoch": 5.063291139240507,
"grad_norm": 0.38643401861190796,
"learning_rate": 0.00011518504347825145,
"loss": 0.661,
"step": 1200
},
{
"epoch": 5.084388185654008,
"grad_norm": 0.3515397012233734,
"learning_rate": 0.00011445675242932457,
"loss": 0.6455,
"step": 1205
},
{
"epoch": 5.105485232067511,
"grad_norm": 0.37624698877334595,
"learning_rate": 0.00011372767736097039,
"loss": 0.6628,
"step": 1210
},
{
"epoch": 5.1265822784810124,
"grad_norm": 0.3468095660209656,
"learning_rate": 0.00011299785781242982,
"loss": 0.6591,
"step": 1215
},
{
"epoch": 5.147679324894515,
"grad_norm": 0.3849187195301056,
"learning_rate": 0.00011226733336331855,
"loss": 0.6726,
"step": 1220
},
{
"epoch": 5.168776371308017,
"grad_norm": 0.36786338686943054,
"learning_rate": 0.00011153614363148032,
"loss": 0.6795,
"step": 1225
},
{
"epoch": 5.189873417721519,
"grad_norm": 0.35997211933135986,
"learning_rate": 0.00011080432827083873,
"loss": 0.676,
"step": 1230
},
{
"epoch": 5.210970464135021,
"grad_norm": 0.3702506721019745,
"learning_rate": 0.00011007192696924638,
"loss": 0.6734,
"step": 1235
},
{
"epoch": 5.232067510548523,
"grad_norm": 0.35727155208587646,
"learning_rate": 0.00010933897944633265,
"loss": 0.6719,
"step": 1240
},
{
"epoch": 5.253164556962025,
"grad_norm": 0.35158923268318176,
"learning_rate": 0.0001086055254513497,
"loss": 0.6572,
"step": 1245
},
{
"epoch": 5.274261603375527,
"grad_norm": 0.3676392734050751,
"learning_rate": 0.00010787160476101668,
"loss": 0.663,
"step": 1250
},
{
"epoch": 5.29535864978903,
"grad_norm": 0.35416457056999207,
"learning_rate": 0.00010713725717736254,
"loss": 0.6619,
"step": 1255
},
{
"epoch": 5.3164556962025316,
"grad_norm": 0.36827412247657776,
"learning_rate": 0.00010640252252556759,
"loss": 0.6861,
"step": 1260
},
{
"epoch": 5.337552742616034,
"grad_norm": 0.37270885705947876,
"learning_rate": 0.00010566744065180368,
"loss": 0.6842,
"step": 1265
},
{
"epoch": 5.358649789029536,
"grad_norm": 0.6396368741989136,
"learning_rate": 0.00010493205142107312,
"loss": 0.6648,
"step": 1270
},
{
"epoch": 5.379746835443038,
"grad_norm": 0.3901231288909912,
"learning_rate": 0.00010419639471504682,
"loss": 0.6682,
"step": 1275
},
{
"epoch": 5.40084388185654,
"grad_norm": 0.3932683765888214,
"learning_rate": 0.0001034605104299016,
"loss": 0.6715,
"step": 1280
},
{
"epoch": 5.421940928270042,
"grad_norm": 0.3795235753059387,
"learning_rate": 0.00010272443847415615,
"loss": 0.6826,
"step": 1285
},
{
"epoch": 5.443037974683544,
"grad_norm": 0.3844228982925415,
"learning_rate": 0.00010198821876650701,
"loss": 0.6624,
"step": 1290
},
{
"epoch": 5.4641350210970465,
"grad_norm": 0.37277084589004517,
"learning_rate": 0.00010125189123366368,
"loss": 0.6818,
"step": 1295
},
{
"epoch": 5.485232067510548,
"grad_norm": 0.3795084059238434,
"learning_rate": 0.0001005154958081831,
"loss": 0.6688,
"step": 1300
},
{
"epoch": 5.506329113924051,
"grad_norm": 0.37196341156959534,
"learning_rate": 9.977907242630426e-05,
"loss": 0.6627,
"step": 1305
},
{
"epoch": 5.527426160337553,
"grad_norm": 0.3792167603969574,
"learning_rate": 9.904266102578231e-05,
"loss": 0.6768,
"step": 1310
},
{
"epoch": 5.548523206751055,
"grad_norm": 0.3688276410102844,
"learning_rate": 9.830630154372252e-05,
"loss": 0.6663,
"step": 1315
},
{
"epoch": 5.569620253164557,
"grad_norm": 0.3876282870769501,
"learning_rate": 9.75700339144146e-05,
"loss": 0.6757,
"step": 1320
},
{
"epoch": 5.590717299578059,
"grad_norm": 0.35115256905555725,
"learning_rate": 9.68338980671669e-05,
"loss": 0.6846,
"step": 1325
},
{
"epoch": 5.6118143459915615,
"grad_norm": 0.3650346100330353,
"learning_rate": 9.609793392414086e-05,
"loss": 0.6948,
"step": 1330
},
{
"epoch": 5.632911392405063,
"grad_norm": 0.3864571750164032,
"learning_rate": 9.536218139818614e-05,
"loss": 0.6712,
"step": 1335
},
{
"epoch": 5.654008438818566,
"grad_norm": 0.36888009309768677,
"learning_rate": 9.462668039067602e-05,
"loss": 0.6705,
"step": 1340
},
{
"epoch": 5.675105485232067,
"grad_norm": 0.36247017979621887,
"learning_rate": 9.389147078934329e-05,
"loss": 0.6696,
"step": 1345
},
{
"epoch": 5.69620253164557,
"grad_norm": 0.3620111048221588,
"learning_rate": 9.31565924661172e-05,
"loss": 0.6686,
"step": 1350
},
{
"epoch": 5.717299578059071,
"grad_norm": 0.3552044630050659,
"learning_rate": 9.242208527496121e-05,
"loss": 0.6922,
"step": 1355
},
{
"epoch": 5.738396624472574,
"grad_norm": 0.36270490288734436,
"learning_rate": 9.168798904971143e-05,
"loss": 0.6625,
"step": 1360
},
{
"epoch": 5.759493670886076,
"grad_norm": 0.3620161712169647,
"learning_rate": 9.095434360191642e-05,
"loss": 0.6684,
"step": 1365
},
{
"epoch": 5.780590717299578,
"grad_norm": 0.37736937403678894,
"learning_rate": 9.02211887186783e-05,
"loss": 0.6896,
"step": 1370
},
{
"epoch": 5.80168776371308,
"grad_norm": 0.4165714979171753,
"learning_rate": 8.948856416049475e-05,
"loss": 0.6704,
"step": 1375
},
{
"epoch": 5.822784810126582,
"grad_norm": 0.3674893081188202,
"learning_rate": 8.875650965910279e-05,
"loss": 0.6871,
"step": 1380
},
{
"epoch": 5.843881856540085,
"grad_norm": 0.39419984817504883,
"learning_rate": 8.802506491532421e-05,
"loss": 0.6941,
"step": 1385
},
{
"epoch": 5.864978902953586,
"grad_norm": 0.3581133782863617,
"learning_rate": 8.72942695969123e-05,
"loss": 0.6815,
"step": 1390
},
{
"epoch": 5.886075949367089,
"grad_norm": 0.3663847744464874,
"learning_rate": 8.656416333640066e-05,
"loss": 0.6792,
"step": 1395
},
{
"epoch": 5.9071729957805905,
"grad_norm": 0.39068740606307983,
"learning_rate": 8.583478572895394e-05,
"loss": 0.6689,
"step": 1400
},
{
"epoch": 5.928270042194093,
"grad_norm": 0.3782387971878052,
"learning_rate": 8.510617633022044e-05,
"loss": 0.6825,
"step": 1405
},
{
"epoch": 5.949367088607595,
"grad_norm": 0.3802437484264374,
"learning_rate": 8.437837465418684e-05,
"loss": 0.669,
"step": 1410
},
{
"epoch": 5.970464135021097,
"grad_norm": 0.35812443494796753,
"learning_rate": 8.365142017103542e-05,
"loss": 0.6788,
"step": 1415
},
{
"epoch": 5.991561181434599,
"grad_norm": 0.36878547072410583,
"learning_rate": 8.292535230500342e-05,
"loss": 0.6876,
"step": 1420
},
{
"epoch": 6.0,
"eval_loss": 1.782979130744934,
"eval_runtime": 0.5539,
"eval_samples_per_second": 3.611,
"eval_steps_per_second": 1.805,
"step": 1422
},
{
"epoch": 6.012658227848101,
"grad_norm": 0.40747499465942383,
"learning_rate": 8.2200210432245e-05,
"loss": 0.6326,
"step": 1425
},
{
"epoch": 6.033755274261603,
"grad_norm": 0.40314236283302307,
"learning_rate": 8.147603387869582e-05,
"loss": 0.6234,
"step": 1430
},
{
"epoch": 6.0548523206751055,
"grad_norm": 0.3922586739063263,
"learning_rate": 8.075286191794025e-05,
"loss": 0.6238,
"step": 1435
},
{
"epoch": 6.075949367088608,
"grad_norm": 0.41105443239212036,
"learning_rate": 8.003073376908163e-05,
"loss": 0.6312,
"step": 1440
},
{
"epoch": 6.09704641350211,
"grad_norm": 0.3970966339111328,
"learning_rate": 7.930968859461516e-05,
"loss": 0.6233,
"step": 1445
},
{
"epoch": 6.118143459915612,
"grad_norm": 0.42427581548690796,
"learning_rate": 7.85897654983041e-05,
"loss": 0.6348,
"step": 1450
},
{
"epoch": 6.139240506329114,
"grad_norm": 0.38989487290382385,
"learning_rate": 7.787100352305908e-05,
"loss": 0.6237,
"step": 1455
},
{
"epoch": 6.160337552742616,
"grad_norm": 0.4042844772338867,
"learning_rate": 7.715344164882085e-05,
"loss": 0.6232,
"step": 1460
},
{
"epoch": 6.181434599156118,
"grad_norm": 0.40070950984954834,
"learning_rate": 7.643711879044612e-05,
"loss": 0.6173,
"step": 1465
},
{
"epoch": 6.2025316455696204,
"grad_norm": 0.40951260924339294,
"learning_rate": 7.572207379559721e-05,
"loss": 0.6369,
"step": 1470
},
{
"epoch": 6.223628691983122,
"grad_norm": 0.40946945548057556,
"learning_rate": 7.50083454426354e-05,
"loss": 0.6267,
"step": 1475
},
{
"epoch": 6.244725738396625,
"grad_norm": 0.40567830204963684,
"learning_rate": 7.429597243851764e-05,
"loss": 0.616,
"step": 1480
},
{
"epoch": 6.265822784810126,
"grad_norm": 0.4094925820827484,
"learning_rate": 7.358499341669756e-05,
"loss": 0.6231,
"step": 1485
},
{
"epoch": 6.286919831223629,
"grad_norm": 0.396982878446579,
"learning_rate": 7.287544693503028e-05,
"loss": 0.6263,
"step": 1490
},
{
"epoch": 6.308016877637131,
"grad_norm": 0.41034215688705444,
"learning_rate": 7.216737147368127e-05,
"loss": 0.6466,
"step": 1495
},
{
"epoch": 6.329113924050633,
"grad_norm": 0.4219072163105011,
"learning_rate": 7.146080543303965e-05,
"loss": 0.6479,
"step": 1500
},
{
"epoch": 6.350210970464135,
"grad_norm": 0.39759665727615356,
"learning_rate": 7.075578713163541e-05,
"loss": 0.6235,
"step": 1505
},
{
"epoch": 6.371308016877637,
"grad_norm": 0.4137880504131317,
"learning_rate": 7.00523548040616e-05,
"loss": 0.6221,
"step": 1510
},
{
"epoch": 6.3924050632911396,
"grad_norm": 0.4084639847278595,
"learning_rate": 6.935054659890052e-05,
"loss": 0.633,
"step": 1515
},
{
"epoch": 6.413502109704641,
"grad_norm": 0.39727863669395447,
"learning_rate": 6.865040057665506e-05,
"loss": 0.6356,
"step": 1520
},
{
"epoch": 6.434599156118144,
"grad_norm": 0.4197627007961273,
"learning_rate": 6.795195470768444e-05,
"loss": 0.6355,
"step": 1525
},
{
"epoch": 6.455696202531645,
"grad_norm": 0.4036734402179718,
"learning_rate": 6.725524687014514e-05,
"loss": 0.6367,
"step": 1530
},
{
"epoch": 6.476793248945148,
"grad_norm": 0.4073878526687622,
"learning_rate": 6.656031484793657e-05,
"loss": 0.6367,
"step": 1535
},
{
"epoch": 6.4978902953586495,
"grad_norm": 0.4095742702484131,
"learning_rate": 6.586719632865198e-05,
"loss": 0.6292,
"step": 1540
},
{
"epoch": 6.518987341772152,
"grad_norm": 0.408542662858963,
"learning_rate": 6.517592890153476e-05,
"loss": 0.6312,
"step": 1545
},
{
"epoch": 6.540084388185654,
"grad_norm": 0.4064979553222656,
"learning_rate": 6.448655005543969e-05,
"loss": 0.6373,
"step": 1550
},
{
"epoch": 6.561181434599156,
"grad_norm": 0.4208141565322876,
"learning_rate": 6.379909717679985e-05,
"loss": 0.6289,
"step": 1555
},
{
"epoch": 6.582278481012658,
"grad_norm": 0.4085118770599365,
"learning_rate": 6.311360754759923e-05,
"loss": 0.6289,
"step": 1560
},
{
"epoch": 6.60337552742616,
"grad_norm": 0.4019670784473419,
"learning_rate": 6.243011834335075e-05,
"loss": 0.639,
"step": 1565
},
{
"epoch": 6.624472573839663,
"grad_norm": 0.4115982949733734,
"learning_rate": 6.17486666310801e-05,
"loss": 0.6437,
"step": 1570
},
{
"epoch": 6.6455696202531644,
"grad_norm": 0.40410783886909485,
"learning_rate": 6.106928936731571e-05,
"loss": 0.6439,
"step": 1575
},
{
"epoch": 6.666666666666667,
"grad_norm": 0.3954565227031708,
"learning_rate": 6.039202339608432e-05,
"loss": 0.6339,
"step": 1580
},
{
"epoch": 6.687763713080169,
"grad_norm": 0.40417176485061646,
"learning_rate": 5.971690544691294e-05,
"loss": 0.6238,
"step": 1585
},
{
"epoch": 6.708860759493671,
"grad_norm": 0.40106064081192017,
"learning_rate": 5.90439721328369e-05,
"loss": 0.6183,
"step": 1590
},
{
"epoch": 6.729957805907173,
"grad_norm": 0.3997708261013031,
"learning_rate": 5.837325994841434e-05,
"loss": 0.6349,
"step": 1595
},
{
"epoch": 6.751054852320675,
"grad_norm": 0.40423154830932617,
"learning_rate": 5.770480526774693e-05,
"loss": 0.6319,
"step": 1600
},
{
"epoch": 6.772151898734177,
"grad_norm": 0.39728954434394836,
"learning_rate": 5.7038644342507205e-05,
"loss": 0.6454,
"step": 1605
},
{
"epoch": 6.793248945147679,
"grad_norm": 0.4143037497997284,
"learning_rate": 5.6374813299972805e-05,
"loss": 0.6532,
"step": 1610
},
{
"epoch": 6.814345991561181,
"grad_norm": 0.4104886054992676,
"learning_rate": 5.571334814106681e-05,
"loss": 0.6375,
"step": 1615
},
{
"epoch": 6.8354430379746836,
"grad_norm": 0.41742509603500366,
"learning_rate": 5.505428473840576e-05,
"loss": 0.6443,
"step": 1620
},
{
"epoch": 6.856540084388186,
"grad_norm": 0.4019664227962494,
"learning_rate": 5.4397658834353895e-05,
"loss": 0.6207,
"step": 1625
},
{
"epoch": 6.877637130801688,
"grad_norm": 0.4325370490550995,
"learning_rate": 5.3743506039084913e-05,
"loss": 0.6357,
"step": 1630
},
{
"epoch": 6.89873417721519,
"grad_norm": 0.4043619632720947,
"learning_rate": 5.309186182865076e-05,
"loss": 0.646,
"step": 1635
},
{
"epoch": 6.919831223628692,
"grad_norm": 0.4148579239845276,
"learning_rate": 5.244276154305758e-05,
"loss": 0.6417,
"step": 1640
},
{
"epoch": 6.940928270042194,
"grad_norm": 0.41201284527778625,
"learning_rate": 5.179624038434938e-05,
"loss": 0.6396,
"step": 1645
},
{
"epoch": 6.962025316455696,
"grad_norm": 0.4112018644809723,
"learning_rate": 5.115233341469877e-05,
"loss": 0.6391,
"step": 1650
},
{
"epoch": 6.9831223628691985,
"grad_norm": 0.42246147990226746,
"learning_rate": 5.0511075554505426e-05,
"loss": 0.6344,
"step": 1655
},
{
"epoch": 7.0,
"eval_loss": 1.8556586503982544,
"eval_runtime": 0.5548,
"eval_samples_per_second": 3.605,
"eval_steps_per_second": 1.802,
"step": 1659
},
{
"epoch": 7.0042194092827,
"grad_norm": 0.38922393321990967,
"learning_rate": 4.987250158050244e-05,
"loss": 0.6267,
"step": 1660
},
{
"epoch": 7.025316455696203,
"grad_norm": 0.4556201100349426,
"learning_rate": 4.923664612387019e-05,
"loss": 0.5894,
"step": 1665
},
{
"epoch": 7.046413502109704,
"grad_norm": 0.4320254325866699,
"learning_rate": 4.860354366835825e-05,
"loss": 0.6007,
"step": 1670
},
{
"epoch": 7.067510548523207,
"grad_norm": 0.41525062918663025,
"learning_rate": 4.7973228548415385e-05,
"loss": 0.5944,
"step": 1675
},
{
"epoch": 7.0886075949367084,
"grad_norm": 0.46430733799934387,
"learning_rate": 4.734573494732735e-05,
"loss": 0.5945,
"step": 1680
},
{
"epoch": 7.109704641350211,
"grad_norm": 0.421763151884079,
"learning_rate": 4.6721096895363114e-05,
"loss": 0.583,
"step": 1685
},
{
"epoch": 7.1308016877637135,
"grad_norm": 0.44340547919273376,
"learning_rate": 4.6099348267929334e-05,
"loss": 0.6034,
"step": 1690
},
{
"epoch": 7.151898734177215,
"grad_norm": 0.4334201216697693,
"learning_rate": 4.548052278373327e-05,
"loss": 0.592,
"step": 1695
},
{
"epoch": 7.172995780590718,
"grad_norm": 0.4375658631324768,
"learning_rate": 4.486465400295404e-05,
"loss": 0.5942,
"step": 1700
},
{
"epoch": 7.194092827004219,
"grad_norm": 0.4318469762802124,
"learning_rate": 4.4251775325422795e-05,
"loss": 0.6079,
"step": 1705
},
{
"epoch": 7.215189873417722,
"grad_norm": 0.4487842619419098,
"learning_rate": 4.364191998881104e-05,
"loss": 0.5938,
"step": 1710
},
{
"epoch": 7.236286919831223,
"grad_norm": 0.4327734112739563,
"learning_rate": 4.303512106682849e-05,
"loss": 0.5965,
"step": 1715
},
{
"epoch": 7.257383966244726,
"grad_norm": 0.4447080194950104,
"learning_rate": 4.243141146742905e-05,
"loss": 0.5953,
"step": 1720
},
{
"epoch": 7.2784810126582276,
"grad_norm": 0.4422175884246826,
"learning_rate": 4.183082393102636e-05,
"loss": 0.5849,
"step": 1725
},
{
"epoch": 7.29957805907173,
"grad_norm": 0.4476224184036255,
"learning_rate": 4.1233391028718116e-05,
"loss": 0.5962,
"step": 1730
},
{
"epoch": 7.320675105485232,
"grad_norm": 0.4534938931465149,
"learning_rate": 4.063914516051984e-05,
"loss": 0.5838,
"step": 1735
},
{
"epoch": 7.341772151898734,
"grad_norm": 0.45842060446739197,
"learning_rate": 4.004811855360748e-05,
"loss": 0.6046,
"step": 1740
},
{
"epoch": 7.362869198312236,
"grad_norm": 0.43340378999710083,
"learning_rate": 3.9460343260569964e-05,
"loss": 0.5972,
"step": 1745
},
{
"epoch": 7.383966244725738,
"grad_norm": 0.4477992057800293,
"learning_rate": 3.887585115767068e-05,
"loss": 0.6067,
"step": 1750
},
{
"epoch": 7.405063291139241,
"grad_norm": 0.44521939754486084,
"learning_rate": 3.82946739431189e-05,
"loss": 0.5959,
"step": 1755
},
{
"epoch": 7.4261603375527425,
"grad_norm": 0.42936068773269653,
"learning_rate": 3.771684313535062e-05,
"loss": 0.5963,
"step": 1760
},
{
"epoch": 7.447257383966245,
"grad_norm": 0.45330098271369934,
"learning_rate": 3.7142390071319454e-05,
"loss": 0.6001,
"step": 1765
},
{
"epoch": 7.468354430379747,
"grad_norm": 0.451648473739624,
"learning_rate": 3.65713459047969e-05,
"loss": 0.6104,
"step": 1770
},
{
"epoch": 7.489451476793249,
"grad_norm": 0.4406780004501343,
"learning_rate": 3.60037416046829e-05,
"loss": 0.5942,
"step": 1775
},
{
"epoch": 7.510548523206751,
"grad_norm": 0.4443998634815216,
"learning_rate": 3.543960795332653e-05,
"loss": 0.5919,
"step": 1780
},
{
"epoch": 7.531645569620253,
"grad_norm": 0.45104894042015076,
"learning_rate": 3.487897554485628e-05,
"loss": 0.5995,
"step": 1785
},
{
"epoch": 7.552742616033755,
"grad_norm": 0.45314210653305054,
"learning_rate": 3.43218747835211e-05,
"loss": 0.587,
"step": 1790
},
{
"epoch": 7.5738396624472575,
"grad_norm": 0.4450884163379669,
"learning_rate": 3.376833588204148e-05,
"loss": 0.5879,
"step": 1795
},
{
"epoch": 7.594936708860759,
"grad_norm": 0.44848042726516724,
"learning_rate": 3.3218388859970875e-05,
"loss": 0.598,
"step": 1800
},
{
"epoch": 7.616033755274262,
"grad_norm": 0.45061829686164856,
"learning_rate": 3.2672063542067734e-05,
"loss": 0.6111,
"step": 1805
},
{
"epoch": 7.637130801687764,
"grad_norm": 0.43524765968322754,
"learning_rate": 3.2129389556678016e-05,
"loss": 0.6004,
"step": 1810
},
{
"epoch": 7.658227848101266,
"grad_norm": 0.46142658591270447,
"learning_rate": 3.15903963341285e-05,
"loss": 0.594,
"step": 1815
},
{
"epoch": 7.679324894514768,
"grad_norm": 0.45434656739234924,
"learning_rate": 3.1055113105130506e-05,
"loss": 0.6002,
"step": 1820
},
{
"epoch": 7.70042194092827,
"grad_norm": 0.45925071835517883,
"learning_rate": 3.052356889919489e-05,
"loss": 0.5914,
"step": 1825
},
{
"epoch": 7.7215189873417724,
"grad_norm": 0.44464772939682007,
"learning_rate": 2.9995792543057478e-05,
"loss": 0.6064,
"step": 1830
},
{
"epoch": 7.742616033755274,
"grad_norm": 0.44117605686187744,
"learning_rate": 2.9471812659115917e-05,
"loss": 0.5993,
"step": 1835
},
{
"epoch": 7.763713080168777,
"grad_norm": 0.4454299509525299,
"learning_rate": 2.895165766387733e-05,
"loss": 0.5957,
"step": 1840
},
{
"epoch": 7.784810126582278,
"grad_norm": 0.4484660029411316,
"learning_rate": 2.843535576641725e-05,
"loss": 0.5985,
"step": 1845
},
{
"epoch": 7.805907172995781,
"grad_norm": 0.4591384828090668,
"learning_rate": 2.7922934966849823e-05,
"loss": 0.6044,
"step": 1850
},
{
"epoch": 7.827004219409282,
"grad_norm": 0.4372834861278534,
"learning_rate": 2.7414423054809302e-05,
"loss": 0.5958,
"step": 1855
},
{
"epoch": 7.848101265822785,
"grad_norm": 0.44407814741134644,
"learning_rate": 2.690984760794284e-05,
"loss": 0.5965,
"step": 1860
},
{
"epoch": 7.869198312236287,
"grad_norm": 0.44278717041015625,
"learning_rate": 2.6409235990415026e-05,
"loss": 0.6062,
"step": 1865
},
{
"epoch": 7.890295358649789,
"grad_norm": 0.4526854455471039,
"learning_rate": 2.591261535142383e-05,
"loss": 0.6035,
"step": 1870
},
{
"epoch": 7.911392405063291,
"grad_norm": 0.4361235499382019,
"learning_rate": 2.5420012623728208e-05,
"loss": 0.6041,
"step": 1875
},
{
"epoch": 7.932489451476793,
"grad_norm": 0.4319293200969696,
"learning_rate": 2.4931454522187593e-05,
"loss": 0.6005,
"step": 1880
},
{
"epoch": 7.953586497890296,
"grad_norm": 0.44515419006347656,
"learning_rate": 2.4446967542313015e-05,
"loss": 0.614,
"step": 1885
},
{
"epoch": 7.974683544303797,
"grad_norm": 0.4468868672847748,
"learning_rate": 2.3966577958830128e-05,
"loss": 0.5999,
"step": 1890
},
{
"epoch": 7.9957805907173,
"grad_norm": 0.43502089381217957,
"learning_rate": 2.3490311824254386e-05,
"loss": 0.591,
"step": 1895
},
{
"epoch": 8.0,
"eval_loss": 1.9239612817764282,
"eval_runtime": 0.555,
"eval_samples_per_second": 3.604,
"eval_steps_per_second": 1.802,
"step": 1896
},
{
"epoch": 8.016877637130802,
"grad_norm": 0.43088486790657043,
"learning_rate": 2.3018194967478145e-05,
"loss": 0.5772,
"step": 1900
},
{
"epoch": 8.037974683544304,
"grad_norm": 0.4887051582336426,
"learning_rate": 2.2550252992369837e-05,
"loss": 0.5858,
"step": 1905
},
{
"epoch": 8.059071729957806,
"grad_norm": 0.46031367778778076,
"learning_rate": 2.2086511276385556e-05,
"loss": 0.5698,
"step": 1910
},
{
"epoch": 8.080168776371307,
"grad_norm": 0.44916045665740967,
"learning_rate": 2.1626994969192617e-05,
"loss": 0.5832,
"step": 1915
},
{
"epoch": 8.10126582278481,
"grad_norm": 0.45516934990882874,
"learning_rate": 2.1171728991305795e-05,
"loss": 0.5678,
"step": 1920
},
{
"epoch": 8.122362869198312,
"grad_norm": 0.4672262668609619,
"learning_rate": 2.072073803273572e-05,
"loss": 0.5609,
"step": 1925
},
{
"epoch": 8.143459915611814,
"grad_norm": 0.4732205271720886,
"learning_rate": 2.0274046551649918e-05,
"loss": 0.5748,
"step": 1930
},
{
"epoch": 8.164556962025316,
"grad_norm": 0.4523015022277832,
"learning_rate": 1.9831678773046424e-05,
"loss": 0.572,
"step": 1935
},
{
"epoch": 8.185654008438819,
"grad_norm": 0.46077847480773926,
"learning_rate": 1.9393658687439985e-05,
"loss": 0.5734,
"step": 1940
},
{
"epoch": 8.20675105485232,
"grad_norm": 0.48243677616119385,
"learning_rate": 1.8960010049561028e-05,
"loss": 0.5749,
"step": 1945
},
{
"epoch": 8.227848101265822,
"grad_norm": 0.45998242497444153,
"learning_rate": 1.8530756377067394e-05,
"loss": 0.5635,
"step": 1950
},
{
"epoch": 8.248945147679326,
"grad_norm": 0.4926050305366516,
"learning_rate": 1.8105920949268862e-05,
"loss": 0.5656,
"step": 1955
},
{
"epoch": 8.270042194092827,
"grad_norm": 0.44752731919288635,
"learning_rate": 1.7685526805864727e-05,
"loss": 0.5713,
"step": 1960
},
{
"epoch": 8.291139240506329,
"grad_norm": 0.4773804843425751,
"learning_rate": 1.7269596745694295e-05,
"loss": 0.5753,
"step": 1965
},
{
"epoch": 8.31223628691983,
"grad_norm": 0.4709602892398834,
"learning_rate": 1.6858153325500435e-05,
"loss": 0.5604,
"step": 1970
},
{
"epoch": 8.333333333333334,
"grad_norm": 0.46927887201309204,
"learning_rate": 1.6451218858706374e-05,
"loss": 0.578,
"step": 1975
},
{
"epoch": 8.354430379746836,
"grad_norm": 0.467042475938797,
"learning_rate": 1.60488154142054e-05,
"loss": 0.5876,
"step": 1980
},
{
"epoch": 8.375527426160337,
"grad_norm": 0.4665224850177765,
"learning_rate": 1.565096481516427e-05,
"loss": 0.5727,
"step": 1985
},
{
"epoch": 8.396624472573839,
"grad_norm": 0.46915140748023987,
"learning_rate": 1.5257688637839484e-05,
"loss": 0.5744,
"step": 1990
},
{
"epoch": 8.417721518987342,
"grad_norm": 0.4630158841609955,
"learning_rate": 1.4869008210407243e-05,
"loss": 0.5609,
"step": 1995
},
{
"epoch": 8.438818565400844,
"grad_norm": 0.46690833568573,
"learning_rate": 1.4484944611806773e-05,
"loss": 0.5764,
"step": 2000
},
{
"epoch": 8.459915611814345,
"grad_norm": 0.46290239691734314,
"learning_rate": 1.410551867059724e-05,
"loss": 0.5817,
"step": 2005
},
{
"epoch": 8.481012658227849,
"grad_norm": 0.4760874807834625,
"learning_rate": 1.3730750963828032e-05,
"loss": 0.5704,
"step": 2010
},
{
"epoch": 8.50210970464135,
"grad_norm": 0.47287824749946594,
"learning_rate": 1.3360661815922903e-05,
"loss": 0.574,
"step": 2015
},
{
"epoch": 8.523206751054852,
"grad_norm": 0.46598172187805176,
"learning_rate": 1.2995271297577816e-05,
"loss": 0.5792,
"step": 2020
},
{
"epoch": 8.544303797468354,
"grad_norm": 0.4603840410709381,
"learning_rate": 1.2634599224672294e-05,
"loss": 0.5674,
"step": 2025
},
{
"epoch": 8.565400843881857,
"grad_norm": 0.48355668783187866,
"learning_rate": 1.227866515719489e-05,
"loss": 0.5676,
"step": 2030
},
{
"epoch": 8.586497890295359,
"grad_norm": 0.4648090898990631,
"learning_rate": 1.1927488398182395e-05,
"loss": 0.5595,
"step": 2035
},
{
"epoch": 8.60759493670886,
"grad_norm": 0.47335174679756165,
"learning_rate": 1.1581087992672935e-05,
"loss": 0.5743,
"step": 2040
},
{
"epoch": 8.628691983122362,
"grad_norm": 0.46940383315086365,
"learning_rate": 1.1239482726673201e-05,
"loss": 0.5719,
"step": 2045
},
{
"epoch": 8.649789029535865,
"grad_norm": 0.46548011898994446,
"learning_rate": 1.0902691126139542e-05,
"loss": 0.5722,
"step": 2050
},
{
"epoch": 8.670886075949367,
"grad_norm": 0.4601798355579376,
"learning_rate": 1.0570731455973414e-05,
"loss": 0.5752,
"step": 2055
},
{
"epoch": 8.691983122362869,
"grad_norm": 0.47531968355178833,
"learning_rate": 1.024362171903065e-05,
"loss": 0.5833,
"step": 2060
},
{
"epoch": 8.713080168776372,
"grad_norm": 0.46817246079444885,
"learning_rate": 9.921379655145313e-06,
"loss": 0.5716,
"step": 2065
},
{
"epoch": 8.734177215189874,
"grad_norm": 0.47469767928123474,
"learning_rate": 9.604022740167495e-06,
"loss": 0.5825,
"step": 2070
},
{
"epoch": 8.755274261603375,
"grad_norm": 0.4739144444465637,
"learning_rate": 9.29156818501561e-06,
"loss": 0.5669,
"step": 2075
},
{
"epoch": 8.776371308016877,
"grad_norm": 0.5070598721504211,
"learning_rate": 8.984032934743026e-06,
"loss": 0.5797,
"step": 2080
},
{
"epoch": 8.79746835443038,
"grad_norm": 0.4720567464828491,
"learning_rate": 8.681433667619065e-06,
"loss": 0.5635,
"step": 2085
},
{
"epoch": 8.818565400843882,
"grad_norm": 0.45980048179626465,
"learning_rate": 8.383786794224569e-06,
"loss": 0.5715,
"step": 2090
},
{
"epoch": 8.839662447257384,
"grad_norm": 0.4796925187110901,
"learning_rate": 8.09110845656187e-06,
"loss": 0.5785,
"step": 2095
},
{
"epoch": 8.860759493670885,
"grad_norm": 0.4801785349845886,
"learning_rate": 7.803414527179343e-06,
"loss": 0.5772,
"step": 2100
},
{
"epoch": 8.881856540084389,
"grad_norm": 0.4780319631099701,
"learning_rate": 7.520720608310683e-06,
"loss": 0.5726,
"step": 2105
},
{
"epoch": 8.90295358649789,
"grad_norm": 0.4621387720108032,
"learning_rate": 7.243042031028713e-06,
"loss": 0.5752,
"step": 2110
},
{
"epoch": 8.924050632911392,
"grad_norm": 0.4708462059497833,
"learning_rate": 6.9703938544139706e-06,
"loss": 0.5716,
"step": 2115
},
{
"epoch": 8.945147679324894,
"grad_norm": 0.4696125090122223,
"learning_rate": 6.702790864738018e-06,
"loss": 0.5666,
"step": 2120
},
{
"epoch": 8.966244725738397,
"grad_norm": 0.47694242000579834,
"learning_rate": 6.440247574661573e-06,
"loss": 0.568,
"step": 2125
},
{
"epoch": 8.987341772151899,
"grad_norm": 0.4875074326992035,
"learning_rate": 6.182778222447383e-06,
"loss": 0.5677,
"step": 2130
},
{
"epoch": 9.0,
"eval_loss": 1.984204888343811,
"eval_runtime": 0.5545,
"eval_samples_per_second": 3.607,
"eval_steps_per_second": 1.804,
"step": 2133
},
{
"epoch": 9.0084388185654,
"grad_norm": 0.4417002499103546,
"learning_rate": 5.930396771188129e-06,
"loss": 0.566,
"step": 2135
},
{
"epoch": 9.029535864978904,
"grad_norm": 0.4602307677268982,
"learning_rate": 5.683116908049168e-06,
"loss": 0.5625,
"step": 2140
},
{
"epoch": 9.050632911392405,
"grad_norm": 0.4665865898132324,
"learning_rate": 5.440952043526215e-06,
"loss": 0.5584,
"step": 2145
},
{
"epoch": 9.071729957805907,
"grad_norm": 0.47397249937057495,
"learning_rate": 5.203915310718099e-06,
"loss": 0.558,
"step": 2150
},
{
"epoch": 9.092827004219409,
"grad_norm": 0.47351840138435364,
"learning_rate": 4.972019564614539e-06,
"loss": 0.5516,
"step": 2155
},
{
"epoch": 9.113924050632912,
"grad_norm": 0.4746864438056946,
"learning_rate": 4.745277381398938e-06,
"loss": 0.5536,
"step": 2160
},
{
"epoch": 9.135021097046414,
"grad_norm": 0.4737743139266968,
"learning_rate": 4.523701057766361e-06,
"loss": 0.5577,
"step": 2165
},
{
"epoch": 9.156118143459915,
"grad_norm": 0.4778996706008911,
"learning_rate": 4.307302610256736e-06,
"loss": 0.5541,
"step": 2170
},
{
"epoch": 9.177215189873417,
"grad_norm": 0.4771966338157654,
"learning_rate": 4.0960937746030605e-06,
"loss": 0.552,
"step": 2175
},
{
"epoch": 9.19831223628692,
"grad_norm": 0.4708782732486725,
"learning_rate": 3.890086005095051e-06,
"loss": 0.5515,
"step": 2180
},
{
"epoch": 9.219409282700422,
"grad_norm": 0.49065274000167847,
"learning_rate": 3.6892904739578736e-06,
"loss": 0.5593,
"step": 2185
},
{
"epoch": 9.240506329113924,
"grad_norm": 0.47753557562828064,
"learning_rate": 3.493718070746299e-06,
"loss": 0.5558,
"step": 2190
},
{
"epoch": 9.261603375527427,
"grad_norm": 0.4750811755657196,
"learning_rate": 3.3033794017541254e-06,
"loss": 0.5588,
"step": 2195
},
{
"epoch": 9.282700421940929,
"grad_norm": 0.46534910798072815,
"learning_rate": 3.1182847894389634e-06,
"loss": 0.5567,
"step": 2200
},
{
"epoch": 9.30379746835443,
"grad_norm": 0.46732398867607117,
"learning_rate": 2.9384442718624395e-06,
"loss": 0.5712,
"step": 2205
},
{
"epoch": 9.324894514767932,
"grad_norm": 0.4703245759010315,
"learning_rate": 2.763867602145842e-06,
"loss": 0.5566,
"step": 2210
},
{
"epoch": 9.345991561181435,
"grad_norm": 0.4757389426231384,
"learning_rate": 2.5945642479411448e-06,
"loss": 0.5669,
"step": 2215
},
{
"epoch": 9.367088607594937,
"grad_norm": 0.49615395069122314,
"learning_rate": 2.430543390917539e-06,
"loss": 0.5771,
"step": 2220
},
{
"epoch": 9.388185654008439,
"grad_norm": 0.4786005914211273,
"learning_rate": 2.2718139262635775e-06,
"loss": 0.5581,
"step": 2225
},
{
"epoch": 9.40928270042194,
"grad_norm": 0.4749692380428314,
"learning_rate": 2.1183844622047034e-06,
"loss": 0.5566,
"step": 2230
},
{
"epoch": 9.430379746835444,
"grad_norm": 0.4892341196537018,
"learning_rate": 1.9702633195363917e-06,
"loss": 0.5577,
"step": 2235
},
{
"epoch": 9.451476793248945,
"grad_norm": 0.485775351524353,
"learning_rate": 1.8274585311729653e-06,
"loss": 0.5724,
"step": 2240
},
{
"epoch": 9.472573839662447,
"grad_norm": 0.47523242235183716,
"learning_rate": 1.6899778417118983e-06,
"loss": 0.5472,
"step": 2245
},
{
"epoch": 9.49367088607595,
"grad_norm": 0.45913201570510864,
"learning_rate": 1.557828707013831e-06,
"loss": 0.5576,
"step": 2250
},
{
"epoch": 9.514767932489452,
"grad_norm": 0.4752641022205353,
"learning_rate": 1.4310182937982141e-06,
"loss": 0.5605,
"step": 2255
},
{
"epoch": 9.535864978902953,
"grad_norm": 0.47292277216911316,
"learning_rate": 1.309553479254666e-06,
"loss": 0.5653,
"step": 2260
},
{
"epoch": 9.556962025316455,
"grad_norm": 0.4652714133262634,
"learning_rate": 1.1934408506699802e-06,
"loss": 0.5571,
"step": 2265
},
{
"epoch": 9.578059071729959,
"grad_norm": 0.4648183584213257,
"learning_rate": 1.0826867050708678e-06,
"loss": 0.5603,
"step": 2270
},
{
"epoch": 9.59915611814346,
"grad_norm": 0.4922165274620056,
"learning_rate": 9.772970488825417e-07,
"loss": 0.5627,
"step": 2275
},
{
"epoch": 9.620253164556962,
"grad_norm": 0.4720841348171234,
"learning_rate": 8.772775976028546e-07,
"loss": 0.5517,
"step": 2280
},
{
"epoch": 9.641350210970463,
"grad_norm": 0.510443389415741,
"learning_rate": 7.826337754924473e-07,
"loss": 0.5641,
"step": 2285
},
{
"epoch": 9.662447257383967,
"grad_norm": 0.46088987588882446,
"learning_rate": 6.933707152805058e-07,
"loss": 0.5595,
"step": 2290
},
{
"epoch": 9.683544303797468,
"grad_norm": 0.48828017711639404,
"learning_rate": 6.094932578864287e-07,
"loss": 0.565,
"step": 2295
},
{
"epoch": 9.70464135021097,
"grad_norm": 0.4722835421562195,
"learning_rate": 5.31005952157304e-07,
"loss": 0.5611,
"step": 2300
},
{
"epoch": 9.725738396624472,
"grad_norm": 0.4879083037376404,
"learning_rate": 4.5791305462120625e-07,
"loss": 0.5746,
"step": 2305
},
{
"epoch": 9.746835443037975,
"grad_norm": 0.4783223271369934,
"learning_rate": 3.902185292563365e-07,
"loss": 0.569,
"step": 2310
},
{
"epoch": 9.767932489451477,
"grad_norm": 0.4648023247718811,
"learning_rate": 3.2792604727608367e-07,
"loss": 0.5503,
"step": 2315
},
{
"epoch": 9.789029535864978,
"grad_norm": 0.47712016105651855,
"learning_rate": 2.710389869298946e-07,
"loss": 0.5522,
"step": 2320
},
{
"epoch": 9.810126582278482,
"grad_norm": 0.4652068614959717,
"learning_rate": 2.1956043332010955e-07,
"loss": 0.556,
"step": 2325
},
{
"epoch": 9.831223628691983,
"grad_norm": 0.48310577869415283,
"learning_rate": 1.7349317823459609e-07,
"loss": 0.5637,
"step": 2330
},
{
"epoch": 9.852320675105485,
"grad_norm": 0.5295414924621582,
"learning_rate": 1.3283971999537015e-07,
"loss": 0.5559,
"step": 2335
},
{
"epoch": 9.873417721518987,
"grad_norm": 0.4798893928527832,
"learning_rate": 9.76022633231155e-08,
"loss": 0.5543,
"step": 2340
},
{
"epoch": 9.89451476793249,
"grad_norm": 0.48038187623023987,
"learning_rate": 6.778271921760171e-08,
"loss": 0.5626,
"step": 2345
},
{
"epoch": 9.915611814345992,
"grad_norm": 0.47445496916770935,
"learning_rate": 4.338270485405582e-08,
"loss": 0.5545,
"step": 2350
},
{
"epoch": 9.936708860759493,
"grad_norm": 0.47432997822761536,
"learning_rate": 2.4403543495454818e-08,
"loss": 0.5651,
"step": 2355
},
{
"epoch": 9.957805907172995,
"grad_norm": 0.4740554094314575,
"learning_rate": 1.0846264420771857e-08,
"loss": 0.5637,
"step": 2360
},
{
"epoch": 9.978902953586498,
"grad_norm": 0.4762984812259674,
"learning_rate": 2.7116028691431817e-09,
"loss": 0.5611,
"step": 2365
},
{
"epoch": 10.0,
"grad_norm": 0.4564709961414337,
"learning_rate": 0.0,
"loss": 0.5648,
"step": 2370
},
{
"epoch": 10.0,
"eval_loss": 2.0032131671905518,
"eval_runtime": 0.5807,
"eval_samples_per_second": 3.444,
"eval_steps_per_second": 1.722,
"step": 2370
},
{
"epoch": 10.0,
"step": 2370,
"total_flos": 3.5097090775444357e+18,
"train_loss": 0.7390849222110797,
"train_runtime": 8188.9555,
"train_samples_per_second": 9.243,
"train_steps_per_second": 0.289
}
],
"logging_steps": 5,
"max_steps": 2370,
"num_input_tokens_seen": 0,
"num_train_epochs": 10,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 3.5097090775444357e+18,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}