chansung's picture
Model save
d687c01 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 9.979633401221996,
"eval_steps": 500,
"global_step": 2450,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.004073319755600814,
"grad_norm": 12.219289779663086,
"learning_rate": 8.16326530612245e-07,
"loss": 2.0786,
"step": 1
},
{
"epoch": 0.020366598778004074,
"grad_norm": 12.530974388122559,
"learning_rate": 4.081632653061224e-06,
"loss": 2.0708,
"step": 5
},
{
"epoch": 0.04073319755600815,
"grad_norm": 10.991973876953125,
"learning_rate": 8.163265306122448e-06,
"loss": 1.9589,
"step": 10
},
{
"epoch": 0.06109979633401222,
"grad_norm": 11.441458702087402,
"learning_rate": 1.2244897959183674e-05,
"loss": 1.6878,
"step": 15
},
{
"epoch": 0.0814663951120163,
"grad_norm": 4.176149845123291,
"learning_rate": 1.6326530612244897e-05,
"loss": 1.3224,
"step": 20
},
{
"epoch": 0.10183299389002037,
"grad_norm": 2.458606004714966,
"learning_rate": 2.0408163265306123e-05,
"loss": 1.1912,
"step": 25
},
{
"epoch": 0.12219959266802444,
"grad_norm": 1.30384361743927,
"learning_rate": 2.448979591836735e-05,
"loss": 1.1123,
"step": 30
},
{
"epoch": 0.1425661914460285,
"grad_norm": 1.0811594724655151,
"learning_rate": 2.857142857142857e-05,
"loss": 1.0462,
"step": 35
},
{
"epoch": 0.1629327902240326,
"grad_norm": 0.7156168222427368,
"learning_rate": 3.265306122448979e-05,
"loss": 0.992,
"step": 40
},
{
"epoch": 0.18329938900203666,
"grad_norm": 0.6059057116508484,
"learning_rate": 3.673469387755102e-05,
"loss": 0.9393,
"step": 45
},
{
"epoch": 0.20366598778004075,
"grad_norm": 0.5178468227386475,
"learning_rate": 4.0816326530612245e-05,
"loss": 0.8967,
"step": 50
},
{
"epoch": 0.2240325865580448,
"grad_norm": 0.45643067359924316,
"learning_rate": 4.4897959183673474e-05,
"loss": 0.8665,
"step": 55
},
{
"epoch": 0.24439918533604887,
"grad_norm": 0.4627813994884491,
"learning_rate": 4.89795918367347e-05,
"loss": 0.8554,
"step": 60
},
{
"epoch": 0.26476578411405294,
"grad_norm": 0.7451216578483582,
"learning_rate": 5.3061224489795926e-05,
"loss": 0.8335,
"step": 65
},
{
"epoch": 0.285132382892057,
"grad_norm": 0.4662891924381256,
"learning_rate": 5.714285714285714e-05,
"loss": 0.827,
"step": 70
},
{
"epoch": 0.3054989816700611,
"grad_norm": 0.42819878458976746,
"learning_rate": 6.122448979591838e-05,
"loss": 0.816,
"step": 75
},
{
"epoch": 0.3258655804480652,
"grad_norm": 0.4619337022304535,
"learning_rate": 6.530612244897959e-05,
"loss": 0.7968,
"step": 80
},
{
"epoch": 0.34623217922606925,
"grad_norm": 0.4922891855239868,
"learning_rate": 6.938775510204082e-05,
"loss": 0.7954,
"step": 85
},
{
"epoch": 0.3665987780040733,
"grad_norm": 0.5278509259223938,
"learning_rate": 7.346938775510205e-05,
"loss": 0.7953,
"step": 90
},
{
"epoch": 0.3869653767820774,
"grad_norm": 0.912343442440033,
"learning_rate": 7.755102040816327e-05,
"loss": 0.7773,
"step": 95
},
{
"epoch": 0.4073319755600815,
"grad_norm": 0.5746473073959351,
"learning_rate": 8.163265306122449e-05,
"loss": 0.775,
"step": 100
},
{
"epoch": 0.42769857433808556,
"grad_norm": 0.4901811480522156,
"learning_rate": 8.571428571428571e-05,
"loss": 0.7708,
"step": 105
},
{
"epoch": 0.4480651731160896,
"grad_norm": 0.596023678779602,
"learning_rate": 8.979591836734695e-05,
"loss": 0.761,
"step": 110
},
{
"epoch": 0.4684317718940937,
"grad_norm": 0.9318987131118774,
"learning_rate": 9.387755102040817e-05,
"loss": 0.7572,
"step": 115
},
{
"epoch": 0.48879837067209775,
"grad_norm": 0.648429811000824,
"learning_rate": 9.79591836734694e-05,
"loss": 0.7652,
"step": 120
},
{
"epoch": 0.5091649694501018,
"grad_norm": 0.4797033369541168,
"learning_rate": 0.00010204081632653062,
"loss": 0.7563,
"step": 125
},
{
"epoch": 0.5295315682281059,
"grad_norm": 0.9584823250770569,
"learning_rate": 0.00010612244897959185,
"loss": 0.7532,
"step": 130
},
{
"epoch": 0.5498981670061099,
"grad_norm": 0.6396449208259583,
"learning_rate": 0.00011020408163265306,
"loss": 0.7438,
"step": 135
},
{
"epoch": 0.570264765784114,
"grad_norm": 0.8127761483192444,
"learning_rate": 0.00011428571428571428,
"loss": 0.7367,
"step": 140
},
{
"epoch": 0.5906313645621182,
"grad_norm": 0.6745628118515015,
"learning_rate": 0.00011836734693877552,
"loss": 0.7284,
"step": 145
},
{
"epoch": 0.6109979633401222,
"grad_norm": 0.5591239929199219,
"learning_rate": 0.00012244897959183676,
"loss": 0.7337,
"step": 150
},
{
"epoch": 0.6313645621181263,
"grad_norm": 0.7267659902572632,
"learning_rate": 0.00012653061224489798,
"loss": 0.7264,
"step": 155
},
{
"epoch": 0.6517311608961304,
"grad_norm": 0.5548218488693237,
"learning_rate": 0.00013061224489795917,
"loss": 0.7301,
"step": 160
},
{
"epoch": 0.6720977596741344,
"grad_norm": 0.9011363387107849,
"learning_rate": 0.0001346938775510204,
"loss": 0.72,
"step": 165
},
{
"epoch": 0.6924643584521385,
"grad_norm": 0.5213197469711304,
"learning_rate": 0.00013877551020408165,
"loss": 0.7243,
"step": 170
},
{
"epoch": 0.7128309572301426,
"grad_norm": 0.476764440536499,
"learning_rate": 0.00014285714285714287,
"loss": 0.7198,
"step": 175
},
{
"epoch": 0.7331975560081466,
"grad_norm": 0.7126782536506653,
"learning_rate": 0.0001469387755102041,
"loss": 0.7177,
"step": 180
},
{
"epoch": 0.7535641547861507,
"grad_norm": 0.4389437735080719,
"learning_rate": 0.0001510204081632653,
"loss": 0.7143,
"step": 185
},
{
"epoch": 0.7739307535641547,
"grad_norm": 0.5171541571617126,
"learning_rate": 0.00015510204081632654,
"loss": 0.714,
"step": 190
},
{
"epoch": 0.7942973523421588,
"grad_norm": 0.5747414827346802,
"learning_rate": 0.00015918367346938776,
"loss": 0.7192,
"step": 195
},
{
"epoch": 0.814663951120163,
"grad_norm": 0.703008770942688,
"learning_rate": 0.00016326530612244898,
"loss": 0.7094,
"step": 200
},
{
"epoch": 0.835030549898167,
"grad_norm": 0.41933196783065796,
"learning_rate": 0.00016734693877551023,
"loss": 0.7027,
"step": 205
},
{
"epoch": 0.8553971486761711,
"grad_norm": 0.8496442437171936,
"learning_rate": 0.00017142857142857143,
"loss": 0.7086,
"step": 210
},
{
"epoch": 0.8757637474541752,
"grad_norm": 0.44953274726867676,
"learning_rate": 0.00017551020408163265,
"loss": 0.7104,
"step": 215
},
{
"epoch": 0.8961303462321792,
"grad_norm": 0.49215880036354065,
"learning_rate": 0.0001795918367346939,
"loss": 0.6998,
"step": 220
},
{
"epoch": 0.9164969450101833,
"grad_norm": 0.4515603482723236,
"learning_rate": 0.00018367346938775512,
"loss": 0.6994,
"step": 225
},
{
"epoch": 0.9368635437881874,
"grad_norm": 0.48363256454467773,
"learning_rate": 0.00018775510204081634,
"loss": 0.697,
"step": 230
},
{
"epoch": 0.9572301425661914,
"grad_norm": 0.504764974117279,
"learning_rate": 0.00019183673469387756,
"loss": 0.6937,
"step": 235
},
{
"epoch": 0.9775967413441955,
"grad_norm": 0.6177113056182861,
"learning_rate": 0.0001959183673469388,
"loss": 0.702,
"step": 240
},
{
"epoch": 0.9979633401221996,
"grad_norm": 0.6133261919021606,
"learning_rate": 0.0002,
"loss": 0.6982,
"step": 245
},
{
"epoch": 0.9979633401221996,
"eval_loss": 1.8247634172439575,
"eval_runtime": 0.3801,
"eval_samples_per_second": 31.574,
"eval_steps_per_second": 2.631,
"step": 245
},
{
"epoch": 1.0183299389002036,
"grad_norm": 0.6387724876403809,
"learning_rate": 0.00019999746258949147,
"loss": 0.6786,
"step": 250
},
{
"epoch": 1.0386965376782078,
"grad_norm": 0.4179505705833435,
"learning_rate": 0.00019998985048673486,
"loss": 0.6806,
"step": 255
},
{
"epoch": 1.0590631364562118,
"grad_norm": 0.7707039713859558,
"learning_rate": 0.0001999771640780308,
"loss": 0.6828,
"step": 260
},
{
"epoch": 1.079429735234216,
"grad_norm": 0.735935628414154,
"learning_rate": 0.0001999594040071918,
"loss": 0.6832,
"step": 265
},
{
"epoch": 1.0997963340122199,
"grad_norm": 0.6434056758880615,
"learning_rate": 0.00019993657117550973,
"loss": 0.681,
"step": 270
},
{
"epoch": 1.120162932790224,
"grad_norm": 0.7113597393035889,
"learning_rate": 0.00019990866674170983,
"loss": 0.6735,
"step": 275
},
{
"epoch": 1.140529531568228,
"grad_norm": 0.7098330855369568,
"learning_rate": 0.00019987569212189224,
"loss": 0.6786,
"step": 280
},
{
"epoch": 1.1608961303462322,
"grad_norm": 0.38251993060112,
"learning_rate": 0.0001998376489894599,
"loss": 0.6794,
"step": 285
},
{
"epoch": 1.1812627291242364,
"grad_norm": 0.6434310674667358,
"learning_rate": 0.00019979453927503364,
"loss": 0.6846,
"step": 290
},
{
"epoch": 1.2016293279022403,
"grad_norm": 0.5492526292800903,
"learning_rate": 0.00019974636516635434,
"loss": 0.6697,
"step": 295
},
{
"epoch": 1.2219959266802445,
"grad_norm": 0.48028430342674255,
"learning_rate": 0.00019969312910817183,
"loss": 0.6778,
"step": 300
},
{
"epoch": 1.2423625254582484,
"grad_norm": 0.35454341769218445,
"learning_rate": 0.00019963483380212068,
"loss": 0.6671,
"step": 305
},
{
"epoch": 1.2627291242362526,
"grad_norm": 0.3555687665939331,
"learning_rate": 0.00019957148220658345,
"loss": 0.6646,
"step": 310
},
{
"epoch": 1.2830957230142566,
"grad_norm": 0.5027297735214233,
"learning_rate": 0.00019950307753654017,
"loss": 0.6723,
"step": 315
},
{
"epoch": 1.3034623217922607,
"grad_norm": 0.34120357036590576,
"learning_rate": 0.00019942962326340537,
"loss": 0.6669,
"step": 320
},
{
"epoch": 1.3238289205702647,
"grad_norm": 0.3805827498435974,
"learning_rate": 0.000199351123114852,
"loss": 0.6682,
"step": 325
},
{
"epoch": 1.3441955193482689,
"grad_norm": 0.4035264551639557,
"learning_rate": 0.00019926758107462206,
"loss": 0.6665,
"step": 330
},
{
"epoch": 1.364562118126273,
"grad_norm": 0.43114498257637024,
"learning_rate": 0.0001991790013823246,
"loss": 0.6687,
"step": 335
},
{
"epoch": 1.384928716904277,
"grad_norm": 0.43574780225753784,
"learning_rate": 0.00019908538853322048,
"loss": 0.6648,
"step": 340
},
{
"epoch": 1.405295315682281,
"grad_norm": 0.36844050884246826,
"learning_rate": 0.00019898674727799419,
"loss": 0.6635,
"step": 345
},
{
"epoch": 1.4256619144602851,
"grad_norm": 0.360999196767807,
"learning_rate": 0.00019888308262251285,
"loss": 0.6715,
"step": 350
},
{
"epoch": 1.4460285132382893,
"grad_norm": 0.47190627455711365,
"learning_rate": 0.0001987743998275723,
"loss": 0.6595,
"step": 355
},
{
"epoch": 1.4663951120162932,
"grad_norm": 0.4380181133747101,
"learning_rate": 0.00019866070440862976,
"loss": 0.6589,
"step": 360
},
{
"epoch": 1.4867617107942974,
"grad_norm": 0.3873290419578552,
"learning_rate": 0.00019854200213552424,
"loss": 0.6649,
"step": 365
},
{
"epoch": 1.5071283095723014,
"grad_norm": 0.4168381094932556,
"learning_rate": 0.00019841829903218376,
"loss": 0.6673,
"step": 370
},
{
"epoch": 1.5274949083503055,
"grad_norm": 0.4359332025051117,
"learning_rate": 0.00019828960137631928,
"loss": 0.6666,
"step": 375
},
{
"epoch": 1.5478615071283097,
"grad_norm": 0.33967724442481995,
"learning_rate": 0.00019815591569910654,
"loss": 0.6663,
"step": 380
},
{
"epoch": 1.5682281059063137,
"grad_norm": 0.33921679854393005,
"learning_rate": 0.00019801724878485438,
"loss": 0.6557,
"step": 385
},
{
"epoch": 1.5885947046843176,
"grad_norm": 0.3741433024406433,
"learning_rate": 0.00019787360767066053,
"loss": 0.6522,
"step": 390
},
{
"epoch": 1.6089613034623218,
"grad_norm": 0.3441908359527588,
"learning_rate": 0.0001977249996460544,
"loss": 0.653,
"step": 395
},
{
"epoch": 1.629327902240326,
"grad_norm": 0.36859628558158875,
"learning_rate": 0.00019757143225262728,
"loss": 0.6582,
"step": 400
},
{
"epoch": 1.64969450101833,
"grad_norm": 0.3605320453643799,
"learning_rate": 0.00019741291328364955,
"loss": 0.6546,
"step": 405
},
{
"epoch": 1.6700610997963339,
"grad_norm": 0.6110367178916931,
"learning_rate": 0.00019724945078367513,
"loss": 0.6567,
"step": 410
},
{
"epoch": 1.690427698574338,
"grad_norm": 0.3475343585014343,
"learning_rate": 0.00019708105304813332,
"loss": 0.6658,
"step": 415
},
{
"epoch": 1.7107942973523422,
"grad_norm": 0.45948854088783264,
"learning_rate": 0.0001969077286229078,
"loss": 0.6584,
"step": 420
},
{
"epoch": 1.7311608961303462,
"grad_norm": 0.31532639265060425,
"learning_rate": 0.00019672948630390294,
"loss": 0.6579,
"step": 425
},
{
"epoch": 1.7515274949083504,
"grad_norm": 0.3328857421875,
"learning_rate": 0.00019654633513659743,
"loss": 0.6583,
"step": 430
},
{
"epoch": 1.7718940936863543,
"grad_norm": 0.3420257866382599,
"learning_rate": 0.00019635828441558515,
"loss": 0.6493,
"step": 435
},
{
"epoch": 1.7922606924643585,
"grad_norm": 0.31616052985191345,
"learning_rate": 0.00019616534368410365,
"loss": 0.6536,
"step": 440
},
{
"epoch": 1.8126272912423627,
"grad_norm": 0.30928483605384827,
"learning_rate": 0.0001959675227335497,
"loss": 0.6583,
"step": 445
},
{
"epoch": 1.8329938900203666,
"grad_norm": 0.3872964084148407,
"learning_rate": 0.00019576483160298246,
"loss": 0.6467,
"step": 450
},
{
"epoch": 1.8533604887983706,
"grad_norm": 0.3449136018753052,
"learning_rate": 0.0001955572805786141,
"loss": 0.6513,
"step": 455
},
{
"epoch": 1.8737270875763747,
"grad_norm": 0.3003346621990204,
"learning_rate": 0.00019534488019328752,
"loss": 0.6542,
"step": 460
},
{
"epoch": 1.894093686354379,
"grad_norm": 0.3554345965385437,
"learning_rate": 0.0001951276412259422,
"loss": 0.6496,
"step": 465
},
{
"epoch": 1.9144602851323829,
"grad_norm": 0.3109475374221802,
"learning_rate": 0.00019490557470106686,
"loss": 0.6544,
"step": 470
},
{
"epoch": 1.9348268839103868,
"grad_norm": 0.35385647416114807,
"learning_rate": 0.00019467869188814023,
"loss": 0.6456,
"step": 475
},
{
"epoch": 1.955193482688391,
"grad_norm": 0.3411131203174591,
"learning_rate": 0.00019444700430105893,
"loss": 0.6478,
"step": 480
},
{
"epoch": 1.9755600814663952,
"grad_norm": 0.3149469792842865,
"learning_rate": 0.00019421052369755334,
"loss": 0.6541,
"step": 485
},
{
"epoch": 1.9959266802443993,
"grad_norm": 0.4090039134025574,
"learning_rate": 0.00019396926207859084,
"loss": 0.6596,
"step": 490
},
{
"epoch": 2.0,
"eval_loss": 1.833816647529602,
"eval_runtime": 0.3787,
"eval_samples_per_second": 31.686,
"eval_steps_per_second": 2.641,
"step": 491
},
{
"epoch": 2.016293279022403,
"grad_norm": 0.4153566062450409,
"learning_rate": 0.0001937232316877668,
"loss": 0.6285,
"step": 495
},
{
"epoch": 2.0366598778004072,
"grad_norm": 0.41273799538612366,
"learning_rate": 0.00019347244501068312,
"loss": 0.6309,
"step": 500
},
{
"epoch": 2.0570264765784114,
"grad_norm": 0.40306323766708374,
"learning_rate": 0.00019321691477431485,
"loss": 0.6294,
"step": 505
},
{
"epoch": 2.0773930753564156,
"grad_norm": 0.37823185324668884,
"learning_rate": 0.00019295665394636414,
"loss": 0.6333,
"step": 510
},
{
"epoch": 2.0977596741344193,
"grad_norm": 0.31898233294487,
"learning_rate": 0.0001926916757346022,
"loss": 0.6217,
"step": 515
},
{
"epoch": 2.1181262729124235,
"grad_norm": 0.3543696105480194,
"learning_rate": 0.00019242199358619893,
"loss": 0.6268,
"step": 520
},
{
"epoch": 2.1384928716904277,
"grad_norm": 0.4415835738182068,
"learning_rate": 0.00019214762118704076,
"loss": 0.6252,
"step": 525
},
{
"epoch": 2.158859470468432,
"grad_norm": 0.40828588604927063,
"learning_rate": 0.00019186857246103584,
"loss": 0.624,
"step": 530
},
{
"epoch": 2.179226069246436,
"grad_norm": 0.37648454308509827,
"learning_rate": 0.00019158486156940757,
"loss": 0.6228,
"step": 535
},
{
"epoch": 2.1995926680244398,
"grad_norm": 0.3546181321144104,
"learning_rate": 0.0001912965029099759,
"loss": 0.62,
"step": 540
},
{
"epoch": 2.219959266802444,
"grad_norm": 0.3437698185443878,
"learning_rate": 0.00019100351111642666,
"loss": 0.6228,
"step": 545
},
{
"epoch": 2.240325865580448,
"grad_norm": 0.3919028639793396,
"learning_rate": 0.000190705901057569,
"loss": 0.623,
"step": 550
},
{
"epoch": 2.2606924643584523,
"grad_norm": 0.36673861742019653,
"learning_rate": 0.00019040368783658073,
"loss": 0.6222,
"step": 555
},
{
"epoch": 2.281059063136456,
"grad_norm": 0.3100906312465668,
"learning_rate": 0.0001900968867902419,
"loss": 0.6276,
"step": 560
},
{
"epoch": 2.30142566191446,
"grad_norm": 0.33423855900764465,
"learning_rate": 0.00018978551348815652,
"loss": 0.6304,
"step": 565
},
{
"epoch": 2.3217922606924644,
"grad_norm": 0.3217325210571289,
"learning_rate": 0.00018946958373196231,
"loss": 0.6255,
"step": 570
},
{
"epoch": 2.3421588594704685,
"grad_norm": 0.32845550775527954,
"learning_rate": 0.00018914911355452895,
"loss": 0.6276,
"step": 575
},
{
"epoch": 2.3625254582484727,
"grad_norm": 0.3502614498138428,
"learning_rate": 0.0001888241192191444,
"loss": 0.6345,
"step": 580
},
{
"epoch": 2.3828920570264764,
"grad_norm": 0.3093807101249695,
"learning_rate": 0.00018849461721868945,
"loss": 0.6231,
"step": 585
},
{
"epoch": 2.4032586558044806,
"grad_norm": 0.34325292706489563,
"learning_rate": 0.0001881606242748009,
"loss": 0.6356,
"step": 590
},
{
"epoch": 2.423625254582485,
"grad_norm": 0.3460986018180847,
"learning_rate": 0.00018782215733702286,
"loss": 0.6228,
"step": 595
},
{
"epoch": 2.443991853360489,
"grad_norm": 0.31046947836875916,
"learning_rate": 0.00018747923358194662,
"loss": 0.6264,
"step": 600
},
{
"epoch": 2.4643584521384927,
"grad_norm": 0.318236380815506,
"learning_rate": 0.00018713187041233896,
"loss": 0.6251,
"step": 605
},
{
"epoch": 2.484725050916497,
"grad_norm": 0.33597373962402344,
"learning_rate": 0.000186780085456259,
"loss": 0.6333,
"step": 610
},
{
"epoch": 2.505091649694501,
"grad_norm": 0.32059183716773987,
"learning_rate": 0.0001864238965661637,
"loss": 0.6322,
"step": 615
},
{
"epoch": 2.525458248472505,
"grad_norm": 0.31177854537963867,
"learning_rate": 0.00018606332181800165,
"loss": 0.6254,
"step": 620
},
{
"epoch": 2.5458248472505094,
"grad_norm": 0.35773152112960815,
"learning_rate": 0.00018569837951029595,
"loss": 0.6282,
"step": 625
},
{
"epoch": 2.566191446028513,
"grad_norm": 0.31747111678123474,
"learning_rate": 0.00018532908816321558,
"loss": 0.6229,
"step": 630
},
{
"epoch": 2.5865580448065173,
"grad_norm": 0.3057750463485718,
"learning_rate": 0.0001849554665176354,
"loss": 0.6257,
"step": 635
},
{
"epoch": 2.6069246435845215,
"grad_norm": 0.30599597096443176,
"learning_rate": 0.0001845775335341852,
"loss": 0.6199,
"step": 640
},
{
"epoch": 2.627291242362525,
"grad_norm": 0.3067176043987274,
"learning_rate": 0.0001841953083922875,
"loss": 0.6244,
"step": 645
},
{
"epoch": 2.6476578411405294,
"grad_norm": 0.32527825236320496,
"learning_rate": 0.00018380881048918405,
"loss": 0.6258,
"step": 650
},
{
"epoch": 2.6680244399185336,
"grad_norm": 0.3099977970123291,
"learning_rate": 0.00018341805943895178,
"loss": 0.6298,
"step": 655
},
{
"epoch": 2.6883910386965377,
"grad_norm": 0.3218511641025543,
"learning_rate": 0.00018302307507150702,
"loss": 0.624,
"step": 660
},
{
"epoch": 2.708757637474542,
"grad_norm": 0.3705076277256012,
"learning_rate": 0.0001826238774315995,
"loss": 0.6219,
"step": 665
},
{
"epoch": 2.729124236252546,
"grad_norm": 0.4094727337360382,
"learning_rate": 0.00018222048677779494,
"loss": 0.6244,
"step": 670
},
{
"epoch": 2.74949083503055,
"grad_norm": 0.327188640832901,
"learning_rate": 0.00018181292358144703,
"loss": 0.6195,
"step": 675
},
{
"epoch": 2.769857433808554,
"grad_norm": 0.3251926302909851,
"learning_rate": 0.0001814012085256585,
"loss": 0.6235,
"step": 680
},
{
"epoch": 2.790224032586558,
"grad_norm": 0.320218950510025,
"learning_rate": 0.00018098536250423154,
"loss": 0.6236,
"step": 685
},
{
"epoch": 2.810590631364562,
"grad_norm": 0.32766249775886536,
"learning_rate": 0.00018056540662060745,
"loss": 0.6188,
"step": 690
},
{
"epoch": 2.830957230142566,
"grad_norm": 0.35727354884147644,
"learning_rate": 0.00018014136218679567,
"loss": 0.619,
"step": 695
},
{
"epoch": 2.8513238289205702,
"grad_norm": 0.32284310460090637,
"learning_rate": 0.00017971325072229226,
"loss": 0.6224,
"step": 700
},
{
"epoch": 2.8716904276985744,
"grad_norm": 0.34942692518234253,
"learning_rate": 0.00017928109395298777,
"loss": 0.617,
"step": 705
},
{
"epoch": 2.8920570264765786,
"grad_norm": 0.3333948850631714,
"learning_rate": 0.00017884491381006478,
"loss": 0.6269,
"step": 710
},
{
"epoch": 2.9124236252545828,
"grad_norm": 0.3345637321472168,
"learning_rate": 0.00017840473242888486,
"loss": 0.6226,
"step": 715
},
{
"epoch": 2.9327902240325865,
"grad_norm": 0.3559291958808899,
"learning_rate": 0.0001779605721478652,
"loss": 0.6173,
"step": 720
},
{
"epoch": 2.9531568228105907,
"grad_norm": 0.3167003393173218,
"learning_rate": 0.0001775124555073452,
"loss": 0.6224,
"step": 725
},
{
"epoch": 2.973523421588595,
"grad_norm": 0.3369687497615814,
"learning_rate": 0.00017706040524844221,
"loss": 0.6155,
"step": 730
},
{
"epoch": 2.9938900203665986,
"grad_norm": 0.30602747201919556,
"learning_rate": 0.0001766044443118978,
"loss": 0.6197,
"step": 735
},
{
"epoch": 2.9979633401221997,
"eval_loss": 1.8432329893112183,
"eval_runtime": 0.4594,
"eval_samples_per_second": 26.122,
"eval_steps_per_second": 2.177,
"step": 736
},
{
"epoch": 3.0142566191446027,
"grad_norm": 0.3442608416080475,
"learning_rate": 0.00017614459583691346,
"loss": 0.6108,
"step": 740
},
{
"epoch": 3.034623217922607,
"grad_norm": 0.30495738983154297,
"learning_rate": 0.0001756808831599762,
"loss": 0.6031,
"step": 745
},
{
"epoch": 3.054989816700611,
"grad_norm": 0.34981706738471985,
"learning_rate": 0.0001752133298136744,
"loss": 0.5959,
"step": 750
},
{
"epoch": 3.0753564154786153,
"grad_norm": 0.32501649856567383,
"learning_rate": 0.00017474195952550355,
"loss": 0.5964,
"step": 755
},
{
"epoch": 3.095723014256619,
"grad_norm": 0.3137637972831726,
"learning_rate": 0.00017426679621666212,
"loss": 0.5885,
"step": 760
},
{
"epoch": 3.116089613034623,
"grad_norm": 0.3315506875514984,
"learning_rate": 0.00017378786400083757,
"loss": 0.5914,
"step": 765
},
{
"epoch": 3.1364562118126273,
"grad_norm": 0.33116060495376587,
"learning_rate": 0.00017330518718298264,
"loss": 0.5997,
"step": 770
},
{
"epoch": 3.1568228105906315,
"grad_norm": 0.33854183554649353,
"learning_rate": 0.0001728187902580819,
"loss": 0.5971,
"step": 775
},
{
"epoch": 3.1771894093686353,
"grad_norm": 0.31813180446624756,
"learning_rate": 0.00017232869790990879,
"loss": 0.5994,
"step": 780
},
{
"epoch": 3.1975560081466394,
"grad_norm": 0.30412253737449646,
"learning_rate": 0.00017183493500977278,
"loss": 0.5979,
"step": 785
},
{
"epoch": 3.2179226069246436,
"grad_norm": 0.3499339520931244,
"learning_rate": 0.0001713375266152572,
"loss": 0.5981,
"step": 790
},
{
"epoch": 3.2382892057026478,
"grad_norm": 0.3247123956680298,
"learning_rate": 0.00017083649796894795,
"loss": 0.5955,
"step": 795
},
{
"epoch": 3.258655804480652,
"grad_norm": 0.3245786726474762,
"learning_rate": 0.00017033187449715196,
"loss": 0.6019,
"step": 800
},
{
"epoch": 3.2790224032586557,
"grad_norm": 0.3269343674182892,
"learning_rate": 0.00016982368180860728,
"loss": 0.5979,
"step": 805
},
{
"epoch": 3.29938900203666,
"grad_norm": 0.4949992299079895,
"learning_rate": 0.00016931194569318327,
"loss": 0.6037,
"step": 810
},
{
"epoch": 3.319755600814664,
"grad_norm": 0.3468710780143738,
"learning_rate": 0.00016879669212057187,
"loss": 0.5943,
"step": 815
},
{
"epoch": 3.340122199592668,
"grad_norm": 0.323321670293808,
"learning_rate": 0.00016827794723896968,
"loss": 0.6003,
"step": 820
},
{
"epoch": 3.360488798370672,
"grad_norm": 0.3203057646751404,
"learning_rate": 0.00016775573737375096,
"loss": 0.6023,
"step": 825
},
{
"epoch": 3.380855397148676,
"grad_norm": 0.32133370637893677,
"learning_rate": 0.0001672300890261317,
"loss": 0.6043,
"step": 830
},
{
"epoch": 3.4012219959266803,
"grad_norm": 0.32316601276397705,
"learning_rate": 0.0001667010288718247,
"loss": 0.6051,
"step": 835
},
{
"epoch": 3.4215885947046845,
"grad_norm": 0.34438490867614746,
"learning_rate": 0.00016616858375968595,
"loss": 0.6039,
"step": 840
},
{
"epoch": 3.4419551934826886,
"grad_norm": 0.3012474477291107,
"learning_rate": 0.0001656327807103518,
"loss": 0.599,
"step": 845
},
{
"epoch": 3.4623217922606924,
"grad_norm": 0.3228248059749603,
"learning_rate": 0.0001650936469148681,
"loss": 0.604,
"step": 850
},
{
"epoch": 3.4826883910386965,
"grad_norm": 0.3279968798160553,
"learning_rate": 0.00016455120973331,
"loss": 0.5953,
"step": 855
},
{
"epoch": 3.5030549898167007,
"grad_norm": 0.3526194393634796,
"learning_rate": 0.0001640054966933935,
"loss": 0.6029,
"step": 860
},
{
"epoch": 3.5234215885947044,
"grad_norm": 0.37160396575927734,
"learning_rate": 0.00016345653548907873,
"loss": 0.6005,
"step": 865
},
{
"epoch": 3.5437881873727086,
"grad_norm": 0.33020690083503723,
"learning_rate": 0.00016290435397916424,
"loss": 0.6049,
"step": 870
},
{
"epoch": 3.564154786150713,
"grad_norm": 0.33525606989860535,
"learning_rate": 0.00016234898018587337,
"loss": 0.5942,
"step": 875
},
{
"epoch": 3.584521384928717,
"grad_norm": 0.33215492963790894,
"learning_rate": 0.00016179044229343206,
"loss": 0.6,
"step": 880
},
{
"epoch": 3.604887983706721,
"grad_norm": 0.3076721429824829,
"learning_rate": 0.00016122876864663868,
"loss": 0.5984,
"step": 885
},
{
"epoch": 3.6252545824847253,
"grad_norm": 0.379398375749588,
"learning_rate": 0.00016066398774942554,
"loss": 0.6005,
"step": 890
},
{
"epoch": 3.645621181262729,
"grad_norm": 0.3342379629611969,
"learning_rate": 0.00016009612826341227,
"loss": 0.5987,
"step": 895
},
{
"epoch": 3.6659877800407332,
"grad_norm": 0.3370605707168579,
"learning_rate": 0.00015952521900645144,
"loss": 0.6013,
"step": 900
},
{
"epoch": 3.6863543788187374,
"grad_norm": 0.3159192204475403,
"learning_rate": 0.000158951288951166,
"loss": 0.5992,
"step": 905
},
{
"epoch": 3.706720977596741,
"grad_norm": 0.3914301097393036,
"learning_rate": 0.000158374367223479,
"loss": 0.5994,
"step": 910
},
{
"epoch": 3.7270875763747453,
"grad_norm": 0.33679434657096863,
"learning_rate": 0.00015779448310113554,
"loss": 0.5949,
"step": 915
},
{
"epoch": 3.7474541751527495,
"grad_norm": 0.3156144618988037,
"learning_rate": 0.00015721166601221698,
"loss": 0.595,
"step": 920
},
{
"epoch": 3.7678207739307537,
"grad_norm": 0.3130475878715515,
"learning_rate": 0.0001566259455336474,
"loss": 0.596,
"step": 925
},
{
"epoch": 3.788187372708758,
"grad_norm": 0.303627610206604,
"learning_rate": 0.00015603735138969272,
"loss": 0.6036,
"step": 930
},
{
"epoch": 3.8085539714867616,
"grad_norm": 0.32147976756095886,
"learning_rate": 0.0001554459134504523,
"loss": 0.5971,
"step": 935
},
{
"epoch": 3.8289205702647657,
"grad_norm": 0.33086270093917847,
"learning_rate": 0.000154851661730343,
"loss": 0.5979,
"step": 940
},
{
"epoch": 3.84928716904277,
"grad_norm": 0.3000107407569885,
"learning_rate": 0.00015425462638657595,
"loss": 0.5914,
"step": 945
},
{
"epoch": 3.869653767820774,
"grad_norm": 0.292559415102005,
"learning_rate": 0.0001536548377176263,
"loss": 0.5941,
"step": 950
},
{
"epoch": 3.890020366598778,
"grad_norm": 0.30751997232437134,
"learning_rate": 0.00015305232616169548,
"loss": 0.6094,
"step": 955
},
{
"epoch": 3.910386965376782,
"grad_norm": 0.3731157183647156,
"learning_rate": 0.00015244712229516656,
"loss": 0.5955,
"step": 960
},
{
"epoch": 3.930753564154786,
"grad_norm": 0.3387170732021332,
"learning_rate": 0.00015183925683105254,
"loss": 0.5972,
"step": 965
},
{
"epoch": 3.9511201629327903,
"grad_norm": 0.3645428717136383,
"learning_rate": 0.0001512287606174377,
"loss": 0.5998,
"step": 970
},
{
"epoch": 3.9714867617107945,
"grad_norm": 0.29603078961372375,
"learning_rate": 0.0001506156646359123,
"loss": 0.596,
"step": 975
},
{
"epoch": 3.9918533604887982,
"grad_norm": 0.30991217494010925,
"learning_rate": 0.00015000000000000001,
"loss": 0.6011,
"step": 980
},
{
"epoch": 4.0,
"eval_loss": 1.870729923248291,
"eval_runtime": 0.3769,
"eval_samples_per_second": 31.837,
"eval_steps_per_second": 2.653,
"step": 982
},
{
"epoch": 4.012219959266803,
"grad_norm": 0.33605799078941345,
"learning_rate": 0.00014938179795357916,
"loss": 0.5858,
"step": 985
},
{
"epoch": 4.032586558044806,
"grad_norm": 0.3564818203449249,
"learning_rate": 0.00014876108986929717,
"loss": 0.5735,
"step": 990
},
{
"epoch": 4.05295315682281,
"grad_norm": 0.31967151165008545,
"learning_rate": 0.00014813790724697832,
"loss": 0.579,
"step": 995
},
{
"epoch": 4.0733197556008145,
"grad_norm": 0.33131125569343567,
"learning_rate": 0.0001475122817120253,
"loss": 0.57,
"step": 1000
},
{
"epoch": 4.093686354378819,
"grad_norm": 0.3302120864391327,
"learning_rate": 0.00014688424501381424,
"loss": 0.5658,
"step": 1005
},
{
"epoch": 4.114052953156823,
"grad_norm": 0.356545627117157,
"learning_rate": 0.00014625382902408356,
"loss": 0.5722,
"step": 1010
},
{
"epoch": 4.134419551934827,
"grad_norm": 0.30831289291381836,
"learning_rate": 0.0001456210657353163,
"loss": 0.5775,
"step": 1015
},
{
"epoch": 4.154786150712831,
"grad_norm": 0.32479575276374817,
"learning_rate": 0.00014498598725911691,
"loss": 0.5762,
"step": 1020
},
{
"epoch": 4.175152749490835,
"grad_norm": 0.3275298476219177,
"learning_rate": 0.00014434862582458135,
"loss": 0.5742,
"step": 1025
},
{
"epoch": 4.195519348268839,
"grad_norm": 0.32563847303390503,
"learning_rate": 0.00014370901377666167,
"loss": 0.5696,
"step": 1030
},
{
"epoch": 4.215885947046843,
"grad_norm": 0.3277311325073242,
"learning_rate": 0.00014306718357452446,
"loss": 0.5683,
"step": 1035
},
{
"epoch": 4.236252545824847,
"grad_norm": 0.3310577869415283,
"learning_rate": 0.00014242316778990372,
"loss": 0.5724,
"step": 1040
},
{
"epoch": 4.256619144602851,
"grad_norm": 0.3116869330406189,
"learning_rate": 0.00014177699910544793,
"loss": 0.5748,
"step": 1045
},
{
"epoch": 4.276985743380855,
"grad_norm": 0.4322071969509125,
"learning_rate": 0.00014112871031306119,
"loss": 0.5848,
"step": 1050
},
{
"epoch": 4.2973523421588595,
"grad_norm": 0.3503554165363312,
"learning_rate": 0.00014047833431223938,
"loss": 0.5732,
"step": 1055
},
{
"epoch": 4.317718940936864,
"grad_norm": 0.31962850689888,
"learning_rate": 0.00013982590410840056,
"loss": 0.5774,
"step": 1060
},
{
"epoch": 4.338085539714868,
"grad_norm": 0.32895031571388245,
"learning_rate": 0.00013917145281120983,
"loss": 0.5765,
"step": 1065
},
{
"epoch": 4.358452138492872,
"grad_norm": 0.3837604522705078,
"learning_rate": 0.00013851501363289906,
"loss": 0.5817,
"step": 1070
},
{
"epoch": 4.378818737270876,
"grad_norm": 0.35425829887390137,
"learning_rate": 0.0001378566198865818,
"loss": 0.58,
"step": 1075
},
{
"epoch": 4.3991853360488795,
"grad_norm": 0.33126357197761536,
"learning_rate": 0.00013719630498456212,
"loss": 0.5827,
"step": 1080
},
{
"epoch": 4.419551934826884,
"grad_norm": 0.3265593945980072,
"learning_rate": 0.00013653410243663952,
"loss": 0.5826,
"step": 1085
},
{
"epoch": 4.439918533604888,
"grad_norm": 0.4005591869354248,
"learning_rate": 0.00013587004584840804,
"loss": 0.5795,
"step": 1090
},
{
"epoch": 4.460285132382892,
"grad_norm": 0.36212101578712463,
"learning_rate": 0.00013520416891955102,
"loss": 0.5711,
"step": 1095
},
{
"epoch": 4.480651731160896,
"grad_norm": 0.340218186378479,
"learning_rate": 0.00013453650544213076,
"loss": 0.5821,
"step": 1100
},
{
"epoch": 4.5010183299389,
"grad_norm": 0.31296971440315247,
"learning_rate": 0.00013386708929887377,
"loss": 0.5798,
"step": 1105
},
{
"epoch": 4.521384928716905,
"grad_norm": 0.31503763794898987,
"learning_rate": 0.00013319595446145116,
"loss": 0.5757,
"step": 1110
},
{
"epoch": 4.541751527494909,
"grad_norm": 0.3179926574230194,
"learning_rate": 0.00013252313498875472,
"loss": 0.5822,
"step": 1115
},
{
"epoch": 4.562118126272912,
"grad_norm": 0.3273943066596985,
"learning_rate": 0.00013184866502516845,
"loss": 0.5796,
"step": 1120
},
{
"epoch": 4.582484725050916,
"grad_norm": 0.3253306448459625,
"learning_rate": 0.00013117257879883583,
"loss": 0.5789,
"step": 1125
},
{
"epoch": 4.60285132382892,
"grad_norm": 0.34257039427757263,
"learning_rate": 0.00013049491061992274,
"loss": 0.5798,
"step": 1130
},
{
"epoch": 4.6232179226069245,
"grad_norm": 0.32451459765434265,
"learning_rate": 0.00012981569487887637,
"loss": 0.5766,
"step": 1135
},
{
"epoch": 4.643584521384929,
"grad_norm": 0.31603989005088806,
"learning_rate": 0.0001291349660446799,
"loss": 0.5765,
"step": 1140
},
{
"epoch": 4.663951120162933,
"grad_norm": 0.3706742525100708,
"learning_rate": 0.00012845275866310324,
"loss": 0.5809,
"step": 1145
},
{
"epoch": 4.684317718940937,
"grad_norm": 0.32328835129737854,
"learning_rate": 0.00012776910735495003,
"loss": 0.5784,
"step": 1150
},
{
"epoch": 4.704684317718941,
"grad_norm": 0.3305988907814026,
"learning_rate": 0.00012708404681430053,
"loss": 0.5678,
"step": 1155
},
{
"epoch": 4.725050916496945,
"grad_norm": 0.33260881900787354,
"learning_rate": 0.00012639761180675098,
"loss": 0.5895,
"step": 1160
},
{
"epoch": 4.745417515274949,
"grad_norm": 0.34639254212379456,
"learning_rate": 0.0001257098371676495,
"loss": 0.5823,
"step": 1165
},
{
"epoch": 4.765784114052953,
"grad_norm": 0.3303966224193573,
"learning_rate": 0.0001250207578003279,
"loss": 0.5716,
"step": 1170
},
{
"epoch": 4.786150712830957,
"grad_norm": 0.3739718496799469,
"learning_rate": 0.0001243304086743309,
"loss": 0.5811,
"step": 1175
},
{
"epoch": 4.806517311608961,
"grad_norm": 0.36287903785705566,
"learning_rate": 0.0001236388248236409,
"loss": 0.5791,
"step": 1180
},
{
"epoch": 4.826883910386965,
"grad_norm": 0.3199651837348938,
"learning_rate": 0.00012294604134490056,
"loss": 0.5769,
"step": 1185
},
{
"epoch": 4.84725050916497,
"grad_norm": 0.3307904899120331,
"learning_rate": 0.00012225209339563145,
"loss": 0.5781,
"step": 1190
},
{
"epoch": 4.867617107942974,
"grad_norm": 0.3095554709434509,
"learning_rate": 0.00012155701619244997,
"loss": 0.5776,
"step": 1195
},
{
"epoch": 4.887983706720978,
"grad_norm": 0.318012535572052,
"learning_rate": 0.0001208608450092801,
"loss": 0.5816,
"step": 1200
},
{
"epoch": 4.908350305498981,
"grad_norm": 0.31566402316093445,
"learning_rate": 0.00012016361517556334,
"loss": 0.578,
"step": 1205
},
{
"epoch": 4.928716904276985,
"grad_norm": 0.32222023606300354,
"learning_rate": 0.00011946536207446586,
"loss": 0.5776,
"step": 1210
},
{
"epoch": 4.94908350305499,
"grad_norm": 0.31605854630470276,
"learning_rate": 0.00011876612114108277,
"loss": 0.5793,
"step": 1215
},
{
"epoch": 4.969450101832994,
"grad_norm": 0.3147432804107666,
"learning_rate": 0.0001180659278606399,
"loss": 0.5753,
"step": 1220
},
{
"epoch": 4.989816700610998,
"grad_norm": 0.3298138380050659,
"learning_rate": 0.00011736481776669306,
"loss": 0.5805,
"step": 1225
},
{
"epoch": 4.997963340122199,
"eval_loss": 1.9009366035461426,
"eval_runtime": 0.3793,
"eval_samples_per_second": 31.637,
"eval_steps_per_second": 2.636,
"step": 1227
},
{
"epoch": 5.010183299389002,
"grad_norm": 0.3077872395515442,
"learning_rate": 0.00011666282643932458,
"loss": 0.5659,
"step": 1230
},
{
"epoch": 5.030549898167006,
"grad_norm": 0.3151206374168396,
"learning_rate": 0.00011595998950333793,
"loss": 0.5451,
"step": 1235
},
{
"epoch": 5.05091649694501,
"grad_norm": 0.31535080075263977,
"learning_rate": 0.00011525634262644964,
"loss": 0.5478,
"step": 1240
},
{
"epoch": 5.071283095723015,
"grad_norm": 0.3317250907421112,
"learning_rate": 0.00011455192151747932,
"loss": 0.5568,
"step": 1245
},
{
"epoch": 5.091649694501018,
"grad_norm": 0.34287726879119873,
"learning_rate": 0.0001138467619245374,
"loss": 0.5529,
"step": 1250
},
{
"epoch": 5.112016293279022,
"grad_norm": 0.34103691577911377,
"learning_rate": 0.00011314089963321119,
"loss": 0.5546,
"step": 1255
},
{
"epoch": 5.132382892057026,
"grad_norm": 0.34115347266197205,
"learning_rate": 0.00011243437046474853,
"loss": 0.5571,
"step": 1260
},
{
"epoch": 5.15274949083503,
"grad_norm": 0.350697785615921,
"learning_rate": 0.0001117272102742402,
"loss": 0.5534,
"step": 1265
},
{
"epoch": 5.173116089613035,
"grad_norm": 0.34019774198532104,
"learning_rate": 0.00011101945494880012,
"loss": 0.5536,
"step": 1270
},
{
"epoch": 5.193482688391039,
"grad_norm": 0.34322109818458557,
"learning_rate": 0.00011031114040574437,
"loss": 0.5577,
"step": 1275
},
{
"epoch": 5.213849287169043,
"grad_norm": 0.33681756258010864,
"learning_rate": 0.00010960230259076818,
"loss": 0.5555,
"step": 1280
},
{
"epoch": 5.234215885947047,
"grad_norm": 0.335644394159317,
"learning_rate": 0.00010889297747612202,
"loss": 0.5559,
"step": 1285
},
{
"epoch": 5.254582484725051,
"grad_norm": 0.33518001437187195,
"learning_rate": 0.00010818320105878584,
"loss": 0.5547,
"step": 1290
},
{
"epoch": 5.274949083503055,
"grad_norm": 0.32509303092956543,
"learning_rate": 0.00010747300935864243,
"loss": 0.5539,
"step": 1295
},
{
"epoch": 5.295315682281059,
"grad_norm": 0.33121034502983093,
"learning_rate": 0.0001067624384166495,
"loss": 0.5609,
"step": 1300
},
{
"epoch": 5.315682281059063,
"grad_norm": 0.34210404753685,
"learning_rate": 0.00010605152429301055,
"loss": 0.5629,
"step": 1305
},
{
"epoch": 5.336048879837067,
"grad_norm": 0.3202488422393799,
"learning_rate": 0.0001053403030653449,
"loss": 0.5616,
"step": 1310
},
{
"epoch": 5.356415478615071,
"grad_norm": 0.33203092217445374,
"learning_rate": 0.00010462881082685691,
"loss": 0.5625,
"step": 1315
},
{
"epoch": 5.3767820773930755,
"grad_norm": 0.37128540873527527,
"learning_rate": 0.00010391708368450427,
"loss": 0.5531,
"step": 1320
},
{
"epoch": 5.39714867617108,
"grad_norm": 0.32188987731933594,
"learning_rate": 0.00010320515775716555,
"loss": 0.5559,
"step": 1325
},
{
"epoch": 5.417515274949084,
"grad_norm": 0.33244988322257996,
"learning_rate": 0.0001024930691738073,
"loss": 0.5579,
"step": 1330
},
{
"epoch": 5.437881873727088,
"grad_norm": 0.3311491310596466,
"learning_rate": 0.00010178085407165066,
"loss": 0.5545,
"step": 1335
},
{
"epoch": 5.458248472505091,
"grad_norm": 0.3205713629722595,
"learning_rate": 0.00010106854859433734,
"loss": 0.5596,
"step": 1340
},
{
"epoch": 5.478615071283095,
"grad_norm": 0.3317239284515381,
"learning_rate": 0.00010035618889009535,
"loss": 0.5669,
"step": 1345
},
{
"epoch": 5.4989816700611,
"grad_norm": 0.3288976848125458,
"learning_rate": 9.96438111099047e-05,
"loss": 0.5633,
"step": 1350
},
{
"epoch": 5.519348268839104,
"grad_norm": 0.33682990074157715,
"learning_rate": 9.893145140566269e-05,
"loss": 0.5603,
"step": 1355
},
{
"epoch": 5.539714867617108,
"grad_norm": 0.3263765871524811,
"learning_rate": 9.821914592834935e-05,
"loss": 0.5604,
"step": 1360
},
{
"epoch": 5.560081466395112,
"grad_norm": 0.3341580033302307,
"learning_rate": 9.750693082619273e-05,
"loss": 0.5627,
"step": 1365
},
{
"epoch": 5.580448065173116,
"grad_norm": 0.3362567126750946,
"learning_rate": 9.679484224283449e-05,
"loss": 0.5595,
"step": 1370
},
{
"epoch": 5.6008146639511205,
"grad_norm": 0.33394739031791687,
"learning_rate": 9.608291631549574e-05,
"loss": 0.5548,
"step": 1375
},
{
"epoch": 5.621181262729124,
"grad_norm": 0.3413245975971222,
"learning_rate": 9.537118917314311e-05,
"loss": 0.5509,
"step": 1380
},
{
"epoch": 5.641547861507128,
"grad_norm": 0.32499754428863525,
"learning_rate": 9.46596969346551e-05,
"loss": 0.5582,
"step": 1385
},
{
"epoch": 5.661914460285132,
"grad_norm": 0.3440452218055725,
"learning_rate": 9.39484757069895e-05,
"loss": 0.5638,
"step": 1390
},
{
"epoch": 5.682281059063136,
"grad_norm": 0.3295697271823883,
"learning_rate": 9.323756158335053e-05,
"loss": 0.5592,
"step": 1395
},
{
"epoch": 5.7026476578411405,
"grad_norm": 0.32334938645362854,
"learning_rate": 9.252699064135758e-05,
"loss": 0.5609,
"step": 1400
},
{
"epoch": 5.723014256619145,
"grad_norm": 0.33664849400520325,
"learning_rate": 9.181679894121421e-05,
"loss": 0.5584,
"step": 1405
},
{
"epoch": 5.743380855397149,
"grad_norm": 0.3422396779060364,
"learning_rate": 9.1107022523878e-05,
"loss": 0.5553,
"step": 1410
},
{
"epoch": 5.763747454175153,
"grad_norm": 0.3184050917625427,
"learning_rate": 9.039769740923183e-05,
"loss": 0.5604,
"step": 1415
},
{
"epoch": 5.784114052953157,
"grad_norm": 0.326419860124588,
"learning_rate": 8.968885959425567e-05,
"loss": 0.5645,
"step": 1420
},
{
"epoch": 5.804480651731161,
"grad_norm": 0.33074718713760376,
"learning_rate": 8.898054505119989e-05,
"loss": 0.5654,
"step": 1425
},
{
"epoch": 5.824847250509165,
"grad_norm": 0.3398876488208771,
"learning_rate": 8.827278972575983e-05,
"loss": 0.5635,
"step": 1430
},
{
"epoch": 5.845213849287169,
"grad_norm": 0.33468544483184814,
"learning_rate": 8.756562953525152e-05,
"loss": 0.5557,
"step": 1435
},
{
"epoch": 5.865580448065173,
"grad_norm": 0.3315589427947998,
"learning_rate": 8.685910036678883e-05,
"loss": 0.5626,
"step": 1440
},
{
"epoch": 5.885947046843177,
"grad_norm": 0.3360842764377594,
"learning_rate": 8.615323807546258e-05,
"loss": 0.5588,
"step": 1445
},
{
"epoch": 5.906313645621181,
"grad_norm": 0.3325459063053131,
"learning_rate": 8.54480784825207e-05,
"loss": 0.5569,
"step": 1450
},
{
"epoch": 5.9266802443991855,
"grad_norm": 0.3266001343727112,
"learning_rate": 8.474365737355038e-05,
"loss": 0.5617,
"step": 1455
},
{
"epoch": 5.94704684317719,
"grad_norm": 0.3289891481399536,
"learning_rate": 8.404001049666211e-05,
"loss": 0.5572,
"step": 1460
},
{
"epoch": 5.967413441955194,
"grad_norm": 0.32171157002449036,
"learning_rate": 8.333717356067543e-05,
"loss": 0.5571,
"step": 1465
},
{
"epoch": 5.987780040733197,
"grad_norm": 0.3251187205314636,
"learning_rate": 8.263518223330697e-05,
"loss": 0.5585,
"step": 1470
},
{
"epoch": 6.0,
"eval_loss": 1.9298146963119507,
"eval_runtime": 0.3781,
"eval_samples_per_second": 31.738,
"eval_steps_per_second": 2.645,
"step": 1473
},
{
"epoch": 6.008146639511201,
"grad_norm": 0.32297611236572266,
"learning_rate": 8.193407213936012e-05,
"loss": 0.5502,
"step": 1475
},
{
"epoch": 6.0285132382892055,
"grad_norm": 0.32624468207359314,
"learning_rate": 8.123387885891725e-05,
"loss": 0.5355,
"step": 1480
},
{
"epoch": 6.04887983706721,
"grad_norm": 0.34197378158569336,
"learning_rate": 8.053463792553416e-05,
"loss": 0.5415,
"step": 1485
},
{
"epoch": 6.069246435845214,
"grad_norm": 0.3497610092163086,
"learning_rate": 7.98363848244367e-05,
"loss": 0.5418,
"step": 1490
},
{
"epoch": 6.089613034623218,
"grad_norm": 0.35511311888694763,
"learning_rate": 7.913915499071993e-05,
"loss": 0.5353,
"step": 1495
},
{
"epoch": 6.109979633401222,
"grad_norm": 0.33935195207595825,
"learning_rate": 7.844298380755003e-05,
"loss": 0.5371,
"step": 1500
},
{
"epoch": 6.130346232179226,
"grad_norm": 0.33899983763694763,
"learning_rate": 7.774790660436858e-05,
"loss": 0.5379,
"step": 1505
},
{
"epoch": 6.1507128309572305,
"grad_norm": 0.35454773902893066,
"learning_rate": 7.705395865509947e-05,
"loss": 0.5353,
"step": 1510
},
{
"epoch": 6.171079429735234,
"grad_norm": 0.3619644045829773,
"learning_rate": 7.636117517635912e-05,
"loss": 0.5402,
"step": 1515
},
{
"epoch": 6.191446028513238,
"grad_norm": 0.34515708684921265,
"learning_rate": 7.566959132566915e-05,
"loss": 0.5377,
"step": 1520
},
{
"epoch": 6.211812627291242,
"grad_norm": 0.347515344619751,
"learning_rate": 7.497924219967209e-05,
"loss": 0.535,
"step": 1525
},
{
"epoch": 6.232179226069246,
"grad_norm": 0.34020859003067017,
"learning_rate": 7.429016283235053e-05,
"loss": 0.5406,
"step": 1530
},
{
"epoch": 6.2525458248472505,
"grad_norm": 0.34483760595321655,
"learning_rate": 7.360238819324903e-05,
"loss": 0.5377,
"step": 1535
},
{
"epoch": 6.272912423625255,
"grad_norm": 0.34881579875946045,
"learning_rate": 7.291595318569951e-05,
"loss": 0.5372,
"step": 1540
},
{
"epoch": 6.293279022403259,
"grad_norm": 0.37821272015571594,
"learning_rate": 7.223089264505e-05,
"loss": 0.5383,
"step": 1545
},
{
"epoch": 6.313645621181263,
"grad_norm": 0.35509517788887024,
"learning_rate": 7.154724133689677e-05,
"loss": 0.5464,
"step": 1550
},
{
"epoch": 6.334012219959266,
"grad_norm": 0.34138378500938416,
"learning_rate": 7.086503395532012e-05,
"loss": 0.5366,
"step": 1555
},
{
"epoch": 6.3543788187372705,
"grad_norm": 0.3617904484272003,
"learning_rate": 7.018430512112366e-05,
"loss": 0.541,
"step": 1560
},
{
"epoch": 6.374745417515275,
"grad_norm": 0.3442498743534088,
"learning_rate": 6.950508938007729e-05,
"loss": 0.5388,
"step": 1565
},
{
"epoch": 6.395112016293279,
"grad_norm": 0.3558500111103058,
"learning_rate": 6.88274212011642e-05,
"loss": 0.5363,
"step": 1570
},
{
"epoch": 6.415478615071283,
"grad_norm": 0.35669848322868347,
"learning_rate": 6.815133497483157e-05,
"loss": 0.5462,
"step": 1575
},
{
"epoch": 6.435845213849287,
"grad_norm": 0.3383665084838867,
"learning_rate": 6.74768650112453e-05,
"loss": 0.5389,
"step": 1580
},
{
"epoch": 6.456211812627291,
"grad_norm": 0.3450273275375366,
"learning_rate": 6.680404553854885e-05,
"loss": 0.5378,
"step": 1585
},
{
"epoch": 6.4765784114052956,
"grad_norm": 0.34580737352371216,
"learning_rate": 6.613291070112624e-05,
"loss": 0.5401,
"step": 1590
},
{
"epoch": 6.4969450101833,
"grad_norm": 0.34363240003585815,
"learning_rate": 6.546349455786926e-05,
"loss": 0.5378,
"step": 1595
},
{
"epoch": 6.517311608961304,
"grad_norm": 0.35278892517089844,
"learning_rate": 6.479583108044899e-05,
"loss": 0.5387,
"step": 1600
},
{
"epoch": 6.537678207739307,
"grad_norm": 0.36023351550102234,
"learning_rate": 6.412995415159197e-05,
"loss": 0.5421,
"step": 1605
},
{
"epoch": 6.558044806517311,
"grad_norm": 0.3527635633945465,
"learning_rate": 6.34658975633605e-05,
"loss": 0.5469,
"step": 1610
},
{
"epoch": 6.5784114052953155,
"grad_norm": 0.3474122881889343,
"learning_rate": 6.28036950154379e-05,
"loss": 0.5468,
"step": 1615
},
{
"epoch": 6.59877800407332,
"grad_norm": 0.336265504360199,
"learning_rate": 6.214338011341824e-05,
"loss": 0.5459,
"step": 1620
},
{
"epoch": 6.619144602851324,
"grad_norm": 0.3502029776573181,
"learning_rate": 6.148498636710092e-05,
"loss": 0.5399,
"step": 1625
},
{
"epoch": 6.639511201629328,
"grad_norm": 0.3454211950302124,
"learning_rate": 6.082854718879021e-05,
"loss": 0.5376,
"step": 1630
},
{
"epoch": 6.659877800407332,
"grad_norm": 0.3441862165927887,
"learning_rate": 6.017409589159946e-05,
"loss": 0.543,
"step": 1635
},
{
"epoch": 6.680244399185336,
"grad_norm": 0.33739784359931946,
"learning_rate": 5.952166568776062e-05,
"loss": 0.5441,
"step": 1640
},
{
"epoch": 6.70061099796334,
"grad_norm": 0.3396100103855133,
"learning_rate": 5.887128968693887e-05,
"loss": 0.5471,
"step": 1645
},
{
"epoch": 6.720977596741344,
"grad_norm": 0.35324475169181824,
"learning_rate": 5.822300089455211e-05,
"loss": 0.5408,
"step": 1650
},
{
"epoch": 6.741344195519348,
"grad_norm": 0.3377906084060669,
"learning_rate": 5.7576832210096245e-05,
"loss": 0.5416,
"step": 1655
},
{
"epoch": 6.761710794297352,
"grad_norm": 0.3413607180118561,
"learning_rate": 5.6932816425475554e-05,
"loss": 0.5425,
"step": 1660
},
{
"epoch": 6.782077393075356,
"grad_norm": 0.3485510051250458,
"learning_rate": 5.629098622333837e-05,
"loss": 0.5459,
"step": 1665
},
{
"epoch": 6.802443991853361,
"grad_norm": 0.3555828332901001,
"learning_rate": 5.5651374175418656e-05,
"loss": 0.5473,
"step": 1670
},
{
"epoch": 6.822810590631365,
"grad_norm": 0.34203040599823,
"learning_rate": 5.5014012740883115e-05,
"loss": 0.539,
"step": 1675
},
{
"epoch": 6.843177189409369,
"grad_norm": 0.34898054599761963,
"learning_rate": 5.43789342646837e-05,
"loss": 0.5388,
"step": 1680
},
{
"epoch": 6.863543788187373,
"grad_norm": 0.3571433126926422,
"learning_rate": 5.37461709759165e-05,
"loss": 0.5486,
"step": 1685
},
{
"epoch": 6.883910386965377,
"grad_norm": 0.34400326013565063,
"learning_rate": 5.3115754986185774e-05,
"loss": 0.545,
"step": 1690
},
{
"epoch": 6.904276985743381,
"grad_norm": 0.3509522080421448,
"learning_rate": 5.248771828797474e-05,
"loss": 0.5438,
"step": 1695
},
{
"epoch": 6.924643584521385,
"grad_norm": 0.33699506521224976,
"learning_rate": 5.1862092753021754e-05,
"loss": 0.5396,
"step": 1700
},
{
"epoch": 6.945010183299389,
"grad_norm": 0.3432750105857849,
"learning_rate": 5.123891013070288e-05,
"loss": 0.5468,
"step": 1705
},
{
"epoch": 6.965376782077393,
"grad_norm": 0.3491397500038147,
"learning_rate": 5.061820204642085e-05,
"loss": 0.5467,
"step": 1710
},
{
"epoch": 6.985743380855397,
"grad_norm": 0.34977835416793823,
"learning_rate": 5.000000000000002e-05,
"loss": 0.5413,
"step": 1715
},
{
"epoch": 6.997963340122199,
"eval_loss": 1.9540338516235352,
"eval_runtime": 0.3818,
"eval_samples_per_second": 31.43,
"eval_steps_per_second": 2.619,
"step": 1718
},
{
"epoch": 7.006109979633401,
"grad_norm": 0.35306668281555176,
"learning_rate": 4.938433536408771e-05,
"loss": 0.5349,
"step": 1720
},
{
"epoch": 7.026476578411406,
"grad_norm": 0.3659164011478424,
"learning_rate": 4.8771239382562287e-05,
"loss": 0.5277,
"step": 1725
},
{
"epoch": 7.04684317718941,
"grad_norm": 0.35312947630882263,
"learning_rate": 4.8160743168947496e-05,
"loss": 0.5283,
"step": 1730
},
{
"epoch": 7.067209775967413,
"grad_norm": 0.37476396560668945,
"learning_rate": 4.755287770483349e-05,
"loss": 0.5254,
"step": 1735
},
{
"epoch": 7.087576374745417,
"grad_norm": 0.37969648838043213,
"learning_rate": 4.694767383830453e-05,
"loss": 0.5267,
"step": 1740
},
{
"epoch": 7.107942973523421,
"grad_norm": 0.3799395263195038,
"learning_rate": 4.634516228237372e-05,
"loss": 0.5191,
"step": 1745
},
{
"epoch": 7.128309572301426,
"grad_norm": 0.3764006495475769,
"learning_rate": 4.574537361342407e-05,
"loss": 0.5243,
"step": 1750
},
{
"epoch": 7.14867617107943,
"grad_norm": 0.34774523973464966,
"learning_rate": 4.514833826965705e-05,
"loss": 0.5261,
"step": 1755
},
{
"epoch": 7.169042769857434,
"grad_norm": 0.3704371750354767,
"learning_rate": 4.4554086549547715e-05,
"loss": 0.5313,
"step": 1760
},
{
"epoch": 7.189409368635438,
"grad_norm": 0.3685317039489746,
"learning_rate": 4.3962648610307286e-05,
"loss": 0.5301,
"step": 1765
},
{
"epoch": 7.209775967413442,
"grad_norm": 0.35810336470603943,
"learning_rate": 4.337405446635264e-05,
"loss": 0.5253,
"step": 1770
},
{
"epoch": 7.2301425661914465,
"grad_norm": 0.3616923987865448,
"learning_rate": 4.278833398778306e-05,
"loss": 0.5272,
"step": 1775
},
{
"epoch": 7.25050916496945,
"grad_norm": 0.3660266399383545,
"learning_rate": 4.2205516898864463e-05,
"loss": 0.5304,
"step": 1780
},
{
"epoch": 7.270875763747454,
"grad_norm": 0.36601680517196655,
"learning_rate": 4.1625632776521037e-05,
"loss": 0.5327,
"step": 1785
},
{
"epoch": 7.291242362525458,
"grad_norm": 0.36126160621643066,
"learning_rate": 4.1048711048834033e-05,
"loss": 0.5256,
"step": 1790
},
{
"epoch": 7.311608961303462,
"grad_norm": 0.3732444941997528,
"learning_rate": 4.0474780993548566e-05,
"loss": 0.5261,
"step": 1795
},
{
"epoch": 7.3319755600814664,
"grad_norm": 0.36716046929359436,
"learning_rate": 3.990387173658774e-05,
"loss": 0.53,
"step": 1800
},
{
"epoch": 7.352342158859471,
"grad_norm": 0.3701815903186798,
"learning_rate": 3.933601225057446e-05,
"loss": 0.532,
"step": 1805
},
{
"epoch": 7.372708757637475,
"grad_norm": 0.3797333836555481,
"learning_rate": 3.8771231353361326e-05,
"loss": 0.524,
"step": 1810
},
{
"epoch": 7.393075356415479,
"grad_norm": 0.364581435918808,
"learning_rate": 3.820955770656798e-05,
"loss": 0.524,
"step": 1815
},
{
"epoch": 7.413441955193482,
"grad_norm": 0.36119377613067627,
"learning_rate": 3.7651019814126654e-05,
"loss": 0.5268,
"step": 1820
},
{
"epoch": 7.433808553971486,
"grad_norm": 0.35705694556236267,
"learning_rate": 3.7095646020835754e-05,
"loss": 0.524,
"step": 1825
},
{
"epoch": 7.454175152749491,
"grad_norm": 0.3591172993183136,
"learning_rate": 3.654346451092129e-05,
"loss": 0.5306,
"step": 1830
},
{
"epoch": 7.474541751527495,
"grad_norm": 0.3496874272823334,
"learning_rate": 3.5994503306606497e-05,
"loss": 0.5292,
"step": 1835
},
{
"epoch": 7.494908350305499,
"grad_norm": 0.3553576171398163,
"learning_rate": 3.544879026669005e-05,
"loss": 0.5273,
"step": 1840
},
{
"epoch": 7.515274949083503,
"grad_norm": 0.3603392541408539,
"learning_rate": 3.4906353085131914e-05,
"loss": 0.5367,
"step": 1845
},
{
"epoch": 7.535641547861507,
"grad_norm": 0.3681369721889496,
"learning_rate": 3.436721928964819e-05,
"loss": 0.5243,
"step": 1850
},
{
"epoch": 7.5560081466395115,
"grad_norm": 0.36058658361434937,
"learning_rate": 3.383141624031408e-05,
"loss": 0.5177,
"step": 1855
},
{
"epoch": 7.576374745417516,
"grad_norm": 0.3690408170223236,
"learning_rate": 3.329897112817529e-05,
"loss": 0.5168,
"step": 1860
},
{
"epoch": 7.59674134419552,
"grad_norm": 0.3683675229549408,
"learning_rate": 3.276991097386831e-05,
"loss": 0.5259,
"step": 1865
},
{
"epoch": 7.617107942973523,
"grad_norm": 0.3538174033164978,
"learning_rate": 3.2244262626249075e-05,
"loss": 0.5241,
"step": 1870
},
{
"epoch": 7.637474541751527,
"grad_norm": 0.357802152633667,
"learning_rate": 3.172205276103033e-05,
"loss": 0.5203,
"step": 1875
},
{
"epoch": 7.6578411405295315,
"grad_norm": 0.3668569028377533,
"learning_rate": 3.120330787942815e-05,
"loss": 0.5295,
"step": 1880
},
{
"epoch": 7.678207739307536,
"grad_norm": 0.3501560389995575,
"learning_rate": 3.068805430681675e-05,
"loss": 0.534,
"step": 1885
},
{
"epoch": 7.69857433808554,
"grad_norm": 0.35364505648612976,
"learning_rate": 3.0176318191392726e-05,
"loss": 0.5311,
"step": 1890
},
{
"epoch": 7.718940936863544,
"grad_norm": 0.3579385578632355,
"learning_rate": 2.966812550284803e-05,
"loss": 0.5261,
"step": 1895
},
{
"epoch": 7.739307535641548,
"grad_norm": 0.3566991090774536,
"learning_rate": 2.916350203105207e-05,
"loss": 0.5238,
"step": 1900
},
{
"epoch": 7.7596741344195515,
"grad_norm": 0.3725852966308594,
"learning_rate": 2.8662473384742773e-05,
"loss": 0.5279,
"step": 1905
},
{
"epoch": 7.780040733197556,
"grad_norm": 0.3531646132469177,
"learning_rate": 2.8165064990227252e-05,
"loss": 0.5266,
"step": 1910
},
{
"epoch": 7.80040733197556,
"grad_norm": 0.36464017629623413,
"learning_rate": 2.76713020900912e-05,
"loss": 0.5345,
"step": 1915
},
{
"epoch": 7.820773930753564,
"grad_norm": 0.36327338218688965,
"learning_rate": 2.718120974191809e-05,
"loss": 0.5211,
"step": 1920
},
{
"epoch": 7.841140529531568,
"grad_norm": 0.3583681583404541,
"learning_rate": 2.669481281701739e-05,
"loss": 0.531,
"step": 1925
},
{
"epoch": 7.861507128309572,
"grad_norm": 0.36573490500450134,
"learning_rate": 2.6212135999162445e-05,
"loss": 0.5236,
"step": 1930
},
{
"epoch": 7.8818737270875765,
"grad_norm": 0.3562600910663605,
"learning_rate": 2.573320378333789e-05,
"loss": 0.5273,
"step": 1935
},
{
"epoch": 7.902240325865581,
"grad_norm": 0.3540510833263397,
"learning_rate": 2.525804047449648e-05,
"loss": 0.53,
"step": 1940
},
{
"epoch": 7.922606924643585,
"grad_norm": 0.35902467370033264,
"learning_rate": 2.478667018632562e-05,
"loss": 0.5273,
"step": 1945
},
{
"epoch": 7.942973523421589,
"grad_norm": 0.37109845876693726,
"learning_rate": 2.4319116840023813e-05,
"loss": 0.5305,
"step": 1950
},
{
"epoch": 7.963340122199592,
"grad_norm": 0.36434054374694824,
"learning_rate": 2.3855404163086558e-05,
"loss": 0.5287,
"step": 1955
},
{
"epoch": 7.9837067209775965,
"grad_norm": 0.35995838046073914,
"learning_rate": 2.339555568810221e-05,
"loss": 0.5295,
"step": 1960
},
{
"epoch": 8.0,
"eval_loss": 1.981424331665039,
"eval_runtime": 0.3785,
"eval_samples_per_second": 31.705,
"eval_steps_per_second": 2.642,
"step": 1964
},
{
"epoch": 8.004073319755602,
"grad_norm": 0.3557198941707611,
"learning_rate": 2.2939594751557802e-05,
"loss": 0.5225,
"step": 1965
},
{
"epoch": 8.024439918533606,
"grad_norm": 0.362589955329895,
"learning_rate": 2.248754449265483e-05,
"loss": 0.5222,
"step": 1970
},
{
"epoch": 8.044806517311608,
"grad_norm": 0.36456504464149475,
"learning_rate": 2.2039427852134788e-05,
"loss": 0.5174,
"step": 1975
},
{
"epoch": 8.065173116089612,
"grad_norm": 0.35784122347831726,
"learning_rate": 2.1595267571115163e-05,
"loss": 0.5102,
"step": 1980
},
{
"epoch": 8.085539714867616,
"grad_norm": 0.3696603775024414,
"learning_rate": 2.1155086189935224e-05,
"loss": 0.5194,
"step": 1985
},
{
"epoch": 8.10590631364562,
"grad_norm": 0.3742619454860687,
"learning_rate": 2.0718906047012242e-05,
"loss": 0.5176,
"step": 1990
},
{
"epoch": 8.126272912423625,
"grad_norm": 0.368327260017395,
"learning_rate": 2.0286749277707782e-05,
"loss": 0.5175,
"step": 1995
},
{
"epoch": 8.146639511201629,
"grad_norm": 0.3673515319824219,
"learning_rate": 1.985863781320435e-05,
"loss": 0.5188,
"step": 2000
},
{
"epoch": 8.167006109979633,
"grad_norm": 0.3596293330192566,
"learning_rate": 1.9434593379392562e-05,
"loss": 0.5179,
"step": 2005
},
{
"epoch": 8.187372708757637,
"grad_norm": 0.36756983399391174,
"learning_rate": 1.9014637495768483e-05,
"loss": 0.5137,
"step": 2010
},
{
"epoch": 8.207739307535642,
"grad_norm": 0.3634319603443146,
"learning_rate": 1.8598791474341514e-05,
"loss": 0.5173,
"step": 2015
},
{
"epoch": 8.228105906313646,
"grad_norm": 0.37804871797561646,
"learning_rate": 1.8187076418552974e-05,
"loss": 0.5177,
"step": 2020
},
{
"epoch": 8.24847250509165,
"grad_norm": 0.3717120885848999,
"learning_rate": 1.777951322220508e-05,
"loss": 0.5151,
"step": 2025
},
{
"epoch": 8.268839103869654,
"grad_norm": 0.3710237443447113,
"learning_rate": 1.7376122568400532e-05,
"loss": 0.5186,
"step": 2030
},
{
"epoch": 8.289205702647658,
"grad_norm": 0.369852215051651,
"learning_rate": 1.697692492849299e-05,
"loss": 0.5146,
"step": 2035
},
{
"epoch": 8.309572301425662,
"grad_norm": 0.3662493824958801,
"learning_rate": 1.658194056104825e-05,
"loss": 0.5163,
"step": 2040
},
{
"epoch": 8.329938900203667,
"grad_norm": 0.3654170334339142,
"learning_rate": 1.619118951081594e-05,
"loss": 0.5182,
"step": 2045
},
{
"epoch": 8.35030549898167,
"grad_norm": 0.3634709119796753,
"learning_rate": 1.580469160771253e-05,
"loss": 0.5182,
"step": 2050
},
{
"epoch": 8.370672097759675,
"grad_norm": 0.3728494942188263,
"learning_rate": 1.54224664658148e-05,
"loss": 0.5193,
"step": 2055
},
{
"epoch": 8.391038696537677,
"grad_norm": 0.3711146116256714,
"learning_rate": 1.504453348236461e-05,
"loss": 0.5183,
"step": 2060
},
{
"epoch": 8.411405295315681,
"grad_norm": 0.36539244651794434,
"learning_rate": 1.467091183678444e-05,
"loss": 0.5167,
"step": 2065
},
{
"epoch": 8.431771894093686,
"grad_norm": 0.367108017206192,
"learning_rate": 1.430162048970407e-05,
"loss": 0.525,
"step": 2070
},
{
"epoch": 8.45213849287169,
"grad_norm": 0.3688015341758728,
"learning_rate": 1.3936678181998374e-05,
"loss": 0.5202,
"step": 2075
},
{
"epoch": 8.472505091649694,
"grad_norm": 0.36969494819641113,
"learning_rate": 1.357610343383634e-05,
"loss": 0.5214,
"step": 2080
},
{
"epoch": 8.492871690427698,
"grad_norm": 0.36953508853912354,
"learning_rate": 1.3219914543741008e-05,
"loss": 0.5217,
"step": 2085
},
{
"epoch": 8.513238289205702,
"grad_norm": 0.3687889575958252,
"learning_rate": 1.286812958766106e-05,
"loss": 0.5153,
"step": 2090
},
{
"epoch": 8.533604887983707,
"grad_norm": 0.36163973808288574,
"learning_rate": 1.2520766418053408e-05,
"loss": 0.5167,
"step": 2095
},
{
"epoch": 8.55397148676171,
"grad_norm": 0.37320175766944885,
"learning_rate": 1.2177842662977135e-05,
"loss": 0.5144,
"step": 2100
},
{
"epoch": 8.574338085539715,
"grad_norm": 0.36868005990982056,
"learning_rate": 1.1839375725199098e-05,
"loss": 0.5227,
"step": 2105
},
{
"epoch": 8.594704684317719,
"grad_norm": 0.36514589190483093,
"learning_rate": 1.1505382781310559e-05,
"loss": 0.5157,
"step": 2110
},
{
"epoch": 8.615071283095723,
"grad_norm": 0.3700886070728302,
"learning_rate": 1.1175880780855608e-05,
"loss": 0.5187,
"step": 2115
},
{
"epoch": 8.635437881873727,
"grad_norm": 0.3674730956554413,
"learning_rate": 1.0850886445471054e-05,
"loss": 0.5153,
"step": 2120
},
{
"epoch": 8.655804480651732,
"grad_norm": 0.37101948261260986,
"learning_rate": 1.0530416268037702e-05,
"loss": 0.5149,
"step": 2125
},
{
"epoch": 8.676171079429736,
"grad_norm": 0.36893007159233093,
"learning_rate": 1.021448651184349e-05,
"loss": 0.5183,
"step": 2130
},
{
"epoch": 8.69653767820774,
"grad_norm": 0.3680102527141571,
"learning_rate": 9.903113209758096e-06,
"loss": 0.5163,
"step": 2135
},
{
"epoch": 8.716904276985744,
"grad_norm": 0.36099961400032043,
"learning_rate": 9.596312163419274e-06,
"loss": 0.5188,
"step": 2140
},
{
"epoch": 8.737270875763748,
"grad_norm": 0.36749574542045593,
"learning_rate": 9.294098942430996e-06,
"loss": 0.5181,
"step": 2145
},
{
"epoch": 8.757637474541752,
"grad_norm": 0.36886829137802124,
"learning_rate": 8.99648888357335e-06,
"loss": 0.5171,
"step": 2150
},
{
"epoch": 8.778004073319755,
"grad_norm": 0.3683699071407318,
"learning_rate": 8.703497090024116e-06,
"loss": 0.5118,
"step": 2155
},
{
"epoch": 8.798370672097759,
"grad_norm": 0.3667095899581909,
"learning_rate": 8.415138430592428e-06,
"loss": 0.5157,
"step": 2160
},
{
"epoch": 8.818737270875763,
"grad_norm": 0.36200326681137085,
"learning_rate": 8.131427538964164e-06,
"loss": 0.5167,
"step": 2165
},
{
"epoch": 8.839103869653767,
"grad_norm": 0.3657187521457672,
"learning_rate": 7.852378812959227e-06,
"loss": 0.5155,
"step": 2170
},
{
"epoch": 8.859470468431772,
"grad_norm": 0.3640839457511902,
"learning_rate": 7.578006413801075e-06,
"loss": 0.5179,
"step": 2175
},
{
"epoch": 8.879837067209776,
"grad_norm": 0.3747439980506897,
"learning_rate": 7.308324265397836e-06,
"loss": 0.5154,
"step": 2180
},
{
"epoch": 8.90020366598778,
"grad_norm": 0.366439551115036,
"learning_rate": 7.0433460536358685e-06,
"loss": 0.5147,
"step": 2185
},
{
"epoch": 8.920570264765784,
"grad_norm": 0.37617239356040955,
"learning_rate": 6.783085225685148e-06,
"loss": 0.52,
"step": 2190
},
{
"epoch": 8.940936863543788,
"grad_norm": 0.3725389540195465,
"learning_rate": 6.527554989316897e-06,
"loss": 0.517,
"step": 2195
},
{
"epoch": 8.961303462321792,
"grad_norm": 0.3703247606754303,
"learning_rate": 6.276768312233228e-06,
"loss": 0.52,
"step": 2200
},
{
"epoch": 8.981670061099797,
"grad_norm": 0.35923415422439575,
"learning_rate": 6.030737921409169e-06,
"loss": 0.5154,
"step": 2205
},
{
"epoch": 8.9979633401222,
"eval_loss": 1.9978961944580078,
"eval_runtime": 0.3817,
"eval_samples_per_second": 31.441,
"eval_steps_per_second": 2.62,
"step": 2209
},
{
"epoch": 9.0020366598778,
"grad_norm": 0.35924673080444336,
"learning_rate": 5.789476302446662e-06,
"loss": 0.5142,
"step": 2210
},
{
"epoch": 9.022403258655805,
"grad_norm": 0.3611084818840027,
"learning_rate": 5.552995698941088e-06,
"loss": 0.5042,
"step": 2215
},
{
"epoch": 9.04276985743381,
"grad_norm": 0.3653542101383209,
"learning_rate": 5.321308111859791e-06,
"loss": 0.514,
"step": 2220
},
{
"epoch": 9.063136456211813,
"grad_norm": 0.36559605598449707,
"learning_rate": 5.094425298933136e-06,
"loss": 0.508,
"step": 2225
},
{
"epoch": 9.083503054989817,
"grad_norm": 0.3634701669216156,
"learning_rate": 4.872358774057806e-06,
"loss": 0.5073,
"step": 2230
},
{
"epoch": 9.103869653767822,
"grad_norm": 0.36651164293289185,
"learning_rate": 4.655119806712482e-06,
"loss": 0.5087,
"step": 2235
},
{
"epoch": 9.124236252545824,
"grad_norm": 0.36734244227409363,
"learning_rate": 4.442719421385922e-06,
"loss": 0.5099,
"step": 2240
},
{
"epoch": 9.144602851323828,
"grad_norm": 0.3621695339679718,
"learning_rate": 4.235168397017541e-06,
"loss": 0.5093,
"step": 2245
},
{
"epoch": 9.164969450101832,
"grad_norm": 0.36267393827438354,
"learning_rate": 4.0324772664503296e-06,
"loss": 0.5087,
"step": 2250
},
{
"epoch": 9.185336048879837,
"grad_norm": 0.36996185779571533,
"learning_rate": 3.8346563158963785e-06,
"loss": 0.5129,
"step": 2255
},
{
"epoch": 9.20570264765784,
"grad_norm": 0.3650117814540863,
"learning_rate": 3.641715584414862e-06,
"loss": 0.5145,
"step": 2260
},
{
"epoch": 9.226069246435845,
"grad_norm": 0.37506943941116333,
"learning_rate": 3.453664863402595e-06,
"loss": 0.5174,
"step": 2265
},
{
"epoch": 9.246435845213849,
"grad_norm": 0.37266090512275696,
"learning_rate": 3.270513696097055e-06,
"loss": 0.5149,
"step": 2270
},
{
"epoch": 9.266802443991853,
"grad_norm": 0.371489554643631,
"learning_rate": 3.092271377092215e-06,
"loss": 0.5155,
"step": 2275
},
{
"epoch": 9.287169042769857,
"grad_norm": 0.36911362409591675,
"learning_rate": 2.9189469518666967e-06,
"loss": 0.5125,
"step": 2280
},
{
"epoch": 9.307535641547862,
"grad_norm": 0.37086358666419983,
"learning_rate": 2.7505492163248934e-06,
"loss": 0.5116,
"step": 2285
},
{
"epoch": 9.327902240325866,
"grad_norm": 0.37481385469436646,
"learning_rate": 2.587086716350473e-06,
"loss": 0.5127,
"step": 2290
},
{
"epoch": 9.34826883910387,
"grad_norm": 0.3686206638813019,
"learning_rate": 2.4285677473727118e-06,
"loss": 0.5168,
"step": 2295
},
{
"epoch": 9.368635437881874,
"grad_norm": 0.37265482544898987,
"learning_rate": 2.2750003539455998e-06,
"loss": 0.5067,
"step": 2300
},
{
"epoch": 9.389002036659878,
"grad_norm": 0.36506715416908264,
"learning_rate": 2.1263923293394774e-06,
"loss": 0.5135,
"step": 2305
},
{
"epoch": 9.409368635437882,
"grad_norm": 0.36821919679641724,
"learning_rate": 1.9827512151456173e-06,
"loss": 0.514,
"step": 2310
},
{
"epoch": 9.429735234215887,
"grad_norm": 0.3649906814098358,
"learning_rate": 1.8440843008934561e-06,
"loss": 0.5127,
"step": 2315
},
{
"epoch": 9.45010183299389,
"grad_norm": 0.36437344551086426,
"learning_rate": 1.7103986236807313e-06,
"loss": 0.5087,
"step": 2320
},
{
"epoch": 9.470468431771893,
"grad_norm": 0.36268892884254456,
"learning_rate": 1.5817009678162685e-06,
"loss": 0.5157,
"step": 2325
},
{
"epoch": 9.490835030549897,
"grad_norm": 0.37203824520111084,
"learning_rate": 1.4579978644757464e-06,
"loss": 0.5097,
"step": 2330
},
{
"epoch": 9.511201629327902,
"grad_norm": 0.3669569492340088,
"learning_rate": 1.339295591370271e-06,
"loss": 0.5134,
"step": 2335
},
{
"epoch": 9.531568228105906,
"grad_norm": 0.37304890155792236,
"learning_rate": 1.2256001724277321e-06,
"loss": 0.5177,
"step": 2340
},
{
"epoch": 9.55193482688391,
"grad_norm": 0.36801040172576904,
"learning_rate": 1.1169173774871478e-06,
"loss": 0.5135,
"step": 2345
},
{
"epoch": 9.572301425661914,
"grad_norm": 0.3649773895740509,
"learning_rate": 1.013252722005842e-06,
"loss": 0.5096,
"step": 2350
},
{
"epoch": 9.592668024439918,
"grad_norm": 0.3708580732345581,
"learning_rate": 9.146114667795358e-07,
"loss": 0.5084,
"step": 2355
},
{
"epoch": 9.613034623217922,
"grad_norm": 0.3649328351020813,
"learning_rate": 8.209986176753948e-07,
"loss": 0.5143,
"step": 2360
},
{
"epoch": 9.633401221995927,
"grad_norm": 0.3666220009326935,
"learning_rate": 7.324189253779312e-07,
"loss": 0.508,
"step": 2365
},
{
"epoch": 9.65376782077393,
"grad_norm": 0.36750587821006775,
"learning_rate": 6.488768851480087e-07,
"loss": 0.51,
"step": 2370
},
{
"epoch": 9.674134419551935,
"grad_norm": 0.36811190843582153,
"learning_rate": 5.703767365946466e-07,
"loss": 0.512,
"step": 2375
},
{
"epoch": 9.69450101832994,
"grad_norm": 0.37304404377937317,
"learning_rate": 4.969224634598591e-07,
"loss": 0.5131,
"step": 2380
},
{
"epoch": 9.714867617107943,
"grad_norm": 0.37467247247695923,
"learning_rate": 4.2851779341654964e-07,
"loss": 0.5167,
"step": 2385
},
{
"epoch": 9.735234215885948,
"grad_norm": 0.37464529275894165,
"learning_rate": 3.651661978793075e-07,
"loss": 0.5116,
"step": 2390
},
{
"epoch": 9.755600814663952,
"grad_norm": 0.36434635519981384,
"learning_rate": 3.068708918281926e-07,
"loss": 0.5088,
"step": 2395
},
{
"epoch": 9.775967413441956,
"grad_norm": 0.3670872747898102,
"learning_rate": 2.536348336456551e-07,
"loss": 0.5172,
"step": 2400
},
{
"epoch": 9.79633401221996,
"grad_norm": 0.36443930864334106,
"learning_rate": 2.054607249663665e-07,
"loss": 0.5135,
"step": 2405
},
{
"epoch": 9.816700610997962,
"grad_norm": 0.3700142204761505,
"learning_rate": 1.6235101054011824e-07,
"loss": 0.5085,
"step": 2410
},
{
"epoch": 9.837067209775967,
"grad_norm": 0.3709539473056793,
"learning_rate": 1.2430787810776555e-07,
"loss": 0.5164,
"step": 2415
},
{
"epoch": 9.85743380855397,
"grad_norm": 0.36750784516334534,
"learning_rate": 9.133325829017158e-08,
"loss": 0.5163,
"step": 2420
},
{
"epoch": 9.877800407331975,
"grad_norm": 0.3683048188686371,
"learning_rate": 6.342882449029696e-08,
"loss": 0.5113,
"step": 2425
},
{
"epoch": 9.89816700610998,
"grad_norm": 0.37231889367103577,
"learning_rate": 4.059599280819004e-08,
"loss": 0.5136,
"step": 2430
},
{
"epoch": 9.918533604887983,
"grad_norm": 0.36660727858543396,
"learning_rate": 2.2835921969210917e-08,
"loss": 0.5183,
"step": 2435
},
{
"epoch": 9.938900203665987,
"grad_norm": 0.3696270287036896,
"learning_rate": 1.0149513265145238e-08,
"loss": 0.5079,
"step": 2440
},
{
"epoch": 9.959266802443992,
"grad_norm": 0.36083751916885376,
"learning_rate": 2.5374105085518295e-09,
"loss": 0.5151,
"step": 2445
},
{
"epoch": 9.979633401221996,
"grad_norm": 0.3677830100059509,
"learning_rate": 0.0,
"loss": 0.508,
"step": 2450
},
{
"epoch": 9.979633401221996,
"eval_loss": 2.0012366771698,
"eval_runtime": 0.3701,
"eval_samples_per_second": 32.425,
"eval_steps_per_second": 2.702,
"step": 2450
},
{
"epoch": 9.979633401221996,
"step": 2450,
"total_flos": 6.866432485157241e+18,
"train_loss": 0.5986210168137842,
"train_runtime": 9994.3917,
"train_samples_per_second": 15.694,
"train_steps_per_second": 0.245
}
],
"logging_steps": 5,
"max_steps": 2450,
"num_input_tokens_seen": 0,
"num_train_epochs": 10,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 6.866432485157241e+18,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}