nthakur's picture
Model save
53eb465 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 200,
"global_step": 3348,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0002986857825567503,
"grad_norm": 0.3188975789150384,
"learning_rate": 5.970149253731343e-07,
"loss": 0.2045,
"step": 1
},
{
"epoch": 0.0014934289127837516,
"grad_norm": 0.5466095149619057,
"learning_rate": 2.9850746268656716e-06,
"loss": 0.5187,
"step": 5
},
{
"epoch": 0.002986857825567503,
"grad_norm": 0.46728261805758203,
"learning_rate": 5.970149253731343e-06,
"loss": 0.4648,
"step": 10
},
{
"epoch": 0.004480286738351254,
"grad_norm": 0.6171065397357978,
"learning_rate": 8.955223880597016e-06,
"loss": 0.475,
"step": 15
},
{
"epoch": 0.005973715651135006,
"grad_norm": 0.49334289479515664,
"learning_rate": 1.1940298507462686e-05,
"loss": 0.4602,
"step": 20
},
{
"epoch": 0.007467144563918757,
"grad_norm": 0.5174483814171045,
"learning_rate": 1.4925373134328357e-05,
"loss": 0.4816,
"step": 25
},
{
"epoch": 0.008960573476702509,
"grad_norm": 0.4471748736139888,
"learning_rate": 1.791044776119403e-05,
"loss": 0.3534,
"step": 30
},
{
"epoch": 0.01045400238948626,
"grad_norm": 0.5227128696793767,
"learning_rate": 2.0895522388059702e-05,
"loss": 0.455,
"step": 35
},
{
"epoch": 0.011947431302270013,
"grad_norm": 0.57335506943901,
"learning_rate": 2.3880597014925373e-05,
"loss": 0.4381,
"step": 40
},
{
"epoch": 0.013440860215053764,
"grad_norm": 0.5672666580878948,
"learning_rate": 2.6865671641791047e-05,
"loss": 0.4183,
"step": 45
},
{
"epoch": 0.014934289127837515,
"grad_norm": 0.615033725558188,
"learning_rate": 2.9850746268656714e-05,
"loss": 0.4493,
"step": 50
},
{
"epoch": 0.016427718040621268,
"grad_norm": 0.5694295537449918,
"learning_rate": 3.283582089552239e-05,
"loss": 0.3214,
"step": 55
},
{
"epoch": 0.017921146953405017,
"grad_norm": 0.4317314011547438,
"learning_rate": 3.582089552238806e-05,
"loss": 0.3697,
"step": 60
},
{
"epoch": 0.01941457586618877,
"grad_norm": 0.5686093679349059,
"learning_rate": 3.8805970149253736e-05,
"loss": 0.3266,
"step": 65
},
{
"epoch": 0.02090800477897252,
"grad_norm": 0.5271452706616764,
"learning_rate": 4.1791044776119404e-05,
"loss": 0.365,
"step": 70
},
{
"epoch": 0.022401433691756272,
"grad_norm": 0.7400114199406753,
"learning_rate": 4.477611940298508e-05,
"loss": 0.4381,
"step": 75
},
{
"epoch": 0.023894862604540025,
"grad_norm": 0.5987079894007333,
"learning_rate": 4.7761194029850745e-05,
"loss": 0.3233,
"step": 80
},
{
"epoch": 0.025388291517323774,
"grad_norm": 0.5231052349056124,
"learning_rate": 5.074626865671642e-05,
"loss": 0.3543,
"step": 85
},
{
"epoch": 0.026881720430107527,
"grad_norm": 0.6540900409905596,
"learning_rate": 5.373134328358209e-05,
"loss": 0.3387,
"step": 90
},
{
"epoch": 0.028375149342891277,
"grad_norm": 0.5922650571689663,
"learning_rate": 5.671641791044776e-05,
"loss": 0.3648,
"step": 95
},
{
"epoch": 0.02986857825567503,
"grad_norm": 0.671441503007203,
"learning_rate": 5.970149253731343e-05,
"loss": 0.3842,
"step": 100
},
{
"epoch": 0.03136200716845878,
"grad_norm": 0.5589374162458473,
"learning_rate": 6.268656716417911e-05,
"loss": 0.3074,
"step": 105
},
{
"epoch": 0.032855436081242535,
"grad_norm": 0.5380602393505818,
"learning_rate": 6.567164179104478e-05,
"loss": 0.3507,
"step": 110
},
{
"epoch": 0.034348864994026285,
"grad_norm": 0.6063877341870142,
"learning_rate": 6.865671641791044e-05,
"loss": 0.3401,
"step": 115
},
{
"epoch": 0.035842293906810034,
"grad_norm": 0.5260512769152438,
"learning_rate": 7.164179104477612e-05,
"loss": 0.3724,
"step": 120
},
{
"epoch": 0.03733572281959379,
"grad_norm": 0.5959345681808844,
"learning_rate": 7.46268656716418e-05,
"loss": 0.3713,
"step": 125
},
{
"epoch": 0.03882915173237754,
"grad_norm": 0.5141504485060038,
"learning_rate": 7.761194029850747e-05,
"loss": 0.306,
"step": 130
},
{
"epoch": 0.04032258064516129,
"grad_norm": 0.45162013987431543,
"learning_rate": 8.059701492537314e-05,
"loss": 0.3279,
"step": 135
},
{
"epoch": 0.04181600955794504,
"grad_norm": 0.5664477964459901,
"learning_rate": 8.358208955223881e-05,
"loss": 0.2853,
"step": 140
},
{
"epoch": 0.043309438470728795,
"grad_norm": 0.4935236502406577,
"learning_rate": 8.656716417910447e-05,
"loss": 0.3515,
"step": 145
},
{
"epoch": 0.044802867383512544,
"grad_norm": 0.49883746692891684,
"learning_rate": 8.955223880597016e-05,
"loss": 0.3112,
"step": 150
},
{
"epoch": 0.046296296296296294,
"grad_norm": 0.5370459530188539,
"learning_rate": 9.253731343283582e-05,
"loss": 0.2918,
"step": 155
},
{
"epoch": 0.04778972520908005,
"grad_norm": 0.5147364002910404,
"learning_rate": 9.552238805970149e-05,
"loss": 0.3427,
"step": 160
},
{
"epoch": 0.0492831541218638,
"grad_norm": 0.48526830628523326,
"learning_rate": 9.850746268656717e-05,
"loss": 0.3272,
"step": 165
},
{
"epoch": 0.05077658303464755,
"grad_norm": 0.44536442368732077,
"learning_rate": 0.00010149253731343284,
"loss": 0.3216,
"step": 170
},
{
"epoch": 0.052270011947431305,
"grad_norm": 0.6529775264003637,
"learning_rate": 0.0001044776119402985,
"loss": 0.3343,
"step": 175
},
{
"epoch": 0.053763440860215055,
"grad_norm": 0.500538842360273,
"learning_rate": 0.00010746268656716419,
"loss": 0.2882,
"step": 180
},
{
"epoch": 0.055256869772998804,
"grad_norm": 0.45720922405183984,
"learning_rate": 0.00011044776119402987,
"loss": 0.3565,
"step": 185
},
{
"epoch": 0.05675029868578255,
"grad_norm": 0.395572707884918,
"learning_rate": 0.00011343283582089552,
"loss": 0.2879,
"step": 190
},
{
"epoch": 0.05824372759856631,
"grad_norm": 0.38317766271575404,
"learning_rate": 0.0001164179104477612,
"loss": 0.326,
"step": 195
},
{
"epoch": 0.05973715651135006,
"grad_norm": 0.5636861131014499,
"learning_rate": 0.00011940298507462686,
"loss": 0.3403,
"step": 200
},
{
"epoch": 0.05973715651135006,
"eval_loss": 0.3074450194835663,
"eval_runtime": 675.2136,
"eval_samples_per_second": 5.924,
"eval_steps_per_second": 0.741,
"step": 200
},
{
"epoch": 0.06123058542413381,
"grad_norm": 0.5285903241151306,
"learning_rate": 0.00012238805970149255,
"loss": 0.3054,
"step": 205
},
{
"epoch": 0.06272401433691756,
"grad_norm": 0.462285447656946,
"learning_rate": 0.00012537313432835822,
"loss": 0.3289,
"step": 210
},
{
"epoch": 0.06421744324970131,
"grad_norm": 0.41949653076343973,
"learning_rate": 0.00012835820895522389,
"loss": 0.3235,
"step": 215
},
{
"epoch": 0.06571087216248507,
"grad_norm": 0.37413394493320784,
"learning_rate": 0.00013134328358208955,
"loss": 0.3284,
"step": 220
},
{
"epoch": 0.06720430107526881,
"grad_norm": 0.43599604159082916,
"learning_rate": 0.00013432835820895525,
"loss": 0.3509,
"step": 225
},
{
"epoch": 0.06869772998805257,
"grad_norm": 0.45911550206829044,
"learning_rate": 0.0001373134328358209,
"loss": 0.2856,
"step": 230
},
{
"epoch": 0.07019115890083633,
"grad_norm": 0.6197373846263158,
"learning_rate": 0.00014029850746268658,
"loss": 0.3381,
"step": 235
},
{
"epoch": 0.07168458781362007,
"grad_norm": 0.5162001410157283,
"learning_rate": 0.00014328358208955225,
"loss": 0.2987,
"step": 240
},
{
"epoch": 0.07317801672640382,
"grad_norm": 0.3756162304697717,
"learning_rate": 0.00014626865671641792,
"loss": 0.3016,
"step": 245
},
{
"epoch": 0.07467144563918758,
"grad_norm": 0.4427925525549972,
"learning_rate": 0.0001492537313432836,
"loss": 0.3301,
"step": 250
},
{
"epoch": 0.07616487455197132,
"grad_norm": 0.4587549108919026,
"learning_rate": 0.00015223880597014925,
"loss": 0.2864,
"step": 255
},
{
"epoch": 0.07765830346475508,
"grad_norm": 0.4539343579685493,
"learning_rate": 0.00015522388059701495,
"loss": 0.3263,
"step": 260
},
{
"epoch": 0.07915173237753884,
"grad_norm": 0.4601786532582476,
"learning_rate": 0.00015820895522388059,
"loss": 0.3257,
"step": 265
},
{
"epoch": 0.08064516129032258,
"grad_norm": 0.4952572370631647,
"learning_rate": 0.00016119402985074628,
"loss": 0.3227,
"step": 270
},
{
"epoch": 0.08213859020310633,
"grad_norm": 0.5594600091045848,
"learning_rate": 0.00016417910447761195,
"loss": 0.3662,
"step": 275
},
{
"epoch": 0.08363201911589008,
"grad_norm": 0.4150317719098373,
"learning_rate": 0.00016716417910447761,
"loss": 0.251,
"step": 280
},
{
"epoch": 0.08512544802867383,
"grad_norm": 0.45530314509050923,
"learning_rate": 0.00017014925373134328,
"loss": 0.3277,
"step": 285
},
{
"epoch": 0.08661887694145759,
"grad_norm": 0.3707658637725497,
"learning_rate": 0.00017313432835820895,
"loss": 0.2958,
"step": 290
},
{
"epoch": 0.08811230585424133,
"grad_norm": 0.3532362466068546,
"learning_rate": 0.00017611940298507464,
"loss": 0.3275,
"step": 295
},
{
"epoch": 0.08960573476702509,
"grad_norm": 0.41975328143934704,
"learning_rate": 0.0001791044776119403,
"loss": 0.361,
"step": 300
},
{
"epoch": 0.09109916367980884,
"grad_norm": 0.3889061650864088,
"learning_rate": 0.00018208955223880598,
"loss": 0.2696,
"step": 305
},
{
"epoch": 0.09259259259259259,
"grad_norm": 0.46543359027880427,
"learning_rate": 0.00018507462686567165,
"loss": 0.3443,
"step": 310
},
{
"epoch": 0.09408602150537634,
"grad_norm": 0.5714057611200172,
"learning_rate": 0.00018805970149253734,
"loss": 0.3374,
"step": 315
},
{
"epoch": 0.0955794504181601,
"grad_norm": 0.3948901958880527,
"learning_rate": 0.00019104477611940298,
"loss": 0.321,
"step": 320
},
{
"epoch": 0.09707287933094384,
"grad_norm": 0.5540915005920514,
"learning_rate": 0.00019402985074626867,
"loss": 0.3277,
"step": 325
},
{
"epoch": 0.0985663082437276,
"grad_norm": 0.48042615725477694,
"learning_rate": 0.00019701492537313434,
"loss": 0.2833,
"step": 330
},
{
"epoch": 0.10005973715651136,
"grad_norm": 0.35451829573422367,
"learning_rate": 0.0002,
"loss": 0.2882,
"step": 335
},
{
"epoch": 0.1015531660692951,
"grad_norm": 0.37754764974968846,
"learning_rate": 0.00019999864102799163,
"loss": 0.2981,
"step": 340
},
{
"epoch": 0.10304659498207885,
"grad_norm": 0.41754484552953547,
"learning_rate": 0.0001999945641489025,
"loss": 0.3385,
"step": 345
},
{
"epoch": 0.10454002389486261,
"grad_norm": 0.48242379028499566,
"learning_rate": 0.00019998776947353995,
"loss": 0.3462,
"step": 350
},
{
"epoch": 0.10603345280764635,
"grad_norm": 0.43615681396287914,
"learning_rate": 0.00019997825718657945,
"loss": 0.2829,
"step": 355
},
{
"epoch": 0.10752688172043011,
"grad_norm": 0.5382667372932559,
"learning_rate": 0.0001999660275465596,
"loss": 0.3086,
"step": 360
},
{
"epoch": 0.10902031063321387,
"grad_norm": 0.4221095850996317,
"learning_rate": 0.00019995108088587528,
"loss": 0.2853,
"step": 365
},
{
"epoch": 0.11051373954599761,
"grad_norm": 0.4049511576507522,
"learning_rate": 0.00019993341761076824,
"loss": 0.3144,
"step": 370
},
{
"epoch": 0.11200716845878136,
"grad_norm": 0.41672237725910044,
"learning_rate": 0.00019991303820131645,
"loss": 0.3262,
"step": 375
},
{
"epoch": 0.1135005973715651,
"grad_norm": 0.37665369388894876,
"learning_rate": 0.00019988994321142088,
"loss": 0.2746,
"step": 380
},
{
"epoch": 0.11499402628434886,
"grad_norm": 0.38890047129761735,
"learning_rate": 0.00019986413326879035,
"loss": 0.3194,
"step": 385
},
{
"epoch": 0.11648745519713262,
"grad_norm": 0.37534439629256694,
"learning_rate": 0.00019983560907492476,
"loss": 0.3134,
"step": 390
},
{
"epoch": 0.11798088410991636,
"grad_norm": 0.3401949072629029,
"learning_rate": 0.00019980437140509563,
"loss": 0.3198,
"step": 395
},
{
"epoch": 0.11947431302270012,
"grad_norm": 0.4404312617632526,
"learning_rate": 0.00019977042110832537,
"loss": 0.3224,
"step": 400
},
{
"epoch": 0.11947431302270012,
"eval_loss": 0.29538044333457947,
"eval_runtime": 671.0671,
"eval_samples_per_second": 5.961,
"eval_steps_per_second": 0.745,
"step": 400
},
{
"epoch": 0.12096774193548387,
"grad_norm": 0.3877778527340937,
"learning_rate": 0.00019973375910736408,
"loss": 0.2605,
"step": 405
},
{
"epoch": 0.12246117084826762,
"grad_norm": 0.42449178944770155,
"learning_rate": 0.00019969438639866436,
"loss": 0.3292,
"step": 410
},
{
"epoch": 0.12395459976105137,
"grad_norm": 0.4421146156020614,
"learning_rate": 0.00019965230405235443,
"loss": 0.2847,
"step": 415
},
{
"epoch": 0.12544802867383512,
"grad_norm": 0.33313138965230243,
"learning_rate": 0.00019960751321220887,
"loss": 0.3346,
"step": 420
},
{
"epoch": 0.12694145758661887,
"grad_norm": 0.35768831528441764,
"learning_rate": 0.0001995600150956177,
"loss": 0.2896,
"step": 425
},
{
"epoch": 0.12843488649940263,
"grad_norm": 0.4073503627739374,
"learning_rate": 0.0001995098109935531,
"loss": 0.2963,
"step": 430
},
{
"epoch": 0.12992831541218638,
"grad_norm": 0.4634204409239979,
"learning_rate": 0.00019945690227053445,
"loss": 0.309,
"step": 435
},
{
"epoch": 0.13142174432497014,
"grad_norm": 0.48789887803899984,
"learning_rate": 0.00019940129036459121,
"loss": 0.2778,
"step": 440
},
{
"epoch": 0.13291517323775387,
"grad_norm": 0.41135714129715417,
"learning_rate": 0.0001993429767872239,
"loss": 0.312,
"step": 445
},
{
"epoch": 0.13440860215053763,
"grad_norm": 0.48911130889740795,
"learning_rate": 0.00019928196312336285,
"loss": 0.354,
"step": 450
},
{
"epoch": 0.13590203106332138,
"grad_norm": 0.31087210214054994,
"learning_rate": 0.00019921825103132531,
"loss": 0.2581,
"step": 455
},
{
"epoch": 0.13739545997610514,
"grad_norm": 0.34319695770593983,
"learning_rate": 0.00019915184224277032,
"loss": 0.2861,
"step": 460
},
{
"epoch": 0.1388888888888889,
"grad_norm": 0.39072017479958165,
"learning_rate": 0.00019908273856265152,
"loss": 0.28,
"step": 465
},
{
"epoch": 0.14038231780167265,
"grad_norm": 0.38108092212692735,
"learning_rate": 0.00019901094186916825,
"loss": 0.2938,
"step": 470
},
{
"epoch": 0.14187574671445638,
"grad_norm": 0.40284399729512566,
"learning_rate": 0.00019893645411371447,
"loss": 0.3232,
"step": 475
},
{
"epoch": 0.14336917562724014,
"grad_norm": 0.41081349055067273,
"learning_rate": 0.00019885927732082563,
"loss": 0.2682,
"step": 480
},
{
"epoch": 0.1448626045400239,
"grad_norm": 0.35453420174268324,
"learning_rate": 0.00019877941358812382,
"loss": 0.3099,
"step": 485
},
{
"epoch": 0.14635603345280765,
"grad_norm": 0.4219522172174877,
"learning_rate": 0.00019869686508626054,
"loss": 0.2929,
"step": 490
},
{
"epoch": 0.1478494623655914,
"grad_norm": 0.433146389991006,
"learning_rate": 0.00019861163405885787,
"loss": 0.3255,
"step": 495
},
{
"epoch": 0.14934289127837516,
"grad_norm": 0.43636524610968963,
"learning_rate": 0.0001985237228224474,
"loss": 0.3251,
"step": 500
},
{
"epoch": 0.1508363201911589,
"grad_norm": 0.39903987473575914,
"learning_rate": 0.00019843313376640732,
"loss": 0.2786,
"step": 505
},
{
"epoch": 0.15232974910394265,
"grad_norm": 0.42757498238366765,
"learning_rate": 0.0001983398693528975,
"loss": 0.3365,
"step": 510
},
{
"epoch": 0.1538231780167264,
"grad_norm": 0.4030570437823937,
"learning_rate": 0.00019824393211679246,
"loss": 0.2501,
"step": 515
},
{
"epoch": 0.15531660692951016,
"grad_norm": 0.3624497193987114,
"learning_rate": 0.00019814532466561259,
"loss": 0.2987,
"step": 520
},
{
"epoch": 0.15681003584229392,
"grad_norm": 0.4694503509924643,
"learning_rate": 0.00019804404967945315,
"loss": 0.3244,
"step": 525
},
{
"epoch": 0.15830346475507767,
"grad_norm": 0.43327424018713545,
"learning_rate": 0.00019794010991091164,
"loss": 0.2656,
"step": 530
},
{
"epoch": 0.1597968936678614,
"grad_norm": 0.36124038110578643,
"learning_rate": 0.00019783350818501272,
"loss": 0.2949,
"step": 535
},
{
"epoch": 0.16129032258064516,
"grad_norm": 0.4813284517384988,
"learning_rate": 0.00019772424739913168,
"loss": 0.2527,
"step": 540
},
{
"epoch": 0.1627837514934289,
"grad_norm": 0.517676779630334,
"learning_rate": 0.00019761233052291544,
"loss": 0.3322,
"step": 545
},
{
"epoch": 0.16427718040621267,
"grad_norm": 0.41586609149970155,
"learning_rate": 0.0001974977605982021,
"loss": 0.3058,
"step": 550
},
{
"epoch": 0.16577060931899643,
"grad_norm": 0.38142158790681063,
"learning_rate": 0.00019738054073893807,
"loss": 0.2617,
"step": 555
},
{
"epoch": 0.16726403823178015,
"grad_norm": 0.4508850152729104,
"learning_rate": 0.00019726067413109347,
"loss": 0.3124,
"step": 560
},
{
"epoch": 0.1687574671445639,
"grad_norm": 0.38711879954880984,
"learning_rate": 0.0001971381640325756,
"loss": 0.2886,
"step": 565
},
{
"epoch": 0.17025089605734767,
"grad_norm": 0.3713159311530026,
"learning_rate": 0.00019701301377314038,
"loss": 0.2884,
"step": 570
},
{
"epoch": 0.17174432497013142,
"grad_norm": 0.4253789489790291,
"learning_rate": 0.0001968852267543018,
"loss": 0.3403,
"step": 575
},
{
"epoch": 0.17323775388291518,
"grad_norm": 0.3716154947554444,
"learning_rate": 0.00019675480644923944,
"loss": 0.2563,
"step": 580
},
{
"epoch": 0.17473118279569894,
"grad_norm": 0.4067400864065092,
"learning_rate": 0.00019662175640270424,
"loss": 0.3139,
"step": 585
},
{
"epoch": 0.17622461170848266,
"grad_norm": 0.3587456607358574,
"learning_rate": 0.00019648608023092195,
"loss": 0.2851,
"step": 590
},
{
"epoch": 0.17771804062126642,
"grad_norm": 0.35869894756290216,
"learning_rate": 0.00019634778162149497,
"loss": 0.299,
"step": 595
},
{
"epoch": 0.17921146953405018,
"grad_norm": 0.4291965063428789,
"learning_rate": 0.00019620686433330207,
"loss": 0.3055,
"step": 600
},
{
"epoch": 0.17921146953405018,
"eval_loss": 0.28864020109176636,
"eval_runtime": 670.6057,
"eval_samples_per_second": 5.965,
"eval_steps_per_second": 0.746,
"step": 600
},
{
"epoch": 0.18070489844683393,
"grad_norm": 0.4107636120154602,
"learning_rate": 0.00019606333219639624,
"loss": 0.2674,
"step": 605
},
{
"epoch": 0.1821983273596177,
"grad_norm": 0.4094821475481708,
"learning_rate": 0.00019591718911190066,
"loss": 0.2748,
"step": 610
},
{
"epoch": 0.18369175627240145,
"grad_norm": 0.42712744392327107,
"learning_rate": 0.00019576843905190253,
"loss": 0.2913,
"step": 615
},
{
"epoch": 0.18518518518518517,
"grad_norm": 0.3838363588569511,
"learning_rate": 0.00019561708605934515,
"loss": 0.3019,
"step": 620
},
{
"epoch": 0.18667861409796893,
"grad_norm": 0.4060361946228662,
"learning_rate": 0.0001954631342479182,
"loss": 0.3028,
"step": 625
},
{
"epoch": 0.1881720430107527,
"grad_norm": 0.36895107070060834,
"learning_rate": 0.0001953065878019457,
"loss": 0.2912,
"step": 630
},
{
"epoch": 0.18966547192353644,
"grad_norm": 0.5389015474723242,
"learning_rate": 0.0001951474509762724,
"loss": 0.3277,
"step": 635
},
{
"epoch": 0.1911589008363202,
"grad_norm": 0.3713681816986798,
"learning_rate": 0.0001949857280961481,
"loss": 0.2564,
"step": 640
},
{
"epoch": 0.19265232974910393,
"grad_norm": 0.39787955963512267,
"learning_rate": 0.00019482142355711023,
"loss": 0.3167,
"step": 645
},
{
"epoch": 0.19414575866188769,
"grad_norm": 0.4160144539607265,
"learning_rate": 0.0001946545418248641,
"loss": 0.3473,
"step": 650
},
{
"epoch": 0.19563918757467144,
"grad_norm": 0.4611729979684992,
"learning_rate": 0.00019448508743516186,
"loss": 0.2579,
"step": 655
},
{
"epoch": 0.1971326164874552,
"grad_norm": 0.37926676547852983,
"learning_rate": 0.00019431306499367886,
"loss": 0.2995,
"step": 660
},
{
"epoch": 0.19862604540023895,
"grad_norm": 0.41885156529130685,
"learning_rate": 0.00019413847917588878,
"loss": 0.3042,
"step": 665
},
{
"epoch": 0.2001194743130227,
"grad_norm": 0.427624169132239,
"learning_rate": 0.00019396133472693642,
"loss": 0.3173,
"step": 670
},
{
"epoch": 0.20161290322580644,
"grad_norm": 0.38475710691405496,
"learning_rate": 0.00019378163646150876,
"loss": 0.3021,
"step": 675
},
{
"epoch": 0.2031063321385902,
"grad_norm": 0.4138321670681785,
"learning_rate": 0.000193599389263704,
"loss": 0.274,
"step": 680
},
{
"epoch": 0.20459976105137395,
"grad_norm": 0.355884010075492,
"learning_rate": 0.00019341459808689898,
"loss": 0.3393,
"step": 685
},
{
"epoch": 0.2060931899641577,
"grad_norm": 0.3631472347986104,
"learning_rate": 0.00019322726795361443,
"loss": 0.2949,
"step": 690
},
{
"epoch": 0.20758661887694146,
"grad_norm": 0.4773277681201716,
"learning_rate": 0.0001930374039553785,
"loss": 0.301,
"step": 695
},
{
"epoch": 0.20908004778972522,
"grad_norm": 0.448239132237338,
"learning_rate": 0.00019284501125258835,
"loss": 0.3332,
"step": 700
},
{
"epoch": 0.21057347670250895,
"grad_norm": 0.38679234194418,
"learning_rate": 0.00019265009507436997,
"loss": 0.2821,
"step": 705
},
{
"epoch": 0.2120669056152927,
"grad_norm": 0.34050133768678587,
"learning_rate": 0.00019245266071843596,
"loss": 0.2971,
"step": 710
},
{
"epoch": 0.21356033452807646,
"grad_norm": 0.377904449126756,
"learning_rate": 0.00019225271355094155,
"loss": 0.2644,
"step": 715
},
{
"epoch": 0.21505376344086022,
"grad_norm": 0.43091731111413145,
"learning_rate": 0.00019205025900633884,
"loss": 0.3028,
"step": 720
},
{
"epoch": 0.21654719235364397,
"grad_norm": 0.35823568635784103,
"learning_rate": 0.00019184530258722899,
"loss": 0.2941,
"step": 725
},
{
"epoch": 0.21804062126642773,
"grad_norm": 0.3871906521592633,
"learning_rate": 0.00019163784986421276,
"loss": 0.2453,
"step": 730
},
{
"epoch": 0.21953405017921146,
"grad_norm": 0.3870097665982384,
"learning_rate": 0.00019142790647573902,
"loss": 0.3163,
"step": 735
},
{
"epoch": 0.22102747909199522,
"grad_norm": 0.4043861350049887,
"learning_rate": 0.00019121547812795152,
"loss": 0.2867,
"step": 740
},
{
"epoch": 0.22252090800477897,
"grad_norm": 0.3472296001968241,
"learning_rate": 0.00019100057059453381,
"loss": 0.308,
"step": 745
},
{
"epoch": 0.22401433691756273,
"grad_norm": 0.4555451484017787,
"learning_rate": 0.00019078318971655237,
"loss": 0.3094,
"step": 750
},
{
"epoch": 0.22550776583034648,
"grad_norm": 0.39417763113385473,
"learning_rate": 0.00019056334140229777,
"loss": 0.2673,
"step": 755
},
{
"epoch": 0.2270011947431302,
"grad_norm": 0.33903321452810364,
"learning_rate": 0.00019034103162712408,
"loss": 0.318,
"step": 760
},
{
"epoch": 0.22849462365591397,
"grad_norm": 0.37450897393444954,
"learning_rate": 0.0001901162664332866,
"loss": 0.2543,
"step": 765
},
{
"epoch": 0.22998805256869773,
"grad_norm": 0.3798157275859589,
"learning_rate": 0.0001898890519297774,
"loss": 0.2945,
"step": 770
},
{
"epoch": 0.23148148148148148,
"grad_norm": 0.3682423785435909,
"learning_rate": 0.00018965939429215948,
"loss": 0.3127,
"step": 775
},
{
"epoch": 0.23297491039426524,
"grad_norm": 0.3771090235525355,
"learning_rate": 0.0001894272997623989,
"loss": 0.2722,
"step": 780
},
{
"epoch": 0.234468339307049,
"grad_norm": 0.33941544966890785,
"learning_rate": 0.00018919277464869504,
"loss": 0.3163,
"step": 785
},
{
"epoch": 0.23596176821983272,
"grad_norm": 0.37376605453569783,
"learning_rate": 0.0001889558253253092,
"loss": 0.2598,
"step": 790
},
{
"epoch": 0.23745519713261648,
"grad_norm": 0.38959423980866825,
"learning_rate": 0.00018871645823239128,
"loss": 0.2983,
"step": 795
},
{
"epoch": 0.23894862604540024,
"grad_norm": 0.38139963613722794,
"learning_rate": 0.00018847467987580493,
"loss": 0.2899,
"step": 800
},
{
"epoch": 0.23894862604540024,
"eval_loss": 0.2804073989391327,
"eval_runtime": 670.1656,
"eval_samples_per_second": 5.969,
"eval_steps_per_second": 0.746,
"step": 800
},
{
"epoch": 0.240442054958184,
"grad_norm": 0.40419199411712203,
"learning_rate": 0.00018823049682695052,
"loss": 0.3008,
"step": 805
},
{
"epoch": 0.24193548387096775,
"grad_norm": 0.3612377971129905,
"learning_rate": 0.0001879839157225866,
"loss": 0.316,
"step": 810
},
{
"epoch": 0.2434289127837515,
"grad_norm": 0.4605084068043848,
"learning_rate": 0.0001877349432646495,
"loss": 0.2776,
"step": 815
},
{
"epoch": 0.24492234169653523,
"grad_norm": 0.39686115392847526,
"learning_rate": 0.0001874835862200713,
"loss": 0.2664,
"step": 820
},
{
"epoch": 0.246415770609319,
"grad_norm": 0.38181276219849275,
"learning_rate": 0.00018722985142059572,
"loss": 0.2896,
"step": 825
},
{
"epoch": 0.24790919952210275,
"grad_norm": 0.36529483511621264,
"learning_rate": 0.0001869737457625926,
"loss": 0.2939,
"step": 830
},
{
"epoch": 0.2494026284348865,
"grad_norm": 0.4035594565736224,
"learning_rate": 0.00018671527620687034,
"loss": 0.3102,
"step": 835
},
{
"epoch": 0.25089605734767023,
"grad_norm": 0.3854541007588229,
"learning_rate": 0.00018645444977848677,
"loss": 0.2769,
"step": 840
},
{
"epoch": 0.252389486260454,
"grad_norm": 0.42123392936346155,
"learning_rate": 0.00018619127356655813,
"loss": 0.308,
"step": 845
},
{
"epoch": 0.25388291517323774,
"grad_norm": 0.49950892415773196,
"learning_rate": 0.0001859257547240666,
"loss": 0.314,
"step": 850
},
{
"epoch": 0.2553763440860215,
"grad_norm": 0.40906372202780966,
"learning_rate": 0.00018565790046766564,
"loss": 0.2881,
"step": 855
},
{
"epoch": 0.25686977299880526,
"grad_norm": 0.38764669514338385,
"learning_rate": 0.000185387718077484,
"loss": 0.3027,
"step": 860
},
{
"epoch": 0.258363201911589,
"grad_norm": 0.42262538628601043,
"learning_rate": 0.00018511521489692775,
"loss": 0.2545,
"step": 865
},
{
"epoch": 0.25985663082437277,
"grad_norm": 0.41209507390817657,
"learning_rate": 0.00018484039833248085,
"loss": 0.2828,
"step": 870
},
{
"epoch": 0.2613500597371565,
"grad_norm": 0.4469642279365962,
"learning_rate": 0.0001845632758535036,
"loss": 0.332,
"step": 875
},
{
"epoch": 0.2628434886499403,
"grad_norm": 0.36473061276227664,
"learning_rate": 0.00018428385499202988,
"loss": 0.2647,
"step": 880
},
{
"epoch": 0.26433691756272404,
"grad_norm": 0.4433634956798964,
"learning_rate": 0.00018400214334256227,
"loss": 0.2952,
"step": 885
},
{
"epoch": 0.26583034647550774,
"grad_norm": 0.3494167236815255,
"learning_rate": 0.00018371814856186572,
"loss": 0.2825,
"step": 890
},
{
"epoch": 0.2673237753882915,
"grad_norm": 0.38005445773872815,
"learning_rate": 0.00018343187836875928,
"loss": 0.3096,
"step": 895
},
{
"epoch": 0.26881720430107525,
"grad_norm": 0.5133589136001121,
"learning_rate": 0.00018314334054390664,
"loss": 0.3078,
"step": 900
},
{
"epoch": 0.270310633213859,
"grad_norm": 0.37745450032232525,
"learning_rate": 0.00018285254292960433,
"loss": 0.2687,
"step": 905
},
{
"epoch": 0.27180406212664276,
"grad_norm": 0.40501108505960953,
"learning_rate": 0.00018255949342956863,
"loss": 0.2969,
"step": 910
},
{
"epoch": 0.2732974910394265,
"grad_norm": 0.4103922981810604,
"learning_rate": 0.000182264200008721,
"loss": 0.2635,
"step": 915
},
{
"epoch": 0.2747909199522103,
"grad_norm": 0.8269261010911084,
"learning_rate": 0.00018196667069297123,
"loss": 0.2894,
"step": 920
},
{
"epoch": 0.27628434886499403,
"grad_norm": 0.5797869922585875,
"learning_rate": 0.0001816669135689996,
"loss": 0.3175,
"step": 925
},
{
"epoch": 0.2777777777777778,
"grad_norm": 0.3960727461959874,
"learning_rate": 0.00018136493678403686,
"loss": 0.2122,
"step": 930
},
{
"epoch": 0.27927120669056155,
"grad_norm": 0.3281979427197208,
"learning_rate": 0.00018106074854564306,
"loss": 0.2772,
"step": 935
},
{
"epoch": 0.2807646356033453,
"grad_norm": 0.3180009593314954,
"learning_rate": 0.00018075435712148417,
"loss": 0.2494,
"step": 940
},
{
"epoch": 0.28225806451612906,
"grad_norm": 0.5630380699907694,
"learning_rate": 0.00018044577083910758,
"loss": 0.287,
"step": 945
},
{
"epoch": 0.28375149342891276,
"grad_norm": 0.5821492853630149,
"learning_rate": 0.00018013499808571567,
"loss": 0.3064,
"step": 950
},
{
"epoch": 0.2852449223416965,
"grad_norm": 0.34343056721676674,
"learning_rate": 0.00017982204730793795,
"loss": 0.2479,
"step": 955
},
{
"epoch": 0.2867383512544803,
"grad_norm": 0.4045824911297299,
"learning_rate": 0.0001795069270116013,
"loss": 0.3042,
"step": 960
},
{
"epoch": 0.28823178016726403,
"grad_norm": 0.4423077224611283,
"learning_rate": 0.000179189645761499,
"loss": 0.2587,
"step": 965
},
{
"epoch": 0.2897252090800478,
"grad_norm": 0.3881310780181844,
"learning_rate": 0.00017887021218115782,
"loss": 0.2883,
"step": 970
},
{
"epoch": 0.29121863799283154,
"grad_norm": 0.4157883620231867,
"learning_rate": 0.00017854863495260354,
"loss": 0.3163,
"step": 975
},
{
"epoch": 0.2927120669056153,
"grad_norm": 0.5286260329472577,
"learning_rate": 0.00017822492281612532,
"loss": 0.2784,
"step": 980
},
{
"epoch": 0.29420549581839905,
"grad_norm": 0.43556243885834467,
"learning_rate": 0.00017789908457003777,
"loss": 0.2934,
"step": 985
},
{
"epoch": 0.2956989247311828,
"grad_norm": 0.46879974456717627,
"learning_rate": 0.000177571129070442,
"loss": 0.273,
"step": 990
},
{
"epoch": 0.29719235364396657,
"grad_norm": 0.3913970910916109,
"learning_rate": 0.00017724106523098486,
"loss": 0.2864,
"step": 995
},
{
"epoch": 0.2986857825567503,
"grad_norm": 0.5166170033604143,
"learning_rate": 0.00017690890202261676,
"loss": 0.3116,
"step": 1000
},
{
"epoch": 0.2986857825567503,
"eval_loss": 0.27715662121772766,
"eval_runtime": 670.7329,
"eval_samples_per_second": 5.964,
"eval_steps_per_second": 0.745,
"step": 1000
},
{
"epoch": 0.300179211469534,
"grad_norm": 0.49927259244291966,
"learning_rate": 0.00017657464847334775,
"loss": 0.2295,
"step": 1005
},
{
"epoch": 0.3016726403823178,
"grad_norm": 0.4015981411197979,
"learning_rate": 0.0001762383136680022,
"loss": 0.2807,
"step": 1010
},
{
"epoch": 0.30316606929510154,
"grad_norm": 0.33536145313741844,
"learning_rate": 0.00017589990674797171,
"loss": 0.2683,
"step": 1015
},
{
"epoch": 0.3046594982078853,
"grad_norm": 0.4291404869332869,
"learning_rate": 0.00017555943691096706,
"loss": 0.2704,
"step": 1020
},
{
"epoch": 0.30615292712066905,
"grad_norm": 0.47290922277355507,
"learning_rate": 0.00017521691341076774,
"loss": 0.3136,
"step": 1025
},
{
"epoch": 0.3076463560334528,
"grad_norm": 0.4688378459680159,
"learning_rate": 0.00017487234555697072,
"loss": 0.272,
"step": 1030
},
{
"epoch": 0.30913978494623656,
"grad_norm": 0.36206038691433684,
"learning_rate": 0.0001745257427147374,
"loss": 0.2949,
"step": 1035
},
{
"epoch": 0.3106332138590203,
"grad_norm": 0.34718258949903247,
"learning_rate": 0.00017417711430453897,
"loss": 0.2724,
"step": 1040
},
{
"epoch": 0.3121266427718041,
"grad_norm": 0.35896609288857995,
"learning_rate": 0.00017382646980190048,
"loss": 0.2893,
"step": 1045
},
{
"epoch": 0.31362007168458783,
"grad_norm": 0.4371043853081932,
"learning_rate": 0.00017347381873714316,
"loss": 0.2874,
"step": 1050
},
{
"epoch": 0.3151135005973716,
"grad_norm": 0.33305502199179604,
"learning_rate": 0.00017311917069512555,
"loss": 0.2329,
"step": 1055
},
{
"epoch": 0.31660692951015534,
"grad_norm": 0.43013382325716004,
"learning_rate": 0.00017276253531498293,
"loss": 0.312,
"step": 1060
},
{
"epoch": 0.31810035842293904,
"grad_norm": 0.347929106798288,
"learning_rate": 0.00017240392228986518,
"loss": 0.2648,
"step": 1065
},
{
"epoch": 0.3195937873357228,
"grad_norm": 0.3492181786101714,
"learning_rate": 0.00017204334136667365,
"loss": 0.2821,
"step": 1070
},
{
"epoch": 0.32108721624850656,
"grad_norm": 0.45792503261610795,
"learning_rate": 0.0001716808023457959,
"loss": 0.3277,
"step": 1075
},
{
"epoch": 0.3225806451612903,
"grad_norm": 0.3654545092753955,
"learning_rate": 0.00017131631508083962,
"loss": 0.2675,
"step": 1080
},
{
"epoch": 0.32407407407407407,
"grad_norm": 0.37947758138241466,
"learning_rate": 0.0001709498894783646,
"loss": 0.2893,
"step": 1085
},
{
"epoch": 0.3255675029868578,
"grad_norm": 0.4068572595536966,
"learning_rate": 0.0001705815354976135,
"loss": 0.2435,
"step": 1090
},
{
"epoch": 0.3270609318996416,
"grad_norm": 0.4809381438376721,
"learning_rate": 0.00017021126315024145,
"loss": 0.2985,
"step": 1095
},
{
"epoch": 0.32855436081242534,
"grad_norm": 0.40059623315230986,
"learning_rate": 0.00016983908250004344,
"loss": 0.273,
"step": 1100
},
{
"epoch": 0.3300477897252091,
"grad_norm": 0.4052544383737238,
"learning_rate": 0.00016946500366268123,
"loss": 0.2433,
"step": 1105
},
{
"epoch": 0.33154121863799285,
"grad_norm": 0.3560399096095347,
"learning_rate": 0.0001690890368054082,
"loss": 0.2932,
"step": 1110
},
{
"epoch": 0.3330346475507766,
"grad_norm": 0.4156230656107325,
"learning_rate": 0.00016871119214679304,
"loss": 0.2685,
"step": 1115
},
{
"epoch": 0.3345280764635603,
"grad_norm": 0.3271097281515262,
"learning_rate": 0.00016833147995644202,
"loss": 0.3102,
"step": 1120
},
{
"epoch": 0.33602150537634407,
"grad_norm": 0.4217178620957521,
"learning_rate": 0.00016794991055471992,
"loss": 0.293,
"step": 1125
},
{
"epoch": 0.3375149342891278,
"grad_norm": 0.37416501140273456,
"learning_rate": 0.00016756649431246953,
"loss": 0.2479,
"step": 1130
},
{
"epoch": 0.3390083632019116,
"grad_norm": 0.45055007223505494,
"learning_rate": 0.00016718124165072953,
"loss": 0.2836,
"step": 1135
},
{
"epoch": 0.34050179211469533,
"grad_norm": 0.40163491862797124,
"learning_rate": 0.0001667941630404517,
"loss": 0.2484,
"step": 1140
},
{
"epoch": 0.3419952210274791,
"grad_norm": 0.41798757856275687,
"learning_rate": 0.00016640526900221593,
"loss": 0.2881,
"step": 1145
},
{
"epoch": 0.34348864994026285,
"grad_norm": 0.38258918057820673,
"learning_rate": 0.00016601457010594447,
"loss": 0.3107,
"step": 1150
},
{
"epoch": 0.3449820788530466,
"grad_norm": 0.3845197614655172,
"learning_rate": 0.0001656220769706146,
"loss": 0.2848,
"step": 1155
},
{
"epoch": 0.34647550776583036,
"grad_norm": 0.3985931675974656,
"learning_rate": 0.00016522780026397,
"loss": 0.2887,
"step": 1160
},
{
"epoch": 0.3479689366786141,
"grad_norm": 0.35770417481369043,
"learning_rate": 0.00016483175070223081,
"loss": 0.2504,
"step": 1165
},
{
"epoch": 0.34946236559139787,
"grad_norm": 0.4093106514843705,
"learning_rate": 0.00016443393904980242,
"loss": 0.2898,
"step": 1170
},
{
"epoch": 0.35095579450418163,
"grad_norm": 0.4539044394175168,
"learning_rate": 0.00016403437611898282,
"loss": 0.2809,
"step": 1175
},
{
"epoch": 0.35244922341696533,
"grad_norm": 0.36797578012073034,
"learning_rate": 0.0001636330727696688,
"loss": 0.2529,
"step": 1180
},
{
"epoch": 0.3539426523297491,
"grad_norm": 0.3182084185670439,
"learning_rate": 0.00016323003990906072,
"loss": 0.3142,
"step": 1185
},
{
"epoch": 0.35543608124253284,
"grad_norm": 0.33781054620192646,
"learning_rate": 0.00016282528849136612,
"loss": 0.2403,
"step": 1190
},
{
"epoch": 0.3569295101553166,
"grad_norm": 0.3880915127154374,
"learning_rate": 0.0001624188295175019,
"loss": 0.2999,
"step": 1195
},
{
"epoch": 0.35842293906810035,
"grad_norm": 0.359774759483394,
"learning_rate": 0.00016201067403479543,
"loss": 0.3101,
"step": 1200
},
{
"epoch": 0.35842293906810035,
"eval_loss": 0.2727881968021393,
"eval_runtime": 672.0857,
"eval_samples_per_second": 5.952,
"eval_steps_per_second": 0.744,
"step": 1200
},
{
"epoch": 0.3599163679808841,
"grad_norm": 0.388508285355121,
"learning_rate": 0.0001616008331366843,
"loss": 0.2564,
"step": 1205
},
{
"epoch": 0.36140979689366787,
"grad_norm": 0.36408140890196194,
"learning_rate": 0.00016118931796241457,
"loss": 0.3252,
"step": 1210
},
{
"epoch": 0.3629032258064516,
"grad_norm": 0.41167492762052504,
"learning_rate": 0.0001607761396967384,
"loss": 0.26,
"step": 1215
},
{
"epoch": 0.3643966547192354,
"grad_norm": 0.5248867549730605,
"learning_rate": 0.00016036130956960967,
"loss": 0.2655,
"step": 1220
},
{
"epoch": 0.36589008363201914,
"grad_norm": 0.49678835879333444,
"learning_rate": 0.00015994483885587902,
"loss": 0.3085,
"step": 1225
},
{
"epoch": 0.3673835125448029,
"grad_norm": 0.3988902390012257,
"learning_rate": 0.0001595267388749873,
"loss": 0.2607,
"step": 1230
},
{
"epoch": 0.3688769414575866,
"grad_norm": 0.3925797006591169,
"learning_rate": 0.0001591070209906579,
"loss": 0.2997,
"step": 1235
},
{
"epoch": 0.37037037037037035,
"grad_norm": 0.3107954625386247,
"learning_rate": 0.000158685696610588,
"loss": 0.2537,
"step": 1240
},
{
"epoch": 0.3718637992831541,
"grad_norm": 0.41712837863937413,
"learning_rate": 0.00015826277718613833,
"loss": 0.3003,
"step": 1245
},
{
"epoch": 0.37335722819593786,
"grad_norm": 0.3630150573236221,
"learning_rate": 0.00015783827421202214,
"loss": 0.3056,
"step": 1250
},
{
"epoch": 0.3748506571087216,
"grad_norm": 0.38258678014490427,
"learning_rate": 0.00015741219922599253,
"loss": 0.2641,
"step": 1255
},
{
"epoch": 0.3763440860215054,
"grad_norm": 0.42701857998693166,
"learning_rate": 0.00015698456380852915,
"loss": 0.3068,
"step": 1260
},
{
"epoch": 0.37783751493428913,
"grad_norm": 0.4380649167022928,
"learning_rate": 0.00015655537958252324,
"loss": 0.2576,
"step": 1265
},
{
"epoch": 0.3793309438470729,
"grad_norm": 0.4082068880960036,
"learning_rate": 0.00015612465821296175,
"loss": 0.2953,
"step": 1270
},
{
"epoch": 0.38082437275985664,
"grad_norm": 0.4190875138856109,
"learning_rate": 0.00015569241140661047,
"loss": 0.3032,
"step": 1275
},
{
"epoch": 0.3823178016726404,
"grad_norm": 0.36568309245691916,
"learning_rate": 0.0001552586509116955,
"loss": 0.2195,
"step": 1280
},
{
"epoch": 0.38381123058542416,
"grad_norm": 0.44216085615223133,
"learning_rate": 0.0001548233885175843,
"loss": 0.3073,
"step": 1285
},
{
"epoch": 0.38530465949820786,
"grad_norm": 0.24922873576347065,
"learning_rate": 0.00015438663605446507,
"loss": 0.2472,
"step": 1290
},
{
"epoch": 0.3867980884109916,
"grad_norm": 0.41662099994810825,
"learning_rate": 0.00015394840539302527,
"loss": 0.2868,
"step": 1295
},
{
"epoch": 0.38829151732377537,
"grad_norm": 0.39646098577140865,
"learning_rate": 0.0001535087084441289,
"loss": 0.2855,
"step": 1300
},
{
"epoch": 0.3897849462365591,
"grad_norm": 0.36239571418608835,
"learning_rate": 0.00015306755715849293,
"loss": 0.227,
"step": 1305
},
{
"epoch": 0.3912783751493429,
"grad_norm": 0.35793958618986693,
"learning_rate": 0.0001526249635263623,
"loss": 0.2888,
"step": 1310
},
{
"epoch": 0.39277180406212664,
"grad_norm": 0.3783233493544514,
"learning_rate": 0.00015218093957718415,
"loss": 0.2605,
"step": 1315
},
{
"epoch": 0.3942652329749104,
"grad_norm": 0.37987257135077446,
"learning_rate": 0.00015173549737928084,
"loss": 0.2854,
"step": 1320
},
{
"epoch": 0.39575866188769415,
"grad_norm": 0.3892605729869642,
"learning_rate": 0.0001512886490395219,
"loss": 0.2918,
"step": 1325
},
{
"epoch": 0.3972520908004779,
"grad_norm": 0.41196692553982217,
"learning_rate": 0.00015084040670299516,
"loss": 0.2557,
"step": 1330
},
{
"epoch": 0.39874551971326166,
"grad_norm": 0.3254462520604089,
"learning_rate": 0.00015039078255267628,
"loss": 0.3088,
"step": 1335
},
{
"epoch": 0.4002389486260454,
"grad_norm": 0.40781344675681797,
"learning_rate": 0.00014993978880909796,
"loss": 0.2805,
"step": 1340
},
{
"epoch": 0.4017323775388292,
"grad_norm": 0.39304465117259413,
"learning_rate": 0.00014948743773001772,
"loss": 0.2843,
"step": 1345
},
{
"epoch": 0.4032258064516129,
"grad_norm": 0.48556741767644995,
"learning_rate": 0.00014903374161008464,
"loss": 0.3051,
"step": 1350
},
{
"epoch": 0.40471923536439663,
"grad_norm": 0.3663553575319031,
"learning_rate": 0.0001485787127805052,
"loss": 0.2336,
"step": 1355
},
{
"epoch": 0.4062126642771804,
"grad_norm": 0.4541411950152405,
"learning_rate": 0.00014812236360870834,
"loss": 0.2854,
"step": 1360
},
{
"epoch": 0.40770609318996415,
"grad_norm": 0.34795001206833953,
"learning_rate": 0.00014766470649800904,
"loss": 0.26,
"step": 1365
},
{
"epoch": 0.4091995221027479,
"grad_norm": 0.35146295322347754,
"learning_rate": 0.00014720575388727132,
"loss": 0.266,
"step": 1370
},
{
"epoch": 0.41069295101553166,
"grad_norm": 0.44633677983363984,
"learning_rate": 0.00014674551825057024,
"loss": 0.2853,
"step": 1375
},
{
"epoch": 0.4121863799283154,
"grad_norm": 0.3830303196674387,
"learning_rate": 0.0001462840120968527,
"loss": 0.2515,
"step": 1380
},
{
"epoch": 0.4136798088410992,
"grad_norm": 0.39604955470206826,
"learning_rate": 0.00014582124796959765,
"loss": 0.2751,
"step": 1385
},
{
"epoch": 0.41517323775388293,
"grad_norm": 0.3633797549153069,
"learning_rate": 0.00014535723844647503,
"loss": 0.2531,
"step": 1390
},
{
"epoch": 0.4166666666666667,
"grad_norm": 0.6020608532163392,
"learning_rate": 0.00014489199613900385,
"loss": 0.2766,
"step": 1395
},
{
"epoch": 0.41816009557945044,
"grad_norm": 0.38363605454823274,
"learning_rate": 0.0001444255336922095,
"loss": 0.2913,
"step": 1400
},
{
"epoch": 0.41816009557945044,
"eval_loss": 0.2678522765636444,
"eval_runtime": 671.8286,
"eval_samples_per_second": 5.954,
"eval_steps_per_second": 0.744,
"step": 1400
},
{
"epoch": 0.41965352449223414,
"grad_norm": 0.34726800590023943,
"learning_rate": 0.00014395786378428033,
"loss": 0.2308,
"step": 1405
},
{
"epoch": 0.4211469534050179,
"grad_norm": 0.3445589635670954,
"learning_rate": 0.00014348899912622248,
"loss": 0.306,
"step": 1410
},
{
"epoch": 0.42264038231780165,
"grad_norm": 0.3573121569173274,
"learning_rate": 0.0001430189524615149,
"loss": 0.2669,
"step": 1415
},
{
"epoch": 0.4241338112305854,
"grad_norm": 0.40102571413373606,
"learning_rate": 0.0001425477365657628,
"loss": 0.2701,
"step": 1420
},
{
"epoch": 0.42562724014336917,
"grad_norm": 0.3591108084356434,
"learning_rate": 0.0001420753642463504,
"loss": 0.2866,
"step": 1425
},
{
"epoch": 0.4271206690561529,
"grad_norm": 0.38049883203567986,
"learning_rate": 0.00014160184834209296,
"loss": 0.267,
"step": 1430
},
{
"epoch": 0.4286140979689367,
"grad_norm": 0.34901096097918405,
"learning_rate": 0.0001411272017228876,
"loss": 0.2772,
"step": 1435
},
{
"epoch": 0.43010752688172044,
"grad_norm": 0.43110083401740446,
"learning_rate": 0.0001406514372893637,
"loss": 0.2773,
"step": 1440
},
{
"epoch": 0.4316009557945042,
"grad_norm": 0.49928896541185935,
"learning_rate": 0.0001401745679725323,
"loss": 0.3121,
"step": 1445
},
{
"epoch": 0.43309438470728795,
"grad_norm": 0.38753869075919367,
"learning_rate": 0.00013969660673343435,
"loss": 0.2809,
"step": 1450
},
{
"epoch": 0.4345878136200717,
"grad_norm": 0.431047229984796,
"learning_rate": 0.00013921756656278884,
"loss": 0.246,
"step": 1455
},
{
"epoch": 0.43608124253285546,
"grad_norm": 0.33787432704790127,
"learning_rate": 0.00013873746048063943,
"loss": 0.2905,
"step": 1460
},
{
"epoch": 0.43757467144563916,
"grad_norm": 0.5450443166622904,
"learning_rate": 0.00013825630153600058,
"loss": 0.2439,
"step": 1465
},
{
"epoch": 0.4390681003584229,
"grad_norm": 0.46991031442183334,
"learning_rate": 0.0001377741028065031,
"loss": 0.2916,
"step": 1470
},
{
"epoch": 0.4405615292712067,
"grad_norm": 0.43373701568669226,
"learning_rate": 0.00013729087739803846,
"loss": 0.3018,
"step": 1475
},
{
"epoch": 0.44205495818399043,
"grad_norm": 0.41610864937340164,
"learning_rate": 0.0001368066384444028,
"loss": 0.2252,
"step": 1480
},
{
"epoch": 0.4435483870967742,
"grad_norm": 0.3708263427774675,
"learning_rate": 0.0001363213991069397,
"loss": 0.2644,
"step": 1485
},
{
"epoch": 0.44504181600955794,
"grad_norm": 0.37850068431288725,
"learning_rate": 0.00013583517257418278,
"loss": 0.2641,
"step": 1490
},
{
"epoch": 0.4465352449223417,
"grad_norm": 0.4878849200428849,
"learning_rate": 0.000135347972061497,
"loss": 0.2722,
"step": 1495
},
{
"epoch": 0.44802867383512546,
"grad_norm": 0.3890166450566535,
"learning_rate": 0.00013485981081071949,
"loss": 0.2876,
"step": 1500
},
{
"epoch": 0.4495221027479092,
"grad_norm": 0.4698494336715817,
"learning_rate": 0.00013437070208979974,
"loss": 0.2618,
"step": 1505
},
{
"epoch": 0.45101553166069297,
"grad_norm": 0.42945392994548376,
"learning_rate": 0.000133880659192439,
"loss": 0.2887,
"step": 1510
},
{
"epoch": 0.4525089605734767,
"grad_norm": 0.36608609898718975,
"learning_rate": 0.00013338969543772892,
"loss": 0.2309,
"step": 1515
},
{
"epoch": 0.4540023894862604,
"grad_norm": 0.31719911951791674,
"learning_rate": 0.00013289782416978947,
"loss": 0.2669,
"step": 1520
},
{
"epoch": 0.4554958183990442,
"grad_norm": 0.44642248141095303,
"learning_rate": 0.0001324050587574063,
"loss": 0.2638,
"step": 1525
},
{
"epoch": 0.45698924731182794,
"grad_norm": 0.4094419878172264,
"learning_rate": 0.0001319114125936675,
"loss": 0.2647,
"step": 1530
},
{
"epoch": 0.4584826762246117,
"grad_norm": 0.48369368762161424,
"learning_rate": 0.00013141689909559943,
"loss": 0.298,
"step": 1535
},
{
"epoch": 0.45997610513739545,
"grad_norm": 0.4137701154348147,
"learning_rate": 0.00013092153170380212,
"loss": 0.2706,
"step": 1540
},
{
"epoch": 0.4614695340501792,
"grad_norm": 0.4125118524580438,
"learning_rate": 0.00013042532388208398,
"loss": 0.2769,
"step": 1545
},
{
"epoch": 0.46296296296296297,
"grad_norm": 0.3835331195333388,
"learning_rate": 0.0001299282891170958,
"loss": 0.29,
"step": 1550
},
{
"epoch": 0.4644563918757467,
"grad_norm": 0.4050806911084048,
"learning_rate": 0.00012943044091796418,
"loss": 0.2733,
"step": 1555
},
{
"epoch": 0.4659498207885305,
"grad_norm": 0.3686929748618163,
"learning_rate": 0.00012893179281592453,
"loss": 0.299,
"step": 1560
},
{
"epoch": 0.46744324970131423,
"grad_norm": 0.3713013220008561,
"learning_rate": 0.0001284323583639531,
"loss": 0.2514,
"step": 1565
},
{
"epoch": 0.468936678614098,
"grad_norm": 0.37363712594365256,
"learning_rate": 0.00012793215113639862,
"loss": 0.278,
"step": 1570
},
{
"epoch": 0.47043010752688175,
"grad_norm": 0.3759963829551066,
"learning_rate": 0.00012743118472861365,
"loss": 0.2819,
"step": 1575
},
{
"epoch": 0.47192353643966545,
"grad_norm": 0.38520117361685985,
"learning_rate": 0.00012692947275658455,
"loss": 0.2345,
"step": 1580
},
{
"epoch": 0.4734169653524492,
"grad_norm": 0.3075498245821686,
"learning_rate": 0.00012642702885656192,
"loss": 0.263,
"step": 1585
},
{
"epoch": 0.47491039426523296,
"grad_norm": 0.3841403461793403,
"learning_rate": 0.00012592386668468968,
"loss": 0.2913,
"step": 1590
},
{
"epoch": 0.4764038231780167,
"grad_norm": 0.3367807617111592,
"learning_rate": 0.00012541999991663388,
"loss": 0.2801,
"step": 1595
},
{
"epoch": 0.4778972520908005,
"grad_norm": 0.378016075471954,
"learning_rate": 0.00012491544224721136,
"loss": 0.2765,
"step": 1600
},
{
"epoch": 0.4778972520908005,
"eval_loss": 0.2625010311603546,
"eval_runtime": 671.1626,
"eval_samples_per_second": 5.96,
"eval_steps_per_second": 0.745,
"step": 1600
},
{
"epoch": 0.47939068100358423,
"grad_norm": 0.3691633610486676,
"learning_rate": 0.00012441020739001698,
"loss": 0.2793,
"step": 1605
},
{
"epoch": 0.480884109916368,
"grad_norm": 0.3264004788515136,
"learning_rate": 0.00012390430907705134,
"loss": 0.2683,
"step": 1610
},
{
"epoch": 0.48237753882915174,
"grad_norm": 0.35913080769203726,
"learning_rate": 0.00012339776105834744,
"loss": 0.229,
"step": 1615
},
{
"epoch": 0.4838709677419355,
"grad_norm": 0.45935643829710515,
"learning_rate": 0.00012289057710159672,
"loss": 0.2714,
"step": 1620
},
{
"epoch": 0.48536439665471925,
"grad_norm": 0.3879403177186618,
"learning_rate": 0.00012238277099177526,
"loss": 0.2829,
"step": 1625
},
{
"epoch": 0.486857825567503,
"grad_norm": 0.3514395181766118,
"learning_rate": 0.00012187435653076889,
"loss": 0.249,
"step": 1630
},
{
"epoch": 0.4883512544802867,
"grad_norm": 0.3914277195367492,
"learning_rate": 0.0001213653475369979,
"loss": 0.2862,
"step": 1635
},
{
"epoch": 0.48984468339307047,
"grad_norm": 0.468390381134599,
"learning_rate": 0.00012085575784504191,
"loss": 0.2307,
"step": 1640
},
{
"epoch": 0.4913381123058542,
"grad_norm": 0.42676427345496687,
"learning_rate": 0.0001203456013052634,
"loss": 0.2963,
"step": 1645
},
{
"epoch": 0.492831541218638,
"grad_norm": 0.4042777140715105,
"learning_rate": 0.0001198348917834315,
"loss": 0.3155,
"step": 1650
},
{
"epoch": 0.49432497013142174,
"grad_norm": 0.333904149664281,
"learning_rate": 0.00011932364316034514,
"loss": 0.2433,
"step": 1655
},
{
"epoch": 0.4958183990442055,
"grad_norm": 0.41506264824806743,
"learning_rate": 0.0001188118693314557,
"loss": 0.2806,
"step": 1660
},
{
"epoch": 0.49731182795698925,
"grad_norm": 0.4293221243565824,
"learning_rate": 0.0001182995842064893,
"loss": 0.2582,
"step": 1665
},
{
"epoch": 0.498805256869773,
"grad_norm": 0.39173839921627196,
"learning_rate": 0.00011778680170906888,
"loss": 0.2886,
"step": 1670
},
{
"epoch": 0.5002986857825568,
"grad_norm": 0.47362247526518175,
"learning_rate": 0.0001172735357763356,
"loss": 0.293,
"step": 1675
},
{
"epoch": 0.5017921146953405,
"grad_norm": 0.4176472433369269,
"learning_rate": 0.00011675980035857019,
"loss": 0.2091,
"step": 1680
},
{
"epoch": 0.5032855436081243,
"grad_norm": 0.39087371326818915,
"learning_rate": 0.00011624560941881371,
"loss": 0.3164,
"step": 1685
},
{
"epoch": 0.504778972520908,
"grad_norm": 0.3444686695813055,
"learning_rate": 0.00011573097693248805,
"loss": 0.2601,
"step": 1690
},
{
"epoch": 0.5062724014336918,
"grad_norm": 0.3629126474420232,
"learning_rate": 0.00011521591688701605,
"loss": 0.2677,
"step": 1695
},
{
"epoch": 0.5077658303464755,
"grad_norm": 0.3677730257235259,
"learning_rate": 0.00011470044328144143,
"loss": 0.2845,
"step": 1700
},
{
"epoch": 0.5092592592592593,
"grad_norm": 0.4484524597840905,
"learning_rate": 0.0001141845701260482,
"loss": 0.2535,
"step": 1705
},
{
"epoch": 0.510752688172043,
"grad_norm": 0.44520906295668633,
"learning_rate": 0.0001136683114419799,
"loss": 0.2923,
"step": 1710
},
{
"epoch": 0.5122461170848268,
"grad_norm": 0.3843678110888455,
"learning_rate": 0.00011315168126085857,
"loss": 0.2419,
"step": 1715
},
{
"epoch": 0.5137395459976105,
"grad_norm": 0.3424866175556353,
"learning_rate": 0.00011263469362440331,
"loss": 0.2722,
"step": 1720
},
{
"epoch": 0.5152329749103942,
"grad_norm": 0.3664503507774818,
"learning_rate": 0.00011211736258404864,
"loss": 0.2652,
"step": 1725
},
{
"epoch": 0.516726403823178,
"grad_norm": 0.3882647311598004,
"learning_rate": 0.00011159970220056262,
"loss": 0.2613,
"step": 1730
},
{
"epoch": 0.5182198327359617,
"grad_norm": 0.3934259702377591,
"learning_rate": 0.00011108172654366467,
"loss": 0.2744,
"step": 1735
},
{
"epoch": 0.5197132616487455,
"grad_norm": 0.38348245744708515,
"learning_rate": 0.00011056344969164317,
"loss": 0.2484,
"step": 1740
},
{
"epoch": 0.5212066905615292,
"grad_norm": 0.4552854124473697,
"learning_rate": 0.0001100448857309728,
"loss": 0.2559,
"step": 1745
},
{
"epoch": 0.522700119474313,
"grad_norm": 0.39439256103140385,
"learning_rate": 0.00010952604875593171,
"loss": 0.2724,
"step": 1750
},
{
"epoch": 0.5241935483870968,
"grad_norm": 0.42701555614159337,
"learning_rate": 0.00010900695286821843,
"loss": 0.2284,
"step": 1755
},
{
"epoch": 0.5256869772998806,
"grad_norm": 0.3685725297263498,
"learning_rate": 0.00010848761217656856,
"loss": 0.2897,
"step": 1760
},
{
"epoch": 0.5271804062126643,
"grad_norm": 0.40798420628174437,
"learning_rate": 0.00010796804079637137,
"loss": 0.2689,
"step": 1765
},
{
"epoch": 0.5286738351254481,
"grad_norm": 0.4141834988836477,
"learning_rate": 0.0001074482528492861,
"loss": 0.2833,
"step": 1770
},
{
"epoch": 0.5301672640382318,
"grad_norm": 0.391053467201034,
"learning_rate": 0.00010692826246285815,
"loss": 0.2783,
"step": 1775
},
{
"epoch": 0.5316606929510155,
"grad_norm": 0.3660707237272847,
"learning_rate": 0.00010640808377013509,
"loss": 0.2504,
"step": 1780
},
{
"epoch": 0.5331541218637993,
"grad_norm": 0.29960550920174905,
"learning_rate": 0.00010588773090928268,
"loss": 0.2773,
"step": 1785
},
{
"epoch": 0.534647550776583,
"grad_norm": 0.49371001212196336,
"learning_rate": 0.00010536721802320027,
"loss": 0.2543,
"step": 1790
},
{
"epoch": 0.5361409796893668,
"grad_norm": 0.381676599509551,
"learning_rate": 0.00010484655925913669,
"loss": 0.3047,
"step": 1795
},
{
"epoch": 0.5376344086021505,
"grad_norm": 0.4165684073566037,
"learning_rate": 0.00010432576876830576,
"loss": 0.2697,
"step": 1800
},
{
"epoch": 0.5376344086021505,
"eval_loss": 0.2601146101951599,
"eval_runtime": 671.0756,
"eval_samples_per_second": 5.961,
"eval_steps_per_second": 0.745,
"step": 1800
},
{
"epoch": 0.5391278375149343,
"grad_norm": 0.34164713060339763,
"learning_rate": 0.00010380486070550135,
"loss": 0.2613,
"step": 1805
},
{
"epoch": 0.540621266427718,
"grad_norm": 0.3608068205239928,
"learning_rate": 0.00010328384922871307,
"loss": 0.3023,
"step": 1810
},
{
"epoch": 0.5421146953405018,
"grad_norm": 0.4134220541761705,
"learning_rate": 0.0001027627484987412,
"loss": 0.2759,
"step": 1815
},
{
"epoch": 0.5436081242532855,
"grad_norm": 0.4380879213819027,
"learning_rate": 0.00010224157267881176,
"loss": 0.2999,
"step": 1820
},
{
"epoch": 0.5451015531660693,
"grad_norm": 0.41334705402523975,
"learning_rate": 0.00010172033593419184,
"loss": 0.2905,
"step": 1825
},
{
"epoch": 0.546594982078853,
"grad_norm": 0.33573837522797895,
"learning_rate": 0.00010119905243180432,
"loss": 0.2197,
"step": 1830
},
{
"epoch": 0.5480884109916367,
"grad_norm": 0.34530007854548944,
"learning_rate": 0.00010067773633984294,
"loss": 0.2838,
"step": 1835
},
{
"epoch": 0.5495818399044206,
"grad_norm": 0.39282261233632176,
"learning_rate": 0.00010015640182738733,
"loss": 0.2449,
"step": 1840
},
{
"epoch": 0.5510752688172043,
"grad_norm": 0.369443654894832,
"learning_rate": 9.963506306401753e-05,
"loss": 0.2641,
"step": 1845
},
{
"epoch": 0.5525686977299881,
"grad_norm": 0.3736606416875578,
"learning_rate": 9.911373421942939e-05,
"loss": 0.2877,
"step": 1850
},
{
"epoch": 0.5540621266427718,
"grad_norm": 0.4082824771004434,
"learning_rate": 9.859242946304903e-05,
"loss": 0.2462,
"step": 1855
},
{
"epoch": 0.5555555555555556,
"grad_norm": 0.37234188557058073,
"learning_rate": 9.807116296364783e-05,
"loss": 0.2647,
"step": 1860
},
{
"epoch": 0.5570489844683393,
"grad_norm": 0.3373920413107433,
"learning_rate": 9.754994888895744e-05,
"loss": 0.2481,
"step": 1865
},
{
"epoch": 0.5585424133811231,
"grad_norm": 0.3691166711599759,
"learning_rate": 9.702880140528466e-05,
"loss": 0.2651,
"step": 1870
},
{
"epoch": 0.5600358422939068,
"grad_norm": 0.4237027962737052,
"learning_rate": 9.650773467712632e-05,
"loss": 0.2976,
"step": 1875
},
{
"epoch": 0.5615292712066906,
"grad_norm": 0.3628296244980769,
"learning_rate": 9.598676286678434e-05,
"loss": 0.233,
"step": 1880
},
{
"epoch": 0.5630227001194743,
"grad_norm": 0.32168405334156497,
"learning_rate": 9.546590013398091e-05,
"loss": 0.2875,
"step": 1885
},
{
"epoch": 0.5645161290322581,
"grad_norm": 0.4011755917649591,
"learning_rate": 9.494516063547353e-05,
"loss": 0.2556,
"step": 1890
},
{
"epoch": 0.5660095579450418,
"grad_norm": 0.4531648719387392,
"learning_rate": 9.442455852467019e-05,
"loss": 0.2674,
"step": 1895
},
{
"epoch": 0.5675029868578255,
"grad_norm": 0.4603202269611572,
"learning_rate": 9.390410795124484e-05,
"loss": 0.2902,
"step": 1900
},
{
"epoch": 0.5689964157706093,
"grad_norm": 0.3873591567262209,
"learning_rate": 9.338382306075274e-05,
"loss": 0.2272,
"step": 1905
},
{
"epoch": 0.570489844683393,
"grad_norm": 0.41809794251121246,
"learning_rate": 9.286371799424584e-05,
"loss": 0.309,
"step": 1910
},
{
"epoch": 0.5719832735961768,
"grad_norm": 0.41735588072251223,
"learning_rate": 9.234380688788879e-05,
"loss": 0.2422,
"step": 1915
},
{
"epoch": 0.5734767025089605,
"grad_norm": 0.35987074653242784,
"learning_rate": 9.182410387257434e-05,
"loss": 0.2868,
"step": 1920
},
{
"epoch": 0.5749701314217444,
"grad_norm": 0.3381000166870348,
"learning_rate": 9.130462307353945e-05,
"loss": 0.2794,
"step": 1925
},
{
"epoch": 0.5764635603345281,
"grad_norm": 0.3412056255088943,
"learning_rate": 9.078537860998155e-05,
"loss": 0.2391,
"step": 1930
},
{
"epoch": 0.5779569892473119,
"grad_norm": 0.40246570101157014,
"learning_rate": 9.026638459467435e-05,
"loss": 0.2757,
"step": 1935
},
{
"epoch": 0.5794504181600956,
"grad_norm": 0.29830016511773766,
"learning_rate": 8.974765513358466e-05,
"loss": 0.2403,
"step": 1940
},
{
"epoch": 0.5809438470728794,
"grad_norm": 0.3411808872796586,
"learning_rate": 8.92292043254889e-05,
"loss": 0.2708,
"step": 1945
},
{
"epoch": 0.5824372759856631,
"grad_norm": 0.4133927108399512,
"learning_rate": 8.871104626158972e-05,
"loss": 0.2873,
"step": 1950
},
{
"epoch": 0.5839307048984468,
"grad_norm": 0.39290137528322705,
"learning_rate": 8.81931950251332e-05,
"loss": 0.2103,
"step": 1955
},
{
"epoch": 0.5854241338112306,
"grad_norm": 0.45501182269960977,
"learning_rate": 8.767566469102613e-05,
"loss": 0.2781,
"step": 1960
},
{
"epoch": 0.5869175627240143,
"grad_norm": 0.4189475808727597,
"learning_rate": 8.715846932545317e-05,
"loss": 0.2378,
"step": 1965
},
{
"epoch": 0.5884109916367981,
"grad_norm": 0.4085112647747901,
"learning_rate": 8.664162298549483e-05,
"loss": 0.2806,
"step": 1970
},
{
"epoch": 0.5899044205495818,
"grad_norm": 0.3732596924042426,
"learning_rate": 8.612513971874534e-05,
"loss": 0.2932,
"step": 1975
},
{
"epoch": 0.5913978494623656,
"grad_norm": 0.35250262119931725,
"learning_rate": 8.560903356293069e-05,
"loss": 0.2418,
"step": 1980
},
{
"epoch": 0.5928912783751493,
"grad_norm": 0.41366159645877715,
"learning_rate": 8.509331854552724e-05,
"loss": 0.2743,
"step": 1985
},
{
"epoch": 0.5943847072879331,
"grad_norm": 0.3464575423602374,
"learning_rate": 8.457800868338051e-05,
"loss": 0.2222,
"step": 1990
},
{
"epoch": 0.5958781362007168,
"grad_norm": 0.360808643210085,
"learning_rate": 8.406311798232408e-05,
"loss": 0.2625,
"step": 1995
},
{
"epoch": 0.5973715651135006,
"grad_norm": 0.4208234863436116,
"learning_rate": 8.354866043679887e-05,
"loss": 0.2759,
"step": 2000
},
{
"epoch": 0.5973715651135006,
"eval_loss": 0.25568902492523193,
"eval_runtime": 670.9224,
"eval_samples_per_second": 5.962,
"eval_steps_per_second": 0.745,
"step": 2000
},
{
"epoch": 0.5988649940262843,
"grad_norm": 0.5128576614444417,
"learning_rate": 8.303465002947302e-05,
"loss": 0.2265,
"step": 2005
},
{
"epoch": 0.600358422939068,
"grad_norm": 0.4246792251104683,
"learning_rate": 8.252110073086165e-05,
"loss": 0.2591,
"step": 2010
},
{
"epoch": 0.6018518518518519,
"grad_norm": 0.36587149536697655,
"learning_rate": 8.200802649894713e-05,
"loss": 0.2598,
"step": 2015
},
{
"epoch": 0.6033452807646356,
"grad_norm": 0.3560153774291601,
"learning_rate": 8.149544127879988e-05,
"loss": 0.26,
"step": 2020
},
{
"epoch": 0.6048387096774194,
"grad_norm": 0.3820268167566159,
"learning_rate": 8.098335900219929e-05,
"loss": 0.2812,
"step": 2025
},
{
"epoch": 0.6063321385902031,
"grad_norm": 0.4670838320418128,
"learning_rate": 8.047179358725487e-05,
"loss": 0.2398,
"step": 2030
},
{
"epoch": 0.6078255675029869,
"grad_norm": 0.4359410281219078,
"learning_rate": 7.99607589380282e-05,
"loss": 0.2973,
"step": 2035
},
{
"epoch": 0.6093189964157706,
"grad_norm": 0.3406561530367968,
"learning_rate": 7.945026894415504e-05,
"loss": 0.251,
"step": 2040
},
{
"epoch": 0.6108124253285544,
"grad_norm": 0.34818876087607525,
"learning_rate": 7.894033748046756e-05,
"loss": 0.2691,
"step": 2045
},
{
"epoch": 0.6123058542413381,
"grad_norm": 0.39942001844135094,
"learning_rate": 7.843097840661748e-05,
"loss": 0.2879,
"step": 2050
},
{
"epoch": 0.6137992831541219,
"grad_norm": 0.3618592441449212,
"learning_rate": 7.792220556669929e-05,
"loss": 0.223,
"step": 2055
},
{
"epoch": 0.6152927120669056,
"grad_norm": 0.4022080561354502,
"learning_rate": 7.741403278887397e-05,
"loss": 0.2785,
"step": 2060
},
{
"epoch": 0.6167861409796893,
"grad_norm": 0.26523185523371673,
"learning_rate": 7.690647388499305e-05,
"loss": 0.2482,
"step": 2065
},
{
"epoch": 0.6182795698924731,
"grad_norm": 0.33411219118251206,
"learning_rate": 7.639954265022349e-05,
"loss": 0.2703,
"step": 2070
},
{
"epoch": 0.6197729988052568,
"grad_norm": 0.35999133636969094,
"learning_rate": 7.589325286267241e-05,
"loss": 0.262,
"step": 2075
},
{
"epoch": 0.6212664277180406,
"grad_norm": 0.3853527518287356,
"learning_rate": 7.538761828301276e-05,
"loss": 0.2261,
"step": 2080
},
{
"epoch": 0.6227598566308243,
"grad_norm": 0.43335074980788474,
"learning_rate": 7.48826526541094e-05,
"loss": 0.254,
"step": 2085
},
{
"epoch": 0.6242532855436081,
"grad_norm": 0.3542418988641108,
"learning_rate": 7.437836970064545e-05,
"loss": 0.2334,
"step": 2090
},
{
"epoch": 0.6257467144563919,
"grad_norm": 0.398192032214991,
"learning_rate": 7.387478312874916e-05,
"loss": 0.2742,
"step": 2095
},
{
"epoch": 0.6272401433691757,
"grad_norm": 0.33547770368905083,
"learning_rate": 7.337190662562174e-05,
"loss": 0.2511,
"step": 2100
},
{
"epoch": 0.6287335722819594,
"grad_norm": 0.4179386434535002,
"learning_rate": 7.2869753859165e-05,
"loss": 0.2667,
"step": 2105
},
{
"epoch": 0.6302270011947432,
"grad_norm": 0.36708628994003195,
"learning_rate": 7.236833847760994e-05,
"loss": 0.2805,
"step": 2110
},
{
"epoch": 0.6317204301075269,
"grad_norm": 0.3128895549559772,
"learning_rate": 7.186767410914601e-05,
"loss": 0.2477,
"step": 2115
},
{
"epoch": 0.6332138590203107,
"grad_norm": 0.3308033239038678,
"learning_rate": 7.136777436155041e-05,
"loss": 0.2559,
"step": 2120
},
{
"epoch": 0.6347072879330944,
"grad_norm": 0.36157519220837353,
"learning_rate": 7.086865282181841e-05,
"loss": 0.2747,
"step": 2125
},
{
"epoch": 0.6362007168458781,
"grad_norm": 0.38491703355424195,
"learning_rate": 7.037032305579409e-05,
"loss": 0.202,
"step": 2130
},
{
"epoch": 0.6376941457586619,
"grad_norm": 0.29165681376515235,
"learning_rate": 6.987279860780148e-05,
"loss": 0.286,
"step": 2135
},
{
"epoch": 0.6391875746714456,
"grad_norm": 0.3691280781000132,
"learning_rate": 6.93760930002765e-05,
"loss": 0.2628,
"step": 2140
},
{
"epoch": 0.6406810035842294,
"grad_norm": 0.33721048070187576,
"learning_rate": 6.888021973339958e-05,
"loss": 0.2629,
"step": 2145
},
{
"epoch": 0.6421744324970131,
"grad_norm": 0.40230756965561154,
"learning_rate": 6.838519228472845e-05,
"loss": 0.2919,
"step": 2150
},
{
"epoch": 0.6436678614097969,
"grad_norm": 0.3832722542456181,
"learning_rate": 6.789102410883201e-05,
"loss": 0.2258,
"step": 2155
},
{
"epoch": 0.6451612903225806,
"grad_norm": 0.3679304149313773,
"learning_rate": 6.739772863692471e-05,
"loss": 0.2684,
"step": 2160
},
{
"epoch": 0.6466547192353644,
"grad_norm": 0.40756568956863426,
"learning_rate": 6.690531927650126e-05,
"loss": 0.2583,
"step": 2165
},
{
"epoch": 0.6481481481481481,
"grad_norm": 0.3980654356736509,
"learning_rate": 6.641380941097237e-05,
"loss": 0.26,
"step": 2170
},
{
"epoch": 0.649641577060932,
"grad_norm": 0.4297937372635352,
"learning_rate": 6.592321239930112e-05,
"loss": 0.2736,
"step": 2175
},
{
"epoch": 0.6511350059737157,
"grad_norm": 0.358991865502727,
"learning_rate": 6.54335415756396e-05,
"loss": 0.2458,
"step": 2180
},
{
"epoch": 0.6526284348864994,
"grad_norm": 0.35194743625956787,
"learning_rate": 6.49448102489666e-05,
"loss": 0.2731,
"step": 2185
},
{
"epoch": 0.6541218637992832,
"grad_norm": 0.3344682852694039,
"learning_rate": 6.445703170272603e-05,
"loss": 0.2392,
"step": 2190
},
{
"epoch": 0.6556152927120669,
"grad_norm": 0.36819816176424924,
"learning_rate": 6.39702191944657e-05,
"loss": 0.2742,
"step": 2195
},
{
"epoch": 0.6571087216248507,
"grad_norm": 0.41481385734175824,
"learning_rate": 6.348438595547705e-05,
"loss": 0.264,
"step": 2200
},
{
"epoch": 0.6571087216248507,
"eval_loss": 0.25236326456069946,
"eval_runtime": 670.9223,
"eval_samples_per_second": 5.962,
"eval_steps_per_second": 0.745,
"step": 2200
},
{
"epoch": 0.6586021505376344,
"grad_norm": 0.37828046273666066,
"learning_rate": 6.299954519043552e-05,
"loss": 0.2045,
"step": 2205
},
{
"epoch": 0.6600955794504182,
"grad_norm": 0.3647111671687454,
"learning_rate": 6.25157100770417e-05,
"loss": 0.2757,
"step": 2210
},
{
"epoch": 0.6615890083632019,
"grad_norm": 0.333924367566417,
"learning_rate": 6.203289376566307e-05,
"loss": 0.2334,
"step": 2215
},
{
"epoch": 0.6630824372759857,
"grad_norm": 0.3731260608605994,
"learning_rate": 6.155110937897667e-05,
"loss": 0.2359,
"step": 2220
},
{
"epoch": 0.6645758661887694,
"grad_norm": 0.36399635762642046,
"learning_rate": 6.107037001161248e-05,
"loss": 0.2509,
"step": 2225
},
{
"epoch": 0.6660692951015532,
"grad_norm": 0.3990653767812247,
"learning_rate": 6.0590688729797295e-05,
"loss": 0.2521,
"step": 2230
},
{
"epoch": 0.6675627240143369,
"grad_norm": 0.3608936095400094,
"learning_rate": 6.011207857099983e-05,
"loss": 0.2543,
"step": 2235
},
{
"epoch": 0.6690561529271206,
"grad_norm": 0.31260618340047636,
"learning_rate": 5.963455254357631e-05,
"loss": 0.2318,
"step": 2240
},
{
"epoch": 0.6705495818399044,
"grad_norm": 0.4642758261006183,
"learning_rate": 5.9158123626416794e-05,
"loss": 0.2754,
"step": 2245
},
{
"epoch": 0.6720430107526881,
"grad_norm": 0.37892791122014974,
"learning_rate": 5.868280476859249e-05,
"loss": 0.2821,
"step": 2250
},
{
"epoch": 0.6735364396654719,
"grad_norm": 0.317919026225424,
"learning_rate": 5.820860888900392e-05,
"loss": 0.2021,
"step": 2255
},
{
"epoch": 0.6750298685782556,
"grad_norm": 0.3638464340699625,
"learning_rate": 5.7735548876029597e-05,
"loss": 0.2758,
"step": 2260
},
{
"epoch": 0.6765232974910395,
"grad_norm": 0.4089378874713291,
"learning_rate": 5.7263637587175836e-05,
"loss": 0.2319,
"step": 2265
},
{
"epoch": 0.6780167264038232,
"grad_norm": 0.36255601718231956,
"learning_rate": 5.679288784872727e-05,
"loss": 0.2736,
"step": 2270
},
{
"epoch": 0.679510155316607,
"grad_norm": 0.4450766858081928,
"learning_rate": 5.632331245539826e-05,
"loss": 0.2682,
"step": 2275
},
{
"epoch": 0.6810035842293907,
"grad_norm": 0.38190771534156615,
"learning_rate": 5.585492416998511e-05,
"loss": 0.2289,
"step": 2280
},
{
"epoch": 0.6824970131421745,
"grad_norm": 0.4405779247739025,
"learning_rate": 5.5387735723019207e-05,
"loss": 0.2928,
"step": 2285
},
{
"epoch": 0.6839904420549582,
"grad_norm": 0.3774472213099317,
"learning_rate": 5.492175981242097e-05,
"loss": 0.2514,
"step": 2290
},
{
"epoch": 0.6854838709677419,
"grad_norm": 0.3377801375518953,
"learning_rate": 5.445700910315481e-05,
"loss": 0.2655,
"step": 2295
},
{
"epoch": 0.6869772998805257,
"grad_norm": 0.39587982015232115,
"learning_rate": 5.399349622688479e-05,
"loss": 0.2946,
"step": 2300
},
{
"epoch": 0.6884707287933094,
"grad_norm": 0.43119763327491156,
"learning_rate": 5.353123378163143e-05,
"loss": 0.2407,
"step": 2305
},
{
"epoch": 0.6899641577060932,
"grad_norm": 0.36132290047057575,
"learning_rate": 5.307023433142919e-05,
"loss": 0.2807,
"step": 2310
},
{
"epoch": 0.6914575866188769,
"grad_norm": 0.38966391503202974,
"learning_rate": 5.2610510405985034e-05,
"loss": 0.2187,
"step": 2315
},
{
"epoch": 0.6929510155316607,
"grad_norm": 0.36997141842452064,
"learning_rate": 5.21520745003379e-05,
"loss": 0.2772,
"step": 2320
},
{
"epoch": 0.6944444444444444,
"grad_norm": 0.4830585130129818,
"learning_rate": 5.169493907451906e-05,
"loss": 0.2846,
"step": 2325
},
{
"epoch": 0.6959378733572282,
"grad_norm": 0.4663739708408958,
"learning_rate": 5.123911655321345e-05,
"loss": 0.2334,
"step": 2330
},
{
"epoch": 0.6974313022700119,
"grad_norm": 0.38317110709020197,
"learning_rate": 5.078461932542204e-05,
"loss": 0.2903,
"step": 2335
},
{
"epoch": 0.6989247311827957,
"grad_norm": 0.44204152745518926,
"learning_rate": 5.033145974412502e-05,
"loss": 0.2418,
"step": 2340
},
{
"epoch": 0.7004181600955794,
"grad_norm": 0.3562540524906317,
"learning_rate": 4.987965012594612e-05,
"loss": 0.2429,
"step": 2345
},
{
"epoch": 0.7019115890083633,
"grad_norm": 0.415882439063641,
"learning_rate": 4.942920275081784e-05,
"loss": 0.2716,
"step": 2350
},
{
"epoch": 0.703405017921147,
"grad_norm": 0.4349967578566546,
"learning_rate": 4.898012986164764e-05,
"loss": 0.227,
"step": 2355
},
{
"epoch": 0.7048984468339307,
"grad_norm": 0.38888281415830517,
"learning_rate": 4.853244366398524e-05,
"loss": 0.2727,
"step": 2360
},
{
"epoch": 0.7063918757467145,
"grad_norm": 0.48748379250803325,
"learning_rate": 4.808615632569089e-05,
"loss": 0.2237,
"step": 2365
},
{
"epoch": 0.7078853046594982,
"grad_norm": 0.33545458271479295,
"learning_rate": 4.764127997660457e-05,
"loss": 0.2512,
"step": 2370
},
{
"epoch": 0.709378733572282,
"grad_norm": 0.4038148334793521,
"learning_rate": 4.719782670821642e-05,
"loss": 0.2729,
"step": 2375
},
{
"epoch": 0.7108721624850657,
"grad_norm": 0.34079737026356866,
"learning_rate": 4.675580857333799e-05,
"loss": 0.2256,
"step": 2380
},
{
"epoch": 0.7123655913978495,
"grad_norm": 0.46179757394318344,
"learning_rate": 4.631523758577475e-05,
"loss": 0.2722,
"step": 2385
},
{
"epoch": 0.7138590203106332,
"grad_norm": 0.39463185516160715,
"learning_rate": 4.5876125719999474e-05,
"loss": 0.2632,
"step": 2390
},
{
"epoch": 0.715352449223417,
"grad_norm": 0.370808339928303,
"learning_rate": 4.543848491082687e-05,
"loss": 0.2381,
"step": 2395
},
{
"epoch": 0.7168458781362007,
"grad_norm": 0.3791124095428876,
"learning_rate": 4.500232705308911e-05,
"loss": 0.2705,
"step": 2400
},
{
"epoch": 0.7168458781362007,
"eval_loss": 0.24903884530067444,
"eval_runtime": 670.4879,
"eval_samples_per_second": 5.966,
"eval_steps_per_second": 0.746,
"step": 2400
},
{
"epoch": 0.7183393070489845,
"grad_norm": 0.3775218725937434,
"learning_rate": 4.45676640013126e-05,
"loss": 0.2268,
"step": 2405
},
{
"epoch": 0.7198327359617682,
"grad_norm": 0.3827463104057589,
"learning_rate": 4.413450756939574e-05,
"loss": 0.2536,
"step": 2410
},
{
"epoch": 0.7213261648745519,
"grad_norm": 0.3648532710176232,
"learning_rate": 4.370286953028787e-05,
"loss": 0.2504,
"step": 2415
},
{
"epoch": 0.7228195937873357,
"grad_norm": 0.3055219924416998,
"learning_rate": 4.32727616156692e-05,
"loss": 0.2324,
"step": 2420
},
{
"epoch": 0.7243130227001194,
"grad_norm": 0.4438604649063271,
"learning_rate": 4.2844195515632166e-05,
"loss": 0.2589,
"step": 2425
},
{
"epoch": 0.7258064516129032,
"grad_norm": 0.3795959005263556,
"learning_rate": 4.2417182878363336e-05,
"loss": 0.2476,
"step": 2430
},
{
"epoch": 0.727299880525687,
"grad_norm": 0.3427145661133711,
"learning_rate": 4.199173530982715e-05,
"loss": 0.2829,
"step": 2435
},
{
"epoch": 0.7287933094384708,
"grad_norm": 0.39445938386306895,
"learning_rate": 4.156786437345044e-05,
"loss": 0.2703,
"step": 2440
},
{
"epoch": 0.7302867383512545,
"grad_norm": 0.39923700769389914,
"learning_rate": 4.114558158980785e-05,
"loss": 0.285,
"step": 2445
},
{
"epoch": 0.7317801672640383,
"grad_norm": 0.4038674564871101,
"learning_rate": 4.072489843630903e-05,
"loss": 0.2814,
"step": 2450
},
{
"epoch": 0.733273596176822,
"grad_norm": 0.36737797848087894,
"learning_rate": 4.030582634688669e-05,
"loss": 0.2397,
"step": 2455
},
{
"epoch": 0.7347670250896058,
"grad_norm": 0.33021542644527924,
"learning_rate": 3.98883767116855e-05,
"loss": 0.2415,
"step": 2460
},
{
"epoch": 0.7362604540023895,
"grad_norm": 0.4047149773056364,
"learning_rate": 3.947256087675286e-05,
"loss": 0.2636,
"step": 2465
},
{
"epoch": 0.7377538829151732,
"grad_norm": 0.418965527019863,
"learning_rate": 3.905839014373047e-05,
"loss": 0.2818,
"step": 2470
},
{
"epoch": 0.739247311827957,
"grad_norm": 0.44764561078371773,
"learning_rate": 3.864587576954688e-05,
"loss": 0.2781,
"step": 2475
},
{
"epoch": 0.7407407407407407,
"grad_norm": 0.39101875860430557,
"learning_rate": 3.823502896611184e-05,
"loss": 0.2577,
"step": 2480
},
{
"epoch": 0.7422341696535245,
"grad_norm": 0.3922118709138838,
"learning_rate": 3.782586090001155e-05,
"loss": 0.2563,
"step": 2485
},
{
"epoch": 0.7437275985663082,
"grad_norm": 0.379882936966773,
"learning_rate": 3.741838269220496e-05,
"loss": 0.2517,
"step": 2490
},
{
"epoch": 0.745221027479092,
"grad_norm": 0.37880769907144246,
"learning_rate": 3.7012605417721534e-05,
"loss": 0.2908,
"step": 2495
},
{
"epoch": 0.7467144563918757,
"grad_norm": 0.4030754847132898,
"learning_rate": 3.660854010536055e-05,
"loss": 0.2634,
"step": 2500
},
{
"epoch": 0.7482078853046595,
"grad_norm": 0.32788742027641565,
"learning_rate": 3.620619773739097e-05,
"loss": 0.2183,
"step": 2505
},
{
"epoch": 0.7497013142174432,
"grad_norm": 0.3263923214186051,
"learning_rate": 3.580558924925305e-05,
"loss": 0.2751,
"step": 2510
},
{
"epoch": 0.751194743130227,
"grad_norm": 0.450636347100308,
"learning_rate": 3.540672552926122e-05,
"loss": 0.2625,
"step": 2515
},
{
"epoch": 0.7526881720430108,
"grad_norm": 0.33757741246501627,
"learning_rate": 3.500961741830821e-05,
"loss": 0.2575,
"step": 2520
},
{
"epoch": 0.7541816009557945,
"grad_norm": 0.4249647628134897,
"learning_rate": 3.4614275709570033e-05,
"loss": 0.2619,
"step": 2525
},
{
"epoch": 0.7556750298685783,
"grad_norm": 0.37310284633404023,
"learning_rate": 3.422071114821304e-05,
"loss": 0.2466,
"step": 2530
},
{
"epoch": 0.757168458781362,
"grad_norm": 0.3827899754640218,
"learning_rate": 3.382893443110175e-05,
"loss": 0.2808,
"step": 2535
},
{
"epoch": 0.7586618876941458,
"grad_norm": 0.34434815179683986,
"learning_rate": 3.343895620650793e-05,
"loss": 0.2369,
"step": 2540
},
{
"epoch": 0.7601553166069295,
"grad_norm": 0.38063545879124566,
"learning_rate": 3.305078707382138e-05,
"loss": 0.277,
"step": 2545
},
{
"epoch": 0.7616487455197133,
"grad_norm": 0.39371915129642,
"learning_rate": 3.26644375832619e-05,
"loss": 0.2721,
"step": 2550
},
{
"epoch": 0.763142174432497,
"grad_norm": 0.4068353272903545,
"learning_rate": 3.2279918235592346e-05,
"loss": 0.23,
"step": 2555
},
{
"epoch": 0.7646356033452808,
"grad_norm": 0.40612317566802253,
"learning_rate": 3.189723948183322e-05,
"loss": 0.2665,
"step": 2560
},
{
"epoch": 0.7661290322580645,
"grad_norm": 0.36414950817333475,
"learning_rate": 3.151641172297891e-05,
"loss": 0.2534,
"step": 2565
},
{
"epoch": 0.7676224611708483,
"grad_norm": 0.36001677147408806,
"learning_rate": 3.1137445309714705e-05,
"loss": 0.2537,
"step": 2570
},
{
"epoch": 0.769115890083632,
"grad_norm": 0.4064294097626817,
"learning_rate": 3.076035054213548e-05,
"loss": 0.2628,
"step": 2575
},
{
"epoch": 0.7706093189964157,
"grad_norm": 0.38634424971456655,
"learning_rate": 3.0385137669465967e-05,
"loss": 0.2475,
"step": 2580
},
{
"epoch": 0.7721027479091995,
"grad_norm": 0.3868359752050659,
"learning_rate": 3.001181688978203e-05,
"loss": 0.257,
"step": 2585
},
{
"epoch": 0.7735961768219832,
"grad_norm": 0.31847898193899454,
"learning_rate": 2.9640398349733334e-05,
"loss": 0.243,
"step": 2590
},
{
"epoch": 0.775089605734767,
"grad_norm": 0.4155060821736252,
"learning_rate": 2.9270892144267993e-05,
"loss": 0.2739,
"step": 2595
},
{
"epoch": 0.7765830346475507,
"grad_norm": 0.36416964732293394,
"learning_rate": 2.890330831635778e-05,
"loss": 0.2694,
"step": 2600
},
{
"epoch": 0.7765830346475507,
"eval_loss": 0.24662606418132782,
"eval_runtime": 671.3061,
"eval_samples_per_second": 5.959,
"eval_steps_per_second": 0.745,
"step": 2600
},
{
"epoch": 0.7780764635603346,
"grad_norm": 0.34102575217523634,
"learning_rate": 2.853765685672528e-05,
"loss": 0.2345,
"step": 2605
},
{
"epoch": 0.7795698924731183,
"grad_norm": 0.3812471975597926,
"learning_rate": 2.8173947703572546e-05,
"loss": 0.2619,
"step": 2610
},
{
"epoch": 0.7810633213859021,
"grad_norm": 0.4709017717954452,
"learning_rate": 2.7812190742310707e-05,
"loss": 0.2359,
"step": 2615
},
{
"epoch": 0.7825567502986858,
"grad_norm": 0.3815586898993409,
"learning_rate": 2.7452395805291463e-05,
"loss": 0.2485,
"step": 2620
},
{
"epoch": 0.7840501792114696,
"grad_norm": 0.44121044789192165,
"learning_rate": 2.7094572671539765e-05,
"loss": 0.2615,
"step": 2625
},
{
"epoch": 0.7855436081242533,
"grad_norm": 0.35382301353141093,
"learning_rate": 2.6738731066488075e-05,
"loss": 0.2154,
"step": 2630
},
{
"epoch": 0.7870370370370371,
"grad_norm": 0.33814033552150285,
"learning_rate": 2.638488066171201e-05,
"loss": 0.2569,
"step": 2635
},
{
"epoch": 0.7885304659498208,
"grad_norm": 0.37849934442478683,
"learning_rate": 2.603303107466747e-05,
"loss": 0.2616,
"step": 2640
},
{
"epoch": 0.7900238948626045,
"grad_norm": 0.3904880202778289,
"learning_rate": 2.5683191868429247e-05,
"loss": 0.2587,
"step": 2645
},
{
"epoch": 0.7915173237753883,
"grad_norm": 0.3646861718649452,
"learning_rate": 2.5335372551431125e-05,
"loss": 0.2766,
"step": 2650
},
{
"epoch": 0.793010752688172,
"grad_norm": 0.38220642098483965,
"learning_rate": 2.4989582577207417e-05,
"loss": 0.2506,
"step": 2655
},
{
"epoch": 0.7945041816009558,
"grad_norm": 0.34734076514892087,
"learning_rate": 2.4645831344136037e-05,
"loss": 0.2676,
"step": 2660
},
{
"epoch": 0.7959976105137395,
"grad_norm": 0.33364705048661314,
"learning_rate": 2.4304128195183086e-05,
"loss": 0.2144,
"step": 2665
},
{
"epoch": 0.7974910394265233,
"grad_norm": 0.41830275352951174,
"learning_rate": 2.396448241764877e-05,
"loss": 0.281,
"step": 2670
},
{
"epoch": 0.798984468339307,
"grad_norm": 0.4095216224292848,
"learning_rate": 2.3626903242915267e-05,
"loss": 0.2855,
"step": 2675
},
{
"epoch": 0.8004778972520908,
"grad_norm": 0.3532672344479507,
"learning_rate": 2.329139984619553e-05,
"loss": 0.2391,
"step": 2680
},
{
"epoch": 0.8019713261648745,
"grad_norm": 0.3758073226768294,
"learning_rate": 2.295798134628404e-05,
"loss": 0.2929,
"step": 2685
},
{
"epoch": 0.8034647550776584,
"grad_norm": 0.35018181270060134,
"learning_rate": 2.2626656805309e-05,
"loss": 0.2357,
"step": 2690
},
{
"epoch": 0.804958183990442,
"grad_norm": 0.4220541464364436,
"learning_rate": 2.2297435228485918e-05,
"loss": 0.246,
"step": 2695
},
{
"epoch": 0.8064516129032258,
"grad_norm": 0.3870739284799057,
"learning_rate": 2.197032556387295e-05,
"loss": 0.2766,
"step": 2700
},
{
"epoch": 0.8079450418160096,
"grad_norm": 0.41053303775154976,
"learning_rate": 2.164533670212766e-05,
"loss": 0.2407,
"step": 2705
},
{
"epoch": 0.8094384707287933,
"grad_norm": 0.3731772239188146,
"learning_rate": 2.1322477476265367e-05,
"loss": 0.2864,
"step": 2710
},
{
"epoch": 0.8109318996415771,
"grad_norm": 0.37634222348253304,
"learning_rate": 2.1001756661419093e-05,
"loss": 0.2424,
"step": 2715
},
{
"epoch": 0.8124253285543608,
"grad_norm": 0.3644559156551827,
"learning_rate": 2.068318297460102e-05,
"loss": 0.2839,
"step": 2720
},
{
"epoch": 0.8139187574671446,
"grad_norm": 0.39313980741067645,
"learning_rate": 2.036676507446562e-05,
"loss": 0.2642,
"step": 2725
},
{
"epoch": 0.8154121863799283,
"grad_norm": 0.3850789535231269,
"learning_rate": 2.005251156107426e-05,
"loss": 0.2104,
"step": 2730
},
{
"epoch": 0.8169056152927121,
"grad_norm": 0.39312531551082797,
"learning_rate": 1.9740430975661528e-05,
"loss": 0.264,
"step": 2735
},
{
"epoch": 0.8183990442054958,
"grad_norm": 0.36295621136819534,
"learning_rate": 1.943053180040302e-05,
"loss": 0.2307,
"step": 2740
},
{
"epoch": 0.8198924731182796,
"grad_norm": 0.3725954567916029,
"learning_rate": 1.9122822458184807e-05,
"loss": 0.2479,
"step": 2745
},
{
"epoch": 0.8213859020310633,
"grad_norm": 0.3713829913873105,
"learning_rate": 1.8817311312374564e-05,
"loss": 0.2575,
"step": 2750
},
{
"epoch": 0.822879330943847,
"grad_norm": 0.3202055385441327,
"learning_rate": 1.8514006666594197e-05,
"loss": 0.2308,
"step": 2755
},
{
"epoch": 0.8243727598566308,
"grad_norm": 0.3833958130308794,
"learning_rate": 1.821291676449417e-05,
"loss": 0.2713,
"step": 2760
},
{
"epoch": 0.8258661887694145,
"grad_norm": 0.37070939699033145,
"learning_rate": 1.7914049789529465e-05,
"loss": 0.2253,
"step": 2765
},
{
"epoch": 0.8273596176821983,
"grad_norm": 0.3091049252789486,
"learning_rate": 1.7617413864737153e-05,
"loss": 0.2491,
"step": 2770
},
{
"epoch": 0.828853046594982,
"grad_norm": 0.3863395305062906,
"learning_rate": 1.73230170525156e-05,
"loss": 0.3097,
"step": 2775
},
{
"epoch": 0.8303464755077659,
"grad_norm": 0.3398351066167249,
"learning_rate": 1.7030867354405354e-05,
"loss": 0.2175,
"step": 2780
},
{
"epoch": 0.8318399044205496,
"grad_norm": 0.353170388342167,
"learning_rate": 1.674097271087165e-05,
"loss": 0.2526,
"step": 2785
},
{
"epoch": 0.8333333333333334,
"grad_norm": 0.37416682070994395,
"learning_rate": 1.6453341001088572e-05,
"loss": 0.2155,
"step": 2790
},
{
"epoch": 0.8348267622461171,
"grad_norm": 0.38759381091463324,
"learning_rate": 1.616798004272506e-05,
"loss": 0.2516,
"step": 2795
},
{
"epoch": 0.8363201911589009,
"grad_norm": 0.3774042044993102,
"learning_rate": 1.5884897591732127e-05,
"loss": 0.2639,
"step": 2800
},
{
"epoch": 0.8363201911589009,
"eval_loss": 0.24498099088668823,
"eval_runtime": 670.7113,
"eval_samples_per_second": 5.964,
"eval_steps_per_second": 0.745,
"step": 2800
},
{
"epoch": 0.8378136200716846,
"grad_norm": 0.47282083906303013,
"learning_rate": 1.5604101342132306e-05,
"loss": 0.1951,
"step": 2805
},
{
"epoch": 0.8393070489844683,
"grad_norm": 0.3464605159651398,
"learning_rate": 1.5325598925810548e-05,
"loss": 0.265,
"step": 2810
},
{
"epoch": 0.8408004778972521,
"grad_norm": 0.4146312634378869,
"learning_rate": 1.5049397912306518e-05,
"loss": 0.2574,
"step": 2815
},
{
"epoch": 0.8422939068100358,
"grad_norm": 0.40351451569319263,
"learning_rate": 1.4775505808609191e-05,
"loss": 0.2733,
"step": 2820
},
{
"epoch": 0.8437873357228196,
"grad_norm": 0.3716555969690491,
"learning_rate": 1.4503930058952586e-05,
"loss": 0.277,
"step": 2825
},
{
"epoch": 0.8452807646356033,
"grad_norm": 0.34726179203272095,
"learning_rate": 1.423467804461357e-05,
"loss": 0.2247,
"step": 2830
},
{
"epoch": 0.8467741935483871,
"grad_norm": 0.3357554714579106,
"learning_rate": 1.396775708371113e-05,
"loss": 0.2656,
"step": 2835
},
{
"epoch": 0.8482676224611708,
"grad_norm": 0.3575152949170555,
"learning_rate": 1.3703174431007559e-05,
"loss": 0.2352,
"step": 2840
},
{
"epoch": 0.8497610513739546,
"grad_norm": 0.3501080536067048,
"learning_rate": 1.344093727771124e-05,
"loss": 0.2406,
"step": 2845
},
{
"epoch": 0.8512544802867383,
"grad_norm": 0.5043952724631786,
"learning_rate": 1.3181052751281197e-05,
"loss": 0.2768,
"step": 2850
},
{
"epoch": 0.8527479091995221,
"grad_norm": 0.3575895422258797,
"learning_rate": 1.2923527915233336e-05,
"loss": 0.2125,
"step": 2855
},
{
"epoch": 0.8542413381123058,
"grad_norm": 0.3407245335177286,
"learning_rate": 1.2668369768948608e-05,
"loss": 0.2661,
"step": 2860
},
{
"epoch": 0.8557347670250897,
"grad_norm": 0.4458153062323654,
"learning_rate": 1.2415585247482498e-05,
"loss": 0.2226,
"step": 2865
},
{
"epoch": 0.8572281959378734,
"grad_norm": 0.4369000451331262,
"learning_rate": 1.2165181221376787e-05,
"loss": 0.272,
"step": 2870
},
{
"epoch": 0.8587216248506571,
"grad_norm": 0.48118794169539114,
"learning_rate": 1.1917164496472799e-05,
"loss": 0.2616,
"step": 2875
},
{
"epoch": 0.8602150537634409,
"grad_norm": 0.3744622943326254,
"learning_rate": 1.1671541813726194e-05,
"loss": 0.264,
"step": 2880
},
{
"epoch": 0.8617084826762246,
"grad_norm": 0.4536685590862434,
"learning_rate": 1.1428319849023984e-05,
"loss": 0.2557,
"step": 2885
},
{
"epoch": 0.8632019115890084,
"grad_norm": 0.34146777737976525,
"learning_rate": 1.1187505213003068e-05,
"loss": 0.2228,
"step": 2890
},
{
"epoch": 0.8646953405017921,
"grad_norm": 0.4000563885902401,
"learning_rate": 1.0949104450870384e-05,
"loss": 0.2672,
"step": 2895
},
{
"epoch": 0.8661887694145759,
"grad_norm": 0.380767424516121,
"learning_rate": 1.0713124042225165e-05,
"loss": 0.2704,
"step": 2900
},
{
"epoch": 0.8676821983273596,
"grad_norm": 0.3742084189949366,
"learning_rate": 1.0479570400882888e-05,
"loss": 0.2166,
"step": 2905
},
{
"epoch": 0.8691756272401434,
"grad_norm": 0.4568051868923601,
"learning_rate": 1.0248449874700705e-05,
"loss": 0.2639,
"step": 2910
},
{
"epoch": 0.8706690561529271,
"grad_norm": 0.3959326669620433,
"learning_rate": 1.0019768745405122e-05,
"loss": 0.224,
"step": 2915
},
{
"epoch": 0.8721624850657109,
"grad_norm": 0.3843360418655597,
"learning_rate": 9.79353322842127e-06,
"loss": 0.2596,
"step": 2920
},
{
"epoch": 0.8736559139784946,
"grad_norm": 0.3404605327204112,
"learning_rate": 9.569749472703816e-06,
"loss": 0.2532,
"step": 2925
},
{
"epoch": 0.8751493428912783,
"grad_norm": 0.4023542619848836,
"learning_rate": 9.34842356056993e-06,
"loss": 0.2461,
"step": 2930
},
{
"epoch": 0.8766427718040621,
"grad_norm": 0.34874999168812615,
"learning_rate": 9.129561507534046e-06,
"loss": 0.2625,
"step": 2935
},
{
"epoch": 0.8781362007168458,
"grad_norm": 0.363113367835939,
"learning_rate": 8.913169262144239e-06,
"loss": 0.2298,
"step": 2940
},
{
"epoch": 0.8796296296296297,
"grad_norm": 0.4366858825613128,
"learning_rate": 8.699252705820526e-06,
"loss": 0.266,
"step": 2945
},
{
"epoch": 0.8811230585424134,
"grad_norm": 0.3631698452397834,
"learning_rate": 8.487817652695229e-06,
"loss": 0.2884,
"step": 2950
},
{
"epoch": 0.8826164874551972,
"grad_norm": 0.3576538561239022,
"learning_rate": 8.278869849454718e-06,
"loss": 0.242,
"step": 2955
},
{
"epoch": 0.8841099163679809,
"grad_norm": 0.4230742530870855,
"learning_rate": 8.07241497518324e-06,
"loss": 0.2733,
"step": 2960
},
{
"epoch": 0.8856033452807647,
"grad_norm": 0.4193451166331605,
"learning_rate": 7.868458641208765e-06,
"loss": 0.2314,
"step": 2965
},
{
"epoch": 0.8870967741935484,
"grad_norm": 0.3701152603399041,
"learning_rate": 7.667006390950248e-06,
"loss": 0.2767,
"step": 2970
},
{
"epoch": 0.8885902031063322,
"grad_norm": 0.47090020438598185,
"learning_rate": 7.468063699767081e-06,
"loss": 0.2731,
"step": 2975
},
{
"epoch": 0.8900836320191159,
"grad_norm": 0.4420262562340979,
"learning_rate": 7.27163597481022e-06,
"loss": 0.2428,
"step": 2980
},
{
"epoch": 0.8915770609318996,
"grad_norm": 0.4555285504311469,
"learning_rate": 7.077728554875263e-06,
"loss": 0.2637,
"step": 2985
},
{
"epoch": 0.8930704898446834,
"grad_norm": 0.32544659259467257,
"learning_rate": 6.886346710257363e-06,
"loss": 0.2332,
"step": 2990
},
{
"epoch": 0.8945639187574671,
"grad_norm": 0.39917120432545034,
"learning_rate": 6.697495642607854e-06,
"loss": 0.2774,
"step": 2995
},
{
"epoch": 0.8960573476702509,
"grad_norm": 0.4322326823522948,
"learning_rate": 6.511180484793067e-06,
"loss": 0.2598,
"step": 3000
},
{
"epoch": 0.8960573476702509,
"eval_loss": 0.24354791641235352,
"eval_runtime": 670.7371,
"eval_samples_per_second": 5.964,
"eval_steps_per_second": 0.745,
"step": 3000
},
{
"epoch": 0.8975507765830346,
"grad_norm": 0.41924813736657407,
"learning_rate": 6.3274063007547125e-06,
"loss": 0.249,
"step": 3005
},
{
"epoch": 0.8990442054958184,
"grad_norm": 0.31419280668927857,
"learning_rate": 6.146178085372156e-06,
"loss": 0.2753,
"step": 3010
},
{
"epoch": 0.9005376344086021,
"grad_norm": 0.36068884644950366,
"learning_rate": 5.9675007643269054e-06,
"loss": 0.2412,
"step": 3015
},
{
"epoch": 0.9020310633213859,
"grad_norm": 0.3391790680684511,
"learning_rate": 5.791379193968505e-06,
"loss": 0.2583,
"step": 3020
},
{
"epoch": 0.9035244922341696,
"grad_norm": 0.362307062919707,
"learning_rate": 5.617818161182586e-06,
"loss": 0.2867,
"step": 3025
},
{
"epoch": 0.9050179211469535,
"grad_norm": 0.38255580305837367,
"learning_rate": 5.446822383260908e-06,
"loss": 0.2392,
"step": 3030
},
{
"epoch": 0.9065113500597372,
"grad_norm": 0.41055570601524083,
"learning_rate": 5.2783965077729755e-06,
"loss": 0.2414,
"step": 3035
},
{
"epoch": 0.9080047789725209,
"grad_norm": 0.4885743038657098,
"learning_rate": 5.112545112439782e-06,
"loss": 0.2534,
"step": 3040
},
{
"epoch": 0.9094982078853047,
"grad_norm": 0.42278986461993495,
"learning_rate": 4.949272705009411e-06,
"loss": 0.2591,
"step": 3045
},
{
"epoch": 0.9109916367980884,
"grad_norm": 0.4025383892648368,
"learning_rate": 4.788583723134498e-06,
"loss": 0.2887,
"step": 3050
},
{
"epoch": 0.9124850657108722,
"grad_norm": 0.42940556785118494,
"learning_rate": 4.630482534251601e-06,
"loss": 0.2239,
"step": 3055
},
{
"epoch": 0.9139784946236559,
"grad_norm": 0.4252373447840496,
"learning_rate": 4.474973435462526e-06,
"loss": 0.2648,
"step": 3060
},
{
"epoch": 0.9154719235364397,
"grad_norm": 0.34717731405681657,
"learning_rate": 4.322060653417525e-06,
"loss": 0.2359,
"step": 3065
},
{
"epoch": 0.9169653524492234,
"grad_norm": 0.3932159130540263,
"learning_rate": 4.171748344200399e-06,
"loss": 0.2844,
"step": 3070
},
{
"epoch": 0.9184587813620072,
"grad_norm": 0.3818285602145142,
"learning_rate": 4.0240405932155725e-06,
"loss": 0.2601,
"step": 3075
},
{
"epoch": 0.9199522102747909,
"grad_norm": 0.3967329457656309,
"learning_rate": 3.87894141507702e-06,
"loss": 0.2485,
"step": 3080
},
{
"epoch": 0.9214456391875747,
"grad_norm": 0.3646492393972917,
"learning_rate": 3.7364547534991745e-06,
"loss": 0.2921,
"step": 3085
},
{
"epoch": 0.9229390681003584,
"grad_norm": 0.31319503804493787,
"learning_rate": 3.5965844811897243e-06,
"loss": 0.2294,
"step": 3090
},
{
"epoch": 0.9244324970131422,
"grad_norm": 0.36230176443854195,
"learning_rate": 3.459334399744374e-06,
"loss": 0.2616,
"step": 3095
},
{
"epoch": 0.9259259259259259,
"grad_norm": 0.46214491992499196,
"learning_rate": 3.3247082395434835e-06,
"loss": 0.2821,
"step": 3100
},
{
"epoch": 0.9274193548387096,
"grad_norm": 0.32464884452568865,
"learning_rate": 3.19270965965075e-06,
"loss": 0.2249,
"step": 3105
},
{
"epoch": 0.9289127837514934,
"grad_norm": 0.38961006991574,
"learning_rate": 3.063342247713652e-06,
"loss": 0.2814,
"step": 3110
},
{
"epoch": 0.9304062126642771,
"grad_norm": 0.3430373709286391,
"learning_rate": 2.9366095198660292e-06,
"loss": 0.2424,
"step": 3115
},
{
"epoch": 0.931899641577061,
"grad_norm": 0.3754679213082568,
"learning_rate": 2.812514920632481e-06,
"loss": 0.2458,
"step": 3120
},
{
"epoch": 0.9333930704898447,
"grad_norm": 0.32857648616715374,
"learning_rate": 2.6910618228347397e-06,
"loss": 0.2648,
"step": 3125
},
{
"epoch": 0.9348864994026285,
"grad_norm": 0.38704797315639966,
"learning_rate": 2.5722535275000014e-06,
"loss": 0.2576,
"step": 3130
},
{
"epoch": 0.9363799283154122,
"grad_norm": 0.347816629229733,
"learning_rate": 2.456093263771242e-06,
"loss": 0.2579,
"step": 3135
},
{
"epoch": 0.937873357228196,
"grad_norm": 0.3285229980093725,
"learning_rate": 2.3425841888193744e-06,
"loss": 0.226,
"step": 3140
},
{
"epoch": 0.9393667861409797,
"grad_norm": 0.35095116254795605,
"learning_rate": 2.2317293877575195e-06,
"loss": 0.2715,
"step": 3145
},
{
"epoch": 0.9408602150537635,
"grad_norm": 0.43133915586490196,
"learning_rate": 2.1235318735571164e-06,
"loss": 0.2743,
"step": 3150
},
{
"epoch": 0.9423536439665472,
"grad_norm": 0.39808144981035765,
"learning_rate": 2.0179945869660098e-06,
"loss": 0.2133,
"step": 3155
},
{
"epoch": 0.9438470728793309,
"grad_norm": 0.37858205437619763,
"learning_rate": 1.9151203964285936e-06,
"loss": 0.2744,
"step": 3160
},
{
"epoch": 0.9453405017921147,
"grad_norm": 0.33007798696691204,
"learning_rate": 1.8149120980077595e-06,
"loss": 0.2364,
"step": 3165
},
{
"epoch": 0.9468339307048984,
"grad_norm": 0.40641752996751496,
"learning_rate": 1.7173724153089598e-06,
"loss": 0.2546,
"step": 3170
},
{
"epoch": 0.9483273596176822,
"grad_norm": 0.41533410990816155,
"learning_rate": 1.6225039994061552e-06,
"loss": 0.2551,
"step": 3175
},
{
"epoch": 0.9498207885304659,
"grad_norm": 0.31219456691145986,
"learning_rate": 1.5303094287697938e-06,
"loss": 0.2152,
"step": 3180
},
{
"epoch": 0.9513142174432497,
"grad_norm": 0.394461459007535,
"learning_rate": 1.4407912091966902e-06,
"loss": 0.2989,
"step": 3185
},
{
"epoch": 0.9528076463560334,
"grad_norm": 0.3577386265843814,
"learning_rate": 1.353951773741935e-06,
"loss": 0.2229,
"step": 3190
},
{
"epoch": 0.9543010752688172,
"grad_norm": 0.3868747591156057,
"learning_rate": 1.2697934826527701e-06,
"loss": 0.249,
"step": 3195
},
{
"epoch": 0.955794504181601,
"grad_norm": 0.3780574063379163,
"learning_rate": 1.1883186233044186e-06,
"loss": 0.2483,
"step": 3200
},
{
"epoch": 0.955794504181601,
"eval_loss": 0.2432253062725067,
"eval_runtime": 671.4179,
"eval_samples_per_second": 5.958,
"eval_steps_per_second": 0.745,
"step": 3200
},
{
"epoch": 0.9572879330943848,
"grad_norm": 0.3892026722877345,
"learning_rate": 1.109529410137966e-06,
"loss": 0.2305,
"step": 3205
},
{
"epoch": 0.9587813620071685,
"grad_norm": 0.3931487469870653,
"learning_rate": 1.0334279846001106e-06,
"loss": 0.2544,
"step": 3210
},
{
"epoch": 0.9602747909199522,
"grad_norm": 0.36069640184936463,
"learning_rate": 9.600164150849854e-07,
"loss": 0.2575,
"step": 3215
},
{
"epoch": 0.961768219832736,
"grad_norm": 0.44849903301516003,
"learning_rate": 8.892966968779615e-07,
"loss": 0.2503,
"step": 3220
},
{
"epoch": 0.9632616487455197,
"grad_norm": 0.411995160828815,
"learning_rate": 8.212707521013774e-07,
"loss": 0.2612,
"step": 3225
},
{
"epoch": 0.9647550776583035,
"grad_norm": 0.5604597904188017,
"learning_rate": 7.559404296623495e-07,
"loss": 0.2106,
"step": 3230
},
{
"epoch": 0.9662485065710872,
"grad_norm": 0.3789859551503982,
"learning_rate": 6.933075052024562e-07,
"loss": 0.2597,
"step": 3235
},
{
"epoch": 0.967741935483871,
"grad_norm": 0.3029358775615665,
"learning_rate": 6.333736810495095e-07,
"loss": 0.22,
"step": 3240
},
{
"epoch": 0.9692353643966547,
"grad_norm": 0.43863865163510735,
"learning_rate": 5.761405861713142e-07,
"loss": 0.2888,
"step": 3245
},
{
"epoch": 0.9707287933094385,
"grad_norm": 0.3410374697833037,
"learning_rate": 5.216097761313377e-07,
"loss": 0.2615,
"step": 3250
},
{
"epoch": 0.9722222222222222,
"grad_norm": 0.3972032404026882,
"learning_rate": 4.6978273304646434e-07,
"loss": 0.2084,
"step": 3255
},
{
"epoch": 0.973715651135006,
"grad_norm": 0.43978255054644894,
"learning_rate": 4.206608655467403e-07,
"loss": 0.2626,
"step": 3260
},
{
"epoch": 0.9752090800477897,
"grad_norm": 0.4054251372400272,
"learning_rate": 3.7424550873699227e-07,
"loss": 0.2104,
"step": 3265
},
{
"epoch": 0.9767025089605734,
"grad_norm": 0.3250043163561357,
"learning_rate": 3.305379241606343e-07,
"loss": 0.2575,
"step": 3270
},
{
"epoch": 0.9781959378733572,
"grad_norm": 0.34288193953849005,
"learning_rate": 2.8953929976536233e-07,
"loss": 0.2521,
"step": 3275
},
{
"epoch": 0.9796893667861409,
"grad_norm": 0.3438606265427078,
"learning_rate": 2.51250749870835e-07,
"loss": 0.2223,
"step": 3280
},
{
"epoch": 0.9811827956989247,
"grad_norm": 0.45991832962727547,
"learning_rate": 2.156733151383872e-07,
"loss": 0.2615,
"step": 3285
},
{
"epoch": 0.9826762246117084,
"grad_norm": 0.34236518853477144,
"learning_rate": 1.8280796254279698e-07,
"loss": 0.2332,
"step": 3290
},
{
"epoch": 0.9841696535244923,
"grad_norm": 0.3358771867133346,
"learning_rate": 1.5265558534592883e-07,
"loss": 0.2436,
"step": 3295
},
{
"epoch": 0.985663082437276,
"grad_norm": 0.393065156149057,
"learning_rate": 1.252170030725308e-07,
"loss": 0.2707,
"step": 3300
},
{
"epoch": 0.9871565113500598,
"grad_norm": 0.3901577896476038,
"learning_rate": 1.0049296148790799e-07,
"loss": 0.2241,
"step": 3305
},
{
"epoch": 0.9886499402628435,
"grad_norm": 0.39775264512368114,
"learning_rate": 7.848413257766085e-08,
"loss": 0.2768,
"step": 3310
},
{
"epoch": 0.9901433691756273,
"grad_norm": 0.4277804217503546,
"learning_rate": 5.9191114529433266e-08,
"loss": 0.2351,
"step": 3315
},
{
"epoch": 0.991636798088411,
"grad_norm": 0.326337320725509,
"learning_rate": 4.26144317166699e-08,
"loss": 0.2526,
"step": 3320
},
{
"epoch": 0.9931302270011948,
"grad_norm": 0.4455257895279916,
"learning_rate": 2.8754534684316547e-08,
"loss": 0.267,
"step": 3325
},
{
"epoch": 0.9946236559139785,
"grad_norm": 0.3562424876586232,
"learning_rate": 1.7611800136618783e-08,
"loss": 0.2358,
"step": 3330
},
{
"epoch": 0.9961170848267622,
"grad_norm": 0.36120057011336665,
"learning_rate": 9.186530926874604e-09,
"loss": 0.2675,
"step": 3335
},
{
"epoch": 0.997610513739546,
"grad_norm": 0.3421927721985385,
"learning_rate": 3.4789560491743643e-09,
"loss": 0.2207,
"step": 3340
},
{
"epoch": 0.9991039426523297,
"grad_norm": 0.3603187953155133,
"learning_rate": 4.892306322057394e-10,
"loss": 0.2668,
"step": 3345
},
{
"epoch": 1.0,
"step": 3348,
"total_flos": 1.6414748941746176e+16,
"train_loss": 0.27713136451503567,
"train_runtime": 31532.8493,
"train_samples_per_second": 1.699,
"train_steps_per_second": 0.106
}
],
"logging_steps": 5,
"max_steps": 3348,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 200,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.6414748941746176e+16,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}