cfb-drives-lora / checkpoint-632 /trainer_state.json
awilliamson's picture
Upload folder using huggingface_hub
7f9b55d verified
raw
history blame
111 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.6000474721101353,
"eval_steps": 500,
"global_step": 632,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0009494422027059103,
"grad_norm": 5.364119052886963,
"learning_rate": 6.329113924050633e-07,
"loss": 2.9616,
"step": 1
},
{
"epoch": 0.0018988844054118206,
"grad_norm": 5.426063060760498,
"learning_rate": 1.2658227848101265e-06,
"loss": 2.9664,
"step": 2
},
{
"epoch": 0.0028483266081177306,
"grad_norm": 5.507386684417725,
"learning_rate": 1.8987341772151901e-06,
"loss": 2.96,
"step": 3
},
{
"epoch": 0.0037977688108236413,
"grad_norm": 5.467552185058594,
"learning_rate": 2.531645569620253e-06,
"loss": 2.975,
"step": 4
},
{
"epoch": 0.004747211013529551,
"grad_norm": 5.386384963989258,
"learning_rate": 3.1645569620253167e-06,
"loss": 2.9473,
"step": 5
},
{
"epoch": 0.005696653216235461,
"grad_norm": 4.9166951179504395,
"learning_rate": 3.7974683544303802e-06,
"loss": 2.8597,
"step": 6
},
{
"epoch": 0.006646095418941372,
"grad_norm": 5.469020843505859,
"learning_rate": 4.430379746835443e-06,
"loss": 2.9633,
"step": 7
},
{
"epoch": 0.0075955376216472826,
"grad_norm": 5.380453586578369,
"learning_rate": 5.063291139240506e-06,
"loss": 2.8931,
"step": 8
},
{
"epoch": 0.008544979824353193,
"grad_norm": 4.922253131866455,
"learning_rate": 5.69620253164557e-06,
"loss": 2.7452,
"step": 9
},
{
"epoch": 0.009494422027059102,
"grad_norm": 5.517508029937744,
"learning_rate": 6.329113924050633e-06,
"loss": 2.8326,
"step": 10
},
{
"epoch": 0.010443864229765013,
"grad_norm": 5.810976982116699,
"learning_rate": 6.9620253164556965e-06,
"loss": 2.7854,
"step": 11
},
{
"epoch": 0.011393306432470923,
"grad_norm": 5.690661430358887,
"learning_rate": 7.5949367088607605e-06,
"loss": 2.7069,
"step": 12
},
{
"epoch": 0.012342748635176834,
"grad_norm": 5.994122505187988,
"learning_rate": 8.227848101265822e-06,
"loss": 2.5705,
"step": 13
},
{
"epoch": 0.013292190837882745,
"grad_norm": 5.86803674697876,
"learning_rate": 8.860759493670886e-06,
"loss": 2.4461,
"step": 14
},
{
"epoch": 0.014241633040588654,
"grad_norm": 5.448781490325928,
"learning_rate": 9.49367088607595e-06,
"loss": 2.2408,
"step": 15
},
{
"epoch": 0.015191075243294565,
"grad_norm": 6.4004902839660645,
"learning_rate": 1.0126582278481012e-05,
"loss": 2.1205,
"step": 16
},
{
"epoch": 0.016140517446000476,
"grad_norm": 6.970590591430664,
"learning_rate": 1.0759493670886076e-05,
"loss": 1.9474,
"step": 17
},
{
"epoch": 0.017089959648706386,
"grad_norm": 7.423785209655762,
"learning_rate": 1.139240506329114e-05,
"loss": 1.7348,
"step": 18
},
{
"epoch": 0.018039401851412295,
"grad_norm": 7.429481029510498,
"learning_rate": 1.2025316455696203e-05,
"loss": 1.4835,
"step": 19
},
{
"epoch": 0.018988844054118204,
"grad_norm": 6.7193284034729,
"learning_rate": 1.2658227848101267e-05,
"loss": 1.2373,
"step": 20
},
{
"epoch": 0.019938286256824117,
"grad_norm": 4.46099853515625,
"learning_rate": 1.3291139240506329e-05,
"loss": 1.1095,
"step": 21
},
{
"epoch": 0.020887728459530026,
"grad_norm": 3.001573085784912,
"learning_rate": 1.3924050632911393e-05,
"loss": 0.8642,
"step": 22
},
{
"epoch": 0.021837170662235936,
"grad_norm": 2.197000026702881,
"learning_rate": 1.4556962025316457e-05,
"loss": 0.7734,
"step": 23
},
{
"epoch": 0.022786612864941845,
"grad_norm": 1.8113943338394165,
"learning_rate": 1.5189873417721521e-05,
"loss": 0.7341,
"step": 24
},
{
"epoch": 0.023736055067647758,
"grad_norm": 1.7461305856704712,
"learning_rate": 1.5822784810126583e-05,
"loss": 0.6743,
"step": 25
},
{
"epoch": 0.024685497270353667,
"grad_norm": 1.3315849304199219,
"learning_rate": 1.6455696202531644e-05,
"loss": 0.5975,
"step": 26
},
{
"epoch": 0.025634939473059577,
"grad_norm": 0.726314127445221,
"learning_rate": 1.7088607594936708e-05,
"loss": 0.5659,
"step": 27
},
{
"epoch": 0.02658438167576549,
"grad_norm": 0.6269010901451111,
"learning_rate": 1.7721518987341772e-05,
"loss": 0.5226,
"step": 28
},
{
"epoch": 0.0275338238784714,
"grad_norm": 0.5819966197013855,
"learning_rate": 1.8354430379746836e-05,
"loss": 0.6132,
"step": 29
},
{
"epoch": 0.028483266081177308,
"grad_norm": 0.6247850060462952,
"learning_rate": 1.89873417721519e-05,
"loss": 0.5,
"step": 30
},
{
"epoch": 0.029432708283883217,
"grad_norm": 0.702621579170227,
"learning_rate": 1.962025316455696e-05,
"loss": 0.4958,
"step": 31
},
{
"epoch": 0.03038215048658913,
"grad_norm": 0.6045309901237488,
"learning_rate": 2.0253164556962025e-05,
"loss": 0.4405,
"step": 32
},
{
"epoch": 0.031331592689295036,
"grad_norm": 0.5436626076698303,
"learning_rate": 2.088607594936709e-05,
"loss": 0.5607,
"step": 33
},
{
"epoch": 0.03228103489200095,
"grad_norm": 0.43146297335624695,
"learning_rate": 2.1518987341772153e-05,
"loss": 0.3987,
"step": 34
},
{
"epoch": 0.03323047709470686,
"grad_norm": 0.5124548673629761,
"learning_rate": 2.2151898734177217e-05,
"loss": 0.5084,
"step": 35
},
{
"epoch": 0.03417991929741277,
"grad_norm": 0.4466649293899536,
"learning_rate": 2.278481012658228e-05,
"loss": 0.3761,
"step": 36
},
{
"epoch": 0.03512936150011868,
"grad_norm": 0.41221529245376587,
"learning_rate": 2.341772151898734e-05,
"loss": 0.3859,
"step": 37
},
{
"epoch": 0.03607880370282459,
"grad_norm": 0.3802257180213928,
"learning_rate": 2.4050632911392405e-05,
"loss": 0.3447,
"step": 38
},
{
"epoch": 0.0370282459055305,
"grad_norm": 0.47727710008621216,
"learning_rate": 2.468354430379747e-05,
"loss": 0.3914,
"step": 39
},
{
"epoch": 0.03797768810823641,
"grad_norm": 0.41048529744148254,
"learning_rate": 2.5316455696202533e-05,
"loss": 0.2988,
"step": 40
},
{
"epoch": 0.038927130310942325,
"grad_norm": 0.5019667744636536,
"learning_rate": 2.5949367088607597e-05,
"loss": 0.2938,
"step": 41
},
{
"epoch": 0.039876572513648234,
"grad_norm": 0.42121732234954834,
"learning_rate": 2.6582278481012658e-05,
"loss": 0.2579,
"step": 42
},
{
"epoch": 0.04082601471635414,
"grad_norm": 0.4193897247314453,
"learning_rate": 2.7215189873417722e-05,
"loss": 0.3262,
"step": 43
},
{
"epoch": 0.04177545691906005,
"grad_norm": 0.2978931665420532,
"learning_rate": 2.7848101265822786e-05,
"loss": 0.2365,
"step": 44
},
{
"epoch": 0.04272489912176596,
"grad_norm": 0.34771448373794556,
"learning_rate": 2.848101265822785e-05,
"loss": 0.2364,
"step": 45
},
{
"epoch": 0.04367434132447187,
"grad_norm": 0.3881576955318451,
"learning_rate": 2.9113924050632914e-05,
"loss": 0.286,
"step": 46
},
{
"epoch": 0.04462378352717778,
"grad_norm": 0.33863797783851624,
"learning_rate": 2.9746835443037974e-05,
"loss": 0.2739,
"step": 47
},
{
"epoch": 0.04557322572988369,
"grad_norm": 0.2894616723060608,
"learning_rate": 3.0379746835443042e-05,
"loss": 0.2587,
"step": 48
},
{
"epoch": 0.046522667932589606,
"grad_norm": 0.22292694449424744,
"learning_rate": 3.10126582278481e-05,
"loss": 0.1861,
"step": 49
},
{
"epoch": 0.047472110135295516,
"grad_norm": 0.21907460689544678,
"learning_rate": 3.1645569620253167e-05,
"loss": 0.1755,
"step": 50
},
{
"epoch": 0.048421552338001425,
"grad_norm": 0.29593944549560547,
"learning_rate": 3.227848101265823e-05,
"loss": 0.1856,
"step": 51
},
{
"epoch": 0.049370994540707334,
"grad_norm": 0.23055657744407654,
"learning_rate": 3.291139240506329e-05,
"loss": 0.2102,
"step": 52
},
{
"epoch": 0.050320436743413244,
"grad_norm": 0.18929323554039001,
"learning_rate": 3.354430379746836e-05,
"loss": 0.1909,
"step": 53
},
{
"epoch": 0.05126987894611915,
"grad_norm": 0.15004883706569672,
"learning_rate": 3.4177215189873416e-05,
"loss": 0.1619,
"step": 54
},
{
"epoch": 0.05221932114882506,
"grad_norm": 0.15621644258499146,
"learning_rate": 3.4810126582278487e-05,
"loss": 0.1759,
"step": 55
},
{
"epoch": 0.05316876335153098,
"grad_norm": 0.16266578435897827,
"learning_rate": 3.5443037974683544e-05,
"loss": 0.1657,
"step": 56
},
{
"epoch": 0.05411820555423689,
"grad_norm": 0.14417718350887299,
"learning_rate": 3.607594936708861e-05,
"loss": 0.1698,
"step": 57
},
{
"epoch": 0.0550676477569428,
"grad_norm": 0.21402889490127563,
"learning_rate": 3.670886075949367e-05,
"loss": 0.2185,
"step": 58
},
{
"epoch": 0.05601708995964871,
"grad_norm": 0.1997889280319214,
"learning_rate": 3.7341772151898736e-05,
"loss": 0.2143,
"step": 59
},
{
"epoch": 0.056966532162354616,
"grad_norm": 0.13755086064338684,
"learning_rate": 3.79746835443038e-05,
"loss": 0.1677,
"step": 60
},
{
"epoch": 0.057915974365060525,
"grad_norm": 0.19304363429546356,
"learning_rate": 3.8607594936708864e-05,
"loss": 0.2113,
"step": 61
},
{
"epoch": 0.058865416567766435,
"grad_norm": 0.14066031575202942,
"learning_rate": 3.924050632911392e-05,
"loss": 0.1612,
"step": 62
},
{
"epoch": 0.059814858770472344,
"grad_norm": 0.13375213742256165,
"learning_rate": 3.987341772151899e-05,
"loss": 0.164,
"step": 63
},
{
"epoch": 0.06076430097317826,
"grad_norm": 0.15216922760009766,
"learning_rate": 4.050632911392405e-05,
"loss": 0.16,
"step": 64
},
{
"epoch": 0.06171374317588417,
"grad_norm": 0.16130389273166656,
"learning_rate": 4.113924050632912e-05,
"loss": 0.1957,
"step": 65
},
{
"epoch": 0.06266318537859007,
"grad_norm": 0.1791229248046875,
"learning_rate": 4.177215189873418e-05,
"loss": 0.1993,
"step": 66
},
{
"epoch": 0.06361262758129599,
"grad_norm": 0.11038907617330551,
"learning_rate": 4.240506329113924e-05,
"loss": 0.1517,
"step": 67
},
{
"epoch": 0.0645620697840019,
"grad_norm": 0.13327902555465698,
"learning_rate": 4.3037974683544305e-05,
"loss": 0.1501,
"step": 68
},
{
"epoch": 0.06551151198670781,
"grad_norm": 0.13731731474399567,
"learning_rate": 4.367088607594937e-05,
"loss": 0.1596,
"step": 69
},
{
"epoch": 0.06646095418941372,
"grad_norm": 0.13924308121204376,
"learning_rate": 4.430379746835443e-05,
"loss": 0.152,
"step": 70
},
{
"epoch": 0.06741039639211963,
"grad_norm": 0.1482289433479309,
"learning_rate": 4.49367088607595e-05,
"loss": 0.1536,
"step": 71
},
{
"epoch": 0.06835983859482554,
"grad_norm": 0.10759364813566208,
"learning_rate": 4.556962025316456e-05,
"loss": 0.1543,
"step": 72
},
{
"epoch": 0.06930928079753144,
"grad_norm": 0.12899678945541382,
"learning_rate": 4.6202531645569625e-05,
"loss": 0.165,
"step": 73
},
{
"epoch": 0.07025872300023736,
"grad_norm": 0.11689919233322144,
"learning_rate": 4.683544303797468e-05,
"loss": 0.1564,
"step": 74
},
{
"epoch": 0.07120816520294328,
"grad_norm": 0.12697139382362366,
"learning_rate": 4.7468354430379746e-05,
"loss": 0.162,
"step": 75
},
{
"epoch": 0.07215760740564918,
"grad_norm": 0.12069376558065414,
"learning_rate": 4.810126582278481e-05,
"loss": 0.1467,
"step": 76
},
{
"epoch": 0.0731070496083551,
"grad_norm": 0.10199815034866333,
"learning_rate": 4.8734177215189874e-05,
"loss": 0.1528,
"step": 77
},
{
"epoch": 0.074056491811061,
"grad_norm": 0.1142750009894371,
"learning_rate": 4.936708860759494e-05,
"loss": 0.1574,
"step": 78
},
{
"epoch": 0.07500593401376691,
"grad_norm": 0.11019093543291092,
"learning_rate": 5e-05,
"loss": 0.1512,
"step": 79
},
{
"epoch": 0.07595537621647282,
"grad_norm": 0.09426973015069962,
"learning_rate": 5.0632911392405066e-05,
"loss": 0.1481,
"step": 80
},
{
"epoch": 0.07690481841917873,
"grad_norm": 0.09757663309574127,
"learning_rate": 5.1265822784810124e-05,
"loss": 0.1484,
"step": 81
},
{
"epoch": 0.07785426062188465,
"grad_norm": 0.10646392405033112,
"learning_rate": 5.1898734177215194e-05,
"loss": 0.1549,
"step": 82
},
{
"epoch": 0.07880370282459055,
"grad_norm": 0.12109784036874771,
"learning_rate": 5.253164556962026e-05,
"loss": 0.1448,
"step": 83
},
{
"epoch": 0.07975314502729647,
"grad_norm": 0.12039211392402649,
"learning_rate": 5.3164556962025316e-05,
"loss": 0.1538,
"step": 84
},
{
"epoch": 0.08070258723000237,
"grad_norm": 0.16873961687088013,
"learning_rate": 5.379746835443038e-05,
"loss": 0.1971,
"step": 85
},
{
"epoch": 0.08165202943270829,
"grad_norm": 0.12140022218227386,
"learning_rate": 5.4430379746835444e-05,
"loss": 0.1497,
"step": 86
},
{
"epoch": 0.08260147163541419,
"grad_norm": 0.14637599885463715,
"learning_rate": 5.5063291139240514e-05,
"loss": 0.1958,
"step": 87
},
{
"epoch": 0.0835509138381201,
"grad_norm": 0.1141396313905716,
"learning_rate": 5.569620253164557e-05,
"loss": 0.1457,
"step": 88
},
{
"epoch": 0.08450035604082601,
"grad_norm": 0.2128390371799469,
"learning_rate": 5.6329113924050636e-05,
"loss": 0.2339,
"step": 89
},
{
"epoch": 0.08544979824353192,
"grad_norm": 0.18838858604431152,
"learning_rate": 5.69620253164557e-05,
"loss": 0.2029,
"step": 90
},
{
"epoch": 0.08639924044623784,
"grad_norm": 0.19592566788196564,
"learning_rate": 5.759493670886076e-05,
"loss": 0.2276,
"step": 91
},
{
"epoch": 0.08734868264894374,
"grad_norm": 0.14753012359142303,
"learning_rate": 5.822784810126583e-05,
"loss": 0.1916,
"step": 92
},
{
"epoch": 0.08829812485164966,
"grad_norm": 0.1494351178407669,
"learning_rate": 5.886075949367089e-05,
"loss": 0.1913,
"step": 93
},
{
"epoch": 0.08924756705435556,
"grad_norm": 0.1173478439450264,
"learning_rate": 5.949367088607595e-05,
"loss": 0.1438,
"step": 94
},
{
"epoch": 0.09019700925706148,
"grad_norm": 0.12023188918828964,
"learning_rate": 6.012658227848101e-05,
"loss": 0.1516,
"step": 95
},
{
"epoch": 0.09114645145976738,
"grad_norm": 0.1275833547115326,
"learning_rate": 6.0759493670886084e-05,
"loss": 0.1492,
"step": 96
},
{
"epoch": 0.0920958936624733,
"grad_norm": 0.1360282599925995,
"learning_rate": 6.139240506329115e-05,
"loss": 0.1507,
"step": 97
},
{
"epoch": 0.09304533586517921,
"grad_norm": 0.1586841195821762,
"learning_rate": 6.20253164556962e-05,
"loss": 0.1956,
"step": 98
},
{
"epoch": 0.09399477806788512,
"grad_norm": 0.14281995594501495,
"learning_rate": 6.265822784810128e-05,
"loss": 0.1774,
"step": 99
},
{
"epoch": 0.09494422027059103,
"grad_norm": 0.12553077936172485,
"learning_rate": 6.329113924050633e-05,
"loss": 0.148,
"step": 100
},
{
"epoch": 0.09589366247329693,
"grad_norm": 0.1117570698261261,
"learning_rate": 6.392405063291139e-05,
"loss": 0.16,
"step": 101
},
{
"epoch": 0.09684310467600285,
"grad_norm": 0.13955281674861908,
"learning_rate": 6.455696202531646e-05,
"loss": 0.1464,
"step": 102
},
{
"epoch": 0.09779254687870875,
"grad_norm": 0.10990285873413086,
"learning_rate": 6.518987341772153e-05,
"loss": 0.147,
"step": 103
},
{
"epoch": 0.09874198908141467,
"grad_norm": 0.10545991361141205,
"learning_rate": 6.582278481012658e-05,
"loss": 0.1436,
"step": 104
},
{
"epoch": 0.09969143128412059,
"grad_norm": 0.1717437207698822,
"learning_rate": 6.645569620253165e-05,
"loss": 0.2278,
"step": 105
},
{
"epoch": 0.10064087348682649,
"grad_norm": 0.10950994491577148,
"learning_rate": 6.708860759493672e-05,
"loss": 0.1493,
"step": 106
},
{
"epoch": 0.1015903156895324,
"grad_norm": 0.11200258880853653,
"learning_rate": 6.772151898734177e-05,
"loss": 0.1536,
"step": 107
},
{
"epoch": 0.1025397578922383,
"grad_norm": 0.10955105721950531,
"learning_rate": 6.835443037974683e-05,
"loss": 0.1483,
"step": 108
},
{
"epoch": 0.10348920009494422,
"grad_norm": 0.11920775473117828,
"learning_rate": 6.89873417721519e-05,
"loss": 0.1492,
"step": 109
},
{
"epoch": 0.10443864229765012,
"grad_norm": 0.1390092819929123,
"learning_rate": 6.962025316455697e-05,
"loss": 0.1849,
"step": 110
},
{
"epoch": 0.10538808450035604,
"grad_norm": 0.1363140493631363,
"learning_rate": 7.025316455696203e-05,
"loss": 0.1849,
"step": 111
},
{
"epoch": 0.10633752670306196,
"grad_norm": 0.09190025180578232,
"learning_rate": 7.088607594936709e-05,
"loss": 0.1587,
"step": 112
},
{
"epoch": 0.10728696890576786,
"grad_norm": 0.09020426124334335,
"learning_rate": 7.151898734177216e-05,
"loss": 0.1377,
"step": 113
},
{
"epoch": 0.10823641110847378,
"grad_norm": 0.10544883459806442,
"learning_rate": 7.215189873417722e-05,
"loss": 0.1516,
"step": 114
},
{
"epoch": 0.10918585331117968,
"grad_norm": 0.12401281297206879,
"learning_rate": 7.278481012658229e-05,
"loss": 0.154,
"step": 115
},
{
"epoch": 0.1101352955138856,
"grad_norm": 0.1008707657456398,
"learning_rate": 7.341772151898734e-05,
"loss": 0.1448,
"step": 116
},
{
"epoch": 0.1110847377165915,
"grad_norm": 0.10302747040987015,
"learning_rate": 7.40506329113924e-05,
"loss": 0.1451,
"step": 117
},
{
"epoch": 0.11203417991929741,
"grad_norm": 0.12748293578624725,
"learning_rate": 7.468354430379747e-05,
"loss": 0.1829,
"step": 118
},
{
"epoch": 0.11298362212200333,
"grad_norm": 0.10413361340761185,
"learning_rate": 7.531645569620254e-05,
"loss": 0.1371,
"step": 119
},
{
"epoch": 0.11393306432470923,
"grad_norm": 0.1243433803319931,
"learning_rate": 7.59493670886076e-05,
"loss": 0.1409,
"step": 120
},
{
"epoch": 0.11488250652741515,
"grad_norm": 0.11630933731794357,
"learning_rate": 7.658227848101266e-05,
"loss": 0.1372,
"step": 121
},
{
"epoch": 0.11583194873012105,
"grad_norm": 0.17981529235839844,
"learning_rate": 7.721518987341773e-05,
"loss": 0.2257,
"step": 122
},
{
"epoch": 0.11678139093282697,
"grad_norm": 0.14063452184200287,
"learning_rate": 7.78481012658228e-05,
"loss": 0.1841,
"step": 123
},
{
"epoch": 0.11773083313553287,
"grad_norm": 0.1264188438653946,
"learning_rate": 7.848101265822784e-05,
"loss": 0.1471,
"step": 124
},
{
"epoch": 0.11868027533823879,
"grad_norm": 0.12827955186367035,
"learning_rate": 7.911392405063291e-05,
"loss": 0.1493,
"step": 125
},
{
"epoch": 0.11962971754094469,
"grad_norm": 0.09800329059362411,
"learning_rate": 7.974683544303798e-05,
"loss": 0.1414,
"step": 126
},
{
"epoch": 0.1205791597436506,
"grad_norm": 0.09902197122573853,
"learning_rate": 8.037974683544304e-05,
"loss": 0.1462,
"step": 127
},
{
"epoch": 0.12152860194635652,
"grad_norm": 0.09450504928827286,
"learning_rate": 8.10126582278481e-05,
"loss": 0.1484,
"step": 128
},
{
"epoch": 0.12247804414906242,
"grad_norm": 0.11012883484363556,
"learning_rate": 8.164556962025317e-05,
"loss": 0.1437,
"step": 129
},
{
"epoch": 0.12342748635176834,
"grad_norm": 0.11717642843723297,
"learning_rate": 8.227848101265824e-05,
"loss": 0.1478,
"step": 130
},
{
"epoch": 0.12437692855447424,
"grad_norm": 0.08754123747348785,
"learning_rate": 8.29113924050633e-05,
"loss": 0.1408,
"step": 131
},
{
"epoch": 0.12532637075718014,
"grad_norm": 0.10017862170934677,
"learning_rate": 8.354430379746835e-05,
"loss": 0.1476,
"step": 132
},
{
"epoch": 0.12627581295988607,
"grad_norm": 0.08994068205356598,
"learning_rate": 8.417721518987342e-05,
"loss": 0.1478,
"step": 133
},
{
"epoch": 0.12722525516259198,
"grad_norm": 0.09894968569278717,
"learning_rate": 8.481012658227848e-05,
"loss": 0.1309,
"step": 134
},
{
"epoch": 0.12817469736529788,
"grad_norm": 0.10028701275587082,
"learning_rate": 8.544303797468355e-05,
"loss": 0.1433,
"step": 135
},
{
"epoch": 0.1291241395680038,
"grad_norm": 0.0897536426782608,
"learning_rate": 8.607594936708861e-05,
"loss": 0.1459,
"step": 136
},
{
"epoch": 0.1300735817707097,
"grad_norm": 0.10435349494218826,
"learning_rate": 8.670886075949367e-05,
"loss": 0.1434,
"step": 137
},
{
"epoch": 0.13102302397341561,
"grad_norm": 0.11718117445707321,
"learning_rate": 8.734177215189874e-05,
"loss": 0.1509,
"step": 138
},
{
"epoch": 0.13197246617612152,
"grad_norm": 0.14426474273204803,
"learning_rate": 8.797468354430381e-05,
"loss": 0.1373,
"step": 139
},
{
"epoch": 0.13292190837882745,
"grad_norm": 0.13101965188980103,
"learning_rate": 8.860759493670887e-05,
"loss": 0.1358,
"step": 140
},
{
"epoch": 0.13387135058153335,
"grad_norm": 0.11235956102609634,
"learning_rate": 8.924050632911392e-05,
"loss": 0.1394,
"step": 141
},
{
"epoch": 0.13482079278423925,
"grad_norm": 0.11327100545167923,
"learning_rate": 8.9873417721519e-05,
"loss": 0.1443,
"step": 142
},
{
"epoch": 0.13577023498694518,
"grad_norm": 0.10912016034126282,
"learning_rate": 9.050632911392407e-05,
"loss": 0.1698,
"step": 143
},
{
"epoch": 0.13671967718965108,
"grad_norm": 0.16535617411136627,
"learning_rate": 9.113924050632912e-05,
"loss": 0.2255,
"step": 144
},
{
"epoch": 0.137669119392357,
"grad_norm": 0.10184327512979507,
"learning_rate": 9.177215189873418e-05,
"loss": 0.1371,
"step": 145
},
{
"epoch": 0.1386185615950629,
"grad_norm": 0.10998040437698364,
"learning_rate": 9.240506329113925e-05,
"loss": 0.1794,
"step": 146
},
{
"epoch": 0.13956800379776882,
"grad_norm": 0.08974044770002365,
"learning_rate": 9.303797468354431e-05,
"loss": 0.144,
"step": 147
},
{
"epoch": 0.14051744600047472,
"grad_norm": 0.12724193930625916,
"learning_rate": 9.367088607594936e-05,
"loss": 0.1794,
"step": 148
},
{
"epoch": 0.14146688820318062,
"grad_norm": 0.1079091802239418,
"learning_rate": 9.430379746835444e-05,
"loss": 0.1399,
"step": 149
},
{
"epoch": 0.14241633040588655,
"grad_norm": 0.09480807185173035,
"learning_rate": 9.493670886075949e-05,
"loss": 0.1395,
"step": 150
},
{
"epoch": 0.14336577260859246,
"grad_norm": 0.08620745688676834,
"learning_rate": 9.556962025316456e-05,
"loss": 0.1415,
"step": 151
},
{
"epoch": 0.14431521481129836,
"grad_norm": 0.10517002642154694,
"learning_rate": 9.620253164556962e-05,
"loss": 0.1723,
"step": 152
},
{
"epoch": 0.14526465701400426,
"grad_norm": 0.0956311896443367,
"learning_rate": 9.683544303797469e-05,
"loss": 0.1515,
"step": 153
},
{
"epoch": 0.1462140992167102,
"grad_norm": 0.08050324022769928,
"learning_rate": 9.746835443037975e-05,
"loss": 0.1322,
"step": 154
},
{
"epoch": 0.1471635414194161,
"grad_norm": 0.0853201299905777,
"learning_rate": 9.810126582278482e-05,
"loss": 0.142,
"step": 155
},
{
"epoch": 0.148112983622122,
"grad_norm": 0.09991180151700974,
"learning_rate": 9.873417721518988e-05,
"loss": 0.1348,
"step": 156
},
{
"epoch": 0.14906242582482793,
"grad_norm": 0.08640603721141815,
"learning_rate": 9.936708860759493e-05,
"loss": 0.1397,
"step": 157
},
{
"epoch": 0.15001186802753383,
"grad_norm": 0.09057717025279999,
"learning_rate": 0.0001,
"loss": 0.1381,
"step": 158
},
{
"epoch": 0.15096131023023973,
"grad_norm": 0.09916041046380997,
"learning_rate": 0.00010063291139240508,
"loss": 0.1509,
"step": 159
},
{
"epoch": 0.15191075243294563,
"grad_norm": 0.09434045851230621,
"learning_rate": 0.00010126582278481013,
"loss": 0.1388,
"step": 160
},
{
"epoch": 0.15286019463565156,
"grad_norm": 0.1273377537727356,
"learning_rate": 0.0001018987341772152,
"loss": 0.1401,
"step": 161
},
{
"epoch": 0.15380963683835747,
"grad_norm": 0.1297912299633026,
"learning_rate": 0.00010253164556962025,
"loss": 0.1852,
"step": 162
},
{
"epoch": 0.15475907904106337,
"grad_norm": 0.1151595488190651,
"learning_rate": 0.00010316455696202532,
"loss": 0.1848,
"step": 163
},
{
"epoch": 0.1557085212437693,
"grad_norm": 0.13381290435791016,
"learning_rate": 0.00010379746835443039,
"loss": 0.1438,
"step": 164
},
{
"epoch": 0.1566579634464752,
"grad_norm": 0.07880119979381561,
"learning_rate": 0.00010443037974683545,
"loss": 0.1327,
"step": 165
},
{
"epoch": 0.1576074056491811,
"grad_norm": 0.0843740776181221,
"learning_rate": 0.00010506329113924052,
"loss": 0.1398,
"step": 166
},
{
"epoch": 0.158556847851887,
"grad_norm": 0.0981813594698906,
"learning_rate": 0.00010569620253164559,
"loss": 0.1409,
"step": 167
},
{
"epoch": 0.15950629005459294,
"grad_norm": 0.10005304962396622,
"learning_rate": 0.00010632911392405063,
"loss": 0.1783,
"step": 168
},
{
"epoch": 0.16045573225729884,
"grad_norm": 0.08365727961063385,
"learning_rate": 0.00010696202531645569,
"loss": 0.1275,
"step": 169
},
{
"epoch": 0.16140517446000474,
"grad_norm": 0.1017635315656662,
"learning_rate": 0.00010759493670886076,
"loss": 0.1792,
"step": 170
},
{
"epoch": 0.16235461666271067,
"grad_norm": 0.07007888704538345,
"learning_rate": 0.00010822784810126583,
"loss": 0.1473,
"step": 171
},
{
"epoch": 0.16330405886541657,
"grad_norm": 0.07718679308891296,
"learning_rate": 0.00010886075949367089,
"loss": 0.1396,
"step": 172
},
{
"epoch": 0.16425350106812248,
"grad_norm": 0.07228100299835205,
"learning_rate": 0.00010949367088607596,
"loss": 0.1398,
"step": 173
},
{
"epoch": 0.16520294327082838,
"grad_norm": 0.07955378293991089,
"learning_rate": 0.00011012658227848103,
"loss": 0.1402,
"step": 174
},
{
"epoch": 0.1661523854735343,
"grad_norm": 0.0816427692770958,
"learning_rate": 0.00011075949367088607,
"loss": 0.1345,
"step": 175
},
{
"epoch": 0.1671018276762402,
"grad_norm": 0.07641757279634476,
"learning_rate": 0.00011139240506329114,
"loss": 0.1373,
"step": 176
},
{
"epoch": 0.1680512698789461,
"grad_norm": 0.07354450225830078,
"learning_rate": 0.0001120253164556962,
"loss": 0.1394,
"step": 177
},
{
"epoch": 0.16900071208165202,
"grad_norm": 0.08322398364543915,
"learning_rate": 0.00011265822784810127,
"loss": 0.138,
"step": 178
},
{
"epoch": 0.16995015428435795,
"grad_norm": 0.13528607785701752,
"learning_rate": 0.00011329113924050634,
"loss": 0.2188,
"step": 179
},
{
"epoch": 0.17089959648706385,
"grad_norm": 0.10803692042827606,
"learning_rate": 0.0001139240506329114,
"loss": 0.1782,
"step": 180
},
{
"epoch": 0.17184903868976975,
"grad_norm": 0.08404573053121567,
"learning_rate": 0.00011455696202531647,
"loss": 0.1394,
"step": 181
},
{
"epoch": 0.17279848089247568,
"grad_norm": 0.12790893018245697,
"learning_rate": 0.00011518987341772151,
"loss": 0.2157,
"step": 182
},
{
"epoch": 0.17374792309518158,
"grad_norm": 0.09879907220602036,
"learning_rate": 0.00011582278481012658,
"loss": 0.1693,
"step": 183
},
{
"epoch": 0.17469736529788749,
"grad_norm": 0.08092228323221207,
"learning_rate": 0.00011645569620253166,
"loss": 0.136,
"step": 184
},
{
"epoch": 0.1756468075005934,
"grad_norm": 0.07660632580518723,
"learning_rate": 0.00011708860759493671,
"loss": 0.1332,
"step": 185
},
{
"epoch": 0.17659624970329932,
"grad_norm": 0.07474201172590256,
"learning_rate": 0.00011772151898734178,
"loss": 0.1301,
"step": 186
},
{
"epoch": 0.17754569190600522,
"grad_norm": 0.09162931889295578,
"learning_rate": 0.00011835443037974685,
"loss": 0.1407,
"step": 187
},
{
"epoch": 0.17849513410871112,
"grad_norm": 0.08646775782108307,
"learning_rate": 0.0001189873417721519,
"loss": 0.139,
"step": 188
},
{
"epoch": 0.17944457631141705,
"grad_norm": 0.0759253203868866,
"learning_rate": 0.00011962025316455696,
"loss": 0.1342,
"step": 189
},
{
"epoch": 0.18039401851412296,
"grad_norm": 0.08292865008115768,
"learning_rate": 0.00012025316455696203,
"loss": 0.1389,
"step": 190
},
{
"epoch": 0.18134346071682886,
"grad_norm": 0.12379574030637741,
"learning_rate": 0.0001208860759493671,
"loss": 0.1795,
"step": 191
},
{
"epoch": 0.18229290291953476,
"grad_norm": 0.10240278393030167,
"learning_rate": 0.00012151898734177217,
"loss": 0.1721,
"step": 192
},
{
"epoch": 0.1832423451222407,
"grad_norm": 0.09666036069393158,
"learning_rate": 0.00012215189873417722,
"loss": 0.1783,
"step": 193
},
{
"epoch": 0.1841917873249466,
"grad_norm": 0.08314768224954605,
"learning_rate": 0.0001227848101265823,
"loss": 0.1429,
"step": 194
},
{
"epoch": 0.1851412295276525,
"grad_norm": 0.07590368390083313,
"learning_rate": 0.00012341772151898734,
"loss": 0.1393,
"step": 195
},
{
"epoch": 0.18609067173035843,
"grad_norm": 0.10585250705480576,
"learning_rate": 0.0001240506329113924,
"loss": 0.2155,
"step": 196
},
{
"epoch": 0.18704011393306433,
"grad_norm": 0.06995555013418198,
"learning_rate": 0.00012468354430379748,
"loss": 0.1374,
"step": 197
},
{
"epoch": 0.18798955613577023,
"grad_norm": 0.07370735704898834,
"learning_rate": 0.00012531645569620255,
"loss": 0.1367,
"step": 198
},
{
"epoch": 0.18893899833847613,
"grad_norm": 0.07194443792104721,
"learning_rate": 0.0001259493670886076,
"loss": 0.1437,
"step": 199
},
{
"epoch": 0.18988844054118206,
"grad_norm": 0.06982647627592087,
"learning_rate": 0.00012658227848101267,
"loss": 0.1358,
"step": 200
},
{
"epoch": 0.19083788274388797,
"grad_norm": 0.06538347154855728,
"learning_rate": 0.0001272151898734177,
"loss": 0.1354,
"step": 201
},
{
"epoch": 0.19178732494659387,
"grad_norm": 0.07789324969053268,
"learning_rate": 0.00012784810126582278,
"loss": 0.178,
"step": 202
},
{
"epoch": 0.1927367671492998,
"grad_norm": 0.07376820594072342,
"learning_rate": 0.00012848101265822785,
"loss": 0.1621,
"step": 203
},
{
"epoch": 0.1936862093520057,
"grad_norm": 0.0720745250582695,
"learning_rate": 0.00012911392405063292,
"loss": 0.132,
"step": 204
},
{
"epoch": 0.1946356515547116,
"grad_norm": 0.06211116537451744,
"learning_rate": 0.000129746835443038,
"loss": 0.1387,
"step": 205
},
{
"epoch": 0.1955850937574175,
"grad_norm": 0.06701771914958954,
"learning_rate": 0.00013037974683544306,
"loss": 0.14,
"step": 206
},
{
"epoch": 0.19653453596012344,
"grad_norm": 0.07692532986402512,
"learning_rate": 0.0001310126582278481,
"loss": 0.1322,
"step": 207
},
{
"epoch": 0.19748397816282934,
"grad_norm": 0.07763269543647766,
"learning_rate": 0.00013164556962025315,
"loss": 0.1393,
"step": 208
},
{
"epoch": 0.19843342036553524,
"grad_norm": 0.08769022673368454,
"learning_rate": 0.00013227848101265822,
"loss": 0.1489,
"step": 209
},
{
"epoch": 0.19938286256824117,
"grad_norm": 0.08881859481334686,
"learning_rate": 0.0001329113924050633,
"loss": 0.1765,
"step": 210
},
{
"epoch": 0.20033230477094707,
"grad_norm": 0.06811822950839996,
"learning_rate": 0.00013354430379746836,
"loss": 0.1332,
"step": 211
},
{
"epoch": 0.20128174697365298,
"grad_norm": 0.06390922516584396,
"learning_rate": 0.00013417721518987343,
"loss": 0.1343,
"step": 212
},
{
"epoch": 0.20223118917635888,
"grad_norm": 0.06630406528711319,
"learning_rate": 0.0001348101265822785,
"loss": 0.1329,
"step": 213
},
{
"epoch": 0.2031806313790648,
"grad_norm": 0.0730772465467453,
"learning_rate": 0.00013544303797468355,
"loss": 0.1354,
"step": 214
},
{
"epoch": 0.2041300735817707,
"grad_norm": 0.06487323343753815,
"learning_rate": 0.00013607594936708862,
"loss": 0.1297,
"step": 215
},
{
"epoch": 0.2050795157844766,
"grad_norm": 0.06967955082654953,
"learning_rate": 0.00013670886075949366,
"loss": 0.1398,
"step": 216
},
{
"epoch": 0.20602895798718254,
"grad_norm": 0.08531820774078369,
"learning_rate": 0.00013734177215189873,
"loss": 0.1336,
"step": 217
},
{
"epoch": 0.20697840018988845,
"grad_norm": 0.0757659375667572,
"learning_rate": 0.0001379746835443038,
"loss": 0.1606,
"step": 218
},
{
"epoch": 0.20792784239259435,
"grad_norm": 0.060206469148397446,
"learning_rate": 0.00013860759493670888,
"loss": 0.1337,
"step": 219
},
{
"epoch": 0.20887728459530025,
"grad_norm": 0.07996556162834167,
"learning_rate": 0.00013924050632911395,
"loss": 0.1308,
"step": 220
},
{
"epoch": 0.20982672679800618,
"grad_norm": 0.06206861138343811,
"learning_rate": 0.000139873417721519,
"loss": 0.1347,
"step": 221
},
{
"epoch": 0.21077616900071208,
"grad_norm": 0.08736416697502136,
"learning_rate": 0.00014050632911392406,
"loss": 0.1768,
"step": 222
},
{
"epoch": 0.21172561120341798,
"grad_norm": 0.06427916139364243,
"learning_rate": 0.00014113924050632913,
"loss": 0.1374,
"step": 223
},
{
"epoch": 0.21267505340612392,
"grad_norm": 0.10996536910533905,
"learning_rate": 0.00014177215189873418,
"loss": 0.222,
"step": 224
},
{
"epoch": 0.21362449560882982,
"grad_norm": 0.08439125120639801,
"learning_rate": 0.00014240506329113925,
"loss": 0.1854,
"step": 225
},
{
"epoch": 0.21457393781153572,
"grad_norm": 0.06892693787813187,
"learning_rate": 0.00014303797468354432,
"loss": 0.139,
"step": 226
},
{
"epoch": 0.21552338001424162,
"grad_norm": 0.08241122961044312,
"learning_rate": 0.0001436708860759494,
"loss": 0.173,
"step": 227
},
{
"epoch": 0.21647282221694755,
"grad_norm": 0.07911046594381332,
"learning_rate": 0.00014430379746835443,
"loss": 0.1418,
"step": 228
},
{
"epoch": 0.21742226441965345,
"grad_norm": 0.06346064805984497,
"learning_rate": 0.0001449367088607595,
"loss": 0.1406,
"step": 229
},
{
"epoch": 0.21837170662235936,
"grad_norm": 0.060393668711185455,
"learning_rate": 0.00014556962025316457,
"loss": 0.1417,
"step": 230
},
{
"epoch": 0.2193211488250653,
"grad_norm": 0.05912507325410843,
"learning_rate": 0.00014620253164556962,
"loss": 0.1298,
"step": 231
},
{
"epoch": 0.2202705910277712,
"grad_norm": 0.07730337232351303,
"learning_rate": 0.0001468354430379747,
"loss": 0.1769,
"step": 232
},
{
"epoch": 0.2212200332304771,
"grad_norm": 0.07612381875514984,
"learning_rate": 0.00014746835443037976,
"loss": 0.1338,
"step": 233
},
{
"epoch": 0.222169475433183,
"grad_norm": 0.055311791598796844,
"learning_rate": 0.0001481012658227848,
"loss": 0.1313,
"step": 234
},
{
"epoch": 0.22311891763588892,
"grad_norm": 0.08492033183574677,
"learning_rate": 0.00014873417721518987,
"loss": 0.1367,
"step": 235
},
{
"epoch": 0.22406835983859483,
"grad_norm": 0.07133237272500992,
"learning_rate": 0.00014936708860759494,
"loss": 0.1308,
"step": 236
},
{
"epoch": 0.22501780204130073,
"grad_norm": 0.07148605585098267,
"learning_rate": 0.00015000000000000001,
"loss": 0.133,
"step": 237
},
{
"epoch": 0.22596724424400666,
"grad_norm": 0.06900472939014435,
"learning_rate": 0.00015063291139240508,
"loss": 0.138,
"step": 238
},
{
"epoch": 0.22691668644671256,
"grad_norm": 0.062325432896614075,
"learning_rate": 0.00015126582278481013,
"loss": 0.1338,
"step": 239
},
{
"epoch": 0.22786612864941846,
"grad_norm": 0.06719667464494705,
"learning_rate": 0.0001518987341772152,
"loss": 0.1316,
"step": 240
},
{
"epoch": 0.22881557085212437,
"grad_norm": 0.07456009089946747,
"learning_rate": 0.00015253164556962024,
"loss": 0.1412,
"step": 241
},
{
"epoch": 0.2297650130548303,
"grad_norm": 0.05619575083255768,
"learning_rate": 0.00015316455696202531,
"loss": 0.1342,
"step": 242
},
{
"epoch": 0.2307144552575362,
"grad_norm": 0.06157098710536957,
"learning_rate": 0.00015379746835443038,
"loss": 0.1329,
"step": 243
},
{
"epoch": 0.2316638974602421,
"grad_norm": 0.06759827584028244,
"learning_rate": 0.00015443037974683546,
"loss": 0.1411,
"step": 244
},
{
"epoch": 0.232613339662948,
"grad_norm": 0.06892479956150055,
"learning_rate": 0.00015506329113924053,
"loss": 0.1484,
"step": 245
},
{
"epoch": 0.23356278186565393,
"grad_norm": 0.08536699414253235,
"learning_rate": 0.0001556962025316456,
"loss": 0.1855,
"step": 246
},
{
"epoch": 0.23451222406835984,
"grad_norm": 0.06800314784049988,
"learning_rate": 0.00015632911392405064,
"loss": 0.1379,
"step": 247
},
{
"epoch": 0.23546166627106574,
"grad_norm": 0.0625622496008873,
"learning_rate": 0.00015696202531645568,
"loss": 0.1344,
"step": 248
},
{
"epoch": 0.23641110847377167,
"grad_norm": 0.06030593812465668,
"learning_rate": 0.00015759493670886075,
"loss": 0.1254,
"step": 249
},
{
"epoch": 0.23736055067647757,
"grad_norm": 0.06694353371858597,
"learning_rate": 0.00015822784810126583,
"loss": 0.1413,
"step": 250
},
{
"epoch": 0.23830999287918347,
"grad_norm": 0.06594134122133255,
"learning_rate": 0.0001588607594936709,
"loss": 0.1394,
"step": 251
},
{
"epoch": 0.23925943508188938,
"grad_norm": 0.09062930941581726,
"learning_rate": 0.00015949367088607597,
"loss": 0.1883,
"step": 252
},
{
"epoch": 0.2402088772845953,
"grad_norm": 0.06029089167714119,
"learning_rate": 0.00016012658227848104,
"loss": 0.1271,
"step": 253
},
{
"epoch": 0.2411583194873012,
"grad_norm": 0.08471622318029404,
"learning_rate": 0.00016075949367088608,
"loss": 0.172,
"step": 254
},
{
"epoch": 0.2421077616900071,
"grad_norm": 0.061710160225629807,
"learning_rate": 0.00016139240506329115,
"loss": 0.1348,
"step": 255
},
{
"epoch": 0.24305720389271304,
"grad_norm": 0.0812671035528183,
"learning_rate": 0.0001620253164556962,
"loss": 0.1312,
"step": 256
},
{
"epoch": 0.24400664609541894,
"grad_norm": 0.06917005032300949,
"learning_rate": 0.00016265822784810127,
"loss": 0.1464,
"step": 257
},
{
"epoch": 0.24495608829812485,
"grad_norm": 0.0905887708067894,
"learning_rate": 0.00016329113924050634,
"loss": 0.1759,
"step": 258
},
{
"epoch": 0.24590553050083075,
"grad_norm": 0.05976787209510803,
"learning_rate": 0.0001639240506329114,
"loss": 0.1404,
"step": 259
},
{
"epoch": 0.24685497270353668,
"grad_norm": 0.07545675337314606,
"learning_rate": 0.00016455696202531648,
"loss": 0.1322,
"step": 260
},
{
"epoch": 0.24780441490624258,
"grad_norm": 0.07035024464130402,
"learning_rate": 0.00016518987341772152,
"loss": 0.1378,
"step": 261
},
{
"epoch": 0.24875385710894848,
"grad_norm": 0.07665737718343735,
"learning_rate": 0.0001658227848101266,
"loss": 0.1827,
"step": 262
},
{
"epoch": 0.24970329931165441,
"grad_norm": 0.06619013845920563,
"learning_rate": 0.00016645569620253166,
"loss": 0.1284,
"step": 263
},
{
"epoch": 0.2506527415143603,
"grad_norm": 0.0647001713514328,
"learning_rate": 0.0001670886075949367,
"loss": 0.133,
"step": 264
},
{
"epoch": 0.2516021837170662,
"grad_norm": 0.060702718794345856,
"learning_rate": 0.00016772151898734178,
"loss": 0.1335,
"step": 265
},
{
"epoch": 0.25255162591977215,
"grad_norm": 0.0508468896150589,
"learning_rate": 0.00016835443037974685,
"loss": 0.1333,
"step": 266
},
{
"epoch": 0.253501068122478,
"grad_norm": 0.09877864271402359,
"learning_rate": 0.0001689873417721519,
"loss": 0.2031,
"step": 267
},
{
"epoch": 0.25445051032518395,
"grad_norm": 0.06673337519168854,
"learning_rate": 0.00016962025316455696,
"loss": 0.1356,
"step": 268
},
{
"epoch": 0.2553999525278899,
"grad_norm": 0.10604165494441986,
"learning_rate": 0.00017025316455696204,
"loss": 0.2517,
"step": 269
},
{
"epoch": 0.25634939473059576,
"grad_norm": 0.07689858227968216,
"learning_rate": 0.0001708860759493671,
"loss": 0.1761,
"step": 270
},
{
"epoch": 0.2572988369333017,
"grad_norm": 0.05482449755072594,
"learning_rate": 0.00017151898734177218,
"loss": 0.131,
"step": 271
},
{
"epoch": 0.2582482791360076,
"grad_norm": 0.08622145652770996,
"learning_rate": 0.00017215189873417722,
"loss": 0.1335,
"step": 272
},
{
"epoch": 0.2591977213387135,
"grad_norm": 0.0748213455080986,
"learning_rate": 0.0001727848101265823,
"loss": 0.176,
"step": 273
},
{
"epoch": 0.2601471635414194,
"grad_norm": 0.06163305416703224,
"learning_rate": 0.00017341772151898733,
"loss": 0.1381,
"step": 274
},
{
"epoch": 0.26109660574412535,
"grad_norm": 0.06141841039061546,
"learning_rate": 0.0001740506329113924,
"loss": 0.1353,
"step": 275
},
{
"epoch": 0.26204604794683123,
"grad_norm": 0.07326913625001907,
"learning_rate": 0.00017468354430379748,
"loss": 0.1441,
"step": 276
},
{
"epoch": 0.26299549014953716,
"grad_norm": 0.05951124057173729,
"learning_rate": 0.00017531645569620255,
"loss": 0.1321,
"step": 277
},
{
"epoch": 0.26394493235224303,
"grad_norm": 0.08364073932170868,
"learning_rate": 0.00017594936708860762,
"loss": 0.187,
"step": 278
},
{
"epoch": 0.26489437455494896,
"grad_norm": 0.05849132314324379,
"learning_rate": 0.00017658227848101266,
"loss": 0.1393,
"step": 279
},
{
"epoch": 0.2658438167576549,
"grad_norm": 0.05452360957860947,
"learning_rate": 0.00017721518987341773,
"loss": 0.1342,
"step": 280
},
{
"epoch": 0.26679325896036077,
"grad_norm": 0.04878188297152519,
"learning_rate": 0.00017784810126582278,
"loss": 0.1445,
"step": 281
},
{
"epoch": 0.2677427011630667,
"grad_norm": 0.06066753342747688,
"learning_rate": 0.00017848101265822785,
"loss": 0.1423,
"step": 282
},
{
"epoch": 0.26869214336577263,
"grad_norm": 0.04918207973241806,
"learning_rate": 0.00017911392405063292,
"loss": 0.1316,
"step": 283
},
{
"epoch": 0.2696415855684785,
"grad_norm": 0.05103525519371033,
"learning_rate": 0.000179746835443038,
"loss": 0.1313,
"step": 284
},
{
"epoch": 0.27059102777118443,
"grad_norm": 0.05667628347873688,
"learning_rate": 0.00018037974683544306,
"loss": 0.1434,
"step": 285
},
{
"epoch": 0.27154046997389036,
"grad_norm": 0.06226016581058502,
"learning_rate": 0.00018101265822784813,
"loss": 0.1357,
"step": 286
},
{
"epoch": 0.27248991217659624,
"grad_norm": 0.04695293679833412,
"learning_rate": 0.00018164556962025317,
"loss": 0.1314,
"step": 287
},
{
"epoch": 0.27343935437930217,
"grad_norm": 0.05762844532728195,
"learning_rate": 0.00018227848101265824,
"loss": 0.1349,
"step": 288
},
{
"epoch": 0.27438879658200804,
"grad_norm": 0.05454534292221069,
"learning_rate": 0.0001829113924050633,
"loss": 0.1432,
"step": 289
},
{
"epoch": 0.275338238784714,
"grad_norm": 0.050270579755306244,
"learning_rate": 0.00018354430379746836,
"loss": 0.1272,
"step": 290
},
{
"epoch": 0.2762876809874199,
"grad_norm": 0.0688452497124672,
"learning_rate": 0.00018417721518987343,
"loss": 0.1708,
"step": 291
},
{
"epoch": 0.2772371231901258,
"grad_norm": 0.06213200092315674,
"learning_rate": 0.0001848101265822785,
"loss": 0.1674,
"step": 292
},
{
"epoch": 0.2781865653928317,
"grad_norm": 0.059717319905757904,
"learning_rate": 0.00018544303797468354,
"loss": 0.169,
"step": 293
},
{
"epoch": 0.27913600759553764,
"grad_norm": 0.06223325803875923,
"learning_rate": 0.00018607594936708861,
"loss": 0.1369,
"step": 294
},
{
"epoch": 0.2800854497982435,
"grad_norm": 0.053163208067417145,
"learning_rate": 0.00018670886075949369,
"loss": 0.133,
"step": 295
},
{
"epoch": 0.28103489200094944,
"grad_norm": 0.06647945195436478,
"learning_rate": 0.00018734177215189873,
"loss": 0.1438,
"step": 296
},
{
"epoch": 0.2819843342036554,
"grad_norm": 0.0588272288441658,
"learning_rate": 0.0001879746835443038,
"loss": 0.1338,
"step": 297
},
{
"epoch": 0.28293377640636125,
"grad_norm": 0.05841274932026863,
"learning_rate": 0.00018860759493670887,
"loss": 0.1329,
"step": 298
},
{
"epoch": 0.2838832186090672,
"grad_norm": 0.09033369272947311,
"learning_rate": 0.00018924050632911394,
"loss": 0.1747,
"step": 299
},
{
"epoch": 0.2848326608117731,
"grad_norm": 0.052215326577425,
"learning_rate": 0.00018987341772151899,
"loss": 0.1296,
"step": 300
},
{
"epoch": 0.285782103014479,
"grad_norm": 0.05880101025104523,
"learning_rate": 0.00019050632911392406,
"loss": 0.1287,
"step": 301
},
{
"epoch": 0.2867315452171849,
"grad_norm": 0.0691700354218483,
"learning_rate": 0.00019113924050632913,
"loss": 0.1676,
"step": 302
},
{
"epoch": 0.2876809874198908,
"grad_norm": 0.057025909423828125,
"learning_rate": 0.0001917721518987342,
"loss": 0.1346,
"step": 303
},
{
"epoch": 0.2886304296225967,
"grad_norm": 0.04936329275369644,
"learning_rate": 0.00019240506329113924,
"loss": 0.1354,
"step": 304
},
{
"epoch": 0.28957987182530265,
"grad_norm": 0.0680055245757103,
"learning_rate": 0.0001930379746835443,
"loss": 0.1344,
"step": 305
},
{
"epoch": 0.2905293140280085,
"grad_norm": 0.07374466210603714,
"learning_rate": 0.00019367088607594938,
"loss": 0.1428,
"step": 306
},
{
"epoch": 0.29147875623071445,
"grad_norm": 0.061204761266708374,
"learning_rate": 0.00019430379746835443,
"loss": 0.1246,
"step": 307
},
{
"epoch": 0.2924281984334204,
"grad_norm": 0.053467705845832825,
"learning_rate": 0.0001949367088607595,
"loss": 0.1342,
"step": 308
},
{
"epoch": 0.29337764063612626,
"grad_norm": 0.057525087147951126,
"learning_rate": 0.00019556962025316457,
"loss": 0.1377,
"step": 309
},
{
"epoch": 0.2943270828388322,
"grad_norm": 0.07857844978570938,
"learning_rate": 0.00019620253164556964,
"loss": 0.2076,
"step": 310
},
{
"epoch": 0.2952765250415381,
"grad_norm": 0.05250545218586922,
"learning_rate": 0.0001968354430379747,
"loss": 0.1432,
"step": 311
},
{
"epoch": 0.296225967244244,
"grad_norm": 0.07495012134313583,
"learning_rate": 0.00019746835443037975,
"loss": 0.1766,
"step": 312
},
{
"epoch": 0.2971754094469499,
"grad_norm": 0.04692578688263893,
"learning_rate": 0.0001981012658227848,
"loss": 0.1408,
"step": 313
},
{
"epoch": 0.29812485164965585,
"grad_norm": 0.055666085332632065,
"learning_rate": 0.00019873417721518987,
"loss": 0.1391,
"step": 314
},
{
"epoch": 0.29907429385236173,
"grad_norm": 0.050465911626815796,
"learning_rate": 0.00019936708860759494,
"loss": 0.1415,
"step": 315
},
{
"epoch": 0.30002373605506766,
"grad_norm": 0.051260240375995636,
"learning_rate": 0.0002,
"loss": 0.1423,
"step": 316
},
{
"epoch": 0.30097317825777353,
"grad_norm": 0.0503215529024601,
"learning_rate": 0.000199999938945738,
"loss": 0.1348,
"step": 317
},
{
"epoch": 0.30192262046047946,
"grad_norm": 0.04917483776807785,
"learning_rate": 0.0001999997557830265,
"loss": 0.1342,
"step": 318
},
{
"epoch": 0.3028720626631854,
"grad_norm": 0.06354209035634995,
"learning_rate": 0.00019999945051208916,
"loss": 0.1365,
"step": 319
},
{
"epoch": 0.30382150486589127,
"grad_norm": 0.04878314957022667,
"learning_rate": 0.0001999990231332988,
"loss": 0.13,
"step": 320
},
{
"epoch": 0.3047709470685972,
"grad_norm": 0.07046223431825638,
"learning_rate": 0.0001999984736471772,
"loss": 0.1394,
"step": 321
},
{
"epoch": 0.30572038927130313,
"grad_norm": 0.04456232488155365,
"learning_rate": 0.00019999780205439538,
"loss": 0.1278,
"step": 322
},
{
"epoch": 0.306669831474009,
"grad_norm": 0.06280628591775894,
"learning_rate": 0.00019999700835577342,
"loss": 0.1715,
"step": 323
},
{
"epoch": 0.30761927367671493,
"grad_norm": 0.07462131977081299,
"learning_rate": 0.00019999609255228046,
"loss": 0.1772,
"step": 324
},
{
"epoch": 0.30856871587942086,
"grad_norm": 0.059642352163791656,
"learning_rate": 0.00019999505464503482,
"loss": 0.1294,
"step": 325
},
{
"epoch": 0.30951815808212674,
"grad_norm": 0.06458820402622223,
"learning_rate": 0.00019999389463530383,
"loss": 0.173,
"step": 326
},
{
"epoch": 0.31046760028483267,
"grad_norm": 0.05901939421892166,
"learning_rate": 0.00019999261252450396,
"loss": 0.1419,
"step": 327
},
{
"epoch": 0.3114170424875386,
"grad_norm": 0.055540215224027634,
"learning_rate": 0.00019999120831420083,
"loss": 0.1314,
"step": 328
},
{
"epoch": 0.3123664846902445,
"grad_norm": 0.0546739287674427,
"learning_rate": 0.00019998968200610903,
"loss": 0.1354,
"step": 329
},
{
"epoch": 0.3133159268929504,
"grad_norm": 0.0689477026462555,
"learning_rate": 0.00019998803360209234,
"loss": 0.132,
"step": 330
},
{
"epoch": 0.3142653690956563,
"grad_norm": 0.05279696360230446,
"learning_rate": 0.00019998626310416365,
"loss": 0.1424,
"step": 331
},
{
"epoch": 0.3152148112983622,
"grad_norm": 0.055384278297424316,
"learning_rate": 0.00019998437051448482,
"loss": 0.141,
"step": 332
},
{
"epoch": 0.31616425350106814,
"grad_norm": 0.04636182263493538,
"learning_rate": 0.0001999823558353669,
"loss": 0.1414,
"step": 333
},
{
"epoch": 0.317113695703774,
"grad_norm": 0.04795726016163826,
"learning_rate": 0.00019998021906926993,
"loss": 0.1255,
"step": 334
},
{
"epoch": 0.31806313790647994,
"grad_norm": 0.05326540395617485,
"learning_rate": 0.00019997796021880318,
"loss": 0.1309,
"step": 335
},
{
"epoch": 0.3190125801091859,
"grad_norm": 0.0684736892580986,
"learning_rate": 0.00019997557928672484,
"loss": 0.1825,
"step": 336
},
{
"epoch": 0.31996202231189175,
"grad_norm": 0.042282164096832275,
"learning_rate": 0.0001999730762759422,
"loss": 0.12,
"step": 337
},
{
"epoch": 0.3209114645145977,
"grad_norm": 0.05297423154115677,
"learning_rate": 0.00019997045118951175,
"loss": 0.1309,
"step": 338
},
{
"epoch": 0.3218609067173036,
"grad_norm": 0.080621138215065,
"learning_rate": 0.00019996770403063883,
"loss": 0.2134,
"step": 339
},
{
"epoch": 0.3228103489200095,
"grad_norm": 0.05552308261394501,
"learning_rate": 0.00019996483480267803,
"loss": 0.1361,
"step": 340
},
{
"epoch": 0.3237597911227154,
"grad_norm": 0.05070111155509949,
"learning_rate": 0.00019996184350913287,
"loss": 0.1314,
"step": 341
},
{
"epoch": 0.32470923332542134,
"grad_norm": 0.04412266984581947,
"learning_rate": 0.00019995873015365601,
"loss": 0.1299,
"step": 342
},
{
"epoch": 0.3256586755281272,
"grad_norm": 0.0445338599383831,
"learning_rate": 0.00019995549474004917,
"loss": 0.1313,
"step": 343
},
{
"epoch": 0.32660811773083315,
"grad_norm": 0.08224980533123016,
"learning_rate": 0.000199952137272263,
"loss": 0.1844,
"step": 344
},
{
"epoch": 0.327557559933539,
"grad_norm": 0.04331446811556816,
"learning_rate": 0.0001999486577543972,
"loss": 0.133,
"step": 345
},
{
"epoch": 0.32850700213624495,
"grad_norm": 0.049314577132463455,
"learning_rate": 0.00019994505619070068,
"loss": 0.1351,
"step": 346
},
{
"epoch": 0.3294564443389509,
"grad_norm": 0.0697011798620224,
"learning_rate": 0.00019994133258557117,
"loss": 0.1709,
"step": 347
},
{
"epoch": 0.33040588654165676,
"grad_norm": 0.0510990135371685,
"learning_rate": 0.00019993748694355557,
"loss": 0.1365,
"step": 348
},
{
"epoch": 0.3313553287443627,
"grad_norm": 0.05100785568356514,
"learning_rate": 0.00019993351926934967,
"loss": 0.1302,
"step": 349
},
{
"epoch": 0.3323047709470686,
"grad_norm": 0.08001980185508728,
"learning_rate": 0.00019992942956779838,
"loss": 0.1736,
"step": 350
},
{
"epoch": 0.3332542131497745,
"grad_norm": 0.05298507958650589,
"learning_rate": 0.00019992521784389559,
"loss": 0.159,
"step": 351
},
{
"epoch": 0.3342036553524804,
"grad_norm": 0.04655485600233078,
"learning_rate": 0.00019992088410278414,
"loss": 0.1401,
"step": 352
},
{
"epoch": 0.33515309755518635,
"grad_norm": 0.047509439289569855,
"learning_rate": 0.00019991642834975594,
"loss": 0.1369,
"step": 353
},
{
"epoch": 0.3361025397578922,
"grad_norm": 0.046006906777620316,
"learning_rate": 0.0001999118505902518,
"loss": 0.1384,
"step": 354
},
{
"epoch": 0.33705198196059816,
"grad_norm": 0.07522892951965332,
"learning_rate": 0.00019990715082986155,
"loss": 0.2254,
"step": 355
},
{
"epoch": 0.33800142416330403,
"grad_norm": 0.048646144568920135,
"learning_rate": 0.00019990232907432404,
"loss": 0.1355,
"step": 356
},
{
"epoch": 0.33895086636600996,
"grad_norm": 0.03941798582673073,
"learning_rate": 0.000199897385329527,
"loss": 0.1242,
"step": 357
},
{
"epoch": 0.3399003085687159,
"grad_norm": 0.04582727700471878,
"learning_rate": 0.0001998923196015072,
"loss": 0.1347,
"step": 358
},
{
"epoch": 0.34084975077142177,
"grad_norm": 0.05890033766627312,
"learning_rate": 0.00019988713189645027,
"loss": 0.1356,
"step": 359
},
{
"epoch": 0.3417991929741277,
"grad_norm": 0.050398606806993484,
"learning_rate": 0.00019988182222069093,
"loss": 0.1379,
"step": 360
},
{
"epoch": 0.3427486351768336,
"grad_norm": 0.053657352924346924,
"learning_rate": 0.00019987639058071267,
"loss": 0.1417,
"step": 361
},
{
"epoch": 0.3436980773795395,
"grad_norm": 0.04928993433713913,
"learning_rate": 0.00019987083698314804,
"loss": 0.1269,
"step": 362
},
{
"epoch": 0.34464751958224543,
"grad_norm": 0.04932550713419914,
"learning_rate": 0.0001998651614347784,
"loss": 0.1429,
"step": 363
},
{
"epoch": 0.34559696178495136,
"grad_norm": 0.0531768873333931,
"learning_rate": 0.00019985936394253413,
"loss": 0.1367,
"step": 364
},
{
"epoch": 0.34654640398765724,
"grad_norm": 0.05342009291052818,
"learning_rate": 0.00019985344451349443,
"loss": 0.1365,
"step": 365
},
{
"epoch": 0.34749584619036317,
"grad_norm": 0.04960772022604942,
"learning_rate": 0.00019984740315488742,
"loss": 0.133,
"step": 366
},
{
"epoch": 0.3484452883930691,
"grad_norm": 0.04490765556693077,
"learning_rate": 0.00019984123987409013,
"loss": 0.1347,
"step": 367
},
{
"epoch": 0.34939473059577497,
"grad_norm": 0.05546121671795845,
"learning_rate": 0.0001998349546786285,
"loss": 0.169,
"step": 368
},
{
"epoch": 0.3503441727984809,
"grad_norm": 0.04962169751524925,
"learning_rate": 0.0001998285475761772,
"loss": 0.1325,
"step": 369
},
{
"epoch": 0.3512936150011868,
"grad_norm": 0.0451858825981617,
"learning_rate": 0.00019982201857455988,
"loss": 0.1291,
"step": 370
},
{
"epoch": 0.3522430572038927,
"grad_norm": 0.07738906145095825,
"learning_rate": 0.00019981536768174903,
"loss": 0.1841,
"step": 371
},
{
"epoch": 0.35319249940659864,
"grad_norm": 0.05104148015379906,
"learning_rate": 0.000199808594905866,
"loss": 0.1375,
"step": 372
},
{
"epoch": 0.3541419416093045,
"grad_norm": 0.04850155860185623,
"learning_rate": 0.00019980170025518082,
"loss": 0.1335,
"step": 373
},
{
"epoch": 0.35509138381201044,
"grad_norm": 0.050271324813365936,
"learning_rate": 0.00019979468373811248,
"loss": 0.1394,
"step": 374
},
{
"epoch": 0.35604082601471637,
"grad_norm": 0.050799645483493805,
"learning_rate": 0.0001997875453632288,
"loss": 0.135,
"step": 375
},
{
"epoch": 0.35699026821742225,
"grad_norm": 0.05703526735305786,
"learning_rate": 0.00019978028513924627,
"loss": 0.1371,
"step": 376
},
{
"epoch": 0.3579397104201282,
"grad_norm": 0.06665853410959244,
"learning_rate": 0.00019977290307503028,
"loss": 0.1837,
"step": 377
},
{
"epoch": 0.3588891526228341,
"grad_norm": 0.04639972746372223,
"learning_rate": 0.000199765399179595,
"loss": 0.1315,
"step": 378
},
{
"epoch": 0.35983859482554,
"grad_norm": 0.07625308632850647,
"learning_rate": 0.00019975777346210326,
"loss": 0.2064,
"step": 379
},
{
"epoch": 0.3607880370282459,
"grad_norm": 0.048770248889923096,
"learning_rate": 0.00019975002593186674,
"loss": 0.1363,
"step": 380
},
{
"epoch": 0.36173747923095184,
"grad_norm": 0.04932136833667755,
"learning_rate": 0.00019974215659834582,
"loss": 0.1374,
"step": 381
},
{
"epoch": 0.3626869214336577,
"grad_norm": 0.03848756104707718,
"learning_rate": 0.00019973416547114964,
"loss": 0.1333,
"step": 382
},
{
"epoch": 0.36363636363636365,
"grad_norm": 0.04468891769647598,
"learning_rate": 0.00019972605256003605,
"loss": 0.129,
"step": 383
},
{
"epoch": 0.3645858058390695,
"grad_norm": 0.048413511365652084,
"learning_rate": 0.0001997178178749116,
"loss": 0.1314,
"step": 384
},
{
"epoch": 0.36553524804177545,
"grad_norm": 0.045054856687784195,
"learning_rate": 0.00019970946142583155,
"loss": 0.1323,
"step": 385
},
{
"epoch": 0.3664846902444814,
"grad_norm": 0.05541200935840607,
"learning_rate": 0.00019970098322299982,
"loss": 0.1342,
"step": 386
},
{
"epoch": 0.36743413244718726,
"grad_norm": 0.06861472874879837,
"learning_rate": 0.00019969238327676906,
"loss": 0.1347,
"step": 387
},
{
"epoch": 0.3683835746498932,
"grad_norm": 0.043996453285217285,
"learning_rate": 0.00019968366159764047,
"loss": 0.132,
"step": 388
},
{
"epoch": 0.3693330168525991,
"grad_norm": 0.06562239676713943,
"learning_rate": 0.000199674818196264,
"loss": 0.1759,
"step": 389
},
{
"epoch": 0.370282459055305,
"grad_norm": 0.04714899882674217,
"learning_rate": 0.00019966585308343822,
"loss": 0.1274,
"step": 390
},
{
"epoch": 0.3712319012580109,
"grad_norm": 0.04736959934234619,
"learning_rate": 0.00019965676627011026,
"loss": 0.1265,
"step": 391
},
{
"epoch": 0.37218134346071685,
"grad_norm": 0.056829433888196945,
"learning_rate": 0.0001996475577673759,
"loss": 0.1402,
"step": 392
},
{
"epoch": 0.3731307856634227,
"grad_norm": 0.0426231250166893,
"learning_rate": 0.00019963822758647953,
"loss": 0.1364,
"step": 393
},
{
"epoch": 0.37408022786612866,
"grad_norm": 0.07376877963542938,
"learning_rate": 0.00019962877573881404,
"loss": 0.2042,
"step": 394
},
{
"epoch": 0.3750296700688346,
"grad_norm": 0.043273668736219406,
"learning_rate": 0.00019961920223592104,
"loss": 0.132,
"step": 395
},
{
"epoch": 0.37597911227154046,
"grad_norm": 0.044406965374946594,
"learning_rate": 0.00019960950708949052,
"loss": 0.1344,
"step": 396
},
{
"epoch": 0.3769285544742464,
"grad_norm": 0.040342606604099274,
"learning_rate": 0.00019959969031136106,
"loss": 0.1214,
"step": 397
},
{
"epoch": 0.37787799667695227,
"grad_norm": 0.05118388682603836,
"learning_rate": 0.00019958975191351983,
"loss": 0.14,
"step": 398
},
{
"epoch": 0.3788274388796582,
"grad_norm": 0.045876793563365936,
"learning_rate": 0.00019957969190810245,
"loss": 0.1335,
"step": 399
},
{
"epoch": 0.3797768810823641,
"grad_norm": 0.0645332932472229,
"learning_rate": 0.00019956951030739308,
"loss": 0.1702,
"step": 400
},
{
"epoch": 0.38072632328507,
"grad_norm": 0.05039132386445999,
"learning_rate": 0.00019955920712382423,
"loss": 0.136,
"step": 401
},
{
"epoch": 0.38167576548777593,
"grad_norm": 0.052004653960466385,
"learning_rate": 0.00019954878236997704,
"loss": 0.1386,
"step": 402
},
{
"epoch": 0.38262520769048186,
"grad_norm": 0.05021458491683006,
"learning_rate": 0.00019953823605858105,
"loss": 0.1378,
"step": 403
},
{
"epoch": 0.38357464989318774,
"grad_norm": 0.058653559535741806,
"learning_rate": 0.0001995275682025141,
"loss": 0.1437,
"step": 404
},
{
"epoch": 0.38452409209589367,
"grad_norm": 0.04466673359274864,
"learning_rate": 0.00019951677881480264,
"loss": 0.1334,
"step": 405
},
{
"epoch": 0.3854735342985996,
"grad_norm": 0.06119415909051895,
"learning_rate": 0.00019950586790862138,
"loss": 0.1296,
"step": 406
},
{
"epoch": 0.38642297650130547,
"grad_norm": 0.04749077931046486,
"learning_rate": 0.0001994948354972935,
"loss": 0.1341,
"step": 407
},
{
"epoch": 0.3873724187040114,
"grad_norm": 0.037752799689769745,
"learning_rate": 0.00019948368159429053,
"loss": 0.134,
"step": 408
},
{
"epoch": 0.38832186090671733,
"grad_norm": 0.08903038501739502,
"learning_rate": 0.00019947240621323226,
"loss": 0.2155,
"step": 409
},
{
"epoch": 0.3892713031094232,
"grad_norm": 0.03878140076994896,
"learning_rate": 0.00019946100936788698,
"loss": 0.1176,
"step": 410
},
{
"epoch": 0.39022074531212914,
"grad_norm": 0.04927309602499008,
"learning_rate": 0.00019944949107217113,
"loss": 0.1344,
"step": 411
},
{
"epoch": 0.391170187514835,
"grad_norm": 0.04933890327811241,
"learning_rate": 0.00019943785134014962,
"loss": 0.1315,
"step": 412
},
{
"epoch": 0.39211962971754094,
"grad_norm": 0.06702516227960587,
"learning_rate": 0.0001994260901860355,
"loss": 0.1826,
"step": 413
},
{
"epoch": 0.39306907192024687,
"grad_norm": 0.048132237046957016,
"learning_rate": 0.00019941420762419014,
"loss": 0.1436,
"step": 414
},
{
"epoch": 0.39401851412295275,
"grad_norm": 0.07756894826889038,
"learning_rate": 0.00019940220366912318,
"loss": 0.2162,
"step": 415
},
{
"epoch": 0.3949679563256587,
"grad_norm": 0.04789011925458908,
"learning_rate": 0.00019939007833549242,
"loss": 0.1295,
"step": 416
},
{
"epoch": 0.3959173985283646,
"grad_norm": 0.04369444027543068,
"learning_rate": 0.000199377831638104,
"loss": 0.1322,
"step": 417
},
{
"epoch": 0.3968668407310705,
"grad_norm": 0.05376122146844864,
"learning_rate": 0.00019936546359191216,
"loss": 0.1743,
"step": 418
},
{
"epoch": 0.3978162829337764,
"grad_norm": 0.045930229127407074,
"learning_rate": 0.0001993529742120193,
"loss": 0.1336,
"step": 419
},
{
"epoch": 0.39876572513648234,
"grad_norm": 0.039980966597795486,
"learning_rate": 0.00019934036351367606,
"loss": 0.1349,
"step": 420
},
{
"epoch": 0.3997151673391882,
"grad_norm": 0.03797341510653496,
"learning_rate": 0.00019932763151228115,
"loss": 0.1256,
"step": 421
},
{
"epoch": 0.40066460954189415,
"grad_norm": 0.04779914394021034,
"learning_rate": 0.00019931477822338146,
"loss": 0.1411,
"step": 422
},
{
"epoch": 0.4016140517446,
"grad_norm": 0.040458668023347855,
"learning_rate": 0.00019930180366267193,
"loss": 0.126,
"step": 423
},
{
"epoch": 0.40256349394730595,
"grad_norm": 0.04114462807774544,
"learning_rate": 0.0001992887078459956,
"loss": 0.127,
"step": 424
},
{
"epoch": 0.4035129361500119,
"grad_norm": 0.048119012266397476,
"learning_rate": 0.00019927549078934358,
"loss": 0.1346,
"step": 425
},
{
"epoch": 0.40446237835271776,
"grad_norm": 0.0545562319457531,
"learning_rate": 0.00019926215250885504,
"loss": 0.1387,
"step": 426
},
{
"epoch": 0.4054118205554237,
"grad_norm": 0.052092909812927246,
"learning_rate": 0.00019924869302081715,
"loss": 0.1389,
"step": 427
},
{
"epoch": 0.4063612627581296,
"grad_norm": 0.03847799077630043,
"learning_rate": 0.0001992351123416651,
"loss": 0.1234,
"step": 428
},
{
"epoch": 0.4073107049608355,
"grad_norm": 0.0436912477016449,
"learning_rate": 0.000199221410487982,
"loss": 0.1362,
"step": 429
},
{
"epoch": 0.4082601471635414,
"grad_norm": 0.04420888423919678,
"learning_rate": 0.00019920758747649908,
"loss": 0.1243,
"step": 430
},
{
"epoch": 0.40920958936624735,
"grad_norm": 0.037297070026397705,
"learning_rate": 0.00019919364332409535,
"loss": 0.1331,
"step": 431
},
{
"epoch": 0.4101590315689532,
"grad_norm": 0.03854360058903694,
"learning_rate": 0.00019917957804779782,
"loss": 0.1266,
"step": 432
},
{
"epoch": 0.41110847377165916,
"grad_norm": 0.04071418195962906,
"learning_rate": 0.00019916539166478137,
"loss": 0.1292,
"step": 433
},
{
"epoch": 0.4120579159743651,
"grad_norm": 0.04560808837413788,
"learning_rate": 0.00019915108419236882,
"loss": 0.1381,
"step": 434
},
{
"epoch": 0.41300735817707096,
"grad_norm": 0.06313233822584152,
"learning_rate": 0.00019913665564803078,
"loss": 0.2031,
"step": 435
},
{
"epoch": 0.4139568003797769,
"grad_norm": 0.04507524147629738,
"learning_rate": 0.00019912210604938578,
"loss": 0.1277,
"step": 436
},
{
"epoch": 0.41490624258248276,
"grad_norm": 0.05048058554530144,
"learning_rate": 0.00019910743541420007,
"loss": 0.1315,
"step": 437
},
{
"epoch": 0.4158556847851887,
"grad_norm": 0.04872648045420647,
"learning_rate": 0.0001990926437603878,
"loss": 0.1292,
"step": 438
},
{
"epoch": 0.4168051269878946,
"grad_norm": 0.04400710016489029,
"learning_rate": 0.00019907773110601075,
"loss": 0.1236,
"step": 439
},
{
"epoch": 0.4177545691906005,
"grad_norm": 0.051591627299785614,
"learning_rate": 0.00019906269746927863,
"loss": 0.1358,
"step": 440
},
{
"epoch": 0.41870401139330643,
"grad_norm": 0.04288725182414055,
"learning_rate": 0.00019904754286854877,
"loss": 0.126,
"step": 441
},
{
"epoch": 0.41965345359601236,
"grad_norm": 0.04984726384282112,
"learning_rate": 0.00019903226732232622,
"loss": 0.1326,
"step": 442
},
{
"epoch": 0.42060289579871823,
"grad_norm": 0.041585132479667664,
"learning_rate": 0.00019901687084926373,
"loss": 0.136,
"step": 443
},
{
"epoch": 0.42155233800142417,
"grad_norm": 0.05849035084247589,
"learning_rate": 0.0001990013534681617,
"loss": 0.1727,
"step": 444
},
{
"epoch": 0.4225017802041301,
"grad_norm": 0.043387994170188904,
"learning_rate": 0.00019898571519796817,
"loss": 0.1393,
"step": 445
},
{
"epoch": 0.42345122240683597,
"grad_norm": 0.05867496132850647,
"learning_rate": 0.0001989699560577788,
"loss": 0.1664,
"step": 446
},
{
"epoch": 0.4244006646095419,
"grad_norm": 0.07019232958555222,
"learning_rate": 0.00019895407606683685,
"loss": 0.1653,
"step": 447
},
{
"epoch": 0.42535010681224783,
"grad_norm": 0.04676515609025955,
"learning_rate": 0.00019893807524453314,
"loss": 0.1368,
"step": 448
},
{
"epoch": 0.4262995490149537,
"grad_norm": 0.06640240550041199,
"learning_rate": 0.00019892195361040607,
"loss": 0.2089,
"step": 449
},
{
"epoch": 0.42724899121765963,
"grad_norm": 0.044658735394477844,
"learning_rate": 0.00019890571118414148,
"loss": 0.1298,
"step": 450
},
{
"epoch": 0.4281984334203655,
"grad_norm": 0.04810122773051262,
"learning_rate": 0.00019888934798557278,
"loss": 0.1288,
"step": 451
},
{
"epoch": 0.42914787562307144,
"grad_norm": 0.0425436794757843,
"learning_rate": 0.0001988728640346808,
"loss": 0.1354,
"step": 452
},
{
"epoch": 0.43009731782577737,
"grad_norm": 0.04513363912701607,
"learning_rate": 0.0001988562593515939,
"loss": 0.1346,
"step": 453
},
{
"epoch": 0.43104676002848324,
"grad_norm": 0.052022870630025864,
"learning_rate": 0.0001988395339565878,
"loss": 0.1302,
"step": 454
},
{
"epoch": 0.4319962022311892,
"grad_norm": 0.04852641373872757,
"learning_rate": 0.0001988226878700856,
"loss": 0.1388,
"step": 455
},
{
"epoch": 0.4329456444338951,
"grad_norm": 0.04990584775805473,
"learning_rate": 0.00019880572111265785,
"loss": 0.1552,
"step": 456
},
{
"epoch": 0.433895086636601,
"grad_norm": 0.052271679043769836,
"learning_rate": 0.00019878863370502238,
"loss": 0.1404,
"step": 457
},
{
"epoch": 0.4348445288393069,
"grad_norm": 0.04795520752668381,
"learning_rate": 0.00019877142566804436,
"loss": 0.1341,
"step": 458
},
{
"epoch": 0.43579397104201284,
"grad_norm": 0.048165664076805115,
"learning_rate": 0.00019875409702273632,
"loss": 0.1343,
"step": 459
},
{
"epoch": 0.4367434132447187,
"grad_norm": 0.04213611036539078,
"learning_rate": 0.000198736647790258,
"loss": 0.1369,
"step": 460
},
{
"epoch": 0.43769285544742464,
"grad_norm": 0.05819966271519661,
"learning_rate": 0.00019871907799191632,
"loss": 0.1615,
"step": 461
},
{
"epoch": 0.4386422976501306,
"grad_norm": 0.057378821074962616,
"learning_rate": 0.00019870138764916558,
"loss": 0.175,
"step": 462
},
{
"epoch": 0.43959173985283645,
"grad_norm": 0.0432853177189827,
"learning_rate": 0.00019868357678360724,
"loss": 0.1371,
"step": 463
},
{
"epoch": 0.4405411820555424,
"grad_norm": 0.03890872746706009,
"learning_rate": 0.0001986656454169898,
"loss": 0.1332,
"step": 464
},
{
"epoch": 0.44149062425824825,
"grad_norm": 0.04006613418459892,
"learning_rate": 0.00019864759357120896,
"loss": 0.1342,
"step": 465
},
{
"epoch": 0.4424400664609542,
"grad_norm": 0.049053166061639786,
"learning_rate": 0.00019862942126830767,
"loss": 0.1756,
"step": 466
},
{
"epoch": 0.4433895086636601,
"grad_norm": 0.03966079652309418,
"learning_rate": 0.00019861112853047577,
"loss": 0.1303,
"step": 467
},
{
"epoch": 0.444338950866366,
"grad_norm": 0.04506433755159378,
"learning_rate": 0.0001985927153800503,
"loss": 0.136,
"step": 468
},
{
"epoch": 0.4452883930690719,
"grad_norm": 0.04392915591597557,
"learning_rate": 0.00019857418183951526,
"loss": 0.1397,
"step": 469
},
{
"epoch": 0.44623783527177785,
"grad_norm": 0.038007620722055435,
"learning_rate": 0.0001985555279315017,
"loss": 0.1246,
"step": 470
},
{
"epoch": 0.4471872774744837,
"grad_norm": 0.048948097974061966,
"learning_rate": 0.00019853675367878764,
"loss": 0.1329,
"step": 471
},
{
"epoch": 0.44813671967718965,
"grad_norm": 0.04174380376935005,
"learning_rate": 0.00019851785910429806,
"loss": 0.13,
"step": 472
},
{
"epoch": 0.4490861618798956,
"grad_norm": 0.048575468361377716,
"learning_rate": 0.00019849884423110478,
"loss": 0.1385,
"step": 473
},
{
"epoch": 0.45003560408260146,
"grad_norm": 0.05167670175433159,
"learning_rate": 0.00019847970908242664,
"loss": 0.1684,
"step": 474
},
{
"epoch": 0.4509850462853074,
"grad_norm": 0.06849198788404465,
"learning_rate": 0.00019846045368162923,
"loss": 0.1795,
"step": 475
},
{
"epoch": 0.4519344884880133,
"grad_norm": 0.044273603707551956,
"learning_rate": 0.0001984410780522251,
"loss": 0.1246,
"step": 476
},
{
"epoch": 0.4528839306907192,
"grad_norm": 0.048194363713264465,
"learning_rate": 0.00019842158221787353,
"loss": 0.1366,
"step": 477
},
{
"epoch": 0.4538333728934251,
"grad_norm": 0.033906418830156326,
"learning_rate": 0.00019840196620238057,
"loss": 0.1235,
"step": 478
},
{
"epoch": 0.454782815096131,
"grad_norm": 0.043933141976594925,
"learning_rate": 0.00019838223002969905,
"loss": 0.1195,
"step": 479
},
{
"epoch": 0.45573225729883693,
"grad_norm": 0.056823644787073135,
"learning_rate": 0.00019836237372392854,
"loss": 0.1757,
"step": 480
},
{
"epoch": 0.45668169950154286,
"grad_norm": 0.07587820291519165,
"learning_rate": 0.00019834239730931526,
"loss": 0.1784,
"step": 481
},
{
"epoch": 0.45763114170424873,
"grad_norm": 0.04008018970489502,
"learning_rate": 0.0001983223008102521,
"loss": 0.1306,
"step": 482
},
{
"epoch": 0.45858058390695466,
"grad_norm": 0.05180038511753082,
"learning_rate": 0.00019830208425127867,
"loss": 0.1485,
"step": 483
},
{
"epoch": 0.4595300261096606,
"grad_norm": 0.0691617876291275,
"learning_rate": 0.00019828174765708104,
"loss": 0.1249,
"step": 484
},
{
"epoch": 0.46047946831236647,
"grad_norm": 0.0565367266535759,
"learning_rate": 0.00019826129105249195,
"loss": 0.1744,
"step": 485
},
{
"epoch": 0.4614289105150724,
"grad_norm": 0.044927019625902176,
"learning_rate": 0.00019824071446249072,
"loss": 0.1341,
"step": 486
},
{
"epoch": 0.46237835271777833,
"grad_norm": 0.04481721669435501,
"learning_rate": 0.00019822001791220298,
"loss": 0.1354,
"step": 487
},
{
"epoch": 0.4633277949204842,
"grad_norm": 0.05233500525355339,
"learning_rate": 0.0001981992014269011,
"loss": 0.1501,
"step": 488
},
{
"epoch": 0.46427723712319013,
"grad_norm": 0.044350553303956985,
"learning_rate": 0.00019817826503200372,
"loss": 0.1335,
"step": 489
},
{
"epoch": 0.465226679325896,
"grad_norm": 0.03551819548010826,
"learning_rate": 0.000198157208753076,
"loss": 0.1322,
"step": 490
},
{
"epoch": 0.46617612152860194,
"grad_norm": 0.04409592226147652,
"learning_rate": 0.00019813603261582943,
"loss": 0.1561,
"step": 491
},
{
"epoch": 0.46712556373130787,
"grad_norm": 0.04842127487063408,
"learning_rate": 0.0001981147366461219,
"loss": 0.1296,
"step": 492
},
{
"epoch": 0.46807500593401374,
"grad_norm": 0.04349881038069725,
"learning_rate": 0.00019809332086995757,
"loss": 0.1319,
"step": 493
},
{
"epoch": 0.4690244481367197,
"grad_norm": 0.04413028433918953,
"learning_rate": 0.00019807178531348698,
"loss": 0.1321,
"step": 494
},
{
"epoch": 0.4699738903394256,
"grad_norm": 0.03972313553094864,
"learning_rate": 0.00019805013000300683,
"loss": 0.1358,
"step": 495
},
{
"epoch": 0.4709233325421315,
"grad_norm": 0.052269116044044495,
"learning_rate": 0.00019802835496496012,
"loss": 0.1389,
"step": 496
},
{
"epoch": 0.4718727747448374,
"grad_norm": 0.0379653237760067,
"learning_rate": 0.00019800646022593603,
"loss": 0.1283,
"step": 497
},
{
"epoch": 0.47282221694754334,
"grad_norm": 0.04370688647031784,
"learning_rate": 0.0001979844458126699,
"loss": 0.1278,
"step": 498
},
{
"epoch": 0.4737716591502492,
"grad_norm": 0.03912369906902313,
"learning_rate": 0.0001979623117520432,
"loss": 0.1257,
"step": 499
},
{
"epoch": 0.47472110135295514,
"grad_norm": 0.039594005793333054,
"learning_rate": 0.00019794005807108352,
"loss": 0.1375,
"step": 500
},
{
"epoch": 0.4756705435556611,
"grad_norm": 0.03889892250299454,
"learning_rate": 0.00019791768479696448,
"loss": 0.13,
"step": 501
},
{
"epoch": 0.47661998575836695,
"grad_norm": 0.03966660797595978,
"learning_rate": 0.00019789519195700578,
"loss": 0.1268,
"step": 502
},
{
"epoch": 0.4775694279610729,
"grad_norm": 0.04501716047525406,
"learning_rate": 0.00019787257957867306,
"loss": 0.1423,
"step": 503
},
{
"epoch": 0.47851887016377875,
"grad_norm": 0.06255436688661575,
"learning_rate": 0.000197849847689578,
"loss": 0.1799,
"step": 504
},
{
"epoch": 0.4794683123664847,
"grad_norm": 0.050308458507061005,
"learning_rate": 0.00019782699631747813,
"loss": 0.1733,
"step": 505
},
{
"epoch": 0.4804177545691906,
"grad_norm": 0.0357963964343071,
"learning_rate": 0.00019780402549027698,
"loss": 0.1268,
"step": 506
},
{
"epoch": 0.4813671967718965,
"grad_norm": 0.03651968017220497,
"learning_rate": 0.00019778093523602384,
"loss": 0.1267,
"step": 507
},
{
"epoch": 0.4823166389746024,
"grad_norm": 0.043042074888944626,
"learning_rate": 0.0001977577255829139,
"loss": 0.1256,
"step": 508
},
{
"epoch": 0.48326608117730835,
"grad_norm": 0.07031014561653137,
"learning_rate": 0.00019773439655928815,
"loss": 0.1796,
"step": 509
},
{
"epoch": 0.4842155233800142,
"grad_norm": 0.04429268836975098,
"learning_rate": 0.00019771094819363326,
"loss": 0.1298,
"step": 510
},
{
"epoch": 0.48516496558272015,
"grad_norm": 0.0373898483812809,
"learning_rate": 0.00019768738051458172,
"loss": 0.1232,
"step": 511
},
{
"epoch": 0.4861144077854261,
"grad_norm": 0.05853155627846718,
"learning_rate": 0.00019766369355091166,
"loss": 0.1694,
"step": 512
},
{
"epoch": 0.48706384998813196,
"grad_norm": 0.05050895735621452,
"learning_rate": 0.00019763988733154686,
"loss": 0.1665,
"step": 513
},
{
"epoch": 0.4880132921908379,
"grad_norm": 0.04074448347091675,
"learning_rate": 0.0001976159618855568,
"loss": 0.1336,
"step": 514
},
{
"epoch": 0.4889627343935438,
"grad_norm": 0.03826110064983368,
"learning_rate": 0.00019759191724215644,
"loss": 0.132,
"step": 515
},
{
"epoch": 0.4899121765962497,
"grad_norm": 0.04392875358462334,
"learning_rate": 0.0001975677534307064,
"loss": 0.1204,
"step": 516
},
{
"epoch": 0.4908616187989556,
"grad_norm": 0.04615531116724014,
"learning_rate": 0.0001975434704807127,
"loss": 0.1358,
"step": 517
},
{
"epoch": 0.4918110610016615,
"grad_norm": 0.053060565143823624,
"learning_rate": 0.00019751906842182688,
"loss": 0.1299,
"step": 518
},
{
"epoch": 0.49276050320436743,
"grad_norm": 0.04905511438846588,
"learning_rate": 0.00019749454728384594,
"loss": 0.1284,
"step": 519
},
{
"epoch": 0.49370994540707336,
"grad_norm": 0.04257996007800102,
"learning_rate": 0.00019746990709671234,
"loss": 0.1353,
"step": 520
},
{
"epoch": 0.49465938760977923,
"grad_norm": 0.05581909418106079,
"learning_rate": 0.0001974451478905138,
"loss": 0.1594,
"step": 521
},
{
"epoch": 0.49560882981248516,
"grad_norm": 0.04603990167379379,
"learning_rate": 0.00019742026969548338,
"loss": 0.1383,
"step": 522
},
{
"epoch": 0.4965582720151911,
"grad_norm": 0.058511972427368164,
"learning_rate": 0.00019739527254199958,
"loss": 0.1725,
"step": 523
},
{
"epoch": 0.49750771421789697,
"grad_norm": 0.03875808045268059,
"learning_rate": 0.000197370156460586,
"loss": 0.1405,
"step": 524
},
{
"epoch": 0.4984571564206029,
"grad_norm": 0.040860000997781754,
"learning_rate": 0.00019734492148191151,
"loss": 0.139,
"step": 525
},
{
"epoch": 0.49940659862330883,
"grad_norm": 0.06110459193587303,
"learning_rate": 0.00019731956763679014,
"loss": 0.223,
"step": 526
},
{
"epoch": 0.5003560408260147,
"grad_norm": 0.05238598585128784,
"learning_rate": 0.00019729409495618117,
"loss": 0.1681,
"step": 527
},
{
"epoch": 0.5013054830287206,
"grad_norm": 0.05180145800113678,
"learning_rate": 0.00019726850347118885,
"loss": 0.1743,
"step": 528
},
{
"epoch": 0.5022549252314266,
"grad_norm": 0.05066410079598427,
"learning_rate": 0.00019724279321306262,
"loss": 0.1634,
"step": 529
},
{
"epoch": 0.5032043674341324,
"grad_norm": 0.06856084614992142,
"learning_rate": 0.00019721696421319684,
"loss": 0.1685,
"step": 530
},
{
"epoch": 0.5041538096368383,
"grad_norm": 0.045972324907779694,
"learning_rate": 0.00019719101650313096,
"loss": 0.1245,
"step": 531
},
{
"epoch": 0.5051032518395443,
"grad_norm": 0.04522623121738434,
"learning_rate": 0.00019716495011454934,
"loss": 0.1367,
"step": 532
},
{
"epoch": 0.5060526940422502,
"grad_norm": 0.0780516117811203,
"learning_rate": 0.00019713876507928126,
"loss": 0.1351,
"step": 533
},
{
"epoch": 0.507002136244956,
"grad_norm": 0.04264210909605026,
"learning_rate": 0.00019711246142930088,
"loss": 0.1312,
"step": 534
},
{
"epoch": 0.507951578447662,
"grad_norm": 0.059501100331544876,
"learning_rate": 0.00019708603919672718,
"loss": 0.1698,
"step": 535
},
{
"epoch": 0.5089010206503679,
"grad_norm": 0.060105033218860626,
"learning_rate": 0.00019705949841382396,
"loss": 0.1303,
"step": 536
},
{
"epoch": 0.5098504628530738,
"grad_norm": 0.04733967408537865,
"learning_rate": 0.00019703283911299982,
"loss": 0.1245,
"step": 537
},
{
"epoch": 0.5107999050557798,
"grad_norm": 0.04254663735628128,
"learning_rate": 0.00019700606132680798,
"loss": 0.1343,
"step": 538
},
{
"epoch": 0.5117493472584856,
"grad_norm": 0.06302463263273239,
"learning_rate": 0.00019697916508794645,
"loss": 0.1831,
"step": 539
},
{
"epoch": 0.5126987894611915,
"grad_norm": 0.05301344394683838,
"learning_rate": 0.0001969521504292578,
"loss": 0.1316,
"step": 540
},
{
"epoch": 0.5136482316638975,
"grad_norm": 0.04151083528995514,
"learning_rate": 0.00019692501738372922,
"loss": 0.1335,
"step": 541
},
{
"epoch": 0.5145976738666034,
"grad_norm": 0.05647062510251999,
"learning_rate": 0.00019689776598449257,
"loss": 0.1688,
"step": 542
},
{
"epoch": 0.5155471160693093,
"grad_norm": 0.037060294300317764,
"learning_rate": 0.000196870396264824,
"loss": 0.1339,
"step": 543
},
{
"epoch": 0.5164965582720152,
"grad_norm": 0.04036247730255127,
"learning_rate": 0.0001968429082581443,
"loss": 0.1361,
"step": 544
},
{
"epoch": 0.5174460004747211,
"grad_norm": 0.040889665484428406,
"learning_rate": 0.00019681530199801875,
"loss": 0.1356,
"step": 545
},
{
"epoch": 0.518395442677427,
"grad_norm": 0.0538480207324028,
"learning_rate": 0.00019678757751815686,
"loss": 0.1689,
"step": 546
},
{
"epoch": 0.519344884880133,
"grad_norm": 0.04074794426560402,
"learning_rate": 0.0001967597348524126,
"loss": 0.1329,
"step": 547
},
{
"epoch": 0.5202943270828388,
"grad_norm": 0.03896891698241234,
"learning_rate": 0.00019673177403478428,
"loss": 0.1356,
"step": 548
},
{
"epoch": 0.5212437692855447,
"grad_norm": 0.04619259387254715,
"learning_rate": 0.00019670369509941442,
"loss": 0.163,
"step": 549
},
{
"epoch": 0.5221932114882507,
"grad_norm": 0.035968657582998276,
"learning_rate": 0.00019667549808058976,
"loss": 0.1242,
"step": 550
},
{
"epoch": 0.5231426536909566,
"grad_norm": 0.04564007744193077,
"learning_rate": 0.0001966471830127413,
"loss": 0.1364,
"step": 551
},
{
"epoch": 0.5240920958936625,
"grad_norm": 0.03991610184311867,
"learning_rate": 0.00019661874993044415,
"loss": 0.1312,
"step": 552
},
{
"epoch": 0.5250415380963683,
"grad_norm": 0.037240512669086456,
"learning_rate": 0.00019659019886841752,
"loss": 0.1279,
"step": 553
},
{
"epoch": 0.5259909802990743,
"grad_norm": 0.06598762422800064,
"learning_rate": 0.00019656152986152468,
"loss": 0.2165,
"step": 554
},
{
"epoch": 0.5269404225017802,
"grad_norm": 0.03867746889591217,
"learning_rate": 0.00019653274294477292,
"loss": 0.1233,
"step": 555
},
{
"epoch": 0.5278898647044861,
"grad_norm": 0.051915477961301804,
"learning_rate": 0.00019650383815331357,
"loss": 0.168,
"step": 556
},
{
"epoch": 0.528839306907192,
"grad_norm": 0.054896485060453415,
"learning_rate": 0.00019647481552244182,
"loss": 0.1678,
"step": 557
},
{
"epoch": 0.5297887491098979,
"grad_norm": 0.05439051240682602,
"learning_rate": 0.00019644567508759675,
"loss": 0.1607,
"step": 558
},
{
"epoch": 0.5307381913126038,
"grad_norm": 0.03601578250527382,
"learning_rate": 0.00019641641688436135,
"loss": 0.1271,
"step": 559
},
{
"epoch": 0.5316876335153098,
"grad_norm": 0.06025104597210884,
"learning_rate": 0.00019638704094846236,
"loss": 0.176,
"step": 560
},
{
"epoch": 0.5326370757180157,
"grad_norm": 0.04126368835568428,
"learning_rate": 0.00019635754731577032,
"loss": 0.1319,
"step": 561
},
{
"epoch": 0.5335865179207215,
"grad_norm": 0.05305393040180206,
"learning_rate": 0.00019632793602229943,
"loss": 0.1699,
"step": 562
},
{
"epoch": 0.5345359601234275,
"grad_norm": 0.03538331016898155,
"learning_rate": 0.00019629820710420764,
"loss": 0.124,
"step": 563
},
{
"epoch": 0.5354854023261334,
"grad_norm": 0.05861300975084305,
"learning_rate": 0.0001962683605977965,
"loss": 0.1688,
"step": 564
},
{
"epoch": 0.5364348445288393,
"grad_norm": 0.040226079523563385,
"learning_rate": 0.0001962383965395111,
"loss": 0.1334,
"step": 565
},
{
"epoch": 0.5373842867315453,
"grad_norm": 0.035788875073194504,
"learning_rate": 0.00019620831496594017,
"loss": 0.1281,
"step": 566
},
{
"epoch": 0.5383337289342511,
"grad_norm": 0.0334162712097168,
"learning_rate": 0.0001961781159138158,
"loss": 0.1317,
"step": 567
},
{
"epoch": 0.539283171136957,
"grad_norm": 0.03352081775665283,
"learning_rate": 0.00019614779942001364,
"loss": 0.1334,
"step": 568
},
{
"epoch": 0.540232613339663,
"grad_norm": 0.03684060648083687,
"learning_rate": 0.00019611736552155274,
"loss": 0.1349,
"step": 569
},
{
"epoch": 0.5411820555423689,
"grad_norm": 0.03640671446919441,
"learning_rate": 0.00019608681425559542,
"loss": 0.1278,
"step": 570
},
{
"epoch": 0.5421314977450747,
"grad_norm": 0.04167250171303749,
"learning_rate": 0.00019605614565944748,
"loss": 0.1384,
"step": 571
},
{
"epoch": 0.5430809399477807,
"grad_norm": 0.0416824147105217,
"learning_rate": 0.00019602535977055778,
"loss": 0.1319,
"step": 572
},
{
"epoch": 0.5440303821504866,
"grad_norm": 0.03897137567400932,
"learning_rate": 0.00019599445662651861,
"loss": 0.1389,
"step": 573
},
{
"epoch": 0.5449798243531925,
"grad_norm": 0.03894896060228348,
"learning_rate": 0.00019596343626506526,
"loss": 0.1341,
"step": 574
},
{
"epoch": 0.5459292665558985,
"grad_norm": 0.04211690276861191,
"learning_rate": 0.00019593229872407627,
"loss": 0.1377,
"step": 575
},
{
"epoch": 0.5468787087586043,
"grad_norm": 0.04308454692363739,
"learning_rate": 0.00019590104404157327,
"loss": 0.1268,
"step": 576
},
{
"epoch": 0.5478281509613102,
"grad_norm": 0.0525001622736454,
"learning_rate": 0.00019586967225572086,
"loss": 0.1775,
"step": 577
},
{
"epoch": 0.5487775931640161,
"grad_norm": 0.056315965950489044,
"learning_rate": 0.00019583818340482664,
"loss": 0.1688,
"step": 578
},
{
"epoch": 0.5497270353667221,
"grad_norm": 0.03801283985376358,
"learning_rate": 0.0001958065775273412,
"loss": 0.1309,
"step": 579
},
{
"epoch": 0.550676477569428,
"grad_norm": 0.03738854080438614,
"learning_rate": 0.00019577485466185804,
"loss": 0.137,
"step": 580
},
{
"epoch": 0.5516259197721338,
"grad_norm": 0.03772661089897156,
"learning_rate": 0.0001957430148471134,
"loss": 0.1276,
"step": 581
},
{
"epoch": 0.5525753619748398,
"grad_norm": 0.039842378348112106,
"learning_rate": 0.00019571105812198652,
"loss": 0.1329,
"step": 582
},
{
"epoch": 0.5535248041775457,
"grad_norm": 0.033689334988594055,
"learning_rate": 0.0001956789845254992,
"loss": 0.1265,
"step": 583
},
{
"epoch": 0.5544742463802516,
"grad_norm": 0.046588387340307236,
"learning_rate": 0.00019564679409681608,
"loss": 0.1645,
"step": 584
},
{
"epoch": 0.5554236885829575,
"grad_norm": 0.03861064463853836,
"learning_rate": 0.0001956144868752444,
"loss": 0.1267,
"step": 585
},
{
"epoch": 0.5563731307856634,
"grad_norm": 0.03467525169253349,
"learning_rate": 0.000195582062900234,
"loss": 0.1299,
"step": 586
},
{
"epoch": 0.5573225729883693,
"grad_norm": 0.03659389913082123,
"learning_rate": 0.0001955495222113774,
"loss": 0.1286,
"step": 587
},
{
"epoch": 0.5582720151910753,
"grad_norm": 0.03826770931482315,
"learning_rate": 0.0001955168648484095,
"loss": 0.1313,
"step": 588
},
{
"epoch": 0.5592214573937812,
"grad_norm": 0.038110729306936264,
"learning_rate": 0.00019548409085120772,
"loss": 0.137,
"step": 589
},
{
"epoch": 0.560170899596487,
"grad_norm": 0.03989555314183235,
"learning_rate": 0.0001954512002597919,
"loss": 0.132,
"step": 590
},
{
"epoch": 0.561120341799193,
"grad_norm": 0.05395180359482765,
"learning_rate": 0.00019541819311432427,
"loss": 0.1401,
"step": 591
},
{
"epoch": 0.5620697840018989,
"grad_norm": 0.05007918179035187,
"learning_rate": 0.00019538506945510938,
"loss": 0.1584,
"step": 592
},
{
"epoch": 0.5630192262046048,
"grad_norm": 0.047849785536527634,
"learning_rate": 0.00019535182932259404,
"loss": 0.1265,
"step": 593
},
{
"epoch": 0.5639686684073107,
"grad_norm": 0.04303041473031044,
"learning_rate": 0.00019531847275736726,
"loss": 0.1245,
"step": 594
},
{
"epoch": 0.5649181106100166,
"grad_norm": 0.04128289222717285,
"learning_rate": 0.00019528499980016025,
"loss": 0.1317,
"step": 595
},
{
"epoch": 0.5658675528127225,
"grad_norm": 0.04311414808034897,
"learning_rate": 0.00019525141049184637,
"loss": 0.1364,
"step": 596
},
{
"epoch": 0.5668169950154285,
"grad_norm": 0.03765838220715523,
"learning_rate": 0.00019521770487344103,
"loss": 0.1268,
"step": 597
},
{
"epoch": 0.5677664372181344,
"grad_norm": 0.03674585744738579,
"learning_rate": 0.00019518388298610164,
"loss": 0.1297,
"step": 598
},
{
"epoch": 0.5687158794208402,
"grad_norm": 0.036937762051820755,
"learning_rate": 0.0001951499448711276,
"loss": 0.1303,
"step": 599
},
{
"epoch": 0.5696653216235462,
"grad_norm": 0.03748161345720291,
"learning_rate": 0.0001951158905699603,
"loss": 0.1328,
"step": 600
},
{
"epoch": 0.5706147638262521,
"grad_norm": 0.04011257737874985,
"learning_rate": 0.00019508172012418283,
"loss": 0.1346,
"step": 601
},
{
"epoch": 0.571564206028958,
"grad_norm": 0.03853931650519371,
"learning_rate": 0.00019504743357552035,
"loss": 0.1279,
"step": 602
},
{
"epoch": 0.572513648231664,
"grad_norm": 0.03750459849834442,
"learning_rate": 0.0001950130309658396,
"loss": 0.1227,
"step": 603
},
{
"epoch": 0.5734630904343698,
"grad_norm": 0.05542079731822014,
"learning_rate": 0.00019497851233714908,
"loss": 0.1647,
"step": 604
},
{
"epoch": 0.5744125326370757,
"grad_norm": 0.04472218081355095,
"learning_rate": 0.00019494387773159898,
"loss": 0.1416,
"step": 605
},
{
"epoch": 0.5753619748397816,
"grad_norm": 0.052323974668979645,
"learning_rate": 0.00019490912719148114,
"loss": 0.1367,
"step": 606
},
{
"epoch": 0.5763114170424876,
"grad_norm": 0.037580832839012146,
"learning_rate": 0.00019487426075922893,
"loss": 0.131,
"step": 607
},
{
"epoch": 0.5772608592451934,
"grad_norm": 0.03929577395319939,
"learning_rate": 0.0001948392784774172,
"loss": 0.128,
"step": 608
},
{
"epoch": 0.5782103014478993,
"grad_norm": 0.03706606104969978,
"learning_rate": 0.0001948041803887623,
"loss": 0.1316,
"step": 609
},
{
"epoch": 0.5791597436506053,
"grad_norm": 0.038938358426094055,
"learning_rate": 0.00019476896653612203,
"loss": 0.1275,
"step": 610
},
{
"epoch": 0.5801091858533112,
"grad_norm": 0.04818068817257881,
"learning_rate": 0.00019473363696249546,
"loss": 0.1662,
"step": 611
},
{
"epoch": 0.581058628056017,
"grad_norm": 0.03735940158367157,
"learning_rate": 0.00019469819171102304,
"loss": 0.1361,
"step": 612
},
{
"epoch": 0.582008070258723,
"grad_norm": 0.03568827733397484,
"learning_rate": 0.00019466263082498645,
"loss": 0.1216,
"step": 613
},
{
"epoch": 0.5829575124614289,
"grad_norm": 0.03913251310586929,
"learning_rate": 0.0001946269543478085,
"loss": 0.1321,
"step": 614
},
{
"epoch": 0.5839069546641348,
"grad_norm": 0.062009479850530624,
"learning_rate": 0.0001945911623230533,
"loss": 0.1778,
"step": 615
},
{
"epoch": 0.5848563968668408,
"grad_norm": 0.039088111370801926,
"learning_rate": 0.0001945552547944259,
"loss": 0.1352,
"step": 616
},
{
"epoch": 0.5858058390695466,
"grad_norm": 0.041976600885391235,
"learning_rate": 0.0001945192318057725,
"loss": 0.1394,
"step": 617
},
{
"epoch": 0.5867552812722525,
"grad_norm": 0.03723563253879547,
"learning_rate": 0.00019448309340108018,
"loss": 0.1246,
"step": 618
},
{
"epoch": 0.5877047234749585,
"grad_norm": 0.0382399819791317,
"learning_rate": 0.00019444683962447707,
"loss": 0.1232,
"step": 619
},
{
"epoch": 0.5886541656776644,
"grad_norm": 0.03758077695965767,
"learning_rate": 0.0001944104705202321,
"loss": 0.1417,
"step": 620
},
{
"epoch": 0.5896036078803703,
"grad_norm": 0.034823786467313766,
"learning_rate": 0.000194373986132755,
"loss": 0.1304,
"step": 621
},
{
"epoch": 0.5905530500830762,
"grad_norm": 0.03755120187997818,
"learning_rate": 0.00019433738650659641,
"loss": 0.133,
"step": 622
},
{
"epoch": 0.5915024922857821,
"grad_norm": 0.03759913146495819,
"learning_rate": 0.00019430067168644754,
"loss": 0.1222,
"step": 623
},
{
"epoch": 0.592451934488488,
"grad_norm": 0.06232694163918495,
"learning_rate": 0.0001942638417171403,
"loss": 0.1778,
"step": 624
},
{
"epoch": 0.593401376691194,
"grad_norm": 0.05642306059598923,
"learning_rate": 0.00019422689664364725,
"loss": 0.1706,
"step": 625
},
{
"epoch": 0.5943508188938998,
"grad_norm": 0.0827709287405014,
"learning_rate": 0.00019418983651108148,
"loss": 0.2371,
"step": 626
},
{
"epoch": 0.5953002610966057,
"grad_norm": 0.03614366054534912,
"learning_rate": 0.00019415266136469652,
"loss": 0.1225,
"step": 627
},
{
"epoch": 0.5962497032993117,
"grad_norm": 0.042416494339704514,
"learning_rate": 0.00019411537124988643,
"loss": 0.1239,
"step": 628
},
{
"epoch": 0.5971991455020176,
"grad_norm": 0.037246908992528915,
"learning_rate": 0.00019407796621218566,
"loss": 0.1292,
"step": 629
},
{
"epoch": 0.5981485877047235,
"grad_norm": 0.05374092981219292,
"learning_rate": 0.00019404044629726887,
"loss": 0.1782,
"step": 630
},
{
"epoch": 0.5990980299074293,
"grad_norm": 0.052854426205158234,
"learning_rate": 0.00019400281155095112,
"loss": 0.1711,
"step": 631
},
{
"epoch": 0.6000474721101353,
"grad_norm": 0.038800131529569626,
"learning_rate": 0.00019396506201918765,
"loss": 0.1285,
"step": 632
}
],
"logging_steps": 1,
"max_steps": 3159,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 158,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 3.667125138649252e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}