aleynahukmet's picture
Upload folder using huggingface_hub
bf0d06b verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.998067259373792,
"eval_steps": 500,
"global_step": 1292,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0015461925009663702,
"grad_norm": 0.305040568113327,
"learning_rate": 5.000000000000001e-07,
"loss": 0.9801,
"step": 1
},
{
"epoch": 0.0030923850019327404,
"grad_norm": 0.3316026031970978,
"learning_rate": 1.0000000000000002e-06,
"loss": 1.0582,
"step": 2
},
{
"epoch": 0.004638577502899111,
"grad_norm": 0.321511834859848,
"learning_rate": 1.5e-06,
"loss": 1.0877,
"step": 3
},
{
"epoch": 0.006184770003865481,
"grad_norm": 0.32884451746940613,
"learning_rate": 2.0000000000000003e-06,
"loss": 1.1082,
"step": 4
},
{
"epoch": 0.007730962504831852,
"grad_norm": 0.3244039714336395,
"learning_rate": 2.5e-06,
"loss": 1.1352,
"step": 5
},
{
"epoch": 0.009277155005798222,
"grad_norm": 0.319381445646286,
"learning_rate": 3e-06,
"loss": 1.0808,
"step": 6
},
{
"epoch": 0.010823347506764593,
"grad_norm": 0.3265005946159363,
"learning_rate": 3.5000000000000004e-06,
"loss": 1.0554,
"step": 7
},
{
"epoch": 0.012369540007730962,
"grad_norm": 0.3248363137245178,
"learning_rate": 4.000000000000001e-06,
"loss": 1.0834,
"step": 8
},
{
"epoch": 0.013915732508697333,
"grad_norm": 0.3369300365447998,
"learning_rate": 4.5e-06,
"loss": 1.1304,
"step": 9
},
{
"epoch": 0.015461925009663703,
"grad_norm": 0.3302851915359497,
"learning_rate": 5e-06,
"loss": 1.0683,
"step": 10
},
{
"epoch": 0.017008117510630073,
"grad_norm": 0.3330378830432892,
"learning_rate": 5.500000000000001e-06,
"loss": 1.1028,
"step": 11
},
{
"epoch": 0.018554310011596443,
"grad_norm": 0.332900732755661,
"learning_rate": 6e-06,
"loss": 1.1557,
"step": 12
},
{
"epoch": 0.020100502512562814,
"grad_norm": 0.36040395498275757,
"learning_rate": 6.5000000000000004e-06,
"loss": 1.1344,
"step": 13
},
{
"epoch": 0.021646695013529185,
"grad_norm": 0.37165355682373047,
"learning_rate": 7.000000000000001e-06,
"loss": 1.1522,
"step": 14
},
{
"epoch": 0.023192887514495556,
"grad_norm": 0.35057052969932556,
"learning_rate": 7.5e-06,
"loss": 1.1397,
"step": 15
},
{
"epoch": 0.024739080015461924,
"grad_norm": 0.3664647042751312,
"learning_rate": 8.000000000000001e-06,
"loss": 1.1833,
"step": 16
},
{
"epoch": 0.026285272516428294,
"grad_norm": 0.38355353474617004,
"learning_rate": 8.500000000000002e-06,
"loss": 1.1925,
"step": 17
},
{
"epoch": 0.027831465017394665,
"grad_norm": 0.3568861186504364,
"learning_rate": 9e-06,
"loss": 1.1113,
"step": 18
},
{
"epoch": 0.029377657518361036,
"grad_norm": 0.3547118604183197,
"learning_rate": 9.5e-06,
"loss": 1.1403,
"step": 19
},
{
"epoch": 0.030923850019327407,
"grad_norm": 0.3746045231819153,
"learning_rate": 1e-05,
"loss": 1.1356,
"step": 20
},
{
"epoch": 0.03247004252029378,
"grad_norm": 0.38165998458862305,
"learning_rate": 1.05e-05,
"loss": 1.1239,
"step": 21
},
{
"epoch": 0.034016235021260145,
"grad_norm": 0.4228748679161072,
"learning_rate": 1.1000000000000001e-05,
"loss": 1.1263,
"step": 22
},
{
"epoch": 0.03556242752222652,
"grad_norm": 0.36566364765167236,
"learning_rate": 1.1500000000000002e-05,
"loss": 1.1362,
"step": 23
},
{
"epoch": 0.03710862002319289,
"grad_norm": 0.37338364124298096,
"learning_rate": 1.2e-05,
"loss": 1.1816,
"step": 24
},
{
"epoch": 0.038654812524159254,
"grad_norm": 0.3412342667579651,
"learning_rate": 1.25e-05,
"loss": 1.1341,
"step": 25
},
{
"epoch": 0.04020100502512563,
"grad_norm": 0.38790470361709595,
"learning_rate": 1.3000000000000001e-05,
"loss": 1.2407,
"step": 26
},
{
"epoch": 0.041747197526091996,
"grad_norm": 0.38183456659317017,
"learning_rate": 1.3500000000000001e-05,
"loss": 1.1206,
"step": 27
},
{
"epoch": 0.04329339002705837,
"grad_norm": 0.4037930965423584,
"learning_rate": 1.4000000000000001e-05,
"loss": 1.1111,
"step": 28
},
{
"epoch": 0.04483958252802474,
"grad_norm": 0.3747173547744751,
"learning_rate": 1.45e-05,
"loss": 1.14,
"step": 29
},
{
"epoch": 0.04638577502899111,
"grad_norm": 0.37349933385849,
"learning_rate": 1.5e-05,
"loss": 1.1629,
"step": 30
},
{
"epoch": 0.04793196752995748,
"grad_norm": 0.3699789345264435,
"learning_rate": 1.55e-05,
"loss": 1.1417,
"step": 31
},
{
"epoch": 0.04947816003092385,
"grad_norm": 0.37245669960975647,
"learning_rate": 1.6000000000000003e-05,
"loss": 1.1563,
"step": 32
},
{
"epoch": 0.05102435253189022,
"grad_norm": 0.3655848503112793,
"learning_rate": 1.65e-05,
"loss": 1.1556,
"step": 33
},
{
"epoch": 0.05257054503285659,
"grad_norm": 0.3637336194515228,
"learning_rate": 1.7000000000000003e-05,
"loss": 1.1589,
"step": 34
},
{
"epoch": 0.05411673753382296,
"grad_norm": 0.35633373260498047,
"learning_rate": 1.75e-05,
"loss": 1.0969,
"step": 35
},
{
"epoch": 0.05566293003478933,
"grad_norm": 0.35284438729286194,
"learning_rate": 1.8e-05,
"loss": 1.0921,
"step": 36
},
{
"epoch": 0.057209122535755705,
"grad_norm": 0.3530278205871582,
"learning_rate": 1.85e-05,
"loss": 1.0982,
"step": 37
},
{
"epoch": 0.05875531503672207,
"grad_norm": 0.3533940315246582,
"learning_rate": 1.9e-05,
"loss": 1.1329,
"step": 38
},
{
"epoch": 0.06030150753768844,
"grad_norm": 0.35752803087234497,
"learning_rate": 1.9500000000000003e-05,
"loss": 1.1617,
"step": 39
},
{
"epoch": 0.061847700038654814,
"grad_norm": 0.34077584743499756,
"learning_rate": 2e-05,
"loss": 1.0956,
"step": 40
},
{
"epoch": 0.06339389253962119,
"grad_norm": 0.36975380778312683,
"learning_rate": 2.05e-05,
"loss": 1.0757,
"step": 41
},
{
"epoch": 0.06494008504058756,
"grad_norm": 0.36623820662498474,
"learning_rate": 2.1e-05,
"loss": 1.1465,
"step": 42
},
{
"epoch": 0.06648627754155392,
"grad_norm": 0.3988489508628845,
"learning_rate": 2.15e-05,
"loss": 1.1701,
"step": 43
},
{
"epoch": 0.06803247004252029,
"grad_norm": 0.37758833169937134,
"learning_rate": 2.2000000000000003e-05,
"loss": 1.0647,
"step": 44
},
{
"epoch": 0.06957866254348666,
"grad_norm": 0.4047185182571411,
"learning_rate": 2.25e-05,
"loss": 1.1057,
"step": 45
},
{
"epoch": 0.07112485504445304,
"grad_norm": 0.39636239409446716,
"learning_rate": 2.3000000000000003e-05,
"loss": 1.1294,
"step": 46
},
{
"epoch": 0.0726710475454194,
"grad_norm": 0.4347302317619324,
"learning_rate": 2.35e-05,
"loss": 1.152,
"step": 47
},
{
"epoch": 0.07421724004638577,
"grad_norm": 0.4483806788921356,
"learning_rate": 2.4e-05,
"loss": 1.1676,
"step": 48
},
{
"epoch": 0.07576343254735214,
"grad_norm": 0.5394858121871948,
"learning_rate": 2.45e-05,
"loss": 1.2228,
"step": 49
},
{
"epoch": 0.07730962504831851,
"grad_norm": 0.7282954454421997,
"learning_rate": 2.5e-05,
"loss": 1.2397,
"step": 50
},
{
"epoch": 0.07885581754928489,
"grad_norm": 0.3573731482028961,
"learning_rate": 2.5500000000000003e-05,
"loss": 0.9119,
"step": 51
},
{
"epoch": 0.08040201005025126,
"grad_norm": 0.36452800035476685,
"learning_rate": 2.6000000000000002e-05,
"loss": 0.983,
"step": 52
},
{
"epoch": 0.08194820255121762,
"grad_norm": 0.3743703067302704,
"learning_rate": 2.6500000000000004e-05,
"loss": 0.9428,
"step": 53
},
{
"epoch": 0.08349439505218399,
"grad_norm": 0.3590448498725891,
"learning_rate": 2.7000000000000002e-05,
"loss": 0.9967,
"step": 54
},
{
"epoch": 0.08504058755315037,
"grad_norm": 0.35900965332984924,
"learning_rate": 2.7500000000000004e-05,
"loss": 0.9687,
"step": 55
},
{
"epoch": 0.08658678005411674,
"grad_norm": 0.3654124140739441,
"learning_rate": 2.8000000000000003e-05,
"loss": 0.9342,
"step": 56
},
{
"epoch": 0.08813297255508311,
"grad_norm": 0.3421430289745331,
"learning_rate": 2.8499999999999998e-05,
"loss": 0.9207,
"step": 57
},
{
"epoch": 0.08967916505604948,
"grad_norm": 0.3365703523159027,
"learning_rate": 2.9e-05,
"loss": 0.9487,
"step": 58
},
{
"epoch": 0.09122535755701584,
"grad_norm": 0.32828226685523987,
"learning_rate": 2.95e-05,
"loss": 0.9236,
"step": 59
},
{
"epoch": 0.09277155005798222,
"grad_norm": 0.31865429878234863,
"learning_rate": 3e-05,
"loss": 0.9071,
"step": 60
},
{
"epoch": 0.09431774255894859,
"grad_norm": 0.3250292241573334,
"learning_rate": 3.05e-05,
"loss": 0.9609,
"step": 61
},
{
"epoch": 0.09586393505991496,
"grad_norm": 0.3177807033061981,
"learning_rate": 3.1e-05,
"loss": 0.9727,
"step": 62
},
{
"epoch": 0.09741012756088133,
"grad_norm": 0.3273405134677887,
"learning_rate": 3.15e-05,
"loss": 1.0691,
"step": 63
},
{
"epoch": 0.0989563200618477,
"grad_norm": 0.30870744585990906,
"learning_rate": 3.2000000000000005e-05,
"loss": 0.9726,
"step": 64
},
{
"epoch": 0.10050251256281408,
"grad_norm": 0.3052951395511627,
"learning_rate": 3.2500000000000004e-05,
"loss": 0.8804,
"step": 65
},
{
"epoch": 0.10204870506378044,
"grad_norm": 0.2952975928783417,
"learning_rate": 3.3e-05,
"loss": 0.9563,
"step": 66
},
{
"epoch": 0.10359489756474681,
"grad_norm": 0.29250308871269226,
"learning_rate": 3.35e-05,
"loss": 0.9601,
"step": 67
},
{
"epoch": 0.10514109006571318,
"grad_norm": 0.2929418683052063,
"learning_rate": 3.4000000000000007e-05,
"loss": 0.8986,
"step": 68
},
{
"epoch": 0.10668728256667954,
"grad_norm": 0.28594040870666504,
"learning_rate": 3.45e-05,
"loss": 0.8887,
"step": 69
},
{
"epoch": 0.10823347506764593,
"grad_norm": 0.2964857816696167,
"learning_rate": 3.5e-05,
"loss": 0.9407,
"step": 70
},
{
"epoch": 0.1097796675686123,
"grad_norm": 0.31084346771240234,
"learning_rate": 3.55e-05,
"loss": 0.9907,
"step": 71
},
{
"epoch": 0.11132586006957866,
"grad_norm": 0.31780946254730225,
"learning_rate": 3.6e-05,
"loss": 0.9753,
"step": 72
},
{
"epoch": 0.11287205257054503,
"grad_norm": 0.31742170453071594,
"learning_rate": 3.65e-05,
"loss": 0.9411,
"step": 73
},
{
"epoch": 0.11441824507151141,
"grad_norm": 0.3252294361591339,
"learning_rate": 3.7e-05,
"loss": 0.9973,
"step": 74
},
{
"epoch": 0.11596443757247778,
"grad_norm": 0.3107585608959198,
"learning_rate": 3.7500000000000003e-05,
"loss": 0.9186,
"step": 75
},
{
"epoch": 0.11751063007344414,
"grad_norm": 0.32054993510246277,
"learning_rate": 3.8e-05,
"loss": 0.9264,
"step": 76
},
{
"epoch": 0.11905682257441051,
"grad_norm": 0.31898629665374756,
"learning_rate": 3.85e-05,
"loss": 0.9958,
"step": 77
},
{
"epoch": 0.12060301507537688,
"grad_norm": 0.3400716185569763,
"learning_rate": 3.9000000000000006e-05,
"loss": 0.9793,
"step": 78
},
{
"epoch": 0.12214920757634326,
"grad_norm": 0.347918838262558,
"learning_rate": 3.9500000000000005e-05,
"loss": 0.914,
"step": 79
},
{
"epoch": 0.12369540007730963,
"grad_norm": 0.33906570076942444,
"learning_rate": 4e-05,
"loss": 0.93,
"step": 80
},
{
"epoch": 0.125241592578276,
"grad_norm": 0.31842949986457825,
"learning_rate": 4.05e-05,
"loss": 0.8957,
"step": 81
},
{
"epoch": 0.12678778507924238,
"grad_norm": 0.3301747739315033,
"learning_rate": 4.1e-05,
"loss": 0.9606,
"step": 82
},
{
"epoch": 0.12833397758020873,
"grad_norm": 0.33865636587142944,
"learning_rate": 4.15e-05,
"loss": 0.9846,
"step": 83
},
{
"epoch": 0.1298801700811751,
"grad_norm": 0.31991609930992126,
"learning_rate": 4.2e-05,
"loss": 0.9574,
"step": 84
},
{
"epoch": 0.13142636258214146,
"grad_norm": 0.3442661762237549,
"learning_rate": 4.25e-05,
"loss": 0.9398,
"step": 85
},
{
"epoch": 0.13297255508310785,
"grad_norm": 0.33919069170951843,
"learning_rate": 4.3e-05,
"loss": 0.9736,
"step": 86
},
{
"epoch": 0.13451874758407423,
"grad_norm": 0.3447560966014862,
"learning_rate": 4.35e-05,
"loss": 0.9589,
"step": 87
},
{
"epoch": 0.13606494008504058,
"grad_norm": 0.3504810333251953,
"learning_rate": 4.4000000000000006e-05,
"loss": 1.0156,
"step": 88
},
{
"epoch": 0.13761113258600696,
"grad_norm": 0.35089215636253357,
"learning_rate": 4.4500000000000004e-05,
"loss": 0.9732,
"step": 89
},
{
"epoch": 0.13915732508697332,
"grad_norm": 0.346599280834198,
"learning_rate": 4.5e-05,
"loss": 0.9705,
"step": 90
},
{
"epoch": 0.1407035175879397,
"grad_norm": 0.3983485698699951,
"learning_rate": 4.55e-05,
"loss": 1.0236,
"step": 91
},
{
"epoch": 0.14224971008890608,
"grad_norm": 0.38158902525901794,
"learning_rate": 4.600000000000001e-05,
"loss": 1.0174,
"step": 92
},
{
"epoch": 0.14379590258987243,
"grad_norm": 0.3771804869174957,
"learning_rate": 4.6500000000000005e-05,
"loss": 1.012,
"step": 93
},
{
"epoch": 0.1453420950908388,
"grad_norm": 0.38228315114974976,
"learning_rate": 4.7e-05,
"loss": 0.9786,
"step": 94
},
{
"epoch": 0.14688828759180517,
"grad_norm": 0.3989880681037903,
"learning_rate": 4.75e-05,
"loss": 0.98,
"step": 95
},
{
"epoch": 0.14843448009277155,
"grad_norm": 0.41765791177749634,
"learning_rate": 4.8e-05,
"loss": 1.0502,
"step": 96
},
{
"epoch": 0.14998067259373793,
"grad_norm": 0.4242067039012909,
"learning_rate": 4.85e-05,
"loss": 1.0911,
"step": 97
},
{
"epoch": 0.15152686509470428,
"grad_norm": 0.4490616023540497,
"learning_rate": 4.9e-05,
"loss": 1.0791,
"step": 98
},
{
"epoch": 0.15307305759567066,
"grad_norm": 0.4694664478302002,
"learning_rate": 4.9500000000000004e-05,
"loss": 1.0374,
"step": 99
},
{
"epoch": 0.15461925009663702,
"grad_norm": 0.7335464954376221,
"learning_rate": 5e-05,
"loss": 1.093,
"step": 100
},
{
"epoch": 0.1561654425976034,
"grad_norm": 0.3335883915424347,
"learning_rate": 4.995805369127517e-05,
"loss": 0.8461,
"step": 101
},
{
"epoch": 0.15771163509856978,
"grad_norm": 0.3368653357028961,
"learning_rate": 4.9916107382550336e-05,
"loss": 0.8619,
"step": 102
},
{
"epoch": 0.15925782759953613,
"grad_norm": 0.34530943632125854,
"learning_rate": 4.9874161073825505e-05,
"loss": 0.8399,
"step": 103
},
{
"epoch": 0.16080402010050251,
"grad_norm": 0.3303908407688141,
"learning_rate": 4.983221476510067e-05,
"loss": 0.891,
"step": 104
},
{
"epoch": 0.1623502126014689,
"grad_norm": 0.3104991614818573,
"learning_rate": 4.9790268456375845e-05,
"loss": 0.8851,
"step": 105
},
{
"epoch": 0.16389640510243525,
"grad_norm": 0.3138999342918396,
"learning_rate": 4.974832214765101e-05,
"loss": 0.8668,
"step": 106
},
{
"epoch": 0.16544259760340163,
"grad_norm": 0.3183957040309906,
"learning_rate": 4.970637583892618e-05,
"loss": 0.8494,
"step": 107
},
{
"epoch": 0.16698879010436798,
"grad_norm": 0.337296724319458,
"learning_rate": 4.966442953020135e-05,
"loss": 0.872,
"step": 108
},
{
"epoch": 0.16853498260533437,
"grad_norm": 0.29680609703063965,
"learning_rate": 4.962248322147651e-05,
"loss": 0.8473,
"step": 109
},
{
"epoch": 0.17008117510630075,
"grad_norm": 0.31526410579681396,
"learning_rate": 4.958053691275168e-05,
"loss": 0.9042,
"step": 110
},
{
"epoch": 0.1716273676072671,
"grad_norm": 0.3531287610530853,
"learning_rate": 4.9538590604026845e-05,
"loss": 0.9444,
"step": 111
},
{
"epoch": 0.17317356010823348,
"grad_norm": 0.324305921792984,
"learning_rate": 4.9496644295302015e-05,
"loss": 0.8867,
"step": 112
},
{
"epoch": 0.17471975260919984,
"grad_norm": 0.32134464383125305,
"learning_rate": 4.945469798657718e-05,
"loss": 0.8575,
"step": 113
},
{
"epoch": 0.17626594511016622,
"grad_norm": 0.32409024238586426,
"learning_rate": 4.9412751677852355e-05,
"loss": 0.8937,
"step": 114
},
{
"epoch": 0.1778121376111326,
"grad_norm": 0.3491647243499756,
"learning_rate": 4.937080536912752e-05,
"loss": 0.8902,
"step": 115
},
{
"epoch": 0.17935833011209895,
"grad_norm": 0.3334656357765198,
"learning_rate": 4.932885906040269e-05,
"loss": 0.8788,
"step": 116
},
{
"epoch": 0.18090452261306533,
"grad_norm": 0.3333323299884796,
"learning_rate": 4.928691275167786e-05,
"loss": 0.907,
"step": 117
},
{
"epoch": 0.1824507151140317,
"grad_norm": 0.3563230335712433,
"learning_rate": 4.924496644295302e-05,
"loss": 0.9043,
"step": 118
},
{
"epoch": 0.18399690761499807,
"grad_norm": 0.35443171858787537,
"learning_rate": 4.920302013422819e-05,
"loss": 0.8966,
"step": 119
},
{
"epoch": 0.18554310011596445,
"grad_norm": 0.36155420541763306,
"learning_rate": 4.9161073825503354e-05,
"loss": 0.903,
"step": 120
},
{
"epoch": 0.1870892926169308,
"grad_norm": 0.3701721131801605,
"learning_rate": 4.9119127516778524e-05,
"loss": 0.8786,
"step": 121
},
{
"epoch": 0.18863548511789718,
"grad_norm": 0.3697488307952881,
"learning_rate": 4.9077181208053694e-05,
"loss": 0.9279,
"step": 122
},
{
"epoch": 0.19018167761886354,
"grad_norm": 0.35044190287590027,
"learning_rate": 4.9035234899328864e-05,
"loss": 0.8577,
"step": 123
},
{
"epoch": 0.19172787011982992,
"grad_norm": 0.3510192632675171,
"learning_rate": 4.8993288590604034e-05,
"loss": 0.8831,
"step": 124
},
{
"epoch": 0.1932740626207963,
"grad_norm": 0.3509821593761444,
"learning_rate": 4.89513422818792e-05,
"loss": 0.8507,
"step": 125
},
{
"epoch": 0.19482025512176265,
"grad_norm": 0.3761618137359619,
"learning_rate": 4.890939597315437e-05,
"loss": 0.9151,
"step": 126
},
{
"epoch": 0.19636644762272903,
"grad_norm": 0.3758131265640259,
"learning_rate": 4.886744966442953e-05,
"loss": 0.9067,
"step": 127
},
{
"epoch": 0.1979126401236954,
"grad_norm": 0.39383259415626526,
"learning_rate": 4.88255033557047e-05,
"loss": 0.9002,
"step": 128
},
{
"epoch": 0.19945883262466177,
"grad_norm": 0.37084588408470154,
"learning_rate": 4.878355704697986e-05,
"loss": 0.8459,
"step": 129
},
{
"epoch": 0.20100502512562815,
"grad_norm": 0.389466792345047,
"learning_rate": 4.874161073825503e-05,
"loss": 0.9218,
"step": 130
},
{
"epoch": 0.2025512176265945,
"grad_norm": 0.38749396800994873,
"learning_rate": 4.86996644295302e-05,
"loss": 0.892,
"step": 131
},
{
"epoch": 0.20409741012756089,
"grad_norm": 0.36285966634750366,
"learning_rate": 4.865771812080537e-05,
"loss": 0.8458,
"step": 132
},
{
"epoch": 0.20564360262852724,
"grad_norm": 0.38966134190559387,
"learning_rate": 4.861577181208054e-05,
"loss": 0.9189,
"step": 133
},
{
"epoch": 0.20718979512949362,
"grad_norm": 0.3792301118373871,
"learning_rate": 4.8573825503355706e-05,
"loss": 0.891,
"step": 134
},
{
"epoch": 0.20873598763046,
"grad_norm": 0.39519718289375305,
"learning_rate": 4.8531879194630876e-05,
"loss": 0.8946,
"step": 135
},
{
"epoch": 0.21028218013142636,
"grad_norm": 0.399034708738327,
"learning_rate": 4.848993288590604e-05,
"loss": 0.8692,
"step": 136
},
{
"epoch": 0.21182837263239274,
"grad_norm": 0.451168954372406,
"learning_rate": 4.844798657718121e-05,
"loss": 0.9005,
"step": 137
},
{
"epoch": 0.2133745651333591,
"grad_norm": 0.41019031405448914,
"learning_rate": 4.840604026845638e-05,
"loss": 0.9766,
"step": 138
},
{
"epoch": 0.21492075763432547,
"grad_norm": 0.4109886586666107,
"learning_rate": 4.836409395973154e-05,
"loss": 0.9078,
"step": 139
},
{
"epoch": 0.21646695013529185,
"grad_norm": 0.4546094834804535,
"learning_rate": 4.832214765100672e-05,
"loss": 0.9621,
"step": 140
},
{
"epoch": 0.2180131426362582,
"grad_norm": 0.44215071201324463,
"learning_rate": 4.828020134228188e-05,
"loss": 0.9965,
"step": 141
},
{
"epoch": 0.2195593351372246,
"grad_norm": 0.4271162748336792,
"learning_rate": 4.823825503355705e-05,
"loss": 0.9277,
"step": 142
},
{
"epoch": 0.22110552763819097,
"grad_norm": 0.43438541889190674,
"learning_rate": 4.8196308724832215e-05,
"loss": 0.9727,
"step": 143
},
{
"epoch": 0.22265172013915732,
"grad_norm": 0.45643556118011475,
"learning_rate": 4.8154362416107385e-05,
"loss": 0.9284,
"step": 144
},
{
"epoch": 0.2241979126401237,
"grad_norm": 0.5079519748687744,
"learning_rate": 4.8112416107382555e-05,
"loss": 1.004,
"step": 145
},
{
"epoch": 0.22574410514109006,
"grad_norm": 0.4708511531352997,
"learning_rate": 4.807046979865772e-05,
"loss": 1.001,
"step": 146
},
{
"epoch": 0.22729029764205644,
"grad_norm": 0.510336697101593,
"learning_rate": 4.802852348993289e-05,
"loss": 0.9808,
"step": 147
},
{
"epoch": 0.22883649014302282,
"grad_norm": 0.5595572590827942,
"learning_rate": 4.798657718120805e-05,
"loss": 0.9983,
"step": 148
},
{
"epoch": 0.23038268264398917,
"grad_norm": 0.5763404965400696,
"learning_rate": 4.794463087248323e-05,
"loss": 0.9841,
"step": 149
},
{
"epoch": 0.23192887514495555,
"grad_norm": 0.9527286887168884,
"learning_rate": 4.790268456375839e-05,
"loss": 0.9827,
"step": 150
},
{
"epoch": 0.2334750676459219,
"grad_norm": 0.35568949580192566,
"learning_rate": 4.786073825503356e-05,
"loss": 0.8084,
"step": 151
},
{
"epoch": 0.2350212601468883,
"grad_norm": 0.3866717517375946,
"learning_rate": 4.7818791946308725e-05,
"loss": 0.8538,
"step": 152
},
{
"epoch": 0.23656745264785467,
"grad_norm": 0.3884953260421753,
"learning_rate": 4.7776845637583895e-05,
"loss": 0.853,
"step": 153
},
{
"epoch": 0.23811364514882102,
"grad_norm": 0.40036800503730774,
"learning_rate": 4.7734899328859064e-05,
"loss": 0.8604,
"step": 154
},
{
"epoch": 0.2396598376497874,
"grad_norm": 0.4019627273082733,
"learning_rate": 4.769295302013423e-05,
"loss": 0.8409,
"step": 155
},
{
"epoch": 0.24120603015075376,
"grad_norm": 0.3811728358268738,
"learning_rate": 4.76510067114094e-05,
"loss": 0.8602,
"step": 156
},
{
"epoch": 0.24275222265172014,
"grad_norm": 0.39744654297828674,
"learning_rate": 4.760906040268457e-05,
"loss": 0.8011,
"step": 157
},
{
"epoch": 0.24429841515268652,
"grad_norm": 0.38136810064315796,
"learning_rate": 4.756711409395974e-05,
"loss": 0.9004,
"step": 158
},
{
"epoch": 0.24584460765365287,
"grad_norm": 0.3670859932899475,
"learning_rate": 4.75251677852349e-05,
"loss": 0.8506,
"step": 159
},
{
"epoch": 0.24739080015461926,
"grad_norm": 0.4202471375465393,
"learning_rate": 4.748322147651007e-05,
"loss": 0.7832,
"step": 160
},
{
"epoch": 0.2489369926555856,
"grad_norm": 0.40722745656967163,
"learning_rate": 4.744127516778524e-05,
"loss": 0.8813,
"step": 161
},
{
"epoch": 0.250483185156552,
"grad_norm": 0.4037550389766693,
"learning_rate": 4.7399328859060404e-05,
"loss": 0.8328,
"step": 162
},
{
"epoch": 0.25202937765751837,
"grad_norm": 0.4262521266937256,
"learning_rate": 4.7357382550335574e-05,
"loss": 0.9148,
"step": 163
},
{
"epoch": 0.25357557015848475,
"grad_norm": 0.40966907143592834,
"learning_rate": 4.731543624161074e-05,
"loss": 0.8689,
"step": 164
},
{
"epoch": 0.2551217626594511,
"grad_norm": 0.41350632905960083,
"learning_rate": 4.727348993288591e-05,
"loss": 0.8253,
"step": 165
},
{
"epoch": 0.25666795516041746,
"grad_norm": 0.3822355568408966,
"learning_rate": 4.723154362416108e-05,
"loss": 0.8176,
"step": 166
},
{
"epoch": 0.25821414766138384,
"grad_norm": 0.4168022572994232,
"learning_rate": 4.718959731543625e-05,
"loss": 0.8821,
"step": 167
},
{
"epoch": 0.2597603401623502,
"grad_norm": 0.4205927848815918,
"learning_rate": 4.714765100671141e-05,
"loss": 0.9162,
"step": 168
},
{
"epoch": 0.2613065326633166,
"grad_norm": 0.41793638467788696,
"learning_rate": 4.710570469798658e-05,
"loss": 0.9045,
"step": 169
},
{
"epoch": 0.26285272516428293,
"grad_norm": 0.41607388854026794,
"learning_rate": 4.706375838926175e-05,
"loss": 0.8474,
"step": 170
},
{
"epoch": 0.2643989176652493,
"grad_norm": 0.3936135172843933,
"learning_rate": 4.702181208053691e-05,
"loss": 0.8859,
"step": 171
},
{
"epoch": 0.2659451101662157,
"grad_norm": 0.39022210240364075,
"learning_rate": 4.697986577181208e-05,
"loss": 0.8455,
"step": 172
},
{
"epoch": 0.2674913026671821,
"grad_norm": 0.40289798378944397,
"learning_rate": 4.6937919463087246e-05,
"loss": 0.9244,
"step": 173
},
{
"epoch": 0.26903749516814845,
"grad_norm": 0.4135661721229553,
"learning_rate": 4.6895973154362416e-05,
"loss": 0.8427,
"step": 174
},
{
"epoch": 0.2705836876691148,
"grad_norm": 0.4445662796497345,
"learning_rate": 4.6854026845637586e-05,
"loss": 0.8042,
"step": 175
},
{
"epoch": 0.27212988017008116,
"grad_norm": 0.42175382375717163,
"learning_rate": 4.6812080536912756e-05,
"loss": 0.8447,
"step": 176
},
{
"epoch": 0.27367607267104754,
"grad_norm": 0.42236700654029846,
"learning_rate": 4.6770134228187926e-05,
"loss": 0.8964,
"step": 177
},
{
"epoch": 0.2752222651720139,
"grad_norm": 0.44100165367126465,
"learning_rate": 4.672818791946309e-05,
"loss": 0.7989,
"step": 178
},
{
"epoch": 0.2767684576729803,
"grad_norm": 0.43265581130981445,
"learning_rate": 4.668624161073826e-05,
"loss": 0.8795,
"step": 179
},
{
"epoch": 0.27831465017394663,
"grad_norm": 0.43812161684036255,
"learning_rate": 4.664429530201342e-05,
"loss": 0.8984,
"step": 180
},
{
"epoch": 0.279860842674913,
"grad_norm": 0.4735831916332245,
"learning_rate": 4.660234899328859e-05,
"loss": 0.8868,
"step": 181
},
{
"epoch": 0.2814070351758794,
"grad_norm": 0.4555661678314209,
"learning_rate": 4.6560402684563755e-05,
"loss": 0.9057,
"step": 182
},
{
"epoch": 0.2829532276768458,
"grad_norm": 0.44536617398262024,
"learning_rate": 4.6518456375838925e-05,
"loss": 0.8956,
"step": 183
},
{
"epoch": 0.28449942017781216,
"grad_norm": 0.4430801272392273,
"learning_rate": 4.6476510067114095e-05,
"loss": 0.9171,
"step": 184
},
{
"epoch": 0.2860456126787785,
"grad_norm": 0.4534691572189331,
"learning_rate": 4.6434563758389265e-05,
"loss": 0.7871,
"step": 185
},
{
"epoch": 0.28759180517974486,
"grad_norm": 0.46288105845451355,
"learning_rate": 4.6392617449664435e-05,
"loss": 0.8819,
"step": 186
},
{
"epoch": 0.28913799768071125,
"grad_norm": 0.4744343161582947,
"learning_rate": 4.63506711409396e-05,
"loss": 0.9415,
"step": 187
},
{
"epoch": 0.2906841901816776,
"grad_norm": 0.46916016936302185,
"learning_rate": 4.630872483221477e-05,
"loss": 0.9436,
"step": 188
},
{
"epoch": 0.292230382682644,
"grad_norm": 0.4746655821800232,
"learning_rate": 4.626677852348993e-05,
"loss": 0.9039,
"step": 189
},
{
"epoch": 0.29377657518361033,
"grad_norm": 0.48333853483200073,
"learning_rate": 4.62248322147651e-05,
"loss": 0.9466,
"step": 190
},
{
"epoch": 0.2953227676845767,
"grad_norm": 0.4735361933708191,
"learning_rate": 4.618288590604027e-05,
"loss": 0.9088,
"step": 191
},
{
"epoch": 0.2968689601855431,
"grad_norm": 0.484523206949234,
"learning_rate": 4.6140939597315434e-05,
"loss": 0.909,
"step": 192
},
{
"epoch": 0.2984151526865095,
"grad_norm": 0.5111984610557556,
"learning_rate": 4.609899328859061e-05,
"loss": 0.9451,
"step": 193
},
{
"epoch": 0.29996134518747586,
"grad_norm": 0.5157277584075928,
"learning_rate": 4.6057046979865774e-05,
"loss": 0.9768,
"step": 194
},
{
"epoch": 0.3015075376884422,
"grad_norm": 0.5363951325416565,
"learning_rate": 4.6015100671140944e-05,
"loss": 0.9411,
"step": 195
},
{
"epoch": 0.30305373018940857,
"grad_norm": 0.5306119918823242,
"learning_rate": 4.597315436241611e-05,
"loss": 0.9238,
"step": 196
},
{
"epoch": 0.30459992269037495,
"grad_norm": 0.5576231479644775,
"learning_rate": 4.593120805369128e-05,
"loss": 0.9866,
"step": 197
},
{
"epoch": 0.30614611519134133,
"grad_norm": 0.6054885983467102,
"learning_rate": 4.588926174496645e-05,
"loss": 0.9987,
"step": 198
},
{
"epoch": 0.3076923076923077,
"grad_norm": 0.6553575396537781,
"learning_rate": 4.584731543624161e-05,
"loss": 1.0413,
"step": 199
},
{
"epoch": 0.30923850019327404,
"grad_norm": 1.023345708847046,
"learning_rate": 4.580536912751678e-05,
"loss": 1.0355,
"step": 200
},
{
"epoch": 0.3107846926942404,
"grad_norm": 0.4243873059749603,
"learning_rate": 4.576342281879195e-05,
"loss": 0.7642,
"step": 201
},
{
"epoch": 0.3123308851952068,
"grad_norm": 0.4183518886566162,
"learning_rate": 4.572147651006712e-05,
"loss": 0.8115,
"step": 202
},
{
"epoch": 0.3138770776961732,
"grad_norm": 0.43979495763778687,
"learning_rate": 4.5679530201342284e-05,
"loss": 0.7705,
"step": 203
},
{
"epoch": 0.31542327019713956,
"grad_norm": 0.44140326976776123,
"learning_rate": 4.5637583892617453e-05,
"loss": 0.7784,
"step": 204
},
{
"epoch": 0.31696946269810594,
"grad_norm": 0.4253062903881073,
"learning_rate": 4.559563758389262e-05,
"loss": 0.8006,
"step": 205
},
{
"epoch": 0.31851565519907227,
"grad_norm": 0.41898655891418457,
"learning_rate": 4.5553691275167787e-05,
"loss": 0.8304,
"step": 206
},
{
"epoch": 0.32006184770003865,
"grad_norm": 0.3986593186855316,
"learning_rate": 4.5511744966442957e-05,
"loss": 0.7815,
"step": 207
},
{
"epoch": 0.32160804020100503,
"grad_norm": 0.42180249094963074,
"learning_rate": 4.546979865771812e-05,
"loss": 0.859,
"step": 208
},
{
"epoch": 0.3231542327019714,
"grad_norm": 0.429800808429718,
"learning_rate": 4.542785234899329e-05,
"loss": 0.8188,
"step": 209
},
{
"epoch": 0.3247004252029378,
"grad_norm": 0.4582015573978424,
"learning_rate": 4.538590604026846e-05,
"loss": 0.8069,
"step": 210
},
{
"epoch": 0.3262466177039041,
"grad_norm": 0.44153645634651184,
"learning_rate": 4.534395973154363e-05,
"loss": 0.8107,
"step": 211
},
{
"epoch": 0.3277928102048705,
"grad_norm": 0.4466538429260254,
"learning_rate": 4.530201342281879e-05,
"loss": 0.8076,
"step": 212
},
{
"epoch": 0.3293390027058369,
"grad_norm": 0.4263439178466797,
"learning_rate": 4.526006711409396e-05,
"loss": 0.8532,
"step": 213
},
{
"epoch": 0.33088519520680326,
"grad_norm": 0.469927042722702,
"learning_rate": 4.521812080536913e-05,
"loss": 0.8264,
"step": 214
},
{
"epoch": 0.33243138770776964,
"grad_norm": 0.45597830414772034,
"learning_rate": 4.5176174496644296e-05,
"loss": 0.8125,
"step": 215
},
{
"epoch": 0.33397758020873597,
"grad_norm": 0.43289533257484436,
"learning_rate": 4.5134228187919466e-05,
"loss": 0.8223,
"step": 216
},
{
"epoch": 0.33552377270970235,
"grad_norm": 0.4476693272590637,
"learning_rate": 4.509228187919463e-05,
"loss": 0.8206,
"step": 217
},
{
"epoch": 0.33706996521066873,
"grad_norm": 0.43755078315734863,
"learning_rate": 4.50503355704698e-05,
"loss": 0.8084,
"step": 218
},
{
"epoch": 0.3386161577116351,
"grad_norm": 0.4320957064628601,
"learning_rate": 4.500838926174497e-05,
"loss": 0.8127,
"step": 219
},
{
"epoch": 0.3401623502126015,
"grad_norm": 0.45105233788490295,
"learning_rate": 4.496644295302014e-05,
"loss": 0.8208,
"step": 220
},
{
"epoch": 0.3417085427135678,
"grad_norm": 0.42372819781303406,
"learning_rate": 4.49244966442953e-05,
"loss": 0.8685,
"step": 221
},
{
"epoch": 0.3432547352145342,
"grad_norm": 0.4444120526313782,
"learning_rate": 4.488255033557047e-05,
"loss": 0.8501,
"step": 222
},
{
"epoch": 0.3448009277155006,
"grad_norm": 0.4355757236480713,
"learning_rate": 4.484060402684564e-05,
"loss": 0.8507,
"step": 223
},
{
"epoch": 0.34634712021646696,
"grad_norm": 0.45914098620414734,
"learning_rate": 4.4798657718120805e-05,
"loss": 0.9056,
"step": 224
},
{
"epoch": 0.34789331271743335,
"grad_norm": 0.4455774128437042,
"learning_rate": 4.4756711409395975e-05,
"loss": 0.8716,
"step": 225
},
{
"epoch": 0.34943950521839967,
"grad_norm": 0.4548904299736023,
"learning_rate": 4.471476510067114e-05,
"loss": 0.7938,
"step": 226
},
{
"epoch": 0.35098569771936605,
"grad_norm": 0.44709107279777527,
"learning_rate": 4.467281879194631e-05,
"loss": 0.8248,
"step": 227
},
{
"epoch": 0.35253189022033243,
"grad_norm": 0.4717913866043091,
"learning_rate": 4.463087248322148e-05,
"loss": 0.864,
"step": 228
},
{
"epoch": 0.3540780827212988,
"grad_norm": 0.48902633786201477,
"learning_rate": 4.458892617449665e-05,
"loss": 0.7545,
"step": 229
},
{
"epoch": 0.3556242752222652,
"grad_norm": 0.4466339945793152,
"learning_rate": 4.454697986577182e-05,
"loss": 0.8869,
"step": 230
},
{
"epoch": 0.3571704677232315,
"grad_norm": 0.4503350257873535,
"learning_rate": 4.450503355704698e-05,
"loss": 0.8719,
"step": 231
},
{
"epoch": 0.3587166602241979,
"grad_norm": 0.4927978217601776,
"learning_rate": 4.446308724832215e-05,
"loss": 0.9214,
"step": 232
},
{
"epoch": 0.3602628527251643,
"grad_norm": 0.4751432240009308,
"learning_rate": 4.4421140939597314e-05,
"loss": 0.8735,
"step": 233
},
{
"epoch": 0.36180904522613067,
"grad_norm": 0.4933432638645172,
"learning_rate": 4.4379194630872484e-05,
"loss": 0.8353,
"step": 234
},
{
"epoch": 0.36335523772709705,
"grad_norm": 0.5034083127975464,
"learning_rate": 4.4337248322147654e-05,
"loss": 0.9168,
"step": 235
},
{
"epoch": 0.3649014302280634,
"grad_norm": 0.5039856433868408,
"learning_rate": 4.4295302013422824e-05,
"loss": 0.9183,
"step": 236
},
{
"epoch": 0.36644762272902975,
"grad_norm": 0.47999468445777893,
"learning_rate": 4.4253355704697994e-05,
"loss": 0.9402,
"step": 237
},
{
"epoch": 0.36799381522999614,
"grad_norm": 0.5011245608329773,
"learning_rate": 4.421140939597316e-05,
"loss": 0.9587,
"step": 238
},
{
"epoch": 0.3695400077309625,
"grad_norm": 0.49425458908081055,
"learning_rate": 4.416946308724833e-05,
"loss": 0.8474,
"step": 239
},
{
"epoch": 0.3710862002319289,
"grad_norm": 0.514880359172821,
"learning_rate": 4.412751677852349e-05,
"loss": 0.9016,
"step": 240
},
{
"epoch": 0.3726323927328952,
"grad_norm": 0.5200977325439453,
"learning_rate": 4.408557046979866e-05,
"loss": 0.9552,
"step": 241
},
{
"epoch": 0.3741785852338616,
"grad_norm": 0.5420643091201782,
"learning_rate": 4.4043624161073823e-05,
"loss": 0.9023,
"step": 242
},
{
"epoch": 0.375724777734828,
"grad_norm": 0.5034189820289612,
"learning_rate": 4.4001677852348993e-05,
"loss": 0.9255,
"step": 243
},
{
"epoch": 0.37727097023579437,
"grad_norm": 0.529698371887207,
"learning_rate": 4.395973154362416e-05,
"loss": 0.9337,
"step": 244
},
{
"epoch": 0.37881716273676075,
"grad_norm": 0.5248023867607117,
"learning_rate": 4.391778523489933e-05,
"loss": 0.9626,
"step": 245
},
{
"epoch": 0.3803633552377271,
"grad_norm": 0.5809698700904846,
"learning_rate": 4.38758389261745e-05,
"loss": 0.9598,
"step": 246
},
{
"epoch": 0.38190954773869346,
"grad_norm": 0.5663429498672485,
"learning_rate": 4.3833892617449666e-05,
"loss": 0.979,
"step": 247
},
{
"epoch": 0.38345574023965984,
"grad_norm": 0.6147710680961609,
"learning_rate": 4.3791946308724836e-05,
"loss": 0.9747,
"step": 248
},
{
"epoch": 0.3850019327406262,
"grad_norm": 0.7502771615982056,
"learning_rate": 4.375e-05,
"loss": 0.9344,
"step": 249
},
{
"epoch": 0.3865481252415926,
"grad_norm": 1.0911471843719482,
"learning_rate": 4.370805369127517e-05,
"loss": 1.0075,
"step": 250
},
{
"epoch": 0.3880943177425589,
"grad_norm": 0.4232114255428314,
"learning_rate": 4.366610738255034e-05,
"loss": 0.7643,
"step": 251
},
{
"epoch": 0.3896405102435253,
"grad_norm": 0.423627644777298,
"learning_rate": 4.36241610738255e-05,
"loss": 0.7551,
"step": 252
},
{
"epoch": 0.3911867027444917,
"grad_norm": 0.4198389947414398,
"learning_rate": 4.358221476510067e-05,
"loss": 0.7888,
"step": 253
},
{
"epoch": 0.39273289524545807,
"grad_norm": 0.43714022636413574,
"learning_rate": 4.354026845637584e-05,
"loss": 0.7958,
"step": 254
},
{
"epoch": 0.39427908774642445,
"grad_norm": 0.4278389513492584,
"learning_rate": 4.349832214765101e-05,
"loss": 0.7959,
"step": 255
},
{
"epoch": 0.3958252802473908,
"grad_norm": 0.45499542355537415,
"learning_rate": 4.3456375838926176e-05,
"loss": 0.7461,
"step": 256
},
{
"epoch": 0.39737147274835716,
"grad_norm": 0.46723300218582153,
"learning_rate": 4.3414429530201346e-05,
"loss": 0.8312,
"step": 257
},
{
"epoch": 0.39891766524932354,
"grad_norm": 0.44024282693862915,
"learning_rate": 4.337248322147651e-05,
"loss": 0.8517,
"step": 258
},
{
"epoch": 0.4004638577502899,
"grad_norm": 0.42471960186958313,
"learning_rate": 4.333053691275168e-05,
"loss": 0.8312,
"step": 259
},
{
"epoch": 0.4020100502512563,
"grad_norm": 0.47083520889282227,
"learning_rate": 4.328859060402685e-05,
"loss": 0.8618,
"step": 260
},
{
"epoch": 0.4035562427522226,
"grad_norm": 0.45681217312812805,
"learning_rate": 4.324664429530201e-05,
"loss": 0.8986,
"step": 261
},
{
"epoch": 0.405102435253189,
"grad_norm": 0.42533236742019653,
"learning_rate": 4.320469798657718e-05,
"loss": 0.824,
"step": 262
},
{
"epoch": 0.4066486277541554,
"grad_norm": 0.4372871220111847,
"learning_rate": 4.316275167785235e-05,
"loss": 0.8077,
"step": 263
},
{
"epoch": 0.40819482025512177,
"grad_norm": 0.45690760016441345,
"learning_rate": 4.312080536912752e-05,
"loss": 0.8735,
"step": 264
},
{
"epoch": 0.40974101275608815,
"grad_norm": 0.46814388036727905,
"learning_rate": 4.3078859060402685e-05,
"loss": 0.8161,
"step": 265
},
{
"epoch": 0.4112872052570545,
"grad_norm": 0.4442387521266937,
"learning_rate": 4.3036912751677855e-05,
"loss": 0.8352,
"step": 266
},
{
"epoch": 0.41283339775802086,
"grad_norm": 0.4814305305480957,
"learning_rate": 4.2994966442953025e-05,
"loss": 0.7508,
"step": 267
},
{
"epoch": 0.41437959025898724,
"grad_norm": 0.4563637673854828,
"learning_rate": 4.295302013422819e-05,
"loss": 0.8639,
"step": 268
},
{
"epoch": 0.4159257827599536,
"grad_norm": 0.4439164698123932,
"learning_rate": 4.291107382550336e-05,
"loss": 0.8466,
"step": 269
},
{
"epoch": 0.41747197526092,
"grad_norm": 0.45698022842407227,
"learning_rate": 4.286912751677852e-05,
"loss": 0.8784,
"step": 270
},
{
"epoch": 0.41901816776188633,
"grad_norm": 0.46446678042411804,
"learning_rate": 4.28271812080537e-05,
"loss": 0.7856,
"step": 271
},
{
"epoch": 0.4205643602628527,
"grad_norm": 0.4599681496620178,
"learning_rate": 4.278523489932886e-05,
"loss": 0.8241,
"step": 272
},
{
"epoch": 0.4221105527638191,
"grad_norm": 0.4912761449813843,
"learning_rate": 4.274328859060403e-05,
"loss": 0.8697,
"step": 273
},
{
"epoch": 0.4236567452647855,
"grad_norm": 0.4875286817550659,
"learning_rate": 4.27013422818792e-05,
"loss": 0.7999,
"step": 274
},
{
"epoch": 0.42520293776575185,
"grad_norm": 0.4821873903274536,
"learning_rate": 4.2659395973154364e-05,
"loss": 0.7944,
"step": 275
},
{
"epoch": 0.4267491302667182,
"grad_norm": 0.4670597016811371,
"learning_rate": 4.2617449664429534e-05,
"loss": 0.9423,
"step": 276
},
{
"epoch": 0.42829532276768456,
"grad_norm": 0.5004227757453918,
"learning_rate": 4.25755033557047e-05,
"loss": 0.8073,
"step": 277
},
{
"epoch": 0.42984151526865094,
"grad_norm": 0.510180652141571,
"learning_rate": 4.253355704697987e-05,
"loss": 0.8658,
"step": 278
},
{
"epoch": 0.4313877077696173,
"grad_norm": 0.4866536855697632,
"learning_rate": 4.249161073825503e-05,
"loss": 0.7532,
"step": 279
},
{
"epoch": 0.4329339002705837,
"grad_norm": 0.5010313987731934,
"learning_rate": 4.244966442953021e-05,
"loss": 0.9368,
"step": 280
},
{
"epoch": 0.43448009277155003,
"grad_norm": 0.5455735325813293,
"learning_rate": 4.240771812080537e-05,
"loss": 0.9058,
"step": 281
},
{
"epoch": 0.4360262852725164,
"grad_norm": 0.47950518131256104,
"learning_rate": 4.236577181208054e-05,
"loss": 0.8709,
"step": 282
},
{
"epoch": 0.4375724777734828,
"grad_norm": 0.5143994092941284,
"learning_rate": 4.232382550335571e-05,
"loss": 0.8528,
"step": 283
},
{
"epoch": 0.4391186702744492,
"grad_norm": 0.5421144366264343,
"learning_rate": 4.228187919463087e-05,
"loss": 0.8989,
"step": 284
},
{
"epoch": 0.44066486277541556,
"grad_norm": 0.5018422603607178,
"learning_rate": 4.223993288590604e-05,
"loss": 0.8903,
"step": 285
},
{
"epoch": 0.44221105527638194,
"grad_norm": 0.4975266754627228,
"learning_rate": 4.2197986577181206e-05,
"loss": 0.9568,
"step": 286
},
{
"epoch": 0.44375724777734826,
"grad_norm": 0.4982987642288208,
"learning_rate": 4.2156040268456376e-05,
"loss": 0.946,
"step": 287
},
{
"epoch": 0.44530344027831464,
"grad_norm": 0.5241461992263794,
"learning_rate": 4.2114093959731546e-05,
"loss": 0.9359,
"step": 288
},
{
"epoch": 0.446849632779281,
"grad_norm": 0.4898720681667328,
"learning_rate": 4.2072147651006716e-05,
"loss": 0.9798,
"step": 289
},
{
"epoch": 0.4483958252802474,
"grad_norm": 0.5274215936660767,
"learning_rate": 4.2030201342281886e-05,
"loss": 0.8603,
"step": 290
},
{
"epoch": 0.4499420177812138,
"grad_norm": 0.5538789629936218,
"learning_rate": 4.198825503355705e-05,
"loss": 0.8873,
"step": 291
},
{
"epoch": 0.4514882102821801,
"grad_norm": 0.5374253392219543,
"learning_rate": 4.194630872483222e-05,
"loss": 1.0174,
"step": 292
},
{
"epoch": 0.4530344027831465,
"grad_norm": 0.5286028385162354,
"learning_rate": 4.190436241610738e-05,
"loss": 0.9891,
"step": 293
},
{
"epoch": 0.4545805952841129,
"grad_norm": 0.5447548627853394,
"learning_rate": 4.186241610738255e-05,
"loss": 0.9587,
"step": 294
},
{
"epoch": 0.45612678778507926,
"grad_norm": 0.5515534281730652,
"learning_rate": 4.1820469798657716e-05,
"loss": 0.9663,
"step": 295
},
{
"epoch": 0.45767298028604564,
"grad_norm": 0.6086519956588745,
"learning_rate": 4.1778523489932886e-05,
"loss": 1.0357,
"step": 296
},
{
"epoch": 0.45921917278701196,
"grad_norm": 0.6332388520240784,
"learning_rate": 4.1736577181208055e-05,
"loss": 0.9999,
"step": 297
},
{
"epoch": 0.46076536528797835,
"grad_norm": 0.6928420066833496,
"learning_rate": 4.1694630872483225e-05,
"loss": 1.0243,
"step": 298
},
{
"epoch": 0.4623115577889447,
"grad_norm": 0.7226160168647766,
"learning_rate": 4.1652684563758395e-05,
"loss": 0.9209,
"step": 299
},
{
"epoch": 0.4638577502899111,
"grad_norm": 1.0083510875701904,
"learning_rate": 4.161073825503356e-05,
"loss": 0.9493,
"step": 300
},
{
"epoch": 0.4654039427908775,
"grad_norm": 0.4649132192134857,
"learning_rate": 4.156879194630873e-05,
"loss": 0.7743,
"step": 301
},
{
"epoch": 0.4669501352918438,
"grad_norm": 0.43430233001708984,
"learning_rate": 4.152684563758389e-05,
"loss": 0.8197,
"step": 302
},
{
"epoch": 0.4684963277928102,
"grad_norm": 0.44487112760543823,
"learning_rate": 4.148489932885906e-05,
"loss": 0.7632,
"step": 303
},
{
"epoch": 0.4700425202937766,
"grad_norm": 0.4654642343521118,
"learning_rate": 4.144295302013423e-05,
"loss": 0.7637,
"step": 304
},
{
"epoch": 0.47158871279474296,
"grad_norm": 0.48062166571617126,
"learning_rate": 4.1401006711409395e-05,
"loss": 0.7771,
"step": 305
},
{
"epoch": 0.47313490529570934,
"grad_norm": 0.4307418167591095,
"learning_rate": 4.135906040268457e-05,
"loss": 0.8336,
"step": 306
},
{
"epoch": 0.47468109779667567,
"grad_norm": 0.4675520658493042,
"learning_rate": 4.1317114093959735e-05,
"loss": 0.7969,
"step": 307
},
{
"epoch": 0.47622729029764205,
"grad_norm": 0.48613134026527405,
"learning_rate": 4.1275167785234905e-05,
"loss": 0.8821,
"step": 308
},
{
"epoch": 0.47777348279860843,
"grad_norm": 0.46094274520874023,
"learning_rate": 4.123322147651007e-05,
"loss": 0.8103,
"step": 309
},
{
"epoch": 0.4793196752995748,
"grad_norm": 0.47018101811408997,
"learning_rate": 4.119127516778524e-05,
"loss": 0.8515,
"step": 310
},
{
"epoch": 0.4808658678005412,
"grad_norm": 0.43754515051841736,
"learning_rate": 4.11493288590604e-05,
"loss": 0.8218,
"step": 311
},
{
"epoch": 0.4824120603015075,
"grad_norm": 0.49217140674591064,
"learning_rate": 4.110738255033557e-05,
"loss": 0.7866,
"step": 312
},
{
"epoch": 0.4839582528024739,
"grad_norm": 0.47612541913986206,
"learning_rate": 4.106543624161074e-05,
"loss": 0.8088,
"step": 313
},
{
"epoch": 0.4855044453034403,
"grad_norm": 0.475328654050827,
"learning_rate": 4.1023489932885904e-05,
"loss": 0.8165,
"step": 314
},
{
"epoch": 0.48705063780440666,
"grad_norm": 0.4439482092857361,
"learning_rate": 4.098154362416108e-05,
"loss": 0.7833,
"step": 315
},
{
"epoch": 0.48859683030537304,
"grad_norm": 0.5019133687019348,
"learning_rate": 4.0939597315436244e-05,
"loss": 0.7942,
"step": 316
},
{
"epoch": 0.49014302280633937,
"grad_norm": 0.451492577791214,
"learning_rate": 4.0897651006711414e-05,
"loss": 0.8039,
"step": 317
},
{
"epoch": 0.49168921530730575,
"grad_norm": 0.46894291043281555,
"learning_rate": 4.085570469798658e-05,
"loss": 0.843,
"step": 318
},
{
"epoch": 0.49323540780827213,
"grad_norm": 0.49084803462028503,
"learning_rate": 4.081375838926175e-05,
"loss": 0.8121,
"step": 319
},
{
"epoch": 0.4947816003092385,
"grad_norm": 0.46862879395484924,
"learning_rate": 4.077181208053692e-05,
"loss": 0.901,
"step": 320
},
{
"epoch": 0.4963277928102049,
"grad_norm": 0.4881606698036194,
"learning_rate": 4.072986577181208e-05,
"loss": 0.9001,
"step": 321
},
{
"epoch": 0.4978739853111712,
"grad_norm": 0.5040379166603088,
"learning_rate": 4.068791946308725e-05,
"loss": 0.801,
"step": 322
},
{
"epoch": 0.4994201778121376,
"grad_norm": 0.46498745679855347,
"learning_rate": 4.064597315436241e-05,
"loss": 0.8473,
"step": 323
},
{
"epoch": 0.500966370313104,
"grad_norm": 0.4876708388328552,
"learning_rate": 4.060402684563759e-05,
"loss": 0.8418,
"step": 324
},
{
"epoch": 0.5025125628140703,
"grad_norm": 0.4962271451950073,
"learning_rate": 4.056208053691275e-05,
"loss": 0.7611,
"step": 325
},
{
"epoch": 0.5040587553150367,
"grad_norm": 0.48774364590644836,
"learning_rate": 4.052013422818792e-05,
"loss": 0.8825,
"step": 326
},
{
"epoch": 0.5056049478160031,
"grad_norm": 0.5011964440345764,
"learning_rate": 4.047818791946309e-05,
"loss": 0.7659,
"step": 327
},
{
"epoch": 0.5071511403169695,
"grad_norm": 0.5307598114013672,
"learning_rate": 4.0436241610738256e-05,
"loss": 0.8629,
"step": 328
},
{
"epoch": 0.5086973328179358,
"grad_norm": 0.49158573150634766,
"learning_rate": 4.0394295302013426e-05,
"loss": 0.9071,
"step": 329
},
{
"epoch": 0.5102435253189022,
"grad_norm": 0.5126063227653503,
"learning_rate": 4.035234899328859e-05,
"loss": 0.7998,
"step": 330
},
{
"epoch": 0.5117897178198686,
"grad_norm": 0.5247028470039368,
"learning_rate": 4.031040268456376e-05,
"loss": 0.8443,
"step": 331
},
{
"epoch": 0.5133359103208349,
"grad_norm": 0.5511295199394226,
"learning_rate": 4.026845637583892e-05,
"loss": 0.8946,
"step": 332
},
{
"epoch": 0.5148821028218014,
"grad_norm": 0.5266992449760437,
"learning_rate": 4.02265100671141e-05,
"loss": 0.8172,
"step": 333
},
{
"epoch": 0.5164282953227677,
"grad_norm": 0.5326645374298096,
"learning_rate": 4.018456375838926e-05,
"loss": 0.8407,
"step": 334
},
{
"epoch": 0.517974487823734,
"grad_norm": 0.5065395832061768,
"learning_rate": 4.014261744966443e-05,
"loss": 0.8751,
"step": 335
},
{
"epoch": 0.5195206803247004,
"grad_norm": 0.53061443567276,
"learning_rate": 4.01006711409396e-05,
"loss": 0.9204,
"step": 336
},
{
"epoch": 0.5210668728256668,
"grad_norm": 0.5223432779312134,
"learning_rate": 4.0058724832214765e-05,
"loss": 0.8933,
"step": 337
},
{
"epoch": 0.5226130653266332,
"grad_norm": 0.5290891528129578,
"learning_rate": 4.0016778523489935e-05,
"loss": 0.977,
"step": 338
},
{
"epoch": 0.5241592578275995,
"grad_norm": 0.5419663786888123,
"learning_rate": 3.99748322147651e-05,
"loss": 0.9521,
"step": 339
},
{
"epoch": 0.5257054503285659,
"grad_norm": 0.5404860973358154,
"learning_rate": 3.993288590604027e-05,
"loss": 0.8907,
"step": 340
},
{
"epoch": 0.5272516428295323,
"grad_norm": 0.5783061385154724,
"learning_rate": 3.989093959731544e-05,
"loss": 0.9162,
"step": 341
},
{
"epoch": 0.5287978353304986,
"grad_norm": 0.5513641834259033,
"learning_rate": 3.984899328859061e-05,
"loss": 0.8912,
"step": 342
},
{
"epoch": 0.5303440278314651,
"grad_norm": 0.5670628547668457,
"learning_rate": 3.980704697986578e-05,
"loss": 0.9502,
"step": 343
},
{
"epoch": 0.5318902203324314,
"grad_norm": 0.5960517525672913,
"learning_rate": 3.976510067114094e-05,
"loss": 0.9976,
"step": 344
},
{
"epoch": 0.5334364128333977,
"grad_norm": 0.6074213981628418,
"learning_rate": 3.972315436241611e-05,
"loss": 0.9861,
"step": 345
},
{
"epoch": 0.5349826053343641,
"grad_norm": 0.6100422143936157,
"learning_rate": 3.9681208053691275e-05,
"loss": 0.9483,
"step": 346
},
{
"epoch": 0.5365287978353305,
"grad_norm": 0.6308846473693848,
"learning_rate": 3.9639261744966445e-05,
"loss": 0.9455,
"step": 347
},
{
"epoch": 0.5380749903362969,
"grad_norm": 0.6239264607429504,
"learning_rate": 3.959731543624161e-05,
"loss": 0.9551,
"step": 348
},
{
"epoch": 0.5396211828372632,
"grad_norm": 0.6669812798500061,
"learning_rate": 3.955536912751678e-05,
"loss": 0.9251,
"step": 349
},
{
"epoch": 0.5411673753382296,
"grad_norm": 0.8175333738327026,
"learning_rate": 3.951342281879195e-05,
"loss": 0.9795,
"step": 350
},
{
"epoch": 0.542713567839196,
"grad_norm": 0.44942405819892883,
"learning_rate": 3.947147651006712e-05,
"loss": 0.7184,
"step": 351
},
{
"epoch": 0.5442597603401623,
"grad_norm": 0.46283286809921265,
"learning_rate": 3.942953020134229e-05,
"loss": 0.7251,
"step": 352
},
{
"epoch": 0.5458059528411288,
"grad_norm": 0.47923749685287476,
"learning_rate": 3.938758389261745e-05,
"loss": 0.766,
"step": 353
},
{
"epoch": 0.5473521453420951,
"grad_norm": 0.4575538635253906,
"learning_rate": 3.934563758389262e-05,
"loss": 0.7539,
"step": 354
},
{
"epoch": 0.5488983378430614,
"grad_norm": 0.48839524388313293,
"learning_rate": 3.9303691275167784e-05,
"loss": 0.7633,
"step": 355
},
{
"epoch": 0.5504445303440278,
"grad_norm": 0.45980286598205566,
"learning_rate": 3.9261744966442954e-05,
"loss": 0.774,
"step": 356
},
{
"epoch": 0.5519907228449942,
"grad_norm": 0.4767687916755676,
"learning_rate": 3.9219798657718124e-05,
"loss": 0.7588,
"step": 357
},
{
"epoch": 0.5535369153459606,
"grad_norm": 0.4907895028591156,
"learning_rate": 3.917785234899329e-05,
"loss": 0.7305,
"step": 358
},
{
"epoch": 0.5550831078469269,
"grad_norm": 0.46510565280914307,
"learning_rate": 3.9135906040268464e-05,
"loss": 0.8727,
"step": 359
},
{
"epoch": 0.5566293003478933,
"grad_norm": 0.4917638599872589,
"learning_rate": 3.909395973154363e-05,
"loss": 0.8013,
"step": 360
},
{
"epoch": 0.5581754928488597,
"grad_norm": 0.49160364270210266,
"learning_rate": 3.90520134228188e-05,
"loss": 0.8871,
"step": 361
},
{
"epoch": 0.559721685349826,
"grad_norm": 0.5044489502906799,
"learning_rate": 3.901006711409396e-05,
"loss": 0.7945,
"step": 362
},
{
"epoch": 0.5612678778507925,
"grad_norm": 0.49622687697410583,
"learning_rate": 3.896812080536913e-05,
"loss": 0.7616,
"step": 363
},
{
"epoch": 0.5628140703517588,
"grad_norm": 0.5114305019378662,
"learning_rate": 3.89261744966443e-05,
"loss": 0.7593,
"step": 364
},
{
"epoch": 0.5643602628527251,
"grad_norm": 0.499931663274765,
"learning_rate": 3.888422818791946e-05,
"loss": 0.7935,
"step": 365
},
{
"epoch": 0.5659064553536916,
"grad_norm": 0.511566698551178,
"learning_rate": 3.884228187919463e-05,
"loss": 0.7665,
"step": 366
},
{
"epoch": 0.5674526478546579,
"grad_norm": 0.49260082840919495,
"learning_rate": 3.8800335570469796e-05,
"loss": 0.8207,
"step": 367
},
{
"epoch": 0.5689988403556243,
"grad_norm": 0.48072633147239685,
"learning_rate": 3.875838926174497e-05,
"loss": 0.7739,
"step": 368
},
{
"epoch": 0.5705450328565906,
"grad_norm": 0.4969918727874756,
"learning_rate": 3.8716442953020136e-05,
"loss": 0.8155,
"step": 369
},
{
"epoch": 0.572091225357557,
"grad_norm": 0.49780312180519104,
"learning_rate": 3.8674496644295306e-05,
"loss": 0.8397,
"step": 370
},
{
"epoch": 0.5736374178585234,
"grad_norm": 0.5126276612281799,
"learning_rate": 3.863255033557047e-05,
"loss": 0.8101,
"step": 371
},
{
"epoch": 0.5751836103594897,
"grad_norm": 0.48839470744132996,
"learning_rate": 3.859060402684564e-05,
"loss": 0.7629,
"step": 372
},
{
"epoch": 0.5767298028604562,
"grad_norm": 0.4904743432998657,
"learning_rate": 3.854865771812081e-05,
"loss": 0.8126,
"step": 373
},
{
"epoch": 0.5782759953614225,
"grad_norm": 0.48648014664649963,
"learning_rate": 3.850671140939597e-05,
"loss": 0.845,
"step": 374
},
{
"epoch": 0.5798221878623888,
"grad_norm": 0.537968635559082,
"learning_rate": 3.846476510067114e-05,
"loss": 0.863,
"step": 375
},
{
"epoch": 0.5813683803633553,
"grad_norm": 0.5034117102622986,
"learning_rate": 3.8422818791946305e-05,
"loss": 0.8333,
"step": 376
},
{
"epoch": 0.5829145728643216,
"grad_norm": 0.529244601726532,
"learning_rate": 3.838087248322148e-05,
"loss": 0.7756,
"step": 377
},
{
"epoch": 0.584460765365288,
"grad_norm": 0.5105958580970764,
"learning_rate": 3.8338926174496645e-05,
"loss": 0.8372,
"step": 378
},
{
"epoch": 0.5860069578662543,
"grad_norm": 0.5209059715270996,
"learning_rate": 3.8296979865771815e-05,
"loss": 0.8347,
"step": 379
},
{
"epoch": 0.5875531503672207,
"grad_norm": 0.5022969841957092,
"learning_rate": 3.8255033557046985e-05,
"loss": 0.8657,
"step": 380
},
{
"epoch": 0.5890993428681871,
"grad_norm": 0.5265902876853943,
"learning_rate": 3.821308724832215e-05,
"loss": 0.8333,
"step": 381
},
{
"epoch": 0.5906455353691534,
"grad_norm": 0.5082066655158997,
"learning_rate": 3.817114093959732e-05,
"loss": 0.8834,
"step": 382
},
{
"epoch": 0.5921917278701199,
"grad_norm": 0.5264511704444885,
"learning_rate": 3.812919463087248e-05,
"loss": 0.908,
"step": 383
},
{
"epoch": 0.5937379203710862,
"grad_norm": 0.5576880574226379,
"learning_rate": 3.808724832214765e-05,
"loss": 0.9174,
"step": 384
},
{
"epoch": 0.5952841128720525,
"grad_norm": 0.5170852541923523,
"learning_rate": 3.804530201342282e-05,
"loss": 0.8983,
"step": 385
},
{
"epoch": 0.596830305373019,
"grad_norm": 0.5212023854255676,
"learning_rate": 3.800335570469799e-05,
"loss": 0.8684,
"step": 386
},
{
"epoch": 0.5983764978739853,
"grad_norm": 0.5252931714057922,
"learning_rate": 3.7961409395973154e-05,
"loss": 0.9016,
"step": 387
},
{
"epoch": 0.5999226903749517,
"grad_norm": 0.5745819807052612,
"learning_rate": 3.7919463087248324e-05,
"loss": 0.8521,
"step": 388
},
{
"epoch": 0.601468882875918,
"grad_norm": 0.5259801149368286,
"learning_rate": 3.7877516778523494e-05,
"loss": 0.9101,
"step": 389
},
{
"epoch": 0.6030150753768844,
"grad_norm": 0.5453861355781555,
"learning_rate": 3.783557046979866e-05,
"loss": 0.8816,
"step": 390
},
{
"epoch": 0.6045612678778508,
"grad_norm": 0.5567570328712463,
"learning_rate": 3.779362416107383e-05,
"loss": 0.9126,
"step": 391
},
{
"epoch": 0.6061074603788171,
"grad_norm": 0.5654810070991516,
"learning_rate": 3.775167785234899e-05,
"loss": 0.8147,
"step": 392
},
{
"epoch": 0.6076536528797836,
"grad_norm": 0.5371072888374329,
"learning_rate": 3.770973154362416e-05,
"loss": 0.936,
"step": 393
},
{
"epoch": 0.6091998453807499,
"grad_norm": 0.6840121746063232,
"learning_rate": 3.766778523489933e-05,
"loss": 1.0121,
"step": 394
},
{
"epoch": 0.6107460378817162,
"grad_norm": 0.6007458567619324,
"learning_rate": 3.76258389261745e-05,
"loss": 0.895,
"step": 395
},
{
"epoch": 0.6122922303826827,
"grad_norm": 0.6644576191902161,
"learning_rate": 3.758389261744967e-05,
"loss": 0.9978,
"step": 396
},
{
"epoch": 0.613838422883649,
"grad_norm": 0.6520909667015076,
"learning_rate": 3.7541946308724834e-05,
"loss": 0.9551,
"step": 397
},
{
"epoch": 0.6153846153846154,
"grad_norm": 0.66536945104599,
"learning_rate": 3.7500000000000003e-05,
"loss": 0.9551,
"step": 398
},
{
"epoch": 0.6169308078855817,
"grad_norm": 0.7479529976844788,
"learning_rate": 3.745805369127517e-05,
"loss": 1.022,
"step": 399
},
{
"epoch": 0.6184770003865481,
"grad_norm": 1.1065740585327148,
"learning_rate": 3.741610738255034e-05,
"loss": 0.9431,
"step": 400
},
{
"epoch": 0.6200231928875145,
"grad_norm": 0.4878085255622864,
"learning_rate": 3.7374161073825507e-05,
"loss": 0.728,
"step": 401
},
{
"epoch": 0.6215693853884808,
"grad_norm": 0.4753767251968384,
"learning_rate": 3.733221476510067e-05,
"loss": 0.7575,
"step": 402
},
{
"epoch": 0.6231155778894473,
"grad_norm": 0.4956800043582916,
"learning_rate": 3.7290268456375846e-05,
"loss": 0.7292,
"step": 403
},
{
"epoch": 0.6246617703904136,
"grad_norm": 0.48405370116233826,
"learning_rate": 3.724832214765101e-05,
"loss": 0.7879,
"step": 404
},
{
"epoch": 0.62620796289138,
"grad_norm": 0.46832674741744995,
"learning_rate": 3.720637583892618e-05,
"loss": 0.7819,
"step": 405
},
{
"epoch": 0.6277541553923464,
"grad_norm": 0.4691636264324188,
"learning_rate": 3.716442953020134e-05,
"loss": 0.7984,
"step": 406
},
{
"epoch": 0.6293003478933127,
"grad_norm": 0.5165214538574219,
"learning_rate": 3.712248322147651e-05,
"loss": 0.7296,
"step": 407
},
{
"epoch": 0.6308465403942791,
"grad_norm": 0.5001758933067322,
"learning_rate": 3.7080536912751676e-05,
"loss": 0.808,
"step": 408
},
{
"epoch": 0.6323927328952454,
"grad_norm": 0.4783785045146942,
"learning_rate": 3.7038590604026846e-05,
"loss": 0.7856,
"step": 409
},
{
"epoch": 0.6339389253962119,
"grad_norm": 0.5511770248413086,
"learning_rate": 3.6996644295302016e-05,
"loss": 0.711,
"step": 410
},
{
"epoch": 0.6354851178971782,
"grad_norm": 0.5434517860412598,
"learning_rate": 3.695469798657718e-05,
"loss": 0.8498,
"step": 411
},
{
"epoch": 0.6370313103981445,
"grad_norm": 0.5018695592880249,
"learning_rate": 3.6912751677852356e-05,
"loss": 0.7898,
"step": 412
},
{
"epoch": 0.638577502899111,
"grad_norm": 0.5000431537628174,
"learning_rate": 3.687080536912752e-05,
"loss": 0.7803,
"step": 413
},
{
"epoch": 0.6401236954000773,
"grad_norm": 0.5340814590454102,
"learning_rate": 3.682885906040269e-05,
"loss": 0.7671,
"step": 414
},
{
"epoch": 0.6416698879010437,
"grad_norm": 0.5046684741973877,
"learning_rate": 3.678691275167785e-05,
"loss": 0.7876,
"step": 415
},
{
"epoch": 0.6432160804020101,
"grad_norm": 0.5036927461624146,
"learning_rate": 3.674496644295302e-05,
"loss": 0.8726,
"step": 416
},
{
"epoch": 0.6447622729029764,
"grad_norm": 0.5092120170593262,
"learning_rate": 3.670302013422819e-05,
"loss": 0.806,
"step": 417
},
{
"epoch": 0.6463084654039428,
"grad_norm": 0.5002840757369995,
"learning_rate": 3.6661073825503355e-05,
"loss": 0.8814,
"step": 418
},
{
"epoch": 0.6478546579049091,
"grad_norm": 0.5107703804969788,
"learning_rate": 3.6619127516778525e-05,
"loss": 0.7984,
"step": 419
},
{
"epoch": 0.6494008504058756,
"grad_norm": 0.5407206416130066,
"learning_rate": 3.6577181208053695e-05,
"loss": 0.7979,
"step": 420
},
{
"epoch": 0.6509470429068419,
"grad_norm": 0.5402006506919861,
"learning_rate": 3.6535234899328865e-05,
"loss": 0.7547,
"step": 421
},
{
"epoch": 0.6524932354078082,
"grad_norm": 0.5221443176269531,
"learning_rate": 3.649328859060403e-05,
"loss": 0.8426,
"step": 422
},
{
"epoch": 0.6540394279087747,
"grad_norm": 0.5319470167160034,
"learning_rate": 3.64513422818792e-05,
"loss": 0.7866,
"step": 423
},
{
"epoch": 0.655585620409741,
"grad_norm": 0.5060355067253113,
"learning_rate": 3.640939597315436e-05,
"loss": 0.8102,
"step": 424
},
{
"epoch": 0.6571318129107074,
"grad_norm": 0.5067129731178284,
"learning_rate": 3.636744966442953e-05,
"loss": 0.8316,
"step": 425
},
{
"epoch": 0.6586780054116738,
"grad_norm": 0.5815935730934143,
"learning_rate": 3.63255033557047e-05,
"loss": 0.7207,
"step": 426
},
{
"epoch": 0.6602241979126401,
"grad_norm": 0.518981397151947,
"learning_rate": 3.6283557046979864e-05,
"loss": 0.7909,
"step": 427
},
{
"epoch": 0.6617703904136065,
"grad_norm": 0.5797117948532104,
"learning_rate": 3.6241610738255034e-05,
"loss": 0.8346,
"step": 428
},
{
"epoch": 0.6633165829145728,
"grad_norm": 0.565744936466217,
"learning_rate": 3.6199664429530204e-05,
"loss": 0.8366,
"step": 429
},
{
"epoch": 0.6648627754155393,
"grad_norm": 0.5538972020149231,
"learning_rate": 3.6157718120805374e-05,
"loss": 0.8346,
"step": 430
},
{
"epoch": 0.6664089679165056,
"grad_norm": 0.561184823513031,
"learning_rate": 3.611577181208054e-05,
"loss": 0.8389,
"step": 431
},
{
"epoch": 0.6679551604174719,
"grad_norm": 0.5446305871009827,
"learning_rate": 3.607382550335571e-05,
"loss": 0.7981,
"step": 432
},
{
"epoch": 0.6695013529184384,
"grad_norm": 0.6099398136138916,
"learning_rate": 3.603187919463088e-05,
"loss": 0.8988,
"step": 433
},
{
"epoch": 0.6710475454194047,
"grad_norm": 0.572850227355957,
"learning_rate": 3.598993288590604e-05,
"loss": 0.8981,
"step": 434
},
{
"epoch": 0.6725937379203711,
"grad_norm": 0.5837070941925049,
"learning_rate": 3.594798657718121e-05,
"loss": 0.9049,
"step": 435
},
{
"epoch": 0.6741399304213375,
"grad_norm": 0.5394341945648193,
"learning_rate": 3.5906040268456373e-05,
"loss": 0.8744,
"step": 436
},
{
"epoch": 0.6756861229223038,
"grad_norm": 0.5494775772094727,
"learning_rate": 3.5864093959731543e-05,
"loss": 0.9612,
"step": 437
},
{
"epoch": 0.6772323154232702,
"grad_norm": 0.5773348808288574,
"learning_rate": 3.582214765100671e-05,
"loss": 0.9005,
"step": 438
},
{
"epoch": 0.6787785079242366,
"grad_norm": 0.5757828950881958,
"learning_rate": 3.578020134228188e-05,
"loss": 0.9121,
"step": 439
},
{
"epoch": 0.680324700425203,
"grad_norm": 0.6010167598724365,
"learning_rate": 3.5738255033557046e-05,
"loss": 0.8575,
"step": 440
},
{
"epoch": 0.6818708929261693,
"grad_norm": 0.621435284614563,
"learning_rate": 3.5696308724832216e-05,
"loss": 0.9356,
"step": 441
},
{
"epoch": 0.6834170854271356,
"grad_norm": 0.5952706933021545,
"learning_rate": 3.5654362416107386e-05,
"loss": 0.9653,
"step": 442
},
{
"epoch": 0.6849632779281021,
"grad_norm": 0.59757000207901,
"learning_rate": 3.561241610738255e-05,
"loss": 0.9147,
"step": 443
},
{
"epoch": 0.6865094704290684,
"grad_norm": 0.6352919936180115,
"learning_rate": 3.557046979865772e-05,
"loss": 0.8931,
"step": 444
},
{
"epoch": 0.6880556629300348,
"grad_norm": 0.5814509987831116,
"learning_rate": 3.552852348993288e-05,
"loss": 0.9158,
"step": 445
},
{
"epoch": 0.6896018554310012,
"grad_norm": 0.7028055787086487,
"learning_rate": 3.548657718120805e-05,
"loss": 0.8891,
"step": 446
},
{
"epoch": 0.6911480479319675,
"grad_norm": 0.659831166267395,
"learning_rate": 3.544463087248322e-05,
"loss": 0.9629,
"step": 447
},
{
"epoch": 0.6926942404329339,
"grad_norm": 0.7297990918159485,
"learning_rate": 3.540268456375839e-05,
"loss": 0.9856,
"step": 448
},
{
"epoch": 0.6942404329339003,
"grad_norm": 0.8682158589363098,
"learning_rate": 3.536073825503356e-05,
"loss": 1.0006,
"step": 449
},
{
"epoch": 0.6957866254348667,
"grad_norm": 1.2786856889724731,
"learning_rate": 3.5318791946308726e-05,
"loss": 0.9672,
"step": 450
},
{
"epoch": 0.697332817935833,
"grad_norm": 0.5570630431175232,
"learning_rate": 3.5276845637583896e-05,
"loss": 0.6907,
"step": 451
},
{
"epoch": 0.6988790104367993,
"grad_norm": 0.4869362413883209,
"learning_rate": 3.523489932885906e-05,
"loss": 0.7276,
"step": 452
},
{
"epoch": 0.7004252029377658,
"grad_norm": 0.4926219582557678,
"learning_rate": 3.519295302013423e-05,
"loss": 0.7988,
"step": 453
},
{
"epoch": 0.7019713954387321,
"grad_norm": 0.49141383171081543,
"learning_rate": 3.51510067114094e-05,
"loss": 0.7625,
"step": 454
},
{
"epoch": 0.7035175879396985,
"grad_norm": 0.48099425435066223,
"learning_rate": 3.510906040268457e-05,
"loss": 0.7932,
"step": 455
},
{
"epoch": 0.7050637804406649,
"grad_norm": 0.48948779702186584,
"learning_rate": 3.506711409395974e-05,
"loss": 0.725,
"step": 456
},
{
"epoch": 0.7066099729416312,
"grad_norm": 0.5123636722564697,
"learning_rate": 3.50251677852349e-05,
"loss": 0.7704,
"step": 457
},
{
"epoch": 0.7081561654425976,
"grad_norm": 0.49425482749938965,
"learning_rate": 3.498322147651007e-05,
"loss": 0.8,
"step": 458
},
{
"epoch": 0.709702357943564,
"grad_norm": 0.5385423898696899,
"learning_rate": 3.4941275167785235e-05,
"loss": 0.8093,
"step": 459
},
{
"epoch": 0.7112485504445304,
"grad_norm": 0.5466816425323486,
"learning_rate": 3.4899328859060405e-05,
"loss": 0.7991,
"step": 460
},
{
"epoch": 0.7127947429454967,
"grad_norm": 0.5069779753684998,
"learning_rate": 3.485738255033557e-05,
"loss": 0.7989,
"step": 461
},
{
"epoch": 0.714340935446463,
"grad_norm": 0.5083027482032776,
"learning_rate": 3.481543624161074e-05,
"loss": 0.7891,
"step": 462
},
{
"epoch": 0.7158871279474295,
"grad_norm": 0.5162233710289001,
"learning_rate": 3.477348993288591e-05,
"loss": 0.7208,
"step": 463
},
{
"epoch": 0.7174333204483958,
"grad_norm": 0.5323002338409424,
"learning_rate": 3.473154362416108e-05,
"loss": 0.7471,
"step": 464
},
{
"epoch": 0.7189795129493622,
"grad_norm": 0.5144820809364319,
"learning_rate": 3.468959731543625e-05,
"loss": 0.8335,
"step": 465
},
{
"epoch": 0.7205257054503286,
"grad_norm": 0.5344799160957336,
"learning_rate": 3.464765100671141e-05,
"loss": 0.7575,
"step": 466
},
{
"epoch": 0.7220718979512949,
"grad_norm": 0.5163140296936035,
"learning_rate": 3.460570469798658e-05,
"loss": 0.7471,
"step": 467
},
{
"epoch": 0.7236180904522613,
"grad_norm": 0.537284791469574,
"learning_rate": 3.4563758389261744e-05,
"loss": 0.8013,
"step": 468
},
{
"epoch": 0.7251642829532277,
"grad_norm": 0.5194945335388184,
"learning_rate": 3.4521812080536914e-05,
"loss": 0.7855,
"step": 469
},
{
"epoch": 0.7267104754541941,
"grad_norm": 0.5371651649475098,
"learning_rate": 3.4479865771812084e-05,
"loss": 0.7345,
"step": 470
},
{
"epoch": 0.7282566679551604,
"grad_norm": 0.5294081568717957,
"learning_rate": 3.443791946308725e-05,
"loss": 0.854,
"step": 471
},
{
"epoch": 0.7298028604561267,
"grad_norm": 0.5522701144218445,
"learning_rate": 3.439597315436242e-05,
"loss": 0.8338,
"step": 472
},
{
"epoch": 0.7313490529570932,
"grad_norm": 0.5342947244644165,
"learning_rate": 3.435402684563759e-05,
"loss": 0.8429,
"step": 473
},
{
"epoch": 0.7328952454580595,
"grad_norm": 0.5579586029052734,
"learning_rate": 3.431208053691276e-05,
"loss": 0.77,
"step": 474
},
{
"epoch": 0.734441437959026,
"grad_norm": 0.5210466384887695,
"learning_rate": 3.427013422818792e-05,
"loss": 0.8538,
"step": 475
},
{
"epoch": 0.7359876304599923,
"grad_norm": 0.5218535661697388,
"learning_rate": 3.422818791946309e-05,
"loss": 0.8771,
"step": 476
},
{
"epoch": 0.7375338229609586,
"grad_norm": 0.5211417078971863,
"learning_rate": 3.418624161073825e-05,
"loss": 0.827,
"step": 477
},
{
"epoch": 0.739080015461925,
"grad_norm": 0.5523113012313843,
"learning_rate": 3.414429530201342e-05,
"loss": 0.8367,
"step": 478
},
{
"epoch": 0.7406262079628914,
"grad_norm": 0.5218014717102051,
"learning_rate": 3.410234899328859e-05,
"loss": 0.8319,
"step": 479
},
{
"epoch": 0.7421724004638578,
"grad_norm": 0.5514734387397766,
"learning_rate": 3.4060402684563756e-05,
"loss": 0.8703,
"step": 480
},
{
"epoch": 0.7437185929648241,
"grad_norm": 0.5442553162574768,
"learning_rate": 3.4018456375838926e-05,
"loss": 0.8094,
"step": 481
},
{
"epoch": 0.7452647854657904,
"grad_norm": 0.5472526550292969,
"learning_rate": 3.3976510067114096e-05,
"loss": 0.8419,
"step": 482
},
{
"epoch": 0.7468109779667569,
"grad_norm": 0.5330756306648254,
"learning_rate": 3.3934563758389266e-05,
"loss": 0.8497,
"step": 483
},
{
"epoch": 0.7483571704677232,
"grad_norm": 0.5805166363716125,
"learning_rate": 3.389261744966443e-05,
"loss": 0.8505,
"step": 484
},
{
"epoch": 0.7499033629686896,
"grad_norm": 0.5910758376121521,
"learning_rate": 3.38506711409396e-05,
"loss": 0.8154,
"step": 485
},
{
"epoch": 0.751449555469656,
"grad_norm": 0.579165518283844,
"learning_rate": 3.380872483221477e-05,
"loss": 0.8398,
"step": 486
},
{
"epoch": 0.7529957479706223,
"grad_norm": 0.5517799258232117,
"learning_rate": 3.376677852348993e-05,
"loss": 0.8426,
"step": 487
},
{
"epoch": 0.7545419404715887,
"grad_norm": 0.588138997554779,
"learning_rate": 3.37248322147651e-05,
"loss": 0.8658,
"step": 488
},
{
"epoch": 0.7560881329725551,
"grad_norm": 0.5962609648704529,
"learning_rate": 3.3682885906040266e-05,
"loss": 0.9157,
"step": 489
},
{
"epoch": 0.7576343254735215,
"grad_norm": 0.5821203589439392,
"learning_rate": 3.3640939597315436e-05,
"loss": 0.9158,
"step": 490
},
{
"epoch": 0.7591805179744878,
"grad_norm": 0.5653342008590698,
"learning_rate": 3.3598993288590605e-05,
"loss": 0.924,
"step": 491
},
{
"epoch": 0.7607267104754541,
"grad_norm": 0.6114529967308044,
"learning_rate": 3.3557046979865775e-05,
"loss": 0.9249,
"step": 492
},
{
"epoch": 0.7622729029764206,
"grad_norm": 0.6291983723640442,
"learning_rate": 3.3515100671140945e-05,
"loss": 0.8665,
"step": 493
},
{
"epoch": 0.7638190954773869,
"grad_norm": 0.6541888117790222,
"learning_rate": 3.347315436241611e-05,
"loss": 0.9678,
"step": 494
},
{
"epoch": 0.7653652879783533,
"grad_norm": 0.6252798438072205,
"learning_rate": 3.343120805369128e-05,
"loss": 0.8691,
"step": 495
},
{
"epoch": 0.7669114804793197,
"grad_norm": 0.6714550852775574,
"learning_rate": 3.338926174496644e-05,
"loss": 0.9581,
"step": 496
},
{
"epoch": 0.768457672980286,
"grad_norm": 0.7336750626564026,
"learning_rate": 3.334731543624161e-05,
"loss": 0.9024,
"step": 497
},
{
"epoch": 0.7700038654812524,
"grad_norm": 0.7459555864334106,
"learning_rate": 3.3305369127516775e-05,
"loss": 0.9153,
"step": 498
},
{
"epoch": 0.7715500579822188,
"grad_norm": 0.7960460186004639,
"learning_rate": 3.326342281879195e-05,
"loss": 0.9776,
"step": 499
},
{
"epoch": 0.7730962504831852,
"grad_norm": 1.188610315322876,
"learning_rate": 3.3221476510067115e-05,
"loss": 0.9963,
"step": 500
},
{
"epoch": 0.7746424429841515,
"grad_norm": 0.5373630523681641,
"learning_rate": 3.3179530201342285e-05,
"loss": 0.7497,
"step": 501
},
{
"epoch": 0.7761886354851179,
"grad_norm": 0.5297138094902039,
"learning_rate": 3.3137583892617455e-05,
"loss": 0.6842,
"step": 502
},
{
"epoch": 0.7777348279860843,
"grad_norm": 0.5116965770721436,
"learning_rate": 3.309563758389262e-05,
"loss": 0.7081,
"step": 503
},
{
"epoch": 0.7792810204870506,
"grad_norm": 0.523707389831543,
"learning_rate": 3.305369127516779e-05,
"loss": 0.7354,
"step": 504
},
{
"epoch": 0.780827212988017,
"grad_norm": 0.5127305388450623,
"learning_rate": 3.301174496644295e-05,
"loss": 0.7478,
"step": 505
},
{
"epoch": 0.7823734054889834,
"grad_norm": 0.5249508619308472,
"learning_rate": 3.296979865771812e-05,
"loss": 0.7198,
"step": 506
},
{
"epoch": 0.7839195979899497,
"grad_norm": 0.4998358190059662,
"learning_rate": 3.292785234899329e-05,
"loss": 0.7726,
"step": 507
},
{
"epoch": 0.7854657904909161,
"grad_norm": 0.48964107036590576,
"learning_rate": 3.288590604026846e-05,
"loss": 0.7643,
"step": 508
},
{
"epoch": 0.7870119829918825,
"grad_norm": 0.4985421299934387,
"learning_rate": 3.284395973154363e-05,
"loss": 0.7807,
"step": 509
},
{
"epoch": 0.7885581754928489,
"grad_norm": 0.5199535489082336,
"learning_rate": 3.2802013422818794e-05,
"loss": 0.7899,
"step": 510
},
{
"epoch": 0.7901043679938152,
"grad_norm": 0.5251364707946777,
"learning_rate": 3.2760067114093964e-05,
"loss": 0.7619,
"step": 511
},
{
"epoch": 0.7916505604947816,
"grad_norm": 0.5302333831787109,
"learning_rate": 3.271812080536913e-05,
"loss": 0.7561,
"step": 512
},
{
"epoch": 0.793196752995748,
"grad_norm": 0.544916033744812,
"learning_rate": 3.26761744966443e-05,
"loss": 0.7497,
"step": 513
},
{
"epoch": 0.7947429454967143,
"grad_norm": 0.5568458437919617,
"learning_rate": 3.263422818791946e-05,
"loss": 0.7872,
"step": 514
},
{
"epoch": 0.7962891379976808,
"grad_norm": 0.5136657357215881,
"learning_rate": 3.259228187919463e-05,
"loss": 0.8583,
"step": 515
},
{
"epoch": 0.7978353304986471,
"grad_norm": 0.5261164903640747,
"learning_rate": 3.25503355704698e-05,
"loss": 0.7806,
"step": 516
},
{
"epoch": 0.7993815229996134,
"grad_norm": 0.5282236933708191,
"learning_rate": 3.250838926174497e-05,
"loss": 0.7391,
"step": 517
},
{
"epoch": 0.8009277155005798,
"grad_norm": 0.5518252849578857,
"learning_rate": 3.246644295302014e-05,
"loss": 0.7981,
"step": 518
},
{
"epoch": 0.8024739080015462,
"grad_norm": 0.5146390199661255,
"learning_rate": 3.24244966442953e-05,
"loss": 0.7831,
"step": 519
},
{
"epoch": 0.8040201005025126,
"grad_norm": 0.5058282017707825,
"learning_rate": 3.238255033557047e-05,
"loss": 0.807,
"step": 520
},
{
"epoch": 0.8055662930034789,
"grad_norm": 0.5313502550125122,
"learning_rate": 3.2340604026845636e-05,
"loss": 0.8176,
"step": 521
},
{
"epoch": 0.8071124855044453,
"grad_norm": 0.5471131801605225,
"learning_rate": 3.2298657718120806e-05,
"loss": 0.81,
"step": 522
},
{
"epoch": 0.8086586780054117,
"grad_norm": 0.5153350830078125,
"learning_rate": 3.2256711409395976e-05,
"loss": 0.8729,
"step": 523
},
{
"epoch": 0.810204870506378,
"grad_norm": 0.5148741006851196,
"learning_rate": 3.221476510067114e-05,
"loss": 0.7776,
"step": 524
},
{
"epoch": 0.8117510630073445,
"grad_norm": 0.5151890516281128,
"learning_rate": 3.217281879194631e-05,
"loss": 0.8355,
"step": 525
},
{
"epoch": 0.8132972555083108,
"grad_norm": 0.5539215207099915,
"learning_rate": 3.213087248322148e-05,
"loss": 0.8074,
"step": 526
},
{
"epoch": 0.8148434480092771,
"grad_norm": 0.5416978597640991,
"learning_rate": 3.208892617449665e-05,
"loss": 0.8077,
"step": 527
},
{
"epoch": 0.8163896405102435,
"grad_norm": 0.5648258924484253,
"learning_rate": 3.204697986577181e-05,
"loss": 0.7928,
"step": 528
},
{
"epoch": 0.8179358330112099,
"grad_norm": 0.549947202205658,
"learning_rate": 3.200503355704698e-05,
"loss": 0.8371,
"step": 529
},
{
"epoch": 0.8194820255121763,
"grad_norm": 0.5815473198890686,
"learning_rate": 3.196308724832215e-05,
"loss": 0.8236,
"step": 530
},
{
"epoch": 0.8210282180131426,
"grad_norm": 0.5924739241600037,
"learning_rate": 3.1921140939597315e-05,
"loss": 0.7953,
"step": 531
},
{
"epoch": 0.822574410514109,
"grad_norm": 0.5691911578178406,
"learning_rate": 3.1879194630872485e-05,
"loss": 0.8312,
"step": 532
},
{
"epoch": 0.8241206030150754,
"grad_norm": 0.5568677186965942,
"learning_rate": 3.183724832214765e-05,
"loss": 0.839,
"step": 533
},
{
"epoch": 0.8256667955160417,
"grad_norm": 0.584894061088562,
"learning_rate": 3.1795302013422825e-05,
"loss": 0.9014,
"step": 534
},
{
"epoch": 0.8272129880170082,
"grad_norm": 0.5695962309837341,
"learning_rate": 3.175335570469799e-05,
"loss": 0.8177,
"step": 535
},
{
"epoch": 0.8287591805179745,
"grad_norm": 0.5662206411361694,
"learning_rate": 3.171140939597316e-05,
"loss": 0.8666,
"step": 536
},
{
"epoch": 0.8303053730189408,
"grad_norm": 0.6044044494628906,
"learning_rate": 3.166946308724832e-05,
"loss": 0.8757,
"step": 537
},
{
"epoch": 0.8318515655199072,
"grad_norm": 0.5584191083908081,
"learning_rate": 3.162751677852349e-05,
"loss": 0.8701,
"step": 538
},
{
"epoch": 0.8333977580208736,
"grad_norm": 0.6086747646331787,
"learning_rate": 3.158557046979866e-05,
"loss": 0.9369,
"step": 539
},
{
"epoch": 0.83494395052184,
"grad_norm": 0.5706461071968079,
"learning_rate": 3.1543624161073825e-05,
"loss": 0.9749,
"step": 540
},
{
"epoch": 0.8364901430228063,
"grad_norm": 0.5996837615966797,
"learning_rate": 3.1501677852348995e-05,
"loss": 0.881,
"step": 541
},
{
"epoch": 0.8380363355237727,
"grad_norm": 0.645283043384552,
"learning_rate": 3.145973154362416e-05,
"loss": 0.9201,
"step": 542
},
{
"epoch": 0.8395825280247391,
"grad_norm": 0.6247162222862244,
"learning_rate": 3.1417785234899334e-05,
"loss": 0.9159,
"step": 543
},
{
"epoch": 0.8411287205257054,
"grad_norm": 0.6042696833610535,
"learning_rate": 3.13758389261745e-05,
"loss": 0.9362,
"step": 544
},
{
"epoch": 0.8426749130266719,
"grad_norm": 0.6593182682991028,
"learning_rate": 3.133389261744967e-05,
"loss": 0.9722,
"step": 545
},
{
"epoch": 0.8442211055276382,
"grad_norm": 0.665708601474762,
"learning_rate": 3.129194630872484e-05,
"loss": 0.9028,
"step": 546
},
{
"epoch": 0.8457672980286045,
"grad_norm": 0.6491990089416504,
"learning_rate": 3.125e-05,
"loss": 0.9169,
"step": 547
},
{
"epoch": 0.847313490529571,
"grad_norm": 0.7431784868240356,
"learning_rate": 3.120805369127517e-05,
"loss": 0.9927,
"step": 548
},
{
"epoch": 0.8488596830305373,
"grad_norm": 0.8240005970001221,
"learning_rate": 3.1166107382550334e-05,
"loss": 1.0051,
"step": 549
},
{
"epoch": 0.8504058755315037,
"grad_norm": 1.1392122507095337,
"learning_rate": 3.1124161073825504e-05,
"loss": 0.9428,
"step": 550
},
{
"epoch": 0.85195206803247,
"grad_norm": 0.48860234022140503,
"learning_rate": 3.108221476510067e-05,
"loss": 0.7069,
"step": 551
},
{
"epoch": 0.8534982605334364,
"grad_norm": 0.4950990378856659,
"learning_rate": 3.1040268456375844e-05,
"loss": 0.7463,
"step": 552
},
{
"epoch": 0.8550444530344028,
"grad_norm": 0.48706600069999695,
"learning_rate": 3.099832214765101e-05,
"loss": 0.786,
"step": 553
},
{
"epoch": 0.8565906455353691,
"grad_norm": 0.5027382373809814,
"learning_rate": 3.095637583892618e-05,
"loss": 0.7882,
"step": 554
},
{
"epoch": 0.8581368380363356,
"grad_norm": 0.5104815363883972,
"learning_rate": 3.091442953020135e-05,
"loss": 0.7301,
"step": 555
},
{
"epoch": 0.8596830305373019,
"grad_norm": 0.49334749579429626,
"learning_rate": 3.087248322147651e-05,
"loss": 0.786,
"step": 556
},
{
"epoch": 0.8612292230382682,
"grad_norm": 0.5204259157180786,
"learning_rate": 3.083053691275168e-05,
"loss": 0.7809,
"step": 557
},
{
"epoch": 0.8627754155392346,
"grad_norm": 0.5358408093452454,
"learning_rate": 3.078859060402684e-05,
"loss": 0.7887,
"step": 558
},
{
"epoch": 0.864321608040201,
"grad_norm": 0.515990674495697,
"learning_rate": 3.074664429530201e-05,
"loss": 0.8452,
"step": 559
},
{
"epoch": 0.8658678005411674,
"grad_norm": 0.5604992508888245,
"learning_rate": 3.070469798657718e-05,
"loss": 0.7178,
"step": 560
},
{
"epoch": 0.8674139930421337,
"grad_norm": 0.5511056184768677,
"learning_rate": 3.066275167785235e-05,
"loss": 0.7577,
"step": 561
},
{
"epoch": 0.8689601855431001,
"grad_norm": 0.5570728778839111,
"learning_rate": 3.062080536912752e-05,
"loss": 0.7799,
"step": 562
},
{
"epoch": 0.8705063780440665,
"grad_norm": 0.535253643989563,
"learning_rate": 3.0578859060402686e-05,
"loss": 0.85,
"step": 563
},
{
"epoch": 0.8720525705450328,
"grad_norm": 0.5327886343002319,
"learning_rate": 3.0536912751677856e-05,
"loss": 0.8057,
"step": 564
},
{
"epoch": 0.8735987630459993,
"grad_norm": 0.5440667867660522,
"learning_rate": 3.0494966442953022e-05,
"loss": 0.7407,
"step": 565
},
{
"epoch": 0.8751449555469656,
"grad_norm": 0.5552430152893066,
"learning_rate": 3.045302013422819e-05,
"loss": 0.8022,
"step": 566
},
{
"epoch": 0.876691148047932,
"grad_norm": 0.523145854473114,
"learning_rate": 3.0411073825503356e-05,
"loss": 0.7694,
"step": 567
},
{
"epoch": 0.8782373405488983,
"grad_norm": 0.5461344122886658,
"learning_rate": 3.0369127516778522e-05,
"loss": 0.7849,
"step": 568
},
{
"epoch": 0.8797835330498647,
"grad_norm": 0.548468828201294,
"learning_rate": 3.0327181208053695e-05,
"loss": 0.8127,
"step": 569
},
{
"epoch": 0.8813297255508311,
"grad_norm": 0.5320367813110352,
"learning_rate": 3.0285234899328862e-05,
"loss": 0.8243,
"step": 570
},
{
"epoch": 0.8828759180517974,
"grad_norm": 0.5147125124931335,
"learning_rate": 3.024328859060403e-05,
"loss": 0.8784,
"step": 571
},
{
"epoch": 0.8844221105527639,
"grad_norm": 0.5353872776031494,
"learning_rate": 3.02013422818792e-05,
"loss": 0.7841,
"step": 572
},
{
"epoch": 0.8859683030537302,
"grad_norm": 0.5554389953613281,
"learning_rate": 3.0159395973154365e-05,
"loss": 0.8601,
"step": 573
},
{
"epoch": 0.8875144955546965,
"grad_norm": 0.5762984752655029,
"learning_rate": 3.011744966442953e-05,
"loss": 0.7359,
"step": 574
},
{
"epoch": 0.889060688055663,
"grad_norm": 0.5287367701530457,
"learning_rate": 3.0075503355704698e-05,
"loss": 0.8063,
"step": 575
},
{
"epoch": 0.8906068805566293,
"grad_norm": 0.5730248689651489,
"learning_rate": 3.0033557046979865e-05,
"loss": 0.735,
"step": 576
},
{
"epoch": 0.8921530730575957,
"grad_norm": 0.557384192943573,
"learning_rate": 2.999161073825503e-05,
"loss": 0.7962,
"step": 577
},
{
"epoch": 0.893699265558562,
"grad_norm": 0.5956771373748779,
"learning_rate": 2.9949664429530205e-05,
"loss": 0.8867,
"step": 578
},
{
"epoch": 0.8952454580595284,
"grad_norm": 0.5796382427215576,
"learning_rate": 2.990771812080537e-05,
"loss": 0.858,
"step": 579
},
{
"epoch": 0.8967916505604948,
"grad_norm": 0.5635401010513306,
"learning_rate": 2.986577181208054e-05,
"loss": 0.7987,
"step": 580
},
{
"epoch": 0.8983378430614611,
"grad_norm": 0.6214528679847717,
"learning_rate": 2.9823825503355708e-05,
"loss": 0.8397,
"step": 581
},
{
"epoch": 0.8998840355624276,
"grad_norm": 0.5945339202880859,
"learning_rate": 2.9781879194630874e-05,
"loss": 0.8428,
"step": 582
},
{
"epoch": 0.9014302280633939,
"grad_norm": 0.5704767107963562,
"learning_rate": 2.973993288590604e-05,
"loss": 0.8291,
"step": 583
},
{
"epoch": 0.9029764205643602,
"grad_norm": 0.5803176164627075,
"learning_rate": 2.9697986577181207e-05,
"loss": 0.8792,
"step": 584
},
{
"epoch": 0.9045226130653267,
"grad_norm": 0.6039038300514221,
"learning_rate": 2.9656040268456374e-05,
"loss": 0.7788,
"step": 585
},
{
"epoch": 0.906068805566293,
"grad_norm": 0.5550165176391602,
"learning_rate": 2.9614093959731544e-05,
"loss": 0.8658,
"step": 586
},
{
"epoch": 0.9076149980672594,
"grad_norm": 0.6096512079238892,
"learning_rate": 2.9572147651006714e-05,
"loss": 0.9237,
"step": 587
},
{
"epoch": 0.9091611905682258,
"grad_norm": 0.5784430503845215,
"learning_rate": 2.9530201342281884e-05,
"loss": 0.8538,
"step": 588
},
{
"epoch": 0.9107073830691921,
"grad_norm": 0.5854966044425964,
"learning_rate": 2.948825503355705e-05,
"loss": 0.9009,
"step": 589
},
{
"epoch": 0.9122535755701585,
"grad_norm": 0.5904499292373657,
"learning_rate": 2.9446308724832217e-05,
"loss": 0.9064,
"step": 590
},
{
"epoch": 0.9137997680711248,
"grad_norm": 0.6125240325927734,
"learning_rate": 2.9404362416107384e-05,
"loss": 0.8539,
"step": 591
},
{
"epoch": 0.9153459605720913,
"grad_norm": 0.6209454536437988,
"learning_rate": 2.936241610738255e-05,
"loss": 0.8864,
"step": 592
},
{
"epoch": 0.9168921530730576,
"grad_norm": 0.634355902671814,
"learning_rate": 2.9320469798657717e-05,
"loss": 0.8999,
"step": 593
},
{
"epoch": 0.9184383455740239,
"grad_norm": 0.644378125667572,
"learning_rate": 2.9278523489932887e-05,
"loss": 0.9351,
"step": 594
},
{
"epoch": 0.9199845380749904,
"grad_norm": 0.638783872127533,
"learning_rate": 2.9236577181208053e-05,
"loss": 0.8522,
"step": 595
},
{
"epoch": 0.9215307305759567,
"grad_norm": 0.6960675716400146,
"learning_rate": 2.9194630872483227e-05,
"loss": 0.8183,
"step": 596
},
{
"epoch": 0.9230769230769231,
"grad_norm": 0.7445054054260254,
"learning_rate": 2.9152684563758393e-05,
"loss": 0.9749,
"step": 597
},
{
"epoch": 0.9246231155778895,
"grad_norm": 0.7296366095542908,
"learning_rate": 2.911073825503356e-05,
"loss": 0.9535,
"step": 598
},
{
"epoch": 0.9261693080788558,
"grad_norm": 0.8242074251174927,
"learning_rate": 2.9068791946308726e-05,
"loss": 0.9661,
"step": 599
},
{
"epoch": 0.9277155005798222,
"grad_norm": 0.983094334602356,
"learning_rate": 2.9026845637583893e-05,
"loss": 0.8454,
"step": 600
},
{
"epoch": 0.9292616930807885,
"grad_norm": 0.4916326403617859,
"learning_rate": 2.898489932885906e-05,
"loss": 0.7178,
"step": 601
},
{
"epoch": 0.930807885581755,
"grad_norm": 0.5236021876335144,
"learning_rate": 2.894295302013423e-05,
"loss": 0.7586,
"step": 602
},
{
"epoch": 0.9323540780827213,
"grad_norm": 0.5226188898086548,
"learning_rate": 2.8901006711409396e-05,
"loss": 0.7287,
"step": 603
},
{
"epoch": 0.9339002705836876,
"grad_norm": 0.5189059376716614,
"learning_rate": 2.885906040268457e-05,
"loss": 0.7176,
"step": 604
},
{
"epoch": 0.9354464630846541,
"grad_norm": 0.5282127857208252,
"learning_rate": 2.8817114093959736e-05,
"loss": 0.7487,
"step": 605
},
{
"epoch": 0.9369926555856204,
"grad_norm": 0.5156176090240479,
"learning_rate": 2.8775167785234902e-05,
"loss": 0.7929,
"step": 606
},
{
"epoch": 0.9385388480865868,
"grad_norm": 0.5219593644142151,
"learning_rate": 2.873322147651007e-05,
"loss": 0.7992,
"step": 607
},
{
"epoch": 0.9400850405875532,
"grad_norm": 0.5400338768959045,
"learning_rate": 2.8691275167785235e-05,
"loss": 0.7738,
"step": 608
},
{
"epoch": 0.9416312330885195,
"grad_norm": 0.5067276954650879,
"learning_rate": 2.8649328859060402e-05,
"loss": 0.8046,
"step": 609
},
{
"epoch": 0.9431774255894859,
"grad_norm": 0.5286040902137756,
"learning_rate": 2.8607382550335572e-05,
"loss": 0.6818,
"step": 610
},
{
"epoch": 0.9447236180904522,
"grad_norm": 0.5325278043746948,
"learning_rate": 2.856543624161074e-05,
"loss": 0.8112,
"step": 611
},
{
"epoch": 0.9462698105914187,
"grad_norm": 0.5434727668762207,
"learning_rate": 2.8523489932885905e-05,
"loss": 0.8245,
"step": 612
},
{
"epoch": 0.947816003092385,
"grad_norm": 0.5526994466781616,
"learning_rate": 2.848154362416108e-05,
"loss": 0.7493,
"step": 613
},
{
"epoch": 0.9493621955933513,
"grad_norm": 0.5655114054679871,
"learning_rate": 2.8439597315436245e-05,
"loss": 0.7514,
"step": 614
},
{
"epoch": 0.9509083880943178,
"grad_norm": 0.5636076331138611,
"learning_rate": 2.839765100671141e-05,
"loss": 0.7992,
"step": 615
},
{
"epoch": 0.9524545805952841,
"grad_norm": 0.5688204169273376,
"learning_rate": 2.8355704697986578e-05,
"loss": 0.7747,
"step": 616
},
{
"epoch": 0.9540007730962505,
"grad_norm": 0.5534058809280396,
"learning_rate": 2.8313758389261748e-05,
"loss": 0.7511,
"step": 617
},
{
"epoch": 0.9555469655972169,
"grad_norm": 0.5112160444259644,
"learning_rate": 2.8271812080536915e-05,
"loss": 0.7875,
"step": 618
},
{
"epoch": 0.9570931580981832,
"grad_norm": 0.553830623626709,
"learning_rate": 2.822986577181208e-05,
"loss": 0.8652,
"step": 619
},
{
"epoch": 0.9586393505991496,
"grad_norm": 0.5614729523658752,
"learning_rate": 2.8187919463087248e-05,
"loss": 0.8693,
"step": 620
},
{
"epoch": 0.960185543100116,
"grad_norm": 0.5519264340400696,
"learning_rate": 2.8145973154362414e-05,
"loss": 0.8102,
"step": 621
},
{
"epoch": 0.9617317356010824,
"grad_norm": 0.5544281601905823,
"learning_rate": 2.8104026845637588e-05,
"loss": 0.8263,
"step": 622
},
{
"epoch": 0.9632779281020487,
"grad_norm": 0.5747584700584412,
"learning_rate": 2.8062080536912754e-05,
"loss": 0.7718,
"step": 623
},
{
"epoch": 0.964824120603015,
"grad_norm": 0.5676540732383728,
"learning_rate": 2.802013422818792e-05,
"loss": 0.7616,
"step": 624
},
{
"epoch": 0.9663703131039815,
"grad_norm": 0.5307291150093079,
"learning_rate": 2.797818791946309e-05,
"loss": 0.8401,
"step": 625
},
{
"epoch": 0.9679165056049478,
"grad_norm": 0.5527417063713074,
"learning_rate": 2.7936241610738257e-05,
"loss": 0.8818,
"step": 626
},
{
"epoch": 0.9694626981059142,
"grad_norm": 0.545058012008667,
"learning_rate": 2.7894295302013424e-05,
"loss": 0.8606,
"step": 627
},
{
"epoch": 0.9710088906068806,
"grad_norm": 0.5928349494934082,
"learning_rate": 2.785234899328859e-05,
"loss": 0.7728,
"step": 628
},
{
"epoch": 0.9725550831078469,
"grad_norm": 0.5348992943763733,
"learning_rate": 2.7810402684563757e-05,
"loss": 0.8101,
"step": 629
},
{
"epoch": 0.9741012756088133,
"grad_norm": 0.583490788936615,
"learning_rate": 2.7768456375838923e-05,
"loss": 0.8648,
"step": 630
},
{
"epoch": 0.9756474681097796,
"grad_norm": 0.6235148906707764,
"learning_rate": 2.7726510067114097e-05,
"loss": 0.8425,
"step": 631
},
{
"epoch": 0.9771936606107461,
"grad_norm": 0.5755742788314819,
"learning_rate": 2.7684563758389263e-05,
"loss": 0.8765,
"step": 632
},
{
"epoch": 0.9787398531117124,
"grad_norm": 0.5636020302772522,
"learning_rate": 2.7642617449664433e-05,
"loss": 0.8114,
"step": 633
},
{
"epoch": 0.9802860456126787,
"grad_norm": 0.5665507316589355,
"learning_rate": 2.76006711409396e-05,
"loss": 0.9008,
"step": 634
},
{
"epoch": 0.9818322381136452,
"grad_norm": 0.5966489911079407,
"learning_rate": 2.7558724832214766e-05,
"loss": 0.846,
"step": 635
},
{
"epoch": 0.9833784306146115,
"grad_norm": 0.5960109829902649,
"learning_rate": 2.7516778523489933e-05,
"loss": 0.8962,
"step": 636
},
{
"epoch": 0.9849246231155779,
"grad_norm": 0.5626753568649292,
"learning_rate": 2.74748322147651e-05,
"loss": 0.9062,
"step": 637
},
{
"epoch": 0.9864708156165443,
"grad_norm": 0.6764492392539978,
"learning_rate": 2.7432885906040266e-05,
"loss": 0.8961,
"step": 638
},
{
"epoch": 0.9880170081175106,
"grad_norm": 0.5922832489013672,
"learning_rate": 2.7390939597315436e-05,
"loss": 0.8525,
"step": 639
},
{
"epoch": 0.989563200618477,
"grad_norm": 0.6102508902549744,
"learning_rate": 2.7348993288590606e-05,
"loss": 0.8887,
"step": 640
},
{
"epoch": 0.9911093931194433,
"grad_norm": 0.6205296516418457,
"learning_rate": 2.7307046979865776e-05,
"loss": 0.9007,
"step": 641
},
{
"epoch": 0.9926555856204098,
"grad_norm": 0.6284985542297363,
"learning_rate": 2.7265100671140943e-05,
"loss": 0.9228,
"step": 642
},
{
"epoch": 0.9942017781213761,
"grad_norm": 0.6276938915252686,
"learning_rate": 2.722315436241611e-05,
"loss": 0.9066,
"step": 643
},
{
"epoch": 0.9957479706223424,
"grad_norm": 0.6849061250686646,
"learning_rate": 2.7181208053691276e-05,
"loss": 0.9204,
"step": 644
},
{
"epoch": 0.9972941631233089,
"grad_norm": 0.7061152458190918,
"learning_rate": 2.7139261744966442e-05,
"loss": 0.9711,
"step": 645
},
{
"epoch": 0.9988403556242752,
"grad_norm": 0.7999619841575623,
"learning_rate": 2.709731543624161e-05,
"loss": 0.9722,
"step": 646
},
{
"epoch": 1.0007730962504833,
"grad_norm": 1.3448657989501953,
"learning_rate": 2.705536912751678e-05,
"loss": 1.2285,
"step": 647
},
{
"epoch": 1.0023192887514496,
"grad_norm": 0.4756757318973541,
"learning_rate": 2.701342281879195e-05,
"loss": 0.7208,
"step": 648
},
{
"epoch": 1.003865481252416,
"grad_norm": 0.495257169008255,
"learning_rate": 2.697147651006712e-05,
"loss": 0.7645,
"step": 649
},
{
"epoch": 1.0054116737533823,
"grad_norm": 0.5022267699241638,
"learning_rate": 2.6929530201342285e-05,
"loss": 0.7099,
"step": 650
},
{
"epoch": 1.0069578662543486,
"grad_norm": 0.5082698464393616,
"learning_rate": 2.6887583892617452e-05,
"loss": 0.7451,
"step": 651
},
{
"epoch": 1.0085040587553151,
"grad_norm": 0.5273095369338989,
"learning_rate": 2.6845637583892618e-05,
"loss": 0.7665,
"step": 652
},
{
"epoch": 1.0100502512562815,
"grad_norm": 0.5531541109085083,
"learning_rate": 2.6803691275167785e-05,
"loss": 0.7041,
"step": 653
},
{
"epoch": 1.0115964437572478,
"grad_norm": 0.5290402173995972,
"learning_rate": 2.6761744966442955e-05,
"loss": 0.812,
"step": 654
},
{
"epoch": 1.013142636258214,
"grad_norm": 0.556932270526886,
"learning_rate": 2.671979865771812e-05,
"loss": 0.7414,
"step": 655
},
{
"epoch": 1.0146888287591804,
"grad_norm": 0.501987874507904,
"learning_rate": 2.6677852348993288e-05,
"loss": 0.7446,
"step": 656
},
{
"epoch": 1.016235021260147,
"grad_norm": 0.5225424766540527,
"learning_rate": 2.663590604026846e-05,
"loss": 0.8439,
"step": 657
},
{
"epoch": 1.0177812137611133,
"grad_norm": 0.5627469420433044,
"learning_rate": 2.6593959731543628e-05,
"loss": 0.6966,
"step": 658
},
{
"epoch": 1.0193274062620796,
"grad_norm": 0.6055929660797119,
"learning_rate": 2.6552013422818794e-05,
"loss": 0.7859,
"step": 659
},
{
"epoch": 1.020873598763046,
"grad_norm": 0.6422880291938782,
"learning_rate": 2.651006711409396e-05,
"loss": 0.7733,
"step": 660
},
{
"epoch": 1.0224197912640123,
"grad_norm": 0.5887860059738159,
"learning_rate": 2.6468120805369128e-05,
"loss": 0.7041,
"step": 661
},
{
"epoch": 1.0239659837649788,
"grad_norm": 0.5853235125541687,
"learning_rate": 2.6426174496644297e-05,
"loss": 0.7268,
"step": 662
},
{
"epoch": 1.0255121762659452,
"grad_norm": 0.5295597910881042,
"learning_rate": 2.6384228187919464e-05,
"loss": 0.7482,
"step": 663
},
{
"epoch": 1.0270583687669115,
"grad_norm": 0.5481401085853577,
"learning_rate": 2.634228187919463e-05,
"loss": 0.7548,
"step": 664
},
{
"epoch": 1.0286045612678778,
"grad_norm": 0.538827121257782,
"learning_rate": 2.6300335570469797e-05,
"loss": 0.764,
"step": 665
},
{
"epoch": 1.0301507537688441,
"grad_norm": 0.577368974685669,
"learning_rate": 2.625838926174497e-05,
"loss": 0.6835,
"step": 666
},
{
"epoch": 1.0316969462698107,
"grad_norm": 0.5659049153327942,
"learning_rate": 2.6216442953020137e-05,
"loss": 0.7264,
"step": 667
},
{
"epoch": 1.033243138770777,
"grad_norm": 0.5179746150970459,
"learning_rate": 2.6174496644295304e-05,
"loss": 0.7911,
"step": 668
},
{
"epoch": 1.0347893312717433,
"grad_norm": 0.5719809532165527,
"learning_rate": 2.613255033557047e-05,
"loss": 0.7479,
"step": 669
},
{
"epoch": 1.0363355237727097,
"grad_norm": 0.5943763852119446,
"learning_rate": 2.609060402684564e-05,
"loss": 0.7802,
"step": 670
},
{
"epoch": 1.037881716273676,
"grad_norm": 0.5392094850540161,
"learning_rate": 2.6048657718120807e-05,
"loss": 0.7625,
"step": 671
},
{
"epoch": 1.0394279087746425,
"grad_norm": 0.5679749250411987,
"learning_rate": 2.6006711409395973e-05,
"loss": 0.7911,
"step": 672
},
{
"epoch": 1.0409741012756089,
"grad_norm": 0.5740141272544861,
"learning_rate": 2.596476510067114e-05,
"loss": 0.7477,
"step": 673
},
{
"epoch": 1.0425202937765752,
"grad_norm": 0.607397198677063,
"learning_rate": 2.5922818791946306e-05,
"loss": 0.852,
"step": 674
},
{
"epoch": 1.0440664862775415,
"grad_norm": 0.5706917643547058,
"learning_rate": 2.588087248322148e-05,
"loss": 0.6778,
"step": 675
},
{
"epoch": 1.0456126787785078,
"grad_norm": 0.5882996320724487,
"learning_rate": 2.5838926174496646e-05,
"loss": 0.7071,
"step": 676
},
{
"epoch": 1.0471588712794744,
"grad_norm": 0.5978296995162964,
"learning_rate": 2.5796979865771813e-05,
"loss": 0.7593,
"step": 677
},
{
"epoch": 1.0487050637804407,
"grad_norm": 0.6237056255340576,
"learning_rate": 2.5755033557046983e-05,
"loss": 0.8099,
"step": 678
},
{
"epoch": 1.050251256281407,
"grad_norm": 0.6156934499740601,
"learning_rate": 2.571308724832215e-05,
"loss": 0.8256,
"step": 679
},
{
"epoch": 1.0517974487823734,
"grad_norm": 0.6217848062515259,
"learning_rate": 2.5671140939597316e-05,
"loss": 0.8445,
"step": 680
},
{
"epoch": 1.0533436412833397,
"grad_norm": 0.6550363898277283,
"learning_rate": 2.5629194630872482e-05,
"loss": 0.7931,
"step": 681
},
{
"epoch": 1.0548898337843062,
"grad_norm": 0.6072224378585815,
"learning_rate": 2.558724832214765e-05,
"loss": 0.7985,
"step": 682
},
{
"epoch": 1.0564360262852726,
"grad_norm": 0.6478685140609741,
"learning_rate": 2.5545302013422822e-05,
"loss": 0.8055,
"step": 683
},
{
"epoch": 1.0579822187862389,
"grad_norm": 0.6033689975738525,
"learning_rate": 2.550335570469799e-05,
"loss": 0.9145,
"step": 684
},
{
"epoch": 1.0595284112872052,
"grad_norm": 0.5850486755371094,
"learning_rate": 2.5461409395973155e-05,
"loss": 0.8129,
"step": 685
},
{
"epoch": 1.0610746037881715,
"grad_norm": 0.6233928799629211,
"learning_rate": 2.5419463087248325e-05,
"loss": 0.9122,
"step": 686
},
{
"epoch": 1.062620796289138,
"grad_norm": 0.6058603525161743,
"learning_rate": 2.5377516778523492e-05,
"loss": 0.7968,
"step": 687
},
{
"epoch": 1.0641669887901044,
"grad_norm": 0.6053382754325867,
"learning_rate": 2.533557046979866e-05,
"loss": 0.8349,
"step": 688
},
{
"epoch": 1.0657131812910707,
"grad_norm": 0.6354022026062012,
"learning_rate": 2.5293624161073825e-05,
"loss": 0.8879,
"step": 689
},
{
"epoch": 1.067259373792037,
"grad_norm": 0.6736825108528137,
"learning_rate": 2.525167785234899e-05,
"loss": 0.8317,
"step": 690
},
{
"epoch": 1.0688055662930034,
"grad_norm": 0.6729496717453003,
"learning_rate": 2.5209731543624158e-05,
"loss": 0.8247,
"step": 691
},
{
"epoch": 1.07035175879397,
"grad_norm": 0.7000686526298523,
"learning_rate": 2.516778523489933e-05,
"loss": 0.9076,
"step": 692
},
{
"epoch": 1.0718979512949363,
"grad_norm": 0.7471379041671753,
"learning_rate": 2.5125838926174498e-05,
"loss": 0.861,
"step": 693
},
{
"epoch": 1.0734441437959026,
"grad_norm": 0.7505892515182495,
"learning_rate": 2.5083892617449668e-05,
"loss": 0.9747,
"step": 694
},
{
"epoch": 1.074990336296869,
"grad_norm": 0.7872920632362366,
"learning_rate": 2.5041946308724835e-05,
"loss": 0.8623,
"step": 695
},
{
"epoch": 1.0765365287978352,
"grad_norm": 0.97450190782547,
"learning_rate": 2.5e-05,
"loss": 0.8574,
"step": 696
},
{
"epoch": 1.0780827212988018,
"grad_norm": 0.6834471225738525,
"learning_rate": 2.4958053691275168e-05,
"loss": 0.7284,
"step": 697
},
{
"epoch": 1.079628913799768,
"grad_norm": 0.5371273756027222,
"learning_rate": 2.4916107382550334e-05,
"loss": 0.6604,
"step": 698
},
{
"epoch": 1.0811751063007344,
"grad_norm": 0.5436398983001709,
"learning_rate": 2.4874161073825504e-05,
"loss": 0.7106,
"step": 699
},
{
"epoch": 1.0827212988017008,
"grad_norm": 0.5626257061958313,
"learning_rate": 2.4832214765100674e-05,
"loss": 0.7199,
"step": 700
},
{
"epoch": 1.084267491302667,
"grad_norm": 0.5859701037406921,
"learning_rate": 2.479026845637584e-05,
"loss": 0.727,
"step": 701
},
{
"epoch": 1.0858136838036336,
"grad_norm": 0.5619367957115173,
"learning_rate": 2.4748322147651007e-05,
"loss": 0.7721,
"step": 702
},
{
"epoch": 1.0873598763046,
"grad_norm": 0.5503877997398376,
"learning_rate": 2.4706375838926177e-05,
"loss": 0.6971,
"step": 703
},
{
"epoch": 1.0889060688055663,
"grad_norm": 0.5587684512138367,
"learning_rate": 2.4664429530201344e-05,
"loss": 0.7868,
"step": 704
},
{
"epoch": 1.0904522613065326,
"grad_norm": 0.5771764516830444,
"learning_rate": 2.462248322147651e-05,
"loss": 0.7831,
"step": 705
},
{
"epoch": 1.091998453807499,
"grad_norm": 0.5702334046363831,
"learning_rate": 2.4580536912751677e-05,
"loss": 0.6751,
"step": 706
},
{
"epoch": 1.0935446463084655,
"grad_norm": 0.5503116250038147,
"learning_rate": 2.4538590604026847e-05,
"loss": 0.7732,
"step": 707
},
{
"epoch": 1.0950908388094318,
"grad_norm": 0.5595600008964539,
"learning_rate": 2.4496644295302017e-05,
"loss": 0.7168,
"step": 708
},
{
"epoch": 1.0966370313103981,
"grad_norm": 0.5581438541412354,
"learning_rate": 2.4454697986577183e-05,
"loss": 0.7462,
"step": 709
},
{
"epoch": 1.0981832238113645,
"grad_norm": 0.5454738736152649,
"learning_rate": 2.441275167785235e-05,
"loss": 0.8809,
"step": 710
},
{
"epoch": 1.0997294163123308,
"grad_norm": 0.5763128399848938,
"learning_rate": 2.4370805369127517e-05,
"loss": 0.6942,
"step": 711
},
{
"epoch": 1.1012756088132973,
"grad_norm": 0.5813524127006531,
"learning_rate": 2.4328859060402687e-05,
"loss": 0.6836,
"step": 712
},
{
"epoch": 1.1028218013142637,
"grad_norm": 0.5554409623146057,
"learning_rate": 2.4286912751677853e-05,
"loss": 0.7616,
"step": 713
},
{
"epoch": 1.10436799381523,
"grad_norm": 0.5576358437538147,
"learning_rate": 2.424496644295302e-05,
"loss": 0.7835,
"step": 714
},
{
"epoch": 1.1059141863161963,
"grad_norm": 0.5760726928710938,
"learning_rate": 2.420302013422819e-05,
"loss": 0.7866,
"step": 715
},
{
"epoch": 1.1074603788171626,
"grad_norm": 0.5910109877586365,
"learning_rate": 2.416107382550336e-05,
"loss": 0.7893,
"step": 716
},
{
"epoch": 1.1090065713181292,
"grad_norm": 0.6322896480560303,
"learning_rate": 2.4119127516778526e-05,
"loss": 0.7235,
"step": 717
},
{
"epoch": 1.1105527638190955,
"grad_norm": 0.5939295887947083,
"learning_rate": 2.4077181208053693e-05,
"loss": 0.7816,
"step": 718
},
{
"epoch": 1.1120989563200618,
"grad_norm": 0.5953226089477539,
"learning_rate": 2.403523489932886e-05,
"loss": 0.7444,
"step": 719
},
{
"epoch": 1.1136451488210282,
"grad_norm": 0.60584956407547,
"learning_rate": 2.3993288590604026e-05,
"loss": 0.7525,
"step": 720
},
{
"epoch": 1.1151913413219945,
"grad_norm": 0.6163296699523926,
"learning_rate": 2.3951342281879196e-05,
"loss": 0.772,
"step": 721
},
{
"epoch": 1.116737533822961,
"grad_norm": 0.5978072881698608,
"learning_rate": 2.3909395973154362e-05,
"loss": 0.7421,
"step": 722
},
{
"epoch": 1.1182837263239274,
"grad_norm": 0.5980417132377625,
"learning_rate": 2.3867449664429532e-05,
"loss": 0.7942,
"step": 723
},
{
"epoch": 1.1198299188248937,
"grad_norm": 0.6046664118766785,
"learning_rate": 2.38255033557047e-05,
"loss": 0.8131,
"step": 724
},
{
"epoch": 1.12137611132586,
"grad_norm": 0.5973670482635498,
"learning_rate": 2.378355704697987e-05,
"loss": 0.7857,
"step": 725
},
{
"epoch": 1.1229223038268263,
"grad_norm": 0.6115438342094421,
"learning_rate": 2.3741610738255035e-05,
"loss": 0.7942,
"step": 726
},
{
"epoch": 1.1244684963277929,
"grad_norm": 0.5946105718612671,
"learning_rate": 2.3699664429530202e-05,
"loss": 0.8088,
"step": 727
},
{
"epoch": 1.1260146888287592,
"grad_norm": 0.6095959544181824,
"learning_rate": 2.365771812080537e-05,
"loss": 0.8034,
"step": 728
},
{
"epoch": 1.1275608813297255,
"grad_norm": 0.6506893634796143,
"learning_rate": 2.361577181208054e-05,
"loss": 0.8469,
"step": 729
},
{
"epoch": 1.1291070738306919,
"grad_norm": 0.6179336309432983,
"learning_rate": 2.3573825503355705e-05,
"loss": 0.7918,
"step": 730
},
{
"epoch": 1.1306532663316582,
"grad_norm": 0.6253457069396973,
"learning_rate": 2.3531879194630875e-05,
"loss": 0.845,
"step": 731
},
{
"epoch": 1.1321994588326247,
"grad_norm": 0.6577156186103821,
"learning_rate": 2.348993288590604e-05,
"loss": 0.7475,
"step": 732
},
{
"epoch": 1.133745651333591,
"grad_norm": 0.6993891596794128,
"learning_rate": 2.3447986577181208e-05,
"loss": 0.7988,
"step": 733
},
{
"epoch": 1.1352918438345574,
"grad_norm": 0.6977733373641968,
"learning_rate": 2.3406040268456378e-05,
"loss": 0.7899,
"step": 734
},
{
"epoch": 1.1368380363355237,
"grad_norm": 0.6664114594459534,
"learning_rate": 2.3364093959731545e-05,
"loss": 0.8609,
"step": 735
},
{
"epoch": 1.13838422883649,
"grad_norm": 0.6565979719161987,
"learning_rate": 2.332214765100671e-05,
"loss": 0.8592,
"step": 736
},
{
"epoch": 1.1399304213374566,
"grad_norm": 0.6794628500938416,
"learning_rate": 2.3280201342281878e-05,
"loss": 0.8709,
"step": 737
},
{
"epoch": 1.141476613838423,
"grad_norm": 0.6833226084709167,
"learning_rate": 2.3238255033557048e-05,
"loss": 0.8121,
"step": 738
},
{
"epoch": 1.1430228063393892,
"grad_norm": 0.6899168491363525,
"learning_rate": 2.3196308724832218e-05,
"loss": 0.8362,
"step": 739
},
{
"epoch": 1.1445689988403556,
"grad_norm": 0.7028947472572327,
"learning_rate": 2.3154362416107384e-05,
"loss": 0.7995,
"step": 740
},
{
"epoch": 1.146115191341322,
"grad_norm": 0.7305999994277954,
"learning_rate": 2.311241610738255e-05,
"loss": 0.832,
"step": 741
},
{
"epoch": 1.1476613838422884,
"grad_norm": 0.7404617667198181,
"learning_rate": 2.3070469798657717e-05,
"loss": 0.9275,
"step": 742
},
{
"epoch": 1.1492075763432548,
"grad_norm": 0.7118339538574219,
"learning_rate": 2.3028523489932887e-05,
"loss": 0.8852,
"step": 743
},
{
"epoch": 1.150753768844221,
"grad_norm": 0.8329970240592957,
"learning_rate": 2.2986577181208054e-05,
"loss": 0.8482,
"step": 744
},
{
"epoch": 1.1522999613451874,
"grad_norm": 0.8370991945266724,
"learning_rate": 2.2944630872483224e-05,
"loss": 1.0003,
"step": 745
},
{
"epoch": 1.1538461538461537,
"grad_norm": 1.0701762437820435,
"learning_rate": 2.290268456375839e-05,
"loss": 1.0102,
"step": 746
},
{
"epoch": 1.1553923463471203,
"grad_norm": 0.7740350961685181,
"learning_rate": 2.286073825503356e-05,
"loss": 0.617,
"step": 747
},
{
"epoch": 1.1569385388480866,
"grad_norm": 0.5996577739715576,
"learning_rate": 2.2818791946308727e-05,
"loss": 0.7698,
"step": 748
},
{
"epoch": 1.158484731349053,
"grad_norm": 0.597896933555603,
"learning_rate": 2.2776845637583893e-05,
"loss": 0.7018,
"step": 749
},
{
"epoch": 1.1600309238500193,
"grad_norm": 0.5840704441070557,
"learning_rate": 2.273489932885906e-05,
"loss": 0.7634,
"step": 750
},
{
"epoch": 1.1615771163509856,
"grad_norm": 0.5987460613250732,
"learning_rate": 2.269295302013423e-05,
"loss": 0.7973,
"step": 751
},
{
"epoch": 1.1631233088519521,
"grad_norm": 0.5817953944206238,
"learning_rate": 2.2651006711409396e-05,
"loss": 0.7703,
"step": 752
},
{
"epoch": 1.1646695013529185,
"grad_norm": 0.5792746543884277,
"learning_rate": 2.2609060402684566e-05,
"loss": 0.7425,
"step": 753
},
{
"epoch": 1.1662156938538848,
"grad_norm": 0.5840611457824707,
"learning_rate": 2.2567114093959733e-05,
"loss": 0.715,
"step": 754
},
{
"epoch": 1.1677618863548511,
"grad_norm": 0.534000039100647,
"learning_rate": 2.25251677852349e-05,
"loss": 0.7764,
"step": 755
},
{
"epoch": 1.1693080788558174,
"grad_norm": 0.6020839214324951,
"learning_rate": 2.248322147651007e-05,
"loss": 0.7502,
"step": 756
},
{
"epoch": 1.170854271356784,
"grad_norm": 0.5764395594596863,
"learning_rate": 2.2441275167785236e-05,
"loss": 0.7298,
"step": 757
},
{
"epoch": 1.1724004638577503,
"grad_norm": 0.5840582251548767,
"learning_rate": 2.2399328859060403e-05,
"loss": 0.7836,
"step": 758
},
{
"epoch": 1.1739466563587166,
"grad_norm": 0.5759351253509521,
"learning_rate": 2.235738255033557e-05,
"loss": 0.7646,
"step": 759
},
{
"epoch": 1.175492848859683,
"grad_norm": 0.5777841806411743,
"learning_rate": 2.231543624161074e-05,
"loss": 0.8351,
"step": 760
},
{
"epoch": 1.1770390413606493,
"grad_norm": 0.6419858932495117,
"learning_rate": 2.227348993288591e-05,
"loss": 0.6833,
"step": 761
},
{
"epoch": 1.1785852338616158,
"grad_norm": 0.5851649045944214,
"learning_rate": 2.2231543624161076e-05,
"loss": 0.7559,
"step": 762
},
{
"epoch": 1.1801314263625822,
"grad_norm": 0.6243789196014404,
"learning_rate": 2.2189597315436242e-05,
"loss": 0.6587,
"step": 763
},
{
"epoch": 1.1816776188635485,
"grad_norm": 0.6107107996940613,
"learning_rate": 2.2147651006711412e-05,
"loss": 0.7724,
"step": 764
},
{
"epoch": 1.1832238113645148,
"grad_norm": 0.6164106130599976,
"learning_rate": 2.210570469798658e-05,
"loss": 0.6949,
"step": 765
},
{
"epoch": 1.1847700038654811,
"grad_norm": 0.6295919418334961,
"learning_rate": 2.2063758389261745e-05,
"loss": 0.7239,
"step": 766
},
{
"epoch": 1.1863161963664477,
"grad_norm": 0.582129955291748,
"learning_rate": 2.2021812080536912e-05,
"loss": 0.7619,
"step": 767
},
{
"epoch": 1.187862388867414,
"grad_norm": 0.639700174331665,
"learning_rate": 2.197986577181208e-05,
"loss": 0.7788,
"step": 768
},
{
"epoch": 1.1894085813683803,
"grad_norm": 0.6275160908699036,
"learning_rate": 2.193791946308725e-05,
"loss": 0.7457,
"step": 769
},
{
"epoch": 1.1909547738693467,
"grad_norm": 0.5969827175140381,
"learning_rate": 2.1895973154362418e-05,
"loss": 0.7981,
"step": 770
},
{
"epoch": 1.192500966370313,
"grad_norm": 0.5809218883514404,
"learning_rate": 2.1854026845637585e-05,
"loss": 0.8194,
"step": 771
},
{
"epoch": 1.1940471588712795,
"grad_norm": 0.5926761627197266,
"learning_rate": 2.181208053691275e-05,
"loss": 0.7808,
"step": 772
},
{
"epoch": 1.1955933513722459,
"grad_norm": 0.6315684914588928,
"learning_rate": 2.177013422818792e-05,
"loss": 0.8342,
"step": 773
},
{
"epoch": 1.1971395438732122,
"grad_norm": 0.5824242830276489,
"learning_rate": 2.1728187919463088e-05,
"loss": 0.8041,
"step": 774
},
{
"epoch": 1.1986857363741785,
"grad_norm": 0.6280264258384705,
"learning_rate": 2.1686241610738254e-05,
"loss": 0.7976,
"step": 775
},
{
"epoch": 1.2002319288751448,
"grad_norm": 0.6388978958129883,
"learning_rate": 2.1644295302013424e-05,
"loss": 0.7253,
"step": 776
},
{
"epoch": 1.2017781213761114,
"grad_norm": 0.6550062894821167,
"learning_rate": 2.160234899328859e-05,
"loss": 0.7531,
"step": 777
},
{
"epoch": 1.2033243138770777,
"grad_norm": 0.6625634431838989,
"learning_rate": 2.156040268456376e-05,
"loss": 0.7869,
"step": 778
},
{
"epoch": 1.204870506378044,
"grad_norm": 0.6421430706977844,
"learning_rate": 2.1518456375838927e-05,
"loss": 0.8198,
"step": 779
},
{
"epoch": 1.2064166988790104,
"grad_norm": 0.6213207840919495,
"learning_rate": 2.1476510067114094e-05,
"loss": 0.8565,
"step": 780
},
{
"epoch": 1.2079628913799767,
"grad_norm": 0.6527850031852722,
"learning_rate": 2.143456375838926e-05,
"loss": 0.8084,
"step": 781
},
{
"epoch": 1.2095090838809432,
"grad_norm": 0.6330167651176453,
"learning_rate": 2.139261744966443e-05,
"loss": 0.8062,
"step": 782
},
{
"epoch": 1.2110552763819096,
"grad_norm": 0.6649383902549744,
"learning_rate": 2.13506711409396e-05,
"loss": 0.7586,
"step": 783
},
{
"epoch": 1.212601468882876,
"grad_norm": 0.6256992220878601,
"learning_rate": 2.1308724832214767e-05,
"loss": 0.7785,
"step": 784
},
{
"epoch": 1.2141476613838422,
"grad_norm": 0.622163712978363,
"learning_rate": 2.1266778523489934e-05,
"loss": 0.9157,
"step": 785
},
{
"epoch": 1.2156938538848086,
"grad_norm": 0.6411583423614502,
"learning_rate": 2.1224832214765103e-05,
"loss": 0.8916,
"step": 786
},
{
"epoch": 1.217240046385775,
"grad_norm": 0.6612896919250488,
"learning_rate": 2.118288590604027e-05,
"loss": 0.8234,
"step": 787
},
{
"epoch": 1.2187862388867414,
"grad_norm": 0.7050024271011353,
"learning_rate": 2.1140939597315437e-05,
"loss": 0.9097,
"step": 788
},
{
"epoch": 1.2203324313877078,
"grad_norm": 0.7046983242034912,
"learning_rate": 2.1098993288590603e-05,
"loss": 0.8503,
"step": 789
},
{
"epoch": 1.221878623888674,
"grad_norm": 0.7282384634017944,
"learning_rate": 2.1057046979865773e-05,
"loss": 0.8317,
"step": 790
},
{
"epoch": 1.2234248163896404,
"grad_norm": 0.7514353394508362,
"learning_rate": 2.1015100671140943e-05,
"loss": 0.8562,
"step": 791
},
{
"epoch": 1.224971008890607,
"grad_norm": 0.6923422813415527,
"learning_rate": 2.097315436241611e-05,
"loss": 0.8234,
"step": 792
},
{
"epoch": 1.2265172013915733,
"grad_norm": 0.7773630023002625,
"learning_rate": 2.0931208053691276e-05,
"loss": 0.8988,
"step": 793
},
{
"epoch": 1.2280633938925396,
"grad_norm": 0.8075311183929443,
"learning_rate": 2.0889261744966443e-05,
"loss": 0.8421,
"step": 794
},
{
"epoch": 1.229609586393506,
"grad_norm": 0.8537722826004028,
"learning_rate": 2.0847315436241613e-05,
"loss": 0.835,
"step": 795
},
{
"epoch": 1.2311557788944723,
"grad_norm": 1.1471562385559082,
"learning_rate": 2.080536912751678e-05,
"loss": 0.9985,
"step": 796
},
{
"epoch": 1.2327019713954388,
"grad_norm": 0.8220142126083374,
"learning_rate": 2.0763422818791946e-05,
"loss": 0.6295,
"step": 797
},
{
"epoch": 1.2342481638964051,
"grad_norm": 0.6230762600898743,
"learning_rate": 2.0721476510067116e-05,
"loss": 0.6921,
"step": 798
},
{
"epoch": 1.2357943563973715,
"grad_norm": 0.596222996711731,
"learning_rate": 2.0679530201342286e-05,
"loss": 0.7503,
"step": 799
},
{
"epoch": 1.2373405488983378,
"grad_norm": 0.5620123744010925,
"learning_rate": 2.0637583892617452e-05,
"loss": 0.6849,
"step": 800
},
{
"epoch": 1.238886741399304,
"grad_norm": 0.5710748434066772,
"learning_rate": 2.059563758389262e-05,
"loss": 0.7431,
"step": 801
},
{
"epoch": 1.2404329339002707,
"grad_norm": 0.584814727306366,
"learning_rate": 2.0553691275167785e-05,
"loss": 0.721,
"step": 802
},
{
"epoch": 1.241979126401237,
"grad_norm": 0.576964795589447,
"learning_rate": 2.0511744966442952e-05,
"loss": 0.6897,
"step": 803
},
{
"epoch": 1.2435253189022033,
"grad_norm": 0.6063031554222107,
"learning_rate": 2.0469798657718122e-05,
"loss": 0.7516,
"step": 804
},
{
"epoch": 1.2450715114031696,
"grad_norm": 0.596976101398468,
"learning_rate": 2.042785234899329e-05,
"loss": 0.7488,
"step": 805
},
{
"epoch": 1.246617703904136,
"grad_norm": 0.6499916911125183,
"learning_rate": 2.038590604026846e-05,
"loss": 0.7554,
"step": 806
},
{
"epoch": 1.2481638964051025,
"grad_norm": 0.5964358448982239,
"learning_rate": 2.0343959731543625e-05,
"loss": 0.7417,
"step": 807
},
{
"epoch": 1.2497100889060688,
"grad_norm": 0.5935381650924683,
"learning_rate": 2.0302013422818795e-05,
"loss": 0.7308,
"step": 808
},
{
"epoch": 1.2512562814070352,
"grad_norm": 0.6293373703956604,
"learning_rate": 2.026006711409396e-05,
"loss": 0.734,
"step": 809
},
{
"epoch": 1.2528024739080015,
"grad_norm": 0.592042088508606,
"learning_rate": 2.0218120805369128e-05,
"loss": 0.7412,
"step": 810
},
{
"epoch": 1.2543486664089678,
"grad_norm": 0.5861983895301819,
"learning_rate": 2.0176174496644295e-05,
"loss": 0.8101,
"step": 811
},
{
"epoch": 1.2558948589099344,
"grad_norm": 0.6101320385932922,
"learning_rate": 2.013422818791946e-05,
"loss": 0.7394,
"step": 812
},
{
"epoch": 1.2574410514109007,
"grad_norm": 0.6127662062644958,
"learning_rate": 2.009228187919463e-05,
"loss": 0.7941,
"step": 813
},
{
"epoch": 1.258987243911867,
"grad_norm": 0.5885155200958252,
"learning_rate": 2.00503355704698e-05,
"loss": 0.7987,
"step": 814
},
{
"epoch": 1.2605334364128333,
"grad_norm": 0.6043751835823059,
"learning_rate": 2.0008389261744968e-05,
"loss": 0.7918,
"step": 815
},
{
"epoch": 1.2620796289137997,
"grad_norm": 0.6351797580718994,
"learning_rate": 1.9966442953020134e-05,
"loss": 0.7218,
"step": 816
},
{
"epoch": 1.2636258214147662,
"grad_norm": 0.6218336224555969,
"learning_rate": 1.9924496644295304e-05,
"loss": 0.8294,
"step": 817
},
{
"epoch": 1.2651720139157325,
"grad_norm": 0.6148021817207336,
"learning_rate": 1.988255033557047e-05,
"loss": 0.7767,
"step": 818
},
{
"epoch": 1.2667182064166989,
"grad_norm": 0.6303946375846863,
"learning_rate": 1.9840604026845637e-05,
"loss": 0.7816,
"step": 819
},
{
"epoch": 1.2682643989176652,
"grad_norm": 0.6076005101203918,
"learning_rate": 1.9798657718120804e-05,
"loss": 0.8023,
"step": 820
},
{
"epoch": 1.2698105914186315,
"grad_norm": 0.6453294157981873,
"learning_rate": 1.9756711409395974e-05,
"loss": 0.7728,
"step": 821
},
{
"epoch": 1.271356783919598,
"grad_norm": 0.6279742121696472,
"learning_rate": 1.9714765100671144e-05,
"loss": 0.7564,
"step": 822
},
{
"epoch": 1.2729029764205644,
"grad_norm": 0.6434690356254578,
"learning_rate": 1.967281879194631e-05,
"loss": 0.7989,
"step": 823
},
{
"epoch": 1.2744491689215307,
"grad_norm": 0.6816707253456116,
"learning_rate": 1.9630872483221477e-05,
"loss": 0.7964,
"step": 824
},
{
"epoch": 1.275995361422497,
"grad_norm": 0.6387498378753662,
"learning_rate": 1.9588926174496643e-05,
"loss": 0.7621,
"step": 825
},
{
"epoch": 1.2775415539234634,
"grad_norm": 0.6609524488449097,
"learning_rate": 1.9546979865771813e-05,
"loss": 0.8252,
"step": 826
},
{
"epoch": 1.27908774642443,
"grad_norm": 0.6324920654296875,
"learning_rate": 1.950503355704698e-05,
"loss": 0.7502,
"step": 827
},
{
"epoch": 1.2806339389253962,
"grad_norm": 0.6569236516952515,
"learning_rate": 1.946308724832215e-05,
"loss": 0.8148,
"step": 828
},
{
"epoch": 1.2821801314263626,
"grad_norm": 0.625497043132782,
"learning_rate": 1.9421140939597316e-05,
"loss": 0.8663,
"step": 829
},
{
"epoch": 1.2837263239273289,
"grad_norm": 0.6833832263946533,
"learning_rate": 1.9379194630872486e-05,
"loss": 0.692,
"step": 830
},
{
"epoch": 1.2852725164282952,
"grad_norm": 0.6534375548362732,
"learning_rate": 1.9337248322147653e-05,
"loss": 0.7731,
"step": 831
},
{
"epoch": 1.2868187089292618,
"grad_norm": 0.6668411493301392,
"learning_rate": 1.929530201342282e-05,
"loss": 0.758,
"step": 832
},
{
"epoch": 1.288364901430228,
"grad_norm": 0.6552398204803467,
"learning_rate": 1.9253355704697986e-05,
"loss": 0.8387,
"step": 833
},
{
"epoch": 1.2899110939311944,
"grad_norm": 0.6804989576339722,
"learning_rate": 1.9211409395973153e-05,
"loss": 0.823,
"step": 834
},
{
"epoch": 1.2914572864321607,
"grad_norm": 0.6738516092300415,
"learning_rate": 1.9169463087248323e-05,
"loss": 0.8759,
"step": 835
},
{
"epoch": 1.293003478933127,
"grad_norm": 0.6957226991653442,
"learning_rate": 1.9127516778523493e-05,
"loss": 0.8465,
"step": 836
},
{
"epoch": 1.2945496714340936,
"grad_norm": 0.7255749106407166,
"learning_rate": 1.908557046979866e-05,
"loss": 0.7872,
"step": 837
},
{
"epoch": 1.29609586393506,
"grad_norm": 0.7283822894096375,
"learning_rate": 1.9043624161073826e-05,
"loss": 0.8759,
"step": 838
},
{
"epoch": 1.2976420564360263,
"grad_norm": 0.7126689553260803,
"learning_rate": 1.9001677852348996e-05,
"loss": 0.8252,
"step": 839
},
{
"epoch": 1.2991882489369926,
"grad_norm": 0.7381671071052551,
"learning_rate": 1.8959731543624162e-05,
"loss": 0.9227,
"step": 840
},
{
"epoch": 1.300734441437959,
"grad_norm": 0.721825897693634,
"learning_rate": 1.891778523489933e-05,
"loss": 0.9147,
"step": 841
},
{
"epoch": 1.3022806339389255,
"grad_norm": 0.719792902469635,
"learning_rate": 1.8875838926174495e-05,
"loss": 0.9038,
"step": 842
},
{
"epoch": 1.3038268264398918,
"grad_norm": 0.7600436806678772,
"learning_rate": 1.8833892617449665e-05,
"loss": 0.903,
"step": 843
},
{
"epoch": 1.305373018940858,
"grad_norm": 0.743240475654602,
"learning_rate": 1.8791946308724835e-05,
"loss": 0.8799,
"step": 844
},
{
"epoch": 1.3069192114418244,
"grad_norm": 0.8364670872688293,
"learning_rate": 1.8750000000000002e-05,
"loss": 0.9059,
"step": 845
},
{
"epoch": 1.3084654039427908,
"grad_norm": 0.9684616327285767,
"learning_rate": 1.870805369127517e-05,
"loss": 0.9468,
"step": 846
},
{
"epoch": 1.3100115964437573,
"grad_norm": 0.7947605848312378,
"learning_rate": 1.8666107382550335e-05,
"loss": 0.6681,
"step": 847
},
{
"epoch": 1.3115577889447236,
"grad_norm": 0.5737844109535217,
"learning_rate": 1.8624161073825505e-05,
"loss": 0.6837,
"step": 848
},
{
"epoch": 1.31310398144569,
"grad_norm": 0.5890055298805237,
"learning_rate": 1.858221476510067e-05,
"loss": 0.7115,
"step": 849
},
{
"epoch": 1.3146501739466563,
"grad_norm": 0.6102372407913208,
"learning_rate": 1.8540268456375838e-05,
"loss": 0.7131,
"step": 850
},
{
"epoch": 1.3161963664476226,
"grad_norm": 0.5825210809707642,
"learning_rate": 1.8498322147651008e-05,
"loss": 0.6873,
"step": 851
},
{
"epoch": 1.3177425589485892,
"grad_norm": 0.5857095718383789,
"learning_rate": 1.8456375838926178e-05,
"loss": 0.7435,
"step": 852
},
{
"epoch": 1.3192887514495555,
"grad_norm": 0.60005122423172,
"learning_rate": 1.8414429530201344e-05,
"loss": 0.6988,
"step": 853
},
{
"epoch": 1.3208349439505218,
"grad_norm": 0.6294332146644592,
"learning_rate": 1.837248322147651e-05,
"loss": 0.7796,
"step": 854
},
{
"epoch": 1.3223811364514881,
"grad_norm": 0.6006381511688232,
"learning_rate": 1.8330536912751678e-05,
"loss": 0.7677,
"step": 855
},
{
"epoch": 1.3239273289524545,
"grad_norm": 0.5754826664924622,
"learning_rate": 1.8288590604026847e-05,
"loss": 0.7845,
"step": 856
},
{
"epoch": 1.325473521453421,
"grad_norm": 0.6118499040603638,
"learning_rate": 1.8246644295302014e-05,
"loss": 0.6819,
"step": 857
},
{
"epoch": 1.3270197139543873,
"grad_norm": 0.5924245119094849,
"learning_rate": 1.820469798657718e-05,
"loss": 0.8092,
"step": 858
},
{
"epoch": 1.3285659064553537,
"grad_norm": 0.6202556490898132,
"learning_rate": 1.816275167785235e-05,
"loss": 0.7992,
"step": 859
},
{
"epoch": 1.33011209895632,
"grad_norm": 0.6356550455093384,
"learning_rate": 1.8120805369127517e-05,
"loss": 0.7731,
"step": 860
},
{
"epoch": 1.3316582914572863,
"grad_norm": 0.6205728650093079,
"learning_rate": 1.8078859060402687e-05,
"loss": 0.7951,
"step": 861
},
{
"epoch": 1.3332044839582529,
"grad_norm": 0.6306980848312378,
"learning_rate": 1.8036912751677854e-05,
"loss": 0.7904,
"step": 862
},
{
"epoch": 1.3347506764592192,
"grad_norm": 0.5935050249099731,
"learning_rate": 1.799496644295302e-05,
"loss": 0.8029,
"step": 863
},
{
"epoch": 1.3362968689601855,
"grad_norm": 0.6275442838668823,
"learning_rate": 1.7953020134228187e-05,
"loss": 0.7376,
"step": 864
},
{
"epoch": 1.3378430614611518,
"grad_norm": 0.6528346538543701,
"learning_rate": 1.7911073825503357e-05,
"loss": 0.7286,
"step": 865
},
{
"epoch": 1.3393892539621182,
"grad_norm": 0.6417142748832703,
"learning_rate": 1.7869127516778523e-05,
"loss": 0.7632,
"step": 866
},
{
"epoch": 1.3409354464630847,
"grad_norm": 0.5968343615531921,
"learning_rate": 1.7827181208053693e-05,
"loss": 0.8156,
"step": 867
},
{
"epoch": 1.342481638964051,
"grad_norm": 0.5820614695549011,
"learning_rate": 1.778523489932886e-05,
"loss": 0.7957,
"step": 868
},
{
"epoch": 1.3440278314650174,
"grad_norm": 0.576564610004425,
"learning_rate": 1.7743288590604026e-05,
"loss": 0.7799,
"step": 869
},
{
"epoch": 1.3455740239659837,
"grad_norm": 0.6096657514572144,
"learning_rate": 1.7701342281879196e-05,
"loss": 0.7929,
"step": 870
},
{
"epoch": 1.34712021646695,
"grad_norm": 0.6413806080818176,
"learning_rate": 1.7659395973154363e-05,
"loss": 0.7468,
"step": 871
},
{
"epoch": 1.3486664089679166,
"grad_norm": 0.6387087106704712,
"learning_rate": 1.761744966442953e-05,
"loss": 0.7342,
"step": 872
},
{
"epoch": 1.350212601468883,
"grad_norm": 0.6342037320137024,
"learning_rate": 1.75755033557047e-05,
"loss": 0.829,
"step": 873
},
{
"epoch": 1.3517587939698492,
"grad_norm": 0.6201750636100769,
"learning_rate": 1.753355704697987e-05,
"loss": 0.7289,
"step": 874
},
{
"epoch": 1.3533049864708155,
"grad_norm": 0.6321792006492615,
"learning_rate": 1.7491610738255036e-05,
"loss": 0.7253,
"step": 875
},
{
"epoch": 1.3548511789717819,
"grad_norm": 0.6561968922615051,
"learning_rate": 1.7449664429530202e-05,
"loss": 0.7157,
"step": 876
},
{
"epoch": 1.3563973714727484,
"grad_norm": 0.6112854480743408,
"learning_rate": 1.740771812080537e-05,
"loss": 0.7421,
"step": 877
},
{
"epoch": 1.3579435639737147,
"grad_norm": 0.6668578386306763,
"learning_rate": 1.736577181208054e-05,
"loss": 0.7849,
"step": 878
},
{
"epoch": 1.359489756474681,
"grad_norm": 0.6496410965919495,
"learning_rate": 1.7323825503355705e-05,
"loss": 0.8031,
"step": 879
},
{
"epoch": 1.3610359489756474,
"grad_norm": 0.6386198401451111,
"learning_rate": 1.7281879194630872e-05,
"loss": 0.8812,
"step": 880
},
{
"epoch": 1.3625821414766137,
"grad_norm": 0.6844744682312012,
"learning_rate": 1.7239932885906042e-05,
"loss": 0.8245,
"step": 881
},
{
"epoch": 1.3641283339775803,
"grad_norm": 0.6815952658653259,
"learning_rate": 1.719798657718121e-05,
"loss": 0.8751,
"step": 882
},
{
"epoch": 1.3656745264785466,
"grad_norm": 0.6746396422386169,
"learning_rate": 1.715604026845638e-05,
"loss": 0.812,
"step": 883
},
{
"epoch": 1.367220718979513,
"grad_norm": 0.6984376907348633,
"learning_rate": 1.7114093959731545e-05,
"loss": 0.8418,
"step": 884
},
{
"epoch": 1.3687669114804792,
"grad_norm": 0.6949034929275513,
"learning_rate": 1.707214765100671e-05,
"loss": 0.8032,
"step": 885
},
{
"epoch": 1.3703131039814456,
"grad_norm": 0.6689850091934204,
"learning_rate": 1.7030201342281878e-05,
"loss": 0.8363,
"step": 886
},
{
"epoch": 1.3718592964824121,
"grad_norm": 0.6734853386878967,
"learning_rate": 1.6988255033557048e-05,
"loss": 0.8643,
"step": 887
},
{
"epoch": 1.3734054889833784,
"grad_norm": 0.7207046747207642,
"learning_rate": 1.6946308724832215e-05,
"loss": 0.853,
"step": 888
},
{
"epoch": 1.3749516814843448,
"grad_norm": 0.7100691199302673,
"learning_rate": 1.6904362416107385e-05,
"loss": 0.8776,
"step": 889
},
{
"epoch": 1.376497873985311,
"grad_norm": 0.7267347574234009,
"learning_rate": 1.686241610738255e-05,
"loss": 0.817,
"step": 890
},
{
"epoch": 1.3780440664862774,
"grad_norm": 0.7453994154930115,
"learning_rate": 1.6820469798657718e-05,
"loss": 0.7864,
"step": 891
},
{
"epoch": 1.379590258987244,
"grad_norm": 0.7423779368400574,
"learning_rate": 1.6778523489932888e-05,
"loss": 0.8475,
"step": 892
},
{
"epoch": 1.3811364514882103,
"grad_norm": 0.7435901165008545,
"learning_rate": 1.6736577181208054e-05,
"loss": 0.8409,
"step": 893
},
{
"epoch": 1.3826826439891766,
"grad_norm": 0.8185101747512817,
"learning_rate": 1.669463087248322e-05,
"loss": 0.9255,
"step": 894
},
{
"epoch": 1.384228836490143,
"grad_norm": 0.8886857628822327,
"learning_rate": 1.6652684563758387e-05,
"loss": 0.9095,
"step": 895
},
{
"epoch": 1.3857750289911093,
"grad_norm": 1.2426859140396118,
"learning_rate": 1.6610738255033557e-05,
"loss": 0.9899,
"step": 896
},
{
"epoch": 1.3873212214920758,
"grad_norm": 0.7389244437217712,
"learning_rate": 1.6568791946308727e-05,
"loss": 0.5189,
"step": 897
},
{
"epoch": 1.3888674139930421,
"grad_norm": 0.5934030413627625,
"learning_rate": 1.6526845637583894e-05,
"loss": 0.6868,
"step": 898
},
{
"epoch": 1.3904136064940085,
"grad_norm": 0.5986191630363464,
"learning_rate": 1.648489932885906e-05,
"loss": 0.7075,
"step": 899
},
{
"epoch": 1.3919597989949748,
"grad_norm": 0.6108096241950989,
"learning_rate": 1.644295302013423e-05,
"loss": 0.6855,
"step": 900
},
{
"epoch": 1.3935059914959411,
"grad_norm": 0.5924285650253296,
"learning_rate": 1.6401006711409397e-05,
"loss": 0.7459,
"step": 901
},
{
"epoch": 1.3950521839969077,
"grad_norm": 0.627100944519043,
"learning_rate": 1.6359060402684563e-05,
"loss": 0.71,
"step": 902
},
{
"epoch": 1.396598376497874,
"grad_norm": 0.6151503920555115,
"learning_rate": 1.631711409395973e-05,
"loss": 0.6844,
"step": 903
},
{
"epoch": 1.3981445689988403,
"grad_norm": 0.6181209683418274,
"learning_rate": 1.62751677852349e-05,
"loss": 0.7635,
"step": 904
},
{
"epoch": 1.3996907614998066,
"grad_norm": 0.5866500735282898,
"learning_rate": 1.623322147651007e-05,
"loss": 0.742,
"step": 905
},
{
"epoch": 1.401236954000773,
"grad_norm": 0.6311929225921631,
"learning_rate": 1.6191275167785237e-05,
"loss": 0.734,
"step": 906
},
{
"epoch": 1.4027831465017395,
"grad_norm": 0.6104720830917358,
"learning_rate": 1.6149328859060403e-05,
"loss": 0.7187,
"step": 907
},
{
"epoch": 1.4043293390027058,
"grad_norm": 0.6044167280197144,
"learning_rate": 1.610738255033557e-05,
"loss": 0.717,
"step": 908
},
{
"epoch": 1.4058755315036722,
"grad_norm": 0.5984567999839783,
"learning_rate": 1.606543624161074e-05,
"loss": 0.8021,
"step": 909
},
{
"epoch": 1.4074217240046385,
"grad_norm": 0.6355494260787964,
"learning_rate": 1.6023489932885906e-05,
"loss": 0.8507,
"step": 910
},
{
"epoch": 1.4089679165056048,
"grad_norm": 0.6177673935890198,
"learning_rate": 1.5981543624161076e-05,
"loss": 0.7205,
"step": 911
},
{
"epoch": 1.4105141090065714,
"grad_norm": 0.6486150026321411,
"learning_rate": 1.5939597315436243e-05,
"loss": 0.8078,
"step": 912
},
{
"epoch": 1.4120603015075377,
"grad_norm": 0.6070188879966736,
"learning_rate": 1.5897651006711413e-05,
"loss": 0.7067,
"step": 913
},
{
"epoch": 1.413606494008504,
"grad_norm": 0.6363996267318726,
"learning_rate": 1.585570469798658e-05,
"loss": 0.7621,
"step": 914
},
{
"epoch": 1.4151526865094703,
"grad_norm": 0.6805376410484314,
"learning_rate": 1.5813758389261746e-05,
"loss": 0.7354,
"step": 915
},
{
"epoch": 1.4166988790104367,
"grad_norm": 0.6029914617538452,
"learning_rate": 1.5771812080536912e-05,
"loss": 0.7425,
"step": 916
},
{
"epoch": 1.4182450715114032,
"grad_norm": 0.6197159290313721,
"learning_rate": 1.572986577181208e-05,
"loss": 0.769,
"step": 917
},
{
"epoch": 1.4197912640123695,
"grad_norm": 0.6450138092041016,
"learning_rate": 1.568791946308725e-05,
"loss": 0.7455,
"step": 918
},
{
"epoch": 1.4213374565133359,
"grad_norm": 0.6024583578109741,
"learning_rate": 1.564597315436242e-05,
"loss": 0.783,
"step": 919
},
{
"epoch": 1.4228836490143022,
"grad_norm": 0.6291201114654541,
"learning_rate": 1.5604026845637585e-05,
"loss": 0.7888,
"step": 920
},
{
"epoch": 1.4244298415152685,
"grad_norm": 0.6113951802253723,
"learning_rate": 1.5562080536912752e-05,
"loss": 0.7682,
"step": 921
},
{
"epoch": 1.425976034016235,
"grad_norm": 0.6421045064926147,
"learning_rate": 1.5520134228187922e-05,
"loss": 0.7691,
"step": 922
},
{
"epoch": 1.4275222265172014,
"grad_norm": 0.6352230906486511,
"learning_rate": 1.547818791946309e-05,
"loss": 0.7902,
"step": 923
},
{
"epoch": 1.4290684190181677,
"grad_norm": 0.677621066570282,
"learning_rate": 1.5436241610738255e-05,
"loss": 0.765,
"step": 924
},
{
"epoch": 1.430614611519134,
"grad_norm": 0.6393842697143555,
"learning_rate": 1.539429530201342e-05,
"loss": 0.6943,
"step": 925
},
{
"epoch": 1.4321608040201004,
"grad_norm": 0.6991260051727295,
"learning_rate": 1.535234899328859e-05,
"loss": 0.7966,
"step": 926
},
{
"epoch": 1.433706996521067,
"grad_norm": 0.642305314540863,
"learning_rate": 1.531040268456376e-05,
"loss": 0.8245,
"step": 927
},
{
"epoch": 1.4352531890220332,
"grad_norm": 0.6572315692901611,
"learning_rate": 1.5268456375838928e-05,
"loss": 0.7562,
"step": 928
},
{
"epoch": 1.4367993815229996,
"grad_norm": 0.6849440336227417,
"learning_rate": 1.5226510067114095e-05,
"loss": 0.8522,
"step": 929
},
{
"epoch": 1.4383455740239661,
"grad_norm": 0.7153504490852356,
"learning_rate": 1.5184563758389261e-05,
"loss": 0.7843,
"step": 930
},
{
"epoch": 1.4398917665249322,
"grad_norm": 0.6780909895896912,
"learning_rate": 1.5142617449664431e-05,
"loss": 0.8062,
"step": 931
},
{
"epoch": 1.4414379590258988,
"grad_norm": 0.6425846815109253,
"learning_rate": 1.51006711409396e-05,
"loss": 0.7866,
"step": 932
},
{
"epoch": 1.442984151526865,
"grad_norm": 0.6629062294960022,
"learning_rate": 1.5058724832214766e-05,
"loss": 0.8047,
"step": 933
},
{
"epoch": 1.4445303440278314,
"grad_norm": 0.6615095734596252,
"learning_rate": 1.5016778523489932e-05,
"loss": 0.8117,
"step": 934
},
{
"epoch": 1.446076536528798,
"grad_norm": 0.718272864818573,
"learning_rate": 1.4974832214765102e-05,
"loss": 0.8416,
"step": 935
},
{
"epoch": 1.447622729029764,
"grad_norm": 0.7286227941513062,
"learning_rate": 1.493288590604027e-05,
"loss": 0.8347,
"step": 936
},
{
"epoch": 1.4491689215307306,
"grad_norm": 0.7250097990036011,
"learning_rate": 1.4890939597315437e-05,
"loss": 0.8095,
"step": 937
},
{
"epoch": 1.450715114031697,
"grad_norm": 0.7696204781532288,
"learning_rate": 1.4848993288590604e-05,
"loss": 0.8611,
"step": 938
},
{
"epoch": 1.4522613065326633,
"grad_norm": 0.7375919222831726,
"learning_rate": 1.4807046979865772e-05,
"loss": 0.8903,
"step": 939
},
{
"epoch": 1.4538074990336298,
"grad_norm": 0.7759246230125427,
"learning_rate": 1.4765100671140942e-05,
"loss": 0.8005,
"step": 940
},
{
"epoch": 1.455353691534596,
"grad_norm": 0.8237205743789673,
"learning_rate": 1.4723154362416108e-05,
"loss": 0.8204,
"step": 941
},
{
"epoch": 1.4568998840355625,
"grad_norm": 0.7622566819190979,
"learning_rate": 1.4681208053691275e-05,
"loss": 0.8673,
"step": 942
},
{
"epoch": 1.4584460765365288,
"grad_norm": 0.7782172560691833,
"learning_rate": 1.4639261744966443e-05,
"loss": 0.9434,
"step": 943
},
{
"epoch": 1.4599922690374951,
"grad_norm": 0.856338381767273,
"learning_rate": 1.4597315436241613e-05,
"loss": 0.9032,
"step": 944
},
{
"epoch": 1.4615384615384617,
"grad_norm": 0.9459933042526245,
"learning_rate": 1.455536912751678e-05,
"loss": 0.8599,
"step": 945
},
{
"epoch": 1.4630846540394278,
"grad_norm": 1.1424592733383179,
"learning_rate": 1.4513422818791946e-05,
"loss": 0.879,
"step": 946
},
{
"epoch": 1.4646308465403943,
"grad_norm": 0.9243542551994324,
"learning_rate": 1.4471476510067115e-05,
"loss": 0.6613,
"step": 947
},
{
"epoch": 1.4661770390413607,
"grad_norm": 0.5729179978370667,
"learning_rate": 1.4429530201342285e-05,
"loss": 0.6693,
"step": 948
},
{
"epoch": 1.467723231542327,
"grad_norm": 0.6040563583374023,
"learning_rate": 1.4387583892617451e-05,
"loss": 0.663,
"step": 949
},
{
"epoch": 1.4692694240432935,
"grad_norm": 0.6559063196182251,
"learning_rate": 1.4345637583892618e-05,
"loss": 0.691,
"step": 950
},
{
"epoch": 1.4708156165442596,
"grad_norm": 0.5916107296943665,
"learning_rate": 1.4303691275167786e-05,
"loss": 0.7474,
"step": 951
},
{
"epoch": 1.4723618090452262,
"grad_norm": 0.5876263380050659,
"learning_rate": 1.4261744966442953e-05,
"loss": 0.7287,
"step": 952
},
{
"epoch": 1.4739080015461925,
"grad_norm": 0.5855628848075867,
"learning_rate": 1.4219798657718122e-05,
"loss": 0.7254,
"step": 953
},
{
"epoch": 1.4754541940471588,
"grad_norm": 0.5839198231697083,
"learning_rate": 1.4177852348993289e-05,
"loss": 0.7462,
"step": 954
},
{
"epoch": 1.4770003865481254,
"grad_norm": 0.5933511257171631,
"learning_rate": 1.4135906040268457e-05,
"loss": 0.7303,
"step": 955
},
{
"epoch": 1.4785465790490915,
"grad_norm": 0.620617687702179,
"learning_rate": 1.4093959731543624e-05,
"loss": 0.7287,
"step": 956
},
{
"epoch": 1.480092771550058,
"grad_norm": 0.6549968719482422,
"learning_rate": 1.4052013422818794e-05,
"loss": 0.7426,
"step": 957
},
{
"epoch": 1.4816389640510244,
"grad_norm": 0.6639398336410522,
"learning_rate": 1.401006711409396e-05,
"loss": 0.7329,
"step": 958
},
{
"epoch": 1.4831851565519907,
"grad_norm": 0.6285966634750366,
"learning_rate": 1.3968120805369129e-05,
"loss": 0.7302,
"step": 959
},
{
"epoch": 1.4847313490529572,
"grad_norm": 0.6673871278762817,
"learning_rate": 1.3926174496644295e-05,
"loss": 0.7165,
"step": 960
},
{
"epoch": 1.4862775415539233,
"grad_norm": 0.6557399034500122,
"learning_rate": 1.3884228187919462e-05,
"loss": 0.7465,
"step": 961
},
{
"epoch": 1.4878237340548899,
"grad_norm": 0.615143358707428,
"learning_rate": 1.3842281879194632e-05,
"loss": 0.7531,
"step": 962
},
{
"epoch": 1.4893699265558562,
"grad_norm": 0.6445402503013611,
"learning_rate": 1.38003355704698e-05,
"loss": 0.7088,
"step": 963
},
{
"epoch": 1.4909161190568225,
"grad_norm": 0.6428129076957703,
"learning_rate": 1.3758389261744966e-05,
"loss": 0.7011,
"step": 964
},
{
"epoch": 1.492462311557789,
"grad_norm": 0.6472491025924683,
"learning_rate": 1.3716442953020133e-05,
"loss": 0.8012,
"step": 965
},
{
"epoch": 1.4940085040587552,
"grad_norm": 0.6467485427856445,
"learning_rate": 1.3674496644295303e-05,
"loss": 0.7108,
"step": 966
},
{
"epoch": 1.4955546965597217,
"grad_norm": 0.648501455783844,
"learning_rate": 1.3632550335570471e-05,
"loss": 0.7786,
"step": 967
},
{
"epoch": 1.497100889060688,
"grad_norm": 0.6614555716514587,
"learning_rate": 1.3590604026845638e-05,
"loss": 0.6745,
"step": 968
},
{
"epoch": 1.4986470815616544,
"grad_norm": 0.6633491516113281,
"learning_rate": 1.3548657718120804e-05,
"loss": 0.7064,
"step": 969
},
{
"epoch": 1.500193274062621,
"grad_norm": 0.658253014087677,
"learning_rate": 1.3506711409395974e-05,
"loss": 0.7438,
"step": 970
},
{
"epoch": 1.501739466563587,
"grad_norm": 0.6717640161514282,
"learning_rate": 1.3464765100671143e-05,
"loss": 0.7792,
"step": 971
},
{
"epoch": 1.5032856590645536,
"grad_norm": 0.6208140850067139,
"learning_rate": 1.3422818791946309e-05,
"loss": 0.7973,
"step": 972
},
{
"epoch": 1.50483185156552,
"grad_norm": 0.6302463412284851,
"learning_rate": 1.3380872483221477e-05,
"loss": 0.785,
"step": 973
},
{
"epoch": 1.5063780440664862,
"grad_norm": 0.6479527950286865,
"learning_rate": 1.3338926174496644e-05,
"loss": 0.8266,
"step": 974
},
{
"epoch": 1.5079242365674528,
"grad_norm": 0.6340357661247253,
"learning_rate": 1.3296979865771814e-05,
"loss": 0.7134,
"step": 975
},
{
"epoch": 1.5094704290684189,
"grad_norm": 0.6478318572044373,
"learning_rate": 1.325503355704698e-05,
"loss": 0.7638,
"step": 976
},
{
"epoch": 1.5110166215693854,
"grad_norm": 0.6769170761108398,
"learning_rate": 1.3213087248322149e-05,
"loss": 0.7754,
"step": 977
},
{
"epoch": 1.5125628140703518,
"grad_norm": 0.6835642457008362,
"learning_rate": 1.3171140939597315e-05,
"loss": 0.7856,
"step": 978
},
{
"epoch": 1.514109006571318,
"grad_norm": 0.700343668460846,
"learning_rate": 1.3129194630872485e-05,
"loss": 0.7428,
"step": 979
},
{
"epoch": 1.5156551990722846,
"grad_norm": 0.6960814595222473,
"learning_rate": 1.3087248322147652e-05,
"loss": 0.7649,
"step": 980
},
{
"epoch": 1.5172013915732507,
"grad_norm": 0.6891623735427856,
"learning_rate": 1.304530201342282e-05,
"loss": 0.8444,
"step": 981
},
{
"epoch": 1.5187475840742173,
"grad_norm": 0.6881600618362427,
"learning_rate": 1.3003355704697987e-05,
"loss": 0.7956,
"step": 982
},
{
"epoch": 1.5202937765751836,
"grad_norm": 0.6770979762077332,
"learning_rate": 1.2961409395973153e-05,
"loss": 0.8173,
"step": 983
},
{
"epoch": 1.52183996907615,
"grad_norm": 0.6920093894004822,
"learning_rate": 1.2919463087248323e-05,
"loss": 0.8522,
"step": 984
},
{
"epoch": 1.5233861615771165,
"grad_norm": 0.6803298592567444,
"learning_rate": 1.2877516778523491e-05,
"loss": 0.8345,
"step": 985
},
{
"epoch": 1.5249323540780826,
"grad_norm": 0.6997124552726746,
"learning_rate": 1.2835570469798658e-05,
"loss": 0.8141,
"step": 986
},
{
"epoch": 1.5264785465790491,
"grad_norm": 0.7112342715263367,
"learning_rate": 1.2793624161073825e-05,
"loss": 0.8196,
"step": 987
},
{
"epoch": 1.5280247390800155,
"grad_norm": 0.7443994879722595,
"learning_rate": 1.2751677852348994e-05,
"loss": 0.8575,
"step": 988
},
{
"epoch": 1.5295709315809818,
"grad_norm": 0.6785229444503784,
"learning_rate": 1.2709731543624163e-05,
"loss": 0.8843,
"step": 989
},
{
"epoch": 1.5311171240819483,
"grad_norm": 0.7394726276397705,
"learning_rate": 1.266778523489933e-05,
"loss": 0.9496,
"step": 990
},
{
"epoch": 1.5326633165829144,
"grad_norm": 0.738186776638031,
"learning_rate": 1.2625838926174496e-05,
"loss": 0.9172,
"step": 991
},
{
"epoch": 1.534209509083881,
"grad_norm": 0.7867004871368408,
"learning_rate": 1.2583892617449666e-05,
"loss": 0.8743,
"step": 992
},
{
"epoch": 1.5357557015848473,
"grad_norm": 0.8075705170631409,
"learning_rate": 1.2541946308724834e-05,
"loss": 0.9021,
"step": 993
},
{
"epoch": 1.5373018940858136,
"grad_norm": 0.8144130110740662,
"learning_rate": 1.25e-05,
"loss": 0.8258,
"step": 994
},
{
"epoch": 1.5388480865867802,
"grad_norm": 0.8460942506790161,
"learning_rate": 1.2458053691275167e-05,
"loss": 0.8051,
"step": 995
},
{
"epoch": 1.5403942790877463,
"grad_norm": 1.0733917951583862,
"learning_rate": 1.2416107382550337e-05,
"loss": 0.9469,
"step": 996
},
{
"epoch": 1.5419404715887128,
"grad_norm": 0.6984366178512573,
"learning_rate": 1.2374161073825504e-05,
"loss": 0.6271,
"step": 997
},
{
"epoch": 1.5434866640896792,
"grad_norm": 0.5974945425987244,
"learning_rate": 1.2332214765100672e-05,
"loss": 0.6767,
"step": 998
},
{
"epoch": 1.5450328565906455,
"grad_norm": 0.5917332172393799,
"learning_rate": 1.2290268456375838e-05,
"loss": 0.7027,
"step": 999
},
{
"epoch": 1.546579049091612,
"grad_norm": 0.5879162549972534,
"learning_rate": 1.2248322147651008e-05,
"loss": 0.6444,
"step": 1000
},
{
"epoch": 1.5481252415925781,
"grad_norm": 0.5611408948898315,
"learning_rate": 1.2206375838926175e-05,
"loss": 0.6743,
"step": 1001
},
{
"epoch": 1.5496714340935447,
"grad_norm": 0.6579827666282654,
"learning_rate": 1.2164429530201343e-05,
"loss": 0.6863,
"step": 1002
},
{
"epoch": 1.551217626594511,
"grad_norm": 0.6543861031532288,
"learning_rate": 1.212248322147651e-05,
"loss": 0.6911,
"step": 1003
},
{
"epoch": 1.5527638190954773,
"grad_norm": 0.6316918134689331,
"learning_rate": 1.208053691275168e-05,
"loss": 0.7126,
"step": 1004
},
{
"epoch": 1.5543100115964439,
"grad_norm": 0.6070011258125305,
"learning_rate": 1.2038590604026846e-05,
"loss": 0.5802,
"step": 1005
},
{
"epoch": 1.55585620409741,
"grad_norm": 0.6225163340568542,
"learning_rate": 1.1996644295302013e-05,
"loss": 0.7565,
"step": 1006
},
{
"epoch": 1.5574023965983765,
"grad_norm": 0.6303391456604004,
"learning_rate": 1.1954697986577181e-05,
"loss": 0.7537,
"step": 1007
},
{
"epoch": 1.5589485890993429,
"grad_norm": 0.6572604179382324,
"learning_rate": 1.191275167785235e-05,
"loss": 0.7332,
"step": 1008
},
{
"epoch": 1.5604947816003092,
"grad_norm": 0.6323292255401611,
"learning_rate": 1.1870805369127518e-05,
"loss": 0.7537,
"step": 1009
},
{
"epoch": 1.5620409741012757,
"grad_norm": 0.621051549911499,
"learning_rate": 1.1828859060402684e-05,
"loss": 0.7371,
"step": 1010
},
{
"epoch": 1.5635871666022418,
"grad_norm": 0.6632164120674133,
"learning_rate": 1.1786912751677852e-05,
"loss": 0.7169,
"step": 1011
},
{
"epoch": 1.5651333591032084,
"grad_norm": 0.6632513999938965,
"learning_rate": 1.174496644295302e-05,
"loss": 0.726,
"step": 1012
},
{
"epoch": 1.5666795516041747,
"grad_norm": 0.6338618397712708,
"learning_rate": 1.1703020134228189e-05,
"loss": 0.7774,
"step": 1013
},
{
"epoch": 1.568225744105141,
"grad_norm": 0.6426889896392822,
"learning_rate": 1.1661073825503356e-05,
"loss": 0.712,
"step": 1014
},
{
"epoch": 1.5697719366061076,
"grad_norm": 0.6418123841285706,
"learning_rate": 1.1619127516778524e-05,
"loss": 0.692,
"step": 1015
},
{
"epoch": 1.5713181291070737,
"grad_norm": 0.6338704228401184,
"learning_rate": 1.1577181208053692e-05,
"loss": 0.7498,
"step": 1016
},
{
"epoch": 1.5728643216080402,
"grad_norm": 0.6129948496818542,
"learning_rate": 1.1535234899328859e-05,
"loss": 0.8721,
"step": 1017
},
{
"epoch": 1.5744105141090066,
"grad_norm": 0.6413542032241821,
"learning_rate": 1.1493288590604027e-05,
"loss": 0.8381,
"step": 1018
},
{
"epoch": 1.575956706609973,
"grad_norm": 0.6431681513786316,
"learning_rate": 1.1451342281879195e-05,
"loss": 0.7365,
"step": 1019
},
{
"epoch": 1.5775028991109394,
"grad_norm": 0.6414735317230225,
"learning_rate": 1.1409395973154363e-05,
"loss": 0.8006,
"step": 1020
},
{
"epoch": 1.5790490916119055,
"grad_norm": 0.6431716680526733,
"learning_rate": 1.136744966442953e-05,
"loss": 0.7759,
"step": 1021
},
{
"epoch": 1.580595284112872,
"grad_norm": 0.6646420359611511,
"learning_rate": 1.1325503355704698e-05,
"loss": 0.7038,
"step": 1022
},
{
"epoch": 1.5821414766138384,
"grad_norm": 0.7008923292160034,
"learning_rate": 1.1283557046979866e-05,
"loss": 0.719,
"step": 1023
},
{
"epoch": 1.5836876691148047,
"grad_norm": 0.6393096446990967,
"learning_rate": 1.1241610738255035e-05,
"loss": 0.7841,
"step": 1024
},
{
"epoch": 1.5852338616157713,
"grad_norm": 0.6879417300224304,
"learning_rate": 1.1199664429530201e-05,
"loss": 0.8218,
"step": 1025
},
{
"epoch": 1.5867800541167374,
"grad_norm": 0.6500439047813416,
"learning_rate": 1.115771812080537e-05,
"loss": 0.8026,
"step": 1026
},
{
"epoch": 1.588326246617704,
"grad_norm": 0.6967812180519104,
"learning_rate": 1.1115771812080538e-05,
"loss": 0.7494,
"step": 1027
},
{
"epoch": 1.5898724391186703,
"grad_norm": 0.6527352929115295,
"learning_rate": 1.1073825503355706e-05,
"loss": 0.8472,
"step": 1028
},
{
"epoch": 1.5914186316196366,
"grad_norm": 0.7033571004867554,
"learning_rate": 1.1031879194630873e-05,
"loss": 0.7553,
"step": 1029
},
{
"epoch": 1.5929648241206031,
"grad_norm": 0.6790737509727478,
"learning_rate": 1.098993288590604e-05,
"loss": 0.8045,
"step": 1030
},
{
"epoch": 1.5945110166215692,
"grad_norm": 0.7032893300056458,
"learning_rate": 1.0947986577181209e-05,
"loss": 0.8292,
"step": 1031
},
{
"epoch": 1.5960572091225358,
"grad_norm": 0.6649371981620789,
"learning_rate": 1.0906040268456376e-05,
"loss": 0.8032,
"step": 1032
},
{
"epoch": 1.5976034016235021,
"grad_norm": 0.7009012699127197,
"learning_rate": 1.0864093959731544e-05,
"loss": 0.9107,
"step": 1033
},
{
"epoch": 1.5991495941244684,
"grad_norm": 0.7047061324119568,
"learning_rate": 1.0822147651006712e-05,
"loss": 0.904,
"step": 1034
},
{
"epoch": 1.600695786625435,
"grad_norm": 0.7520581483840942,
"learning_rate": 1.078020134228188e-05,
"loss": 0.8972,
"step": 1035
},
{
"epoch": 1.602241979126401,
"grad_norm": 0.710966944694519,
"learning_rate": 1.0738255033557047e-05,
"loss": 0.8133,
"step": 1036
},
{
"epoch": 1.6037881716273676,
"grad_norm": 0.7045226693153381,
"learning_rate": 1.0696308724832215e-05,
"loss": 0.8474,
"step": 1037
},
{
"epoch": 1.605334364128334,
"grad_norm": 0.7222912311553955,
"learning_rate": 1.0654362416107383e-05,
"loss": 0.8577,
"step": 1038
},
{
"epoch": 1.6068805566293003,
"grad_norm": 0.7768924236297607,
"learning_rate": 1.0612416107382552e-05,
"loss": 0.8906,
"step": 1039
},
{
"epoch": 1.6084267491302668,
"grad_norm": 0.7447935342788696,
"learning_rate": 1.0570469798657718e-05,
"loss": 0.8571,
"step": 1040
},
{
"epoch": 1.609972941631233,
"grad_norm": 0.7860161662101746,
"learning_rate": 1.0528523489932887e-05,
"loss": 0.868,
"step": 1041
},
{
"epoch": 1.6115191341321995,
"grad_norm": 0.772621750831604,
"learning_rate": 1.0486577181208055e-05,
"loss": 0.9064,
"step": 1042
},
{
"epoch": 1.6130653266331658,
"grad_norm": 0.7927963733673096,
"learning_rate": 1.0444630872483221e-05,
"loss": 0.8628,
"step": 1043
},
{
"epoch": 1.6146115191341321,
"grad_norm": 0.8811314105987549,
"learning_rate": 1.040268456375839e-05,
"loss": 0.9203,
"step": 1044
},
{
"epoch": 1.6161577116350987,
"grad_norm": 1.0155308246612549,
"learning_rate": 1.0360738255033558e-05,
"loss": 0.9134,
"step": 1045
},
{
"epoch": 1.6177039041360648,
"grad_norm": 1.3558322191238403,
"learning_rate": 1.0318791946308726e-05,
"loss": 1.007,
"step": 1046
},
{
"epoch": 1.6192500966370313,
"grad_norm": 0.771702766418457,
"learning_rate": 1.0276845637583893e-05,
"loss": 0.5492,
"step": 1047
},
{
"epoch": 1.6207962891379977,
"grad_norm": 0.5778831243515015,
"learning_rate": 1.0234899328859061e-05,
"loss": 0.5969,
"step": 1048
},
{
"epoch": 1.622342481638964,
"grad_norm": 0.5768334865570068,
"learning_rate": 1.019295302013423e-05,
"loss": 0.7561,
"step": 1049
},
{
"epoch": 1.6238886741399305,
"grad_norm": 0.6801527142524719,
"learning_rate": 1.0151006711409397e-05,
"loss": 0.6649,
"step": 1050
},
{
"epoch": 1.6254348666408966,
"grad_norm": 0.6493854522705078,
"learning_rate": 1.0109060402684564e-05,
"loss": 0.6788,
"step": 1051
},
{
"epoch": 1.6269810591418632,
"grad_norm": 0.6070857644081116,
"learning_rate": 1.006711409395973e-05,
"loss": 0.6509,
"step": 1052
},
{
"epoch": 1.6285272516428295,
"grad_norm": 0.6425489783287048,
"learning_rate": 1.00251677852349e-05,
"loss": 0.7303,
"step": 1053
},
{
"epoch": 1.6300734441437958,
"grad_norm": 0.6311664581298828,
"learning_rate": 9.983221476510067e-06,
"loss": 0.6492,
"step": 1054
},
{
"epoch": 1.6316196366447624,
"grad_norm": 0.5983323454856873,
"learning_rate": 9.941275167785235e-06,
"loss": 0.7778,
"step": 1055
},
{
"epoch": 1.6331658291457285,
"grad_norm": 0.5775598287582397,
"learning_rate": 9.899328859060402e-06,
"loss": 0.7594,
"step": 1056
},
{
"epoch": 1.634712021646695,
"grad_norm": 0.6420764327049255,
"learning_rate": 9.857382550335572e-06,
"loss": 0.7812,
"step": 1057
},
{
"epoch": 1.6362582141476614,
"grad_norm": 0.6438013911247253,
"learning_rate": 9.815436241610738e-06,
"loss": 0.7245,
"step": 1058
},
{
"epoch": 1.6378044066486277,
"grad_norm": 0.6310963034629822,
"learning_rate": 9.773489932885907e-06,
"loss": 0.7321,
"step": 1059
},
{
"epoch": 1.6393505991495942,
"grad_norm": 0.6203198432922363,
"learning_rate": 9.731543624161075e-06,
"loss": 0.6914,
"step": 1060
},
{
"epoch": 1.6408967916505603,
"grad_norm": 0.638454258441925,
"learning_rate": 9.689597315436243e-06,
"loss": 0.7478,
"step": 1061
},
{
"epoch": 1.642442984151527,
"grad_norm": 0.6612226963043213,
"learning_rate": 9.64765100671141e-06,
"loss": 0.7209,
"step": 1062
},
{
"epoch": 1.6439891766524932,
"grad_norm": 0.6566651463508606,
"learning_rate": 9.605704697986576e-06,
"loss": 0.8045,
"step": 1063
},
{
"epoch": 1.6455353691534595,
"grad_norm": 0.6543892621994019,
"learning_rate": 9.563758389261746e-06,
"loss": 0.7727,
"step": 1064
},
{
"epoch": 1.647081561654426,
"grad_norm": 0.6707619428634644,
"learning_rate": 9.521812080536913e-06,
"loss": 0.7271,
"step": 1065
},
{
"epoch": 1.6486277541553922,
"grad_norm": 0.6582595705986023,
"learning_rate": 9.479865771812081e-06,
"loss": 0.7332,
"step": 1066
},
{
"epoch": 1.6501739466563587,
"grad_norm": 0.6757826209068298,
"learning_rate": 9.437919463087248e-06,
"loss": 0.7663,
"step": 1067
},
{
"epoch": 1.651720139157325,
"grad_norm": 0.698391318321228,
"learning_rate": 9.395973154362418e-06,
"loss": 0.7277,
"step": 1068
},
{
"epoch": 1.6532663316582914,
"grad_norm": 0.6885313987731934,
"learning_rate": 9.354026845637584e-06,
"loss": 0.7959,
"step": 1069
},
{
"epoch": 1.654812524159258,
"grad_norm": 0.664113461971283,
"learning_rate": 9.312080536912752e-06,
"loss": 0.7571,
"step": 1070
},
{
"epoch": 1.656358716660224,
"grad_norm": 0.6730982065200806,
"learning_rate": 9.270134228187919e-06,
"loss": 0.709,
"step": 1071
},
{
"epoch": 1.6579049091611906,
"grad_norm": 0.684267520904541,
"learning_rate": 9.228187919463089e-06,
"loss": 0.6932,
"step": 1072
},
{
"epoch": 1.659451101662157,
"grad_norm": 0.631272554397583,
"learning_rate": 9.186241610738255e-06,
"loss": 0.8569,
"step": 1073
},
{
"epoch": 1.6609972941631233,
"grad_norm": 0.7173067927360535,
"learning_rate": 9.144295302013424e-06,
"loss": 0.777,
"step": 1074
},
{
"epoch": 1.6625434866640898,
"grad_norm": 0.6663726568222046,
"learning_rate": 9.10234899328859e-06,
"loss": 0.7523,
"step": 1075
},
{
"epoch": 1.664089679165056,
"grad_norm": 0.689304530620575,
"learning_rate": 9.060402684563759e-06,
"loss": 0.7922,
"step": 1076
},
{
"epoch": 1.6656358716660224,
"grad_norm": 0.6620500683784485,
"learning_rate": 9.018456375838927e-06,
"loss": 0.8587,
"step": 1077
},
{
"epoch": 1.6671820641669888,
"grad_norm": 0.6905922889709473,
"learning_rate": 8.976510067114093e-06,
"loss": 0.7897,
"step": 1078
},
{
"epoch": 1.668728256667955,
"grad_norm": 0.6763781309127808,
"learning_rate": 8.934563758389262e-06,
"loss": 0.7886,
"step": 1079
},
{
"epoch": 1.6702744491689216,
"grad_norm": 0.684766948223114,
"learning_rate": 8.89261744966443e-06,
"loss": 0.8122,
"step": 1080
},
{
"epoch": 1.6718206416698878,
"grad_norm": 0.6934885382652283,
"learning_rate": 8.850671140939598e-06,
"loss": 0.7894,
"step": 1081
},
{
"epoch": 1.6733668341708543,
"grad_norm": 0.7051049470901489,
"learning_rate": 8.808724832214765e-06,
"loss": 0.8047,
"step": 1082
},
{
"epoch": 1.6749130266718206,
"grad_norm": 0.7332410216331482,
"learning_rate": 8.766778523489935e-06,
"loss": 0.7371,
"step": 1083
},
{
"epoch": 1.676459219172787,
"grad_norm": 0.7231751084327698,
"learning_rate": 8.724832214765101e-06,
"loss": 0.7804,
"step": 1084
},
{
"epoch": 1.6780054116737535,
"grad_norm": 0.7421995997428894,
"learning_rate": 8.68288590604027e-06,
"loss": 0.8364,
"step": 1085
},
{
"epoch": 1.6795516041747196,
"grad_norm": 0.7182170152664185,
"learning_rate": 8.640939597315436e-06,
"loss": 0.848,
"step": 1086
},
{
"epoch": 1.6810977966756862,
"grad_norm": 0.7546189427375793,
"learning_rate": 8.598993288590604e-06,
"loss": 0.8336,
"step": 1087
},
{
"epoch": 1.6826439891766525,
"grad_norm": 0.7409399747848511,
"learning_rate": 8.557046979865773e-06,
"loss": 0.8151,
"step": 1088
},
{
"epoch": 1.6841901816776188,
"grad_norm": 0.7315151691436768,
"learning_rate": 8.515100671140939e-06,
"loss": 0.8126,
"step": 1089
},
{
"epoch": 1.6857363741785854,
"grad_norm": 0.7949945330619812,
"learning_rate": 8.473154362416107e-06,
"loss": 0.8533,
"step": 1090
},
{
"epoch": 1.6872825666795515,
"grad_norm": 0.7690507769584656,
"learning_rate": 8.431208053691276e-06,
"loss": 0.846,
"step": 1091
},
{
"epoch": 1.688828759180518,
"grad_norm": 0.7926616072654724,
"learning_rate": 8.389261744966444e-06,
"loss": 0.8402,
"step": 1092
},
{
"epoch": 1.6903749516814843,
"grad_norm": 0.8056994676589966,
"learning_rate": 8.34731543624161e-06,
"loss": 0.8191,
"step": 1093
},
{
"epoch": 1.6919211441824507,
"grad_norm": 0.8308284282684326,
"learning_rate": 8.305369127516779e-06,
"loss": 0.9042,
"step": 1094
},
{
"epoch": 1.6934673366834172,
"grad_norm": 0.9019404053688049,
"learning_rate": 8.263422818791947e-06,
"loss": 0.8962,
"step": 1095
},
{
"epoch": 1.6950135291843833,
"grad_norm": 1.144425630569458,
"learning_rate": 8.221476510067115e-06,
"loss": 1.0569,
"step": 1096
},
{
"epoch": 1.6965597216853499,
"grad_norm": 0.7823913097381592,
"learning_rate": 8.179530201342282e-06,
"loss": 0.5956,
"step": 1097
},
{
"epoch": 1.6981059141863162,
"grad_norm": 0.5778486132621765,
"learning_rate": 8.13758389261745e-06,
"loss": 0.6246,
"step": 1098
},
{
"epoch": 1.6996521066872825,
"grad_norm": 0.6111435890197754,
"learning_rate": 8.095637583892618e-06,
"loss": 0.6681,
"step": 1099
},
{
"epoch": 1.701198299188249,
"grad_norm": 0.6174157857894897,
"learning_rate": 8.053691275167785e-06,
"loss": 0.6635,
"step": 1100
},
{
"epoch": 1.7027444916892152,
"grad_norm": 0.660089910030365,
"learning_rate": 8.011744966442953e-06,
"loss": 0.6994,
"step": 1101
},
{
"epoch": 1.7042906841901817,
"grad_norm": 0.6478524208068848,
"learning_rate": 7.969798657718121e-06,
"loss": 0.6157,
"step": 1102
},
{
"epoch": 1.705836876691148,
"grad_norm": 0.6385321617126465,
"learning_rate": 7.92785234899329e-06,
"loss": 0.6967,
"step": 1103
},
{
"epoch": 1.7073830691921144,
"grad_norm": 0.6223974227905273,
"learning_rate": 7.885906040268456e-06,
"loss": 0.7572,
"step": 1104
},
{
"epoch": 1.708929261693081,
"grad_norm": 0.6408361792564392,
"learning_rate": 7.843959731543624e-06,
"loss": 0.8266,
"step": 1105
},
{
"epoch": 1.710475454194047,
"grad_norm": 0.6401566863059998,
"learning_rate": 7.802013422818793e-06,
"loss": 0.7073,
"step": 1106
},
{
"epoch": 1.7120216466950136,
"grad_norm": 0.6204596757888794,
"learning_rate": 7.760067114093961e-06,
"loss": 0.6831,
"step": 1107
},
{
"epoch": 1.7135678391959799,
"grad_norm": 0.6418159604072571,
"learning_rate": 7.718120805369127e-06,
"loss": 0.6925,
"step": 1108
},
{
"epoch": 1.7151140316969462,
"grad_norm": 0.6294611692428589,
"learning_rate": 7.676174496644296e-06,
"loss": 0.7594,
"step": 1109
},
{
"epoch": 1.7166602241979128,
"grad_norm": 0.6831735968589783,
"learning_rate": 7.634228187919464e-06,
"loss": 0.6813,
"step": 1110
},
{
"epoch": 1.7182064166988789,
"grad_norm": 0.6496366262435913,
"learning_rate": 7.5922818791946305e-06,
"loss": 0.7343,
"step": 1111
},
{
"epoch": 1.7197526091998454,
"grad_norm": 0.6550260782241821,
"learning_rate": 7.5503355704698e-06,
"loss": 0.7389,
"step": 1112
},
{
"epoch": 1.7212988017008117,
"grad_norm": 0.6319887638092041,
"learning_rate": 7.508389261744966e-06,
"loss": 0.7011,
"step": 1113
},
{
"epoch": 1.722844994201778,
"grad_norm": 0.6578625440597534,
"learning_rate": 7.466442953020135e-06,
"loss": 0.7221,
"step": 1114
},
{
"epoch": 1.7243911867027446,
"grad_norm": 0.6553155779838562,
"learning_rate": 7.424496644295302e-06,
"loss": 0.7053,
"step": 1115
},
{
"epoch": 1.7259373792037107,
"grad_norm": 0.6502689719200134,
"learning_rate": 7.382550335570471e-06,
"loss": 0.8073,
"step": 1116
},
{
"epoch": 1.7274835717046773,
"grad_norm": 0.6422849893569946,
"learning_rate": 7.3406040268456375e-06,
"loss": 0.6863,
"step": 1117
},
{
"epoch": 1.7290297642056436,
"grad_norm": 0.6866220831871033,
"learning_rate": 7.298657718120807e-06,
"loss": 0.7959,
"step": 1118
},
{
"epoch": 1.73057595670661,
"grad_norm": 0.645169198513031,
"learning_rate": 7.256711409395973e-06,
"loss": 0.7729,
"step": 1119
},
{
"epoch": 1.7321221492075765,
"grad_norm": 0.654244601726532,
"learning_rate": 7.214765100671142e-06,
"loss": 0.7707,
"step": 1120
},
{
"epoch": 1.7336683417085426,
"grad_norm": 0.6695640087127686,
"learning_rate": 7.172818791946309e-06,
"loss": 0.714,
"step": 1121
},
{
"epoch": 1.735214534209509,
"grad_norm": 0.6387913227081299,
"learning_rate": 7.130872483221476e-06,
"loss": 0.7654,
"step": 1122
},
{
"epoch": 1.7367607267104754,
"grad_norm": 0.664368748664856,
"learning_rate": 7.0889261744966445e-06,
"loss": 0.7525,
"step": 1123
},
{
"epoch": 1.7383069192114418,
"grad_norm": 0.7358962893486023,
"learning_rate": 7.046979865771812e-06,
"loss": 0.747,
"step": 1124
},
{
"epoch": 1.7398531117124083,
"grad_norm": 0.6654751896858215,
"learning_rate": 7.00503355704698e-06,
"loss": 0.7669,
"step": 1125
},
{
"epoch": 1.7413993042133744,
"grad_norm": 0.6763463020324707,
"learning_rate": 6.963087248322148e-06,
"loss": 0.8077,
"step": 1126
},
{
"epoch": 1.742945496714341,
"grad_norm": 0.6764401197433472,
"learning_rate": 6.921140939597316e-06,
"loss": 0.7438,
"step": 1127
},
{
"epoch": 1.7444916892153073,
"grad_norm": 0.7182475924491882,
"learning_rate": 6.879194630872483e-06,
"loss": 0.728,
"step": 1128
},
{
"epoch": 1.7460378817162736,
"grad_norm": 0.6785516142845154,
"learning_rate": 6.8372483221476515e-06,
"loss": 0.7939,
"step": 1129
},
{
"epoch": 1.7475840742172402,
"grad_norm": 0.7003461122512817,
"learning_rate": 6.795302013422819e-06,
"loss": 0.7729,
"step": 1130
},
{
"epoch": 1.7491302667182063,
"grad_norm": 0.6882821917533875,
"learning_rate": 6.753355704697987e-06,
"loss": 0.9204,
"step": 1131
},
{
"epoch": 1.7506764592191728,
"grad_norm": 0.7342216372489929,
"learning_rate": 6.7114093959731546e-06,
"loss": 0.7385,
"step": 1132
},
{
"epoch": 1.7522226517201391,
"grad_norm": 0.7073139548301697,
"learning_rate": 6.669463087248322e-06,
"loss": 0.9268,
"step": 1133
},
{
"epoch": 1.7537688442211055,
"grad_norm": 0.7231860756874084,
"learning_rate": 6.62751677852349e-06,
"loss": 0.8596,
"step": 1134
},
{
"epoch": 1.755315036722072,
"grad_norm": 0.7145645618438721,
"learning_rate": 6.585570469798658e-06,
"loss": 0.8466,
"step": 1135
},
{
"epoch": 1.7568612292230381,
"grad_norm": 0.7454160451889038,
"learning_rate": 6.543624161073826e-06,
"loss": 0.7629,
"step": 1136
},
{
"epoch": 1.7584074217240047,
"grad_norm": 0.7323503494262695,
"learning_rate": 6.501677852348993e-06,
"loss": 0.8666,
"step": 1137
},
{
"epoch": 1.759953614224971,
"grad_norm": 0.7377263307571411,
"learning_rate": 6.4597315436241616e-06,
"loss": 0.8515,
"step": 1138
},
{
"epoch": 1.7614998067259373,
"grad_norm": 0.7750667929649353,
"learning_rate": 6.417785234899329e-06,
"loss": 0.8033,
"step": 1139
},
{
"epoch": 1.7630459992269039,
"grad_norm": 0.7829360365867615,
"learning_rate": 6.375838926174497e-06,
"loss": 0.8903,
"step": 1140
},
{
"epoch": 1.76459219172787,
"grad_norm": 0.7346689701080322,
"learning_rate": 6.333892617449665e-06,
"loss": 0.9772,
"step": 1141
},
{
"epoch": 1.7661383842288365,
"grad_norm": 0.7309790253639221,
"learning_rate": 6.291946308724833e-06,
"loss": 0.8317,
"step": 1142
},
{
"epoch": 1.7676845767298028,
"grad_norm": 0.8332825899124146,
"learning_rate": 6.25e-06,
"loss": 0.7692,
"step": 1143
},
{
"epoch": 1.7692307692307692,
"grad_norm": 0.8708672523498535,
"learning_rate": 6.2080536912751686e-06,
"loss": 0.946,
"step": 1144
},
{
"epoch": 1.7707769617317357,
"grad_norm": 0.9887493848800659,
"learning_rate": 6.166107382550336e-06,
"loss": 0.9102,
"step": 1145
},
{
"epoch": 1.7723231542327018,
"grad_norm": 1.3964698314666748,
"learning_rate": 6.124161073825504e-06,
"loss": 0.8942,
"step": 1146
},
{
"epoch": 1.7738693467336684,
"grad_norm": 0.9236196875572205,
"learning_rate": 6.082214765100672e-06,
"loss": 0.6239,
"step": 1147
},
{
"epoch": 1.7754155392346347,
"grad_norm": 0.5753183960914612,
"learning_rate": 6.04026845637584e-06,
"loss": 0.6674,
"step": 1148
},
{
"epoch": 1.776961731735601,
"grad_norm": 0.6132087111473083,
"learning_rate": 5.9983221476510064e-06,
"loss": 0.7186,
"step": 1149
},
{
"epoch": 1.7785079242365676,
"grad_norm": 0.586824893951416,
"learning_rate": 5.956375838926175e-06,
"loss": 0.7077,
"step": 1150
},
{
"epoch": 1.7800541167375337,
"grad_norm": 0.6105204224586487,
"learning_rate": 5.914429530201342e-06,
"loss": 0.7448,
"step": 1151
},
{
"epoch": 1.7816003092385002,
"grad_norm": 0.6133350133895874,
"learning_rate": 5.87248322147651e-06,
"loss": 0.7761,
"step": 1152
},
{
"epoch": 1.7831465017394665,
"grad_norm": 0.6162737011909485,
"learning_rate": 5.830536912751678e-06,
"loss": 0.7204,
"step": 1153
},
{
"epoch": 1.7846926942404329,
"grad_norm": 0.6450937390327454,
"learning_rate": 5.788590604026846e-06,
"loss": 0.7163,
"step": 1154
},
{
"epoch": 1.7862388867413994,
"grad_norm": 0.6387403011322021,
"learning_rate": 5.7466442953020134e-06,
"loss": 0.7454,
"step": 1155
},
{
"epoch": 1.7877850792423655,
"grad_norm": 0.6444848775863647,
"learning_rate": 5.704697986577182e-06,
"loss": 0.7048,
"step": 1156
},
{
"epoch": 1.789331271743332,
"grad_norm": 0.6471322774887085,
"learning_rate": 5.662751677852349e-06,
"loss": 0.75,
"step": 1157
},
{
"epoch": 1.7908774642442984,
"grad_norm": 0.6798081398010254,
"learning_rate": 5.620805369127517e-06,
"loss": 0.7096,
"step": 1158
},
{
"epoch": 1.7924236567452647,
"grad_norm": 0.6646496653556824,
"learning_rate": 5.578859060402685e-06,
"loss": 0.8017,
"step": 1159
},
{
"epoch": 1.7939698492462313,
"grad_norm": 0.6455252170562744,
"learning_rate": 5.536912751677853e-06,
"loss": 0.7323,
"step": 1160
},
{
"epoch": 1.7955160417471974,
"grad_norm": 0.6841678023338318,
"learning_rate": 5.49496644295302e-06,
"loss": 0.7373,
"step": 1161
},
{
"epoch": 1.797062234248164,
"grad_norm": 0.6608906388282776,
"learning_rate": 5.453020134228188e-06,
"loss": 0.7039,
"step": 1162
},
{
"epoch": 1.7986084267491302,
"grad_norm": 0.6611918807029724,
"learning_rate": 5.411073825503356e-06,
"loss": 0.768,
"step": 1163
},
{
"epoch": 1.8001546192500966,
"grad_norm": 0.6479055881500244,
"learning_rate": 5.3691275167785235e-06,
"loss": 0.7672,
"step": 1164
},
{
"epoch": 1.8017008117510631,
"grad_norm": 0.6645311117172241,
"learning_rate": 5.327181208053692e-06,
"loss": 0.8146,
"step": 1165
},
{
"epoch": 1.8032470042520292,
"grad_norm": 0.6700178980827332,
"learning_rate": 5.285234899328859e-06,
"loss": 0.7019,
"step": 1166
},
{
"epoch": 1.8047931967529958,
"grad_norm": 0.6892951130867004,
"learning_rate": 5.243288590604027e-06,
"loss": 0.7982,
"step": 1167
},
{
"epoch": 1.806339389253962,
"grad_norm": 0.6578270196914673,
"learning_rate": 5.201342281879195e-06,
"loss": 0.7844,
"step": 1168
},
{
"epoch": 1.8078855817549284,
"grad_norm": 0.6844699382781982,
"learning_rate": 5.159395973154363e-06,
"loss": 0.73,
"step": 1169
},
{
"epoch": 1.809431774255895,
"grad_norm": 0.67879718542099,
"learning_rate": 5.1174496644295305e-06,
"loss": 0.8256,
"step": 1170
},
{
"epoch": 1.810977966756861,
"grad_norm": 0.6748877167701721,
"learning_rate": 5.075503355704699e-06,
"loss": 0.7733,
"step": 1171
},
{
"epoch": 1.8125241592578276,
"grad_norm": 0.6248130798339844,
"learning_rate": 5.033557046979865e-06,
"loss": 0.7489,
"step": 1172
},
{
"epoch": 1.814070351758794,
"grad_norm": 0.6643280386924744,
"learning_rate": 4.9916107382550336e-06,
"loss": 0.726,
"step": 1173
},
{
"epoch": 1.8156165442597603,
"grad_norm": 0.6891630291938782,
"learning_rate": 4.949664429530201e-06,
"loss": 0.7442,
"step": 1174
},
{
"epoch": 1.8171627367607268,
"grad_norm": 0.6768823862075806,
"learning_rate": 4.907718120805369e-06,
"loss": 0.7376,
"step": 1175
},
{
"epoch": 1.818708929261693,
"grad_norm": 0.669154167175293,
"learning_rate": 4.8657718120805375e-06,
"loss": 0.7775,
"step": 1176
},
{
"epoch": 1.8202551217626595,
"grad_norm": 0.7224996089935303,
"learning_rate": 4.823825503355705e-06,
"loss": 0.7535,
"step": 1177
},
{
"epoch": 1.8218013142636258,
"grad_norm": 0.6954566240310669,
"learning_rate": 4.781879194630873e-06,
"loss": 0.77,
"step": 1178
},
{
"epoch": 1.8233475067645921,
"grad_norm": 0.6938985586166382,
"learning_rate": 4.7399328859060405e-06,
"loss": 0.7963,
"step": 1179
},
{
"epoch": 1.8248936992655587,
"grad_norm": 0.6518607139587402,
"learning_rate": 4.697986577181209e-06,
"loss": 0.7652,
"step": 1180
},
{
"epoch": 1.8264398917665248,
"grad_norm": 0.7072224617004395,
"learning_rate": 4.656040268456376e-06,
"loss": 0.7875,
"step": 1181
},
{
"epoch": 1.8279860842674913,
"grad_norm": 0.689007580280304,
"learning_rate": 4.6140939597315445e-06,
"loss": 0.8659,
"step": 1182
},
{
"epoch": 1.8295322767684576,
"grad_norm": 0.7117170095443726,
"learning_rate": 4.572147651006712e-06,
"loss": 0.7761,
"step": 1183
},
{
"epoch": 1.831078469269424,
"grad_norm": 0.6940242052078247,
"learning_rate": 4.530201342281879e-06,
"loss": 0.8305,
"step": 1184
},
{
"epoch": 1.8326246617703905,
"grad_norm": 0.7297004461288452,
"learning_rate": 4.488255033557047e-06,
"loss": 0.8451,
"step": 1185
},
{
"epoch": 1.8341708542713566,
"grad_norm": 0.7414118647575378,
"learning_rate": 4.446308724832215e-06,
"loss": 0.7932,
"step": 1186
},
{
"epoch": 1.8357170467723232,
"grad_norm": 0.7365830540657043,
"learning_rate": 4.404362416107382e-06,
"loss": 0.8153,
"step": 1187
},
{
"epoch": 1.8372632392732895,
"grad_norm": 0.7378683686256409,
"learning_rate": 4.362416107382551e-06,
"loss": 0.8795,
"step": 1188
},
{
"epoch": 1.8388094317742558,
"grad_norm": 0.7360830307006836,
"learning_rate": 4.320469798657718e-06,
"loss": 0.863,
"step": 1189
},
{
"epoch": 1.8403556242752224,
"grad_norm": 0.7365934252738953,
"learning_rate": 4.278523489932886e-06,
"loss": 0.8836,
"step": 1190
},
{
"epoch": 1.8419018167761885,
"grad_norm": 0.7860958576202393,
"learning_rate": 4.236577181208054e-06,
"loss": 0.8834,
"step": 1191
},
{
"epoch": 1.843448009277155,
"grad_norm": 0.8424365520477295,
"learning_rate": 4.194630872483222e-06,
"loss": 0.8303,
"step": 1192
},
{
"epoch": 1.8449942017781213,
"grad_norm": 0.8641267418861389,
"learning_rate": 4.152684563758389e-06,
"loss": 0.7928,
"step": 1193
},
{
"epoch": 1.8465403942790877,
"grad_norm": 0.8995117545127869,
"learning_rate": 4.110738255033558e-06,
"loss": 0.8474,
"step": 1194
},
{
"epoch": 1.8480865867800542,
"grad_norm": 0.9355427026748657,
"learning_rate": 4.068791946308725e-06,
"loss": 0.8516,
"step": 1195
},
{
"epoch": 1.8496327792810203,
"grad_norm": 1.3238755464553833,
"learning_rate": 4.026845637583892e-06,
"loss": 1.0038,
"step": 1196
},
{
"epoch": 1.8511789717819869,
"grad_norm": 0.8634078502655029,
"learning_rate": 3.984899328859061e-06,
"loss": 0.6067,
"step": 1197
},
{
"epoch": 1.8527251642829532,
"grad_norm": 0.5876993536949158,
"learning_rate": 3.942953020134228e-06,
"loss": 0.6623,
"step": 1198
},
{
"epoch": 1.8542713567839195,
"grad_norm": 0.6208466291427612,
"learning_rate": 3.901006711409396e-06,
"loss": 0.6686,
"step": 1199
},
{
"epoch": 1.855817549284886,
"grad_norm": 0.5899839401245117,
"learning_rate": 3.859060402684564e-06,
"loss": 0.7108,
"step": 1200
},
{
"epoch": 1.8573637417858522,
"grad_norm": 0.6139258742332458,
"learning_rate": 3.817114093959732e-06,
"loss": 0.6853,
"step": 1201
},
{
"epoch": 1.8589099342868187,
"grad_norm": 0.6186558604240417,
"learning_rate": 3.7751677852349e-06,
"loss": 0.6619,
"step": 1202
},
{
"epoch": 1.860456126787785,
"grad_norm": 0.6083664298057556,
"learning_rate": 3.7332214765100677e-06,
"loss": 0.7119,
"step": 1203
},
{
"epoch": 1.8620023192887514,
"grad_norm": 0.6227363348007202,
"learning_rate": 3.6912751677852355e-06,
"loss": 0.7251,
"step": 1204
},
{
"epoch": 1.863548511789718,
"grad_norm": 0.6329985857009888,
"learning_rate": 3.6493288590604033e-06,
"loss": 0.7796,
"step": 1205
},
{
"epoch": 1.865094704290684,
"grad_norm": 0.644888162612915,
"learning_rate": 3.607382550335571e-06,
"loss": 0.7747,
"step": 1206
},
{
"epoch": 1.8666408967916506,
"grad_norm": 0.6022448539733887,
"learning_rate": 3.565436241610738e-06,
"loss": 0.7027,
"step": 1207
},
{
"epoch": 1.868187089292617,
"grad_norm": 0.6584210991859436,
"learning_rate": 3.523489932885906e-06,
"loss": 0.6816,
"step": 1208
},
{
"epoch": 1.8697332817935832,
"grad_norm": 0.6227755546569824,
"learning_rate": 3.481543624161074e-06,
"loss": 0.7245,
"step": 1209
},
{
"epoch": 1.8712794742945498,
"grad_norm": 0.6115254759788513,
"learning_rate": 3.4395973154362416e-06,
"loss": 0.7511,
"step": 1210
},
{
"epoch": 1.8728256667955159,
"grad_norm": 0.638425886631012,
"learning_rate": 3.3976510067114095e-06,
"loss": 0.742,
"step": 1211
},
{
"epoch": 1.8743718592964824,
"grad_norm": 0.654451310634613,
"learning_rate": 3.3557046979865773e-06,
"loss": 0.7323,
"step": 1212
},
{
"epoch": 1.8759180517974487,
"grad_norm": 0.6516563892364502,
"learning_rate": 3.313758389261745e-06,
"loss": 0.6654,
"step": 1213
},
{
"epoch": 1.877464244298415,
"grad_norm": 0.6713635325431824,
"learning_rate": 3.271812080536913e-06,
"loss": 0.7639,
"step": 1214
},
{
"epoch": 1.8790104367993816,
"grad_norm": 0.6944072246551514,
"learning_rate": 3.2298657718120808e-06,
"loss": 0.7274,
"step": 1215
},
{
"epoch": 1.880556629300348,
"grad_norm": 0.6565424799919128,
"learning_rate": 3.1879194630872486e-06,
"loss": 0.7757,
"step": 1216
},
{
"epoch": 1.8821028218013143,
"grad_norm": 0.6998873949050903,
"learning_rate": 3.1459731543624164e-06,
"loss": 0.8059,
"step": 1217
},
{
"epoch": 1.8836490143022806,
"grad_norm": 0.6804831624031067,
"learning_rate": 3.1040268456375843e-06,
"loss": 0.7777,
"step": 1218
},
{
"epoch": 1.885195206803247,
"grad_norm": 0.6547567248344421,
"learning_rate": 3.062080536912752e-06,
"loss": 0.746,
"step": 1219
},
{
"epoch": 1.8867413993042135,
"grad_norm": 0.6822516322135925,
"learning_rate": 3.02013422818792e-06,
"loss": 0.7518,
"step": 1220
},
{
"epoch": 1.8882875918051798,
"grad_norm": 0.6750561594963074,
"learning_rate": 2.9781879194630873e-06,
"loss": 0.7261,
"step": 1221
},
{
"epoch": 1.8898337843061461,
"grad_norm": 0.6937829256057739,
"learning_rate": 2.936241610738255e-06,
"loss": 0.7716,
"step": 1222
},
{
"epoch": 1.8913799768071125,
"grad_norm": 0.6683257818222046,
"learning_rate": 2.894295302013423e-06,
"loss": 0.7657,
"step": 1223
},
{
"epoch": 1.8929261693080788,
"grad_norm": 0.6856616735458374,
"learning_rate": 2.852348993288591e-06,
"loss": 0.7833,
"step": 1224
},
{
"epoch": 1.8944723618090453,
"grad_norm": 0.719022274017334,
"learning_rate": 2.8104026845637587e-06,
"loss": 0.7462,
"step": 1225
},
{
"epoch": 1.8960185543100117,
"grad_norm": 0.6999730467796326,
"learning_rate": 2.7684563758389265e-06,
"loss": 0.7777,
"step": 1226
},
{
"epoch": 1.897564746810978,
"grad_norm": 0.710363507270813,
"learning_rate": 2.726510067114094e-06,
"loss": 0.7624,
"step": 1227
},
{
"epoch": 1.8991109393119443,
"grad_norm": 0.6629136204719543,
"learning_rate": 2.6845637583892617e-06,
"loss": 0.7578,
"step": 1228
},
{
"epoch": 1.9006571318129106,
"grad_norm": 0.6639029383659363,
"learning_rate": 2.6426174496644296e-06,
"loss": 0.8438,
"step": 1229
},
{
"epoch": 1.9022033243138772,
"grad_norm": 0.6924868226051331,
"learning_rate": 2.6006711409395974e-06,
"loss": 0.7844,
"step": 1230
},
{
"epoch": 1.9037495168148435,
"grad_norm": 0.7059327960014343,
"learning_rate": 2.5587248322147652e-06,
"loss": 0.7654,
"step": 1231
},
{
"epoch": 1.9052957093158098,
"grad_norm": 0.6732707619667053,
"learning_rate": 2.5167785234899326e-06,
"loss": 0.7158,
"step": 1232
},
{
"epoch": 1.9068419018167762,
"grad_norm": 0.712011456489563,
"learning_rate": 2.4748322147651005e-06,
"loss": 0.8323,
"step": 1233
},
{
"epoch": 1.9083880943177425,
"grad_norm": 0.7283968329429626,
"learning_rate": 2.4328859060402687e-06,
"loss": 0.86,
"step": 1234
},
{
"epoch": 1.909934286818709,
"grad_norm": 0.736113429069519,
"learning_rate": 2.3909395973154366e-06,
"loss": 0.8779,
"step": 1235
},
{
"epoch": 1.9114804793196754,
"grad_norm": 0.7328662276268005,
"learning_rate": 2.3489932885906044e-06,
"loss": 0.8673,
"step": 1236
},
{
"epoch": 1.9130266718206417,
"grad_norm": 0.7270045280456543,
"learning_rate": 2.3070469798657722e-06,
"loss": 0.8348,
"step": 1237
},
{
"epoch": 1.914572864321608,
"grad_norm": 0.7681441903114319,
"learning_rate": 2.2651006711409396e-06,
"loss": 0.818,
"step": 1238
},
{
"epoch": 1.9161190568225743,
"grad_norm": 0.7744415998458862,
"learning_rate": 2.2231543624161075e-06,
"loss": 0.8879,
"step": 1239
},
{
"epoch": 1.9176652493235409,
"grad_norm": 0.7644962072372437,
"learning_rate": 2.1812080536912753e-06,
"loss": 0.8545,
"step": 1240
},
{
"epoch": 1.9192114418245072,
"grad_norm": 0.7949373722076416,
"learning_rate": 2.139261744966443e-06,
"loss": 0.8603,
"step": 1241
},
{
"epoch": 1.9207576343254735,
"grad_norm": 0.8180006146430969,
"learning_rate": 2.097315436241611e-06,
"loss": 0.8146,
"step": 1242
},
{
"epoch": 1.9223038268264399,
"grad_norm": 0.8280307650566101,
"learning_rate": 2.055369127516779e-06,
"loss": 0.9628,
"step": 1243
},
{
"epoch": 1.9238500193274062,
"grad_norm": 0.8613501787185669,
"learning_rate": 2.013422818791946e-06,
"loss": 0.8326,
"step": 1244
},
{
"epoch": 1.9253962118283727,
"grad_norm": 0.9388262629508972,
"learning_rate": 1.971476510067114e-06,
"loss": 0.9428,
"step": 1245
},
{
"epoch": 1.926942404329339,
"grad_norm": 1.25608491897583,
"learning_rate": 1.929530201342282e-06,
"loss": 0.8971,
"step": 1246
},
{
"epoch": 1.9284885968303054,
"grad_norm": 0.8791813254356384,
"learning_rate": 1.88758389261745e-06,
"loss": 0.6034,
"step": 1247
},
{
"epoch": 1.9300347893312717,
"grad_norm": 0.5444540977478027,
"learning_rate": 1.8456375838926177e-06,
"loss": 0.6993,
"step": 1248
},
{
"epoch": 1.931580981832238,
"grad_norm": 0.6073552966117859,
"learning_rate": 1.8036912751677856e-06,
"loss": 0.7178,
"step": 1249
},
{
"epoch": 1.9331271743332046,
"grad_norm": 0.6113264560699463,
"learning_rate": 1.761744966442953e-06,
"loss": 0.6966,
"step": 1250
},
{
"epoch": 1.934673366834171,
"grad_norm": 0.5979108810424805,
"learning_rate": 1.7197986577181208e-06,
"loss": 0.7133,
"step": 1251
},
{
"epoch": 1.9362195593351372,
"grad_norm": 0.6315338611602783,
"learning_rate": 1.6778523489932886e-06,
"loss": 0.7721,
"step": 1252
},
{
"epoch": 1.9377657518361036,
"grad_norm": 0.6291902661323547,
"learning_rate": 1.6359060402684565e-06,
"loss": 0.7006,
"step": 1253
},
{
"epoch": 1.9393119443370699,
"grad_norm": 0.6238860487937927,
"learning_rate": 1.5939597315436243e-06,
"loss": 0.7787,
"step": 1254
},
{
"epoch": 1.9408581368380364,
"grad_norm": 0.6401642560958862,
"learning_rate": 1.5520134228187921e-06,
"loss": 0.7603,
"step": 1255
},
{
"epoch": 1.9424043293390028,
"grad_norm": 0.6174636483192444,
"learning_rate": 1.51006711409396e-06,
"loss": 0.7259,
"step": 1256
},
{
"epoch": 1.943950521839969,
"grad_norm": 0.6495472192764282,
"learning_rate": 1.4681208053691276e-06,
"loss": 0.6654,
"step": 1257
},
{
"epoch": 1.9454967143409354,
"grad_norm": 0.667812168598175,
"learning_rate": 1.4261744966442954e-06,
"loss": 0.7147,
"step": 1258
},
{
"epoch": 1.9470429068419017,
"grad_norm": 0.6613562703132629,
"learning_rate": 1.3842281879194633e-06,
"loss": 0.7437,
"step": 1259
},
{
"epoch": 1.9485890993428683,
"grad_norm": 0.7061516046524048,
"learning_rate": 1.3422818791946309e-06,
"loss": 0.6831,
"step": 1260
},
{
"epoch": 1.9501352918438346,
"grad_norm": 0.6739703416824341,
"learning_rate": 1.3003355704697987e-06,
"loss": 0.747,
"step": 1261
},
{
"epoch": 1.951681484344801,
"grad_norm": 0.6422568559646606,
"learning_rate": 1.2583892617449663e-06,
"loss": 0.7569,
"step": 1262
},
{
"epoch": 1.9532276768457673,
"grad_norm": 0.6601455211639404,
"learning_rate": 1.2164429530201344e-06,
"loss": 0.7461,
"step": 1263
},
{
"epoch": 1.9547738693467336,
"grad_norm": 0.6846182942390442,
"learning_rate": 1.1744966442953022e-06,
"loss": 0.7185,
"step": 1264
},
{
"epoch": 1.9563200618477001,
"grad_norm": 0.6746420860290527,
"learning_rate": 1.1325503355704698e-06,
"loss": 0.6806,
"step": 1265
},
{
"epoch": 1.9578662543486665,
"grad_norm": 0.6510487198829651,
"learning_rate": 1.0906040268456377e-06,
"loss": 0.7693,
"step": 1266
},
{
"epoch": 1.9594124468496328,
"grad_norm": 0.6752596497535706,
"learning_rate": 1.0486577181208055e-06,
"loss": 0.899,
"step": 1267
},
{
"epoch": 1.960958639350599,
"grad_norm": 0.698874294757843,
"learning_rate": 1.006711409395973e-06,
"loss": 0.7521,
"step": 1268
},
{
"epoch": 1.9625048318515654,
"grad_norm": 0.6726242899894714,
"learning_rate": 9.64765100671141e-07,
"loss": 0.7782,
"step": 1269
},
{
"epoch": 1.964051024352532,
"grad_norm": 0.6857203245162964,
"learning_rate": 9.228187919463089e-07,
"loss": 0.7344,
"step": 1270
},
{
"epoch": 1.9655972168534983,
"grad_norm": 0.7023079991340637,
"learning_rate": 8.808724832214765e-07,
"loss": 0.7533,
"step": 1271
},
{
"epoch": 1.9671434093544646,
"grad_norm": 0.6352121233940125,
"learning_rate": 8.389261744966443e-07,
"loss": 0.7554,
"step": 1272
},
{
"epoch": 1.968689601855431,
"grad_norm": 0.6913352608680725,
"learning_rate": 7.969798657718122e-07,
"loss": 0.7468,
"step": 1273
},
{
"epoch": 1.9702357943563973,
"grad_norm": 0.6950103640556335,
"learning_rate": 7.5503355704698e-07,
"loss": 0.8532,
"step": 1274
},
{
"epoch": 1.9717819868573638,
"grad_norm": 0.6684133410453796,
"learning_rate": 7.130872483221477e-07,
"loss": 0.8302,
"step": 1275
},
{
"epoch": 1.9733281793583302,
"grad_norm": 0.6860172152519226,
"learning_rate": 6.711409395973154e-07,
"loss": 0.7342,
"step": 1276
},
{
"epoch": 1.9748743718592965,
"grad_norm": 0.6672773957252502,
"learning_rate": 6.291946308724832e-07,
"loss": 0.7901,
"step": 1277
},
{
"epoch": 1.9764205643602628,
"grad_norm": 0.7022169828414917,
"learning_rate": 5.872483221476511e-07,
"loss": 0.7796,
"step": 1278
},
{
"epoch": 1.9779667568612291,
"grad_norm": 0.6778275370597839,
"learning_rate": 5.453020134228188e-07,
"loss": 0.8389,
"step": 1279
},
{
"epoch": 1.9795129493621957,
"grad_norm": 0.721585750579834,
"learning_rate": 5.033557046979866e-07,
"loss": 0.848,
"step": 1280
},
{
"epoch": 1.981059141863162,
"grad_norm": 0.6909308433532715,
"learning_rate": 4.6140939597315444e-07,
"loss": 0.8214,
"step": 1281
},
{
"epoch": 1.9826053343641283,
"grad_norm": 0.7320754528045654,
"learning_rate": 4.1946308724832216e-07,
"loss": 0.8179,
"step": 1282
},
{
"epoch": 1.9841515268650947,
"grad_norm": 0.7341198325157166,
"learning_rate": 3.7751677852349e-07,
"loss": 0.8212,
"step": 1283
},
{
"epoch": 1.985697719366061,
"grad_norm": 0.7166808843612671,
"learning_rate": 3.355704697986577e-07,
"loss": 0.8472,
"step": 1284
},
{
"epoch": 1.9872439118670275,
"grad_norm": 0.7207593321800232,
"learning_rate": 2.9362416107382555e-07,
"loss": 0.8408,
"step": 1285
},
{
"epoch": 1.9887901043679939,
"grad_norm": 0.7933880090713501,
"learning_rate": 2.516778523489933e-07,
"loss": 0.8781,
"step": 1286
},
{
"epoch": 1.9903362968689602,
"grad_norm": 0.7923394441604614,
"learning_rate": 2.0973154362416108e-07,
"loss": 0.9313,
"step": 1287
},
{
"epoch": 1.9918824893699265,
"grad_norm": 0.7550710439682007,
"learning_rate": 1.6778523489932886e-07,
"loss": 0.8893,
"step": 1288
},
{
"epoch": 1.9934286818708928,
"grad_norm": 0.8748595118522644,
"learning_rate": 1.2583892617449664e-07,
"loss": 0.9071,
"step": 1289
},
{
"epoch": 1.9949748743718594,
"grad_norm": 0.8392748236656189,
"learning_rate": 8.389261744966443e-08,
"loss": 0.9107,
"step": 1290
},
{
"epoch": 1.9965210668728257,
"grad_norm": 0.8647322058677673,
"learning_rate": 4.1946308724832215e-08,
"loss": 0.8547,
"step": 1291
},
{
"epoch": 1.998067259373792,
"grad_norm": 1.0320862531661987,
"learning_rate": 0.0,
"loss": 0.8732,
"step": 1292
}
],
"logging_steps": 1,
"max_steps": 1292,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.697955930285742e+17,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}