bigstorm's picture
Upload folder using huggingface_hub
1d697a7 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9996163752945689,
"eval_steps": 500,
"global_step": 1140,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0008768564695566394,
"grad_norm": 3.8354088038954104,
"learning_rate": 5.0000000000000004e-08,
"loss": 0.8827,
"step": 1
},
{
"epoch": 0.0017537129391132788,
"grad_norm": 3.854484535409196,
"learning_rate": 1.0000000000000001e-07,
"loss": 0.8816,
"step": 2
},
{
"epoch": 0.0026305694086699184,
"grad_norm": 3.871894613191576,
"learning_rate": 1.5000000000000002e-07,
"loss": 0.8801,
"step": 3
},
{
"epoch": 0.0035074258782265577,
"grad_norm": 4.015192807591418,
"learning_rate": 2.0000000000000002e-07,
"loss": 0.8778,
"step": 4
},
{
"epoch": 0.004384282347783197,
"grad_norm": 3.8093684146898625,
"learning_rate": 2.5000000000000004e-07,
"loss": 0.8711,
"step": 5
},
{
"epoch": 0.005261138817339837,
"grad_norm": 3.8610474891808035,
"learning_rate": 3.0000000000000004e-07,
"loss": 0.8774,
"step": 6
},
{
"epoch": 0.0061379952868964765,
"grad_norm": 3.7967273935876027,
"learning_rate": 3.5000000000000004e-07,
"loss": 0.8669,
"step": 7
},
{
"epoch": 0.007014851756453115,
"grad_norm": 3.6775126026184703,
"learning_rate": 4.0000000000000003e-07,
"loss": 0.8605,
"step": 8
},
{
"epoch": 0.007891708226009755,
"grad_norm": 3.8340713786963674,
"learning_rate": 4.5000000000000003e-07,
"loss": 0.8735,
"step": 9
},
{
"epoch": 0.008768564695566394,
"grad_norm": 3.7479501504503463,
"learning_rate": 5.000000000000001e-07,
"loss": 0.8843,
"step": 10
},
{
"epoch": 0.009645421165123035,
"grad_norm": 3.6317203672346734,
"learning_rate": 5.5e-07,
"loss": 0.8637,
"step": 11
},
{
"epoch": 0.010522277634679673,
"grad_norm": 3.512911808429478,
"learning_rate": 6.000000000000001e-07,
"loss": 0.8649,
"step": 12
},
{
"epoch": 0.011399134104236312,
"grad_norm": 3.5056527507086486,
"learning_rate": 6.5e-07,
"loss": 0.8514,
"step": 13
},
{
"epoch": 0.012275990573792953,
"grad_norm": 3.150666271402955,
"learning_rate": 7.000000000000001e-07,
"loss": 0.844,
"step": 14
},
{
"epoch": 0.013152847043349592,
"grad_norm": 2.92608322776606,
"learning_rate": 7.5e-07,
"loss": 0.8382,
"step": 15
},
{
"epoch": 0.01402970351290623,
"grad_norm": 3.0202821236842246,
"learning_rate": 8.000000000000001e-07,
"loss": 0.8419,
"step": 16
},
{
"epoch": 0.014906559982462871,
"grad_norm": 2.9419098502173515,
"learning_rate": 8.500000000000001e-07,
"loss": 0.8362,
"step": 17
},
{
"epoch": 0.01578341645201951,
"grad_norm": 2.7926753613205433,
"learning_rate": 9.000000000000001e-07,
"loss": 0.825,
"step": 18
},
{
"epoch": 0.01666027292157615,
"grad_norm": 2.4471605086654096,
"learning_rate": 9.500000000000001e-07,
"loss": 0.7904,
"step": 19
},
{
"epoch": 0.017537129391132788,
"grad_norm": 1.8918627793518321,
"learning_rate": 1.0000000000000002e-06,
"loss": 0.7968,
"step": 20
},
{
"epoch": 0.018413985860689427,
"grad_norm": 1.713937144355921,
"learning_rate": 1.0500000000000001e-06,
"loss": 0.7828,
"step": 21
},
{
"epoch": 0.01929084233024607,
"grad_norm": 1.4451729443975803,
"learning_rate": 1.1e-06,
"loss": 0.78,
"step": 22
},
{
"epoch": 0.020167698799802708,
"grad_norm": 1.0866085026095695,
"learning_rate": 1.1500000000000002e-06,
"loss": 0.7807,
"step": 23
},
{
"epoch": 0.021044555269359347,
"grad_norm": 1.022948274017058,
"learning_rate": 1.2000000000000002e-06,
"loss": 0.758,
"step": 24
},
{
"epoch": 0.021921411738915986,
"grad_norm": 0.976807823206357,
"learning_rate": 1.25e-06,
"loss": 0.7783,
"step": 25
},
{
"epoch": 0.022798268208472625,
"grad_norm": 2.5562950715507275,
"learning_rate": 1.3e-06,
"loss": 0.7815,
"step": 26
},
{
"epoch": 0.023675124678029263,
"grad_norm": 1.7956421603987698,
"learning_rate": 1.3500000000000002e-06,
"loss": 0.759,
"step": 27
},
{
"epoch": 0.024551981147585906,
"grad_norm": 1.3622207205502601,
"learning_rate": 1.4000000000000001e-06,
"loss": 0.7551,
"step": 28
},
{
"epoch": 0.025428837617142545,
"grad_norm": 0.9842354354215974,
"learning_rate": 1.45e-06,
"loss": 0.7625,
"step": 29
},
{
"epoch": 0.026305694086699184,
"grad_norm": 0.7679059075291825,
"learning_rate": 1.5e-06,
"loss": 0.7513,
"step": 30
},
{
"epoch": 0.027182550556255822,
"grad_norm": 0.709914193704945,
"learning_rate": 1.5500000000000002e-06,
"loss": 0.7309,
"step": 31
},
{
"epoch": 0.02805940702581246,
"grad_norm": 0.5711165082308596,
"learning_rate": 1.6000000000000001e-06,
"loss": 0.7358,
"step": 32
},
{
"epoch": 0.0289362634953691,
"grad_norm": 0.6732600160748007,
"learning_rate": 1.6500000000000003e-06,
"loss": 0.746,
"step": 33
},
{
"epoch": 0.029813119964925743,
"grad_norm": 0.519623223105866,
"learning_rate": 1.7000000000000002e-06,
"loss": 0.7408,
"step": 34
},
{
"epoch": 0.03068997643448238,
"grad_norm": 0.4967853550459734,
"learning_rate": 1.75e-06,
"loss": 0.7284,
"step": 35
},
{
"epoch": 0.03156683290403902,
"grad_norm": 0.4558474579400771,
"learning_rate": 1.8000000000000001e-06,
"loss": 0.7337,
"step": 36
},
{
"epoch": 0.03244368937359566,
"grad_norm": 0.5187940265183988,
"learning_rate": 1.85e-06,
"loss": 0.7459,
"step": 37
},
{
"epoch": 0.0333205458431523,
"grad_norm": 0.46649520265418404,
"learning_rate": 1.9000000000000002e-06,
"loss": 0.7238,
"step": 38
},
{
"epoch": 0.03419740231270894,
"grad_norm": 0.4621107554297482,
"learning_rate": 1.9500000000000004e-06,
"loss": 0.7243,
"step": 39
},
{
"epoch": 0.035074258782265576,
"grad_norm": 0.4493723053379801,
"learning_rate": 2.0000000000000003e-06,
"loss": 0.7395,
"step": 40
},
{
"epoch": 0.035951115251822215,
"grad_norm": 0.4196555282378131,
"learning_rate": 2.05e-06,
"loss": 0.7371,
"step": 41
},
{
"epoch": 0.036827971721378853,
"grad_norm": 0.3836269605839978,
"learning_rate": 2.1000000000000002e-06,
"loss": 0.7172,
"step": 42
},
{
"epoch": 0.0377048281909355,
"grad_norm": 0.38056806308372326,
"learning_rate": 2.15e-06,
"loss": 0.7163,
"step": 43
},
{
"epoch": 0.03858168466049214,
"grad_norm": 0.3561457145290273,
"learning_rate": 2.2e-06,
"loss": 0.6986,
"step": 44
},
{
"epoch": 0.03945854113004878,
"grad_norm": 0.3723153937166507,
"learning_rate": 2.25e-06,
"loss": 0.7154,
"step": 45
},
{
"epoch": 0.040335397599605416,
"grad_norm": 0.36630666691552083,
"learning_rate": 2.3000000000000004e-06,
"loss": 0.7201,
"step": 46
},
{
"epoch": 0.041212254069162055,
"grad_norm": 0.3482645877468935,
"learning_rate": 2.35e-06,
"loss": 0.7213,
"step": 47
},
{
"epoch": 0.042089110538718694,
"grad_norm": 0.35892687942862245,
"learning_rate": 2.4000000000000003e-06,
"loss": 0.7167,
"step": 48
},
{
"epoch": 0.04296596700827533,
"grad_norm": 0.3353339246028489,
"learning_rate": 2.4500000000000003e-06,
"loss": 0.7154,
"step": 49
},
{
"epoch": 0.04384282347783197,
"grad_norm": 0.3327601533732165,
"learning_rate": 2.5e-06,
"loss": 0.7149,
"step": 50
},
{
"epoch": 0.04471967994738861,
"grad_norm": 0.31047839521651305,
"learning_rate": 2.55e-06,
"loss": 0.7022,
"step": 51
},
{
"epoch": 0.04559653641694525,
"grad_norm": 0.3140715368302216,
"learning_rate": 2.6e-06,
"loss": 0.7024,
"step": 52
},
{
"epoch": 0.04647339288650189,
"grad_norm": 0.3070088967685052,
"learning_rate": 2.6500000000000005e-06,
"loss": 0.7116,
"step": 53
},
{
"epoch": 0.04735024935605853,
"grad_norm": 0.29688015435603987,
"learning_rate": 2.7000000000000004e-06,
"loss": 0.7068,
"step": 54
},
{
"epoch": 0.04822710582561517,
"grad_norm": 0.312569173156887,
"learning_rate": 2.7500000000000004e-06,
"loss": 0.708,
"step": 55
},
{
"epoch": 0.04910396229517181,
"grad_norm": 0.3212155084231398,
"learning_rate": 2.8000000000000003e-06,
"loss": 0.6895,
"step": 56
},
{
"epoch": 0.04998081876472845,
"grad_norm": 0.30141336197411556,
"learning_rate": 2.85e-06,
"loss": 0.714,
"step": 57
},
{
"epoch": 0.05085767523428509,
"grad_norm": 0.2678799864293998,
"learning_rate": 2.9e-06,
"loss": 0.6864,
"step": 58
},
{
"epoch": 0.05173453170384173,
"grad_norm": 0.2763602360222888,
"learning_rate": 2.95e-06,
"loss": 0.6955,
"step": 59
},
{
"epoch": 0.05261138817339837,
"grad_norm": 0.2960116429627635,
"learning_rate": 3e-06,
"loss": 0.69,
"step": 60
},
{
"epoch": 0.053488244642955006,
"grad_norm": 0.3126860845251708,
"learning_rate": 3.05e-06,
"loss": 0.7008,
"step": 61
},
{
"epoch": 0.054365101112511645,
"grad_norm": 0.2684477743603555,
"learning_rate": 3.1000000000000004e-06,
"loss": 0.7065,
"step": 62
},
{
"epoch": 0.055241957582068284,
"grad_norm": 0.2831279869843839,
"learning_rate": 3.1500000000000003e-06,
"loss": 0.6908,
"step": 63
},
{
"epoch": 0.05611881405162492,
"grad_norm": 0.28914936357131454,
"learning_rate": 3.2000000000000003e-06,
"loss": 0.6847,
"step": 64
},
{
"epoch": 0.05699567052118156,
"grad_norm": 0.2664694092243829,
"learning_rate": 3.2500000000000002e-06,
"loss": 0.6975,
"step": 65
},
{
"epoch": 0.0578725269907382,
"grad_norm": 0.2670931319561963,
"learning_rate": 3.3000000000000006e-06,
"loss": 0.6957,
"step": 66
},
{
"epoch": 0.058749383460294846,
"grad_norm": 0.25481964712146327,
"learning_rate": 3.3500000000000005e-06,
"loss": 0.6907,
"step": 67
},
{
"epoch": 0.059626239929851485,
"grad_norm": 0.2917224006438053,
"learning_rate": 3.4000000000000005e-06,
"loss": 0.6889,
"step": 68
},
{
"epoch": 0.060503096399408124,
"grad_norm": 0.27794604488949715,
"learning_rate": 3.45e-06,
"loss": 0.6815,
"step": 69
},
{
"epoch": 0.06137995286896476,
"grad_norm": 0.24963117175569036,
"learning_rate": 3.5e-06,
"loss": 0.6883,
"step": 70
},
{
"epoch": 0.0622568093385214,
"grad_norm": 0.2893133633641976,
"learning_rate": 3.5500000000000003e-06,
"loss": 0.6792,
"step": 71
},
{
"epoch": 0.06313366580807804,
"grad_norm": 0.2826308836822568,
"learning_rate": 3.6000000000000003e-06,
"loss": 0.7028,
"step": 72
},
{
"epoch": 0.06401052227763468,
"grad_norm": 0.2640935466003184,
"learning_rate": 3.65e-06,
"loss": 0.6916,
"step": 73
},
{
"epoch": 0.06488737874719132,
"grad_norm": 0.24415033172628944,
"learning_rate": 3.7e-06,
"loss": 0.6839,
"step": 74
},
{
"epoch": 0.06576423521674796,
"grad_norm": 0.3112401087242733,
"learning_rate": 3.7500000000000005e-06,
"loss": 0.7021,
"step": 75
},
{
"epoch": 0.0666410916863046,
"grad_norm": 0.2875281112172732,
"learning_rate": 3.8000000000000005e-06,
"loss": 0.6929,
"step": 76
},
{
"epoch": 0.06751794815586123,
"grad_norm": 0.2874092373703745,
"learning_rate": 3.85e-06,
"loss": 0.6788,
"step": 77
},
{
"epoch": 0.06839480462541787,
"grad_norm": 0.26681007920352356,
"learning_rate": 3.900000000000001e-06,
"loss": 0.6881,
"step": 78
},
{
"epoch": 0.06927166109497451,
"grad_norm": 0.25207102904583284,
"learning_rate": 3.95e-06,
"loss": 0.6852,
"step": 79
},
{
"epoch": 0.07014851756453115,
"grad_norm": 0.2747607135538642,
"learning_rate": 4.000000000000001e-06,
"loss": 0.6864,
"step": 80
},
{
"epoch": 0.07102537403408779,
"grad_norm": 0.26361955079133653,
"learning_rate": 4.05e-06,
"loss": 0.685,
"step": 81
},
{
"epoch": 0.07190223050364443,
"grad_norm": 0.33310729956901713,
"learning_rate": 4.1e-06,
"loss": 0.6803,
"step": 82
},
{
"epoch": 0.07277908697320107,
"grad_norm": 0.2453664087918243,
"learning_rate": 4.15e-06,
"loss": 0.6761,
"step": 83
},
{
"epoch": 0.07365594344275771,
"grad_norm": 0.2908734202511105,
"learning_rate": 4.2000000000000004e-06,
"loss": 0.6931,
"step": 84
},
{
"epoch": 0.07453279991231436,
"grad_norm": 0.2786719287704165,
"learning_rate": 4.25e-06,
"loss": 0.6874,
"step": 85
},
{
"epoch": 0.075409656381871,
"grad_norm": 0.271512101257661,
"learning_rate": 4.3e-06,
"loss": 0.6775,
"step": 86
},
{
"epoch": 0.07628651285142764,
"grad_norm": 0.2947304767213564,
"learning_rate": 4.350000000000001e-06,
"loss": 0.6865,
"step": 87
},
{
"epoch": 0.07716336932098428,
"grad_norm": 0.25160176616217883,
"learning_rate": 4.4e-06,
"loss": 0.6785,
"step": 88
},
{
"epoch": 0.07804022579054092,
"grad_norm": 0.32459153781403244,
"learning_rate": 4.450000000000001e-06,
"loss": 0.6773,
"step": 89
},
{
"epoch": 0.07891708226009755,
"grad_norm": 0.2487028104553641,
"learning_rate": 4.5e-06,
"loss": 0.6812,
"step": 90
},
{
"epoch": 0.07979393872965419,
"grad_norm": 0.2925038544983962,
"learning_rate": 4.5500000000000005e-06,
"loss": 0.6791,
"step": 91
},
{
"epoch": 0.08067079519921083,
"grad_norm": 0.28005649996035475,
"learning_rate": 4.600000000000001e-06,
"loss": 0.6704,
"step": 92
},
{
"epoch": 0.08154765166876747,
"grad_norm": 0.3264776457957641,
"learning_rate": 4.65e-06,
"loss": 0.6772,
"step": 93
},
{
"epoch": 0.08242450813832411,
"grad_norm": 0.2533079586966528,
"learning_rate": 4.7e-06,
"loss": 0.6792,
"step": 94
},
{
"epoch": 0.08330136460788075,
"grad_norm": 0.25651763696878965,
"learning_rate": 4.75e-06,
"loss": 0.6607,
"step": 95
},
{
"epoch": 0.08417822107743739,
"grad_norm": 0.2546288408258964,
"learning_rate": 4.800000000000001e-06,
"loss": 0.6669,
"step": 96
},
{
"epoch": 0.08505507754699403,
"grad_norm": 0.25215356470309513,
"learning_rate": 4.85e-06,
"loss": 0.6846,
"step": 97
},
{
"epoch": 0.08593193401655067,
"grad_norm": 0.28631928221309494,
"learning_rate": 4.9000000000000005e-06,
"loss": 0.6717,
"step": 98
},
{
"epoch": 0.0868087904861073,
"grad_norm": 0.27212851090592044,
"learning_rate": 4.95e-06,
"loss": 0.6804,
"step": 99
},
{
"epoch": 0.08768564695566394,
"grad_norm": 0.29348118762199116,
"learning_rate": 5e-06,
"loss": 0.669,
"step": 100
},
{
"epoch": 0.08856250342522058,
"grad_norm": 0.30678288402779474,
"learning_rate": 4.999998880733363e-06,
"loss": 0.6631,
"step": 101
},
{
"epoch": 0.08943935989477722,
"grad_norm": 0.3011120934546324,
"learning_rate": 4.999995522934454e-06,
"loss": 0.679,
"step": 102
},
{
"epoch": 0.09031621636433386,
"grad_norm": 0.31706623056013666,
"learning_rate": 4.9999899266062804e-06,
"loss": 0.6723,
"step": 103
},
{
"epoch": 0.0911930728338905,
"grad_norm": 0.3120471729111099,
"learning_rate": 4.999982091753851e-06,
"loss": 0.6613,
"step": 104
},
{
"epoch": 0.09206992930344714,
"grad_norm": 0.2905613969012575,
"learning_rate": 4.999972018384183e-06,
"loss": 0.6611,
"step": 105
},
{
"epoch": 0.09294678577300378,
"grad_norm": 0.28925318733211003,
"learning_rate": 4.999959706506297e-06,
"loss": 0.6695,
"step": 106
},
{
"epoch": 0.09382364224256041,
"grad_norm": 0.28085987028825943,
"learning_rate": 4.999945156131215e-06,
"loss": 0.6502,
"step": 107
},
{
"epoch": 0.09470049871211705,
"grad_norm": 0.30971852568333075,
"learning_rate": 4.9999283672719665e-06,
"loss": 0.672,
"step": 108
},
{
"epoch": 0.0955773551816737,
"grad_norm": 0.32363303577963826,
"learning_rate": 4.999909339943585e-06,
"loss": 0.673,
"step": 109
},
{
"epoch": 0.09645421165123035,
"grad_norm": 0.29549042512555623,
"learning_rate": 4.999888074163108e-06,
"loss": 0.6591,
"step": 110
},
{
"epoch": 0.09733106812078698,
"grad_norm": 0.33514032815726946,
"learning_rate": 4.999864569949576e-06,
"loss": 0.6673,
"step": 111
},
{
"epoch": 0.09820792459034362,
"grad_norm": 0.3092438114721304,
"learning_rate": 4.999838827324036e-06,
"loss": 0.6641,
"step": 112
},
{
"epoch": 0.09908478105990026,
"grad_norm": 0.35403209993563217,
"learning_rate": 4.999810846309539e-06,
"loss": 0.6597,
"step": 113
},
{
"epoch": 0.0999616375294569,
"grad_norm": 0.2964896689419525,
"learning_rate": 4.999780626931136e-06,
"loss": 0.67,
"step": 114
},
{
"epoch": 0.10083849399901354,
"grad_norm": 0.3484706075226941,
"learning_rate": 4.999748169215891e-06,
"loss": 0.6745,
"step": 115
},
{
"epoch": 0.10171535046857018,
"grad_norm": 0.33505074735981694,
"learning_rate": 4.999713473192863e-06,
"loss": 0.6591,
"step": 116
},
{
"epoch": 0.10259220693812682,
"grad_norm": 0.27082614750107925,
"learning_rate": 4.999676538893121e-06,
"loss": 0.6621,
"step": 117
},
{
"epoch": 0.10346906340768346,
"grad_norm": 0.3506965847465109,
"learning_rate": 4.999637366349736e-06,
"loss": 0.6733,
"step": 118
},
{
"epoch": 0.1043459198772401,
"grad_norm": 0.27422374937685745,
"learning_rate": 4.999595955597784e-06,
"loss": 0.655,
"step": 119
},
{
"epoch": 0.10522277634679673,
"grad_norm": 0.33620430443399,
"learning_rate": 4.999552306674345e-06,
"loss": 0.6755,
"step": 120
},
{
"epoch": 0.10609963281635337,
"grad_norm": 0.2837804889330797,
"learning_rate": 4.999506419618502e-06,
"loss": 0.6579,
"step": 121
},
{
"epoch": 0.10697648928591001,
"grad_norm": 0.37952040871876175,
"learning_rate": 4.999458294471342e-06,
"loss": 0.6692,
"step": 122
},
{
"epoch": 0.10785334575546665,
"grad_norm": 0.2690864525050558,
"learning_rate": 4.99940793127596e-06,
"loss": 0.6494,
"step": 123
},
{
"epoch": 0.10873020222502329,
"grad_norm": 0.3635002166658454,
"learning_rate": 4.999355330077449e-06,
"loss": 0.6611,
"step": 124
},
{
"epoch": 0.10960705869457993,
"grad_norm": 0.29302462194523843,
"learning_rate": 4.999300490922911e-06,
"loss": 0.6526,
"step": 125
},
{
"epoch": 0.11048391516413657,
"grad_norm": 0.3058787861740299,
"learning_rate": 4.999243413861447e-06,
"loss": 0.659,
"step": 126
},
{
"epoch": 0.1113607716336932,
"grad_norm": 0.332548080761125,
"learning_rate": 4.9991840989441665e-06,
"loss": 0.6659,
"step": 127
},
{
"epoch": 0.11223762810324985,
"grad_norm": 0.29432766212441813,
"learning_rate": 4.999122546224181e-06,
"loss": 0.6447,
"step": 128
},
{
"epoch": 0.11311448457280648,
"grad_norm": 0.29523416391879537,
"learning_rate": 4.999058755756605e-06,
"loss": 0.6587,
"step": 129
},
{
"epoch": 0.11399134104236312,
"grad_norm": 0.32423165831626255,
"learning_rate": 4.998992727598557e-06,
"loss": 0.6564,
"step": 130
},
{
"epoch": 0.11486819751191976,
"grad_norm": 0.34859884756639065,
"learning_rate": 4.99892446180916e-06,
"loss": 0.653,
"step": 131
},
{
"epoch": 0.1157450539814764,
"grad_norm": 0.30133447855543133,
"learning_rate": 4.99885395844954e-06,
"loss": 0.647,
"step": 132
},
{
"epoch": 0.11662191045103305,
"grad_norm": 0.3600942516700186,
"learning_rate": 4.998781217582827e-06,
"loss": 0.6581,
"step": 133
},
{
"epoch": 0.11749876692058969,
"grad_norm": 0.29960571448156953,
"learning_rate": 4.998706239274153e-06,
"loss": 0.6623,
"step": 134
},
{
"epoch": 0.11837562339014633,
"grad_norm": 0.2992208264370026,
"learning_rate": 4.998629023590656e-06,
"loss": 0.6538,
"step": 135
},
{
"epoch": 0.11925247985970297,
"grad_norm": 0.36522912538035174,
"learning_rate": 4.998549570601475e-06,
"loss": 0.6566,
"step": 136
},
{
"epoch": 0.12012933632925961,
"grad_norm": 0.2988448634710597,
"learning_rate": 4.998467880377754e-06,
"loss": 0.673,
"step": 137
},
{
"epoch": 0.12100619279881625,
"grad_norm": 0.32912250244162505,
"learning_rate": 4.998383952992639e-06,
"loss": 0.6482,
"step": 138
},
{
"epoch": 0.12188304926837289,
"grad_norm": 0.37178534793553225,
"learning_rate": 4.998297788521279e-06,
"loss": 0.6546,
"step": 139
},
{
"epoch": 0.12275990573792953,
"grad_norm": 0.28062782891296695,
"learning_rate": 4.998209387040829e-06,
"loss": 0.6527,
"step": 140
},
{
"epoch": 0.12363676220748616,
"grad_norm": 0.33723394797540485,
"learning_rate": 4.998118748630443e-06,
"loss": 0.6391,
"step": 141
},
{
"epoch": 0.1245136186770428,
"grad_norm": 0.2834572318610097,
"learning_rate": 4.99802587337128e-06,
"loss": 0.6443,
"step": 142
},
{
"epoch": 0.12539047514659943,
"grad_norm": 0.321495289367043,
"learning_rate": 4.997930761346502e-06,
"loss": 0.6507,
"step": 143
},
{
"epoch": 0.12626733161615608,
"grad_norm": 0.3419910878952078,
"learning_rate": 4.997833412641274e-06,
"loss": 0.6543,
"step": 144
},
{
"epoch": 0.1271441880857127,
"grad_norm": 0.28772221770446305,
"learning_rate": 4.9977338273427625e-06,
"loss": 0.6522,
"step": 145
},
{
"epoch": 0.12802104455526936,
"grad_norm": 0.29706932671928316,
"learning_rate": 4.997632005540139e-06,
"loss": 0.6677,
"step": 146
},
{
"epoch": 0.128897901024826,
"grad_norm": 0.29918610448467253,
"learning_rate": 4.997527947324573e-06,
"loss": 0.6475,
"step": 147
},
{
"epoch": 0.12977475749438264,
"grad_norm": 0.33103419851925103,
"learning_rate": 4.997421652789243e-06,
"loss": 0.67,
"step": 148
},
{
"epoch": 0.1306516139639393,
"grad_norm": 0.27012500247528487,
"learning_rate": 4.9973131220293255e-06,
"loss": 0.647,
"step": 149
},
{
"epoch": 0.13152847043349591,
"grad_norm": 0.297677443804652,
"learning_rate": 4.9972023551419995e-06,
"loss": 0.6519,
"step": 150
},
{
"epoch": 0.13240532690305257,
"grad_norm": 0.27386600476743567,
"learning_rate": 4.997089352226448e-06,
"loss": 0.6562,
"step": 151
},
{
"epoch": 0.1332821833726092,
"grad_norm": 0.3025435071675535,
"learning_rate": 4.996974113383854e-06,
"loss": 0.6485,
"step": 152
},
{
"epoch": 0.13415903984216584,
"grad_norm": 0.2928572797854547,
"learning_rate": 4.996856638717406e-06,
"loss": 0.641,
"step": 153
},
{
"epoch": 0.13503589631172247,
"grad_norm": 0.28232417223789874,
"learning_rate": 4.996736928332292e-06,
"loss": 0.6358,
"step": 154
},
{
"epoch": 0.13591275278127912,
"grad_norm": 0.33877806926878856,
"learning_rate": 4.9966149823357e-06,
"loss": 0.6558,
"step": 155
},
{
"epoch": 0.13678960925083575,
"grad_norm": 0.27274924720742,
"learning_rate": 4.996490800836825e-06,
"loss": 0.6553,
"step": 156
},
{
"epoch": 0.1376664657203924,
"grad_norm": 0.3145522020468823,
"learning_rate": 4.996364383946859e-06,
"loss": 0.6458,
"step": 157
},
{
"epoch": 0.13854332218994903,
"grad_norm": 0.28298098932682264,
"learning_rate": 4.996235731778997e-06,
"loss": 0.6467,
"step": 158
},
{
"epoch": 0.13942017865950568,
"grad_norm": 0.3289393703740858,
"learning_rate": 4.996104844448438e-06,
"loss": 0.6522,
"step": 159
},
{
"epoch": 0.1402970351290623,
"grad_norm": 0.3242491154179804,
"learning_rate": 4.995971722072379e-06,
"loss": 0.6579,
"step": 160
},
{
"epoch": 0.14117389159861896,
"grad_norm": 0.350063023556927,
"learning_rate": 4.995836364770018e-06,
"loss": 0.6639,
"step": 161
},
{
"epoch": 0.14205074806817558,
"grad_norm": 0.26800977502782475,
"learning_rate": 4.995698772662558e-06,
"loss": 0.6564,
"step": 162
},
{
"epoch": 0.14292760453773223,
"grad_norm": 0.37123972908338404,
"learning_rate": 4.9955589458732e-06,
"loss": 0.6521,
"step": 163
},
{
"epoch": 0.14380446100728886,
"grad_norm": 0.25568101611736427,
"learning_rate": 4.995416884527147e-06,
"loss": 0.6489,
"step": 164
},
{
"epoch": 0.1446813174768455,
"grad_norm": 0.3502739955437778,
"learning_rate": 4.9952725887516015e-06,
"loss": 0.6389,
"step": 165
},
{
"epoch": 0.14555817394640214,
"grad_norm": 0.2695951493086468,
"learning_rate": 4.99512605867577e-06,
"loss": 0.6409,
"step": 166
},
{
"epoch": 0.1464350304159588,
"grad_norm": 0.33224546665642934,
"learning_rate": 4.994977294430856e-06,
"loss": 0.6478,
"step": 167
},
{
"epoch": 0.14731188688551541,
"grad_norm": 0.26336591640433304,
"learning_rate": 4.994826296150064e-06,
"loss": 0.6416,
"step": 168
},
{
"epoch": 0.14818874335507207,
"grad_norm": 0.3158628283831438,
"learning_rate": 4.9946730639686025e-06,
"loss": 0.6397,
"step": 169
},
{
"epoch": 0.14906559982462872,
"grad_norm": 0.29572803602407627,
"learning_rate": 4.9945175980236745e-06,
"loss": 0.6356,
"step": 170
},
{
"epoch": 0.14994245629418534,
"grad_norm": 0.3344536076519792,
"learning_rate": 4.99435989845449e-06,
"loss": 0.6494,
"step": 171
},
{
"epoch": 0.150819312763742,
"grad_norm": 0.2811402499936693,
"learning_rate": 4.994199965402252e-06,
"loss": 0.6472,
"step": 172
},
{
"epoch": 0.15169616923329862,
"grad_norm": 0.30351530565920815,
"learning_rate": 4.994037799010168e-06,
"loss": 0.6514,
"step": 173
},
{
"epoch": 0.15257302570285527,
"grad_norm": 0.2667020904201129,
"learning_rate": 4.993873399423445e-06,
"loss": 0.642,
"step": 174
},
{
"epoch": 0.1534498821724119,
"grad_norm": 0.3062654941965369,
"learning_rate": 4.993706766789287e-06,
"loss": 0.6398,
"step": 175
},
{
"epoch": 0.15432673864196855,
"grad_norm": 0.28228507467929365,
"learning_rate": 4.993537901256898e-06,
"loss": 0.6446,
"step": 176
},
{
"epoch": 0.15520359511152518,
"grad_norm": 0.3157908119401443,
"learning_rate": 4.993366802977486e-06,
"loss": 0.645,
"step": 177
},
{
"epoch": 0.15608045158108183,
"grad_norm": 0.29612114085869035,
"learning_rate": 4.993193472104253e-06,
"loss": 0.6379,
"step": 178
},
{
"epoch": 0.15695730805063846,
"grad_norm": 0.31715005105530436,
"learning_rate": 4.9930179087924e-06,
"loss": 0.6446,
"step": 179
},
{
"epoch": 0.1578341645201951,
"grad_norm": 0.3010974405602859,
"learning_rate": 4.992840113199131e-06,
"loss": 0.6273,
"step": 180
},
{
"epoch": 0.15871102098975173,
"grad_norm": 0.3097310667014726,
"learning_rate": 4.992660085483645e-06,
"loss": 0.6477,
"step": 181
},
{
"epoch": 0.15958787745930839,
"grad_norm": 0.25428924204211556,
"learning_rate": 4.992477825807142e-06,
"loss": 0.6562,
"step": 182
},
{
"epoch": 0.160464733928865,
"grad_norm": 0.30870425916577926,
"learning_rate": 4.992293334332821e-06,
"loss": 0.6528,
"step": 183
},
{
"epoch": 0.16134159039842166,
"grad_norm": 0.2915653234864446,
"learning_rate": 4.992106611225875e-06,
"loss": 0.6491,
"step": 184
},
{
"epoch": 0.1622184468679783,
"grad_norm": 0.3032380988277513,
"learning_rate": 4.991917656653501e-06,
"loss": 0.6523,
"step": 185
},
{
"epoch": 0.16309530333753494,
"grad_norm": 0.2986663700583823,
"learning_rate": 4.991726470784891e-06,
"loss": 0.6333,
"step": 186
},
{
"epoch": 0.16397215980709157,
"grad_norm": 0.28321065505069615,
"learning_rate": 4.9915330537912346e-06,
"loss": 0.6411,
"step": 187
},
{
"epoch": 0.16484901627664822,
"grad_norm": 0.358610834369166,
"learning_rate": 4.99133740584572e-06,
"loss": 0.6404,
"step": 188
},
{
"epoch": 0.16572587274620484,
"grad_norm": 0.30976208589225795,
"learning_rate": 4.991139527123534e-06,
"loss": 0.6405,
"step": 189
},
{
"epoch": 0.1666027292157615,
"grad_norm": 0.34149502314365515,
"learning_rate": 4.990939417801859e-06,
"loss": 0.6384,
"step": 190
},
{
"epoch": 0.16747958568531812,
"grad_norm": 0.2959951500432587,
"learning_rate": 4.9907370780598754e-06,
"loss": 0.6469,
"step": 191
},
{
"epoch": 0.16835644215487477,
"grad_norm": 0.3302476980977895,
"learning_rate": 4.990532508078761e-06,
"loss": 0.6359,
"step": 192
},
{
"epoch": 0.1692332986244314,
"grad_norm": 0.3944297035939378,
"learning_rate": 4.990325708041691e-06,
"loss": 0.6502,
"step": 193
},
{
"epoch": 0.17011015509398805,
"grad_norm": 0.360231124267091,
"learning_rate": 4.990116678133836e-06,
"loss": 0.6424,
"step": 194
},
{
"epoch": 0.1709870115635447,
"grad_norm": 0.33832741778437936,
"learning_rate": 4.989905418542366e-06,
"loss": 0.6352,
"step": 195
},
{
"epoch": 0.17186386803310133,
"grad_norm": 0.36238295597291414,
"learning_rate": 4.989691929456443e-06,
"loss": 0.6499,
"step": 196
},
{
"epoch": 0.17274072450265798,
"grad_norm": 0.32684488652867627,
"learning_rate": 4.98947621106723e-06,
"loss": 0.6475,
"step": 197
},
{
"epoch": 0.1736175809722146,
"grad_norm": 0.2757346118610075,
"learning_rate": 4.989258263567884e-06,
"loss": 0.6355,
"step": 198
},
{
"epoch": 0.17449443744177126,
"grad_norm": 0.29755713041423115,
"learning_rate": 4.989038087153556e-06,
"loss": 0.6336,
"step": 199
},
{
"epoch": 0.17537129391132789,
"grad_norm": 0.29151765698243737,
"learning_rate": 4.988815682021398e-06,
"loss": 0.6471,
"step": 200
},
{
"epoch": 0.17624815038088454,
"grad_norm": 0.28111823253643253,
"learning_rate": 4.988591048370552e-06,
"loss": 0.6407,
"step": 201
},
{
"epoch": 0.17712500685044116,
"grad_norm": 0.2656165957748681,
"learning_rate": 4.988364186402159e-06,
"loss": 0.6326,
"step": 202
},
{
"epoch": 0.17800186331999782,
"grad_norm": 0.3028986715129606,
"learning_rate": 4.988135096319355e-06,
"loss": 0.6348,
"step": 203
},
{
"epoch": 0.17887871978955444,
"grad_norm": 0.29924585956112065,
"learning_rate": 4.987903778327269e-06,
"loss": 0.6488,
"step": 204
},
{
"epoch": 0.1797555762591111,
"grad_norm": 0.2747438588784908,
"learning_rate": 4.987670232633027e-06,
"loss": 0.6353,
"step": 205
},
{
"epoch": 0.18063243272866772,
"grad_norm": 0.30887265845064044,
"learning_rate": 4.987434459445748e-06,
"loss": 0.6428,
"step": 206
},
{
"epoch": 0.18150928919822437,
"grad_norm": 0.3193061834187564,
"learning_rate": 4.987196458976548e-06,
"loss": 0.6467,
"step": 207
},
{
"epoch": 0.182386145667781,
"grad_norm": 0.2769424032566695,
"learning_rate": 4.9869562314385335e-06,
"loss": 0.6407,
"step": 208
},
{
"epoch": 0.18326300213733765,
"grad_norm": 0.3406015148633883,
"learning_rate": 4.986713777046809e-06,
"loss": 0.6443,
"step": 209
},
{
"epoch": 0.18413985860689427,
"grad_norm": 0.271878066659463,
"learning_rate": 4.986469096018472e-06,
"loss": 0.6328,
"step": 210
},
{
"epoch": 0.18501671507645093,
"grad_norm": 0.2987491049335003,
"learning_rate": 4.9862221885726115e-06,
"loss": 0.6478,
"step": 211
},
{
"epoch": 0.18589357154600755,
"grad_norm": 0.3087618217189243,
"learning_rate": 4.985973054930313e-06,
"loss": 0.6363,
"step": 212
},
{
"epoch": 0.1867704280155642,
"grad_norm": 0.28612704652497223,
"learning_rate": 4.985721695314653e-06,
"loss": 0.6409,
"step": 213
},
{
"epoch": 0.18764728448512083,
"grad_norm": 0.26033127989473615,
"learning_rate": 4.985468109950704e-06,
"loss": 0.6495,
"step": 214
},
{
"epoch": 0.18852414095467748,
"grad_norm": 0.29345494621139656,
"learning_rate": 4.985212299065528e-06,
"loss": 0.648,
"step": 215
},
{
"epoch": 0.1894009974242341,
"grad_norm": 0.30811406203792147,
"learning_rate": 4.984954262888182e-06,
"loss": 0.639,
"step": 216
},
{
"epoch": 0.19027785389379076,
"grad_norm": 0.3312828084167346,
"learning_rate": 4.9846940016497146e-06,
"loss": 0.6403,
"step": 217
},
{
"epoch": 0.1911547103633474,
"grad_norm": 0.29106752415257064,
"learning_rate": 4.984431515583169e-06,
"loss": 0.6457,
"step": 218
},
{
"epoch": 0.19203156683290404,
"grad_norm": 0.2950307203873666,
"learning_rate": 4.984166804923576e-06,
"loss": 0.6366,
"step": 219
},
{
"epoch": 0.1929084233024607,
"grad_norm": 0.33001978484003053,
"learning_rate": 4.983899869907963e-06,
"loss": 0.6519,
"step": 220
},
{
"epoch": 0.19378527977201732,
"grad_norm": 0.25712182858786903,
"learning_rate": 4.983630710775346e-06,
"loss": 0.6302,
"step": 221
},
{
"epoch": 0.19466213624157397,
"grad_norm": 0.33700258932320354,
"learning_rate": 4.983359327766735e-06,
"loss": 0.6382,
"step": 222
},
{
"epoch": 0.1955389927111306,
"grad_norm": 0.3195952299259763,
"learning_rate": 4.983085721125128e-06,
"loss": 0.6408,
"step": 223
},
{
"epoch": 0.19641584918068725,
"grad_norm": 0.2820582636542398,
"learning_rate": 4.982809891095519e-06,
"loss": 0.6196,
"step": 224
},
{
"epoch": 0.19729270565024387,
"grad_norm": 0.30343326038998625,
"learning_rate": 4.982531837924887e-06,
"loss": 0.6361,
"step": 225
},
{
"epoch": 0.19816956211980052,
"grad_norm": 0.2724213298701267,
"learning_rate": 4.9822515618622055e-06,
"loss": 0.6455,
"step": 226
},
{
"epoch": 0.19904641858935715,
"grad_norm": 0.28433275446155476,
"learning_rate": 4.9819690631584375e-06,
"loss": 0.6329,
"step": 227
},
{
"epoch": 0.1999232750589138,
"grad_norm": 0.2641523923467397,
"learning_rate": 4.981684342066536e-06,
"loss": 0.6301,
"step": 228
},
{
"epoch": 0.20080013152847043,
"grad_norm": 0.29243768749633176,
"learning_rate": 4.9813973988414454e-06,
"loss": 0.6369,
"step": 229
},
{
"epoch": 0.20167698799802708,
"grad_norm": 0.27139535071517695,
"learning_rate": 4.981108233740096e-06,
"loss": 0.6279,
"step": 230
},
{
"epoch": 0.2025538444675837,
"grad_norm": 0.27525475223350887,
"learning_rate": 4.980816847021412e-06,
"loss": 0.6429,
"step": 231
},
{
"epoch": 0.20343070093714036,
"grad_norm": 0.3427701449667448,
"learning_rate": 4.980523238946304e-06,
"loss": 0.6438,
"step": 232
},
{
"epoch": 0.20430755740669698,
"grad_norm": 0.2574596630900604,
"learning_rate": 4.980227409777673e-06,
"loss": 0.6278,
"step": 233
},
{
"epoch": 0.20518441387625364,
"grad_norm": 0.3069435432493287,
"learning_rate": 4.9799293597804086e-06,
"loss": 0.645,
"step": 234
},
{
"epoch": 0.20606127034581026,
"grad_norm": 0.2861360169316533,
"learning_rate": 4.979629089221387e-06,
"loss": 0.646,
"step": 235
},
{
"epoch": 0.2069381268153669,
"grad_norm": 0.258606470239814,
"learning_rate": 4.9793265983694775e-06,
"loss": 0.638,
"step": 236
},
{
"epoch": 0.20781498328492354,
"grad_norm": 0.2852233202848665,
"learning_rate": 4.9790218874955325e-06,
"loss": 0.6233,
"step": 237
},
{
"epoch": 0.2086918397544802,
"grad_norm": 0.27593128237727194,
"learning_rate": 4.978714956872394e-06,
"loss": 0.64,
"step": 238
},
{
"epoch": 0.20956869622403682,
"grad_norm": 0.2721892419938629,
"learning_rate": 4.978405806774892e-06,
"loss": 0.6242,
"step": 239
},
{
"epoch": 0.21044555269359347,
"grad_norm": 0.26477694173686633,
"learning_rate": 4.978094437479843e-06,
"loss": 0.6409,
"step": 240
},
{
"epoch": 0.2113224091631501,
"grad_norm": 0.29511740452877416,
"learning_rate": 4.977780849266054e-06,
"loss": 0.6397,
"step": 241
},
{
"epoch": 0.21219926563270675,
"grad_norm": 0.3137075106480887,
"learning_rate": 4.977465042414314e-06,
"loss": 0.6185,
"step": 242
},
{
"epoch": 0.2130761221022634,
"grad_norm": 0.2841757272525764,
"learning_rate": 4.9771470172073985e-06,
"loss": 0.6394,
"step": 243
},
{
"epoch": 0.21395297857182002,
"grad_norm": 0.289636229771129,
"learning_rate": 4.976826773930076e-06,
"loss": 0.6314,
"step": 244
},
{
"epoch": 0.21482983504137668,
"grad_norm": 0.30163996035868273,
"learning_rate": 4.976504312869093e-06,
"loss": 0.6347,
"step": 245
},
{
"epoch": 0.2157066915109333,
"grad_norm": 0.261372963985366,
"learning_rate": 4.976179634313187e-06,
"loss": 0.6378,
"step": 246
},
{
"epoch": 0.21658354798048995,
"grad_norm": 0.3277256326536918,
"learning_rate": 4.97585273855308e-06,
"loss": 0.6326,
"step": 247
},
{
"epoch": 0.21746040445004658,
"grad_norm": 0.2609300415027874,
"learning_rate": 4.975523625881478e-06,
"loss": 0.643,
"step": 248
},
{
"epoch": 0.21833726091960323,
"grad_norm": 0.360435554160976,
"learning_rate": 4.975192296593072e-06,
"loss": 0.6301,
"step": 249
},
{
"epoch": 0.21921411738915986,
"grad_norm": 0.33545569496984357,
"learning_rate": 4.97485875098454e-06,
"loss": 0.6263,
"step": 250
},
{
"epoch": 0.2200909738587165,
"grad_norm": 0.3109257543138659,
"learning_rate": 4.974522989354544e-06,
"loss": 0.6409,
"step": 251
},
{
"epoch": 0.22096783032827313,
"grad_norm": 0.324992218124581,
"learning_rate": 4.974185012003727e-06,
"loss": 0.634,
"step": 252
},
{
"epoch": 0.2218446867978298,
"grad_norm": 0.32486130027399085,
"learning_rate": 4.97384481923472e-06,
"loss": 0.6164,
"step": 253
},
{
"epoch": 0.2227215432673864,
"grad_norm": 0.37258515700556377,
"learning_rate": 4.973502411352136e-06,
"loss": 0.6387,
"step": 254
},
{
"epoch": 0.22359839973694307,
"grad_norm": 0.29043553996012594,
"learning_rate": 4.97315778866257e-06,
"loss": 0.6287,
"step": 255
},
{
"epoch": 0.2244752562064997,
"grad_norm": 0.36257038619483317,
"learning_rate": 4.972810951474605e-06,
"loss": 0.6343,
"step": 256
},
{
"epoch": 0.22535211267605634,
"grad_norm": 0.2772793728031826,
"learning_rate": 4.972461900098801e-06,
"loss": 0.6289,
"step": 257
},
{
"epoch": 0.22622896914561297,
"grad_norm": 0.35920004083908574,
"learning_rate": 4.972110634847703e-06,
"loss": 0.6532,
"step": 258
},
{
"epoch": 0.22710582561516962,
"grad_norm": 0.29471007707943336,
"learning_rate": 4.97175715603584e-06,
"loss": 0.6431,
"step": 259
},
{
"epoch": 0.22798268208472625,
"grad_norm": 0.3052965075835166,
"learning_rate": 4.971401463979722e-06,
"loss": 0.6373,
"step": 260
},
{
"epoch": 0.2288595385542829,
"grad_norm": 0.27702925326859024,
"learning_rate": 4.971043558997839e-06,
"loss": 0.6254,
"step": 261
},
{
"epoch": 0.22973639502383952,
"grad_norm": 0.30905022457424325,
"learning_rate": 4.9706834414106645e-06,
"loss": 0.6377,
"step": 262
},
{
"epoch": 0.23061325149339618,
"grad_norm": 0.2820956276882666,
"learning_rate": 4.970321111540652e-06,
"loss": 0.6303,
"step": 263
},
{
"epoch": 0.2314901079629528,
"grad_norm": 0.3394900289735489,
"learning_rate": 4.969956569712238e-06,
"loss": 0.6394,
"step": 264
},
{
"epoch": 0.23236696443250945,
"grad_norm": 0.26647926556067275,
"learning_rate": 4.969589816251837e-06,
"loss": 0.6202,
"step": 265
},
{
"epoch": 0.2332438209020661,
"grad_norm": 0.3281231898594553,
"learning_rate": 4.9692208514878445e-06,
"loss": 0.6343,
"step": 266
},
{
"epoch": 0.23412067737162273,
"grad_norm": 0.32675488207496506,
"learning_rate": 4.968849675750638e-06,
"loss": 0.6106,
"step": 267
},
{
"epoch": 0.23499753384117938,
"grad_norm": 0.28838375524590465,
"learning_rate": 4.9684762893725715e-06,
"loss": 0.6191,
"step": 268
},
{
"epoch": 0.235874390310736,
"grad_norm": 0.3568027126734991,
"learning_rate": 4.968100692687981e-06,
"loss": 0.6492,
"step": 269
},
{
"epoch": 0.23675124678029266,
"grad_norm": 0.28443576918161984,
"learning_rate": 4.967722886033181e-06,
"loss": 0.6332,
"step": 270
},
{
"epoch": 0.2376281032498493,
"grad_norm": 0.34347891151295074,
"learning_rate": 4.967342869746463e-06,
"loss": 0.6302,
"step": 271
},
{
"epoch": 0.23850495971940594,
"grad_norm": 0.26856199334324765,
"learning_rate": 4.9669606441681005e-06,
"loss": 0.6253,
"step": 272
},
{
"epoch": 0.23938181618896257,
"grad_norm": 0.28792821400673596,
"learning_rate": 4.966576209640344e-06,
"loss": 0.617,
"step": 273
},
{
"epoch": 0.24025867265851922,
"grad_norm": 0.2749481611356667,
"learning_rate": 4.966189566507418e-06,
"loss": 0.6386,
"step": 274
},
{
"epoch": 0.24113552912807584,
"grad_norm": 0.2499995559979677,
"learning_rate": 4.965800715115531e-06,
"loss": 0.6281,
"step": 275
},
{
"epoch": 0.2420123855976325,
"grad_norm": 0.2802197876098476,
"learning_rate": 4.965409655812865e-06,
"loss": 0.6356,
"step": 276
},
{
"epoch": 0.24288924206718912,
"grad_norm": 0.27112050232805884,
"learning_rate": 4.965016388949579e-06,
"loss": 0.6366,
"step": 277
},
{
"epoch": 0.24376609853674577,
"grad_norm": 0.28745747065199806,
"learning_rate": 4.96462091487781e-06,
"loss": 0.6245,
"step": 278
},
{
"epoch": 0.2446429550063024,
"grad_norm": 0.29635776688822807,
"learning_rate": 4.96422323395167e-06,
"loss": 0.6413,
"step": 279
},
{
"epoch": 0.24551981147585905,
"grad_norm": 0.3376283192201481,
"learning_rate": 4.963823346527249e-06,
"loss": 0.6322,
"step": 280
},
{
"epoch": 0.24639666794541568,
"grad_norm": 0.30520044326595835,
"learning_rate": 4.96342125296261e-06,
"loss": 0.6173,
"step": 281
},
{
"epoch": 0.24727352441497233,
"grad_norm": 0.34476437566601653,
"learning_rate": 4.963016953617794e-06,
"loss": 0.6172,
"step": 282
},
{
"epoch": 0.24815038088452895,
"grad_norm": 0.2611205789369605,
"learning_rate": 4.962610448854816e-06,
"loss": 0.6246,
"step": 283
},
{
"epoch": 0.2490272373540856,
"grad_norm": 0.3294938430549001,
"learning_rate": 4.962201739037665e-06,
"loss": 0.632,
"step": 284
},
{
"epoch": 0.24990409382364223,
"grad_norm": 0.2716869569081184,
"learning_rate": 4.961790824532306e-06,
"loss": 0.6285,
"step": 285
},
{
"epoch": 0.25078095029319886,
"grad_norm": 0.33415021484488,
"learning_rate": 4.961377705706677e-06,
"loss": 0.6295,
"step": 286
},
{
"epoch": 0.2516578067627555,
"grad_norm": 0.3077857421614378,
"learning_rate": 4.960962382930691e-06,
"loss": 0.6273,
"step": 287
},
{
"epoch": 0.25253466323231216,
"grad_norm": 0.3027918805177667,
"learning_rate": 4.960544856576232e-06,
"loss": 0.629,
"step": 288
},
{
"epoch": 0.2534115197018688,
"grad_norm": 0.2916258020649895,
"learning_rate": 4.960125127017159e-06,
"loss": 0.6427,
"step": 289
},
{
"epoch": 0.2542883761714254,
"grad_norm": 0.3152484231550671,
"learning_rate": 4.959703194629304e-06,
"loss": 0.6348,
"step": 290
},
{
"epoch": 0.25516523264098206,
"grad_norm": 0.32915709407999866,
"learning_rate": 4.959279059790471e-06,
"loss": 0.632,
"step": 291
},
{
"epoch": 0.2560420891105387,
"grad_norm": 0.2817567268029023,
"learning_rate": 4.958852722880435e-06,
"loss": 0.6112,
"step": 292
},
{
"epoch": 0.25691894558009537,
"grad_norm": 0.3538236182060425,
"learning_rate": 4.958424184280946e-06,
"loss": 0.6241,
"step": 293
},
{
"epoch": 0.257795802049652,
"grad_norm": 0.2864183700965389,
"learning_rate": 4.957993444375719e-06,
"loss": 0.6277,
"step": 294
},
{
"epoch": 0.2586726585192086,
"grad_norm": 0.33515303575483923,
"learning_rate": 4.95756050355045e-06,
"loss": 0.6277,
"step": 295
},
{
"epoch": 0.2595495149887653,
"grad_norm": 0.31975746198582533,
"learning_rate": 4.957125362192794e-06,
"loss": 0.6114,
"step": 296
},
{
"epoch": 0.2604263714583219,
"grad_norm": 0.34329553758734277,
"learning_rate": 4.956688020692386e-06,
"loss": 0.6457,
"step": 297
},
{
"epoch": 0.2613032279278786,
"grad_norm": 0.3122307785419701,
"learning_rate": 4.956248479440827e-06,
"loss": 0.6272,
"step": 298
},
{
"epoch": 0.2621800843974352,
"grad_norm": 0.3126439049869492,
"learning_rate": 4.955806738831687e-06,
"loss": 0.634,
"step": 299
},
{
"epoch": 0.26305694086699183,
"grad_norm": 0.30725526373905826,
"learning_rate": 4.955362799260507e-06,
"loss": 0.6269,
"step": 300
},
{
"epoch": 0.2639337973365485,
"grad_norm": 0.2952615284346605,
"learning_rate": 4.954916661124797e-06,
"loss": 0.6129,
"step": 301
},
{
"epoch": 0.26481065380610513,
"grad_norm": 0.3284069744839045,
"learning_rate": 4.954468324824035e-06,
"loss": 0.613,
"step": 302
},
{
"epoch": 0.26568751027566173,
"grad_norm": 0.34051928196991404,
"learning_rate": 4.954017790759666e-06,
"loss": 0.6192,
"step": 303
},
{
"epoch": 0.2665643667452184,
"grad_norm": 0.30608255552211977,
"learning_rate": 4.953565059335104e-06,
"loss": 0.6244,
"step": 304
},
{
"epoch": 0.26744122321477504,
"grad_norm": 0.31501722301988566,
"learning_rate": 4.953110130955733e-06,
"loss": 0.6236,
"step": 305
},
{
"epoch": 0.2683180796843317,
"grad_norm": 0.2978345978834651,
"learning_rate": 4.9526530060289e-06,
"loss": 0.6254,
"step": 306
},
{
"epoch": 0.2691949361538883,
"grad_norm": 0.2935986604058687,
"learning_rate": 4.952193684963922e-06,
"loss": 0.6113,
"step": 307
},
{
"epoch": 0.27007179262344494,
"grad_norm": 0.294670736028252,
"learning_rate": 4.95173216817208e-06,
"loss": 0.6335,
"step": 308
},
{
"epoch": 0.2709486490930016,
"grad_norm": 0.2746280487759909,
"learning_rate": 4.951268456066623e-06,
"loss": 0.6211,
"step": 309
},
{
"epoch": 0.27182550556255825,
"grad_norm": 0.2823209312944346,
"learning_rate": 4.950802549062764e-06,
"loss": 0.621,
"step": 310
},
{
"epoch": 0.27270236203211484,
"grad_norm": 0.2811005060766513,
"learning_rate": 4.950334447577685e-06,
"loss": 0.6291,
"step": 311
},
{
"epoch": 0.2735792185016715,
"grad_norm": 0.31377780747479117,
"learning_rate": 4.9498641520305264e-06,
"loss": 0.6308,
"step": 312
},
{
"epoch": 0.27445607497122815,
"grad_norm": 0.263859895152384,
"learning_rate": 4.949391662842401e-06,
"loss": 0.6238,
"step": 313
},
{
"epoch": 0.2753329314407848,
"grad_norm": 0.3124591272767995,
"learning_rate": 4.948916980436379e-06,
"loss": 0.6254,
"step": 314
},
{
"epoch": 0.27620978791034145,
"grad_norm": 0.2762091249470148,
"learning_rate": 4.948440105237499e-06,
"loss": 0.6297,
"step": 315
},
{
"epoch": 0.27708664437989805,
"grad_norm": 0.30510467983773004,
"learning_rate": 4.947961037672761e-06,
"loss": 0.6301,
"step": 316
},
{
"epoch": 0.2779635008494547,
"grad_norm": 0.2894218681866538,
"learning_rate": 4.947479778171127e-06,
"loss": 0.6215,
"step": 317
},
{
"epoch": 0.27884035731901136,
"grad_norm": 0.278604444379188,
"learning_rate": 4.946996327163526e-06,
"loss": 0.6193,
"step": 318
},
{
"epoch": 0.279717213788568,
"grad_norm": 0.29226196825962947,
"learning_rate": 4.946510685082844e-06,
"loss": 0.6205,
"step": 319
},
{
"epoch": 0.2805940702581246,
"grad_norm": 0.2956824922950759,
"learning_rate": 4.946022852363932e-06,
"loss": 0.6238,
"step": 320
},
{
"epoch": 0.28147092672768126,
"grad_norm": 0.28796938907697983,
"learning_rate": 4.945532829443604e-06,
"loss": 0.6176,
"step": 321
},
{
"epoch": 0.2823477831972379,
"grad_norm": 0.2688847498978228,
"learning_rate": 4.945040616760629e-06,
"loss": 0.6178,
"step": 322
},
{
"epoch": 0.28322463966679456,
"grad_norm": 0.3167327299209847,
"learning_rate": 4.944546214755744e-06,
"loss": 0.6315,
"step": 323
},
{
"epoch": 0.28410149613635116,
"grad_norm": 0.28346482132020456,
"learning_rate": 4.9440496238716415e-06,
"loss": 0.6281,
"step": 324
},
{
"epoch": 0.2849783526059078,
"grad_norm": 0.2862108698161924,
"learning_rate": 4.943550844552978e-06,
"loss": 0.6445,
"step": 325
},
{
"epoch": 0.28585520907546447,
"grad_norm": 0.3168994194030117,
"learning_rate": 4.943049877246363e-06,
"loss": 0.6336,
"step": 326
},
{
"epoch": 0.2867320655450211,
"grad_norm": 0.3098419113094991,
"learning_rate": 4.942546722400373e-06,
"loss": 0.6194,
"step": 327
},
{
"epoch": 0.2876089220145777,
"grad_norm": 0.3076330226750193,
"learning_rate": 4.942041380465539e-06,
"loss": 0.6332,
"step": 328
},
{
"epoch": 0.28848577848413437,
"grad_norm": 0.3073675940253473,
"learning_rate": 4.941533851894349e-06,
"loss": 0.6329,
"step": 329
},
{
"epoch": 0.289362634953691,
"grad_norm": 0.27407015238515836,
"learning_rate": 4.9410241371412525e-06,
"loss": 0.6292,
"step": 330
},
{
"epoch": 0.2902394914232477,
"grad_norm": 0.3233677059379673,
"learning_rate": 4.9405122366626545e-06,
"loss": 0.6407,
"step": 331
},
{
"epoch": 0.2911163478928043,
"grad_norm": 0.3056326849325438,
"learning_rate": 4.939998150916917e-06,
"loss": 0.6314,
"step": 332
},
{
"epoch": 0.2919932043623609,
"grad_norm": 0.3140138519054107,
"learning_rate": 4.93948188036436e-06,
"loss": 0.6583,
"step": 333
},
{
"epoch": 0.2928700608319176,
"grad_norm": 0.2967689552064628,
"learning_rate": 4.938963425467258e-06,
"loss": 0.6349,
"step": 334
},
{
"epoch": 0.29374691730147423,
"grad_norm": 0.35320572702474673,
"learning_rate": 4.938442786689843e-06,
"loss": 0.6248,
"step": 335
},
{
"epoch": 0.29462377377103083,
"grad_norm": 0.2958836632865014,
"learning_rate": 4.9379199644983025e-06,
"loss": 0.6255,
"step": 336
},
{
"epoch": 0.2955006302405875,
"grad_norm": 0.3054952399371344,
"learning_rate": 4.937394959360777e-06,
"loss": 0.6119,
"step": 337
},
{
"epoch": 0.29637748671014413,
"grad_norm": 0.34308383177638463,
"learning_rate": 4.9368677717473645e-06,
"loss": 0.6468,
"step": 338
},
{
"epoch": 0.2972543431797008,
"grad_norm": 0.2648620374237178,
"learning_rate": 4.936338402130115e-06,
"loss": 0.6203,
"step": 339
},
{
"epoch": 0.29813119964925744,
"grad_norm": 0.2976099930186866,
"learning_rate": 4.935806850983034e-06,
"loss": 0.6348,
"step": 340
},
{
"epoch": 0.29900805611881404,
"grad_norm": 0.285144357181017,
"learning_rate": 4.935273118782078e-06,
"loss": 0.6115,
"step": 341
},
{
"epoch": 0.2998849125883707,
"grad_norm": 0.3079688238524965,
"learning_rate": 4.934737206005159e-06,
"loss": 0.6254,
"step": 342
},
{
"epoch": 0.30076176905792734,
"grad_norm": 0.27719094781494596,
"learning_rate": 4.93419911313214e-06,
"loss": 0.6386,
"step": 343
},
{
"epoch": 0.301638625527484,
"grad_norm": 0.29796636665366355,
"learning_rate": 4.933658840644837e-06,
"loss": 0.6268,
"step": 344
},
{
"epoch": 0.3025154819970406,
"grad_norm": 0.27509893042636935,
"learning_rate": 4.933116389027017e-06,
"loss": 0.621,
"step": 345
},
{
"epoch": 0.30339233846659724,
"grad_norm": 0.31224342373584874,
"learning_rate": 4.932571758764398e-06,
"loss": 0.6312,
"step": 346
},
{
"epoch": 0.3042691949361539,
"grad_norm": 0.2689144896057607,
"learning_rate": 4.93202495034465e-06,
"loss": 0.6115,
"step": 347
},
{
"epoch": 0.30514605140571055,
"grad_norm": 0.2558266510993566,
"learning_rate": 4.931475964257391e-06,
"loss": 0.6245,
"step": 348
},
{
"epoch": 0.30602290787526715,
"grad_norm": 0.25500762407211314,
"learning_rate": 4.930924800994192e-06,
"loss": 0.6091,
"step": 349
},
{
"epoch": 0.3068997643448238,
"grad_norm": 0.2717131638453367,
"learning_rate": 4.9303714610485705e-06,
"loss": 0.6281,
"step": 350
},
{
"epoch": 0.30777662081438045,
"grad_norm": 0.2729400616989181,
"learning_rate": 4.929815944915997e-06,
"loss": 0.6083,
"step": 351
},
{
"epoch": 0.3086534772839371,
"grad_norm": 0.26000631857019024,
"learning_rate": 4.929258253093885e-06,
"loss": 0.6198,
"step": 352
},
{
"epoch": 0.3095303337534937,
"grad_norm": 0.2740884453189882,
"learning_rate": 4.9286983860816e-06,
"loss": 0.6338,
"step": 353
},
{
"epoch": 0.31040719022305036,
"grad_norm": 0.27150990388252366,
"learning_rate": 4.928136344380457e-06,
"loss": 0.6162,
"step": 354
},
{
"epoch": 0.311284046692607,
"grad_norm": 0.26286571771385,
"learning_rate": 4.9275721284937115e-06,
"loss": 0.629,
"step": 355
},
{
"epoch": 0.31216090316216366,
"grad_norm": 0.27510252961865267,
"learning_rate": 4.9270057389265734e-06,
"loss": 0.633,
"step": 356
},
{
"epoch": 0.31303775963172026,
"grad_norm": 0.2825214790660817,
"learning_rate": 4.926437176186193e-06,
"loss": 0.6263,
"step": 357
},
{
"epoch": 0.3139146161012769,
"grad_norm": 0.29292375908331497,
"learning_rate": 4.92586644078167e-06,
"loss": 0.6313,
"step": 358
},
{
"epoch": 0.31479147257083356,
"grad_norm": 0.2760563004495057,
"learning_rate": 4.925293533224049e-06,
"loss": 0.6174,
"step": 359
},
{
"epoch": 0.3156683290403902,
"grad_norm": 0.29078508943452525,
"learning_rate": 4.924718454026318e-06,
"loss": 0.6156,
"step": 360
},
{
"epoch": 0.3165451855099468,
"grad_norm": 0.2878769173523044,
"learning_rate": 4.924141203703412e-06,
"loss": 0.6047,
"step": 361
},
{
"epoch": 0.31742204197950347,
"grad_norm": 0.27485843884417593,
"learning_rate": 4.923561782772206e-06,
"loss": 0.6293,
"step": 362
},
{
"epoch": 0.3182988984490601,
"grad_norm": 0.2865164028316351,
"learning_rate": 4.922980191751524e-06,
"loss": 0.6269,
"step": 363
},
{
"epoch": 0.31917575491861677,
"grad_norm": 0.27991173694279825,
"learning_rate": 4.922396431162129e-06,
"loss": 0.6143,
"step": 364
},
{
"epoch": 0.3200526113881734,
"grad_norm": 0.279639353480309,
"learning_rate": 4.921810501526728e-06,
"loss": 0.635,
"step": 365
},
{
"epoch": 0.32092946785773,
"grad_norm": 0.2830142803081013,
"learning_rate": 4.921222403369971e-06,
"loss": 0.6157,
"step": 366
},
{
"epoch": 0.3218063243272867,
"grad_norm": 0.2684155306717856,
"learning_rate": 4.920632137218447e-06,
"loss": 0.6294,
"step": 367
},
{
"epoch": 0.3226831807968433,
"grad_norm": 0.2983455576981931,
"learning_rate": 4.920039703600691e-06,
"loss": 0.624,
"step": 368
},
{
"epoch": 0.3235600372664,
"grad_norm": 0.2948947231333358,
"learning_rate": 4.9194451030471735e-06,
"loss": 0.6102,
"step": 369
},
{
"epoch": 0.3244368937359566,
"grad_norm": 0.2826890911442374,
"learning_rate": 4.918848336090309e-06,
"loss": 0.6236,
"step": 370
},
{
"epoch": 0.32531375020551323,
"grad_norm": 0.32269493597939386,
"learning_rate": 4.91824940326445e-06,
"loss": 0.6139,
"step": 371
},
{
"epoch": 0.3261906066750699,
"grad_norm": 0.2734983777513044,
"learning_rate": 4.91764830510589e-06,
"loss": 0.6166,
"step": 372
},
{
"epoch": 0.32706746314462654,
"grad_norm": 0.36983262498880637,
"learning_rate": 4.917045042152858e-06,
"loss": 0.6186,
"step": 373
},
{
"epoch": 0.32794431961418313,
"grad_norm": 0.2751996219950251,
"learning_rate": 4.916439614945527e-06,
"loss": 0.6412,
"step": 374
},
{
"epoch": 0.3288211760837398,
"grad_norm": 0.319865198714037,
"learning_rate": 4.915832024026002e-06,
"loss": 0.627,
"step": 375
},
{
"epoch": 0.32969803255329644,
"grad_norm": 0.29823421688781576,
"learning_rate": 4.915222269938328e-06,
"loss": 0.6181,
"step": 376
},
{
"epoch": 0.3305748890228531,
"grad_norm": 0.27335542421500575,
"learning_rate": 4.914610353228488e-06,
"loss": 0.6202,
"step": 377
},
{
"epoch": 0.3314517454924097,
"grad_norm": 0.3824213724235341,
"learning_rate": 4.913996274444401e-06,
"loss": 0.608,
"step": 378
},
{
"epoch": 0.33232860196196634,
"grad_norm": 0.3269271239671324,
"learning_rate": 4.913380034135919e-06,
"loss": 0.6229,
"step": 379
},
{
"epoch": 0.333205458431523,
"grad_norm": 0.2832871290462529,
"learning_rate": 4.912761632854834e-06,
"loss": 0.618,
"step": 380
},
{
"epoch": 0.33408231490107965,
"grad_norm": 0.329936751234759,
"learning_rate": 4.912141071154869e-06,
"loss": 0.6231,
"step": 381
},
{
"epoch": 0.33495917137063624,
"grad_norm": 0.2752693680315103,
"learning_rate": 4.911518349591685e-06,
"loss": 0.6234,
"step": 382
},
{
"epoch": 0.3358360278401929,
"grad_norm": 0.3136704903953731,
"learning_rate": 4.9108934687228735e-06,
"loss": 0.6248,
"step": 383
},
{
"epoch": 0.33671288430974955,
"grad_norm": 0.2947450161853734,
"learning_rate": 4.910266429107962e-06,
"loss": 0.6291,
"step": 384
},
{
"epoch": 0.3375897407793062,
"grad_norm": 0.27963622109645897,
"learning_rate": 4.90963723130841e-06,
"loss": 0.6168,
"step": 385
},
{
"epoch": 0.3384665972488628,
"grad_norm": 0.2755048673546131,
"learning_rate": 4.90900587588761e-06,
"loss": 0.6022,
"step": 386
},
{
"epoch": 0.33934345371841945,
"grad_norm": 0.28857281828902753,
"learning_rate": 4.908372363410886e-06,
"loss": 0.6254,
"step": 387
},
{
"epoch": 0.3402203101879761,
"grad_norm": 0.28648556573019374,
"learning_rate": 4.907736694445492e-06,
"loss": 0.6175,
"step": 388
},
{
"epoch": 0.34109716665753276,
"grad_norm": 0.26925532018377424,
"learning_rate": 4.9070988695606156e-06,
"loss": 0.6176,
"step": 389
},
{
"epoch": 0.3419740231270894,
"grad_norm": 0.2832182299890066,
"learning_rate": 4.906458889327375e-06,
"loss": 0.6291,
"step": 390
},
{
"epoch": 0.342850879596646,
"grad_norm": 0.24545023229724808,
"learning_rate": 4.905816754318815e-06,
"loss": 0.621,
"step": 391
},
{
"epoch": 0.34372773606620266,
"grad_norm": 0.27071805276574584,
"learning_rate": 4.905172465109912e-06,
"loss": 0.6235,
"step": 392
},
{
"epoch": 0.3446045925357593,
"grad_norm": 0.2686211222363871,
"learning_rate": 4.904526022277572e-06,
"loss": 0.6259,
"step": 393
},
{
"epoch": 0.34548144900531597,
"grad_norm": 0.2788582786567745,
"learning_rate": 4.903877426400629e-06,
"loss": 0.6113,
"step": 394
},
{
"epoch": 0.34635830547487256,
"grad_norm": 0.2882303517807228,
"learning_rate": 4.903226678059842e-06,
"loss": 0.6325,
"step": 395
},
{
"epoch": 0.3472351619444292,
"grad_norm": 0.26417391198725343,
"learning_rate": 4.902573777837902e-06,
"loss": 0.6171,
"step": 396
},
{
"epoch": 0.34811201841398587,
"grad_norm": 0.27931172516771346,
"learning_rate": 4.901918726319424e-06,
"loss": 0.6041,
"step": 397
},
{
"epoch": 0.3489888748835425,
"grad_norm": 0.24713049818043734,
"learning_rate": 4.901261524090949e-06,
"loss": 0.6099,
"step": 398
},
{
"epoch": 0.3498657313530991,
"grad_norm": 0.29086241382146505,
"learning_rate": 4.900602171740946e-06,
"loss": 0.6258,
"step": 399
},
{
"epoch": 0.35074258782265577,
"grad_norm": 0.26291418203363,
"learning_rate": 4.899940669859807e-06,
"loss": 0.6117,
"step": 400
},
{
"epoch": 0.3516194442922124,
"grad_norm": 0.3216617316096804,
"learning_rate": 4.89927701903985e-06,
"loss": 0.6187,
"step": 401
},
{
"epoch": 0.3524963007617691,
"grad_norm": 0.27295463776878537,
"learning_rate": 4.898611219875316e-06,
"loss": 0.6132,
"step": 402
},
{
"epoch": 0.3533731572313257,
"grad_norm": 0.2853334578601736,
"learning_rate": 4.897943272962372e-06,
"loss": 0.6148,
"step": 403
},
{
"epoch": 0.3542500137008823,
"grad_norm": 0.31932832747253076,
"learning_rate": 4.897273178899105e-06,
"loss": 0.6187,
"step": 404
},
{
"epoch": 0.355126870170439,
"grad_norm": 0.28031643219296354,
"learning_rate": 4.896600938285526e-06,
"loss": 0.6236,
"step": 405
},
{
"epoch": 0.35600372663999563,
"grad_norm": 0.26831626886851945,
"learning_rate": 4.89592655172357e-06,
"loss": 0.6102,
"step": 406
},
{
"epoch": 0.35688058310955223,
"grad_norm": 0.2951228212133584,
"learning_rate": 4.895250019817089e-06,
"loss": 0.6164,
"step": 407
},
{
"epoch": 0.3577574395791089,
"grad_norm": 0.27330142007513136,
"learning_rate": 4.894571343171862e-06,
"loss": 0.6023,
"step": 408
},
{
"epoch": 0.35863429604866554,
"grad_norm": 0.3204620119402923,
"learning_rate": 4.893890522395582e-06,
"loss": 0.62,
"step": 409
},
{
"epoch": 0.3595111525182222,
"grad_norm": 0.261478566125417,
"learning_rate": 4.893207558097867e-06,
"loss": 0.6294,
"step": 410
},
{
"epoch": 0.36038800898777884,
"grad_norm": 0.250895473885103,
"learning_rate": 4.892522450890251e-06,
"loss": 0.6152,
"step": 411
},
{
"epoch": 0.36126486545733544,
"grad_norm": 0.2634865561040139,
"learning_rate": 4.89183520138619e-06,
"loss": 0.6157,
"step": 412
},
{
"epoch": 0.3621417219268921,
"grad_norm": 0.26459491662331874,
"learning_rate": 4.891145810201054e-06,
"loss": 0.609,
"step": 413
},
{
"epoch": 0.36301857839644874,
"grad_norm": 0.24301745655990745,
"learning_rate": 4.8904542779521346e-06,
"loss": 0.6082,
"step": 414
},
{
"epoch": 0.3638954348660054,
"grad_norm": 0.2692643109083729,
"learning_rate": 4.8897606052586384e-06,
"loss": 0.6226,
"step": 415
},
{
"epoch": 0.364772291335562,
"grad_norm": 0.24024671108707563,
"learning_rate": 4.889064792741689e-06,
"loss": 0.6153,
"step": 416
},
{
"epoch": 0.36564914780511865,
"grad_norm": 0.273288282597359,
"learning_rate": 4.888366841024327e-06,
"loss": 0.6334,
"step": 417
},
{
"epoch": 0.3665260042746753,
"grad_norm": 0.2713735341001686,
"learning_rate": 4.887666750731507e-06,
"loss": 0.6204,
"step": 418
},
{
"epoch": 0.36740286074423195,
"grad_norm": 0.2749014394381958,
"learning_rate": 4.8869645224901e-06,
"loss": 0.6017,
"step": 419
},
{
"epoch": 0.36827971721378855,
"grad_norm": 0.27621114898765087,
"learning_rate": 4.8862601569288885e-06,
"loss": 0.6193,
"step": 420
},
{
"epoch": 0.3691565736833452,
"grad_norm": 0.25931507650511326,
"learning_rate": 4.885553654678573e-06,
"loss": 0.6233,
"step": 421
},
{
"epoch": 0.37003343015290185,
"grad_norm": 0.28686169175433923,
"learning_rate": 4.884845016371763e-06,
"loss": 0.6197,
"step": 422
},
{
"epoch": 0.3709102866224585,
"grad_norm": 0.27025382919889446,
"learning_rate": 4.884134242642985e-06,
"loss": 0.6033,
"step": 423
},
{
"epoch": 0.3717871430920151,
"grad_norm": 0.275669477293775,
"learning_rate": 4.883421334128674e-06,
"loss": 0.6172,
"step": 424
},
{
"epoch": 0.37266399956157176,
"grad_norm": 0.26014021950194516,
"learning_rate": 4.8827062914671775e-06,
"loss": 0.6207,
"step": 425
},
{
"epoch": 0.3735408560311284,
"grad_norm": 0.2986829920255015,
"learning_rate": 4.881989115298755e-06,
"loss": 0.6034,
"step": 426
},
{
"epoch": 0.37441771250068506,
"grad_norm": 0.28151692244357057,
"learning_rate": 4.881269806265575e-06,
"loss": 0.6133,
"step": 427
},
{
"epoch": 0.37529456897024166,
"grad_norm": 0.2932206682237993,
"learning_rate": 4.8805483650117154e-06,
"loss": 0.6132,
"step": 428
},
{
"epoch": 0.3761714254397983,
"grad_norm": 0.3164265338412961,
"learning_rate": 4.879824792183166e-06,
"loss": 0.6077,
"step": 429
},
{
"epoch": 0.37704828190935497,
"grad_norm": 0.3636164115457003,
"learning_rate": 4.879099088427824e-06,
"loss": 0.6179,
"step": 430
},
{
"epoch": 0.3779251383789116,
"grad_norm": 0.2891875334309757,
"learning_rate": 4.878371254395492e-06,
"loss": 0.6197,
"step": 431
},
{
"epoch": 0.3788019948484682,
"grad_norm": 0.3816104662619605,
"learning_rate": 4.8776412907378845e-06,
"loss": 0.6197,
"step": 432
},
{
"epoch": 0.37967885131802487,
"grad_norm": 0.29131497715708005,
"learning_rate": 4.876909198108619e-06,
"loss": 0.6159,
"step": 433
},
{
"epoch": 0.3805557077875815,
"grad_norm": 0.3138520265609416,
"learning_rate": 4.876174977163222e-06,
"loss": 0.6139,
"step": 434
},
{
"epoch": 0.3814325642571382,
"grad_norm": 0.28035852092093033,
"learning_rate": 4.875438628559124e-06,
"loss": 0.6183,
"step": 435
},
{
"epoch": 0.3823094207266948,
"grad_norm": 0.3120106817898386,
"learning_rate": 4.874700152955661e-06,
"loss": 0.6052,
"step": 436
},
{
"epoch": 0.3831862771962514,
"grad_norm": 0.29139666929908226,
"learning_rate": 4.873959551014075e-06,
"loss": 0.6058,
"step": 437
},
{
"epoch": 0.3840631336658081,
"grad_norm": 0.31305383154436955,
"learning_rate": 4.873216823397511e-06,
"loss": 0.6094,
"step": 438
},
{
"epoch": 0.38493999013536473,
"grad_norm": 0.3052879988977325,
"learning_rate": 4.872471970771015e-06,
"loss": 0.6063,
"step": 439
},
{
"epoch": 0.3858168466049214,
"grad_norm": 0.2965934350138861,
"learning_rate": 4.871724993801541e-06,
"loss": 0.6054,
"step": 440
},
{
"epoch": 0.386693703074478,
"grad_norm": 0.26339362714008424,
"learning_rate": 4.870975893157941e-06,
"loss": 0.6152,
"step": 441
},
{
"epoch": 0.38757055954403463,
"grad_norm": 0.27556079714679943,
"learning_rate": 4.870224669510968e-06,
"loss": 0.6158,
"step": 442
},
{
"epoch": 0.3884474160135913,
"grad_norm": 0.29125701036171053,
"learning_rate": 4.86947132353328e-06,
"loss": 0.6202,
"step": 443
},
{
"epoch": 0.38932427248314794,
"grad_norm": 0.2966406156980298,
"learning_rate": 4.868715855899432e-06,
"loss": 0.6265,
"step": 444
},
{
"epoch": 0.39020112895270453,
"grad_norm": 0.27733217518457043,
"learning_rate": 4.867958267285879e-06,
"loss": 0.6068,
"step": 445
},
{
"epoch": 0.3910779854222612,
"grad_norm": 0.2919788828093281,
"learning_rate": 4.8671985583709765e-06,
"loss": 0.6208,
"step": 446
},
{
"epoch": 0.39195484189181784,
"grad_norm": 0.29327731039840055,
"learning_rate": 4.866436729834979e-06,
"loss": 0.6175,
"step": 447
},
{
"epoch": 0.3928316983613745,
"grad_norm": 0.2568832744529454,
"learning_rate": 4.865672782360037e-06,
"loss": 0.6177,
"step": 448
},
{
"epoch": 0.3937085548309311,
"grad_norm": 0.283654204460893,
"learning_rate": 4.8649067166301985e-06,
"loss": 0.6203,
"step": 449
},
{
"epoch": 0.39458541130048774,
"grad_norm": 0.26828805221375346,
"learning_rate": 4.864138533331411e-06,
"loss": 0.6118,
"step": 450
},
{
"epoch": 0.3954622677700444,
"grad_norm": 0.2597158618871073,
"learning_rate": 4.863368233151514e-06,
"loss": 0.6169,
"step": 451
},
{
"epoch": 0.39633912423960105,
"grad_norm": 0.28436035142498156,
"learning_rate": 4.862595816780246e-06,
"loss": 0.632,
"step": 452
},
{
"epoch": 0.39721598070915765,
"grad_norm": 0.2652505819829089,
"learning_rate": 4.861821284909238e-06,
"loss": 0.6289,
"step": 453
},
{
"epoch": 0.3980928371787143,
"grad_norm": 0.29252031992594624,
"learning_rate": 4.861044638232016e-06,
"loss": 0.6328,
"step": 454
},
{
"epoch": 0.39896969364827095,
"grad_norm": 0.2994469365008051,
"learning_rate": 4.860265877444001e-06,
"loss": 0.617,
"step": 455
},
{
"epoch": 0.3998465501178276,
"grad_norm": 0.2776900829822044,
"learning_rate": 4.8594850032425036e-06,
"loss": 0.608,
"step": 456
},
{
"epoch": 0.4007234065873842,
"grad_norm": 0.2753322141436327,
"learning_rate": 4.858702016326731e-06,
"loss": 0.607,
"step": 457
},
{
"epoch": 0.40160026305694085,
"grad_norm": 0.2738219915396828,
"learning_rate": 4.857916917397779e-06,
"loss": 0.6043,
"step": 458
},
{
"epoch": 0.4024771195264975,
"grad_norm": 0.27192665887665013,
"learning_rate": 4.857129707158637e-06,
"loss": 0.6376,
"step": 459
},
{
"epoch": 0.40335397599605416,
"grad_norm": 0.27689826150792163,
"learning_rate": 4.8563403863141825e-06,
"loss": 0.6172,
"step": 460
},
{
"epoch": 0.4042308324656108,
"grad_norm": 0.311644665297658,
"learning_rate": 4.855548955571183e-06,
"loss": 0.6106,
"step": 461
},
{
"epoch": 0.4051076889351674,
"grad_norm": 0.2912453467934098,
"learning_rate": 4.854755415638298e-06,
"loss": 0.6129,
"step": 462
},
{
"epoch": 0.40598454540472406,
"grad_norm": 0.302939167109194,
"learning_rate": 4.853959767226072e-06,
"loss": 0.6301,
"step": 463
},
{
"epoch": 0.4068614018742807,
"grad_norm": 0.261297831693092,
"learning_rate": 4.85316201104694e-06,
"loss": 0.6136,
"step": 464
},
{
"epoch": 0.40773825834383737,
"grad_norm": 0.3154856081824323,
"learning_rate": 4.852362147815225e-06,
"loss": 0.6171,
"step": 465
},
{
"epoch": 0.40861511481339396,
"grad_norm": 0.29411022742744497,
"learning_rate": 4.8515601782471325e-06,
"loss": 0.6085,
"step": 466
},
{
"epoch": 0.4094919712829506,
"grad_norm": 0.3027595832299397,
"learning_rate": 4.8507561030607576e-06,
"loss": 0.6151,
"step": 467
},
{
"epoch": 0.41036882775250727,
"grad_norm": 0.3003092813187261,
"learning_rate": 4.84994992297608e-06,
"loss": 0.6071,
"step": 468
},
{
"epoch": 0.4112456842220639,
"grad_norm": 0.27374249219050456,
"learning_rate": 4.849141638714965e-06,
"loss": 0.6166,
"step": 469
},
{
"epoch": 0.4121225406916205,
"grad_norm": 0.3064667255626573,
"learning_rate": 4.84833125100116e-06,
"loss": 0.6024,
"step": 470
},
{
"epoch": 0.4129993971611772,
"grad_norm": 0.28188617697439766,
"learning_rate": 4.847518760560297e-06,
"loss": 0.6134,
"step": 471
},
{
"epoch": 0.4138762536307338,
"grad_norm": 0.27693005272362925,
"learning_rate": 4.846704168119892e-06,
"loss": 0.5984,
"step": 472
},
{
"epoch": 0.4147531101002905,
"grad_norm": 0.3011450154809493,
"learning_rate": 4.84588747440934e-06,
"loss": 0.5932,
"step": 473
},
{
"epoch": 0.4156299665698471,
"grad_norm": 0.25715138595393167,
"learning_rate": 4.845068680159921e-06,
"loss": 0.6101,
"step": 474
},
{
"epoch": 0.41650682303940373,
"grad_norm": 0.2963493163477849,
"learning_rate": 4.844247786104794e-06,
"loss": 0.6081,
"step": 475
},
{
"epoch": 0.4173836795089604,
"grad_norm": 0.29399759702492007,
"learning_rate": 4.8434247929789975e-06,
"loss": 0.6046,
"step": 476
},
{
"epoch": 0.41826053597851703,
"grad_norm": 0.3126535237916745,
"learning_rate": 4.842599701519451e-06,
"loss": 0.6304,
"step": 477
},
{
"epoch": 0.41913739244807363,
"grad_norm": 0.29299694878032745,
"learning_rate": 4.841772512464953e-06,
"loss": 0.6168,
"step": 478
},
{
"epoch": 0.4200142489176303,
"grad_norm": 0.289486342187316,
"learning_rate": 4.840943226556178e-06,
"loss": 0.6031,
"step": 479
},
{
"epoch": 0.42089110538718694,
"grad_norm": 0.30359254383613277,
"learning_rate": 4.840111844535682e-06,
"loss": 0.5994,
"step": 480
},
{
"epoch": 0.4217679618567436,
"grad_norm": 0.2641793447534652,
"learning_rate": 4.839278367147894e-06,
"loss": 0.6036,
"step": 481
},
{
"epoch": 0.4226448183263002,
"grad_norm": 0.29968320834098117,
"learning_rate": 4.838442795139121e-06,
"loss": 0.6193,
"step": 482
},
{
"epoch": 0.42352167479585684,
"grad_norm": 0.30614554761610074,
"learning_rate": 4.837605129257546e-06,
"loss": 0.6115,
"step": 483
},
{
"epoch": 0.4243985312654135,
"grad_norm": 0.29316129861054724,
"learning_rate": 4.836765370253223e-06,
"loss": 0.6039,
"step": 484
},
{
"epoch": 0.42527538773497015,
"grad_norm": 0.35388210389950725,
"learning_rate": 4.835923518878088e-06,
"loss": 0.6089,
"step": 485
},
{
"epoch": 0.4261522442045268,
"grad_norm": 0.27541931694811506,
"learning_rate": 4.835079575885944e-06,
"loss": 0.6129,
"step": 486
},
{
"epoch": 0.4270291006740834,
"grad_norm": 0.3408256598988536,
"learning_rate": 4.834233542032468e-06,
"loss": 0.6165,
"step": 487
},
{
"epoch": 0.42790595714364005,
"grad_norm": 0.30259946435062773,
"learning_rate": 4.83338541807521e-06,
"loss": 0.6111,
"step": 488
},
{
"epoch": 0.4287828136131967,
"grad_norm": 0.2871132966743198,
"learning_rate": 4.832535204773593e-06,
"loss": 0.6273,
"step": 489
},
{
"epoch": 0.42965967008275335,
"grad_norm": 0.3457337315321895,
"learning_rate": 4.8316829028889076e-06,
"loss": 0.6005,
"step": 490
},
{
"epoch": 0.43053652655230995,
"grad_norm": 0.2668696078107318,
"learning_rate": 4.830828513184317e-06,
"loss": 0.6122,
"step": 491
},
{
"epoch": 0.4314133830218666,
"grad_norm": 0.321068645111551,
"learning_rate": 4.829972036424854e-06,
"loss": 0.6058,
"step": 492
},
{
"epoch": 0.43229023949142326,
"grad_norm": 0.26125737492647644,
"learning_rate": 4.829113473377417e-06,
"loss": 0.6143,
"step": 493
},
{
"epoch": 0.4331670959609799,
"grad_norm": 0.32002755047063874,
"learning_rate": 4.828252824810777e-06,
"loss": 0.6061,
"step": 494
},
{
"epoch": 0.4340439524305365,
"grad_norm": 0.2863878470189295,
"learning_rate": 4.82739009149557e-06,
"loss": 0.5977,
"step": 495
},
{
"epoch": 0.43492080890009316,
"grad_norm": 0.31874371835878795,
"learning_rate": 4.826525274204297e-06,
"loss": 0.608,
"step": 496
},
{
"epoch": 0.4357976653696498,
"grad_norm": 0.2956391151217163,
"learning_rate": 4.825658373711328e-06,
"loss": 0.6107,
"step": 497
},
{
"epoch": 0.43667452183920646,
"grad_norm": 0.288406786632812,
"learning_rate": 4.824789390792899e-06,
"loss": 0.6094,
"step": 498
},
{
"epoch": 0.43755137830876306,
"grad_norm": 0.33737182032602686,
"learning_rate": 4.823918326227106e-06,
"loss": 0.5971,
"step": 499
},
{
"epoch": 0.4384282347783197,
"grad_norm": 0.25632117321609454,
"learning_rate": 4.823045180793914e-06,
"loss": 0.6044,
"step": 500
},
{
"epoch": 0.43930509124787637,
"grad_norm": 0.2978956835348055,
"learning_rate": 4.8221699552751465e-06,
"loss": 0.6009,
"step": 501
},
{
"epoch": 0.440181947717433,
"grad_norm": 0.30339339194561,
"learning_rate": 4.821292650454495e-06,
"loss": 0.6113,
"step": 502
},
{
"epoch": 0.4410588041869896,
"grad_norm": 0.3083549716587437,
"learning_rate": 4.8204132671175085e-06,
"loss": 0.6074,
"step": 503
},
{
"epoch": 0.44193566065654627,
"grad_norm": 0.291272682255802,
"learning_rate": 4.819531806051599e-06,
"loss": 0.606,
"step": 504
},
{
"epoch": 0.4428125171261029,
"grad_norm": 0.3183233272727026,
"learning_rate": 4.818648268046038e-06,
"loss": 0.6145,
"step": 505
},
{
"epoch": 0.4436893735956596,
"grad_norm": 0.27989457450916727,
"learning_rate": 4.817762653891957e-06,
"loss": 0.6095,
"step": 506
},
{
"epoch": 0.4445662300652162,
"grad_norm": 0.32106502207942483,
"learning_rate": 4.816874964382346e-06,
"loss": 0.6096,
"step": 507
},
{
"epoch": 0.4454430865347728,
"grad_norm": 0.2690675603747584,
"learning_rate": 4.815985200312057e-06,
"loss": 0.5986,
"step": 508
},
{
"epoch": 0.4463199430043295,
"grad_norm": 0.2818980909126885,
"learning_rate": 4.815093362477793e-06,
"loss": 0.6136,
"step": 509
},
{
"epoch": 0.44719679947388613,
"grad_norm": 0.29748447845455983,
"learning_rate": 4.8141994516781196e-06,
"loss": 0.6162,
"step": 510
},
{
"epoch": 0.4480736559434428,
"grad_norm": 0.3107094817046459,
"learning_rate": 4.813303468713456e-06,
"loss": 0.5939,
"step": 511
},
{
"epoch": 0.4489505124129994,
"grad_norm": 0.27493905192543294,
"learning_rate": 4.812405414386078e-06,
"loss": 0.6054,
"step": 512
},
{
"epoch": 0.44982736888255603,
"grad_norm": 0.28885594119974684,
"learning_rate": 4.811505289500113e-06,
"loss": 0.611,
"step": 513
},
{
"epoch": 0.4507042253521127,
"grad_norm": 0.2724458036095346,
"learning_rate": 4.810603094861548e-06,
"loss": 0.6296,
"step": 514
},
{
"epoch": 0.45158108182166934,
"grad_norm": 0.3171235548951884,
"learning_rate": 4.809698831278217e-06,
"loss": 0.6137,
"step": 515
},
{
"epoch": 0.45245793829122594,
"grad_norm": 0.2975607228468226,
"learning_rate": 4.808792499559812e-06,
"loss": 0.6081,
"step": 516
},
{
"epoch": 0.4533347947607826,
"grad_norm": 0.29553804453973653,
"learning_rate": 4.807884100517873e-06,
"loss": 0.6106,
"step": 517
},
{
"epoch": 0.45421165123033924,
"grad_norm": 0.29283068458115197,
"learning_rate": 4.8069736349657935e-06,
"loss": 0.6144,
"step": 518
},
{
"epoch": 0.4550885076998959,
"grad_norm": 0.3123674697628625,
"learning_rate": 4.806061103718816e-06,
"loss": 0.6024,
"step": 519
},
{
"epoch": 0.4559653641694525,
"grad_norm": 0.3185535504257689,
"learning_rate": 4.805146507594034e-06,
"loss": 0.6031,
"step": 520
},
{
"epoch": 0.45684222063900914,
"grad_norm": 0.32719458735857726,
"learning_rate": 4.804229847410388e-06,
"loss": 0.614,
"step": 521
},
{
"epoch": 0.4577190771085658,
"grad_norm": 0.2756686412179773,
"learning_rate": 4.803311123988668e-06,
"loss": 0.6143,
"step": 522
},
{
"epoch": 0.45859593357812245,
"grad_norm": 0.3193363571929515,
"learning_rate": 4.802390338151512e-06,
"loss": 0.5962,
"step": 523
},
{
"epoch": 0.45947279004767905,
"grad_norm": 0.27470129307670516,
"learning_rate": 4.801467490723402e-06,
"loss": 0.6118,
"step": 524
},
{
"epoch": 0.4603496465172357,
"grad_norm": 0.3268257836594815,
"learning_rate": 4.800542582530668e-06,
"loss": 0.6091,
"step": 525
},
{
"epoch": 0.46122650298679235,
"grad_norm": 0.2636715015821582,
"learning_rate": 4.799615614401488e-06,
"loss": 0.6113,
"step": 526
},
{
"epoch": 0.462103359456349,
"grad_norm": 0.3309929173426789,
"learning_rate": 4.79868658716588e-06,
"loss": 0.6063,
"step": 527
},
{
"epoch": 0.4629802159259056,
"grad_norm": 0.2705433155095911,
"learning_rate": 4.7977555016557054e-06,
"loss": 0.6115,
"step": 528
},
{
"epoch": 0.46385707239546226,
"grad_norm": 0.2986983107432822,
"learning_rate": 4.796822358704673e-06,
"loss": 0.624,
"step": 529
},
{
"epoch": 0.4647339288650189,
"grad_norm": 0.27153673858142124,
"learning_rate": 4.7958871591483305e-06,
"loss": 0.6144,
"step": 530
},
{
"epoch": 0.46561078533457556,
"grad_norm": 0.2774095045069063,
"learning_rate": 4.794949903824069e-06,
"loss": 0.6082,
"step": 531
},
{
"epoch": 0.4664876418041322,
"grad_norm": 0.28167525290961587,
"learning_rate": 4.794010593571118e-06,
"loss": 0.6106,
"step": 532
},
{
"epoch": 0.4673644982736888,
"grad_norm": 0.2626835693504621,
"learning_rate": 4.793069229230548e-06,
"loss": 0.6142,
"step": 533
},
{
"epoch": 0.46824135474324546,
"grad_norm": 0.27619948959341917,
"learning_rate": 4.792125811645271e-06,
"loss": 0.6073,
"step": 534
},
{
"epoch": 0.4691182112128021,
"grad_norm": 0.2913249262978291,
"learning_rate": 4.791180341660035e-06,
"loss": 0.6034,
"step": 535
},
{
"epoch": 0.46999506768235877,
"grad_norm": 0.2792318560656134,
"learning_rate": 4.790232820121426e-06,
"loss": 0.6002,
"step": 536
},
{
"epoch": 0.47087192415191537,
"grad_norm": 0.2690237732263836,
"learning_rate": 4.789283247877867e-06,
"loss": 0.6128,
"step": 537
},
{
"epoch": 0.471748780621472,
"grad_norm": 0.2875784864108413,
"learning_rate": 4.7883316257796195e-06,
"loss": 0.6125,
"step": 538
},
{
"epoch": 0.47262563709102867,
"grad_norm": 0.3494280106540881,
"learning_rate": 4.787377954678776e-06,
"loss": 0.6079,
"step": 539
},
{
"epoch": 0.4735024935605853,
"grad_norm": 0.27811345732659243,
"learning_rate": 4.786422235429269e-06,
"loss": 0.6118,
"step": 540
},
{
"epoch": 0.4743793500301419,
"grad_norm": 0.33921109846320074,
"learning_rate": 4.785464468886859e-06,
"loss": 0.6176,
"step": 541
},
{
"epoch": 0.4752562064996986,
"grad_norm": 0.29592545517880114,
"learning_rate": 4.784504655909146e-06,
"loss": 0.6131,
"step": 542
},
{
"epoch": 0.4761330629692552,
"grad_norm": 0.29373530511374163,
"learning_rate": 4.783542797355558e-06,
"loss": 0.6082,
"step": 543
},
{
"epoch": 0.4770099194388119,
"grad_norm": 0.2999691792256973,
"learning_rate": 4.782578894087357e-06,
"loss": 0.5981,
"step": 544
},
{
"epoch": 0.4778867759083685,
"grad_norm": 0.2694268894908227,
"learning_rate": 4.781612946967632e-06,
"loss": 0.6055,
"step": 545
},
{
"epoch": 0.47876363237792513,
"grad_norm": 0.2970836241532985,
"learning_rate": 4.780644956861307e-06,
"loss": 0.6002,
"step": 546
},
{
"epoch": 0.4796404888474818,
"grad_norm": 0.3413332201519291,
"learning_rate": 4.7796749246351335e-06,
"loss": 0.6103,
"step": 547
},
{
"epoch": 0.48051734531703844,
"grad_norm": 0.27732196553749033,
"learning_rate": 4.77870285115769e-06,
"loss": 0.5972,
"step": 548
},
{
"epoch": 0.48139420178659503,
"grad_norm": 0.32594912225980904,
"learning_rate": 4.777728737299387e-06,
"loss": 0.6275,
"step": 549
},
{
"epoch": 0.4822710582561517,
"grad_norm": 0.28158230943213153,
"learning_rate": 4.776752583932455e-06,
"loss": 0.6215,
"step": 550
},
{
"epoch": 0.48314791472570834,
"grad_norm": 0.3244722564822324,
"learning_rate": 4.775774391930956e-06,
"loss": 0.5947,
"step": 551
},
{
"epoch": 0.484024771195265,
"grad_norm": 0.26397208532030864,
"learning_rate": 4.774794162170777e-06,
"loss": 0.611,
"step": 552
},
{
"epoch": 0.4849016276648216,
"grad_norm": 0.2816890422555255,
"learning_rate": 4.773811895529629e-06,
"loss": 0.5942,
"step": 553
},
{
"epoch": 0.48577848413437824,
"grad_norm": 0.28224512879430635,
"learning_rate": 4.772827592887046e-06,
"loss": 0.5918,
"step": 554
},
{
"epoch": 0.4866553406039349,
"grad_norm": 0.2978578883597439,
"learning_rate": 4.771841255124385e-06,
"loss": 0.6031,
"step": 555
},
{
"epoch": 0.48753219707349155,
"grad_norm": 0.3212067488646109,
"learning_rate": 4.770852883124827e-06,
"loss": 0.6066,
"step": 556
},
{
"epoch": 0.4884090535430482,
"grad_norm": 0.3047898856904216,
"learning_rate": 4.769862477773374e-06,
"loss": 0.6097,
"step": 557
},
{
"epoch": 0.4892859100126048,
"grad_norm": 0.32816575436148626,
"learning_rate": 4.768870039956846e-06,
"loss": 0.6078,
"step": 558
},
{
"epoch": 0.49016276648216145,
"grad_norm": 0.30333447423661625,
"learning_rate": 4.767875570563887e-06,
"loss": 0.6103,
"step": 559
},
{
"epoch": 0.4910396229517181,
"grad_norm": 0.32463487013229164,
"learning_rate": 4.766879070484957e-06,
"loss": 0.5925,
"step": 560
},
{
"epoch": 0.49191647942127475,
"grad_norm": 0.27125555349656966,
"learning_rate": 4.765880540612336e-06,
"loss": 0.6095,
"step": 561
},
{
"epoch": 0.49279333589083135,
"grad_norm": 0.29571340419933284,
"learning_rate": 4.764879981840121e-06,
"loss": 0.6061,
"step": 562
},
{
"epoch": 0.493670192360388,
"grad_norm": 0.28779220439984465,
"learning_rate": 4.763877395064225e-06,
"loss": 0.6164,
"step": 563
},
{
"epoch": 0.49454704882994466,
"grad_norm": 0.3023002461106019,
"learning_rate": 4.762872781182378e-06,
"loss": 0.6099,
"step": 564
},
{
"epoch": 0.4954239052995013,
"grad_norm": 0.2852998688047179,
"learning_rate": 4.761866141094126e-06,
"loss": 0.6151,
"step": 565
},
{
"epoch": 0.4963007617690579,
"grad_norm": 0.27004415072990756,
"learning_rate": 4.7608574757008245e-06,
"loss": 0.6056,
"step": 566
},
{
"epoch": 0.49717761823861456,
"grad_norm": 0.26583697629837466,
"learning_rate": 4.759846785905649e-06,
"loss": 0.6073,
"step": 567
},
{
"epoch": 0.4980544747081712,
"grad_norm": 0.29963137609858226,
"learning_rate": 4.758834072613583e-06,
"loss": 0.6175,
"step": 568
},
{
"epoch": 0.49893133117772787,
"grad_norm": 0.2777428291092147,
"learning_rate": 4.757819336731424e-06,
"loss": 0.6084,
"step": 569
},
{
"epoch": 0.49980818764728446,
"grad_norm": 0.286537576055084,
"learning_rate": 4.756802579167781e-06,
"loss": 0.6122,
"step": 570
},
{
"epoch": 0.5006850441168411,
"grad_norm": 0.2900434750609322,
"learning_rate": 4.755783800833071e-06,
"loss": 0.61,
"step": 571
},
{
"epoch": 0.5015619005863977,
"grad_norm": 0.29602981997833644,
"learning_rate": 4.754763002639522e-06,
"loss": 0.5979,
"step": 572
},
{
"epoch": 0.5024387570559544,
"grad_norm": 0.2850500950921633,
"learning_rate": 4.75374018550117e-06,
"loss": 0.616,
"step": 573
},
{
"epoch": 0.503315613525511,
"grad_norm": 0.2747595431255721,
"learning_rate": 4.752715350333858e-06,
"loss": 0.6082,
"step": 574
},
{
"epoch": 0.5041924699950677,
"grad_norm": 0.30963433949041175,
"learning_rate": 4.75168849805524e-06,
"loss": 0.6062,
"step": 575
},
{
"epoch": 0.5050693264646243,
"grad_norm": 0.28817154630491854,
"learning_rate": 4.750659629584772e-06,
"loss": 0.615,
"step": 576
},
{
"epoch": 0.5059461829341809,
"grad_norm": 0.29777143797501865,
"learning_rate": 4.749628745843715e-06,
"loss": 0.6093,
"step": 577
},
{
"epoch": 0.5068230394037376,
"grad_norm": 0.2761328411528336,
"learning_rate": 4.748595847755137e-06,
"loss": 0.5949,
"step": 578
},
{
"epoch": 0.5076998958732942,
"grad_norm": 0.27941749417554973,
"learning_rate": 4.74756093624391e-06,
"loss": 0.6165,
"step": 579
},
{
"epoch": 0.5085767523428508,
"grad_norm": 0.28883681834919644,
"learning_rate": 4.746524012236706e-06,
"loss": 0.6012,
"step": 580
},
{
"epoch": 0.5094536088124075,
"grad_norm": 0.2712633209555587,
"learning_rate": 4.7454850766620005e-06,
"loss": 0.5898,
"step": 581
},
{
"epoch": 0.5103304652819641,
"grad_norm": 0.29386364789948854,
"learning_rate": 4.7444441304500714e-06,
"loss": 0.6057,
"step": 582
},
{
"epoch": 0.5112073217515208,
"grad_norm": 0.27998562308750735,
"learning_rate": 4.743401174532994e-06,
"loss": 0.597,
"step": 583
},
{
"epoch": 0.5120841782210774,
"grad_norm": 0.2944531079667381,
"learning_rate": 4.742356209844646e-06,
"loss": 0.5915,
"step": 584
},
{
"epoch": 0.512961034690634,
"grad_norm": 0.29506045387008756,
"learning_rate": 4.741309237320703e-06,
"loss": 0.6178,
"step": 585
},
{
"epoch": 0.5138378911601907,
"grad_norm": 0.299236621784075,
"learning_rate": 4.740260257898638e-06,
"loss": 0.6121,
"step": 586
},
{
"epoch": 0.5147147476297473,
"grad_norm": 0.303688650889379,
"learning_rate": 4.739209272517721e-06,
"loss": 0.5982,
"step": 587
},
{
"epoch": 0.515591604099304,
"grad_norm": 0.2925779066404172,
"learning_rate": 4.738156282119018e-06,
"loss": 0.5936,
"step": 588
},
{
"epoch": 0.5164684605688606,
"grad_norm": 0.3374725318718031,
"learning_rate": 4.73710128764539e-06,
"loss": 0.6001,
"step": 589
},
{
"epoch": 0.5173453170384172,
"grad_norm": 0.28811046561615106,
"learning_rate": 4.736044290041496e-06,
"loss": 0.61,
"step": 590
},
{
"epoch": 0.518222173507974,
"grad_norm": 0.32139851009391945,
"learning_rate": 4.7349852902537814e-06,
"loss": 0.5931,
"step": 591
},
{
"epoch": 0.5190990299775305,
"grad_norm": 0.27307295767087736,
"learning_rate": 4.733924289230493e-06,
"loss": 0.6035,
"step": 592
},
{
"epoch": 0.5199758864470871,
"grad_norm": 0.3098223534082736,
"learning_rate": 4.7328612879216615e-06,
"loss": 0.6082,
"step": 593
},
{
"epoch": 0.5208527429166439,
"grad_norm": 0.2808341207944162,
"learning_rate": 4.731796287279115e-06,
"loss": 0.5965,
"step": 594
},
{
"epoch": 0.5217295993862004,
"grad_norm": 0.3093125993326785,
"learning_rate": 4.730729288256468e-06,
"loss": 0.6018,
"step": 595
},
{
"epoch": 0.5226064558557572,
"grad_norm": 0.30147164249765196,
"learning_rate": 4.729660291809126e-06,
"loss": 0.6072,
"step": 596
},
{
"epoch": 0.5234833123253138,
"grad_norm": 0.2893545075475105,
"learning_rate": 4.728589298894284e-06,
"loss": 0.5894,
"step": 597
},
{
"epoch": 0.5243601687948704,
"grad_norm": 0.29778530349250987,
"learning_rate": 4.72751631047092e-06,
"loss": 0.5941,
"step": 598
},
{
"epoch": 0.5252370252644271,
"grad_norm": 0.2822751104373634,
"learning_rate": 4.726441327499805e-06,
"loss": 0.6056,
"step": 599
},
{
"epoch": 0.5261138817339837,
"grad_norm": 0.30381920940202223,
"learning_rate": 4.725364350943492e-06,
"loss": 0.6016,
"step": 600
},
{
"epoch": 0.5269907382035403,
"grad_norm": 0.2728312952142679,
"learning_rate": 4.72428538176632e-06,
"loss": 0.6033,
"step": 601
},
{
"epoch": 0.527867594673097,
"grad_norm": 0.2920360605636878,
"learning_rate": 4.723204420934413e-06,
"loss": 0.614,
"step": 602
},
{
"epoch": 0.5287444511426536,
"grad_norm": 0.282387818364113,
"learning_rate": 4.722121469415677e-06,
"loss": 0.5901,
"step": 603
},
{
"epoch": 0.5296213076122103,
"grad_norm": 0.2954181717364726,
"learning_rate": 4.721036528179802e-06,
"loss": 0.6043,
"step": 604
},
{
"epoch": 0.5304981640817669,
"grad_norm": 0.3084979402180987,
"learning_rate": 4.719949598198258e-06,
"loss": 0.5931,
"step": 605
},
{
"epoch": 0.5313750205513235,
"grad_norm": 0.3252699365181927,
"learning_rate": 4.718860680444297e-06,
"loss": 0.6181,
"step": 606
},
{
"epoch": 0.5322518770208802,
"grad_norm": 0.28357295095306256,
"learning_rate": 4.717769775892951e-06,
"loss": 0.5903,
"step": 607
},
{
"epoch": 0.5331287334904368,
"grad_norm": 0.3569079908279582,
"learning_rate": 4.7166768855210294e-06,
"loss": 0.5939,
"step": 608
},
{
"epoch": 0.5340055899599935,
"grad_norm": 0.31741200071485426,
"learning_rate": 4.715582010307121e-06,
"loss": 0.5897,
"step": 609
},
{
"epoch": 0.5348824464295501,
"grad_norm": 0.3218789245412814,
"learning_rate": 4.714485151231593e-06,
"loss": 0.5926,
"step": 610
},
{
"epoch": 0.5357593028991067,
"grad_norm": 0.2824610260583936,
"learning_rate": 4.713386309276585e-06,
"loss": 0.6039,
"step": 611
},
{
"epoch": 0.5366361593686634,
"grad_norm": 0.3111981063952015,
"learning_rate": 4.712285485426017e-06,
"loss": 0.6012,
"step": 612
},
{
"epoch": 0.53751301583822,
"grad_norm": 0.2719370118974663,
"learning_rate": 4.7111826806655804e-06,
"loss": 0.5912,
"step": 613
},
{
"epoch": 0.5383898723077766,
"grad_norm": 0.3161533458613161,
"learning_rate": 4.710077895982741e-06,
"loss": 0.5962,
"step": 614
},
{
"epoch": 0.5392667287773333,
"grad_norm": 0.26701338476822095,
"learning_rate": 4.708971132366739e-06,
"loss": 0.6025,
"step": 615
},
{
"epoch": 0.5401435852468899,
"grad_norm": 0.28447205168753736,
"learning_rate": 4.707862390808583e-06,
"loss": 0.5959,
"step": 616
},
{
"epoch": 0.5410204417164466,
"grad_norm": 0.26585350433139904,
"learning_rate": 4.706751672301058e-06,
"loss": 0.5946,
"step": 617
},
{
"epoch": 0.5418972981860032,
"grad_norm": 0.28276117956241253,
"learning_rate": 4.705638977838712e-06,
"loss": 0.5986,
"step": 618
},
{
"epoch": 0.5427741546555598,
"grad_norm": 0.2752743049051474,
"learning_rate": 4.704524308417872e-06,
"loss": 0.6044,
"step": 619
},
{
"epoch": 0.5436510111251165,
"grad_norm": 0.2744635750786116,
"learning_rate": 4.703407665036622e-06,
"loss": 0.6,
"step": 620
},
{
"epoch": 0.5445278675946731,
"grad_norm": 0.2942835089324837,
"learning_rate": 4.702289048694824e-06,
"loss": 0.6163,
"step": 621
},
{
"epoch": 0.5454047240642297,
"grad_norm": 0.29074004193212294,
"learning_rate": 4.7011684603940985e-06,
"loss": 0.61,
"step": 622
},
{
"epoch": 0.5462815805337864,
"grad_norm": 0.265548853050648,
"learning_rate": 4.700045901137838e-06,
"loss": 0.6003,
"step": 623
},
{
"epoch": 0.547158437003343,
"grad_norm": 0.28147341099339,
"learning_rate": 4.6989213719311956e-06,
"loss": 0.6057,
"step": 624
},
{
"epoch": 0.5480352934728997,
"grad_norm": 0.25061686481638634,
"learning_rate": 4.697794873781089e-06,
"loss": 0.6103,
"step": 625
},
{
"epoch": 0.5489121499424563,
"grad_norm": 0.28270079603778164,
"learning_rate": 4.696666407696201e-06,
"loss": 0.5999,
"step": 626
},
{
"epoch": 0.5497890064120129,
"grad_norm": 0.25832596909684546,
"learning_rate": 4.695535974686975e-06,
"loss": 0.5989,
"step": 627
},
{
"epoch": 0.5506658628815696,
"grad_norm": 0.28610489660664173,
"learning_rate": 4.694403575765615e-06,
"loss": 0.6039,
"step": 628
},
{
"epoch": 0.5515427193511262,
"grad_norm": 0.26039812165621273,
"learning_rate": 4.693269211946086e-06,
"loss": 0.5999,
"step": 629
},
{
"epoch": 0.5524195758206829,
"grad_norm": 0.2802813802636672,
"learning_rate": 4.692132884244113e-06,
"loss": 0.5957,
"step": 630
},
{
"epoch": 0.5532964322902395,
"grad_norm": 0.28045233973715045,
"learning_rate": 4.69099459367718e-06,
"loss": 0.6057,
"step": 631
},
{
"epoch": 0.5541732887597961,
"grad_norm": 0.2850165288729873,
"learning_rate": 4.689854341264525e-06,
"loss": 0.6062,
"step": 632
},
{
"epoch": 0.5550501452293528,
"grad_norm": 0.318532937146288,
"learning_rate": 4.688712128027147e-06,
"loss": 0.615,
"step": 633
},
{
"epoch": 0.5559270016989094,
"grad_norm": 0.2700297126701359,
"learning_rate": 4.687567954987798e-06,
"loss": 0.6027,
"step": 634
},
{
"epoch": 0.556803858168466,
"grad_norm": 0.2709567537114069,
"learning_rate": 4.686421823170987e-06,
"loss": 0.606,
"step": 635
},
{
"epoch": 0.5576807146380227,
"grad_norm": 0.30943308206128534,
"learning_rate": 4.685273733602975e-06,
"loss": 0.6122,
"step": 636
},
{
"epoch": 0.5585575711075793,
"grad_norm": 0.2866407684585244,
"learning_rate": 4.6841236873117765e-06,
"loss": 0.5983,
"step": 637
},
{
"epoch": 0.559434427577136,
"grad_norm": 0.30074858616349,
"learning_rate": 4.6829716853271576e-06,
"loss": 0.6112,
"step": 638
},
{
"epoch": 0.5603112840466926,
"grad_norm": 0.27481764632891953,
"learning_rate": 4.681817728680638e-06,
"loss": 0.5923,
"step": 639
},
{
"epoch": 0.5611881405162492,
"grad_norm": 0.30985792219487485,
"learning_rate": 4.680661818405485e-06,
"loss": 0.6083,
"step": 640
},
{
"epoch": 0.5620649969858059,
"grad_norm": 0.30548099410676144,
"learning_rate": 4.679503955536715e-06,
"loss": 0.6105,
"step": 641
},
{
"epoch": 0.5629418534553625,
"grad_norm": 0.27736446160459594,
"learning_rate": 4.678344141111096e-06,
"loss": 0.6176,
"step": 642
},
{
"epoch": 0.5638187099249191,
"grad_norm": 0.313370779146898,
"learning_rate": 4.6771823761671386e-06,
"loss": 0.6035,
"step": 643
},
{
"epoch": 0.5646955663944758,
"grad_norm": 0.27389315771120454,
"learning_rate": 4.676018661745104e-06,
"loss": 0.6118,
"step": 644
},
{
"epoch": 0.5655724228640324,
"grad_norm": 0.3272671136560007,
"learning_rate": 4.674852998886998e-06,
"loss": 0.6059,
"step": 645
},
{
"epoch": 0.5664492793335891,
"grad_norm": 0.29110434636858074,
"learning_rate": 4.6736853886365704e-06,
"loss": 0.5957,
"step": 646
},
{
"epoch": 0.5673261358031457,
"grad_norm": 0.27566640053494834,
"learning_rate": 4.672515832039315e-06,
"loss": 0.5847,
"step": 647
},
{
"epoch": 0.5682029922727023,
"grad_norm": 0.3439499837560115,
"learning_rate": 4.671344330142468e-06,
"loss": 0.6066,
"step": 648
},
{
"epoch": 0.569079848742259,
"grad_norm": 0.2831795036732806,
"learning_rate": 4.670170883995007e-06,
"loss": 0.5875,
"step": 649
},
{
"epoch": 0.5699567052118156,
"grad_norm": 0.3084275937304928,
"learning_rate": 4.668995494647653e-06,
"loss": 0.6046,
"step": 650
},
{
"epoch": 0.5708335616813722,
"grad_norm": 0.2876312566066635,
"learning_rate": 4.667818163152864e-06,
"loss": 0.609,
"step": 651
},
{
"epoch": 0.5717104181509289,
"grad_norm": 0.27641311480374825,
"learning_rate": 4.6666388905648394e-06,
"loss": 0.6084,
"step": 652
},
{
"epoch": 0.5725872746204855,
"grad_norm": 0.2760161681243495,
"learning_rate": 4.665457677939515e-06,
"loss": 0.6036,
"step": 653
},
{
"epoch": 0.5734641310900422,
"grad_norm": 0.2664014070652965,
"learning_rate": 4.664274526334563e-06,
"loss": 0.6047,
"step": 654
},
{
"epoch": 0.5743409875595988,
"grad_norm": 0.27367722811571643,
"learning_rate": 4.663089436809395e-06,
"loss": 0.607,
"step": 655
},
{
"epoch": 0.5752178440291554,
"grad_norm": 0.2971494077897638,
"learning_rate": 4.661902410425156e-06,
"loss": 0.5851,
"step": 656
},
{
"epoch": 0.5760947004987121,
"grad_norm": 0.28359506675344376,
"learning_rate": 4.660713448244723e-06,
"loss": 0.5911,
"step": 657
},
{
"epoch": 0.5769715569682687,
"grad_norm": 0.27646693971859265,
"learning_rate": 4.6595225513327105e-06,
"loss": 0.601,
"step": 658
},
{
"epoch": 0.5778484134378254,
"grad_norm": 0.2707379861432875,
"learning_rate": 4.658329720755464e-06,
"loss": 0.5905,
"step": 659
},
{
"epoch": 0.578725269907382,
"grad_norm": 0.301271851117793,
"learning_rate": 4.657134957581057e-06,
"loss": 0.6023,
"step": 660
},
{
"epoch": 0.5796021263769386,
"grad_norm": 0.30214846729641187,
"learning_rate": 4.6559382628793e-06,
"loss": 0.6095,
"step": 661
},
{
"epoch": 0.5804789828464954,
"grad_norm": 0.2880769859831512,
"learning_rate": 4.6547396377217265e-06,
"loss": 0.6012,
"step": 662
},
{
"epoch": 0.581355839316052,
"grad_norm": 0.3363251460755209,
"learning_rate": 4.653539083181603e-06,
"loss": 0.5963,
"step": 663
},
{
"epoch": 0.5822326957856085,
"grad_norm": 0.3446871487238731,
"learning_rate": 4.652336600333921e-06,
"loss": 0.5992,
"step": 664
},
{
"epoch": 0.5831095522551653,
"grad_norm": 0.3016824402176579,
"learning_rate": 4.651132190255401e-06,
"loss": 0.6016,
"step": 665
},
{
"epoch": 0.5839864087247219,
"grad_norm": 0.31791554379394255,
"learning_rate": 4.649925854024486e-06,
"loss": 0.5943,
"step": 666
},
{
"epoch": 0.5848632651942786,
"grad_norm": 0.3603510668723624,
"learning_rate": 4.648717592721347e-06,
"loss": 0.6086,
"step": 667
},
{
"epoch": 0.5857401216638352,
"grad_norm": 0.25073578292290827,
"learning_rate": 4.647507407427877e-06,
"loss": 0.5965,
"step": 668
},
{
"epoch": 0.5866169781333918,
"grad_norm": 0.3401292596267892,
"learning_rate": 4.646295299227691e-06,
"loss": 0.5896,
"step": 669
},
{
"epoch": 0.5874938346029485,
"grad_norm": 0.26798950974238206,
"learning_rate": 4.645081269206128e-06,
"loss": 0.5913,
"step": 670
},
{
"epoch": 0.5883706910725051,
"grad_norm": 0.2712753517614824,
"learning_rate": 4.643865318450247e-06,
"loss": 0.5948,
"step": 671
},
{
"epoch": 0.5892475475420617,
"grad_norm": 0.31478669896326056,
"learning_rate": 4.642647448048824e-06,
"loss": 0.6036,
"step": 672
},
{
"epoch": 0.5901244040116184,
"grad_norm": 0.2853149586152437,
"learning_rate": 4.641427659092359e-06,
"loss": 0.5852,
"step": 673
},
{
"epoch": 0.591001260481175,
"grad_norm": 0.31928733056145026,
"learning_rate": 4.6402059526730656e-06,
"loss": 0.596,
"step": 674
},
{
"epoch": 0.5918781169507317,
"grad_norm": 0.28886504451895006,
"learning_rate": 4.638982329884878e-06,
"loss": 0.5867,
"step": 675
},
{
"epoch": 0.5927549734202883,
"grad_norm": 0.34332786639440344,
"learning_rate": 4.637756791823443e-06,
"loss": 0.5951,
"step": 676
},
{
"epoch": 0.5936318298898449,
"grad_norm": 0.31536294202913445,
"learning_rate": 4.6365293395861225e-06,
"loss": 0.6005,
"step": 677
},
{
"epoch": 0.5945086863594016,
"grad_norm": 0.36612645695214535,
"learning_rate": 4.6352999742719954e-06,
"loss": 0.6125,
"step": 678
},
{
"epoch": 0.5953855428289582,
"grad_norm": 0.2865910172606529,
"learning_rate": 4.634068696981852e-06,
"loss": 0.6096,
"step": 679
},
{
"epoch": 0.5962623992985149,
"grad_norm": 0.3077121674916666,
"learning_rate": 4.632835508818192e-06,
"loss": 0.5891,
"step": 680
},
{
"epoch": 0.5971392557680715,
"grad_norm": 0.2930520316480949,
"learning_rate": 4.631600410885231e-06,
"loss": 0.5918,
"step": 681
},
{
"epoch": 0.5980161122376281,
"grad_norm": 0.3412197822800723,
"learning_rate": 4.630363404288891e-06,
"loss": 0.5998,
"step": 682
},
{
"epoch": 0.5988929687071848,
"grad_norm": 0.2869686807201651,
"learning_rate": 4.629124490136804e-06,
"loss": 0.5952,
"step": 683
},
{
"epoch": 0.5997698251767414,
"grad_norm": 0.3044523168792968,
"learning_rate": 4.627883669538311e-06,
"loss": 0.6058,
"step": 684
},
{
"epoch": 0.600646681646298,
"grad_norm": 0.298754941767322,
"learning_rate": 4.626640943604459e-06,
"loss": 0.6099,
"step": 685
},
{
"epoch": 0.6015235381158547,
"grad_norm": 0.30823608651620477,
"learning_rate": 4.625396313448e-06,
"loss": 0.5913,
"step": 686
},
{
"epoch": 0.6024003945854113,
"grad_norm": 0.2745802532714142,
"learning_rate": 4.624149780183395e-06,
"loss": 0.5904,
"step": 687
},
{
"epoch": 0.603277251054968,
"grad_norm": 0.2894557068485525,
"learning_rate": 4.622901344926805e-06,
"loss": 0.6006,
"step": 688
},
{
"epoch": 0.6041541075245246,
"grad_norm": 0.2844643276622375,
"learning_rate": 4.621651008796095e-06,
"loss": 0.5972,
"step": 689
},
{
"epoch": 0.6050309639940812,
"grad_norm": 0.3111750841694179,
"learning_rate": 4.620398772910833e-06,
"loss": 0.5911,
"step": 690
},
{
"epoch": 0.6059078204636379,
"grad_norm": 0.30229136138256857,
"learning_rate": 4.619144638392289e-06,
"loss": 0.6063,
"step": 691
},
{
"epoch": 0.6067846769331945,
"grad_norm": 0.2903177693650587,
"learning_rate": 4.6178886063634295e-06,
"loss": 0.6022,
"step": 692
},
{
"epoch": 0.6076615334027511,
"grad_norm": 0.29466063932438424,
"learning_rate": 4.616630677948924e-06,
"loss": 0.609,
"step": 693
},
{
"epoch": 0.6085383898723078,
"grad_norm": 0.29795014881552045,
"learning_rate": 4.615370854275138e-06,
"loss": 0.5923,
"step": 694
},
{
"epoch": 0.6094152463418644,
"grad_norm": 0.2835342651327551,
"learning_rate": 4.614109136470133e-06,
"loss": 0.5941,
"step": 695
},
{
"epoch": 0.6102921028114211,
"grad_norm": 0.2914927284695803,
"learning_rate": 4.612845525663671e-06,
"loss": 0.5915,
"step": 696
},
{
"epoch": 0.6111689592809777,
"grad_norm": 0.27150994490869584,
"learning_rate": 4.611580022987202e-06,
"loss": 0.5903,
"step": 697
},
{
"epoch": 0.6120458157505343,
"grad_norm": 0.27427922033901636,
"learning_rate": 4.610312629573877e-06,
"loss": 0.5826,
"step": 698
},
{
"epoch": 0.612922672220091,
"grad_norm": 0.3257835351903302,
"learning_rate": 4.609043346558536e-06,
"loss": 0.608,
"step": 699
},
{
"epoch": 0.6137995286896476,
"grad_norm": 0.27542786817313375,
"learning_rate": 4.607772175077712e-06,
"loss": 0.5914,
"step": 700
},
{
"epoch": 0.6146763851592043,
"grad_norm": 0.32541464673918596,
"learning_rate": 4.606499116269628e-06,
"loss": 0.6004,
"step": 701
},
{
"epoch": 0.6155532416287609,
"grad_norm": 0.2775394483279354,
"learning_rate": 4.605224171274198e-06,
"loss": 0.6042,
"step": 702
},
{
"epoch": 0.6164300980983175,
"grad_norm": 0.3010566442707075,
"learning_rate": 4.603947341233026e-06,
"loss": 0.5893,
"step": 703
},
{
"epoch": 0.6173069545678742,
"grad_norm": 0.28841806172316603,
"learning_rate": 4.602668627289401e-06,
"loss": 0.5932,
"step": 704
},
{
"epoch": 0.6181838110374308,
"grad_norm": 0.32720143492110876,
"learning_rate": 4.601388030588303e-06,
"loss": 0.594,
"step": 705
},
{
"epoch": 0.6190606675069874,
"grad_norm": 0.2629157828769276,
"learning_rate": 4.600105552276393e-06,
"loss": 0.5962,
"step": 706
},
{
"epoch": 0.6199375239765441,
"grad_norm": 0.2976311641314985,
"learning_rate": 4.598821193502019e-06,
"loss": 0.5993,
"step": 707
},
{
"epoch": 0.6208143804461007,
"grad_norm": 0.3223849407278096,
"learning_rate": 4.597534955415214e-06,
"loss": 0.6023,
"step": 708
},
{
"epoch": 0.6216912369156574,
"grad_norm": 0.3228934470983084,
"learning_rate": 4.596246839167692e-06,
"loss": 0.6058,
"step": 709
},
{
"epoch": 0.622568093385214,
"grad_norm": 0.2842350311614894,
"learning_rate": 4.59495684591285e-06,
"loss": 0.5965,
"step": 710
},
{
"epoch": 0.6234449498547706,
"grad_norm": 0.30037127301855626,
"learning_rate": 4.593664976805765e-06,
"loss": 0.5912,
"step": 711
},
{
"epoch": 0.6243218063243273,
"grad_norm": 0.29537031301186273,
"learning_rate": 4.592371233003195e-06,
"loss": 0.5847,
"step": 712
},
{
"epoch": 0.6251986627938839,
"grad_norm": 0.3099776656835445,
"learning_rate": 4.5910756156635725e-06,
"loss": 0.6061,
"step": 713
},
{
"epoch": 0.6260755192634405,
"grad_norm": 0.3343474177937486,
"learning_rate": 4.589778125947012e-06,
"loss": 0.5775,
"step": 714
},
{
"epoch": 0.6269523757329972,
"grad_norm": 0.26492597760028275,
"learning_rate": 4.588478765015304e-06,
"loss": 0.6008,
"step": 715
},
{
"epoch": 0.6278292322025538,
"grad_norm": 0.2996728173414987,
"learning_rate": 4.587177534031914e-06,
"loss": 0.5868,
"step": 716
},
{
"epoch": 0.6287060886721105,
"grad_norm": 0.269698012084879,
"learning_rate": 4.585874434161979e-06,
"loss": 0.5908,
"step": 717
},
{
"epoch": 0.6295829451416671,
"grad_norm": 0.3120812259438331,
"learning_rate": 4.584569466572313e-06,
"loss": 0.5964,
"step": 718
},
{
"epoch": 0.6304598016112237,
"grad_norm": 0.306605213663903,
"learning_rate": 4.583262632431402e-06,
"loss": 0.587,
"step": 719
},
{
"epoch": 0.6313366580807804,
"grad_norm": 0.31045769873517814,
"learning_rate": 4.581953932909403e-06,
"loss": 0.5924,
"step": 720
},
{
"epoch": 0.632213514550337,
"grad_norm": 0.30956000847409926,
"learning_rate": 4.580643369178142e-06,
"loss": 0.5905,
"step": 721
},
{
"epoch": 0.6330903710198936,
"grad_norm": 0.2980650280091205,
"learning_rate": 4.579330942411115e-06,
"loss": 0.5961,
"step": 722
},
{
"epoch": 0.6339672274894503,
"grad_norm": 0.2784986194522932,
"learning_rate": 4.578016653783488e-06,
"loss": 0.5962,
"step": 723
},
{
"epoch": 0.6348440839590069,
"grad_norm": 0.32816601752120567,
"learning_rate": 4.57670050447209e-06,
"loss": 0.6149,
"step": 724
},
{
"epoch": 0.6357209404285636,
"grad_norm": 0.2822290286934802,
"learning_rate": 4.575382495655421e-06,
"loss": 0.5915,
"step": 725
},
{
"epoch": 0.6365977968981202,
"grad_norm": 0.2993973936416954,
"learning_rate": 4.574062628513643e-06,
"loss": 0.59,
"step": 726
},
{
"epoch": 0.6374746533676768,
"grad_norm": 0.27875804168057794,
"learning_rate": 4.572740904228582e-06,
"loss": 0.6018,
"step": 727
},
{
"epoch": 0.6383515098372335,
"grad_norm": 0.3144256132274513,
"learning_rate": 4.571417323983727e-06,
"loss": 0.6056,
"step": 728
},
{
"epoch": 0.6392283663067901,
"grad_norm": 0.2763723528672814,
"learning_rate": 4.570091888964231e-06,
"loss": 0.5943,
"step": 729
},
{
"epoch": 0.6401052227763468,
"grad_norm": 0.3001278571328794,
"learning_rate": 4.5687646003569055e-06,
"loss": 0.588,
"step": 730
},
{
"epoch": 0.6409820792459034,
"grad_norm": 0.2847820308061442,
"learning_rate": 4.567435459350222e-06,
"loss": 0.5971,
"step": 731
},
{
"epoch": 0.64185893571546,
"grad_norm": 0.292512543142512,
"learning_rate": 4.566104467134311e-06,
"loss": 0.5864,
"step": 732
},
{
"epoch": 0.6427357921850168,
"grad_norm": 0.28968651062565176,
"learning_rate": 4.564771624900961e-06,
"loss": 0.62,
"step": 733
},
{
"epoch": 0.6436126486545733,
"grad_norm": 0.3004795852693458,
"learning_rate": 4.563436933843617e-06,
"loss": 0.5964,
"step": 734
},
{
"epoch": 0.64448950512413,
"grad_norm": 0.2865806085716862,
"learning_rate": 4.562100395157379e-06,
"loss": 0.6026,
"step": 735
},
{
"epoch": 0.6453663615936867,
"grad_norm": 0.2842649974188147,
"learning_rate": 4.560762010039001e-06,
"loss": 0.5913,
"step": 736
},
{
"epoch": 0.6462432180632433,
"grad_norm": 0.28683866497814775,
"learning_rate": 4.5594217796868915e-06,
"loss": 0.5951,
"step": 737
},
{
"epoch": 0.6471200745328,
"grad_norm": 0.2764873070461295,
"learning_rate": 4.558079705301109e-06,
"loss": 0.6053,
"step": 738
},
{
"epoch": 0.6479969310023566,
"grad_norm": 0.27004479414645,
"learning_rate": 4.556735788083366e-06,
"loss": 0.6039,
"step": 739
},
{
"epoch": 0.6488737874719132,
"grad_norm": 0.29052397029213667,
"learning_rate": 4.555390029237026e-06,
"loss": 0.601,
"step": 740
},
{
"epoch": 0.6497506439414699,
"grad_norm": 0.2947691340138793,
"learning_rate": 4.554042429967095e-06,
"loss": 0.6025,
"step": 741
},
{
"epoch": 0.6506275004110265,
"grad_norm": 0.2792458027197797,
"learning_rate": 4.552692991480234e-06,
"loss": 0.6014,
"step": 742
},
{
"epoch": 0.6515043568805831,
"grad_norm": 0.3382217380230472,
"learning_rate": 4.551341714984748e-06,
"loss": 0.5955,
"step": 743
},
{
"epoch": 0.6523812133501398,
"grad_norm": 0.2966197192699023,
"learning_rate": 4.549988601690588e-06,
"loss": 0.5935,
"step": 744
},
{
"epoch": 0.6532580698196964,
"grad_norm": 0.31516646846151397,
"learning_rate": 4.54863365280935e-06,
"loss": 0.597,
"step": 745
},
{
"epoch": 0.6541349262892531,
"grad_norm": 0.28496714910224397,
"learning_rate": 4.547276869554272e-06,
"loss": 0.5814,
"step": 746
},
{
"epoch": 0.6550117827588097,
"grad_norm": 0.30669749001026353,
"learning_rate": 4.545918253140236e-06,
"loss": 0.5952,
"step": 747
},
{
"epoch": 0.6558886392283663,
"grad_norm": 0.2812261666412913,
"learning_rate": 4.544557804783765e-06,
"loss": 0.6162,
"step": 748
},
{
"epoch": 0.656765495697923,
"grad_norm": 0.27761745178740765,
"learning_rate": 4.543195525703024e-06,
"loss": 0.5807,
"step": 749
},
{
"epoch": 0.6576423521674796,
"grad_norm": 0.31002121863979637,
"learning_rate": 4.541831417117815e-06,
"loss": 0.5851,
"step": 750
},
{
"epoch": 0.6585192086370363,
"grad_norm": 0.29034303454873894,
"learning_rate": 4.540465480249579e-06,
"loss": 0.6019,
"step": 751
},
{
"epoch": 0.6593960651065929,
"grad_norm": 0.30559901683462565,
"learning_rate": 4.539097716321394e-06,
"loss": 0.5866,
"step": 752
},
{
"epoch": 0.6602729215761495,
"grad_norm": 0.2641221990159659,
"learning_rate": 4.537728126557974e-06,
"loss": 0.5972,
"step": 753
},
{
"epoch": 0.6611497780457062,
"grad_norm": 0.3227708789669896,
"learning_rate": 4.536356712185668e-06,
"loss": 0.5796,
"step": 754
},
{
"epoch": 0.6620266345152628,
"grad_norm": 0.294701481555053,
"learning_rate": 4.534983474432458e-06,
"loss": 0.6149,
"step": 755
},
{
"epoch": 0.6629034909848194,
"grad_norm": 0.32377533070879033,
"learning_rate": 4.533608414527961e-06,
"loss": 0.5891,
"step": 756
},
{
"epoch": 0.6637803474543761,
"grad_norm": 0.3042889879699245,
"learning_rate": 4.532231533703423e-06,
"loss": 0.5913,
"step": 757
},
{
"epoch": 0.6646572039239327,
"grad_norm": 0.31760559251266973,
"learning_rate": 4.53085283319172e-06,
"loss": 0.6096,
"step": 758
},
{
"epoch": 0.6655340603934894,
"grad_norm": 0.3078941609749165,
"learning_rate": 4.529472314227362e-06,
"loss": 0.5905,
"step": 759
},
{
"epoch": 0.666410916863046,
"grad_norm": 0.30990175786815527,
"learning_rate": 4.528089978046481e-06,
"loss": 0.5991,
"step": 760
},
{
"epoch": 0.6672877733326026,
"grad_norm": 0.32903820758007046,
"learning_rate": 4.5267058258868414e-06,
"loss": 0.5882,
"step": 761
},
{
"epoch": 0.6681646298021593,
"grad_norm": 0.29452587669480845,
"learning_rate": 4.52531985898783e-06,
"loss": 0.5803,
"step": 762
},
{
"epoch": 0.6690414862717159,
"grad_norm": 0.30776706716693625,
"learning_rate": 4.52393207859046e-06,
"loss": 0.577,
"step": 763
},
{
"epoch": 0.6699183427412725,
"grad_norm": 0.31422641761257675,
"learning_rate": 4.522542485937369e-06,
"loss": 0.6018,
"step": 764
},
{
"epoch": 0.6707951992108292,
"grad_norm": 0.3173718550935184,
"learning_rate": 4.521151082272817e-06,
"loss": 0.5882,
"step": 765
},
{
"epoch": 0.6716720556803858,
"grad_norm": 0.2986562015643124,
"learning_rate": 4.519757868842685e-06,
"loss": 0.579,
"step": 766
},
{
"epoch": 0.6725489121499425,
"grad_norm": 0.3090764441547647,
"learning_rate": 4.518362846894475e-06,
"loss": 0.5985,
"step": 767
},
{
"epoch": 0.6734257686194991,
"grad_norm": 0.30790241933986734,
"learning_rate": 4.516966017677308e-06,
"loss": 0.5863,
"step": 768
},
{
"epoch": 0.6743026250890557,
"grad_norm": 0.2994056106304016,
"learning_rate": 4.515567382441923e-06,
"loss": 0.5991,
"step": 769
},
{
"epoch": 0.6751794815586124,
"grad_norm": 0.2958764046270931,
"learning_rate": 4.514166942440679e-06,
"loss": 0.5963,
"step": 770
},
{
"epoch": 0.676056338028169,
"grad_norm": 0.28788185549499157,
"learning_rate": 4.512764698927545e-06,
"loss": 0.6064,
"step": 771
},
{
"epoch": 0.6769331944977256,
"grad_norm": 0.29708423016925406,
"learning_rate": 4.511360653158111e-06,
"loss": 0.5947,
"step": 772
},
{
"epoch": 0.6778100509672823,
"grad_norm": 0.30991902940049315,
"learning_rate": 4.509954806389577e-06,
"loss": 0.5987,
"step": 773
},
{
"epoch": 0.6786869074368389,
"grad_norm": 0.2873916475278516,
"learning_rate": 4.508547159880758e-06,
"loss": 0.5924,
"step": 774
},
{
"epoch": 0.6795637639063956,
"grad_norm": 0.3007245570293541,
"learning_rate": 4.50713771489208e-06,
"loss": 0.6015,
"step": 775
},
{
"epoch": 0.6804406203759522,
"grad_norm": 0.30867041078073276,
"learning_rate": 4.505726472685577e-06,
"loss": 0.5957,
"step": 776
},
{
"epoch": 0.6813174768455088,
"grad_norm": 0.31345922212682475,
"learning_rate": 4.504313434524894e-06,
"loss": 0.6006,
"step": 777
},
{
"epoch": 0.6821943333150655,
"grad_norm": 0.29707717549610757,
"learning_rate": 4.502898601675285e-06,
"loss": 0.5778,
"step": 778
},
{
"epoch": 0.6830711897846221,
"grad_norm": 0.3796068136152165,
"learning_rate": 4.501481975403611e-06,
"loss": 0.5991,
"step": 779
},
{
"epoch": 0.6839480462541788,
"grad_norm": 0.28337342976468866,
"learning_rate": 4.5000635569783365e-06,
"loss": 0.5948,
"step": 780
},
{
"epoch": 0.6848249027237354,
"grad_norm": 0.31230108669893153,
"learning_rate": 4.498643347669533e-06,
"loss": 0.5925,
"step": 781
},
{
"epoch": 0.685701759193292,
"grad_norm": 0.27904331433791485,
"learning_rate": 4.497221348748874e-06,
"loss": 0.5916,
"step": 782
},
{
"epoch": 0.6865786156628487,
"grad_norm": 0.2942542969448629,
"learning_rate": 4.4957975614896386e-06,
"loss": 0.5992,
"step": 783
},
{
"epoch": 0.6874554721324053,
"grad_norm": 0.2908765617548673,
"learning_rate": 4.494371987166703e-06,
"loss": 0.6065,
"step": 784
},
{
"epoch": 0.6883323286019619,
"grad_norm": 0.2840490179126863,
"learning_rate": 4.492944627056544e-06,
"loss": 0.5902,
"step": 785
},
{
"epoch": 0.6892091850715186,
"grad_norm": 0.2727369127304506,
"learning_rate": 4.491515482437242e-06,
"loss": 0.5867,
"step": 786
},
{
"epoch": 0.6900860415410752,
"grad_norm": 0.28769481832954025,
"learning_rate": 4.4900845545884695e-06,
"loss": 0.5922,
"step": 787
},
{
"epoch": 0.6909628980106319,
"grad_norm": 0.2906309237155975,
"learning_rate": 4.4886518447915e-06,
"loss": 0.5887,
"step": 788
},
{
"epoch": 0.6918397544801885,
"grad_norm": 0.2948842293422461,
"learning_rate": 4.487217354329201e-06,
"loss": 0.6006,
"step": 789
},
{
"epoch": 0.6927166109497451,
"grad_norm": 0.302074977476922,
"learning_rate": 4.4857810844860325e-06,
"loss": 0.5866,
"step": 790
},
{
"epoch": 0.6935934674193018,
"grad_norm": 0.32893770275300094,
"learning_rate": 4.484343036548051e-06,
"loss": 0.5976,
"step": 791
},
{
"epoch": 0.6944703238888584,
"grad_norm": 0.2778002794834819,
"learning_rate": 4.482903211802904e-06,
"loss": 0.584,
"step": 792
},
{
"epoch": 0.695347180358415,
"grad_norm": 0.294631010190205,
"learning_rate": 4.481461611539829e-06,
"loss": 0.5796,
"step": 793
},
{
"epoch": 0.6962240368279717,
"grad_norm": 0.26497721691156156,
"learning_rate": 4.480018237049655e-06,
"loss": 0.5921,
"step": 794
},
{
"epoch": 0.6971008932975283,
"grad_norm": 0.2571147884128945,
"learning_rate": 4.4785730896247985e-06,
"loss": 0.5967,
"step": 795
},
{
"epoch": 0.697977749767085,
"grad_norm": 0.27928133327664356,
"learning_rate": 4.477126170559262e-06,
"loss": 0.5933,
"step": 796
},
{
"epoch": 0.6988546062366416,
"grad_norm": 0.2678842819485542,
"learning_rate": 4.475677481148638e-06,
"loss": 0.6041,
"step": 797
},
{
"epoch": 0.6997314627061982,
"grad_norm": 0.2891606093702898,
"learning_rate": 4.474227022690102e-06,
"loss": 0.5957,
"step": 798
},
{
"epoch": 0.700608319175755,
"grad_norm": 0.288045727848727,
"learning_rate": 4.4727747964824135e-06,
"loss": 0.5904,
"step": 799
},
{
"epoch": 0.7014851756453115,
"grad_norm": 0.31585634496103415,
"learning_rate": 4.471320803825915e-06,
"loss": 0.5976,
"step": 800
},
{
"epoch": 0.7023620321148683,
"grad_norm": 0.2748185200755283,
"learning_rate": 4.469865046022531e-06,
"loss": 0.5752,
"step": 801
},
{
"epoch": 0.7032388885844248,
"grad_norm": 0.3355774877957403,
"learning_rate": 4.468407524375767e-06,
"loss": 0.5983,
"step": 802
},
{
"epoch": 0.7041157450539814,
"grad_norm": 0.29100988533473726,
"learning_rate": 4.466948240190707e-06,
"loss": 0.5942,
"step": 803
},
{
"epoch": 0.7049926015235382,
"grad_norm": 0.32395113661904446,
"learning_rate": 4.465487194774012e-06,
"loss": 0.5934,
"step": 804
},
{
"epoch": 0.7058694579930948,
"grad_norm": 0.27010926989878575,
"learning_rate": 4.464024389433924e-06,
"loss": 0.5965,
"step": 805
},
{
"epoch": 0.7067463144626513,
"grad_norm": 0.31589368881558894,
"learning_rate": 4.462559825480257e-06,
"loss": 0.5892,
"step": 806
},
{
"epoch": 0.7076231709322081,
"grad_norm": 0.2696414843727876,
"learning_rate": 4.461093504224401e-06,
"loss": 0.5995,
"step": 807
},
{
"epoch": 0.7085000274017647,
"grad_norm": 0.2953330107498836,
"learning_rate": 4.459625426979319e-06,
"loss": 0.5918,
"step": 808
},
{
"epoch": 0.7093768838713214,
"grad_norm": 0.281894292123873,
"learning_rate": 4.458155595059549e-06,
"loss": 0.5955,
"step": 809
},
{
"epoch": 0.710253740340878,
"grad_norm": 0.27376761478776995,
"learning_rate": 4.4566840097811956e-06,
"loss": 0.5871,
"step": 810
},
{
"epoch": 0.7111305968104346,
"grad_norm": 0.27713167306531405,
"learning_rate": 4.455210672461938e-06,
"loss": 0.595,
"step": 811
},
{
"epoch": 0.7120074532799913,
"grad_norm": 0.27385713088626723,
"learning_rate": 4.453735584421021e-06,
"loss": 0.5899,
"step": 812
},
{
"epoch": 0.7128843097495479,
"grad_norm": 0.29840396727897567,
"learning_rate": 4.452258746979258e-06,
"loss": 0.5844,
"step": 813
},
{
"epoch": 0.7137611662191045,
"grad_norm": 0.28333795883109736,
"learning_rate": 4.4507801614590285e-06,
"loss": 0.5939,
"step": 814
},
{
"epoch": 0.7146380226886612,
"grad_norm": 0.3089268512848077,
"learning_rate": 4.449299829184278e-06,
"loss": 0.5859,
"step": 815
},
{
"epoch": 0.7155148791582178,
"grad_norm": 0.2808961599877815,
"learning_rate": 4.447817751480516e-06,
"loss": 0.5871,
"step": 816
},
{
"epoch": 0.7163917356277745,
"grad_norm": 0.30287533725577037,
"learning_rate": 4.446333929674816e-06,
"loss": 0.593,
"step": 817
},
{
"epoch": 0.7172685920973311,
"grad_norm": 0.30584446638710266,
"learning_rate": 4.444848365095809e-06,
"loss": 0.5917,
"step": 818
},
{
"epoch": 0.7181454485668877,
"grad_norm": 0.27241453105670504,
"learning_rate": 4.44336105907369e-06,
"loss": 0.5896,
"step": 819
},
{
"epoch": 0.7190223050364444,
"grad_norm": 0.36474064413319707,
"learning_rate": 4.4418720129402145e-06,
"loss": 0.5861,
"step": 820
},
{
"epoch": 0.719899161506001,
"grad_norm": 0.2832577542195539,
"learning_rate": 4.4403812280286915e-06,
"loss": 0.5905,
"step": 821
},
{
"epoch": 0.7207760179755577,
"grad_norm": 0.32117553322486775,
"learning_rate": 4.4388887056739926e-06,
"loss": 0.5801,
"step": 822
},
{
"epoch": 0.7216528744451143,
"grad_norm": 0.27537463782509236,
"learning_rate": 4.43739444721254e-06,
"loss": 0.587,
"step": 823
},
{
"epoch": 0.7225297309146709,
"grad_norm": 0.3274304411602489,
"learning_rate": 4.435898453982313e-06,
"loss": 0.6024,
"step": 824
},
{
"epoch": 0.7234065873842276,
"grad_norm": 0.3232032167824163,
"learning_rate": 4.434400727322844e-06,
"loss": 0.6145,
"step": 825
},
{
"epoch": 0.7242834438537842,
"grad_norm": 0.3431783037261662,
"learning_rate": 4.432901268575218e-06,
"loss": 0.5937,
"step": 826
},
{
"epoch": 0.7251603003233408,
"grad_norm": 0.30897032551229503,
"learning_rate": 4.43140007908207e-06,
"loss": 0.598,
"step": 827
},
{
"epoch": 0.7260371567928975,
"grad_norm": 0.2934772547759602,
"learning_rate": 4.429897160187584e-06,
"loss": 0.5918,
"step": 828
},
{
"epoch": 0.7269140132624541,
"grad_norm": 0.31389790755569874,
"learning_rate": 4.4283925132374946e-06,
"loss": 0.5832,
"step": 829
},
{
"epoch": 0.7277908697320108,
"grad_norm": 0.29548260652561004,
"learning_rate": 4.426886139579083e-06,
"loss": 0.5937,
"step": 830
},
{
"epoch": 0.7286677262015674,
"grad_norm": 0.3162599265610075,
"learning_rate": 4.425378040561175e-06,
"loss": 0.5889,
"step": 831
},
{
"epoch": 0.729544582671124,
"grad_norm": 0.3057143041654656,
"learning_rate": 4.423868217534144e-06,
"loss": 0.5848,
"step": 832
},
{
"epoch": 0.7304214391406807,
"grad_norm": 0.29540394945672244,
"learning_rate": 4.4223566718499055e-06,
"loss": 0.5926,
"step": 833
},
{
"epoch": 0.7312982956102373,
"grad_norm": 0.30681513325771914,
"learning_rate": 4.420843404861917e-06,
"loss": 0.5838,
"step": 834
},
{
"epoch": 0.7321751520797939,
"grad_norm": 0.29780757398255076,
"learning_rate": 4.419328417925177e-06,
"loss": 0.5922,
"step": 835
},
{
"epoch": 0.7330520085493506,
"grad_norm": 0.28283439818927025,
"learning_rate": 4.417811712396226e-06,
"loss": 0.5875,
"step": 836
},
{
"epoch": 0.7339288650189072,
"grad_norm": 0.30029201304931724,
"learning_rate": 4.416293289633144e-06,
"loss": 0.5989,
"step": 837
},
{
"epoch": 0.7348057214884639,
"grad_norm": 0.29188774973524867,
"learning_rate": 4.414773150995543e-06,
"loss": 0.5878,
"step": 838
},
{
"epoch": 0.7356825779580205,
"grad_norm": 0.3037257039566602,
"learning_rate": 4.413251297844579e-06,
"loss": 0.5849,
"step": 839
},
{
"epoch": 0.7365594344275771,
"grad_norm": 0.31802355671271254,
"learning_rate": 4.411727731542937e-06,
"loss": 0.5873,
"step": 840
},
{
"epoch": 0.7374362908971338,
"grad_norm": 0.31892860544931334,
"learning_rate": 4.410202453454841e-06,
"loss": 0.5784,
"step": 841
},
{
"epoch": 0.7383131473666904,
"grad_norm": 0.31731371407494563,
"learning_rate": 4.408675464946043e-06,
"loss": 0.5973,
"step": 842
},
{
"epoch": 0.739190003836247,
"grad_norm": 0.2807004884396655,
"learning_rate": 4.40714676738383e-06,
"loss": 0.5842,
"step": 843
},
{
"epoch": 0.7400668603058037,
"grad_norm": 0.3102700515568577,
"learning_rate": 4.405616362137017e-06,
"loss": 0.584,
"step": 844
},
{
"epoch": 0.7409437167753603,
"grad_norm": 0.28221217756766914,
"learning_rate": 4.404084250575952e-06,
"loss": 0.599,
"step": 845
},
{
"epoch": 0.741820573244917,
"grad_norm": 0.284085524365953,
"learning_rate": 4.4025504340725056e-06,
"loss": 0.5799,
"step": 846
},
{
"epoch": 0.7426974297144736,
"grad_norm": 0.35367792241463614,
"learning_rate": 4.401014914000078e-06,
"loss": 0.5724,
"step": 847
},
{
"epoch": 0.7435742861840302,
"grad_norm": 0.26695572041406385,
"learning_rate": 4.3994776917335945e-06,
"loss": 0.5864,
"step": 848
},
{
"epoch": 0.7444511426535869,
"grad_norm": 0.3230503614090004,
"learning_rate": 4.397938768649505e-06,
"loss": 0.5781,
"step": 849
},
{
"epoch": 0.7453279991231435,
"grad_norm": 0.32670313161244324,
"learning_rate": 4.39639814612578e-06,
"loss": 0.5921,
"step": 850
},
{
"epoch": 0.7462048555927002,
"grad_norm": 0.2965265275169285,
"learning_rate": 4.394855825541915e-06,
"loss": 0.5847,
"step": 851
},
{
"epoch": 0.7470817120622568,
"grad_norm": 0.3364787473225747,
"learning_rate": 4.393311808278924e-06,
"loss": 0.6032,
"step": 852
},
{
"epoch": 0.7479585685318134,
"grad_norm": 0.2925797984612242,
"learning_rate": 4.391766095719341e-06,
"loss": 0.5966,
"step": 853
},
{
"epoch": 0.7488354250013701,
"grad_norm": 0.36558987387215064,
"learning_rate": 4.390218689247216e-06,
"loss": 0.5965,
"step": 854
},
{
"epoch": 0.7497122814709267,
"grad_norm": 0.31214927998435166,
"learning_rate": 4.388669590248119e-06,
"loss": 0.5799,
"step": 855
},
{
"epoch": 0.7505891379404833,
"grad_norm": 0.36912682982458045,
"learning_rate": 4.387118800109133e-06,
"loss": 0.5994,
"step": 856
},
{
"epoch": 0.75146599441004,
"grad_norm": 0.33858825867324854,
"learning_rate": 4.385566320218857e-06,
"loss": 0.5894,
"step": 857
},
{
"epoch": 0.7523428508795966,
"grad_norm": 0.3095865037795698,
"learning_rate": 4.384012151967401e-06,
"loss": 0.5808,
"step": 858
},
{
"epoch": 0.7532197073491533,
"grad_norm": 0.3163720033341599,
"learning_rate": 4.382456296746389e-06,
"loss": 0.61,
"step": 859
},
{
"epoch": 0.7540965638187099,
"grad_norm": 0.30746322298068,
"learning_rate": 4.3808987559489536e-06,
"loss": 0.5901,
"step": 860
},
{
"epoch": 0.7549734202882665,
"grad_norm": 0.3216332568956709,
"learning_rate": 4.379339530969738e-06,
"loss": 0.5824,
"step": 861
},
{
"epoch": 0.7558502767578232,
"grad_norm": 0.2924396456503393,
"learning_rate": 4.377778623204894e-06,
"loss": 0.587,
"step": 862
},
{
"epoch": 0.7567271332273798,
"grad_norm": 0.3102518126275497,
"learning_rate": 4.3762160340520765e-06,
"loss": 0.5722,
"step": 863
},
{
"epoch": 0.7576039896969364,
"grad_norm": 0.29990520801248277,
"learning_rate": 4.374651764910452e-06,
"loss": 0.5867,
"step": 864
},
{
"epoch": 0.7584808461664931,
"grad_norm": 0.2742400854190758,
"learning_rate": 4.373085817180684e-06,
"loss": 0.5897,
"step": 865
},
{
"epoch": 0.7593577026360497,
"grad_norm": 0.2966143324054175,
"learning_rate": 4.371518192264946e-06,
"loss": 0.593,
"step": 866
},
{
"epoch": 0.7602345591056064,
"grad_norm": 0.2659050257990803,
"learning_rate": 4.3699488915669106e-06,
"loss": 0.5933,
"step": 867
},
{
"epoch": 0.761111415575163,
"grad_norm": 0.28333909213084835,
"learning_rate": 4.368377916491749e-06,
"loss": 0.5937,
"step": 868
},
{
"epoch": 0.7619882720447196,
"grad_norm": 0.294367790561846,
"learning_rate": 4.366805268446132e-06,
"loss": 0.5908,
"step": 869
},
{
"epoch": 0.7628651285142763,
"grad_norm": 0.2892104769841804,
"learning_rate": 4.365230948838232e-06,
"loss": 0.5749,
"step": 870
},
{
"epoch": 0.7637419849838329,
"grad_norm": 0.2992157610185369,
"learning_rate": 4.3636549590777144e-06,
"loss": 0.6038,
"step": 871
},
{
"epoch": 0.7646188414533897,
"grad_norm": 0.2849149162166013,
"learning_rate": 4.362077300575742e-06,
"loss": 0.5838,
"step": 872
},
{
"epoch": 0.7654956979229462,
"grad_norm": 0.27419838720395556,
"learning_rate": 4.360497974744971e-06,
"loss": 0.5792,
"step": 873
},
{
"epoch": 0.7663725543925028,
"grad_norm": 0.2719357502719954,
"learning_rate": 4.35891698299955e-06,
"loss": 0.5879,
"step": 874
},
{
"epoch": 0.7672494108620596,
"grad_norm": 0.29276621658420166,
"learning_rate": 4.357334326755123e-06,
"loss": 0.5903,
"step": 875
},
{
"epoch": 0.7681262673316162,
"grad_norm": 0.29234711934765684,
"learning_rate": 4.3557500074288175e-06,
"loss": 0.58,
"step": 876
},
{
"epoch": 0.7690031238011727,
"grad_norm": 0.2900743371372321,
"learning_rate": 4.354164026439256e-06,
"loss": 0.5798,
"step": 877
},
{
"epoch": 0.7698799802707295,
"grad_norm": 0.26606697197934875,
"learning_rate": 4.352576385206547e-06,
"loss": 0.6049,
"step": 878
},
{
"epoch": 0.770756836740286,
"grad_norm": 0.30681607920100556,
"learning_rate": 4.350987085152286e-06,
"loss": 0.5963,
"step": 879
},
{
"epoch": 0.7716336932098428,
"grad_norm": 0.28024451945836265,
"learning_rate": 4.349396127699552e-06,
"loss": 0.6063,
"step": 880
},
{
"epoch": 0.7725105496793994,
"grad_norm": 0.284435176139814,
"learning_rate": 4.347803514272911e-06,
"loss": 0.5847,
"step": 881
},
{
"epoch": 0.773387406148956,
"grad_norm": 0.2787875052171573,
"learning_rate": 4.34620924629841e-06,
"loss": 0.5909,
"step": 882
},
{
"epoch": 0.7742642626185127,
"grad_norm": 0.28222554386796406,
"learning_rate": 4.344613325203577e-06,
"loss": 0.5815,
"step": 883
},
{
"epoch": 0.7751411190880693,
"grad_norm": 0.30850175508825417,
"learning_rate": 4.343015752417421e-06,
"loss": 0.5761,
"step": 884
},
{
"epoch": 0.7760179755576259,
"grad_norm": 0.27711497578948074,
"learning_rate": 4.341416529370431e-06,
"loss": 0.5851,
"step": 885
},
{
"epoch": 0.7768948320271826,
"grad_norm": 0.2945928621135004,
"learning_rate": 4.339815657494571e-06,
"loss": 0.5922,
"step": 886
},
{
"epoch": 0.7777716884967392,
"grad_norm": 0.2843169638684151,
"learning_rate": 4.338213138223285e-06,
"loss": 0.5835,
"step": 887
},
{
"epoch": 0.7786485449662959,
"grad_norm": 0.2840612846899258,
"learning_rate": 4.336608972991489e-06,
"loss": 0.596,
"step": 888
},
{
"epoch": 0.7795254014358525,
"grad_norm": 0.2677194609487142,
"learning_rate": 4.335003163235574e-06,
"loss": 0.5794,
"step": 889
},
{
"epoch": 0.7804022579054091,
"grad_norm": 0.31211329913480695,
"learning_rate": 4.3333957103934025e-06,
"loss": 0.5765,
"step": 890
},
{
"epoch": 0.7812791143749658,
"grad_norm": 0.28583623636409483,
"learning_rate": 4.33178661590431e-06,
"loss": 0.6016,
"step": 891
},
{
"epoch": 0.7821559708445224,
"grad_norm": 0.31500304190137224,
"learning_rate": 4.330175881209102e-06,
"loss": 0.5877,
"step": 892
},
{
"epoch": 0.783032827314079,
"grad_norm": 0.2811796495740926,
"learning_rate": 4.32856350775005e-06,
"loss": 0.5881,
"step": 893
},
{
"epoch": 0.7839096837836357,
"grad_norm": 0.29273259848443445,
"learning_rate": 4.3269494969708954e-06,
"loss": 0.5921,
"step": 894
},
{
"epoch": 0.7847865402531923,
"grad_norm": 0.27373150864211443,
"learning_rate": 4.325333850316846e-06,
"loss": 0.6,
"step": 895
},
{
"epoch": 0.785663396722749,
"grad_norm": 0.3128309122282222,
"learning_rate": 4.323716569234572e-06,
"loss": 0.5904,
"step": 896
},
{
"epoch": 0.7865402531923056,
"grad_norm": 0.2825745062634813,
"learning_rate": 4.32209765517221e-06,
"loss": 0.5816,
"step": 897
},
{
"epoch": 0.7874171096618622,
"grad_norm": 0.3282727674741808,
"learning_rate": 4.320477109579354e-06,
"loss": 0.5882,
"step": 898
},
{
"epoch": 0.7882939661314189,
"grad_norm": 0.2940095641373108,
"learning_rate": 4.318854933907065e-06,
"loss": 0.5985,
"step": 899
},
{
"epoch": 0.7891708226009755,
"grad_norm": 0.31182474508449737,
"learning_rate": 4.317231129607859e-06,
"loss": 0.5843,
"step": 900
},
{
"epoch": 0.7900476790705322,
"grad_norm": 0.26489892008261595,
"learning_rate": 4.315605698135714e-06,
"loss": 0.591,
"step": 901
},
{
"epoch": 0.7909245355400888,
"grad_norm": 0.32933790566988397,
"learning_rate": 4.313978640946061e-06,
"loss": 0.5826,
"step": 902
},
{
"epoch": 0.7918013920096454,
"grad_norm": 0.2790564068544957,
"learning_rate": 4.312349959495791e-06,
"loss": 0.5897,
"step": 903
},
{
"epoch": 0.7926782484792021,
"grad_norm": 0.29278849432785253,
"learning_rate": 4.310719655243243e-06,
"loss": 0.5929,
"step": 904
},
{
"epoch": 0.7935551049487587,
"grad_norm": 0.2898094197798441,
"learning_rate": 4.309087729648217e-06,
"loss": 0.575,
"step": 905
},
{
"epoch": 0.7944319614183153,
"grad_norm": 0.2962974584908221,
"learning_rate": 4.30745418417196e-06,
"loss": 0.5874,
"step": 906
},
{
"epoch": 0.795308817887872,
"grad_norm": 0.2894965323690623,
"learning_rate": 4.305819020277169e-06,
"loss": 0.5769,
"step": 907
},
{
"epoch": 0.7961856743574286,
"grad_norm": 0.2744231484838131,
"learning_rate": 4.304182239427992e-06,
"loss": 0.5943,
"step": 908
},
{
"epoch": 0.7970625308269853,
"grad_norm": 0.2766245048172803,
"learning_rate": 4.302543843090026e-06,
"loss": 0.5814,
"step": 909
},
{
"epoch": 0.7979393872965419,
"grad_norm": 0.2842673020480384,
"learning_rate": 4.30090383273031e-06,
"loss": 0.5912,
"step": 910
},
{
"epoch": 0.7988162437660985,
"grad_norm": 0.28199584242917014,
"learning_rate": 4.2992622098173335e-06,
"loss": 0.5809,
"step": 911
},
{
"epoch": 0.7996931002356552,
"grad_norm": 0.2820675876804688,
"learning_rate": 4.297618975821027e-06,
"loss": 0.5917,
"step": 912
},
{
"epoch": 0.8005699567052118,
"grad_norm": 0.2728605500328137,
"learning_rate": 4.2959741322127635e-06,
"loss": 0.5764,
"step": 913
},
{
"epoch": 0.8014468131747684,
"grad_norm": 0.27169399222059704,
"learning_rate": 4.294327680465358e-06,
"loss": 0.5849,
"step": 914
},
{
"epoch": 0.8023236696443251,
"grad_norm": 0.28063665744680427,
"learning_rate": 4.292679622053066e-06,
"loss": 0.58,
"step": 915
},
{
"epoch": 0.8032005261138817,
"grad_norm": 0.25926421536726935,
"learning_rate": 4.29102995845158e-06,
"loss": 0.5787,
"step": 916
},
{
"epoch": 0.8040773825834384,
"grad_norm": 0.29001417666592577,
"learning_rate": 4.289378691138032e-06,
"loss": 0.5868,
"step": 917
},
{
"epoch": 0.804954239052995,
"grad_norm": 0.27215185007216747,
"learning_rate": 4.287725821590987e-06,
"loss": 0.5894,
"step": 918
},
{
"epoch": 0.8058310955225516,
"grad_norm": 0.3050881231274449,
"learning_rate": 4.286071351290447e-06,
"loss": 0.5911,
"step": 919
},
{
"epoch": 0.8067079519921083,
"grad_norm": 0.2873456207891206,
"learning_rate": 4.2844152817178476e-06,
"loss": 0.5835,
"step": 920
},
{
"epoch": 0.8075848084616649,
"grad_norm": 0.2626365139918821,
"learning_rate": 4.282757614356055e-06,
"loss": 0.5794,
"step": 921
},
{
"epoch": 0.8084616649312216,
"grad_norm": 0.28122583577721894,
"learning_rate": 4.281098350689367e-06,
"loss": 0.581,
"step": 922
},
{
"epoch": 0.8093385214007782,
"grad_norm": 0.2955727164056087,
"learning_rate": 4.279437492203509e-06,
"loss": 0.6024,
"step": 923
},
{
"epoch": 0.8102153778703348,
"grad_norm": 0.2928465088558078,
"learning_rate": 4.277775040385636e-06,
"loss": 0.5777,
"step": 924
},
{
"epoch": 0.8110922343398915,
"grad_norm": 0.279748286657514,
"learning_rate": 4.276110996724332e-06,
"loss": 0.5983,
"step": 925
},
{
"epoch": 0.8119690908094481,
"grad_norm": 0.3064104243975942,
"learning_rate": 4.274445362709602e-06,
"loss": 0.5959,
"step": 926
},
{
"epoch": 0.8128459472790047,
"grad_norm": 0.2705400124701495,
"learning_rate": 4.272778139832876e-06,
"loss": 0.5964,
"step": 927
},
{
"epoch": 0.8137228037485614,
"grad_norm": 0.3030828027995252,
"learning_rate": 4.271109329587009e-06,
"loss": 0.5784,
"step": 928
},
{
"epoch": 0.814599660218118,
"grad_norm": 0.2629159770264448,
"learning_rate": 4.2694389334662745e-06,
"loss": 0.5845,
"step": 929
},
{
"epoch": 0.8154765166876747,
"grad_norm": 0.3351422353981342,
"learning_rate": 4.267766952966369e-06,
"loss": 0.5949,
"step": 930
},
{
"epoch": 0.8163533731572313,
"grad_norm": 0.2760441532769009,
"learning_rate": 4.2660933895844055e-06,
"loss": 0.5904,
"step": 931
},
{
"epoch": 0.8172302296267879,
"grad_norm": 0.30558832310943446,
"learning_rate": 4.264418244818914e-06,
"loss": 0.5839,
"step": 932
},
{
"epoch": 0.8181070860963446,
"grad_norm": 0.28070458613560756,
"learning_rate": 4.262741520169844e-06,
"loss": 0.5791,
"step": 933
},
{
"epoch": 0.8189839425659012,
"grad_norm": 0.2735766456330096,
"learning_rate": 4.261063217138554e-06,
"loss": 0.5836,
"step": 934
},
{
"epoch": 0.8198607990354578,
"grad_norm": 0.3038178849716158,
"learning_rate": 4.259383337227821e-06,
"loss": 0.5885,
"step": 935
},
{
"epoch": 0.8207376555050145,
"grad_norm": 0.26590487432268695,
"learning_rate": 4.25770188194183e-06,
"loss": 0.6035,
"step": 936
},
{
"epoch": 0.8216145119745711,
"grad_norm": 0.31271672720672494,
"learning_rate": 4.25601885278618e-06,
"loss": 0.5926,
"step": 937
},
{
"epoch": 0.8224913684441278,
"grad_norm": 0.26261561071530615,
"learning_rate": 4.254334251267877e-06,
"loss": 0.5996,
"step": 938
},
{
"epoch": 0.8233682249136844,
"grad_norm": 0.2891665251939073,
"learning_rate": 4.252648078895336e-06,
"loss": 0.5876,
"step": 939
},
{
"epoch": 0.824245081383241,
"grad_norm": 0.2897735311167941,
"learning_rate": 4.2509603371783776e-06,
"loss": 0.5892,
"step": 940
},
{
"epoch": 0.8251219378527977,
"grad_norm": 0.28026024666883764,
"learning_rate": 4.249271027628228e-06,
"loss": 0.587,
"step": 941
},
{
"epoch": 0.8259987943223543,
"grad_norm": 0.2765283292737123,
"learning_rate": 4.24758015175752e-06,
"loss": 0.5769,
"step": 942
},
{
"epoch": 0.826875650791911,
"grad_norm": 0.2921232680301083,
"learning_rate": 4.245887711080283e-06,
"loss": 0.5854,
"step": 943
},
{
"epoch": 0.8277525072614677,
"grad_norm": 0.3005072830624817,
"learning_rate": 4.2441937071119524e-06,
"loss": 0.5802,
"step": 944
},
{
"epoch": 0.8286293637310242,
"grad_norm": 0.27059131939602343,
"learning_rate": 4.242498141369361e-06,
"loss": 0.5837,
"step": 945
},
{
"epoch": 0.829506220200581,
"grad_norm": 0.3038588097565146,
"learning_rate": 4.240801015370743e-06,
"loss": 0.5869,
"step": 946
},
{
"epoch": 0.8303830766701376,
"grad_norm": 0.31875741653821127,
"learning_rate": 4.239102330635726e-06,
"loss": 0.5836,
"step": 947
},
{
"epoch": 0.8312599331396942,
"grad_norm": 0.26475770270890336,
"learning_rate": 4.2374020886853354e-06,
"loss": 0.5796,
"step": 948
},
{
"epoch": 0.8321367896092509,
"grad_norm": 0.31635648581412845,
"learning_rate": 4.235700291041989e-06,
"loss": 0.5732,
"step": 949
},
{
"epoch": 0.8330136460788075,
"grad_norm": 0.27123635854757305,
"learning_rate": 4.233996939229502e-06,
"loss": 0.5977,
"step": 950
},
{
"epoch": 0.8338905025483642,
"grad_norm": 0.3356358824197267,
"learning_rate": 4.232292034773076e-06,
"loss": 0.5871,
"step": 951
},
{
"epoch": 0.8347673590179208,
"grad_norm": 0.2723531290949244,
"learning_rate": 4.230585579199306e-06,
"loss": 0.5916,
"step": 952
},
{
"epoch": 0.8356442154874774,
"grad_norm": 0.2975424730057694,
"learning_rate": 4.228877574036175e-06,
"loss": 0.592,
"step": 953
},
{
"epoch": 0.8365210719570341,
"grad_norm": 0.28108527975014536,
"learning_rate": 4.227168020813053e-06,
"loss": 0.5788,
"step": 954
},
{
"epoch": 0.8373979284265907,
"grad_norm": 0.26358656072328285,
"learning_rate": 4.225456921060698e-06,
"loss": 0.5728,
"step": 955
},
{
"epoch": 0.8382747848961473,
"grad_norm": 0.2793044648839571,
"learning_rate": 4.223744276311249e-06,
"loss": 0.5714,
"step": 956
},
{
"epoch": 0.839151641365704,
"grad_norm": 0.30214577120239683,
"learning_rate": 4.222030088098233e-06,
"loss": 0.5993,
"step": 957
},
{
"epoch": 0.8400284978352606,
"grad_norm": 0.2639515397393347,
"learning_rate": 4.220314357956557e-06,
"loss": 0.5994,
"step": 958
},
{
"epoch": 0.8409053543048173,
"grad_norm": 0.3298154347341819,
"learning_rate": 4.218597087422508e-06,
"loss": 0.5877,
"step": 959
},
{
"epoch": 0.8417822107743739,
"grad_norm": 0.28203599665081885,
"learning_rate": 4.216878278033753e-06,
"loss": 0.5865,
"step": 960
},
{
"epoch": 0.8426590672439305,
"grad_norm": 0.2746406409148874,
"learning_rate": 4.2151579313293364e-06,
"loss": 0.5881,
"step": 961
},
{
"epoch": 0.8435359237134872,
"grad_norm": 0.33875497622714734,
"learning_rate": 4.2134360488496804e-06,
"loss": 0.6029,
"step": 962
},
{
"epoch": 0.8444127801830438,
"grad_norm": 0.2875141188036911,
"learning_rate": 4.211712632136581e-06,
"loss": 0.5845,
"step": 963
},
{
"epoch": 0.8452896366526004,
"grad_norm": 0.32374197566257723,
"learning_rate": 4.209987682733207e-06,
"loss": 0.589,
"step": 964
},
{
"epoch": 0.8461664931221571,
"grad_norm": 0.26718900480287466,
"learning_rate": 4.208261202184104e-06,
"loss": 0.5844,
"step": 965
},
{
"epoch": 0.8470433495917137,
"grad_norm": 0.29759515513279916,
"learning_rate": 4.206533192035184e-06,
"loss": 0.5817,
"step": 966
},
{
"epoch": 0.8479202060612704,
"grad_norm": 0.28330165664862006,
"learning_rate": 4.20480365383373e-06,
"loss": 0.5853,
"step": 967
},
{
"epoch": 0.848797062530827,
"grad_norm": 0.26991723910735316,
"learning_rate": 4.203072589128394e-06,
"loss": 0.5847,
"step": 968
},
{
"epoch": 0.8496739190003836,
"grad_norm": 0.28120405866784015,
"learning_rate": 4.201339999469194e-06,
"loss": 0.5771,
"step": 969
},
{
"epoch": 0.8505507754699403,
"grad_norm": 0.29731566030764794,
"learning_rate": 4.199605886407515e-06,
"loss": 0.5872,
"step": 970
},
{
"epoch": 0.8514276319394969,
"grad_norm": 0.29823098898704575,
"learning_rate": 4.197870251496104e-06,
"loss": 0.585,
"step": 971
},
{
"epoch": 0.8523044884090536,
"grad_norm": 0.29246400163730035,
"learning_rate": 4.196133096289071e-06,
"loss": 0.5728,
"step": 972
},
{
"epoch": 0.8531813448786102,
"grad_norm": 0.31038345035918974,
"learning_rate": 4.194394422341888e-06,
"loss": 0.588,
"step": 973
},
{
"epoch": 0.8540582013481668,
"grad_norm": 0.29419655403066824,
"learning_rate": 4.192654231211389e-06,
"loss": 0.5802,
"step": 974
},
{
"epoch": 0.8549350578177235,
"grad_norm": 0.28924212129082133,
"learning_rate": 4.190912524455762e-06,
"loss": 0.5957,
"step": 975
},
{
"epoch": 0.8558119142872801,
"grad_norm": 0.3433724407789192,
"learning_rate": 4.189169303634555e-06,
"loss": 0.5943,
"step": 976
},
{
"epoch": 0.8566887707568367,
"grad_norm": 0.3447246872111939,
"learning_rate": 4.187424570308671e-06,
"loss": 0.5679,
"step": 977
},
{
"epoch": 0.8575656272263934,
"grad_norm": 0.2717297839127488,
"learning_rate": 4.185678326040369e-06,
"loss": 0.5839,
"step": 978
},
{
"epoch": 0.85844248369595,
"grad_norm": 0.3149777108439808,
"learning_rate": 4.1839305723932565e-06,
"loss": 0.5684,
"step": 979
},
{
"epoch": 0.8593193401655067,
"grad_norm": 0.3196280126814673,
"learning_rate": 4.1821813109322975e-06,
"loss": 0.5845,
"step": 980
},
{
"epoch": 0.8601961966350633,
"grad_norm": 0.3166850113740036,
"learning_rate": 4.180430543223803e-06,
"loss": 0.5722,
"step": 981
},
{
"epoch": 0.8610730531046199,
"grad_norm": 0.30727325041845543,
"learning_rate": 4.178678270835435e-06,
"loss": 0.582,
"step": 982
},
{
"epoch": 0.8619499095741766,
"grad_norm": 0.34738075452538025,
"learning_rate": 4.1769244953361995e-06,
"loss": 0.5789,
"step": 983
},
{
"epoch": 0.8628267660437332,
"grad_norm": 0.3029018585056203,
"learning_rate": 4.1751692182964524e-06,
"loss": 0.5906,
"step": 984
},
{
"epoch": 0.8637036225132898,
"grad_norm": 0.27172806950560857,
"learning_rate": 4.1734124412878915e-06,
"loss": 0.5864,
"step": 985
},
{
"epoch": 0.8645804789828465,
"grad_norm": 0.3078626255245488,
"learning_rate": 4.171654165883558e-06,
"loss": 0.5961,
"step": 986
},
{
"epoch": 0.8654573354524031,
"grad_norm": 0.28755523271585887,
"learning_rate": 4.169894393657834e-06,
"loss": 0.5881,
"step": 987
},
{
"epoch": 0.8663341919219598,
"grad_norm": 0.3081436303822685,
"learning_rate": 4.168133126186445e-06,
"loss": 0.5818,
"step": 988
},
{
"epoch": 0.8672110483915164,
"grad_norm": 0.2785218381541765,
"learning_rate": 4.166370365046452e-06,
"loss": 0.5828,
"step": 989
},
{
"epoch": 0.868087904861073,
"grad_norm": 0.3391784184001714,
"learning_rate": 4.164606111816256e-06,
"loss": 0.5867,
"step": 990
},
{
"epoch": 0.8689647613306297,
"grad_norm": 0.27636992919331915,
"learning_rate": 4.162840368075591e-06,
"loss": 0.599,
"step": 991
},
{
"epoch": 0.8698416178001863,
"grad_norm": 0.28517927301055196,
"learning_rate": 4.161073135405529e-06,
"loss": 0.5831,
"step": 992
},
{
"epoch": 0.870718474269743,
"grad_norm": 0.29490820494014364,
"learning_rate": 4.1593044153884745e-06,
"loss": 0.5757,
"step": 993
},
{
"epoch": 0.8715953307392996,
"grad_norm": 0.2780476402469785,
"learning_rate": 4.157534209608161e-06,
"loss": 0.5964,
"step": 994
},
{
"epoch": 0.8724721872088562,
"grad_norm": 0.29068689725516644,
"learning_rate": 4.155762519649654e-06,
"loss": 0.5805,
"step": 995
},
{
"epoch": 0.8733490436784129,
"grad_norm": 0.26095614944942314,
"learning_rate": 4.15398934709935e-06,
"loss": 0.5841,
"step": 996
},
{
"epoch": 0.8742259001479695,
"grad_norm": 0.31389428529448765,
"learning_rate": 4.1522146935449705e-06,
"loss": 0.5846,
"step": 997
},
{
"epoch": 0.8751027566175261,
"grad_norm": 0.26816106638671405,
"learning_rate": 4.150438560575563e-06,
"loss": 0.5833,
"step": 998
},
{
"epoch": 0.8759796130870828,
"grad_norm": 0.31604277041792156,
"learning_rate": 4.1486609497815025e-06,
"loss": 0.5888,
"step": 999
},
{
"epoch": 0.8768564695566394,
"grad_norm": 0.3606037237047822,
"learning_rate": 4.146881862754485e-06,
"loss": 0.5942,
"step": 1000
},
{
"epoch": 0.8777333260261961,
"grad_norm": 0.28543513756367406,
"learning_rate": 4.145101301087527e-06,
"loss": 0.5915,
"step": 1001
},
{
"epoch": 0.8786101824957527,
"grad_norm": 0.3462271962536017,
"learning_rate": 4.143319266374969e-06,
"loss": 0.5942,
"step": 1002
},
{
"epoch": 0.8794870389653093,
"grad_norm": 0.2833352289445499,
"learning_rate": 4.141535760212467e-06,
"loss": 0.5863,
"step": 1003
},
{
"epoch": 0.880363895434866,
"grad_norm": 0.35489814354695126,
"learning_rate": 4.139750784196998e-06,
"loss": 0.5924,
"step": 1004
},
{
"epoch": 0.8812407519044226,
"grad_norm": 0.2942335535458572,
"learning_rate": 4.137964339926852e-06,
"loss": 0.5892,
"step": 1005
},
{
"epoch": 0.8821176083739792,
"grad_norm": 0.32828822885224784,
"learning_rate": 4.136176429001634e-06,
"loss": 0.5909,
"step": 1006
},
{
"epoch": 0.8829944648435359,
"grad_norm": 0.3123727759868493,
"learning_rate": 4.134387053022266e-06,
"loss": 0.5845,
"step": 1007
},
{
"epoch": 0.8838713213130925,
"grad_norm": 0.2862421766790686,
"learning_rate": 4.132596213590977e-06,
"loss": 0.5848,
"step": 1008
},
{
"epoch": 0.8847481777826492,
"grad_norm": 0.32232750817039807,
"learning_rate": 4.1308039123113084e-06,
"loss": 0.5869,
"step": 1009
},
{
"epoch": 0.8856250342522058,
"grad_norm": 0.28776404090006724,
"learning_rate": 4.129010150788112e-06,
"loss": 0.5992,
"step": 1010
},
{
"epoch": 0.8865018907217624,
"grad_norm": 0.3257967217812331,
"learning_rate": 4.127214930627545e-06,
"loss": 0.5828,
"step": 1011
},
{
"epoch": 0.8873787471913192,
"grad_norm": 0.3065300730664574,
"learning_rate": 4.125418253437071e-06,
"loss": 0.578,
"step": 1012
},
{
"epoch": 0.8882556036608757,
"grad_norm": 0.29218143100925903,
"learning_rate": 4.123620120825459e-06,
"loss": 0.5939,
"step": 1013
},
{
"epoch": 0.8891324601304323,
"grad_norm": 0.28565794045128473,
"learning_rate": 4.121820534402781e-06,
"loss": 0.5868,
"step": 1014
},
{
"epoch": 0.890009316599989,
"grad_norm": 0.30898296228273797,
"learning_rate": 4.120019495780412e-06,
"loss": 0.582,
"step": 1015
},
{
"epoch": 0.8908861730695457,
"grad_norm": 0.2911662733325922,
"learning_rate": 4.118217006571023e-06,
"loss": 0.5923,
"step": 1016
},
{
"epoch": 0.8917630295391024,
"grad_norm": 0.2843342810887561,
"learning_rate": 4.116413068388589e-06,
"loss": 0.5754,
"step": 1017
},
{
"epoch": 0.892639886008659,
"grad_norm": 0.334401955522752,
"learning_rate": 4.11460768284838e-06,
"loss": 0.5895,
"step": 1018
},
{
"epoch": 0.8935167424782156,
"grad_norm": 0.2600873368987441,
"learning_rate": 4.11280085156696e-06,
"loss": 0.5858,
"step": 1019
},
{
"epoch": 0.8943935989477723,
"grad_norm": 0.3051388251322737,
"learning_rate": 4.110992576162193e-06,
"loss": 0.5861,
"step": 1020
},
{
"epoch": 0.8952704554173289,
"grad_norm": 0.30230682759222505,
"learning_rate": 4.109182858253231e-06,
"loss": 0.5857,
"step": 1021
},
{
"epoch": 0.8961473118868856,
"grad_norm": 0.27145584987414345,
"learning_rate": 4.107371699460521e-06,
"loss": 0.5827,
"step": 1022
},
{
"epoch": 0.8970241683564422,
"grad_norm": 0.2886096599363367,
"learning_rate": 4.1055591014057964e-06,
"loss": 0.5732,
"step": 1023
},
{
"epoch": 0.8979010248259988,
"grad_norm": 0.2643618798342576,
"learning_rate": 4.103745065712083e-06,
"loss": 0.581,
"step": 1024
},
{
"epoch": 0.8987778812955555,
"grad_norm": 0.27612674007258925,
"learning_rate": 4.101929594003694e-06,
"loss": 0.5774,
"step": 1025
},
{
"epoch": 0.8996547377651121,
"grad_norm": 0.2694404941538916,
"learning_rate": 4.100112687906224e-06,
"loss": 0.5792,
"step": 1026
},
{
"epoch": 0.9005315942346687,
"grad_norm": 0.26812897420311116,
"learning_rate": 4.098294349046556e-06,
"loss": 0.5945,
"step": 1027
},
{
"epoch": 0.9014084507042254,
"grad_norm": 0.2744007605554886,
"learning_rate": 4.0964745790528564e-06,
"loss": 0.5712,
"step": 1028
},
{
"epoch": 0.902285307173782,
"grad_norm": 0.2614641549143825,
"learning_rate": 4.09465337955457e-06,
"loss": 0.5756,
"step": 1029
},
{
"epoch": 0.9031621636433387,
"grad_norm": 0.25643605179903173,
"learning_rate": 4.092830752182423e-06,
"loss": 0.593,
"step": 1030
},
{
"epoch": 0.9040390201128953,
"grad_norm": 0.26698048225450505,
"learning_rate": 4.091006698568419e-06,
"loss": 0.5877,
"step": 1031
},
{
"epoch": 0.9049158765824519,
"grad_norm": 0.2655671129093472,
"learning_rate": 4.0891812203458425e-06,
"loss": 0.5701,
"step": 1032
},
{
"epoch": 0.9057927330520086,
"grad_norm": 0.2706223562384906,
"learning_rate": 4.08735431914925e-06,
"loss": 0.5818,
"step": 1033
},
{
"epoch": 0.9066695895215652,
"grad_norm": 0.26684323937974636,
"learning_rate": 4.085525996614472e-06,
"loss": 0.5878,
"step": 1034
},
{
"epoch": 0.9075464459911218,
"grad_norm": 0.24564951471442678,
"learning_rate": 4.083696254378615e-06,
"loss": 0.5967,
"step": 1035
},
{
"epoch": 0.9084233024606785,
"grad_norm": 0.2761933648093443,
"learning_rate": 4.081865094080053e-06,
"loss": 0.576,
"step": 1036
},
{
"epoch": 0.9093001589302351,
"grad_norm": 0.2722027493749199,
"learning_rate": 4.080032517358431e-06,
"loss": 0.579,
"step": 1037
},
{
"epoch": 0.9101770153997918,
"grad_norm": 0.5039307385586534,
"learning_rate": 4.078198525854664e-06,
"loss": 0.5943,
"step": 1038
},
{
"epoch": 0.9110538718693484,
"grad_norm": 0.26519176650439175,
"learning_rate": 4.0763631212109315e-06,
"loss": 0.5893,
"step": 1039
},
{
"epoch": 0.911930728338905,
"grad_norm": 0.2644411261920598,
"learning_rate": 4.074526305070679e-06,
"loss": 0.5791,
"step": 1040
},
{
"epoch": 0.9128075848084617,
"grad_norm": 0.27917354228958563,
"learning_rate": 4.072688079078616e-06,
"loss": 0.5847,
"step": 1041
},
{
"epoch": 0.9136844412780183,
"grad_norm": 0.27274252297201695,
"learning_rate": 4.070848444880716e-06,
"loss": 0.5695,
"step": 1042
},
{
"epoch": 0.914561297747575,
"grad_norm": 0.26541238057197397,
"learning_rate": 4.06900740412421e-06,
"loss": 0.5858,
"step": 1043
},
{
"epoch": 0.9154381542171316,
"grad_norm": 0.2687466193673103,
"learning_rate": 4.0671649584575925e-06,
"loss": 0.5832,
"step": 1044
},
{
"epoch": 0.9163150106866882,
"grad_norm": 0.27584447196087264,
"learning_rate": 4.065321109530612e-06,
"loss": 0.5828,
"step": 1045
},
{
"epoch": 0.9171918671562449,
"grad_norm": 0.27618254494046185,
"learning_rate": 4.063475858994276e-06,
"loss": 0.5829,
"step": 1046
},
{
"epoch": 0.9180687236258015,
"grad_norm": 0.2800627797716068,
"learning_rate": 4.061629208500847e-06,
"loss": 0.5813,
"step": 1047
},
{
"epoch": 0.9189455800953581,
"grad_norm": 0.2731973027581407,
"learning_rate": 4.059781159703839e-06,
"loss": 0.5907,
"step": 1048
},
{
"epoch": 0.9198224365649148,
"grad_norm": 0.2817329916742434,
"learning_rate": 4.057931714258022e-06,
"loss": 0.5845,
"step": 1049
},
{
"epoch": 0.9206992930344714,
"grad_norm": 0.2624010665247189,
"learning_rate": 4.056080873819412e-06,
"loss": 0.579,
"step": 1050
},
{
"epoch": 0.9215761495040281,
"grad_norm": 0.26121937584936983,
"learning_rate": 4.054228640045275e-06,
"loss": 0.5857,
"step": 1051
},
{
"epoch": 0.9224530059735847,
"grad_norm": 0.2832895486337394,
"learning_rate": 4.052375014594129e-06,
"loss": 0.5957,
"step": 1052
},
{
"epoch": 0.9233298624431413,
"grad_norm": 0.27671228904328893,
"learning_rate": 4.0505199991257325e-06,
"loss": 0.5791,
"step": 1053
},
{
"epoch": 0.924206718912698,
"grad_norm": 0.266998502123574,
"learning_rate": 4.048663595301093e-06,
"loss": 0.5896,
"step": 1054
},
{
"epoch": 0.9250835753822546,
"grad_norm": 0.3094016546060802,
"learning_rate": 4.046805804782456e-06,
"loss": 0.5788,
"step": 1055
},
{
"epoch": 0.9259604318518112,
"grad_norm": 0.2782662002801493,
"learning_rate": 4.0449466292333166e-06,
"loss": 0.5888,
"step": 1056
},
{
"epoch": 0.9268372883213679,
"grad_norm": 0.27821869081922773,
"learning_rate": 4.043086070318401e-06,
"loss": 0.5879,
"step": 1057
},
{
"epoch": 0.9277141447909245,
"grad_norm": 0.32143887759720546,
"learning_rate": 4.04122412970368e-06,
"loss": 0.5884,
"step": 1058
},
{
"epoch": 0.9285910012604812,
"grad_norm": 0.2598221780539352,
"learning_rate": 4.039360809056361e-06,
"loss": 0.58,
"step": 1059
},
{
"epoch": 0.9294678577300378,
"grad_norm": 0.3300275262996093,
"learning_rate": 4.037496110044885e-06,
"loss": 0.5963,
"step": 1060
},
{
"epoch": 0.9303447141995944,
"grad_norm": 0.2723517740568475,
"learning_rate": 4.035630034338928e-06,
"loss": 0.5684,
"step": 1061
},
{
"epoch": 0.9312215706691511,
"grad_norm": 0.26174388908838997,
"learning_rate": 4.033762583609398e-06,
"loss": 0.5741,
"step": 1062
},
{
"epoch": 0.9320984271387077,
"grad_norm": 0.2879705808043353,
"learning_rate": 4.031893759528439e-06,
"loss": 0.5651,
"step": 1063
},
{
"epoch": 0.9329752836082644,
"grad_norm": 0.27573911638107307,
"learning_rate": 4.030023563769418e-06,
"loss": 0.5738,
"step": 1064
},
{
"epoch": 0.933852140077821,
"grad_norm": 0.270890009890323,
"learning_rate": 4.028151998006934e-06,
"loss": 0.5748,
"step": 1065
},
{
"epoch": 0.9347289965473776,
"grad_norm": 0.2651359065699047,
"learning_rate": 4.026279063916811e-06,
"loss": 0.5815,
"step": 1066
},
{
"epoch": 0.9356058530169343,
"grad_norm": 0.285792627094006,
"learning_rate": 4.024404763176101e-06,
"loss": 0.5714,
"step": 1067
},
{
"epoch": 0.9364827094864909,
"grad_norm": 0.25220096965602506,
"learning_rate": 4.022529097463076e-06,
"loss": 0.5761,
"step": 1068
},
{
"epoch": 0.9373595659560475,
"grad_norm": 0.2572736434059626,
"learning_rate": 4.020652068457234e-06,
"loss": 0.5813,
"step": 1069
},
{
"epoch": 0.9382364224256042,
"grad_norm": 0.2769717174034421,
"learning_rate": 4.018773677839289e-06,
"loss": 0.5902,
"step": 1070
},
{
"epoch": 0.9391132788951608,
"grad_norm": 0.2638965107730823,
"learning_rate": 4.016893927291179e-06,
"loss": 0.5774,
"step": 1071
},
{
"epoch": 0.9399901353647175,
"grad_norm": 0.26364544697361064,
"learning_rate": 4.015012818496057e-06,
"loss": 0.5885,
"step": 1072
},
{
"epoch": 0.9408669918342741,
"grad_norm": 0.2782490552191973,
"learning_rate": 4.013130353138293e-06,
"loss": 0.5734,
"step": 1073
},
{
"epoch": 0.9417438483038307,
"grad_norm": 0.2939309170345373,
"learning_rate": 4.011246532903472e-06,
"loss": 0.5863,
"step": 1074
},
{
"epoch": 0.9426207047733874,
"grad_norm": 0.27682818038097917,
"learning_rate": 4.00936135947839e-06,
"loss": 0.5878,
"step": 1075
},
{
"epoch": 0.943497561242944,
"grad_norm": 0.27100650217384786,
"learning_rate": 4.007474834551059e-06,
"loss": 0.5788,
"step": 1076
},
{
"epoch": 0.9443744177125006,
"grad_norm": 0.3179264915740243,
"learning_rate": 4.005586959810697e-06,
"loss": 0.5697,
"step": 1077
},
{
"epoch": 0.9452512741820573,
"grad_norm": 0.26927348365153236,
"learning_rate": 4.003697736947731e-06,
"loss": 0.5683,
"step": 1078
},
{
"epoch": 0.9461281306516139,
"grad_norm": 0.2755764124341007,
"learning_rate": 4.001807167653798e-06,
"loss": 0.5794,
"step": 1079
},
{
"epoch": 0.9470049871211706,
"grad_norm": 0.2908090312996085,
"learning_rate": 3.999915253621739e-06,
"loss": 0.586,
"step": 1080
},
{
"epoch": 0.9478818435907272,
"grad_norm": 0.2545666408606057,
"learning_rate": 3.998021996545599e-06,
"loss": 0.5831,
"step": 1081
},
{
"epoch": 0.9487587000602838,
"grad_norm": 0.29377943743323887,
"learning_rate": 3.9961273981206245e-06,
"loss": 0.585,
"step": 1082
},
{
"epoch": 0.9496355565298406,
"grad_norm": 0.26968750170325856,
"learning_rate": 3.994231460043265e-06,
"loss": 0.5782,
"step": 1083
},
{
"epoch": 0.9505124129993971,
"grad_norm": 0.2911018694543167,
"learning_rate": 3.9923341840111675e-06,
"loss": 0.5813,
"step": 1084
},
{
"epoch": 0.9513892694689537,
"grad_norm": 0.32080813736390973,
"learning_rate": 3.99043557172318e-06,
"loss": 0.5836,
"step": 1085
},
{
"epoch": 0.9522661259385105,
"grad_norm": 0.2894185491332872,
"learning_rate": 3.988535624879344e-06,
"loss": 0.583,
"step": 1086
},
{
"epoch": 0.953142982408067,
"grad_norm": 0.3036439907360394,
"learning_rate": 3.986634345180899e-06,
"loss": 0.5753,
"step": 1087
},
{
"epoch": 0.9540198388776238,
"grad_norm": 0.30256015219807453,
"learning_rate": 3.984731734330273e-06,
"loss": 0.5787,
"step": 1088
},
{
"epoch": 0.9548966953471804,
"grad_norm": 0.2684694121785645,
"learning_rate": 3.982827794031091e-06,
"loss": 0.5811,
"step": 1089
},
{
"epoch": 0.955773551816737,
"grad_norm": 0.3047268297869491,
"learning_rate": 3.980922525988167e-06,
"loss": 0.5757,
"step": 1090
},
{
"epoch": 0.9566504082862937,
"grad_norm": 0.2680829692432763,
"learning_rate": 3.979015931907502e-06,
"loss": 0.5938,
"step": 1091
},
{
"epoch": 0.9575272647558503,
"grad_norm": 0.28352806229638294,
"learning_rate": 3.977108013496286e-06,
"loss": 0.5648,
"step": 1092
},
{
"epoch": 0.958404121225407,
"grad_norm": 0.27134893274934896,
"learning_rate": 3.975198772462896e-06,
"loss": 0.5959,
"step": 1093
},
{
"epoch": 0.9592809776949636,
"grad_norm": 0.27670636726963027,
"learning_rate": 3.973288210516889e-06,
"loss": 0.5825,
"step": 1094
},
{
"epoch": 0.9601578341645202,
"grad_norm": 0.27577855913411087,
"learning_rate": 3.971376329369011e-06,
"loss": 0.5763,
"step": 1095
},
{
"epoch": 0.9610346906340769,
"grad_norm": 0.2613562238768912,
"learning_rate": 3.969463130731183e-06,
"loss": 0.587,
"step": 1096
},
{
"epoch": 0.9619115471036335,
"grad_norm": 0.30682832359084977,
"learning_rate": 3.96754861631651e-06,
"loss": 0.6012,
"step": 1097
},
{
"epoch": 0.9627884035731901,
"grad_norm": 0.2753727317824162,
"learning_rate": 3.965632787839274e-06,
"loss": 0.593,
"step": 1098
},
{
"epoch": 0.9636652600427468,
"grad_norm": 0.2896526629743159,
"learning_rate": 3.963715647014932e-06,
"loss": 0.5823,
"step": 1099
},
{
"epoch": 0.9645421165123034,
"grad_norm": 0.28810606366408137,
"learning_rate": 3.961797195560118e-06,
"loss": 0.5844,
"step": 1100
},
{
"epoch": 0.9654189729818601,
"grad_norm": 0.2603559754869869,
"learning_rate": 3.959877435192639e-06,
"loss": 0.5803,
"step": 1101
},
{
"epoch": 0.9662958294514167,
"grad_norm": 0.28655269690518276,
"learning_rate": 3.957956367631475e-06,
"loss": 0.5707,
"step": 1102
},
{
"epoch": 0.9671726859209733,
"grad_norm": 0.3009451530592475,
"learning_rate": 3.956033994596773e-06,
"loss": 0.5771,
"step": 1103
},
{
"epoch": 0.96804954239053,
"grad_norm": 0.2577540703327921,
"learning_rate": 3.954110317809854e-06,
"loss": 0.576,
"step": 1104
},
{
"epoch": 0.9689263988600866,
"grad_norm": 0.29870257898995317,
"learning_rate": 3.952185338993202e-06,
"loss": 0.5872,
"step": 1105
},
{
"epoch": 0.9698032553296432,
"grad_norm": 0.2768702174324288,
"learning_rate": 3.95025905987047e-06,
"loss": 0.5831,
"step": 1106
},
{
"epoch": 0.9706801117991999,
"grad_norm": 0.288774627238478,
"learning_rate": 3.948331482166473e-06,
"loss": 0.5951,
"step": 1107
},
{
"epoch": 0.9715569682687565,
"grad_norm": 0.324678524263679,
"learning_rate": 3.94640260760719e-06,
"loss": 0.5734,
"step": 1108
},
{
"epoch": 0.9724338247383132,
"grad_norm": 0.2777093036856744,
"learning_rate": 3.944472437919761e-06,
"loss": 0.5846,
"step": 1109
},
{
"epoch": 0.9733106812078698,
"grad_norm": 0.337073965677139,
"learning_rate": 3.942540974832486e-06,
"loss": 0.5904,
"step": 1110
},
{
"epoch": 0.9741875376774264,
"grad_norm": 0.2919504390486104,
"learning_rate": 3.9406082200748216e-06,
"loss": 0.5901,
"step": 1111
},
{
"epoch": 0.9750643941469831,
"grad_norm": 0.26917415244282195,
"learning_rate": 3.938674175377383e-06,
"loss": 0.5727,
"step": 1112
},
{
"epoch": 0.9759412506165397,
"grad_norm": 0.2968354712585106,
"learning_rate": 3.93673884247194e-06,
"loss": 0.5684,
"step": 1113
},
{
"epoch": 0.9768181070860964,
"grad_norm": 0.26666333819741744,
"learning_rate": 3.934802223091415e-06,
"loss": 0.582,
"step": 1114
},
{
"epoch": 0.977694963555653,
"grad_norm": 0.2648009228041306,
"learning_rate": 3.932864318969882e-06,
"loss": 0.5732,
"step": 1115
},
{
"epoch": 0.9785718200252096,
"grad_norm": 0.26447715765911384,
"learning_rate": 3.930925131842567e-06,
"loss": 0.581,
"step": 1116
},
{
"epoch": 0.9794486764947663,
"grad_norm": 0.26650421292261106,
"learning_rate": 3.928984663445844e-06,
"loss": 0.578,
"step": 1117
},
{
"epoch": 0.9803255329643229,
"grad_norm": 0.27399427740484344,
"learning_rate": 3.927042915517234e-06,
"loss": 0.5841,
"step": 1118
},
{
"epoch": 0.9812023894338795,
"grad_norm": 0.29486187077568676,
"learning_rate": 3.925099889795404e-06,
"loss": 0.5791,
"step": 1119
},
{
"epoch": 0.9820792459034362,
"grad_norm": 0.27626862187200796,
"learning_rate": 3.9231555880201655e-06,
"loss": 0.5758,
"step": 1120
},
{
"epoch": 0.9829561023729928,
"grad_norm": 0.2709394700881976,
"learning_rate": 3.9212100119324704e-06,
"loss": 0.5725,
"step": 1121
},
{
"epoch": 0.9838329588425495,
"grad_norm": 0.257787971984586,
"learning_rate": 3.919263163274416e-06,
"loss": 0.5733,
"step": 1122
},
{
"epoch": 0.9847098153121061,
"grad_norm": 0.2854496376494655,
"learning_rate": 3.917315043789235e-06,
"loss": 0.5696,
"step": 1123
},
{
"epoch": 0.9855866717816627,
"grad_norm": 0.2566199610678738,
"learning_rate": 3.9153656552212995e-06,
"loss": 0.5813,
"step": 1124
},
{
"epoch": 0.9864635282512194,
"grad_norm": 0.2555880030988225,
"learning_rate": 3.913414999316118e-06,
"loss": 0.5945,
"step": 1125
},
{
"epoch": 0.987340384720776,
"grad_norm": 0.2577195559469773,
"learning_rate": 3.911463077820336e-06,
"loss": 0.5675,
"step": 1126
},
{
"epoch": 0.9882172411903326,
"grad_norm": 0.26851748898394834,
"learning_rate": 3.909509892481726e-06,
"loss": 0.5807,
"step": 1127
},
{
"epoch": 0.9890940976598893,
"grad_norm": 0.2617539578196299,
"learning_rate": 3.907555445049198e-06,
"loss": 0.5684,
"step": 1128
},
{
"epoch": 0.9899709541294459,
"grad_norm": 0.2586839170532308,
"learning_rate": 3.905599737272791e-06,
"loss": 0.5801,
"step": 1129
},
{
"epoch": 0.9908478105990026,
"grad_norm": 0.25049955800874396,
"learning_rate": 3.903642770903671e-06,
"loss": 0.5762,
"step": 1130
},
{
"epoch": 0.9917246670685592,
"grad_norm": 0.27270516361418773,
"learning_rate": 3.901684547694133e-06,
"loss": 0.5878,
"step": 1131
},
{
"epoch": 0.9926015235381158,
"grad_norm": 0.2816673997379789,
"learning_rate": 3.899725069397593e-06,
"loss": 0.5927,
"step": 1132
},
{
"epoch": 0.9934783800076725,
"grad_norm": 0.2679288547921494,
"learning_rate": 3.897764337768597e-06,
"loss": 0.5772,
"step": 1133
},
{
"epoch": 0.9943552364772291,
"grad_norm": 0.27040765991438753,
"learning_rate": 3.895802354562808e-06,
"loss": 0.5623,
"step": 1134
},
{
"epoch": 0.9952320929467857,
"grad_norm": 0.29605913619532825,
"learning_rate": 3.893839121537015e-06,
"loss": 0.5868,
"step": 1135
},
{
"epoch": 0.9961089494163424,
"grad_norm": 0.27461413478738583,
"learning_rate": 3.89187464044912e-06,
"loss": 0.5871,
"step": 1136
},
{
"epoch": 0.996985805885899,
"grad_norm": 0.28648748056684925,
"learning_rate": 3.8899089130581465e-06,
"loss": 0.5753,
"step": 1137
},
{
"epoch": 0.9978626623554557,
"grad_norm": 0.2925165297373746,
"learning_rate": 3.8879419411242335e-06,
"loss": 0.5828,
"step": 1138
},
{
"epoch": 0.9987395188250123,
"grad_norm": 0.29352029461564516,
"learning_rate": 3.885973726408634e-06,
"loss": 0.5842,
"step": 1139
},
{
"epoch": 0.9996163752945689,
"grad_norm": 0.28650442615475913,
"learning_rate": 3.884004270673711e-06,
"loss": 0.5803,
"step": 1140
}
],
"logging_steps": 1,
"max_steps": 3420,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 1140,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 3818875539947520.0,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}