v0.8.18-adapter / 1100 /trainer_state.json
gotzmann's picture
..
070ed0c
raw
history blame
176 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.0036429872495445,
"eval_steps": 500,
"global_step": 1100,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0,
"grad_norm": 0.46431864432643544,
"learning_rate": 1.2121212121212122e-06,
"loss": 1.4151,
"step": 1
},
{
"epoch": 0.0,
"grad_norm": 0.42969176658539837,
"learning_rate": 2.4242424242424244e-06,
"loss": 1.3729,
"step": 2
},
{
"epoch": 0.01,
"grad_norm": 0.5004307936270223,
"learning_rate": 3.636363636363636e-06,
"loss": 1.3989,
"step": 3
},
{
"epoch": 0.01,
"grad_norm": 0.43666634920041486,
"learning_rate": 4.848484848484849e-06,
"loss": 1.3363,
"step": 4
},
{
"epoch": 0.01,
"grad_norm": 0.4691114419353825,
"learning_rate": 6.060606060606061e-06,
"loss": 1.4293,
"step": 5
},
{
"epoch": 0.01,
"grad_norm": 0.4277596061377729,
"learning_rate": 7.272727272727272e-06,
"loss": 1.4343,
"step": 6
},
{
"epoch": 0.01,
"grad_norm": 0.4238339229382504,
"learning_rate": 8.484848484848486e-06,
"loss": 1.4462,
"step": 7
},
{
"epoch": 0.01,
"grad_norm": 0.40992048534183273,
"learning_rate": 9.696969696969698e-06,
"loss": 1.2756,
"step": 8
},
{
"epoch": 0.02,
"grad_norm": 0.37885700313540693,
"learning_rate": 1.0909090909090909e-05,
"loss": 1.3464,
"step": 9
},
{
"epoch": 0.02,
"grad_norm": 0.36066141170123023,
"learning_rate": 1.2121212121212122e-05,
"loss": 1.3419,
"step": 10
},
{
"epoch": 0.02,
"grad_norm": 0.35617169386863406,
"learning_rate": 1.3333333333333333e-05,
"loss": 1.3533,
"step": 11
},
{
"epoch": 0.02,
"grad_norm": 0.3040188564782602,
"learning_rate": 1.4545454545454545e-05,
"loss": 1.2395,
"step": 12
},
{
"epoch": 0.02,
"grad_norm": 0.31038319439216566,
"learning_rate": 1.5757575757575756e-05,
"loss": 1.3082,
"step": 13
},
{
"epoch": 0.03,
"grad_norm": 0.26683768372135835,
"learning_rate": 1.6969696969696972e-05,
"loss": 1.3063,
"step": 14
},
{
"epoch": 0.03,
"grad_norm": 0.3652323682563078,
"learning_rate": 1.8181818181818182e-05,
"loss": 1.3045,
"step": 15
},
{
"epoch": 0.03,
"grad_norm": 0.23559121485457843,
"learning_rate": 1.9393939393939395e-05,
"loss": 1.2366,
"step": 16
},
{
"epoch": 0.03,
"grad_norm": 0.2342299313020104,
"learning_rate": 2.0606060606060608e-05,
"loss": 1.2831,
"step": 17
},
{
"epoch": 0.03,
"grad_norm": 0.2202931700357255,
"learning_rate": 2.1818181818181818e-05,
"loss": 1.3064,
"step": 18
},
{
"epoch": 0.03,
"grad_norm": 0.2097660599292375,
"learning_rate": 2.3030303030303034e-05,
"loss": 1.2376,
"step": 19
},
{
"epoch": 0.04,
"grad_norm": 0.2356785314652122,
"learning_rate": 2.4242424242424244e-05,
"loss": 1.2802,
"step": 20
},
{
"epoch": 0.04,
"grad_norm": 0.24639302530564244,
"learning_rate": 2.5454545454545454e-05,
"loss": 1.3016,
"step": 21
},
{
"epoch": 0.04,
"grad_norm": 0.24373126133228787,
"learning_rate": 2.6666666666666667e-05,
"loss": 1.3407,
"step": 22
},
{
"epoch": 0.04,
"grad_norm": 0.24488805144123432,
"learning_rate": 2.7878787878787883e-05,
"loss": 1.3325,
"step": 23
},
{
"epoch": 0.04,
"grad_norm": 0.2653033507571198,
"learning_rate": 2.909090909090909e-05,
"loss": 1.2811,
"step": 24
},
{
"epoch": 0.05,
"grad_norm": 1.2841724819336817,
"learning_rate": 3.0303030303030306e-05,
"loss": 1.2837,
"step": 25
},
{
"epoch": 0.05,
"grad_norm": 0.2183883020111492,
"learning_rate": 3.151515151515151e-05,
"loss": 1.2472,
"step": 26
},
{
"epoch": 0.05,
"grad_norm": 0.2137995163762026,
"learning_rate": 3.272727272727273e-05,
"loss": 1.2854,
"step": 27
},
{
"epoch": 0.05,
"grad_norm": 0.19499006223503876,
"learning_rate": 3.3939393939393945e-05,
"loss": 1.3018,
"step": 28
},
{
"epoch": 0.05,
"grad_norm": 0.17367919355340256,
"learning_rate": 3.515151515151515e-05,
"loss": 1.2824,
"step": 29
},
{
"epoch": 0.05,
"grad_norm": 0.18326045693683557,
"learning_rate": 3.6363636363636364e-05,
"loss": 1.2192,
"step": 30
},
{
"epoch": 0.06,
"grad_norm": 0.17474388188066411,
"learning_rate": 3.757575757575758e-05,
"loss": 1.2078,
"step": 31
},
{
"epoch": 0.06,
"grad_norm": 0.17856970178098716,
"learning_rate": 3.878787878787879e-05,
"loss": 1.2683,
"step": 32
},
{
"epoch": 0.06,
"grad_norm": 0.18617589704298348,
"learning_rate": 4e-05,
"loss": 1.2265,
"step": 33
},
{
"epoch": 0.06,
"grad_norm": 0.17653209733215317,
"learning_rate": 4.1212121212121216e-05,
"loss": 1.319,
"step": 34
},
{
"epoch": 0.06,
"grad_norm": 0.1722921367585233,
"learning_rate": 4.242424242424243e-05,
"loss": 1.2117,
"step": 35
},
{
"epoch": 0.07,
"grad_norm": 0.176642606378719,
"learning_rate": 4.3636363636363636e-05,
"loss": 1.2512,
"step": 36
},
{
"epoch": 0.07,
"grad_norm": 0.16696442324691066,
"learning_rate": 4.484848484848485e-05,
"loss": 1.2637,
"step": 37
},
{
"epoch": 0.07,
"grad_norm": 0.17035384059517106,
"learning_rate": 4.606060606060607e-05,
"loss": 1.2699,
"step": 38
},
{
"epoch": 0.07,
"grad_norm": 0.15545801881444482,
"learning_rate": 4.7272727272727275e-05,
"loss": 1.2939,
"step": 39
},
{
"epoch": 0.07,
"grad_norm": 0.17111439344347512,
"learning_rate": 4.848484848484849e-05,
"loss": 1.3033,
"step": 40
},
{
"epoch": 0.07,
"grad_norm": 0.16994151343455458,
"learning_rate": 4.9696969696969694e-05,
"loss": 1.2603,
"step": 41
},
{
"epoch": 0.08,
"grad_norm": 0.15929214926453447,
"learning_rate": 5.090909090909091e-05,
"loss": 1.2626,
"step": 42
},
{
"epoch": 0.08,
"grad_norm": 0.16761261516238699,
"learning_rate": 5.212121212121213e-05,
"loss": 1.296,
"step": 43
},
{
"epoch": 0.08,
"grad_norm": 0.15754700542426123,
"learning_rate": 5.333333333333333e-05,
"loss": 1.278,
"step": 44
},
{
"epoch": 0.08,
"grad_norm": 0.15522526683877644,
"learning_rate": 5.4545454545454546e-05,
"loss": 1.2355,
"step": 45
},
{
"epoch": 0.08,
"grad_norm": 0.1577929926930023,
"learning_rate": 5.5757575757575766e-05,
"loss": 1.2879,
"step": 46
},
{
"epoch": 0.09,
"grad_norm": 0.31075066632858317,
"learning_rate": 5.696969696969697e-05,
"loss": 1.2202,
"step": 47
},
{
"epoch": 0.09,
"grad_norm": 0.1663780653395111,
"learning_rate": 5.818181818181818e-05,
"loss": 1.2319,
"step": 48
},
{
"epoch": 0.09,
"grad_norm": 0.16049499655883026,
"learning_rate": 5.93939393939394e-05,
"loss": 1.2801,
"step": 49
},
{
"epoch": 0.09,
"grad_norm": 0.14515773124436285,
"learning_rate": 6.060606060606061e-05,
"loss": 1.2588,
"step": 50
},
{
"epoch": 0.09,
"grad_norm": 0.14653064850325623,
"learning_rate": 6.181818181818182e-05,
"loss": 1.2677,
"step": 51
},
{
"epoch": 0.09,
"grad_norm": 0.17193239746689878,
"learning_rate": 6.303030303030302e-05,
"loss": 1.2742,
"step": 52
},
{
"epoch": 0.1,
"grad_norm": 0.1967020450342533,
"learning_rate": 6.424242424242424e-05,
"loss": 1.1545,
"step": 53
},
{
"epoch": 0.1,
"grad_norm": 0.16247531997247225,
"learning_rate": 6.545454545454546e-05,
"loss": 1.222,
"step": 54
},
{
"epoch": 0.1,
"grad_norm": 0.14990706377244528,
"learning_rate": 6.666666666666667e-05,
"loss": 1.2103,
"step": 55
},
{
"epoch": 0.1,
"grad_norm": 0.1412817445239095,
"learning_rate": 6.787878787878789e-05,
"loss": 1.2169,
"step": 56
},
{
"epoch": 0.1,
"grad_norm": 0.14575971073482757,
"learning_rate": 6.90909090909091e-05,
"loss": 1.2751,
"step": 57
},
{
"epoch": 0.11,
"grad_norm": 0.13714747569950891,
"learning_rate": 7.03030303030303e-05,
"loss": 1.2508,
"step": 58
},
{
"epoch": 0.11,
"grad_norm": 0.14334695156859903,
"learning_rate": 7.151515151515152e-05,
"loss": 1.2721,
"step": 59
},
{
"epoch": 0.11,
"grad_norm": 0.1456824177522916,
"learning_rate": 7.272727272727273e-05,
"loss": 1.2649,
"step": 60
},
{
"epoch": 0.11,
"grad_norm": 0.15030318240210044,
"learning_rate": 7.393939393939395e-05,
"loss": 1.2167,
"step": 61
},
{
"epoch": 0.11,
"grad_norm": 0.1651326066719482,
"learning_rate": 7.515151515151515e-05,
"loss": 1.3126,
"step": 62
},
{
"epoch": 0.11,
"grad_norm": 0.1408250406479118,
"learning_rate": 7.636363636363637e-05,
"loss": 1.2891,
"step": 63
},
{
"epoch": 0.12,
"grad_norm": 0.21501384376905694,
"learning_rate": 7.757575757575758e-05,
"loss": 1.3019,
"step": 64
},
{
"epoch": 0.12,
"grad_norm": 0.1365168726167339,
"learning_rate": 7.878787878787879e-05,
"loss": 1.2498,
"step": 65
},
{
"epoch": 0.12,
"grad_norm": 0.1431463689660936,
"learning_rate": 8e-05,
"loss": 1.2793,
"step": 66
},
{
"epoch": 0.12,
"grad_norm": 0.13689045214286194,
"learning_rate": 8.121212121212121e-05,
"loss": 1.2295,
"step": 67
},
{
"epoch": 0.12,
"grad_norm": 0.13483608710081227,
"learning_rate": 8.242424242424243e-05,
"loss": 1.2258,
"step": 68
},
{
"epoch": 0.13,
"grad_norm": 0.13707618564415613,
"learning_rate": 8.363636363636364e-05,
"loss": 1.2252,
"step": 69
},
{
"epoch": 0.13,
"grad_norm": 0.13780236215967515,
"learning_rate": 8.484848484848486e-05,
"loss": 1.2565,
"step": 70
},
{
"epoch": 0.13,
"grad_norm": 0.14036805493494423,
"learning_rate": 8.606060606060606e-05,
"loss": 1.3023,
"step": 71
},
{
"epoch": 0.13,
"grad_norm": 0.12776919439147982,
"learning_rate": 8.727272727272727e-05,
"loss": 1.2292,
"step": 72
},
{
"epoch": 0.13,
"grad_norm": 0.1289941815481437,
"learning_rate": 8.848484848484849e-05,
"loss": 1.2191,
"step": 73
},
{
"epoch": 0.13,
"grad_norm": 0.13943952294847306,
"learning_rate": 8.96969696969697e-05,
"loss": 1.2915,
"step": 74
},
{
"epoch": 0.14,
"grad_norm": 0.1493528502117281,
"learning_rate": 9.090909090909092e-05,
"loss": 1.2797,
"step": 75
},
{
"epoch": 0.14,
"grad_norm": 0.1252401242451818,
"learning_rate": 9.212121212121214e-05,
"loss": 1.2552,
"step": 76
},
{
"epoch": 0.14,
"grad_norm": 0.13969800467546992,
"learning_rate": 9.333333333333334e-05,
"loss": 1.3147,
"step": 77
},
{
"epoch": 0.14,
"grad_norm": 0.1277258491470434,
"learning_rate": 9.454545454545455e-05,
"loss": 1.2089,
"step": 78
},
{
"epoch": 0.14,
"grad_norm": 0.133041369314817,
"learning_rate": 9.575757575757576e-05,
"loss": 1.2761,
"step": 79
},
{
"epoch": 0.15,
"grad_norm": 0.14564572037181842,
"learning_rate": 9.696969696969698e-05,
"loss": 1.1901,
"step": 80
},
{
"epoch": 0.15,
"grad_norm": 0.13666505656492195,
"learning_rate": 9.818181818181818e-05,
"loss": 1.2615,
"step": 81
},
{
"epoch": 0.15,
"grad_norm": 0.135007805210003,
"learning_rate": 9.939393939393939e-05,
"loss": 1.2669,
"step": 82
},
{
"epoch": 0.15,
"grad_norm": 0.17287563365884975,
"learning_rate": 0.00010060606060606062,
"loss": 1.2669,
"step": 83
},
{
"epoch": 0.15,
"grad_norm": 0.12934306326048103,
"learning_rate": 0.00010181818181818181,
"loss": 1.1979,
"step": 84
},
{
"epoch": 0.15,
"grad_norm": 0.13517436169178096,
"learning_rate": 0.00010303030303030303,
"loss": 1.2226,
"step": 85
},
{
"epoch": 0.16,
"grad_norm": 0.12105351159271568,
"learning_rate": 0.00010424242424242425,
"loss": 1.1172,
"step": 86
},
{
"epoch": 0.16,
"grad_norm": 0.1281676431775383,
"learning_rate": 0.00010545454545454545,
"loss": 1.2046,
"step": 87
},
{
"epoch": 0.16,
"grad_norm": 0.11730963057933333,
"learning_rate": 0.00010666666666666667,
"loss": 1.1883,
"step": 88
},
{
"epoch": 0.16,
"grad_norm": 0.12655235108503246,
"learning_rate": 0.00010787878787878789,
"loss": 1.1331,
"step": 89
},
{
"epoch": 0.16,
"grad_norm": 0.13047560307970027,
"learning_rate": 0.00010909090909090909,
"loss": 1.2731,
"step": 90
},
{
"epoch": 0.17,
"grad_norm": 0.12193522973752649,
"learning_rate": 0.00011030303030303031,
"loss": 1.2161,
"step": 91
},
{
"epoch": 0.17,
"grad_norm": 0.12804360300116346,
"learning_rate": 0.00011151515151515153,
"loss": 1.3062,
"step": 92
},
{
"epoch": 0.17,
"grad_norm": 0.15991741754516206,
"learning_rate": 0.00011272727272727272,
"loss": 1.239,
"step": 93
},
{
"epoch": 0.17,
"grad_norm": 0.15140182244454561,
"learning_rate": 0.00011393939393939394,
"loss": 1.2349,
"step": 94
},
{
"epoch": 0.17,
"grad_norm": 0.12320241076263434,
"learning_rate": 0.00011515151515151516,
"loss": 1.2875,
"step": 95
},
{
"epoch": 0.17,
"grad_norm": 0.13235998458230466,
"learning_rate": 0.00011636363636363636,
"loss": 1.2218,
"step": 96
},
{
"epoch": 0.18,
"grad_norm": 0.11783688734798668,
"learning_rate": 0.00011757575757575758,
"loss": 1.1864,
"step": 97
},
{
"epoch": 0.18,
"grad_norm": 0.3151933420750235,
"learning_rate": 0.0001187878787878788,
"loss": 1.3023,
"step": 98
},
{
"epoch": 0.18,
"grad_norm": 0.12665632567219295,
"learning_rate": 0.00012,
"loss": 1.2249,
"step": 99
},
{
"epoch": 0.18,
"grad_norm": 0.1228886740460738,
"learning_rate": 0.00012121212121212122,
"loss": 1.2517,
"step": 100
},
{
"epoch": 0.18,
"grad_norm": 0.11892005244989344,
"learning_rate": 0.00012242424242424243,
"loss": 1.2586,
"step": 101
},
{
"epoch": 0.19,
"grad_norm": 0.1232340827222201,
"learning_rate": 0.00012363636363636364,
"loss": 1.3217,
"step": 102
},
{
"epoch": 0.19,
"grad_norm": 0.13837226869323116,
"learning_rate": 0.00012484848484848487,
"loss": 1.2693,
"step": 103
},
{
"epoch": 0.19,
"grad_norm": 0.12068217991774362,
"learning_rate": 0.00012606060606060605,
"loss": 1.2623,
"step": 104
},
{
"epoch": 0.19,
"grad_norm": 0.16779277284606545,
"learning_rate": 0.00012727272727272728,
"loss": 1.2415,
"step": 105
},
{
"epoch": 0.19,
"grad_norm": 0.13396891539963085,
"learning_rate": 0.0001284848484848485,
"loss": 1.2313,
"step": 106
},
{
"epoch": 0.19,
"grad_norm": 0.12457104490772812,
"learning_rate": 0.0001296969696969697,
"loss": 1.1758,
"step": 107
},
{
"epoch": 0.2,
"grad_norm": 0.12676816816563452,
"learning_rate": 0.00013090909090909093,
"loss": 1.2478,
"step": 108
},
{
"epoch": 0.2,
"grad_norm": 0.11973639622066906,
"learning_rate": 0.00013212121212121213,
"loss": 1.2335,
"step": 109
},
{
"epoch": 0.2,
"grad_norm": 0.1330159646034068,
"learning_rate": 0.00013333333333333334,
"loss": 1.26,
"step": 110
},
{
"epoch": 0.2,
"grad_norm": 0.1298003025338099,
"learning_rate": 0.00013454545454545455,
"loss": 1.1907,
"step": 111
},
{
"epoch": 0.2,
"grad_norm": 0.1226154813287666,
"learning_rate": 0.00013575757575757578,
"loss": 1.1807,
"step": 112
},
{
"epoch": 0.21,
"grad_norm": 0.12533753244302145,
"learning_rate": 0.00013696969696969696,
"loss": 1.2098,
"step": 113
},
{
"epoch": 0.21,
"grad_norm": 0.12673266503840944,
"learning_rate": 0.0001381818181818182,
"loss": 1.2265,
"step": 114
},
{
"epoch": 0.21,
"grad_norm": 0.1299039569361384,
"learning_rate": 0.0001393939393939394,
"loss": 1.2534,
"step": 115
},
{
"epoch": 0.21,
"grad_norm": 0.13023496663090803,
"learning_rate": 0.0001406060606060606,
"loss": 1.2453,
"step": 116
},
{
"epoch": 0.21,
"grad_norm": 0.12001793500864573,
"learning_rate": 0.00014181818181818184,
"loss": 1.1608,
"step": 117
},
{
"epoch": 0.21,
"grad_norm": 0.14561862193041028,
"learning_rate": 0.00014303030303030304,
"loss": 1.2233,
"step": 118
},
{
"epoch": 0.22,
"grad_norm": 0.12636130876430832,
"learning_rate": 0.00014424242424242425,
"loss": 1.2833,
"step": 119
},
{
"epoch": 0.22,
"grad_norm": 0.189556849271166,
"learning_rate": 0.00014545454545454546,
"loss": 1.3105,
"step": 120
},
{
"epoch": 0.22,
"grad_norm": 0.12409073764495662,
"learning_rate": 0.00014666666666666666,
"loss": 1.1534,
"step": 121
},
{
"epoch": 0.22,
"grad_norm": 0.12149212466969316,
"learning_rate": 0.0001478787878787879,
"loss": 1.3039,
"step": 122
},
{
"epoch": 0.22,
"grad_norm": 0.12147336887953522,
"learning_rate": 0.0001490909090909091,
"loss": 1.326,
"step": 123
},
{
"epoch": 0.23,
"grad_norm": 0.1176585016163167,
"learning_rate": 0.0001503030303030303,
"loss": 1.191,
"step": 124
},
{
"epoch": 0.23,
"grad_norm": 0.2066428974234372,
"learning_rate": 0.00015151515151515152,
"loss": 1.3054,
"step": 125
},
{
"epoch": 0.23,
"grad_norm": 0.29582724255710047,
"learning_rate": 0.00015272727272727275,
"loss": 1.2032,
"step": 126
},
{
"epoch": 0.23,
"grad_norm": 0.13084381204119358,
"learning_rate": 0.00015393939393939393,
"loss": 1.2289,
"step": 127
},
{
"epoch": 0.23,
"grad_norm": 0.1294157600411397,
"learning_rate": 0.00015515151515151516,
"loss": 1.2561,
"step": 128
},
{
"epoch": 0.23,
"grad_norm": 0.14039614543447027,
"learning_rate": 0.00015636363636363637,
"loss": 1.243,
"step": 129
},
{
"epoch": 0.24,
"grad_norm": 0.19939984917282128,
"learning_rate": 0.00015757575757575757,
"loss": 1.1286,
"step": 130
},
{
"epoch": 0.24,
"grad_norm": 0.14402764349968203,
"learning_rate": 0.0001587878787878788,
"loss": 1.1959,
"step": 131
},
{
"epoch": 0.24,
"grad_norm": 0.13970978861500938,
"learning_rate": 0.00016,
"loss": 1.1814,
"step": 132
},
{
"epoch": 0.24,
"grad_norm": 0.14539538472563127,
"learning_rate": 0.00016121212121212122,
"loss": 1.2317,
"step": 133
},
{
"epoch": 0.24,
"grad_norm": 0.13456425455391557,
"learning_rate": 0.00016242424242424243,
"loss": 1.2239,
"step": 134
},
{
"epoch": 0.25,
"grad_norm": 0.1314997837157779,
"learning_rate": 0.00016363636363636366,
"loss": 1.1986,
"step": 135
},
{
"epoch": 0.25,
"grad_norm": 0.14046946525591422,
"learning_rate": 0.00016484848484848487,
"loss": 1.2238,
"step": 136
},
{
"epoch": 0.25,
"grad_norm": 0.6095538041505763,
"learning_rate": 0.00016606060606060607,
"loss": 1.2332,
"step": 137
},
{
"epoch": 0.25,
"grad_norm": 0.17707289712054367,
"learning_rate": 0.00016727272727272728,
"loss": 1.2401,
"step": 138
},
{
"epoch": 0.25,
"grad_norm": 0.19335172179099247,
"learning_rate": 0.00016848484848484848,
"loss": 1.2361,
"step": 139
},
{
"epoch": 0.26,
"grad_norm": 0.13725591818701255,
"learning_rate": 0.00016969696969696972,
"loss": 1.193,
"step": 140
},
{
"epoch": 0.26,
"grad_norm": 0.15535575462507384,
"learning_rate": 0.0001709090909090909,
"loss": 1.2769,
"step": 141
},
{
"epoch": 0.26,
"grad_norm": 0.14909436560898923,
"learning_rate": 0.00017212121212121213,
"loss": 1.2602,
"step": 142
},
{
"epoch": 0.26,
"grad_norm": 0.15054368082407957,
"learning_rate": 0.00017333333333333334,
"loss": 1.2607,
"step": 143
},
{
"epoch": 0.26,
"grad_norm": 0.13386897838741724,
"learning_rate": 0.00017454545454545454,
"loss": 1.168,
"step": 144
},
{
"epoch": 0.26,
"grad_norm": 0.13567889528730145,
"learning_rate": 0.00017575757575757578,
"loss": 1.1984,
"step": 145
},
{
"epoch": 0.27,
"grad_norm": 0.13994382298003089,
"learning_rate": 0.00017696969696969698,
"loss": 1.2795,
"step": 146
},
{
"epoch": 0.27,
"grad_norm": 0.13941573210713187,
"learning_rate": 0.0001781818181818182,
"loss": 1.2303,
"step": 147
},
{
"epoch": 0.27,
"grad_norm": 0.18302605925485763,
"learning_rate": 0.0001793939393939394,
"loss": 1.2696,
"step": 148
},
{
"epoch": 0.27,
"grad_norm": 0.1547402223275396,
"learning_rate": 0.00018060606060606063,
"loss": 1.1276,
"step": 149
},
{
"epoch": 0.27,
"grad_norm": 0.19947594494850646,
"learning_rate": 0.00018181818181818183,
"loss": 1.271,
"step": 150
},
{
"epoch": 0.28,
"grad_norm": 0.1517101450465788,
"learning_rate": 0.00018303030303030304,
"loss": 1.2193,
"step": 151
},
{
"epoch": 0.28,
"grad_norm": 0.19251063116857103,
"learning_rate": 0.00018424242424242427,
"loss": 1.2703,
"step": 152
},
{
"epoch": 0.28,
"grad_norm": 0.16789099560498666,
"learning_rate": 0.00018545454545454545,
"loss": 1.2244,
"step": 153
},
{
"epoch": 0.28,
"grad_norm": 0.14907376557922342,
"learning_rate": 0.0001866666666666667,
"loss": 1.264,
"step": 154
},
{
"epoch": 0.28,
"grad_norm": 0.14276598263036905,
"learning_rate": 0.0001878787878787879,
"loss": 1.2545,
"step": 155
},
{
"epoch": 0.28,
"grad_norm": 0.14526753816999002,
"learning_rate": 0.0001890909090909091,
"loss": 1.2912,
"step": 156
},
{
"epoch": 0.29,
"grad_norm": 0.1627048894660859,
"learning_rate": 0.0001903030303030303,
"loss": 1.2573,
"step": 157
},
{
"epoch": 0.29,
"grad_norm": 0.16405036632332695,
"learning_rate": 0.0001915151515151515,
"loss": 1.2359,
"step": 158
},
{
"epoch": 0.29,
"grad_norm": 0.14533427219788658,
"learning_rate": 0.00019272727272727274,
"loss": 1.1718,
"step": 159
},
{
"epoch": 0.29,
"grad_norm": 0.13802382666732702,
"learning_rate": 0.00019393939393939395,
"loss": 1.2297,
"step": 160
},
{
"epoch": 0.29,
"grad_norm": 0.15620193618511755,
"learning_rate": 0.00019515151515151516,
"loss": 1.2287,
"step": 161
},
{
"epoch": 0.3,
"grad_norm": 0.1401696295700075,
"learning_rate": 0.00019636363636363636,
"loss": 1.2231,
"step": 162
},
{
"epoch": 0.3,
"grad_norm": 0.15816133304035035,
"learning_rate": 0.0001975757575757576,
"loss": 1.2804,
"step": 163
},
{
"epoch": 0.3,
"grad_norm": 0.14626275180535692,
"learning_rate": 0.00019878787878787878,
"loss": 1.2115,
"step": 164
},
{
"epoch": 0.3,
"grad_norm": 0.13100680398305042,
"learning_rate": 0.0002,
"loss": 1.2524,
"step": 165
},
{
"epoch": 0.3,
"grad_norm": 0.14849458896148926,
"learning_rate": 0.00019999977531546566,
"loss": 1.2161,
"step": 166
},
{
"epoch": 0.3,
"grad_norm": 0.13628125499037252,
"learning_rate": 0.0001999991012628722,
"loss": 1.2452,
"step": 167
},
{
"epoch": 0.31,
"grad_norm": 0.18617698759086793,
"learning_rate": 0.00019999797784524866,
"loss": 1.2197,
"step": 168
},
{
"epoch": 0.31,
"grad_norm": 0.14416004826313944,
"learning_rate": 0.00019999640506764336,
"loss": 1.2796,
"step": 169
},
{
"epoch": 0.31,
"grad_norm": 0.13807081386834757,
"learning_rate": 0.0001999943829371238,
"loss": 1.2732,
"step": 170
},
{
"epoch": 0.31,
"grad_norm": 0.16526927436841996,
"learning_rate": 0.0001999919114627769,
"loss": 1.3016,
"step": 171
},
{
"epoch": 0.31,
"grad_norm": 0.14479672734919855,
"learning_rate": 0.0001999889906557086,
"loss": 1.3106,
"step": 172
},
{
"epoch": 0.32,
"grad_norm": 0.13829284006072087,
"learning_rate": 0.00019998562052904418,
"loss": 1.3355,
"step": 173
},
{
"epoch": 0.32,
"grad_norm": 0.13484630104616105,
"learning_rate": 0.0001999818010979279,
"loss": 1.1928,
"step": 174
},
{
"epoch": 0.32,
"grad_norm": 0.14972770674556948,
"learning_rate": 0.00019997753237952317,
"loss": 1.2559,
"step": 175
},
{
"epoch": 0.32,
"grad_norm": 0.13378525020528342,
"learning_rate": 0.00019997281439301218,
"loss": 1.2673,
"step": 176
},
{
"epoch": 0.32,
"grad_norm": 0.13242998699125438,
"learning_rate": 0.00019996764715959618,
"loss": 1.2272,
"step": 177
},
{
"epoch": 0.32,
"grad_norm": 0.12938881004364342,
"learning_rate": 0.00019996203070249516,
"loss": 1.2035,
"step": 178
},
{
"epoch": 0.33,
"grad_norm": 0.13388032350164566,
"learning_rate": 0.00019995596504694763,
"loss": 1.2642,
"step": 179
},
{
"epoch": 0.33,
"grad_norm": 0.13893372222140873,
"learning_rate": 0.00019994945022021082,
"loss": 1.2235,
"step": 180
},
{
"epoch": 0.33,
"grad_norm": 0.14131710715500717,
"learning_rate": 0.00019994248625156038,
"loss": 1.1095,
"step": 181
},
{
"epoch": 0.33,
"grad_norm": 0.13448100369103572,
"learning_rate": 0.0001999350731722902,
"loss": 1.1879,
"step": 182
},
{
"epoch": 0.33,
"grad_norm": 0.13862444003216381,
"learning_rate": 0.00019992721101571236,
"loss": 1.2227,
"step": 183
},
{
"epoch": 0.34,
"grad_norm": 0.13506115547921224,
"learning_rate": 0.00019991889981715698,
"loss": 1.2833,
"step": 184
},
{
"epoch": 0.34,
"grad_norm": 0.13174857502600473,
"learning_rate": 0.00019991013961397197,
"loss": 1.2394,
"step": 185
},
{
"epoch": 0.34,
"grad_norm": 0.1290276308949748,
"learning_rate": 0.00019990093044552304,
"loss": 1.2659,
"step": 186
},
{
"epoch": 0.34,
"grad_norm": 0.1388159912078538,
"learning_rate": 0.0001998912723531933,
"loss": 1.3052,
"step": 187
},
{
"epoch": 0.34,
"grad_norm": 0.1256806205303357,
"learning_rate": 0.00019988116538038325,
"loss": 1.2031,
"step": 188
},
{
"epoch": 0.34,
"grad_norm": 0.13256850855084143,
"learning_rate": 0.00019987060957251047,
"loss": 1.211,
"step": 189
},
{
"epoch": 0.35,
"grad_norm": 0.13197363789890235,
"learning_rate": 0.0001998596049770095,
"loss": 1.2256,
"step": 190
},
{
"epoch": 0.35,
"grad_norm": 0.13277364593883098,
"learning_rate": 0.00019984815164333163,
"loss": 1.2174,
"step": 191
},
{
"epoch": 0.35,
"grad_norm": 0.13838072824574454,
"learning_rate": 0.00019983624962294458,
"loss": 1.3128,
"step": 192
},
{
"epoch": 0.35,
"grad_norm": 0.13524759737199996,
"learning_rate": 0.0001998238989693323,
"loss": 1.1806,
"step": 193
},
{
"epoch": 0.35,
"grad_norm": 0.12669987683723832,
"learning_rate": 0.0001998110997379949,
"loss": 1.2171,
"step": 194
},
{
"epoch": 0.36,
"grad_norm": 0.1461834612451898,
"learning_rate": 0.00019979785198644806,
"loss": 1.2231,
"step": 195
},
{
"epoch": 0.36,
"grad_norm": 0.13265793664862735,
"learning_rate": 0.0001997841557742232,
"loss": 1.1718,
"step": 196
},
{
"epoch": 0.36,
"grad_norm": 0.12842971557690963,
"learning_rate": 0.00019977001116286674,
"loss": 1.2758,
"step": 197
},
{
"epoch": 0.36,
"grad_norm": 0.12188365921206967,
"learning_rate": 0.00019975541821594026,
"loss": 1.2457,
"step": 198
},
{
"epoch": 0.36,
"grad_norm": 0.12679949330022622,
"learning_rate": 0.00019974037699901993,
"loss": 1.1825,
"step": 199
},
{
"epoch": 0.36,
"grad_norm": 0.12949746150357985,
"learning_rate": 0.00019972488757969635,
"loss": 1.2666,
"step": 200
},
{
"epoch": 0.37,
"grad_norm": 0.1363496149379173,
"learning_rate": 0.00019970895002757413,
"loss": 1.2031,
"step": 201
},
{
"epoch": 0.37,
"grad_norm": 0.14218340110669314,
"learning_rate": 0.0001996925644142717,
"loss": 1.3073,
"step": 202
},
{
"epoch": 0.37,
"grad_norm": 0.14234535389443218,
"learning_rate": 0.00019967573081342103,
"loss": 1.2444,
"step": 203
},
{
"epoch": 0.37,
"grad_norm": 0.12866113026310516,
"learning_rate": 0.000199658449300667,
"loss": 1.2257,
"step": 204
},
{
"epoch": 0.37,
"grad_norm": 0.1324053366295965,
"learning_rate": 0.00019964071995366744,
"loss": 1.2374,
"step": 205
},
{
"epoch": 0.38,
"grad_norm": 0.12906841330218152,
"learning_rate": 0.00019962254285209254,
"loss": 1.2334,
"step": 206
},
{
"epoch": 0.38,
"grad_norm": 0.13620873131846425,
"learning_rate": 0.00019960391807762463,
"loss": 1.242,
"step": 207
},
{
"epoch": 0.38,
"grad_norm": 0.14877366842835116,
"learning_rate": 0.00019958484571395757,
"loss": 1.1772,
"step": 208
},
{
"epoch": 0.38,
"grad_norm": 0.13914108740445985,
"learning_rate": 0.00019956532584679675,
"loss": 1.2734,
"step": 209
},
{
"epoch": 0.38,
"grad_norm": 0.13198394930310692,
"learning_rate": 0.00019954535856385837,
"loss": 1.1728,
"step": 210
},
{
"epoch": 0.38,
"grad_norm": 0.3807736597404611,
"learning_rate": 0.0001995249439548693,
"loss": 1.2089,
"step": 211
},
{
"epoch": 0.39,
"grad_norm": 0.1682550557564819,
"learning_rate": 0.00019950408211156636,
"loss": 1.2423,
"step": 212
},
{
"epoch": 0.39,
"grad_norm": 0.2102196862007261,
"learning_rate": 0.0001994827731276963,
"loss": 1.2096,
"step": 213
},
{
"epoch": 0.39,
"grad_norm": 0.154346739470422,
"learning_rate": 0.00019946101709901514,
"loss": 1.2847,
"step": 214
},
{
"epoch": 0.39,
"grad_norm": 0.16416668358293746,
"learning_rate": 0.0001994388141232876,
"loss": 1.2503,
"step": 215
},
{
"epoch": 0.39,
"grad_norm": 0.13134349458231093,
"learning_rate": 0.0001994161643002871,
"loss": 1.1231,
"step": 216
},
{
"epoch": 0.4,
"grad_norm": 0.15083246389185287,
"learning_rate": 0.00019939306773179497,
"loss": 1.1614,
"step": 217
},
{
"epoch": 0.4,
"grad_norm": 0.1742387260929692,
"learning_rate": 0.00019936952452159995,
"loss": 1.3568,
"step": 218
},
{
"epoch": 0.4,
"grad_norm": 0.18146911432436974,
"learning_rate": 0.00019934553477549794,
"loss": 1.2686,
"step": 219
},
{
"epoch": 0.4,
"grad_norm": 0.1393593447949332,
"learning_rate": 0.00019932109860129154,
"loss": 1.1141,
"step": 220
},
{
"epoch": 0.4,
"grad_norm": 0.14856124153987935,
"learning_rate": 0.00019929621610878927,
"loss": 1.234,
"step": 221
},
{
"epoch": 0.4,
"grad_norm": 0.14820851831477327,
"learning_rate": 0.0001992708874098054,
"loss": 1.2069,
"step": 222
},
{
"epoch": 0.41,
"grad_norm": 0.17893142790958147,
"learning_rate": 0.00019924511261815926,
"loss": 1.1278,
"step": 223
},
{
"epoch": 0.41,
"grad_norm": 0.14573658703265605,
"learning_rate": 0.00019921889184967476,
"loss": 1.2292,
"step": 224
},
{
"epoch": 0.41,
"grad_norm": 0.15282321197574994,
"learning_rate": 0.00019919222522217996,
"loss": 1.2482,
"step": 225
},
{
"epoch": 0.41,
"grad_norm": 0.16342112084119492,
"learning_rate": 0.00019916511285550642,
"loss": 1.2172,
"step": 226
},
{
"epoch": 0.41,
"grad_norm": 0.1475889153814455,
"learning_rate": 0.00019913755487148876,
"loss": 1.1747,
"step": 227
},
{
"epoch": 0.42,
"grad_norm": 0.163738064491857,
"learning_rate": 0.00019910955139396396,
"loss": 1.3007,
"step": 228
},
{
"epoch": 0.42,
"grad_norm": 0.14427856196022704,
"learning_rate": 0.00019908110254877106,
"loss": 1.2464,
"step": 229
},
{
"epoch": 0.42,
"grad_norm": 0.20204742660246344,
"learning_rate": 0.00019905220846375032,
"loss": 1.2515,
"step": 230
},
{
"epoch": 0.42,
"grad_norm": 0.15134144918251685,
"learning_rate": 0.0001990228692687429,
"loss": 1.1786,
"step": 231
},
{
"epoch": 0.42,
"grad_norm": 0.1636590177812163,
"learning_rate": 0.00019899308509558998,
"loss": 1.1974,
"step": 232
},
{
"epoch": 0.42,
"grad_norm": 0.15552319776955892,
"learning_rate": 0.00019896285607813244,
"loss": 1.2308,
"step": 233
},
{
"epoch": 0.43,
"grad_norm": 0.17104898009833774,
"learning_rate": 0.00019893218235221015,
"loss": 1.2828,
"step": 234
},
{
"epoch": 0.43,
"grad_norm": 0.16387378763964267,
"learning_rate": 0.00019890106405566138,
"loss": 1.2779,
"step": 235
},
{
"epoch": 0.43,
"grad_norm": 0.14622126798612248,
"learning_rate": 0.00019886950132832207,
"loss": 1.2894,
"step": 236
},
{
"epoch": 0.43,
"grad_norm": 0.16619841547518147,
"learning_rate": 0.0001988374943120254,
"loss": 1.2133,
"step": 237
},
{
"epoch": 0.43,
"grad_norm": 0.12664832399697545,
"learning_rate": 0.00019880504315060096,
"loss": 1.1807,
"step": 238
},
{
"epoch": 0.44,
"grad_norm": 0.2015108381613456,
"learning_rate": 0.00019877214798987426,
"loss": 1.1876,
"step": 239
},
{
"epoch": 0.44,
"grad_norm": 0.14468620723711506,
"learning_rate": 0.00019873880897766598,
"loss": 1.1883,
"step": 240
},
{
"epoch": 0.44,
"grad_norm": 0.1549018650770757,
"learning_rate": 0.00019870502626379127,
"loss": 1.2896,
"step": 241
},
{
"epoch": 0.44,
"grad_norm": 0.1492917963684983,
"learning_rate": 0.0001986708000000593,
"loss": 1.2102,
"step": 242
},
{
"epoch": 0.44,
"grad_norm": 0.178606606459489,
"learning_rate": 0.00019863613034027224,
"loss": 1.2292,
"step": 243
},
{
"epoch": 0.44,
"grad_norm": 0.206170239681528,
"learning_rate": 0.00019860101744022485,
"loss": 1.2666,
"step": 244
},
{
"epoch": 0.45,
"grad_norm": 0.13741043007948167,
"learning_rate": 0.0001985654614577036,
"loss": 1.2022,
"step": 245
},
{
"epoch": 0.45,
"grad_norm": 0.1595080658199459,
"learning_rate": 0.0001985294625524861,
"loss": 1.1203,
"step": 246
},
{
"epoch": 0.45,
"grad_norm": 0.13929705183853777,
"learning_rate": 0.00019849302088634034,
"loss": 1.1505,
"step": 247
},
{
"epoch": 0.45,
"grad_norm": 0.14045247607912964,
"learning_rate": 0.00019845613662302383,
"loss": 1.1897,
"step": 248
},
{
"epoch": 0.45,
"grad_norm": 0.15002651347444407,
"learning_rate": 0.00019841880992828306,
"loss": 1.2133,
"step": 249
},
{
"epoch": 0.46,
"grad_norm": 0.1567929487810952,
"learning_rate": 0.00019838104096985267,
"loss": 1.129,
"step": 250
},
{
"epoch": 0.46,
"grad_norm": 0.15240634543877116,
"learning_rate": 0.00019834282991745464,
"loss": 1.1995,
"step": 251
},
{
"epoch": 0.46,
"grad_norm": 0.151807679821367,
"learning_rate": 0.00019830417694279766,
"loss": 1.25,
"step": 252
},
{
"epoch": 0.46,
"grad_norm": 0.1648599156208311,
"learning_rate": 0.0001982650822195762,
"loss": 1.2511,
"step": 253
},
{
"epoch": 0.46,
"grad_norm": 0.15363401233808713,
"learning_rate": 0.00019822554592346993,
"loss": 1.1794,
"step": 254
},
{
"epoch": 0.46,
"grad_norm": 0.1569644350778875,
"learning_rate": 0.00019818556823214268,
"loss": 1.2033,
"step": 255
},
{
"epoch": 0.47,
"grad_norm": 0.15996552747294254,
"learning_rate": 0.0001981451493252418,
"loss": 1.2809,
"step": 256
},
{
"epoch": 0.47,
"grad_norm": 0.15863104885072635,
"learning_rate": 0.0001981042893843974,
"loss": 1.1667,
"step": 257
},
{
"epoch": 0.47,
"grad_norm": 0.2887466971861171,
"learning_rate": 0.0001980629885932214,
"loss": 1.1915,
"step": 258
},
{
"epoch": 0.47,
"grad_norm": 0.15233015979193984,
"learning_rate": 0.00019802124713730681,
"loss": 1.1734,
"step": 259
},
{
"epoch": 0.47,
"grad_norm": 0.18207884538436447,
"learning_rate": 0.00019797906520422677,
"loss": 1.2575,
"step": 260
},
{
"epoch": 0.48,
"grad_norm": 0.17323546756038308,
"learning_rate": 0.0001979364429835339,
"loss": 1.1704,
"step": 261
},
{
"epoch": 0.48,
"grad_norm": 0.14592153602263633,
"learning_rate": 0.00019789338066675922,
"loss": 1.192,
"step": 262
},
{
"epoch": 0.48,
"grad_norm": 0.19250697792287097,
"learning_rate": 0.0001978498784474115,
"loss": 1.2779,
"step": 263
},
{
"epoch": 0.48,
"grad_norm": 0.1429107680887097,
"learning_rate": 0.0001978059365209762,
"loss": 1.2529,
"step": 264
},
{
"epoch": 0.48,
"grad_norm": 0.48514081074992116,
"learning_rate": 0.00019776155508491482,
"loss": 1.1917,
"step": 265
},
{
"epoch": 0.48,
"grad_norm": 0.1534376167748161,
"learning_rate": 0.0001977167343386638,
"loss": 1.2384,
"step": 266
},
{
"epoch": 0.49,
"grad_norm": 0.16744875760032166,
"learning_rate": 0.00019767147448363366,
"loss": 1.1744,
"step": 267
},
{
"epoch": 0.49,
"grad_norm": 0.29195538170738244,
"learning_rate": 0.00019762577572320824,
"loss": 1.1418,
"step": 268
},
{
"epoch": 0.49,
"grad_norm": 0.1820804717651353,
"learning_rate": 0.00019757963826274357,
"loss": 1.2815,
"step": 269
},
{
"epoch": 0.49,
"grad_norm": 0.17522345110441973,
"learning_rate": 0.00019753306230956718,
"loss": 1.2363,
"step": 270
},
{
"epoch": 0.49,
"grad_norm": 0.16354388270886613,
"learning_rate": 0.000197486048072977,
"loss": 1.2845,
"step": 271
},
{
"epoch": 0.5,
"grad_norm": 0.17590082756401024,
"learning_rate": 0.0001974385957642404,
"loss": 1.192,
"step": 272
},
{
"epoch": 0.5,
"grad_norm": 0.17345720403188775,
"learning_rate": 0.00019739070559659347,
"loss": 1.2068,
"step": 273
},
{
"epoch": 0.5,
"grad_norm": 0.16070434867766506,
"learning_rate": 0.00019734237778523976,
"loss": 1.189,
"step": 274
},
{
"epoch": 0.5,
"grad_norm": 0.18983443066710415,
"learning_rate": 0.0001972936125473495,
"loss": 1.2223,
"step": 275
},
{
"epoch": 0.5,
"grad_norm": 0.15724400187981355,
"learning_rate": 0.00019724441010205863,
"loss": 1.2292,
"step": 276
},
{
"epoch": 0.5,
"grad_norm": 0.14570729442956004,
"learning_rate": 0.00019719477067046766,
"loss": 1.1421,
"step": 277
},
{
"epoch": 0.51,
"grad_norm": 0.1559242881177266,
"learning_rate": 0.00019714469447564088,
"loss": 1.2598,
"step": 278
},
{
"epoch": 0.51,
"grad_norm": 0.16621830243096108,
"learning_rate": 0.0001970941817426052,
"loss": 1.3038,
"step": 279
},
{
"epoch": 0.51,
"grad_norm": 4.675483994100576,
"learning_rate": 0.00019704323269834927,
"loss": 1.2298,
"step": 280
},
{
"epoch": 0.51,
"grad_norm": 0.2769699381619058,
"learning_rate": 0.00019699184757182225,
"loss": 1.2566,
"step": 281
},
{
"epoch": 0.51,
"grad_norm": 0.20189839889100783,
"learning_rate": 0.00019694002659393305,
"loss": 1.3181,
"step": 282
},
{
"epoch": 0.52,
"grad_norm": 0.19497107359413876,
"learning_rate": 0.00019688776999754912,
"loss": 1.1502,
"step": 283
},
{
"epoch": 0.52,
"grad_norm": 0.1982266815755412,
"learning_rate": 0.00019683507801749545,
"loss": 1.2053,
"step": 284
},
{
"epoch": 0.52,
"grad_norm": 0.1924340950322314,
"learning_rate": 0.00019678195089055346,
"loss": 1.2149,
"step": 285
},
{
"epoch": 0.52,
"grad_norm": 0.1725322346446431,
"learning_rate": 0.00019672838885546008,
"loss": 1.2553,
"step": 286
},
{
"epoch": 0.52,
"grad_norm": 0.2535488743520272,
"learning_rate": 0.00019667439215290648,
"loss": 1.2576,
"step": 287
},
{
"epoch": 0.52,
"grad_norm": 0.37837586860064026,
"learning_rate": 0.00019661996102553718,
"loss": 1.1815,
"step": 288
},
{
"epoch": 0.53,
"grad_norm": 0.17520419597901843,
"learning_rate": 0.00019656509571794878,
"loss": 1.1932,
"step": 289
},
{
"epoch": 0.53,
"grad_norm": 0.17056234784450633,
"learning_rate": 0.00019650979647668906,
"loss": 1.163,
"step": 290
},
{
"epoch": 0.53,
"grad_norm": 0.18272246580207432,
"learning_rate": 0.00019645406355025565,
"loss": 1.1887,
"step": 291
},
{
"epoch": 0.53,
"grad_norm": 0.17889037954429915,
"learning_rate": 0.00019639789718909508,
"loss": 1.2126,
"step": 292
},
{
"epoch": 0.53,
"grad_norm": 0.23993734971101424,
"learning_rate": 0.00019634129764560168,
"loss": 1.2485,
"step": 293
},
{
"epoch": 0.54,
"grad_norm": 0.1847578318208199,
"learning_rate": 0.00019628426517411625,
"loss": 1.2549,
"step": 294
},
{
"epoch": 0.54,
"grad_norm": 0.23185098827091005,
"learning_rate": 0.00019622680003092503,
"loss": 1.1599,
"step": 295
},
{
"epoch": 0.54,
"grad_norm": 0.220638044092583,
"learning_rate": 0.00019616890247425866,
"loss": 1.2281,
"step": 296
},
{
"epoch": 0.54,
"grad_norm": 0.2303439219825616,
"learning_rate": 0.00019611057276429085,
"loss": 1.2208,
"step": 297
},
{
"epoch": 0.54,
"grad_norm": 0.1744807302230573,
"learning_rate": 0.00019605181116313724,
"loss": 1.2303,
"step": 298
},
{
"epoch": 0.54,
"grad_norm": 0.17510946821872422,
"learning_rate": 0.0001959926179348543,
"loss": 1.2385,
"step": 299
},
{
"epoch": 0.55,
"grad_norm": 0.2218474349751746,
"learning_rate": 0.00019593299334543808,
"loss": 1.2153,
"step": 300
},
{
"epoch": 0.55,
"grad_norm": 0.1742070481516402,
"learning_rate": 0.00019587293766282308,
"loss": 1.1628,
"step": 301
},
{
"epoch": 0.55,
"grad_norm": 0.15250311715180823,
"learning_rate": 0.00019581245115688094,
"loss": 1.1632,
"step": 302
},
{
"epoch": 0.55,
"grad_norm": 0.1744397677094501,
"learning_rate": 0.0001957515340994193,
"loss": 1.254,
"step": 303
},
{
"epoch": 0.55,
"grad_norm": 0.1686772182789891,
"learning_rate": 0.00019569018676418053,
"loss": 1.2169,
"step": 304
},
{
"epoch": 0.56,
"grad_norm": 0.16404966161017623,
"learning_rate": 0.00019562840942684067,
"loss": 1.2221,
"step": 305
},
{
"epoch": 0.56,
"grad_norm": 0.16052011449463713,
"learning_rate": 0.00019556620236500793,
"loss": 1.2045,
"step": 306
},
{
"epoch": 0.56,
"grad_norm": 0.16343251390831215,
"learning_rate": 0.0001955035658582216,
"loss": 1.2289,
"step": 307
},
{
"epoch": 0.56,
"grad_norm": 0.14387162360389305,
"learning_rate": 0.00019544050018795075,
"loss": 1.1365,
"step": 308
},
{
"epoch": 0.56,
"grad_norm": 0.15304461439740238,
"learning_rate": 0.00019537700563759304,
"loss": 1.1931,
"step": 309
},
{
"epoch": 0.56,
"grad_norm": 0.17059958050065627,
"learning_rate": 0.00019531308249247327,
"loss": 1.2166,
"step": 310
},
{
"epoch": 0.57,
"grad_norm": 0.17633385530926995,
"learning_rate": 0.00019524873103984235,
"loss": 1.2604,
"step": 311
},
{
"epoch": 0.57,
"grad_norm": 0.17855814403303746,
"learning_rate": 0.00019518395156887576,
"loss": 1.1615,
"step": 312
},
{
"epoch": 0.57,
"grad_norm": 0.19823982444256988,
"learning_rate": 0.00019511874437067243,
"loss": 1.2153,
"step": 313
},
{
"epoch": 0.57,
"grad_norm": 0.1570784627362585,
"learning_rate": 0.0001950531097382533,
"loss": 1.2788,
"step": 314
},
{
"epoch": 0.57,
"grad_norm": 0.2183125402112695,
"learning_rate": 0.00019498704796656018,
"loss": 1.2966,
"step": 315
},
{
"epoch": 0.58,
"grad_norm": 0.18173933276147194,
"learning_rate": 0.00019492055935245418,
"loss": 1.2978,
"step": 316
},
{
"epoch": 0.58,
"grad_norm": 0.17483116680914407,
"learning_rate": 0.00019485364419471454,
"loss": 1.258,
"step": 317
},
{
"epoch": 0.58,
"grad_norm": 0.15490767356815494,
"learning_rate": 0.0001947863027940374,
"loss": 1.2088,
"step": 318
},
{
"epoch": 0.58,
"grad_norm": 0.14703966491934156,
"learning_rate": 0.00019471853545303405,
"loss": 1.2355,
"step": 319
},
{
"epoch": 0.58,
"grad_norm": 0.14386689086661608,
"learning_rate": 0.00019465034247623003,
"loss": 1.2583,
"step": 320
},
{
"epoch": 0.58,
"grad_norm": 0.18818904376313625,
"learning_rate": 0.00019458172417006347,
"loss": 1.2181,
"step": 321
},
{
"epoch": 0.59,
"grad_norm": 0.17393313719202513,
"learning_rate": 0.00019451268084288385,
"loss": 1.3453,
"step": 322
},
{
"epoch": 0.59,
"grad_norm": 0.14706823379985753,
"learning_rate": 0.00019444321280495043,
"loss": 1.2234,
"step": 323
},
{
"epoch": 0.59,
"grad_norm": 0.15282014755252687,
"learning_rate": 0.00019437332036843118,
"loss": 1.1262,
"step": 324
},
{
"epoch": 0.59,
"grad_norm": 0.1618727884326225,
"learning_rate": 0.00019430300384740105,
"loss": 1.3136,
"step": 325
},
{
"epoch": 0.59,
"grad_norm": 0.16090758705378874,
"learning_rate": 0.00019423226355784077,
"loss": 1.2055,
"step": 326
},
{
"epoch": 0.6,
"grad_norm": 0.15241156801091013,
"learning_rate": 0.00019416109981763526,
"loss": 1.2678,
"step": 327
},
{
"epoch": 0.6,
"grad_norm": 0.14216697909809062,
"learning_rate": 0.0001940895129465724,
"loss": 1.2841,
"step": 328
},
{
"epoch": 0.6,
"grad_norm": 0.15790232415414485,
"learning_rate": 0.00019401750326634144,
"loss": 1.3119,
"step": 329
},
{
"epoch": 0.6,
"grad_norm": 0.13322691961062616,
"learning_rate": 0.0001939450711005316,
"loss": 1.1293,
"step": 330
},
{
"epoch": 0.6,
"grad_norm": 0.14075018938835404,
"learning_rate": 0.00019387221677463062,
"loss": 1.2176,
"step": 331
},
{
"epoch": 0.6,
"grad_norm": 0.21565975459393052,
"learning_rate": 0.00019379894061602335,
"loss": 1.1723,
"step": 332
},
{
"epoch": 0.61,
"grad_norm": 0.17967631394222838,
"learning_rate": 0.00019372524295399013,
"loss": 1.239,
"step": 333
},
{
"epoch": 0.61,
"grad_norm": 0.21187969201978435,
"learning_rate": 0.0001936511241197055,
"loss": 1.2207,
"step": 334
},
{
"epoch": 0.61,
"grad_norm": 0.16967789022974608,
"learning_rate": 0.00019357658444623654,
"loss": 1.2478,
"step": 335
},
{
"epoch": 0.61,
"grad_norm": 0.14810621660374448,
"learning_rate": 0.0001935016242685415,
"loss": 1.1223,
"step": 336
},
{
"epoch": 0.61,
"grad_norm": 0.1489106421847434,
"learning_rate": 0.00019342624392346824,
"loss": 1.1592,
"step": 337
},
{
"epoch": 0.62,
"grad_norm": 0.17625176068748855,
"learning_rate": 0.0001933504437497527,
"loss": 1.2145,
"step": 338
},
{
"epoch": 0.62,
"grad_norm": 0.17250255512763446,
"learning_rate": 0.00019327422408801744,
"loss": 1.2504,
"step": 339
},
{
"epoch": 0.62,
"grad_norm": 0.16079375745566896,
"learning_rate": 0.00019319758528077,
"loss": 1.1795,
"step": 340
},
{
"epoch": 0.62,
"grad_norm": 0.15454466809245995,
"learning_rate": 0.0001931205276724015,
"loss": 1.2123,
"step": 341
},
{
"epoch": 0.62,
"grad_norm": 0.7021323604447972,
"learning_rate": 0.000193043051609185,
"loss": 1.2239,
"step": 342
},
{
"epoch": 0.62,
"grad_norm": 0.1572764339385847,
"learning_rate": 0.00019296515743927399,
"loss": 1.2516,
"step": 343
},
{
"epoch": 0.63,
"grad_norm": 0.2136637778252246,
"learning_rate": 0.00019288684551270073,
"loss": 1.2321,
"step": 344
},
{
"epoch": 0.63,
"grad_norm": 0.4546540454773654,
"learning_rate": 0.00019280811618137484,
"loss": 1.18,
"step": 345
},
{
"epoch": 0.63,
"grad_norm": 0.9809832576786297,
"learning_rate": 0.00019272896979908154,
"loss": 1.2081,
"step": 346
},
{
"epoch": 0.63,
"grad_norm": 0.5246256133291822,
"learning_rate": 0.00019264940672148018,
"loss": 1.2722,
"step": 347
},
{
"epoch": 0.63,
"grad_norm": 0.24941717134878091,
"learning_rate": 0.00019256942730610268,
"loss": 1.2352,
"step": 348
},
{
"epoch": 0.64,
"grad_norm": 0.3356068462072784,
"learning_rate": 0.00019248903191235176,
"loss": 1.2225,
"step": 349
},
{
"epoch": 0.64,
"grad_norm": 0.19535845221880543,
"learning_rate": 0.00019240822090149944,
"loss": 1.1669,
"step": 350
},
{
"epoch": 0.64,
"grad_norm": 0.22306941566416597,
"learning_rate": 0.00019232699463668542,
"loss": 1.2281,
"step": 351
},
{
"epoch": 0.64,
"grad_norm": 0.2700134013989352,
"learning_rate": 0.00019224535348291542,
"loss": 1.1939,
"step": 352
},
{
"epoch": 0.64,
"grad_norm": 0.24406908935562743,
"learning_rate": 0.00019216329780705953,
"loss": 1.1839,
"step": 353
},
{
"epoch": 0.64,
"grad_norm": 0.20465183000217488,
"learning_rate": 0.00019208082797785055,
"loss": 1.2277,
"step": 354
},
{
"epoch": 0.65,
"grad_norm": 0.21324820828129784,
"learning_rate": 0.00019199794436588243,
"loss": 1.2072,
"step": 355
},
{
"epoch": 0.65,
"grad_norm": 0.1780562512431263,
"learning_rate": 0.00019191464734360844,
"loss": 1.2082,
"step": 356
},
{
"epoch": 0.65,
"grad_norm": 0.16547971467615655,
"learning_rate": 0.00019183093728533966,
"loss": 1.1978,
"step": 357
},
{
"epoch": 0.65,
"grad_norm": 0.22904664933247196,
"learning_rate": 0.00019174681456724318,
"loss": 1.1562,
"step": 358
},
{
"epoch": 0.65,
"grad_norm": 0.1737397860007602,
"learning_rate": 0.00019166227956734052,
"loss": 1.2383,
"step": 359
},
{
"epoch": 0.66,
"grad_norm": 0.1589465455917568,
"learning_rate": 0.00019157733266550575,
"loss": 1.2158,
"step": 360
},
{
"epoch": 0.66,
"grad_norm": 0.16253126221999709,
"learning_rate": 0.00019149197424346405,
"loss": 1.1952,
"step": 361
},
{
"epoch": 0.66,
"grad_norm": 0.22436676243032663,
"learning_rate": 0.00019140620468478968,
"loss": 1.2315,
"step": 362
},
{
"epoch": 0.66,
"grad_norm": 0.19291682612950423,
"learning_rate": 0.00019132002437490458,
"loss": 1.2283,
"step": 363
},
{
"epoch": 0.66,
"grad_norm": 0.1519191258459668,
"learning_rate": 0.00019123343370107637,
"loss": 1.1151,
"step": 364
},
{
"epoch": 0.66,
"grad_norm": 0.17179909633547025,
"learning_rate": 0.00019114643305241676,
"loss": 1.1576,
"step": 365
},
{
"epoch": 0.67,
"grad_norm": 0.17992599023321432,
"learning_rate": 0.00019105902281987976,
"loss": 1.2592,
"step": 366
},
{
"epoch": 0.67,
"grad_norm": 0.17714099390314453,
"learning_rate": 0.00019097120339625994,
"loss": 1.2578,
"step": 367
},
{
"epoch": 0.67,
"grad_norm": 0.2455577642687935,
"learning_rate": 0.00019088297517619055,
"loss": 1.2361,
"step": 368
},
{
"epoch": 0.67,
"grad_norm": 0.18398518628783986,
"learning_rate": 0.00019079433855614201,
"loss": 1.1906,
"step": 369
},
{
"epoch": 0.67,
"grad_norm": 0.18944067022821645,
"learning_rate": 0.00019070529393441985,
"loss": 1.237,
"step": 370
},
{
"epoch": 0.68,
"grad_norm": 0.17639967519781063,
"learning_rate": 0.00019061584171116303,
"loss": 1.1841,
"step": 371
},
{
"epoch": 0.68,
"grad_norm": 0.15947129998283005,
"learning_rate": 0.00019052598228834217,
"loss": 1.1722,
"step": 372
},
{
"epoch": 0.68,
"grad_norm": 0.1693354353719105,
"learning_rate": 0.00019043571606975777,
"loss": 1.2204,
"step": 373
},
{
"epoch": 0.68,
"grad_norm": 0.16236190451963983,
"learning_rate": 0.00019034504346103823,
"loss": 1.1778,
"step": 374
},
{
"epoch": 0.68,
"grad_norm": 0.17702370729269964,
"learning_rate": 0.00019025396486963827,
"loss": 1.2065,
"step": 375
},
{
"epoch": 0.68,
"grad_norm": 0.19388150596154238,
"learning_rate": 0.00019016248070483687,
"loss": 1.2942,
"step": 376
},
{
"epoch": 0.69,
"grad_norm": 0.16152000400319103,
"learning_rate": 0.0001900705913777356,
"loss": 1.1784,
"step": 377
},
{
"epoch": 0.69,
"grad_norm": 0.1545267913996029,
"learning_rate": 0.00018997829730125663,
"loss": 1.1829,
"step": 378
},
{
"epoch": 0.69,
"grad_norm": 0.15421727704318197,
"learning_rate": 0.000189885598890141,
"loss": 1.177,
"step": 379
},
{
"epoch": 0.69,
"grad_norm": 0.1624966073814206,
"learning_rate": 0.00018979249656094673,
"loss": 1.2439,
"step": 380
},
{
"epoch": 0.69,
"grad_norm": 0.9490737312904575,
"learning_rate": 0.00018969899073204686,
"loss": 1.2085,
"step": 381
},
{
"epoch": 0.7,
"grad_norm": 0.8982903208613089,
"learning_rate": 0.00018960508182362768,
"loss": 1.2347,
"step": 382
},
{
"epoch": 0.7,
"grad_norm": 0.3771428474797688,
"learning_rate": 0.00018951077025768678,
"loss": 1.2546,
"step": 383
},
{
"epoch": 0.7,
"grad_norm": 0.4776152950069111,
"learning_rate": 0.00018941605645803115,
"loss": 1.2904,
"step": 384
},
{
"epoch": 0.7,
"grad_norm": 0.18786943849618057,
"learning_rate": 0.00018932094085027533,
"loss": 1.2122,
"step": 385
},
{
"epoch": 0.7,
"grad_norm": 1.6297025984167128,
"learning_rate": 0.0001892254238618394,
"loss": 1.171,
"step": 386
},
{
"epoch": 0.7,
"grad_norm": 0.20382660707264952,
"learning_rate": 0.0001891295059219472,
"loss": 1.1874,
"step": 387
},
{
"epoch": 0.71,
"grad_norm": 1.3580819775908755,
"learning_rate": 0.00018903318746162429,
"loss": 1.1531,
"step": 388
},
{
"epoch": 0.71,
"grad_norm": 0.43619056173016185,
"learning_rate": 0.00018893646891369602,
"loss": 1.2289,
"step": 389
},
{
"epoch": 0.71,
"grad_norm": 0.29385240705823723,
"learning_rate": 0.0001888393507127856,
"loss": 1.2073,
"step": 390
},
{
"epoch": 0.71,
"grad_norm": 0.3136086850525623,
"learning_rate": 0.00018874183329531223,
"loss": 1.1898,
"step": 391
},
{
"epoch": 0.71,
"grad_norm": 0.2307767217662562,
"learning_rate": 0.000188643917099489,
"loss": 1.207,
"step": 392
},
{
"epoch": 0.72,
"grad_norm": 0.18703654518135468,
"learning_rate": 0.000188545602565321,
"loss": 1.1688,
"step": 393
},
{
"epoch": 0.72,
"grad_norm": 0.4809351333934126,
"learning_rate": 0.00018844689013460336,
"loss": 1.2519,
"step": 394
},
{
"epoch": 0.72,
"grad_norm": 0.40370101428544464,
"learning_rate": 0.0001883477802509192,
"loss": 1.2411,
"step": 395
},
{
"epoch": 0.72,
"grad_norm": 0.2858848636432859,
"learning_rate": 0.00018824827335963765,
"loss": 1.194,
"step": 396
},
{
"epoch": 0.72,
"grad_norm": 0.32195602638999565,
"learning_rate": 0.000188148369907912,
"loss": 1.0988,
"step": 397
},
{
"epoch": 0.72,
"grad_norm": 0.23790306908901832,
"learning_rate": 0.00018804807034467733,
"loss": 1.2237,
"step": 398
},
{
"epoch": 0.73,
"grad_norm": 0.20126988767112128,
"learning_rate": 0.0001879473751206489,
"loss": 1.2731,
"step": 399
},
{
"epoch": 0.73,
"grad_norm": 0.3336380339194037,
"learning_rate": 0.00018784628468831996,
"loss": 1.2369,
"step": 400
},
{
"epoch": 0.73,
"grad_norm": 0.5054330893305989,
"learning_rate": 0.0001877447995019596,
"loss": 1.2443,
"step": 401
},
{
"epoch": 0.73,
"grad_norm": 0.2297866279715136,
"learning_rate": 0.0001876429200176108,
"loss": 1.2376,
"step": 402
},
{
"epoch": 0.73,
"grad_norm": 0.39350567174184636,
"learning_rate": 0.00018754064669308858,
"loss": 1.2126,
"step": 403
},
{
"epoch": 0.74,
"grad_norm": 0.2025361091435325,
"learning_rate": 0.00018743797998797753,
"loss": 1.2224,
"step": 404
},
{
"epoch": 0.74,
"grad_norm": 0.31824903419753814,
"learning_rate": 0.00018733492036363005,
"loss": 1.2942,
"step": 405
},
{
"epoch": 0.74,
"grad_norm": 2.4642066748643017,
"learning_rate": 0.00018723146828316428,
"loss": 1.2515,
"step": 406
},
{
"epoch": 0.74,
"grad_norm": 0.7833055646295342,
"learning_rate": 0.00018712762421146183,
"loss": 1.2207,
"step": 407
},
{
"epoch": 0.74,
"grad_norm": 0.2810249021786599,
"learning_rate": 0.00018702338861516587,
"loss": 1.2755,
"step": 408
},
{
"epoch": 0.74,
"grad_norm": 0.460995724241333,
"learning_rate": 0.0001869187619626789,
"loss": 1.2856,
"step": 409
},
{
"epoch": 0.75,
"grad_norm": 0.49139203044984286,
"learning_rate": 0.00018681374472416073,
"loss": 1.2392,
"step": 410
},
{
"epoch": 0.75,
"grad_norm": 2.691604613969173,
"learning_rate": 0.0001867083373715264,
"loss": 1.2992,
"step": 411
},
{
"epoch": 0.75,
"grad_norm": 0.8014112047318501,
"learning_rate": 0.00018660254037844388,
"loss": 1.2683,
"step": 412
},
{
"epoch": 0.75,
"grad_norm": 0.31614342841331383,
"learning_rate": 0.00018649635422033215,
"loss": 1.2356,
"step": 413
},
{
"epoch": 0.75,
"grad_norm": 0.2559855196513244,
"learning_rate": 0.000186389779374359,
"loss": 1.2053,
"step": 414
},
{
"epoch": 0.76,
"grad_norm": 1.6613999986014714,
"learning_rate": 0.0001862828163194388,
"loss": 1.2568,
"step": 415
},
{
"epoch": 0.76,
"grad_norm": 0.27190082167109786,
"learning_rate": 0.0001861754655362304,
"loss": 1.1288,
"step": 416
},
{
"epoch": 0.76,
"grad_norm": 0.43819582203066043,
"learning_rate": 0.00018606772750713504,
"loss": 1.1758,
"step": 417
},
{
"epoch": 0.76,
"grad_norm": 0.41738497400383384,
"learning_rate": 0.0001859596027162941,
"loss": 1.2993,
"step": 418
},
{
"epoch": 0.76,
"grad_norm": 0.2595142634740817,
"learning_rate": 0.000185851091649587,
"loss": 1.269,
"step": 419
},
{
"epoch": 0.77,
"grad_norm": 0.2795314201020271,
"learning_rate": 0.00018574219479462878,
"loss": 1.1915,
"step": 420
},
{
"epoch": 0.77,
"grad_norm": 0.2502992494749938,
"learning_rate": 0.00018563291264076835,
"loss": 1.2157,
"step": 421
},
{
"epoch": 0.77,
"grad_norm": 0.27422512335538374,
"learning_rate": 0.00018552324567908585,
"loss": 1.2541,
"step": 422
},
{
"epoch": 0.77,
"grad_norm": 1.3360989016060905,
"learning_rate": 0.00018541319440239066,
"loss": 1.2666,
"step": 423
},
{
"epoch": 0.77,
"grad_norm": 1.9441131913572127,
"learning_rate": 0.00018530275930521924,
"loss": 1.2924,
"step": 424
},
{
"epoch": 0.77,
"grad_norm": 0.3772373301771213,
"learning_rate": 0.00018519194088383273,
"loss": 1.1952,
"step": 425
},
{
"epoch": 0.78,
"grad_norm": 0.22091753616251295,
"learning_rate": 0.0001850807396362148,
"loss": 1.1858,
"step": 426
},
{
"epoch": 0.78,
"grad_norm": 0.21423504993321807,
"learning_rate": 0.00018496915606206951,
"loss": 1.2245,
"step": 427
},
{
"epoch": 0.78,
"grad_norm": 0.5238946238105926,
"learning_rate": 0.00018485719066281892,
"loss": 1.2351,
"step": 428
},
{
"epoch": 0.78,
"grad_norm": 0.3037858949309141,
"learning_rate": 0.0001847448439416009,
"loss": 1.1669,
"step": 429
},
{
"epoch": 0.78,
"grad_norm": 0.21553286799952254,
"learning_rate": 0.00018463211640326686,
"loss": 1.1454,
"step": 430
},
{
"epoch": 0.79,
"grad_norm": 0.202875547805464,
"learning_rate": 0.0001845190085543795,
"loss": 1.188,
"step": 431
},
{
"epoch": 0.79,
"grad_norm": 0.24385408620619278,
"learning_rate": 0.00018440552090321047,
"loss": 1.2307,
"step": 432
},
{
"epoch": 0.79,
"grad_norm": 0.23793944272430378,
"learning_rate": 0.0001842916539597382,
"loss": 1.2253,
"step": 433
},
{
"epoch": 0.79,
"grad_norm": 0.17062488448810784,
"learning_rate": 0.0001841774082356455,
"loss": 1.2681,
"step": 434
},
{
"epoch": 0.79,
"grad_norm": 0.20003742001916064,
"learning_rate": 0.00018406278424431736,
"loss": 1.2428,
"step": 435
},
{
"epoch": 0.79,
"grad_norm": 0.2696052831337752,
"learning_rate": 0.0001839477825008385,
"loss": 1.2945,
"step": 436
},
{
"epoch": 0.8,
"grad_norm": 0.23302960820538443,
"learning_rate": 0.00018383240352199117,
"loss": 1.1718,
"step": 437
},
{
"epoch": 0.8,
"grad_norm": 0.38187833239777536,
"learning_rate": 0.00018371664782625287,
"loss": 1.2311,
"step": 438
},
{
"epoch": 0.8,
"grad_norm": 0.4052561772533732,
"learning_rate": 0.00018360051593379383,
"loss": 1.1639,
"step": 439
},
{
"epoch": 0.8,
"grad_norm": 0.23379763821020377,
"learning_rate": 0.0001834840083664749,
"loss": 1.1809,
"step": 440
},
{
"epoch": 0.8,
"grad_norm": 0.2368414607613928,
"learning_rate": 0.00018336712564784503,
"loss": 1.2357,
"step": 441
},
{
"epoch": 0.81,
"grad_norm": 0.20230633988510938,
"learning_rate": 0.000183249868303139,
"loss": 1.1851,
"step": 442
},
{
"epoch": 0.81,
"grad_norm": 0.170513157244292,
"learning_rate": 0.00018313223685927505,
"loss": 1.205,
"step": 443
},
{
"epoch": 0.81,
"grad_norm": 0.18082295035256266,
"learning_rate": 0.0001830142318448525,
"loss": 1.2305,
"step": 444
},
{
"epoch": 0.81,
"grad_norm": 0.18286299264146286,
"learning_rate": 0.00018289585379014942,
"loss": 1.23,
"step": 445
},
{
"epoch": 0.81,
"grad_norm": 0.17868104103482751,
"learning_rate": 0.00018277710322712012,
"loss": 1.2894,
"step": 446
},
{
"epoch": 0.81,
"grad_norm": 0.1820411127336495,
"learning_rate": 0.00018265798068939294,
"loss": 1.2395,
"step": 447
},
{
"epoch": 0.82,
"grad_norm": 0.1738237541783663,
"learning_rate": 0.0001825384867122677,
"loss": 1.1576,
"step": 448
},
{
"epoch": 0.82,
"grad_norm": 0.15693445967795147,
"learning_rate": 0.0001824186218327134,
"loss": 1.0809,
"step": 449
},
{
"epoch": 0.82,
"grad_norm": 0.18509145652208978,
"learning_rate": 0.00018229838658936564,
"loss": 1.2717,
"step": 450
},
{
"epoch": 0.82,
"grad_norm": 0.14702488366564262,
"learning_rate": 0.0001821777815225245,
"loss": 1.2236,
"step": 451
},
{
"epoch": 0.82,
"grad_norm": 0.1828399354418095,
"learning_rate": 0.00018205680717415187,
"loss": 1.2565,
"step": 452
},
{
"epoch": 0.83,
"grad_norm": 0.17460984182013486,
"learning_rate": 0.00018193546408786898,
"loss": 1.2474,
"step": 453
},
{
"epoch": 0.83,
"grad_norm": 0.2001623109673152,
"learning_rate": 0.00018181375280895416,
"loss": 1.2544,
"step": 454
},
{
"epoch": 0.83,
"grad_norm": 0.17228631742863837,
"learning_rate": 0.00018169167388434025,
"loss": 1.1851,
"step": 455
},
{
"epoch": 0.83,
"grad_norm": 0.1644862232819482,
"learning_rate": 0.00018156922786261216,
"loss": 1.1817,
"step": 456
},
{
"epoch": 0.83,
"grad_norm": 0.19775186397477057,
"learning_rate": 0.00018144641529400446,
"loss": 1.257,
"step": 457
},
{
"epoch": 0.83,
"grad_norm": 0.1626281991220394,
"learning_rate": 0.00018132323673039885,
"loss": 1.2277,
"step": 458
},
{
"epoch": 0.84,
"grad_norm": 0.16158256707311264,
"learning_rate": 0.00018119969272532166,
"loss": 1.1624,
"step": 459
},
{
"epoch": 0.84,
"grad_norm": 0.17705809207051687,
"learning_rate": 0.00018107578383394146,
"loss": 1.2421,
"step": 460
},
{
"epoch": 0.84,
"grad_norm": 0.17639060401882287,
"learning_rate": 0.00018095151061306645,
"loss": 1.285,
"step": 461
},
{
"epoch": 0.84,
"grad_norm": 0.16918796486576196,
"learning_rate": 0.00018082687362114212,
"loss": 1.2606,
"step": 462
},
{
"epoch": 0.84,
"grad_norm": 0.15968377185965665,
"learning_rate": 0.0001807018734182485,
"loss": 1.194,
"step": 463
},
{
"epoch": 0.85,
"grad_norm": 0.17537027967397978,
"learning_rate": 0.00018057651056609784,
"loss": 1.1594,
"step": 464
},
{
"epoch": 0.85,
"grad_norm": 0.15753665403127565,
"learning_rate": 0.00018045078562803203,
"loss": 1.1382,
"step": 465
},
{
"epoch": 0.85,
"grad_norm": 0.17121200763916436,
"learning_rate": 0.00018032469916902003,
"loss": 1.2286,
"step": 466
},
{
"epoch": 0.85,
"grad_norm": 0.19120510133331003,
"learning_rate": 0.00018019825175565542,
"loss": 1.2835,
"step": 467
},
{
"epoch": 0.85,
"grad_norm": 0.1671735980123817,
"learning_rate": 0.0001800714439561538,
"loss": 1.2201,
"step": 468
},
{
"epoch": 0.85,
"grad_norm": 0.1579098534969056,
"learning_rate": 0.00017994427634035015,
"loss": 1.2156,
"step": 469
},
{
"epoch": 0.86,
"grad_norm": 0.1746075421158512,
"learning_rate": 0.00017981674947969636,
"loss": 1.2049,
"step": 470
},
{
"epoch": 0.86,
"grad_norm": 0.16878182886737042,
"learning_rate": 0.00017968886394725874,
"loss": 1.2204,
"step": 471
},
{
"epoch": 0.86,
"grad_norm": 0.16725956538286493,
"learning_rate": 0.00017956062031771535,
"loss": 1.2091,
"step": 472
},
{
"epoch": 0.86,
"grad_norm": 0.18877845951705005,
"learning_rate": 0.00017943201916735335,
"loss": 1.241,
"step": 473
},
{
"epoch": 0.86,
"grad_norm": 0.180337447476004,
"learning_rate": 0.00017930306107406653,
"loss": 1.2253,
"step": 474
},
{
"epoch": 0.87,
"grad_norm": 0.16688572366717752,
"learning_rate": 0.0001791737466173527,
"loss": 1.239,
"step": 475
},
{
"epoch": 0.87,
"grad_norm": 0.15385917621135983,
"learning_rate": 0.00017904407637831099,
"loss": 1.2476,
"step": 476
},
{
"epoch": 0.87,
"grad_norm": 0.17725645269055587,
"learning_rate": 0.00017891405093963938,
"loss": 1.2599,
"step": 477
},
{
"epoch": 0.87,
"grad_norm": 0.14758551718901028,
"learning_rate": 0.00017878367088563195,
"loss": 1.2249,
"step": 478
},
{
"epoch": 0.87,
"grad_norm": 0.15216962408661316,
"learning_rate": 0.00017865293680217637,
"loss": 1.2346,
"step": 479
},
{
"epoch": 0.87,
"grad_norm": 0.16679282848599514,
"learning_rate": 0.00017852184927675112,
"loss": 1.2443,
"step": 480
},
{
"epoch": 0.88,
"grad_norm": 0.16723562739069214,
"learning_rate": 0.00017839040889842305,
"loss": 1.224,
"step": 481
},
{
"epoch": 0.88,
"grad_norm": 0.15922276239929914,
"learning_rate": 0.00017825861625784455,
"loss": 1.2739,
"step": 482
},
{
"epoch": 0.88,
"grad_norm": 0.1510107938469514,
"learning_rate": 0.00017812647194725094,
"loss": 1.1764,
"step": 483
},
{
"epoch": 0.88,
"grad_norm": 0.16446999054333494,
"learning_rate": 0.00017799397656045792,
"loss": 1.2498,
"step": 484
},
{
"epoch": 0.88,
"grad_norm": 0.18566301651865832,
"learning_rate": 0.00017786113069285874,
"loss": 1.232,
"step": 485
},
{
"epoch": 0.89,
"grad_norm": 0.20592971655306183,
"learning_rate": 0.00017772793494142167,
"loss": 1.1586,
"step": 486
},
{
"epoch": 0.89,
"grad_norm": 0.1581947714375729,
"learning_rate": 0.00017759438990468725,
"loss": 1.2502,
"step": 487
},
{
"epoch": 0.89,
"grad_norm": 0.15466760695169174,
"learning_rate": 0.00017746049618276545,
"loss": 1.1605,
"step": 488
},
{
"epoch": 0.89,
"grad_norm": 0.16041506222444918,
"learning_rate": 0.00017732625437733335,
"loss": 1.2778,
"step": 489
},
{
"epoch": 0.89,
"grad_norm": 0.17168109661676773,
"learning_rate": 0.0001771916650916321,
"loss": 1.262,
"step": 490
},
{
"epoch": 0.89,
"grad_norm": 0.1788973186498254,
"learning_rate": 0.00017705672893046425,
"loss": 1.2111,
"step": 491
},
{
"epoch": 0.9,
"grad_norm": 0.1759644359346382,
"learning_rate": 0.00017692144650019125,
"loss": 1.2546,
"step": 492
},
{
"epoch": 0.9,
"grad_norm": 0.15710749736088767,
"learning_rate": 0.0001767858184087304,
"loss": 1.2487,
"step": 493
},
{
"epoch": 0.9,
"grad_norm": 0.1648235522911144,
"learning_rate": 0.00017664984526555248,
"loss": 1.2469,
"step": 494
},
{
"epoch": 0.9,
"grad_norm": 0.15452607969890703,
"learning_rate": 0.0001765135276816787,
"loss": 1.1855,
"step": 495
},
{
"epoch": 0.9,
"grad_norm": 0.1837695597880219,
"learning_rate": 0.00017637686626967812,
"loss": 1.2185,
"step": 496
},
{
"epoch": 0.91,
"grad_norm": 0.15861390725762364,
"learning_rate": 0.00017623986164366486,
"loss": 1.2056,
"step": 497
},
{
"epoch": 0.91,
"grad_norm": 0.1663260460966887,
"learning_rate": 0.00017610251441929533,
"loss": 1.1242,
"step": 498
},
{
"epoch": 0.91,
"grad_norm": 0.1803309720529981,
"learning_rate": 0.00017596482521376546,
"loss": 1.2938,
"step": 499
},
{
"epoch": 0.91,
"grad_norm": 0.14909085011764342,
"learning_rate": 0.00017582679464580797,
"loss": 1.1953,
"step": 500
},
{
"epoch": 0.91,
"grad_norm": 0.15779022242482527,
"learning_rate": 0.00017568842333568952,
"loss": 1.2792,
"step": 501
},
{
"epoch": 0.91,
"grad_norm": 0.1553327313967345,
"learning_rate": 0.00017554971190520798,
"loss": 1.2286,
"step": 502
},
{
"epoch": 0.92,
"grad_norm": 0.16363964666273684,
"learning_rate": 0.00017541066097768963,
"loss": 1.2753,
"step": 503
},
{
"epoch": 0.92,
"grad_norm": 0.16668099163659675,
"learning_rate": 0.00017527127117798635,
"loss": 1.185,
"step": 504
},
{
"epoch": 0.92,
"grad_norm": 0.13957568397594883,
"learning_rate": 0.0001751315431324727,
"loss": 1.143,
"step": 505
},
{
"epoch": 0.92,
"grad_norm": 0.1553111736740035,
"learning_rate": 0.00017499147746904335,
"loss": 1.2492,
"step": 506
},
{
"epoch": 0.92,
"grad_norm": 0.1691517335818193,
"learning_rate": 0.00017485107481711012,
"loss": 1.2619,
"step": 507
},
{
"epoch": 0.93,
"grad_norm": 0.15480883994395986,
"learning_rate": 0.00017471033580759903,
"loss": 1.2396,
"step": 508
},
{
"epoch": 0.93,
"grad_norm": 0.1451690143792058,
"learning_rate": 0.00017456926107294765,
"loss": 1.1732,
"step": 509
},
{
"epoch": 0.93,
"grad_norm": 0.1524398957482947,
"learning_rate": 0.00017442785124710227,
"loss": 1.2083,
"step": 510
},
{
"epoch": 0.93,
"grad_norm": 0.16790264977550012,
"learning_rate": 0.0001742861069655148,
"loss": 1.2201,
"step": 511
},
{
"epoch": 0.93,
"grad_norm": 0.1529847047636337,
"learning_rate": 0.0001741440288651403,
"loss": 1.243,
"step": 512
},
{
"epoch": 0.93,
"grad_norm": 0.1485875402374676,
"learning_rate": 0.00017400161758443375,
"loss": 1.2053,
"step": 513
},
{
"epoch": 0.94,
"grad_norm": 0.16950094279079617,
"learning_rate": 0.00017385887376334742,
"loss": 1.1944,
"step": 514
},
{
"epoch": 0.94,
"grad_norm": 0.15289337084330445,
"learning_rate": 0.00017371579804332789,
"loss": 1.2503,
"step": 515
},
{
"epoch": 0.94,
"grad_norm": 0.15337063655317973,
"learning_rate": 0.00017357239106731317,
"loss": 1.3092,
"step": 516
},
{
"epoch": 0.94,
"grad_norm": 0.1458937961897621,
"learning_rate": 0.00017342865347972988,
"loss": 1.2244,
"step": 517
},
{
"epoch": 0.94,
"grad_norm": 0.19897118610161338,
"learning_rate": 0.00017328458592649027,
"loss": 1.2238,
"step": 518
},
{
"epoch": 0.95,
"grad_norm": 0.15850805264911003,
"learning_rate": 0.00017314018905498931,
"loss": 1.195,
"step": 519
},
{
"epoch": 0.95,
"grad_norm": 0.14445183074519347,
"learning_rate": 0.00017299546351410197,
"loss": 1.1974,
"step": 520
},
{
"epoch": 0.95,
"grad_norm": 0.18180731722745677,
"learning_rate": 0.00017285040995418,
"loss": 1.2107,
"step": 521
},
{
"epoch": 0.95,
"grad_norm": 0.14943874953193587,
"learning_rate": 0.00017270502902704926,
"loss": 1.1843,
"step": 522
},
{
"epoch": 0.95,
"grad_norm": 0.15767466790910512,
"learning_rate": 0.00017255932138600665,
"loss": 1.1409,
"step": 523
},
{
"epoch": 0.95,
"grad_norm": 0.16402921378654775,
"learning_rate": 0.00017241328768581726,
"loss": 1.2135,
"step": 524
},
{
"epoch": 0.96,
"grad_norm": 0.15526246786505485,
"learning_rate": 0.00017226692858271134,
"loss": 1.2255,
"step": 525
},
{
"epoch": 0.96,
"grad_norm": 0.16608155892622348,
"learning_rate": 0.00017212024473438147,
"loss": 1.2691,
"step": 526
},
{
"epoch": 0.96,
"grad_norm": 0.14913271520144072,
"learning_rate": 0.00017197323679997943,
"loss": 1.1574,
"step": 527
},
{
"epoch": 0.96,
"grad_norm": 0.1471910610421707,
"learning_rate": 0.00017182590544011347,
"loss": 1.2774,
"step": 528
},
{
"epoch": 0.96,
"grad_norm": 0.1417464185073962,
"learning_rate": 0.00017167825131684513,
"loss": 1.2446,
"step": 529
},
{
"epoch": 0.97,
"grad_norm": 0.1610488125634495,
"learning_rate": 0.0001715302750936864,
"loss": 1.2862,
"step": 530
},
{
"epoch": 0.97,
"grad_norm": 0.20227974555123074,
"learning_rate": 0.00017138197743559654,
"loss": 1.207,
"step": 531
},
{
"epoch": 0.97,
"grad_norm": 0.1355502559749413,
"learning_rate": 0.00017123335900897946,
"loss": 1.1019,
"step": 532
},
{
"epoch": 0.97,
"grad_norm": 0.1559423167028215,
"learning_rate": 0.00017108442048168038,
"loss": 1.2549,
"step": 533
},
{
"epoch": 0.97,
"grad_norm": 0.15898973818185586,
"learning_rate": 0.00017093516252298296,
"loss": 1.2705,
"step": 534
},
{
"epoch": 0.97,
"grad_norm": 0.15169569998999652,
"learning_rate": 0.00017078558580360632,
"loss": 1.2454,
"step": 535
},
{
"epoch": 0.98,
"grad_norm": 0.15976111665597925,
"learning_rate": 0.00017063569099570196,
"loss": 1.2585,
"step": 536
},
{
"epoch": 0.98,
"grad_norm": 0.14488877221999352,
"learning_rate": 0.00017048547877285077,
"loss": 1.2169,
"step": 537
},
{
"epoch": 0.98,
"grad_norm": 0.14919533098974924,
"learning_rate": 0.00017033494981006002,
"loss": 1.2358,
"step": 538
},
{
"epoch": 0.98,
"grad_norm": 0.15251746717084805,
"learning_rate": 0.00017018410478376032,
"loss": 1.2241,
"step": 539
},
{
"epoch": 0.98,
"grad_norm": 0.1456060482002663,
"learning_rate": 0.00017003294437180255,
"loss": 1.2298,
"step": 540
},
{
"epoch": 0.99,
"grad_norm": 0.17048886778787248,
"learning_rate": 0.00016988146925345484,
"loss": 1.2707,
"step": 541
},
{
"epoch": 0.99,
"grad_norm": 0.15304381059310815,
"learning_rate": 0.00016972968010939954,
"loss": 1.1498,
"step": 542
},
{
"epoch": 0.99,
"grad_norm": 0.16590055969071696,
"learning_rate": 0.0001695775776217301,
"loss": 1.2481,
"step": 543
},
{
"epoch": 0.99,
"grad_norm": 0.14299575837437278,
"learning_rate": 0.00016942516247394807,
"loss": 1.2058,
"step": 544
},
{
"epoch": 0.99,
"grad_norm": 0.14275107775859475,
"learning_rate": 0.00016927243535095997,
"loss": 1.2178,
"step": 545
},
{
"epoch": 0.99,
"grad_norm": 0.1554250137491414,
"learning_rate": 0.0001691193969390742,
"loss": 1.1197,
"step": 546
},
{
"epoch": 1.0,
"grad_norm": 0.16958418467021688,
"learning_rate": 0.0001689660479259981,
"loss": 1.1768,
"step": 547
},
{
"epoch": 1.0,
"grad_norm": 0.1546216583314497,
"learning_rate": 0.00016881238900083473,
"loss": 1.1741,
"step": 548
},
{
"epoch": 1.0,
"grad_norm": 0.15287056494787424,
"learning_rate": 0.0001686584208540797,
"loss": 1.2328,
"step": 549
},
{
"epoch": 1.0,
"grad_norm": 0.1419329373337611,
"learning_rate": 0.0001685041441776183,
"loss": 1.1743,
"step": 550
},
{
"epoch": 1.0,
"grad_norm": 0.15662192296485464,
"learning_rate": 0.00016834955966472213,
"loss": 1.1861,
"step": 551
},
{
"epoch": 1.01,
"grad_norm": 0.16304778894798697,
"learning_rate": 0.00016819466801004621,
"loss": 1.2045,
"step": 552
},
{
"epoch": 1.01,
"grad_norm": 0.15690090424895087,
"learning_rate": 0.00016803946990962576,
"loss": 1.1553,
"step": 553
},
{
"epoch": 1.01,
"grad_norm": 0.15227523196027068,
"learning_rate": 0.000167883966060873,
"loss": 1.126,
"step": 554
},
{
"epoch": 1.01,
"grad_norm": 0.15442134859978873,
"learning_rate": 0.00016772815716257412,
"loss": 1.169,
"step": 555
},
{
"epoch": 1.01,
"grad_norm": 0.16235578615890994,
"learning_rate": 0.00016757204391488613,
"loss": 1.198,
"step": 556
},
{
"epoch": 1.01,
"grad_norm": 0.16631365293435893,
"learning_rate": 0.00016741562701933367,
"loss": 1.2191,
"step": 557
},
{
"epoch": 1.02,
"grad_norm": 0.15668252106453312,
"learning_rate": 0.0001672589071788059,
"loss": 1.1094,
"step": 558
},
{
"epoch": 1.02,
"grad_norm": 0.14883987789397163,
"learning_rate": 0.00016710188509755329,
"loss": 1.0653,
"step": 559
},
{
"epoch": 1.02,
"grad_norm": 0.14723412204861966,
"learning_rate": 0.00016694456148118452,
"loss": 1.0632,
"step": 560
},
{
"epoch": 1.02,
"grad_norm": 0.2694273773199372,
"learning_rate": 0.00016678693703666325,
"loss": 1.1577,
"step": 561
},
{
"epoch": 1.02,
"grad_norm": 0.3010203762394184,
"learning_rate": 0.00016662901247230502,
"loss": 1.1049,
"step": 562
},
{
"epoch": 1.03,
"grad_norm": 0.16960159731526514,
"learning_rate": 0.0001664707884977739,
"loss": 1.2185,
"step": 563
},
{
"epoch": 1.03,
"grad_norm": 0.1593257551859938,
"learning_rate": 0.00016631226582407952,
"loss": 1.1085,
"step": 564
},
{
"epoch": 1.03,
"grad_norm": 0.1695114549114674,
"learning_rate": 0.00016615344516357378,
"loss": 1.1863,
"step": 565
},
{
"epoch": 1.03,
"grad_norm": 0.17116727204552032,
"learning_rate": 0.00016599432722994755,
"loss": 1.0921,
"step": 566
},
{
"epoch": 1.03,
"grad_norm": 0.16265948588203852,
"learning_rate": 0.00016583491273822765,
"loss": 1.1905,
"step": 567
},
{
"epoch": 1.03,
"grad_norm": 0.16500105151611763,
"learning_rate": 0.00016567520240477344,
"loss": 1.1477,
"step": 568
},
{
"epoch": 1.04,
"grad_norm": 0.15445707774433456,
"learning_rate": 0.00016551519694727381,
"loss": 1.1247,
"step": 569
},
{
"epoch": 1.04,
"grad_norm": 0.17740657850033337,
"learning_rate": 0.0001653548970847438,
"loss": 1.0164,
"step": 570
},
{
"epoch": 1.04,
"grad_norm": 0.16152198390194278,
"learning_rate": 0.0001651943035375214,
"loss": 1.1779,
"step": 571
},
{
"epoch": 1.04,
"grad_norm": 0.16299638453443582,
"learning_rate": 0.00016503341702726426,
"loss": 1.1847,
"step": 572
},
{
"epoch": 1.04,
"grad_norm": 0.1486770667934901,
"learning_rate": 0.00016487223827694672,
"loss": 1.1002,
"step": 573
},
{
"epoch": 1.05,
"grad_norm": 0.15907088594322072,
"learning_rate": 0.00016471076801085615,
"loss": 1.1127,
"step": 574
},
{
"epoch": 1.05,
"grad_norm": 0.15491780431797444,
"learning_rate": 0.00016454900695458998,
"loss": 1.1196,
"step": 575
},
{
"epoch": 1.05,
"grad_norm": 0.16584586610927665,
"learning_rate": 0.00016438695583505242,
"loss": 1.1441,
"step": 576
},
{
"epoch": 1.05,
"grad_norm": 0.15197831968756778,
"learning_rate": 0.00016422461538045103,
"loss": 1.0532,
"step": 577
},
{
"epoch": 1.05,
"grad_norm": 0.15669296934233234,
"learning_rate": 0.00016406198632029357,
"loss": 1.1753,
"step": 578
},
{
"epoch": 1.05,
"grad_norm": 0.1523894998655591,
"learning_rate": 0.0001638990693853848,
"loss": 1.1339,
"step": 579
},
{
"epoch": 1.06,
"grad_norm": 0.1570057583413597,
"learning_rate": 0.000163735865307823,
"loss": 1.1117,
"step": 580
},
{
"epoch": 1.06,
"grad_norm": 0.1637341350288717,
"learning_rate": 0.00016357237482099684,
"loss": 1.1369,
"step": 581
},
{
"epoch": 1.06,
"grad_norm": 0.15703259344496112,
"learning_rate": 0.0001634085986595819,
"loss": 1.1448,
"step": 582
},
{
"epoch": 1.06,
"grad_norm": 0.1672273979359662,
"learning_rate": 0.00016324453755953773,
"loss": 1.1972,
"step": 583
},
{
"epoch": 1.06,
"grad_norm": 0.16032943590057747,
"learning_rate": 0.0001630801922581041,
"loss": 1.1811,
"step": 584
},
{
"epoch": 1.07,
"grad_norm": 0.15414179865764022,
"learning_rate": 0.00016291556349379795,
"loss": 1.1126,
"step": 585
},
{
"epoch": 1.07,
"grad_norm": 0.16133591336152153,
"learning_rate": 0.00016275065200641004,
"loss": 1.0664,
"step": 586
},
{
"epoch": 1.07,
"grad_norm": 0.17496271909044792,
"learning_rate": 0.00016258545853700158,
"loss": 1.2042,
"step": 587
},
{
"epoch": 1.07,
"grad_norm": 0.16189829224809252,
"learning_rate": 0.00016241998382790095,
"loss": 1.1896,
"step": 588
},
{
"epoch": 1.07,
"grad_norm": 0.1508460775909727,
"learning_rate": 0.00016225422862270027,
"loss": 1.1205,
"step": 589
},
{
"epoch": 1.07,
"grad_norm": 0.16848110855261456,
"learning_rate": 0.00016208819366625218,
"loss": 1.1132,
"step": 590
},
{
"epoch": 1.08,
"grad_norm": 0.15996856299765072,
"learning_rate": 0.00016192187970466644,
"loss": 1.1022,
"step": 591
},
{
"epoch": 1.08,
"grad_norm": 0.15697088450293906,
"learning_rate": 0.0001617552874853065,
"loss": 1.0483,
"step": 592
},
{
"epoch": 1.08,
"grad_norm": 0.16338270251123885,
"learning_rate": 0.0001615884177567863,
"loss": 1.1049,
"step": 593
},
{
"epoch": 1.08,
"grad_norm": 0.1605597226615145,
"learning_rate": 0.0001614212712689668,
"loss": 1.1319,
"step": 594
},
{
"epoch": 1.08,
"grad_norm": 0.1694169270615921,
"learning_rate": 0.00016125384877295257,
"loss": 1.2143,
"step": 595
},
{
"epoch": 1.09,
"grad_norm": 0.16306866904399633,
"learning_rate": 0.00016108615102108855,
"loss": 1.164,
"step": 596
},
{
"epoch": 1.09,
"grad_norm": 0.1646758516051161,
"learning_rate": 0.00016091817876695655,
"loss": 1.2176,
"step": 597
},
{
"epoch": 1.09,
"grad_norm": 0.16334394129553162,
"learning_rate": 0.00016074993276537198,
"loss": 1.1526,
"step": 598
},
{
"epoch": 1.09,
"grad_norm": 0.16103491072240317,
"learning_rate": 0.00016058141377238026,
"loss": 1.1022,
"step": 599
},
{
"epoch": 1.09,
"grad_norm": 0.16029258404100036,
"learning_rate": 0.00016041262254525362,
"loss": 1.1508,
"step": 600
},
{
"epoch": 1.09,
"grad_norm": 0.15686217802450533,
"learning_rate": 0.00016024355984248768,
"loss": 1.1471,
"step": 601
},
{
"epoch": 1.1,
"grad_norm": 0.1520812354157866,
"learning_rate": 0.0001600742264237979,
"loss": 1.0994,
"step": 602
},
{
"epoch": 1.1,
"grad_norm": 0.161954690104593,
"learning_rate": 0.0001599046230501163,
"loss": 1.1475,
"step": 603
},
{
"epoch": 1.1,
"grad_norm": 0.16470046818858164,
"learning_rate": 0.00015973475048358795,
"loss": 1.1425,
"step": 604
},
{
"epoch": 1.1,
"grad_norm": 0.17783426542152342,
"learning_rate": 0.00015956460948756765,
"loss": 1.1933,
"step": 605
},
{
"epoch": 1.1,
"grad_norm": 0.20834095672163827,
"learning_rate": 0.0001593942008266164,
"loss": 1.1545,
"step": 606
},
{
"epoch": 1.11,
"grad_norm": 0.1601922371369916,
"learning_rate": 0.00015922352526649803,
"loss": 1.1626,
"step": 607
},
{
"epoch": 1.11,
"grad_norm": 0.17590795577463478,
"learning_rate": 0.00015905258357417569,
"loss": 1.1428,
"step": 608
},
{
"epoch": 1.11,
"grad_norm": 0.16438364538914646,
"learning_rate": 0.00015888137651780845,
"loss": 1.1237,
"step": 609
},
{
"epoch": 1.11,
"grad_norm": 0.1532404311365829,
"learning_rate": 0.00015870990486674792,
"loss": 1.1292,
"step": 610
},
{
"epoch": 1.11,
"grad_norm": 0.1658092976448754,
"learning_rate": 0.0001585381693915346,
"loss": 1.2051,
"step": 611
},
{
"epoch": 1.11,
"grad_norm": 0.17009266166893572,
"learning_rate": 0.00015836617086389468,
"loss": 1.1574,
"step": 612
},
{
"epoch": 1.12,
"grad_norm": 0.17117032466097273,
"learning_rate": 0.00015819391005673626,
"loss": 1.1715,
"step": 613
},
{
"epoch": 1.12,
"grad_norm": 0.16766945870038483,
"learning_rate": 0.00015802138774414622,
"loss": 1.1322,
"step": 614
},
{
"epoch": 1.12,
"grad_norm": 0.16603674100502688,
"learning_rate": 0.00015784860470138633,
"loss": 1.1682,
"step": 615
},
{
"epoch": 1.12,
"grad_norm": 0.1538296997751752,
"learning_rate": 0.00015767556170489025,
"loss": 1.0979,
"step": 616
},
{
"epoch": 1.12,
"grad_norm": 0.16355862351540293,
"learning_rate": 0.00015750225953225968,
"loss": 1.0859,
"step": 617
},
{
"epoch": 1.13,
"grad_norm": 0.17100029950170442,
"learning_rate": 0.00015732869896226094,
"loss": 1.1513,
"step": 618
},
{
"epoch": 1.13,
"grad_norm": 0.16981502045992727,
"learning_rate": 0.0001571548807748215,
"loss": 1.1665,
"step": 619
},
{
"epoch": 1.13,
"grad_norm": 0.15981257682597005,
"learning_rate": 0.00015698080575102661,
"loss": 1.0862,
"step": 620
},
{
"epoch": 1.13,
"grad_norm": 0.17114664647857297,
"learning_rate": 0.00015680647467311557,
"loss": 1.1883,
"step": 621
},
{
"epoch": 1.13,
"grad_norm": 0.16818704328900685,
"learning_rate": 0.00015663188832447833,
"loss": 1.1345,
"step": 622
},
{
"epoch": 1.13,
"grad_norm": 0.16716958487866437,
"learning_rate": 0.0001564570474896519,
"loss": 1.1613,
"step": 623
},
{
"epoch": 1.14,
"grad_norm": 0.16425076739154273,
"learning_rate": 0.00015628195295431697,
"loss": 1.1153,
"step": 624
},
{
"epoch": 1.14,
"grad_norm": 0.1668267227341853,
"learning_rate": 0.0001561066055052941,
"loss": 1.1796,
"step": 625
},
{
"epoch": 1.14,
"grad_norm": 0.15748311738845247,
"learning_rate": 0.00015593100593054064,
"loss": 1.125,
"step": 626
},
{
"epoch": 1.14,
"grad_norm": 0.1578229320002536,
"learning_rate": 0.00015575515501914668,
"loss": 1.1751,
"step": 627
},
{
"epoch": 1.14,
"grad_norm": 0.17990453419427754,
"learning_rate": 0.0001555790535613318,
"loss": 1.2107,
"step": 628
},
{
"epoch": 1.15,
"grad_norm": 0.17170087096955436,
"learning_rate": 0.0001554027023484416,
"loss": 1.1704,
"step": 629
},
{
"epoch": 1.15,
"grad_norm": 0.1641233552033525,
"learning_rate": 0.00015522610217294375,
"loss": 1.1399,
"step": 630
},
{
"epoch": 1.15,
"grad_norm": 0.22816800313793714,
"learning_rate": 0.00015504925382842487,
"loss": 1.1374,
"step": 631
},
{
"epoch": 1.15,
"grad_norm": 0.1699514304776241,
"learning_rate": 0.00015487215810958675,
"loss": 1.1668,
"step": 632
},
{
"epoch": 1.15,
"grad_norm": 0.16580705156733896,
"learning_rate": 0.00015469481581224272,
"loss": 1.2467,
"step": 633
},
{
"epoch": 1.15,
"grad_norm": 0.1621295978537637,
"learning_rate": 0.0001545172277333142,
"loss": 1.095,
"step": 634
},
{
"epoch": 1.16,
"grad_norm": 0.15661235975760687,
"learning_rate": 0.00015433939467082713,
"loss": 1.1437,
"step": 635
},
{
"epoch": 1.16,
"grad_norm": 0.15913894864881573,
"learning_rate": 0.00015416131742390827,
"loss": 1.1682,
"step": 636
},
{
"epoch": 1.16,
"grad_norm": 0.17078546916822446,
"learning_rate": 0.00015398299679278172,
"loss": 1.1673,
"step": 637
},
{
"epoch": 1.16,
"grad_norm": 0.17593093163481005,
"learning_rate": 0.00015380443357876518,
"loss": 1.1588,
"step": 638
},
{
"epoch": 1.16,
"grad_norm": 0.16563306408223222,
"learning_rate": 0.00015362562858426654,
"loss": 1.1542,
"step": 639
},
{
"epoch": 1.17,
"grad_norm": 0.15824017459526013,
"learning_rate": 0.0001534465826127801,
"loss": 1.136,
"step": 640
},
{
"epoch": 1.17,
"grad_norm": 0.1555748015065671,
"learning_rate": 0.00015326729646888314,
"loss": 1.1744,
"step": 641
},
{
"epoch": 1.17,
"grad_norm": 0.15054650982383877,
"learning_rate": 0.0001530877709582321,
"loss": 1.0996,
"step": 642
},
{
"epoch": 1.17,
"grad_norm": 0.1732483344367302,
"learning_rate": 0.00015290800688755907,
"loss": 1.1375,
"step": 643
},
{
"epoch": 1.17,
"grad_norm": 0.17086276533942882,
"learning_rate": 0.0001527280050646682,
"loss": 1.1399,
"step": 644
},
{
"epoch": 1.17,
"grad_norm": 0.14791498583408913,
"learning_rate": 0.00015254776629843205,
"loss": 1.0112,
"step": 645
},
{
"epoch": 1.18,
"grad_norm": 0.16492159023612152,
"learning_rate": 0.00015236729139878782,
"loss": 1.1179,
"step": 646
},
{
"epoch": 1.18,
"grad_norm": 0.15550107018855622,
"learning_rate": 0.0001521865811767339,
"loss": 1.0872,
"step": 647
},
{
"epoch": 1.18,
"grad_norm": 0.17086993477527834,
"learning_rate": 0.00015200563644432612,
"loss": 1.1747,
"step": 648
},
{
"epoch": 1.18,
"grad_norm": 0.16832629109129904,
"learning_rate": 0.0001518244580146742,
"loss": 1.2623,
"step": 649
},
{
"epoch": 1.18,
"grad_norm": 0.15555098188988337,
"learning_rate": 0.00015164304670193792,
"loss": 1.1172,
"step": 650
},
{
"epoch": 1.19,
"grad_norm": 0.16504655864531245,
"learning_rate": 0.00015146140332132358,
"loss": 1.1615,
"step": 651
},
{
"epoch": 1.19,
"grad_norm": 0.15904615747067738,
"learning_rate": 0.00015127952868908043,
"loss": 1.1841,
"step": 652
},
{
"epoch": 1.19,
"grad_norm": 0.1664190253819244,
"learning_rate": 0.00015109742362249672,
"loss": 1.2258,
"step": 653
},
{
"epoch": 1.19,
"grad_norm": 0.17622098883404838,
"learning_rate": 0.00015091508893989633,
"loss": 1.2141,
"step": 654
},
{
"epoch": 1.19,
"grad_norm": 0.17253106779139057,
"learning_rate": 0.00015073252546063493,
"loss": 1.1289,
"step": 655
},
{
"epoch": 1.19,
"grad_norm": 0.15924181512457228,
"learning_rate": 0.0001505497340050963,
"loss": 1.1836,
"step": 656
},
{
"epoch": 1.2,
"grad_norm": 0.1755114666296529,
"learning_rate": 0.00015036671539468878,
"loss": 1.2109,
"step": 657
},
{
"epoch": 1.2,
"grad_norm": 0.1623211151951732,
"learning_rate": 0.00015018347045184132,
"loss": 1.1922,
"step": 658
},
{
"epoch": 1.2,
"grad_norm": 0.16895842955233992,
"learning_rate": 0.00015000000000000001,
"loss": 1.2402,
"step": 659
},
{
"epoch": 1.2,
"grad_norm": 0.17126264186496687,
"learning_rate": 0.00014981630486362435,
"loss": 1.2327,
"step": 660
},
{
"epoch": 1.2,
"grad_norm": 0.19632534974860108,
"learning_rate": 0.00014963238586818345,
"loss": 1.1586,
"step": 661
},
{
"epoch": 1.21,
"grad_norm": 0.16708414744148634,
"learning_rate": 0.00014944824384015236,
"loss": 1.1766,
"step": 662
},
{
"epoch": 1.21,
"grad_norm": 0.1800283830073218,
"learning_rate": 0.00014926387960700842,
"loss": 1.0902,
"step": 663
},
{
"epoch": 1.21,
"grad_norm": 0.17570100482291343,
"learning_rate": 0.0001490792939972275,
"loss": 1.1875,
"step": 664
},
{
"epoch": 1.21,
"grad_norm": 0.15904497800793038,
"learning_rate": 0.0001488944878402802,
"loss": 1.083,
"step": 665
},
{
"epoch": 1.21,
"grad_norm": 0.1629549994998532,
"learning_rate": 0.00014870946196662822,
"loss": 1.1505,
"step": 666
},
{
"epoch": 1.21,
"grad_norm": 0.17495490666041913,
"learning_rate": 0.00014852421720772062,
"loss": 1.1107,
"step": 667
},
{
"epoch": 1.22,
"grad_norm": 0.1814722009088628,
"learning_rate": 0.00014833875439599004,
"loss": 1.2089,
"step": 668
},
{
"epoch": 1.22,
"grad_norm": 0.16312442272141373,
"learning_rate": 0.00014815307436484898,
"loss": 1.1518,
"step": 669
},
{
"epoch": 1.22,
"grad_norm": 0.18878089936272002,
"learning_rate": 0.00014796717794868607,
"loss": 1.1806,
"step": 670
},
{
"epoch": 1.22,
"grad_norm": 0.171347321491223,
"learning_rate": 0.00014778106598286234,
"loss": 1.3189,
"step": 671
},
{
"epoch": 1.22,
"grad_norm": 0.15561777538677532,
"learning_rate": 0.00014759473930370736,
"loss": 1.1071,
"step": 672
},
{
"epoch": 1.23,
"grad_norm": 0.16095434527629326,
"learning_rate": 0.0001474081987485156,
"loss": 1.173,
"step": 673
},
{
"epoch": 1.23,
"grad_norm": 0.1594519070011647,
"learning_rate": 0.00014722144515554264,
"loss": 1.1099,
"step": 674
},
{
"epoch": 1.23,
"grad_norm": 0.15853021638823037,
"learning_rate": 0.00014703447936400134,
"loss": 1.1563,
"step": 675
},
{
"epoch": 1.23,
"grad_norm": 0.21942200368949344,
"learning_rate": 0.00014684730221405814,
"loss": 1.1226,
"step": 676
},
{
"epoch": 1.23,
"grad_norm": 0.16328115550493413,
"learning_rate": 0.00014665991454682924,
"loss": 1.1878,
"step": 677
},
{
"epoch": 1.23,
"grad_norm": 0.16551970389762746,
"learning_rate": 0.00014647231720437686,
"loss": 1.055,
"step": 678
},
{
"epoch": 1.24,
"grad_norm": 0.17275962400330097,
"learning_rate": 0.00014628451102970547,
"loss": 1.1302,
"step": 679
},
{
"epoch": 1.24,
"grad_norm": 0.17269943862734122,
"learning_rate": 0.00014609649686675785,
"loss": 1.1635,
"step": 680
},
{
"epoch": 1.24,
"grad_norm": 0.18932844828433326,
"learning_rate": 0.00014590827556041158,
"loss": 1.0954,
"step": 681
},
{
"epoch": 1.24,
"grad_norm": 0.1670913427771278,
"learning_rate": 0.00014571984795647494,
"loss": 1.1336,
"step": 682
},
{
"epoch": 1.24,
"grad_norm": 0.1727053222121231,
"learning_rate": 0.00014553121490168332,
"loss": 1.1477,
"step": 683
},
{
"epoch": 1.25,
"grad_norm": 0.20281162440108022,
"learning_rate": 0.00014534237724369534,
"loss": 1.1364,
"step": 684
},
{
"epoch": 1.25,
"grad_norm": 0.17595782723751183,
"learning_rate": 0.00014515333583108896,
"loss": 1.1977,
"step": 685
},
{
"epoch": 1.25,
"grad_norm": 0.16755958800968615,
"learning_rate": 0.00014496409151335785,
"loss": 1.1315,
"step": 686
},
{
"epoch": 1.25,
"grad_norm": 0.17908885139779734,
"learning_rate": 0.00014477464514090743,
"loss": 1.1851,
"step": 687
},
{
"epoch": 1.25,
"grad_norm": 0.16047769884625865,
"learning_rate": 0.00014458499756505116,
"loss": 1.1137,
"step": 688
},
{
"epoch": 1.26,
"grad_norm": 0.18864957014640948,
"learning_rate": 0.00014439514963800648,
"loss": 1.1962,
"step": 689
},
{
"epoch": 1.26,
"grad_norm": 0.16074020271372416,
"learning_rate": 0.00014420510221289137,
"loss": 1.1448,
"step": 690
},
{
"epoch": 1.26,
"grad_norm": 0.17166433855606958,
"learning_rate": 0.00014401485614372008,
"loss": 1.0487,
"step": 691
},
{
"epoch": 1.26,
"grad_norm": 0.1773682331611409,
"learning_rate": 0.0001438244122853996,
"loss": 1.16,
"step": 692
},
{
"epoch": 1.26,
"grad_norm": 0.16596601643110168,
"learning_rate": 0.00014363377149372584,
"loss": 1.0843,
"step": 693
},
{
"epoch": 1.26,
"grad_norm": 0.16748936526973612,
"learning_rate": 0.0001434429346253794,
"loss": 1.1619,
"step": 694
},
{
"epoch": 1.27,
"grad_norm": 0.17600274748860967,
"learning_rate": 0.00014325190253792222,
"loss": 1.151,
"step": 695
},
{
"epoch": 1.27,
"grad_norm": 0.1721019830304793,
"learning_rate": 0.0001430606760897934,
"loss": 1.0609,
"step": 696
},
{
"epoch": 1.27,
"grad_norm": 0.172361054646693,
"learning_rate": 0.00014286925614030542,
"loss": 1.1777,
"step": 697
},
{
"epoch": 1.27,
"grad_norm": 0.1938123204770535,
"learning_rate": 0.00014267764354964038,
"loss": 1.1078,
"step": 698
},
{
"epoch": 1.27,
"grad_norm": 0.171152731950173,
"learning_rate": 0.00014248583917884594,
"loss": 1.2344,
"step": 699
},
{
"epoch": 1.28,
"grad_norm": 0.176915280903416,
"learning_rate": 0.00014229384388983167,
"loss": 1.1443,
"step": 700
},
{
"epoch": 1.28,
"grad_norm": 0.16620426221482948,
"learning_rate": 0.00014210165854536494,
"loss": 1.1635,
"step": 701
},
{
"epoch": 1.28,
"grad_norm": 0.17368168932295722,
"learning_rate": 0.0001419092840090673,
"loss": 1.2076,
"step": 702
},
{
"epoch": 1.28,
"grad_norm": 0.16396003568599715,
"learning_rate": 0.0001417167211454104,
"loss": 1.1798,
"step": 703
},
{
"epoch": 1.28,
"grad_norm": 0.17474861300401356,
"learning_rate": 0.0001415239708197122,
"loss": 1.1951,
"step": 704
},
{
"epoch": 1.28,
"grad_norm": 0.17687469132536782,
"learning_rate": 0.00014133103389813302,
"loss": 1.1863,
"step": 705
},
{
"epoch": 1.29,
"grad_norm": 0.1729350163866909,
"learning_rate": 0.0001411379112476717,
"loss": 1.1791,
"step": 706
},
{
"epoch": 1.29,
"grad_norm": 0.16183372577765218,
"learning_rate": 0.0001409446037361617,
"loss": 1.11,
"step": 707
},
{
"epoch": 1.29,
"grad_norm": 0.17468439327379603,
"learning_rate": 0.0001407511122322672,
"loss": 1.1459,
"step": 708
},
{
"epoch": 1.29,
"grad_norm": 0.1746393388604993,
"learning_rate": 0.00014055743760547917,
"loss": 1.1782,
"step": 709
},
{
"epoch": 1.29,
"grad_norm": 0.1709730768079813,
"learning_rate": 0.00014036358072611147,
"loss": 1.1729,
"step": 710
},
{
"epoch": 1.3,
"grad_norm": 0.16338226575160056,
"learning_rate": 0.00014016954246529696,
"loss": 1.1637,
"step": 711
},
{
"epoch": 1.3,
"grad_norm": 0.16522796461612693,
"learning_rate": 0.00013997532369498355,
"loss": 1.1786,
"step": 712
},
{
"epoch": 1.3,
"grad_norm": 0.16443017398263532,
"learning_rate": 0.0001397809252879303,
"loss": 1.1885,
"step": 713
},
{
"epoch": 1.3,
"grad_norm": 0.16678395716368374,
"learning_rate": 0.0001395863481177036,
"loss": 1.1576,
"step": 714
},
{
"epoch": 1.3,
"grad_norm": 0.15796662849679724,
"learning_rate": 0.000139391593058673,
"loss": 1.1719,
"step": 715
},
{
"epoch": 1.3,
"grad_norm": 0.1605406960709786,
"learning_rate": 0.00013919666098600753,
"loss": 1.1442,
"step": 716
},
{
"epoch": 1.31,
"grad_norm": 0.16720350808668913,
"learning_rate": 0.00013900155277567157,
"loss": 1.1231,
"step": 717
},
{
"epoch": 1.31,
"grad_norm": 0.16025549611436293,
"learning_rate": 0.00013880626930442113,
"loss": 1.0959,
"step": 718
},
{
"epoch": 1.31,
"grad_norm": 0.1682327263659866,
"learning_rate": 0.00013861081144979974,
"loss": 1.1816,
"step": 719
},
{
"epoch": 1.31,
"grad_norm": 0.16843691606816277,
"learning_rate": 0.00013841518009013445,
"loss": 1.1607,
"step": 720
},
{
"epoch": 1.31,
"grad_norm": 0.17753480625595375,
"learning_rate": 0.0001382193761045322,
"loss": 1.2085,
"step": 721
},
{
"epoch": 1.32,
"grad_norm": 0.175444570487372,
"learning_rate": 0.0001380234003728754,
"loss": 1.1883,
"step": 722
},
{
"epoch": 1.32,
"grad_norm": 0.17844909537485765,
"learning_rate": 0.00013782725377581848,
"loss": 1.2548,
"step": 723
},
{
"epoch": 1.32,
"grad_norm": 0.17827531614400435,
"learning_rate": 0.00013763093719478358,
"loss": 1.1753,
"step": 724
},
{
"epoch": 1.32,
"grad_norm": 0.16568782268956592,
"learning_rate": 0.00013743445151195657,
"loss": 1.1269,
"step": 725
},
{
"epoch": 1.32,
"grad_norm": 0.16229148230338714,
"learning_rate": 0.00013723779761028347,
"loss": 1.1233,
"step": 726
},
{
"epoch": 1.32,
"grad_norm": 0.18071555500117806,
"learning_rate": 0.000137040976373466,
"loss": 1.1691,
"step": 727
},
{
"epoch": 1.33,
"grad_norm": 0.16528568438217509,
"learning_rate": 0.000136843988685958,
"loss": 1.1044,
"step": 728
},
{
"epoch": 1.33,
"grad_norm": 0.15621961073605914,
"learning_rate": 0.00013664683543296112,
"loss": 1.0585,
"step": 729
},
{
"epoch": 1.33,
"grad_norm": 0.1821012365890871,
"learning_rate": 0.00013644951750042114,
"loss": 1.1061,
"step": 730
},
{
"epoch": 1.33,
"grad_norm": 0.18832728793079637,
"learning_rate": 0.00013625203577502382,
"loss": 1.2088,
"step": 731
},
{
"epoch": 1.33,
"grad_norm": 0.17821269011741103,
"learning_rate": 0.00013605439114419094,
"loss": 1.1076,
"step": 732
},
{
"epoch": 1.34,
"grad_norm": 0.15816928920099374,
"learning_rate": 0.00013585658449607633,
"loss": 1.057,
"step": 733
},
{
"epoch": 1.34,
"grad_norm": 0.1634582673036166,
"learning_rate": 0.00013565861671956187,
"loss": 1.2206,
"step": 734
},
{
"epoch": 1.34,
"grad_norm": 0.16402056158577089,
"learning_rate": 0.00013546048870425356,
"loss": 1.0809,
"step": 735
},
{
"epoch": 1.34,
"grad_norm": 0.18674575425072043,
"learning_rate": 0.0001352622013404774,
"loss": 1.0979,
"step": 736
},
{
"epoch": 1.34,
"grad_norm": 0.16768128529630819,
"learning_rate": 0.00013506375551927547,
"loss": 1.1813,
"step": 737
},
{
"epoch": 1.34,
"grad_norm": 0.17842009039615817,
"learning_rate": 0.00013486515213240188,
"loss": 1.1767,
"step": 738
},
{
"epoch": 1.35,
"grad_norm": 0.16460609654310757,
"learning_rate": 0.0001346663920723188,
"loss": 1.1711,
"step": 739
},
{
"epoch": 1.35,
"grad_norm": 0.16130836204900895,
"learning_rate": 0.00013446747623219255,
"loss": 1.1506,
"step": 740
},
{
"epoch": 1.35,
"grad_norm": 0.19357518099915463,
"learning_rate": 0.00013426840550588933,
"loss": 1.1454,
"step": 741
},
{
"epoch": 1.35,
"grad_norm": 0.16818483070815926,
"learning_rate": 0.0001340691807879714,
"loss": 1.105,
"step": 742
},
{
"epoch": 1.35,
"grad_norm": 0.17634247767232533,
"learning_rate": 0.00013386980297369307,
"loss": 1.1173,
"step": 743
},
{
"epoch": 1.36,
"grad_norm": 0.17275588243641285,
"learning_rate": 0.0001336702729589965,
"loss": 1.1614,
"step": 744
},
{
"epoch": 1.36,
"grad_norm": 0.18630211892037335,
"learning_rate": 0.00013347059164050794,
"loss": 1.0822,
"step": 745
},
{
"epoch": 1.36,
"grad_norm": 0.1610612417401003,
"learning_rate": 0.0001332707599155334,
"loss": 1.11,
"step": 746
},
{
"epoch": 1.36,
"grad_norm": 0.1710900937487699,
"learning_rate": 0.00013307077868205487,
"loss": 1.1556,
"step": 747
},
{
"epoch": 1.36,
"grad_norm": 0.16789088046962233,
"learning_rate": 0.00013287064883872612,
"loss": 1.166,
"step": 748
},
{
"epoch": 1.36,
"grad_norm": 0.17152792861267624,
"learning_rate": 0.00013267037128486883,
"loss": 1.2076,
"step": 749
},
{
"epoch": 1.37,
"grad_norm": 0.16987939988427406,
"learning_rate": 0.00013246994692046836,
"loss": 1.1892,
"step": 750
},
{
"epoch": 1.37,
"grad_norm": 0.17050807620104472,
"learning_rate": 0.00013226937664616976,
"loss": 1.1405,
"step": 751
},
{
"epoch": 1.37,
"grad_norm": 0.1619859173580362,
"learning_rate": 0.00013206866136327388,
"loss": 1.15,
"step": 752
},
{
"epoch": 1.37,
"grad_norm": 0.1693102604231561,
"learning_rate": 0.00013186780197373306,
"loss": 1.1108,
"step": 753
},
{
"epoch": 1.37,
"grad_norm": 0.15751817301555188,
"learning_rate": 0.00013166679938014726,
"loss": 1.1075,
"step": 754
},
{
"epoch": 1.38,
"grad_norm": 0.17344055609711165,
"learning_rate": 0.00013146565448576004,
"loss": 1.2239,
"step": 755
},
{
"epoch": 1.38,
"grad_norm": 0.16419033762795351,
"learning_rate": 0.00013126436819445422,
"loss": 1.2146,
"step": 756
},
{
"epoch": 1.38,
"grad_norm": 0.17241446484871606,
"learning_rate": 0.00013106294141074825,
"loss": 1.1808,
"step": 757
},
{
"epoch": 1.38,
"grad_norm": 0.16299230805525922,
"learning_rate": 0.0001308613750397917,
"loss": 1.1639,
"step": 758
},
{
"epoch": 1.38,
"grad_norm": 0.15993384139183944,
"learning_rate": 0.00013065966998736155,
"loss": 1.0664,
"step": 759
},
{
"epoch": 1.38,
"grad_norm": 0.17951699730636844,
"learning_rate": 0.00013045782715985792,
"loss": 1.1565,
"step": 760
},
{
"epoch": 1.39,
"grad_norm": 0.16420615622643212,
"learning_rate": 0.0001302558474643,
"loss": 1.1675,
"step": 761
},
{
"epoch": 1.39,
"grad_norm": 0.1718558627273243,
"learning_rate": 0.0001300537318083221,
"loss": 1.1758,
"step": 762
},
{
"epoch": 1.39,
"grad_norm": 0.16040004880583997,
"learning_rate": 0.00012985148110016947,
"loss": 1.1286,
"step": 763
},
{
"epoch": 1.39,
"grad_norm": 0.1644924005506398,
"learning_rate": 0.0001296490962486942,
"loss": 1.0901,
"step": 764
},
{
"epoch": 1.39,
"grad_norm": 0.17508769995003778,
"learning_rate": 0.00012944657816335123,
"loss": 1.158,
"step": 765
},
{
"epoch": 1.4,
"grad_norm": 0.16636235063611218,
"learning_rate": 0.0001292439277541942,
"loss": 1.1506,
"step": 766
},
{
"epoch": 1.4,
"grad_norm": 0.1651752039403403,
"learning_rate": 0.00012904114593187136,
"loss": 1.1167,
"step": 767
},
{
"epoch": 1.4,
"grad_norm": 0.1727449182027404,
"learning_rate": 0.0001288382336076215,
"loss": 1.1208,
"step": 768
},
{
"epoch": 1.4,
"grad_norm": 0.1678888505155555,
"learning_rate": 0.00012863519169326984,
"loss": 1.1164,
"step": 769
},
{
"epoch": 1.4,
"grad_norm": 0.1717914946350261,
"learning_rate": 0.0001284320211012239,
"loss": 1.174,
"step": 770
},
{
"epoch": 1.4,
"grad_norm": 0.15833038102003086,
"learning_rate": 0.00012822872274446958,
"loss": 1.0557,
"step": 771
},
{
"epoch": 1.41,
"grad_norm": 0.16693919668597118,
"learning_rate": 0.00012802529753656668,
"loss": 1.2202,
"step": 772
},
{
"epoch": 1.41,
"grad_norm": 0.1608034670310254,
"learning_rate": 0.0001278217463916453,
"loss": 1.0761,
"step": 773
},
{
"epoch": 1.41,
"grad_norm": 0.16522372158185394,
"learning_rate": 0.0001276180702244012,
"loss": 1.114,
"step": 774
},
{
"epoch": 1.41,
"grad_norm": 0.15699403016449,
"learning_rate": 0.00012741426995009213,
"loss": 1.0487,
"step": 775
},
{
"epoch": 1.41,
"grad_norm": 0.1724966872571711,
"learning_rate": 0.00012721034648453353,
"loss": 1.2198,
"step": 776
},
{
"epoch": 1.42,
"grad_norm": 0.16612867386952304,
"learning_rate": 0.00012700630074409427,
"loss": 1.1481,
"step": 777
},
{
"epoch": 1.42,
"grad_norm": 0.1755115998846381,
"learning_rate": 0.0001268021336456929,
"loss": 1.1295,
"step": 778
},
{
"epoch": 1.42,
"grad_norm": 0.16610612624318807,
"learning_rate": 0.00012659784610679318,
"loss": 1.0735,
"step": 779
},
{
"epoch": 1.42,
"grad_norm": 0.17210665028457767,
"learning_rate": 0.0001263934390454001,
"loss": 1.128,
"step": 780
},
{
"epoch": 1.42,
"grad_norm": 0.17115180879996908,
"learning_rate": 0.00012618891338005573,
"loss": 1.2114,
"step": 781
},
{
"epoch": 1.42,
"grad_norm": 0.1708121456807184,
"learning_rate": 0.0001259842700298352,
"loss": 1.2037,
"step": 782
},
{
"epoch": 1.43,
"grad_norm": 0.16413369503599784,
"learning_rate": 0.00012577950991434248,
"loss": 1.115,
"step": 783
},
{
"epoch": 1.43,
"grad_norm": 0.16575779512688402,
"learning_rate": 0.0001255746339537061,
"loss": 1.13,
"step": 784
},
{
"epoch": 1.43,
"grad_norm": 0.16787768159350958,
"learning_rate": 0.00012536964306857526,
"loss": 1.1187,
"step": 785
},
{
"epoch": 1.43,
"grad_norm": 0.161257935533138,
"learning_rate": 0.00012516453818011566,
"loss": 1.1272,
"step": 786
},
{
"epoch": 1.43,
"grad_norm": 0.17163817923820932,
"learning_rate": 0.00012495932021000517,
"loss": 1.1371,
"step": 787
},
{
"epoch": 1.44,
"grad_norm": 0.16451801686420228,
"learning_rate": 0.0001247539900804299,
"loss": 1.1234,
"step": 788
},
{
"epoch": 1.44,
"grad_norm": 0.4897545674498291,
"learning_rate": 0.00012454854871407994,
"loss": 1.1276,
"step": 789
},
{
"epoch": 1.44,
"grad_norm": 0.1674813107615878,
"learning_rate": 0.00012434299703414524,
"loss": 1.0717,
"step": 790
},
{
"epoch": 1.44,
"grad_norm": 0.16889265391095676,
"learning_rate": 0.0001241373359643114,
"loss": 1.1094,
"step": 791
},
{
"epoch": 1.44,
"grad_norm": 0.17807200243744,
"learning_rate": 0.0001239315664287558,
"loss": 1.2072,
"step": 792
},
{
"epoch": 1.44,
"grad_norm": 0.1821416845887793,
"learning_rate": 0.00012372568935214298,
"loss": 1.1574,
"step": 793
},
{
"epoch": 1.45,
"grad_norm": 0.16787645348154986,
"learning_rate": 0.00012351970565962085,
"loss": 1.0884,
"step": 794
},
{
"epoch": 1.45,
"grad_norm": 0.17767787321450104,
"learning_rate": 0.00012331361627681645,
"loss": 1.1561,
"step": 795
},
{
"epoch": 1.45,
"grad_norm": 0.16990542040230855,
"learning_rate": 0.00012310742212983167,
"loss": 1.1454,
"step": 796
},
{
"epoch": 1.45,
"grad_norm": 0.17040623149046621,
"learning_rate": 0.00012290112414523928,
"loss": 1.1143,
"step": 797
},
{
"epoch": 1.45,
"grad_norm": 0.17383642491755713,
"learning_rate": 0.00012269472325007858,
"loss": 1.1502,
"step": 798
},
{
"epoch": 1.46,
"grad_norm": 0.1670271756098212,
"learning_rate": 0.00012248822037185138,
"loss": 1.1872,
"step": 799
},
{
"epoch": 1.46,
"grad_norm": 0.17338597290651167,
"learning_rate": 0.0001222816164385177,
"loss": 1.201,
"step": 800
},
{
"epoch": 1.46,
"grad_norm": 0.16342647338148947,
"learning_rate": 0.00012207491237849172,
"loss": 1.0994,
"step": 801
},
{
"epoch": 1.46,
"grad_norm": 0.15847559756714308,
"learning_rate": 0.0001218681091206376,
"loss": 1.1272,
"step": 802
},
{
"epoch": 1.46,
"grad_norm": 0.17295023549456498,
"learning_rate": 0.00012166120759426514,
"loss": 1.1026,
"step": 803
},
{
"epoch": 1.46,
"grad_norm": 0.17574482838436642,
"learning_rate": 0.00012145420872912585,
"loss": 1.1785,
"step": 804
},
{
"epoch": 1.47,
"grad_norm": 0.1636907763577917,
"learning_rate": 0.0001212471134554086,
"loss": 1.125,
"step": 805
},
{
"epoch": 1.47,
"grad_norm": 0.1683476239550862,
"learning_rate": 0.00012103992270373547,
"loss": 1.1196,
"step": 806
},
{
"epoch": 1.47,
"grad_norm": 0.1634846923426851,
"learning_rate": 0.00012083263740515765,
"loss": 1.1666,
"step": 807
},
{
"epoch": 1.47,
"grad_norm": 0.20362360607016913,
"learning_rate": 0.00012062525849115107,
"loss": 1.2166,
"step": 808
},
{
"epoch": 1.47,
"grad_norm": 0.15520795046759722,
"learning_rate": 0.00012041778689361254,
"loss": 1.1046,
"step": 809
},
{
"epoch": 1.48,
"grad_norm": 0.17498352958051774,
"learning_rate": 0.00012021022354485514,
"loss": 1.1331,
"step": 810
},
{
"epoch": 1.48,
"grad_norm": 0.17105423893572264,
"learning_rate": 0.00012000256937760445,
"loss": 1.0799,
"step": 811
},
{
"epoch": 1.48,
"grad_norm": 0.16837214432904796,
"learning_rate": 0.00011979482532499401,
"loss": 1.1856,
"step": 812
},
{
"epoch": 1.48,
"grad_norm": 0.1717737899447442,
"learning_rate": 0.00011958699232056134,
"loss": 1.1738,
"step": 813
},
{
"epoch": 1.48,
"grad_norm": 0.16901679352588425,
"learning_rate": 0.0001193790712982437,
"loss": 1.1361,
"step": 814
},
{
"epoch": 1.48,
"grad_norm": 0.39457461047161113,
"learning_rate": 0.00011917106319237386,
"loss": 1.1095,
"step": 815
},
{
"epoch": 1.49,
"grad_norm": 0.1689082831860728,
"learning_rate": 0.00011896296893767587,
"loss": 1.1518,
"step": 816
},
{
"epoch": 1.49,
"grad_norm": 0.17381008421938537,
"learning_rate": 0.00011875478946926093,
"loss": 1.1251,
"step": 817
},
{
"epoch": 1.49,
"grad_norm": 0.1723002103848165,
"learning_rate": 0.00011854652572262323,
"loss": 1.1128,
"step": 818
},
{
"epoch": 1.49,
"grad_norm": 0.1799457776202214,
"learning_rate": 0.00011833817863363564,
"loss": 1.1702,
"step": 819
},
{
"epoch": 1.49,
"grad_norm": 0.18319688210909812,
"learning_rate": 0.00011812974913854545,
"loss": 1.1537,
"step": 820
},
{
"epoch": 1.5,
"grad_norm": 0.1716604871697964,
"learning_rate": 0.0001179212381739704,
"loss": 1.1567,
"step": 821
},
{
"epoch": 1.5,
"grad_norm": 0.17098997576829758,
"learning_rate": 0.00011771264667689427,
"loss": 1.1309,
"step": 822
},
{
"epoch": 1.5,
"grad_norm": 0.2607304915983766,
"learning_rate": 0.00011750397558466273,
"loss": 1.1151,
"step": 823
},
{
"epoch": 1.5,
"grad_norm": 0.1869030292628108,
"learning_rate": 0.00011729522583497912,
"loss": 1.0764,
"step": 824
},
{
"epoch": 1.5,
"grad_norm": 0.1969233515350324,
"learning_rate": 0.00011708639836590023,
"loss": 1.0986,
"step": 825
},
{
"epoch": 1.5,
"grad_norm": 0.17794486028770468,
"learning_rate": 0.00011687749411583213,
"loss": 1.1327,
"step": 826
},
{
"epoch": 1.51,
"grad_norm": 0.19408346665780069,
"learning_rate": 0.00011666851402352588,
"loss": 1.0768,
"step": 827
},
{
"epoch": 1.51,
"grad_norm": 0.5122187922492092,
"learning_rate": 0.00011645945902807341,
"loss": 1.1223,
"step": 828
},
{
"epoch": 1.51,
"grad_norm": 0.18948615786386766,
"learning_rate": 0.00011625033006890315,
"loss": 1.1452,
"step": 829
},
{
"epoch": 1.51,
"grad_norm": 0.16917117240636045,
"learning_rate": 0.00011604112808577603,
"loss": 1.131,
"step": 830
},
{
"epoch": 1.51,
"grad_norm": 0.177460873056157,
"learning_rate": 0.00011583185401878101,
"loss": 1.1691,
"step": 831
},
{
"epoch": 1.52,
"grad_norm": 0.2727518607538819,
"learning_rate": 0.00011562250880833104,
"loss": 1.185,
"step": 832
},
{
"epoch": 1.52,
"grad_norm": 0.1722336921590558,
"learning_rate": 0.00011541309339515869,
"loss": 1.2202,
"step": 833
},
{
"epoch": 1.52,
"grad_norm": 0.16731429410912582,
"learning_rate": 0.00011520360872031209,
"loss": 1.1114,
"step": 834
},
{
"epoch": 1.52,
"grad_norm": 0.1747423744661194,
"learning_rate": 0.00011499405572515059,
"loss": 1.0929,
"step": 835
},
{
"epoch": 1.52,
"grad_norm": 0.18705677818301408,
"learning_rate": 0.00011478443535134049,
"loss": 1.101,
"step": 836
},
{
"epoch": 1.52,
"grad_norm": 0.18090585152937122,
"learning_rate": 0.00011457474854085096,
"loss": 1.2376,
"step": 837
},
{
"epoch": 1.53,
"grad_norm": 0.1572625515956593,
"learning_rate": 0.00011436499623594963,
"loss": 1.1039,
"step": 838
},
{
"epoch": 1.53,
"grad_norm": 0.18868561701551964,
"learning_rate": 0.00011415517937919846,
"loss": 1.1091,
"step": 839
},
{
"epoch": 1.53,
"grad_norm": 0.3765201515989829,
"learning_rate": 0.00011394529891344958,
"loss": 1.1904,
"step": 840
},
{
"epoch": 1.53,
"grad_norm": 0.17659930390592027,
"learning_rate": 0.00011373535578184082,
"loss": 1.1767,
"step": 841
},
{
"epoch": 1.53,
"grad_norm": 0.16250915193456436,
"learning_rate": 0.00011352535092779173,
"loss": 1.0396,
"step": 842
},
{
"epoch": 1.54,
"grad_norm": 0.17679862621844297,
"learning_rate": 0.00011331528529499909,
"loss": 1.0899,
"step": 843
},
{
"epoch": 1.54,
"grad_norm": 0.17303502214887417,
"learning_rate": 0.00011310515982743293,
"loss": 1.1623,
"step": 844
},
{
"epoch": 1.54,
"grad_norm": 0.17241019588300063,
"learning_rate": 0.00011289497546933212,
"loss": 1.1761,
"step": 845
},
{
"epoch": 1.54,
"grad_norm": 0.19940689264510733,
"learning_rate": 0.00011268473316520007,
"loss": 1.186,
"step": 846
},
{
"epoch": 1.54,
"grad_norm": 0.17205111928537736,
"learning_rate": 0.00011247443385980078,
"loss": 1.1669,
"step": 847
},
{
"epoch": 1.54,
"grad_norm": 0.168610859884427,
"learning_rate": 0.0001122640784981542,
"loss": 1.1223,
"step": 848
},
{
"epoch": 1.55,
"grad_norm": 0.1571679218925349,
"learning_rate": 0.0001120536680255323,
"loss": 1.1017,
"step": 849
},
{
"epoch": 1.55,
"grad_norm": 0.1731635445936976,
"learning_rate": 0.00011184320338745467,
"loss": 1.1549,
"step": 850
},
{
"epoch": 1.55,
"grad_norm": 0.16986390394426032,
"learning_rate": 0.00011163268552968423,
"loss": 1.1561,
"step": 851
},
{
"epoch": 1.55,
"grad_norm": 0.17807860518694296,
"learning_rate": 0.00011142211539822318,
"loss": 1.1191,
"step": 852
},
{
"epoch": 1.55,
"grad_norm": 0.1848822566355218,
"learning_rate": 0.0001112114939393085,
"loss": 1.1265,
"step": 853
},
{
"epoch": 1.56,
"grad_norm": 0.16154933469362054,
"learning_rate": 0.00011100082209940795,
"loss": 1.1696,
"step": 854
},
{
"epoch": 1.56,
"grad_norm": 0.17539593545795423,
"learning_rate": 0.00011079010082521557,
"loss": 1.2098,
"step": 855
},
{
"epoch": 1.56,
"grad_norm": 0.16955097983529596,
"learning_rate": 0.00011057933106364758,
"loss": 1.134,
"step": 856
},
{
"epoch": 1.56,
"grad_norm": 0.1627286025891699,
"learning_rate": 0.00011036851376183812,
"loss": 1.0968,
"step": 857
},
{
"epoch": 1.56,
"grad_norm": 0.18241009216885773,
"learning_rate": 0.0001101576498671349,
"loss": 1.0784,
"step": 858
},
{
"epoch": 1.56,
"grad_norm": 0.1602213810951013,
"learning_rate": 0.00010994674032709513,
"loss": 1.0272,
"step": 859
},
{
"epoch": 1.57,
"grad_norm": 0.17710273355839293,
"learning_rate": 0.00010973578608948094,
"loss": 1.1807,
"step": 860
},
{
"epoch": 1.57,
"grad_norm": 0.16846043025211901,
"learning_rate": 0.00010952478810225548,
"loss": 1.1237,
"step": 861
},
{
"epoch": 1.57,
"grad_norm": 0.17578602240771765,
"learning_rate": 0.00010931374731357841,
"loss": 1.1305,
"step": 862
},
{
"epoch": 1.57,
"grad_norm": 0.1741114211076835,
"learning_rate": 0.0001091026646718018,
"loss": 1.0663,
"step": 863
},
{
"epoch": 1.57,
"grad_norm": 0.17419627851243186,
"learning_rate": 0.0001088915411254657,
"loss": 1.0864,
"step": 864
},
{
"epoch": 1.58,
"grad_norm": 0.1941495171655647,
"learning_rate": 0.00010868037762329404,
"loss": 1.1471,
"step": 865
},
{
"epoch": 1.58,
"grad_norm": 0.16828482820770493,
"learning_rate": 0.0001084691751141903,
"loss": 1.1386,
"step": 866
},
{
"epoch": 1.58,
"grad_norm": 0.1812212801001379,
"learning_rate": 0.00010825793454723325,
"loss": 1.193,
"step": 867
},
{
"epoch": 1.58,
"grad_norm": 0.31177412914918357,
"learning_rate": 0.00010804665687167262,
"loss": 1.1416,
"step": 868
},
{
"epoch": 1.58,
"grad_norm": 0.16307908317649347,
"learning_rate": 0.00010783534303692493,
"loss": 1.0742,
"step": 869
},
{
"epoch": 1.58,
"grad_norm": 0.17264040918528867,
"learning_rate": 0.00010762399399256917,
"loss": 1.1196,
"step": 870
},
{
"epoch": 1.59,
"grad_norm": 0.16662882952717148,
"learning_rate": 0.00010741261068834265,
"loss": 1.1095,
"step": 871
},
{
"epoch": 1.59,
"grad_norm": 0.16909860632274898,
"learning_rate": 0.00010720119407413647,
"loss": 1.0978,
"step": 872
},
{
"epoch": 1.59,
"grad_norm": 0.1841227749300958,
"learning_rate": 0.00010698974509999158,
"loss": 1.1249,
"step": 873
},
{
"epoch": 1.59,
"grad_norm": 0.163410587005958,
"learning_rate": 0.00010677826471609422,
"loss": 1.1027,
"step": 874
},
{
"epoch": 1.59,
"grad_norm": 0.16473194468337188,
"learning_rate": 0.00010656675387277182,
"loss": 1.0452,
"step": 875
},
{
"epoch": 1.6,
"grad_norm": 0.16465304101505895,
"learning_rate": 0.00010635521352048872,
"loss": 1.0975,
"step": 876
},
{
"epoch": 1.6,
"grad_norm": 0.1648559678877552,
"learning_rate": 0.00010614364460984176,
"loss": 1.1865,
"step": 877
},
{
"epoch": 1.6,
"grad_norm": 0.17182083260240222,
"learning_rate": 0.00010593204809155628,
"loss": 1.1666,
"step": 878
},
{
"epoch": 1.6,
"grad_norm": 0.262043699845629,
"learning_rate": 0.00010572042491648149,
"loss": 1.1705,
"step": 879
},
{
"epoch": 1.6,
"grad_norm": 0.1642146793634745,
"learning_rate": 0.00010550877603558655,
"loss": 1.0943,
"step": 880
},
{
"epoch": 1.6,
"grad_norm": 0.1740460470258003,
"learning_rate": 0.00010529710239995605,
"loss": 1.0492,
"step": 881
},
{
"epoch": 1.61,
"grad_norm": 0.16424504391420278,
"learning_rate": 0.0001050854049607858,
"loss": 1.0877,
"step": 882
},
{
"epoch": 1.61,
"grad_norm": 0.1754464752434157,
"learning_rate": 0.00010487368466937866,
"loss": 1.2004,
"step": 883
},
{
"epoch": 1.61,
"grad_norm": 0.17019599469065266,
"learning_rate": 0.00010466194247714008,
"loss": 1.1736,
"step": 884
},
{
"epoch": 1.61,
"grad_norm": 0.16453531456339282,
"learning_rate": 0.00010445017933557404,
"loss": 1.0682,
"step": 885
},
{
"epoch": 1.61,
"grad_norm": 0.16576841634828685,
"learning_rate": 0.00010423839619627853,
"loss": 1.1239,
"step": 886
},
{
"epoch": 1.62,
"grad_norm": 0.17959933112877494,
"learning_rate": 0.00010402659401094152,
"loss": 1.1741,
"step": 887
},
{
"epoch": 1.62,
"grad_norm": 0.1764446278189948,
"learning_rate": 0.00010381477373133652,
"loss": 1.1465,
"step": 888
},
{
"epoch": 1.62,
"grad_norm": 0.1704634166756464,
"learning_rate": 0.0001036029363093183,
"loss": 1.2092,
"step": 889
},
{
"epoch": 1.62,
"grad_norm": 0.16952721202785456,
"learning_rate": 0.00010339108269681874,
"loss": 1.1478,
"step": 890
},
{
"epoch": 1.62,
"grad_norm": 0.1787903145977452,
"learning_rate": 0.00010317921384584244,
"loss": 1.1818,
"step": 891
},
{
"epoch": 1.62,
"grad_norm": 0.1623533169508836,
"learning_rate": 0.00010296733070846252,
"loss": 1.1352,
"step": 892
},
{
"epoch": 1.63,
"grad_norm": 0.17651784025914782,
"learning_rate": 0.00010275543423681621,
"loss": 1.175,
"step": 893
},
{
"epoch": 1.63,
"grad_norm": 0.16473007072288753,
"learning_rate": 0.00010254352538310075,
"loss": 1.1837,
"step": 894
},
{
"epoch": 1.63,
"grad_norm": 0.16909571695790346,
"learning_rate": 0.00010233160509956894,
"loss": 1.1435,
"step": 895
},
{
"epoch": 1.63,
"grad_norm": 0.1904744004900901,
"learning_rate": 0.000102119674338525,
"loss": 1.1549,
"step": 896
},
{
"epoch": 1.63,
"grad_norm": 0.17373949711988154,
"learning_rate": 0.00010190773405232024,
"loss": 1.202,
"step": 897
},
{
"epoch": 1.64,
"grad_norm": 0.18192103377408994,
"learning_rate": 0.00010169578519334873,
"loss": 1.1851,
"step": 898
},
{
"epoch": 1.64,
"grad_norm": 0.18229237608440976,
"learning_rate": 0.0001014838287140431,
"loss": 1.1233,
"step": 899
},
{
"epoch": 1.64,
"grad_norm": 0.16940278596266165,
"learning_rate": 0.00010127186556687019,
"loss": 1.169,
"step": 900
},
{
"epoch": 1.64,
"grad_norm": 0.18262078712459354,
"learning_rate": 0.00010105989670432681,
"loss": 1.199,
"step": 901
},
{
"epoch": 1.64,
"grad_norm": 0.16747513400418446,
"learning_rate": 0.00010084792307893552,
"loss": 1.0666,
"step": 902
},
{
"epoch": 1.64,
"grad_norm": 0.1746037115339502,
"learning_rate": 0.00010063594564324012,
"loss": 1.1885,
"step": 903
},
{
"epoch": 1.65,
"grad_norm": 0.1732153359208303,
"learning_rate": 0.00010042396534980176,
"loss": 1.1155,
"step": 904
},
{
"epoch": 1.65,
"grad_norm": 0.18911672921606115,
"learning_rate": 0.00010021198315119424,
"loss": 1.1495,
"step": 905
},
{
"epoch": 1.65,
"grad_norm": 0.1810967486335703,
"learning_rate": 0.0001,
"loss": 1.2386,
"step": 906
},
{
"epoch": 1.65,
"grad_norm": 0.16208600336701465,
"learning_rate": 9.978801684880578e-05,
"loss": 1.1016,
"step": 907
},
{
"epoch": 1.65,
"grad_norm": 0.17108242412955507,
"learning_rate": 9.957603465019826e-05,
"loss": 1.1659,
"step": 908
},
{
"epoch": 1.66,
"grad_norm": 0.17777524975299183,
"learning_rate": 9.93640543567599e-05,
"loss": 1.1877,
"step": 909
},
{
"epoch": 1.66,
"grad_norm": 0.32271912031116656,
"learning_rate": 9.91520769210645e-05,
"loss": 1.1254,
"step": 910
},
{
"epoch": 1.66,
"grad_norm": 0.16598806172835462,
"learning_rate": 9.894010329567323e-05,
"loss": 1.187,
"step": 911
},
{
"epoch": 1.66,
"grad_norm": 0.17032736087667158,
"learning_rate": 9.872813443312984e-05,
"loss": 1.2073,
"step": 912
},
{
"epoch": 1.66,
"grad_norm": 0.16171590357915658,
"learning_rate": 9.851617128595694e-05,
"loss": 1.0839,
"step": 913
},
{
"epoch": 1.66,
"grad_norm": 0.22200097454760384,
"learning_rate": 9.830421480665128e-05,
"loss": 1.1299,
"step": 914
},
{
"epoch": 1.67,
"grad_norm": 0.18159512381331983,
"learning_rate": 9.809226594767978e-05,
"loss": 1.1991,
"step": 915
},
{
"epoch": 1.67,
"grad_norm": 0.1729262170321024,
"learning_rate": 9.788032566147505e-05,
"loss": 1.0962,
"step": 916
},
{
"epoch": 1.67,
"grad_norm": 0.17432638214872634,
"learning_rate": 9.766839490043108e-05,
"loss": 1.1669,
"step": 917
},
{
"epoch": 1.67,
"grad_norm": 0.16088749813770933,
"learning_rate": 9.745647461689931e-05,
"loss": 1.0548,
"step": 918
},
{
"epoch": 1.67,
"grad_norm": 0.1741820101835776,
"learning_rate": 9.724456576318381e-05,
"loss": 1.1986,
"step": 919
},
{
"epoch": 1.68,
"grad_norm": 0.1749247436598961,
"learning_rate": 9.70326692915375e-05,
"loss": 1.1474,
"step": 920
},
{
"epoch": 1.68,
"grad_norm": 0.1637038092120215,
"learning_rate": 9.682078615415754e-05,
"loss": 1.1231,
"step": 921
},
{
"epoch": 1.68,
"grad_norm": 0.18807058316399383,
"learning_rate": 9.660891730318128e-05,
"loss": 1.2045,
"step": 922
},
{
"epoch": 1.68,
"grad_norm": 0.1748352594360658,
"learning_rate": 9.639706369068171e-05,
"loss": 1.1689,
"step": 923
},
{
"epoch": 1.68,
"grad_norm": 0.1716288593635444,
"learning_rate": 9.61852262686635e-05,
"loss": 1.1572,
"step": 924
},
{
"epoch": 1.68,
"grad_norm": 0.18271775085545605,
"learning_rate": 9.597340598905852e-05,
"loss": 1.21,
"step": 925
},
{
"epoch": 1.69,
"grad_norm": 0.1717878271479738,
"learning_rate": 9.576160380372149e-05,
"loss": 1.0988,
"step": 926
},
{
"epoch": 1.69,
"grad_norm": 0.1680182822271049,
"learning_rate": 9.5549820664426e-05,
"loss": 1.0968,
"step": 927
},
{
"epoch": 1.69,
"grad_norm": 0.1742620341628087,
"learning_rate": 9.533805752285993e-05,
"loss": 1.1635,
"step": 928
},
{
"epoch": 1.69,
"grad_norm": 0.16338762068272702,
"learning_rate": 9.512631533062138e-05,
"loss": 0.9968,
"step": 929
},
{
"epoch": 1.69,
"grad_norm": 0.17102172627425277,
"learning_rate": 9.491459503921421e-05,
"loss": 1.1481,
"step": 930
},
{
"epoch": 1.7,
"grad_norm": 0.166342175891061,
"learning_rate": 9.470289760004398e-05,
"loss": 1.0721,
"step": 931
},
{
"epoch": 1.7,
"grad_norm": 0.2215993724024064,
"learning_rate": 9.449122396441345e-05,
"loss": 1.1877,
"step": 932
},
{
"epoch": 1.7,
"grad_norm": 0.1659659894828161,
"learning_rate": 9.427957508351852e-05,
"loss": 1.1691,
"step": 933
},
{
"epoch": 1.7,
"grad_norm": 0.1656871284817402,
"learning_rate": 9.406795190844376e-05,
"loss": 1.06,
"step": 934
},
{
"epoch": 1.7,
"grad_norm": 0.1586845482565193,
"learning_rate": 9.385635539015825e-05,
"loss": 1.1,
"step": 935
},
{
"epoch": 1.7,
"grad_norm": 0.16465420669616035,
"learning_rate": 9.364478647951133e-05,
"loss": 1.1403,
"step": 936
},
{
"epoch": 1.71,
"grad_norm": 0.15668641868741373,
"learning_rate": 9.343324612722819e-05,
"loss": 1.1075,
"step": 937
},
{
"epoch": 1.71,
"grad_norm": 0.16611169088286057,
"learning_rate": 9.32217352839058e-05,
"loss": 1.1256,
"step": 938
},
{
"epoch": 1.71,
"grad_norm": 0.17301783500328655,
"learning_rate": 9.301025490000841e-05,
"loss": 1.1797,
"step": 939
},
{
"epoch": 1.71,
"grad_norm": 0.17015053210523748,
"learning_rate": 9.279880592586354e-05,
"loss": 1.1087,
"step": 940
},
{
"epoch": 1.71,
"grad_norm": 0.17913490315482883,
"learning_rate": 9.25873893116574e-05,
"loss": 1.1673,
"step": 941
},
{
"epoch": 1.72,
"grad_norm": 0.18716341523653843,
"learning_rate": 9.237600600743085e-05,
"loss": 1.1448,
"step": 942
},
{
"epoch": 1.72,
"grad_norm": 0.16020067287870843,
"learning_rate": 9.216465696307512e-05,
"loss": 1.1152,
"step": 943
},
{
"epoch": 1.72,
"grad_norm": 0.18082661292174337,
"learning_rate": 9.195334312832742e-05,
"loss": 1.1923,
"step": 944
},
{
"epoch": 1.72,
"grad_norm": 0.16932315705059348,
"learning_rate": 9.174206545276677e-05,
"loss": 1.2025,
"step": 945
},
{
"epoch": 1.72,
"grad_norm": 0.1773723182281052,
"learning_rate": 9.153082488580968e-05,
"loss": 1.1583,
"step": 946
},
{
"epoch": 1.72,
"grad_norm": 0.1650084811263765,
"learning_rate": 9.131962237670598e-05,
"loss": 1.1456,
"step": 947
},
{
"epoch": 1.73,
"grad_norm": 0.17296637806848497,
"learning_rate": 9.11084588745343e-05,
"loss": 1.1188,
"step": 948
},
{
"epoch": 1.73,
"grad_norm": 0.18441757049220042,
"learning_rate": 9.089733532819824e-05,
"loss": 1.1057,
"step": 949
},
{
"epoch": 1.73,
"grad_norm": 0.16479278966433117,
"learning_rate": 9.068625268642161e-05,
"loss": 0.9686,
"step": 950
},
{
"epoch": 1.73,
"grad_norm": 0.1721522611977686,
"learning_rate": 9.047521189774455e-05,
"loss": 1.1593,
"step": 951
},
{
"epoch": 1.73,
"grad_norm": 0.17693809369137542,
"learning_rate": 9.026421391051907e-05,
"loss": 1.141,
"step": 952
},
{
"epoch": 1.74,
"grad_norm": 0.16748786371543328,
"learning_rate": 9.005325967290488e-05,
"loss": 1.0954,
"step": 953
},
{
"epoch": 1.74,
"grad_norm": 0.16738864271495277,
"learning_rate": 8.984235013286511e-05,
"loss": 1.1954,
"step": 954
},
{
"epoch": 1.74,
"grad_norm": 0.17608800377093817,
"learning_rate": 8.963148623816191e-05,
"loss": 1.2054,
"step": 955
},
{
"epoch": 1.74,
"grad_norm": 0.1572056318124515,
"learning_rate": 8.942066893635246e-05,
"loss": 1.062,
"step": 956
},
{
"epoch": 1.74,
"grad_norm": 0.17321331852374994,
"learning_rate": 8.920989917478447e-05,
"loss": 1.1774,
"step": 957
},
{
"epoch": 1.74,
"grad_norm": 0.24022668199221645,
"learning_rate": 8.899917790059208e-05,
"loss": 1.1422,
"step": 958
},
{
"epoch": 1.75,
"grad_norm": 0.17078130002322278,
"learning_rate": 8.878850606069153e-05,
"loss": 1.102,
"step": 959
},
{
"epoch": 1.75,
"grad_norm": 0.18816676453808523,
"learning_rate": 8.857788460177686e-05,
"loss": 1.1228,
"step": 960
},
{
"epoch": 1.75,
"grad_norm": 0.17061330738089536,
"learning_rate": 8.836731447031581e-05,
"loss": 1.1664,
"step": 961
},
{
"epoch": 1.75,
"grad_norm": 0.15744686305416977,
"learning_rate": 8.815679661254537e-05,
"loss": 1.0526,
"step": 962
},
{
"epoch": 1.75,
"grad_norm": 0.17140964720406338,
"learning_rate": 8.79463319744677e-05,
"loss": 1.1015,
"step": 963
},
{
"epoch": 1.76,
"grad_norm": 0.177839632569439,
"learning_rate": 8.77359215018458e-05,
"loss": 1.179,
"step": 964
},
{
"epoch": 1.76,
"grad_norm": 0.18976111884283367,
"learning_rate": 8.752556614019923e-05,
"loss": 1.1901,
"step": 965
},
{
"epoch": 1.76,
"grad_norm": 0.17757617341847579,
"learning_rate": 8.731526683479992e-05,
"loss": 1.1195,
"step": 966
},
{
"epoch": 1.76,
"grad_norm": 0.18025608157061349,
"learning_rate": 8.710502453066792e-05,
"loss": 1.1681,
"step": 967
},
{
"epoch": 1.76,
"grad_norm": 0.17991548690786335,
"learning_rate": 8.689484017256711e-05,
"loss": 1.0382,
"step": 968
},
{
"epoch": 1.77,
"grad_norm": 0.18377930665665893,
"learning_rate": 8.668471470500095e-05,
"loss": 1.2159,
"step": 969
},
{
"epoch": 1.77,
"grad_norm": 0.16785777204268718,
"learning_rate": 8.647464907220832e-05,
"loss": 1.1001,
"step": 970
},
{
"epoch": 1.77,
"grad_norm": 0.1718542715910871,
"learning_rate": 8.626464421815919e-05,
"loss": 1.1362,
"step": 971
},
{
"epoch": 1.77,
"grad_norm": 0.17567452985919224,
"learning_rate": 8.605470108655045e-05,
"loss": 1.0701,
"step": 972
},
{
"epoch": 1.77,
"grad_norm": 0.16645014392512258,
"learning_rate": 8.584482062080154e-05,
"loss": 1.1224,
"step": 973
},
{
"epoch": 1.77,
"grad_norm": 0.1683034297078507,
"learning_rate": 8.563500376405041e-05,
"loss": 1.1833,
"step": 974
},
{
"epoch": 1.78,
"grad_norm": 0.17027006999646288,
"learning_rate": 8.542525145914905e-05,
"loss": 1.2105,
"step": 975
},
{
"epoch": 1.78,
"grad_norm": 0.16661484142693264,
"learning_rate": 8.521556464865954e-05,
"loss": 1.1345,
"step": 976
},
{
"epoch": 1.78,
"grad_norm": 0.16535108955142738,
"learning_rate": 8.500594427484946e-05,
"loss": 1.0801,
"step": 977
},
{
"epoch": 1.78,
"grad_norm": 0.1619767851731314,
"learning_rate": 8.479639127968792e-05,
"loss": 1.1398,
"step": 978
},
{
"epoch": 1.78,
"grad_norm": 0.16148830590153632,
"learning_rate": 8.458690660484134e-05,
"loss": 1.1714,
"step": 979
},
{
"epoch": 1.79,
"grad_norm": 0.1718526387843433,
"learning_rate": 8.4377491191669e-05,
"loss": 1.1196,
"step": 980
},
{
"epoch": 1.79,
"grad_norm": 0.1589791532312397,
"learning_rate": 8.4168145981219e-05,
"loss": 1.0975,
"step": 981
},
{
"epoch": 1.79,
"grad_norm": 0.18228960673193118,
"learning_rate": 8.395887191422397e-05,
"loss": 1.0911,
"step": 982
},
{
"epoch": 1.79,
"grad_norm": 0.17051635436251017,
"learning_rate": 8.374966993109687e-05,
"loss": 1.1156,
"step": 983
},
{
"epoch": 1.79,
"grad_norm": 0.1734092864786433,
"learning_rate": 8.35405409719266e-05,
"loss": 1.1387,
"step": 984
},
{
"epoch": 1.79,
"grad_norm": 0.17960160313198717,
"learning_rate": 8.333148597647414e-05,
"loss": 1.1255,
"step": 985
},
{
"epoch": 1.8,
"grad_norm": 0.1667460148360989,
"learning_rate": 8.312250588416791e-05,
"loss": 1.0877,
"step": 986
},
{
"epoch": 1.8,
"grad_norm": 0.17227304205526361,
"learning_rate": 8.291360163409978e-05,
"loss": 1.1643,
"step": 987
},
{
"epoch": 1.8,
"grad_norm": 0.16172787345643957,
"learning_rate": 8.270477416502091e-05,
"loss": 1.0813,
"step": 988
},
{
"epoch": 1.8,
"grad_norm": 0.17717380607042416,
"learning_rate": 8.249602441533726e-05,
"loss": 1.1653,
"step": 989
},
{
"epoch": 1.8,
"grad_norm": 0.17646325594155327,
"learning_rate": 8.228735332310575e-05,
"loss": 1.1671,
"step": 990
},
{
"epoch": 1.81,
"grad_norm": 0.16889550666185796,
"learning_rate": 8.207876182602958e-05,
"loss": 1.1068,
"step": 991
},
{
"epoch": 1.81,
"grad_norm": 0.1668114849825541,
"learning_rate": 8.187025086145458e-05,
"loss": 1.1467,
"step": 992
},
{
"epoch": 1.81,
"grad_norm": 0.1721245962364379,
"learning_rate": 8.16618213663644e-05,
"loss": 1.1377,
"step": 993
},
{
"epoch": 1.81,
"grad_norm": 0.1929335873892253,
"learning_rate": 8.145347427737678e-05,
"loss": 1.119,
"step": 994
},
{
"epoch": 1.81,
"grad_norm": 0.16623130316399012,
"learning_rate": 8.12452105307391e-05,
"loss": 1.0687,
"step": 995
},
{
"epoch": 1.81,
"grad_norm": 0.18085333131496864,
"learning_rate": 8.103703106232416e-05,
"loss": 1.1517,
"step": 996
},
{
"epoch": 1.82,
"grad_norm": 0.17906886829733978,
"learning_rate": 8.082893680762619e-05,
"loss": 1.1325,
"step": 997
},
{
"epoch": 1.82,
"grad_norm": 0.16359826556853824,
"learning_rate": 8.062092870175628e-05,
"loss": 1.0353,
"step": 998
},
{
"epoch": 1.82,
"grad_norm": 0.1626009671933035,
"learning_rate": 8.041300767943867e-05,
"loss": 1.1598,
"step": 999
},
{
"epoch": 1.82,
"grad_norm": 0.16643269286545453,
"learning_rate": 8.0205174675006e-05,
"loss": 1.1226,
"step": 1000
},
{
"epoch": 1.82,
"grad_norm": 0.1699654548245084,
"learning_rate": 7.999743062239557e-05,
"loss": 1.0999,
"step": 1001
},
{
"epoch": 1.83,
"grad_norm": 0.1827266429905637,
"learning_rate": 7.978977645514487e-05,
"loss": 1.2791,
"step": 1002
},
{
"epoch": 1.83,
"grad_norm": 0.44434388351574605,
"learning_rate": 7.958221310638749e-05,
"loss": 1.1693,
"step": 1003
},
{
"epoch": 1.83,
"grad_norm": 0.17634755268047692,
"learning_rate": 7.937474150884897e-05,
"loss": 1.0845,
"step": 1004
},
{
"epoch": 1.83,
"grad_norm": 0.177726814790075,
"learning_rate": 7.916736259484239e-05,
"loss": 1.1849,
"step": 1005
},
{
"epoch": 1.83,
"grad_norm": 0.16479013332848777,
"learning_rate": 7.896007729626457e-05,
"loss": 1.1296,
"step": 1006
},
{
"epoch": 1.83,
"grad_norm": 0.1607507116543152,
"learning_rate": 7.875288654459144e-05,
"loss": 1.0829,
"step": 1007
},
{
"epoch": 1.84,
"grad_norm": 0.16387244898588724,
"learning_rate": 7.854579127087417e-05,
"loss": 1.1604,
"step": 1008
},
{
"epoch": 1.84,
"grad_norm": 0.163742117997135,
"learning_rate": 7.833879240573487e-05,
"loss": 1.0757,
"step": 1009
},
{
"epoch": 1.84,
"grad_norm": 0.16638334168591146,
"learning_rate": 7.813189087936243e-05,
"loss": 1.1637,
"step": 1010
},
{
"epoch": 1.84,
"grad_norm": 0.16818864513814694,
"learning_rate": 7.792508762150833e-05,
"loss": 1.1564,
"step": 1011
},
{
"epoch": 1.84,
"grad_norm": 0.17901829622442805,
"learning_rate": 7.771838356148232e-05,
"loss": 1.1841,
"step": 1012
},
{
"epoch": 1.85,
"grad_norm": 0.16614583867770816,
"learning_rate": 7.751177962814866e-05,
"loss": 1.1255,
"step": 1013
},
{
"epoch": 1.85,
"grad_norm": 0.19013974114118293,
"learning_rate": 7.730527674992143e-05,
"loss": 1.1488,
"step": 1014
},
{
"epoch": 1.85,
"grad_norm": 0.17442438513245137,
"learning_rate": 7.709887585476075e-05,
"loss": 1.1066,
"step": 1015
},
{
"epoch": 1.85,
"grad_norm": 0.17701832779209928,
"learning_rate": 7.689257787016834e-05,
"loss": 1.1755,
"step": 1016
},
{
"epoch": 1.85,
"grad_norm": 0.16936912919421254,
"learning_rate": 7.668638372318359e-05,
"loss": 1.1057,
"step": 1017
},
{
"epoch": 1.85,
"grad_norm": 0.19112857435989614,
"learning_rate": 7.648029434037915e-05,
"loss": 1.1412,
"step": 1018
},
{
"epoch": 1.86,
"grad_norm": 0.16103962948152212,
"learning_rate": 7.627431064785706e-05,
"loss": 1.0929,
"step": 1019
},
{
"epoch": 1.86,
"grad_norm": 0.16677548612612422,
"learning_rate": 7.606843357124426e-05,
"loss": 1.2046,
"step": 1020
},
{
"epoch": 1.86,
"grad_norm": 0.16233738083605964,
"learning_rate": 7.58626640356886e-05,
"loss": 1.0678,
"step": 1021
},
{
"epoch": 1.86,
"grad_norm": 0.16602043351095722,
"learning_rate": 7.565700296585483e-05,
"loss": 1.1216,
"step": 1022
},
{
"epoch": 1.86,
"grad_norm": 0.16627151579102536,
"learning_rate": 7.54514512859201e-05,
"loss": 1.0911,
"step": 1023
},
{
"epoch": 1.87,
"grad_norm": 0.20281510390834787,
"learning_rate": 7.524600991957012e-05,
"loss": 1.1722,
"step": 1024
},
{
"epoch": 1.87,
"grad_norm": 0.16069933741616732,
"learning_rate": 7.504067978999484e-05,
"loss": 1.0839,
"step": 1025
},
{
"epoch": 1.87,
"grad_norm": 0.18107514785611215,
"learning_rate": 7.483546181988436e-05,
"loss": 1.0717,
"step": 1026
},
{
"epoch": 1.87,
"grad_norm": 0.1721242495617235,
"learning_rate": 7.463035693142473e-05,
"loss": 1.1599,
"step": 1027
},
{
"epoch": 1.87,
"grad_norm": 0.17918262771658913,
"learning_rate": 7.442536604629395e-05,
"loss": 1.089,
"step": 1028
},
{
"epoch": 1.87,
"grad_norm": 0.17784491718923554,
"learning_rate": 7.422049008565757e-05,
"loss": 1.176,
"step": 1029
},
{
"epoch": 1.88,
"grad_norm": 0.15630505382249996,
"learning_rate": 7.401572997016479e-05,
"loss": 1.039,
"step": 1030
},
{
"epoch": 1.88,
"grad_norm": 0.17070239450442767,
"learning_rate": 7.381108661994429e-05,
"loss": 1.1059,
"step": 1031
},
{
"epoch": 1.88,
"grad_norm": 0.16222624124278331,
"learning_rate": 7.360656095459995e-05,
"loss": 1.024,
"step": 1032
},
{
"epoch": 1.88,
"grad_norm": 0.17125506848213806,
"learning_rate": 7.340215389320687e-05,
"loss": 1.1567,
"step": 1033
},
{
"epoch": 1.88,
"grad_norm": 0.17959953833242476,
"learning_rate": 7.31978663543071e-05,
"loss": 1.141,
"step": 1034
},
{
"epoch": 1.89,
"grad_norm": 0.17010870239393844,
"learning_rate": 7.299369925590574e-05,
"loss": 1.1046,
"step": 1035
},
{
"epoch": 1.89,
"grad_norm": 0.15998424148316903,
"learning_rate": 7.278965351546648e-05,
"loss": 1.0627,
"step": 1036
},
{
"epoch": 1.89,
"grad_norm": 0.1710612967988063,
"learning_rate": 7.258573004990788e-05,
"loss": 1.1486,
"step": 1037
},
{
"epoch": 1.89,
"grad_norm": 0.1607397255850716,
"learning_rate": 7.238192977559884e-05,
"loss": 1.0751,
"step": 1038
},
{
"epoch": 1.89,
"grad_norm": 0.17530360487610264,
"learning_rate": 7.217825360835473e-05,
"loss": 1.1281,
"step": 1039
},
{
"epoch": 1.89,
"grad_norm": 0.17374375355320706,
"learning_rate": 7.197470246343333e-05,
"loss": 1.1493,
"step": 1040
},
{
"epoch": 1.9,
"grad_norm": 0.15987413285182245,
"learning_rate": 7.177127725553045e-05,
"loss": 1.1446,
"step": 1041
},
{
"epoch": 1.9,
"grad_norm": 0.1789355664165954,
"learning_rate": 7.156797889877613e-05,
"loss": 1.2304,
"step": 1042
},
{
"epoch": 1.9,
"grad_norm": 0.17282586483463735,
"learning_rate": 7.136480830673019e-05,
"loss": 1.1909,
"step": 1043
},
{
"epoch": 1.9,
"grad_norm": 0.159213004708279,
"learning_rate": 7.116176639237852e-05,
"loss": 1.0624,
"step": 1044
},
{
"epoch": 1.9,
"grad_norm": 0.16197929506166975,
"learning_rate": 7.095885406812866e-05,
"loss": 1.1283,
"step": 1045
},
{
"epoch": 1.91,
"grad_norm": 0.16495179889101508,
"learning_rate": 7.075607224580581e-05,
"loss": 1.1378,
"step": 1046
},
{
"epoch": 1.91,
"grad_norm": 0.1662731898113741,
"learning_rate": 7.05534218366488e-05,
"loss": 1.0542,
"step": 1047
},
{
"epoch": 1.91,
"grad_norm": 0.1650210560996418,
"learning_rate": 7.035090375130581e-05,
"loss": 1.1326,
"step": 1048
},
{
"epoch": 1.91,
"grad_norm": 0.1708049460231556,
"learning_rate": 7.014851889983057e-05,
"loss": 1.137,
"step": 1049
},
{
"epoch": 1.91,
"grad_norm": 0.17558230008855352,
"learning_rate": 6.994626819167789e-05,
"loss": 1.1359,
"step": 1050
},
{
"epoch": 1.91,
"grad_norm": 0.16645826974884173,
"learning_rate": 6.974415253570003e-05,
"loss": 1.1742,
"step": 1051
},
{
"epoch": 1.92,
"grad_norm": 0.17184188151368943,
"learning_rate": 6.954217284014211e-05,
"loss": 1.1953,
"step": 1052
},
{
"epoch": 1.92,
"grad_norm": 0.17068592296841717,
"learning_rate": 6.934033001263847e-05,
"loss": 1.1301,
"step": 1053
},
{
"epoch": 1.92,
"grad_norm": 0.16302700472895656,
"learning_rate": 6.913862496020831e-05,
"loss": 1.1232,
"step": 1054
},
{
"epoch": 1.92,
"grad_norm": 0.17057720351205383,
"learning_rate": 6.893705858925178e-05,
"loss": 1.1247,
"step": 1055
},
{
"epoch": 1.92,
"grad_norm": 0.1582652133644645,
"learning_rate": 6.873563180554583e-05,
"loss": 1.1203,
"step": 1056
},
{
"epoch": 1.93,
"grad_norm": 0.16895339487465857,
"learning_rate": 6.853434551424e-05,
"loss": 1.1846,
"step": 1057
},
{
"epoch": 1.93,
"grad_norm": 0.1726993947504367,
"learning_rate": 6.833320061985277e-05,
"loss": 1.1963,
"step": 1058
},
{
"epoch": 1.93,
"grad_norm": 0.16055788377798388,
"learning_rate": 6.813219802626698e-05,
"loss": 1.1439,
"step": 1059
},
{
"epoch": 1.93,
"grad_norm": 0.17265920805658908,
"learning_rate": 6.793133863672616e-05,
"loss": 1.192,
"step": 1060
},
{
"epoch": 1.93,
"grad_norm": 0.1833601588530699,
"learning_rate": 6.773062335383024e-05,
"loss": 1.1128,
"step": 1061
},
{
"epoch": 1.93,
"grad_norm": 0.1717674054389487,
"learning_rate": 6.753005307953167e-05,
"loss": 1.134,
"step": 1062
},
{
"epoch": 1.94,
"grad_norm": 0.18207071739270747,
"learning_rate": 6.73296287151312e-05,
"loss": 1.1723,
"step": 1063
},
{
"epoch": 1.94,
"grad_norm": 0.1624856029472637,
"learning_rate": 6.712935116127389e-05,
"loss": 1.1144,
"step": 1064
},
{
"epoch": 1.94,
"grad_norm": 0.19915390833072513,
"learning_rate": 6.692922131794517e-05,
"loss": 1.1072,
"step": 1065
},
{
"epoch": 1.94,
"grad_norm": 0.1722567762020263,
"learning_rate": 6.672924008446662e-05,
"loss": 1.1425,
"step": 1066
},
{
"epoch": 1.94,
"grad_norm": 0.15963511981578907,
"learning_rate": 6.652940835949208e-05,
"loss": 1.1211,
"step": 1067
},
{
"epoch": 1.95,
"grad_norm": 0.16578458195408932,
"learning_rate": 6.632972704100349e-05,
"loss": 1.1244,
"step": 1068
},
{
"epoch": 1.95,
"grad_norm": 0.17110384656595054,
"learning_rate": 6.613019702630694e-05,
"loss": 1.1795,
"step": 1069
},
{
"epoch": 1.95,
"grad_norm": 0.16990207952122507,
"learning_rate": 6.593081921202859e-05,
"loss": 1.0817,
"step": 1070
},
{
"epoch": 1.95,
"grad_norm": 0.16467375370533938,
"learning_rate": 6.57315944941107e-05,
"loss": 1.0977,
"step": 1071
},
{
"epoch": 1.95,
"grad_norm": 0.17152460282733728,
"learning_rate": 6.553252376780748e-05,
"loss": 1.1868,
"step": 1072
},
{
"epoch": 1.95,
"grad_norm": 0.18132610186641904,
"learning_rate": 6.533360792768122e-05,
"loss": 1.2005,
"step": 1073
},
{
"epoch": 1.96,
"grad_norm": 0.16698085910075294,
"learning_rate": 6.513484786759818e-05,
"loss": 1.1362,
"step": 1074
},
{
"epoch": 1.96,
"grad_norm": 0.16504199012773973,
"learning_rate": 6.493624448072457e-05,
"loss": 1.1303,
"step": 1075
},
{
"epoch": 1.96,
"grad_norm": 0.1779596734054924,
"learning_rate": 6.473779865952263e-05,
"loss": 1.1894,
"step": 1076
},
{
"epoch": 1.96,
"grad_norm": 0.16364715074236819,
"learning_rate": 6.453951129574644e-05,
"loss": 1.1443,
"step": 1077
},
{
"epoch": 1.96,
"grad_norm": 0.16251550362999542,
"learning_rate": 6.434138328043815e-05,
"loss": 1.1238,
"step": 1078
},
{
"epoch": 1.97,
"grad_norm": 0.1666990603079008,
"learning_rate": 6.414341550392368e-05,
"loss": 1.0909,
"step": 1079
},
{
"epoch": 1.97,
"grad_norm": 0.15779118156939148,
"learning_rate": 6.39456088558091e-05,
"loss": 1.1109,
"step": 1080
},
{
"epoch": 1.97,
"grad_norm": 0.1677767995314692,
"learning_rate": 6.374796422497621e-05,
"loss": 1.2103,
"step": 1081
},
{
"epoch": 1.97,
"grad_norm": 0.17165919802575422,
"learning_rate": 6.355048249957886e-05,
"loss": 1.1879,
"step": 1082
},
{
"epoch": 1.97,
"grad_norm": 0.1918026223221136,
"learning_rate": 6.33531645670389e-05,
"loss": 1.1733,
"step": 1083
},
{
"epoch": 1.97,
"grad_norm": 0.1787527260843118,
"learning_rate": 6.3156011314042e-05,
"loss": 1.1338,
"step": 1084
},
{
"epoch": 1.98,
"grad_norm": 0.1694260433762424,
"learning_rate": 6.2959023626534e-05,
"loss": 1.1537,
"step": 1085
},
{
"epoch": 1.98,
"grad_norm": 0.18311039386845188,
"learning_rate": 6.276220238971652e-05,
"loss": 1.1976,
"step": 1086
},
{
"epoch": 1.98,
"grad_norm": 0.17113962947413622,
"learning_rate": 6.256554848804343e-05,
"loss": 1.0813,
"step": 1087
},
{
"epoch": 1.98,
"grad_norm": 0.1664054538940142,
"learning_rate": 6.236906280521646e-05,
"loss": 1.1514,
"step": 1088
},
{
"epoch": 1.98,
"grad_norm": 0.16227842164425618,
"learning_rate": 6.217274622418153e-05,
"loss": 1.0597,
"step": 1089
},
{
"epoch": 1.99,
"grad_norm": 0.17755371081391708,
"learning_rate": 6.197659962712461e-05,
"loss": 1.19,
"step": 1090
},
{
"epoch": 1.99,
"grad_norm": 0.15844523213403552,
"learning_rate": 6.178062389546784e-05,
"loss": 1.0587,
"step": 1091
},
{
"epoch": 1.99,
"grad_norm": 0.16903107223771263,
"learning_rate": 6.158481990986557e-05,
"loss": 1.1339,
"step": 1092
},
{
"epoch": 1.99,
"grad_norm": 0.17002303254909312,
"learning_rate": 6.138918855020028e-05,
"loss": 1.1158,
"step": 1093
},
{
"epoch": 1.99,
"grad_norm": 0.16613843815618992,
"learning_rate": 6.11937306955789e-05,
"loss": 1.1894,
"step": 1094
},
{
"epoch": 1.99,
"grad_norm": 0.17111783959067672,
"learning_rate": 6.099844722432843e-05,
"loss": 1.1974,
"step": 1095
},
{
"epoch": 2.0,
"grad_norm": 0.1670020953276936,
"learning_rate": 6.080333901399251e-05,
"loss": 1.1018,
"step": 1096
},
{
"epoch": 2.0,
"grad_norm": 0.18080514034863332,
"learning_rate": 6.060840694132701e-05,
"loss": 1.1845,
"step": 1097
},
{
"epoch": 2.0,
"grad_norm": 0.18525347317947505,
"learning_rate": 6.0413651882296406e-05,
"loss": 1.0796,
"step": 1098
},
{
"epoch": 2.0,
"grad_norm": 0.20400403778138837,
"learning_rate": 6.021907471206971e-05,
"loss": 1.0117,
"step": 1099
},
{
"epoch": 2.0,
"grad_norm": 0.1641784925803925,
"learning_rate": 6.002467630501646e-05,
"loss": 1.097,
"step": 1100
}
],
"logging_steps": 1.0,
"max_steps": 1647,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 50,
"total_flos": 1.0209608098381824e+16,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}