sft_conv_llama3-ko / trainer_state.json
JH-debug's picture
Upload folder using huggingface_hub
29f3b3e verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.9995724668661823,
"eval_steps": 500,
"global_step": 877,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.003420265070542967,
"grad_norm": 10.156329759741077,
"learning_rate": 0.0,
"loss": 2.1575,
"step": 1
},
{
"epoch": 0.006840530141085934,
"grad_norm": 184.3295302770049,
"learning_rate": 1.41e-05,
"loss": 4.88,
"step": 2
},
{
"epoch": 0.010260795211628902,
"grad_norm": 182.65124398096702,
"learning_rate": 1.41e-05,
"loss": 4.8994,
"step": 3
},
{
"epoch": 0.013681060282171868,
"grad_norm": 30.32918049437796,
"learning_rate": 1.41e-05,
"loss": 3.5419,
"step": 4
},
{
"epoch": 0.017101325352714837,
"grad_norm": 16.444831424361816,
"learning_rate": 1.41e-05,
"loss": 2.7311,
"step": 5
},
{
"epoch": 0.020521590423257803,
"grad_norm": 6.633419850116959,
"learning_rate": 1.41e-05,
"loss": 2.2268,
"step": 6
},
{
"epoch": 0.02394185549380077,
"grad_norm": 9.401679436345786,
"learning_rate": 1.41e-05,
"loss": 2.1353,
"step": 7
},
{
"epoch": 0.027362120564343735,
"grad_norm": 7.95097945213086,
"learning_rate": 1.41e-05,
"loss": 2.0867,
"step": 8
},
{
"epoch": 0.030782385634886705,
"grad_norm": 3.9617869212350962,
"learning_rate": 1.41e-05,
"loss": 1.9823,
"step": 9
},
{
"epoch": 0.034202650705429674,
"grad_norm": 9.402637091305385,
"learning_rate": 1.41e-05,
"loss": 1.9361,
"step": 10
},
{
"epoch": 0.03762291577597264,
"grad_norm": 3.7983366071849862,
"learning_rate": 1.41e-05,
"loss": 1.9295,
"step": 11
},
{
"epoch": 0.041043180846515606,
"grad_norm": 4.558710902554969,
"learning_rate": 1.41e-05,
"loss": 1.8787,
"step": 12
},
{
"epoch": 0.04446344591705857,
"grad_norm": 3.9168701968301134,
"learning_rate": 1.41e-05,
"loss": 1.7936,
"step": 13
},
{
"epoch": 0.04788371098760154,
"grad_norm": 3.763989925966716,
"learning_rate": 1.41e-05,
"loss": 1.7617,
"step": 14
},
{
"epoch": 0.051303976058144504,
"grad_norm": 3.4821708929604007,
"learning_rate": 1.41e-05,
"loss": 1.6881,
"step": 15
},
{
"epoch": 0.05472424112868747,
"grad_norm": 2.1426267148325997,
"learning_rate": 1.41e-05,
"loss": 1.6905,
"step": 16
},
{
"epoch": 0.05814450619923044,
"grad_norm": 3.6919608929938255,
"learning_rate": 1.41e-05,
"loss": 1.6473,
"step": 17
},
{
"epoch": 0.06156477126977341,
"grad_norm": 2.725651684433171,
"learning_rate": 1.41e-05,
"loss": 1.7064,
"step": 18
},
{
"epoch": 0.06498503634031637,
"grad_norm": 2.9127756737882824,
"learning_rate": 1.41e-05,
"loss": 1.5901,
"step": 19
},
{
"epoch": 0.06840530141085935,
"grad_norm": 2.664503176902388,
"learning_rate": 1.41e-05,
"loss": 1.6196,
"step": 20
},
{
"epoch": 0.07182556648140231,
"grad_norm": 2.5256162285977077,
"learning_rate": 1.41e-05,
"loss": 1.4734,
"step": 21
},
{
"epoch": 0.07524583155194528,
"grad_norm": 2.232926984374301,
"learning_rate": 1.41e-05,
"loss": 1.4862,
"step": 22
},
{
"epoch": 0.07866609662248825,
"grad_norm": 1.8467296080556872,
"learning_rate": 1.41e-05,
"loss": 1.4919,
"step": 23
},
{
"epoch": 0.08208636169303121,
"grad_norm": 2.133405884576976,
"learning_rate": 1.41e-05,
"loss": 1.4962,
"step": 24
},
{
"epoch": 0.08550662676357418,
"grad_norm": 2.0723150961914256,
"learning_rate": 1.41e-05,
"loss": 1.3745,
"step": 25
},
{
"epoch": 0.08892689183411714,
"grad_norm": 2.2473775199610166,
"learning_rate": 1.41e-05,
"loss": 1.4678,
"step": 26
},
{
"epoch": 0.09234715690466011,
"grad_norm": 3.1366913970429366,
"learning_rate": 1.41e-05,
"loss": 1.3231,
"step": 27
},
{
"epoch": 0.09576742197520308,
"grad_norm": 2.2703890784758585,
"learning_rate": 1.41e-05,
"loss": 1.3328,
"step": 28
},
{
"epoch": 0.09918768704574604,
"grad_norm": 2.7735799657439943,
"learning_rate": 1.41e-05,
"loss": 1.1703,
"step": 29
},
{
"epoch": 0.10260795211628901,
"grad_norm": 2.5181744230002203,
"learning_rate": 1.41e-05,
"loss": 1.257,
"step": 30
},
{
"epoch": 0.10602821718683197,
"grad_norm": 2.5187681530121546,
"learning_rate": 1.41e-05,
"loss": 1.1321,
"step": 31
},
{
"epoch": 0.10944848225737494,
"grad_norm": 2.5085106975094678,
"learning_rate": 1.41e-05,
"loss": 1.0929,
"step": 32
},
{
"epoch": 0.11286874732791792,
"grad_norm": 2.1171154989402887,
"learning_rate": 1.41e-05,
"loss": 1.1053,
"step": 33
},
{
"epoch": 0.11628901239846089,
"grad_norm": 2.4016503905187725,
"learning_rate": 1.41e-05,
"loss": 0.9514,
"step": 34
},
{
"epoch": 0.11970927746900385,
"grad_norm": 4.505185827517653,
"learning_rate": 1.41e-05,
"loss": 1.1443,
"step": 35
},
{
"epoch": 0.12312954253954682,
"grad_norm": 2.283031446044169,
"learning_rate": 1.41e-05,
"loss": 1.1263,
"step": 36
},
{
"epoch": 0.12654980761008977,
"grad_norm": 2.474396323802316,
"learning_rate": 1.41e-05,
"loss": 0.9587,
"step": 37
},
{
"epoch": 0.12997007268063274,
"grad_norm": 3.3515325695241667,
"learning_rate": 1.41e-05,
"loss": 0.8727,
"step": 38
},
{
"epoch": 0.1333903377511757,
"grad_norm": 2.3607499413214694,
"learning_rate": 1.41e-05,
"loss": 0.929,
"step": 39
},
{
"epoch": 0.1368106028217187,
"grad_norm": 2.4723425187830537,
"learning_rate": 1.41e-05,
"loss": 1.0473,
"step": 40
},
{
"epoch": 0.14023086789226166,
"grad_norm": 2.183114877668465,
"learning_rate": 1.41e-05,
"loss": 0.854,
"step": 41
},
{
"epoch": 0.14365113296280463,
"grad_norm": 2.3660790712775555,
"learning_rate": 1.41e-05,
"loss": 0.8244,
"step": 42
},
{
"epoch": 0.1470713980333476,
"grad_norm": 1.9551361934244411,
"learning_rate": 1.41e-05,
"loss": 0.8276,
"step": 43
},
{
"epoch": 0.15049166310389056,
"grad_norm": 3.163303383692252,
"learning_rate": 1.41e-05,
"loss": 0.8468,
"step": 44
},
{
"epoch": 0.15391192817443353,
"grad_norm": 3.163426556149129,
"learning_rate": 1.41e-05,
"loss": 0.7213,
"step": 45
},
{
"epoch": 0.1573321932449765,
"grad_norm": 2.7934537634938783,
"learning_rate": 1.41e-05,
"loss": 0.7035,
"step": 46
},
{
"epoch": 0.16075245831551946,
"grad_norm": 1.8747838269439079,
"learning_rate": 1.41e-05,
"loss": 0.7371,
"step": 47
},
{
"epoch": 0.16417272338606242,
"grad_norm": 2.7812346413247484,
"learning_rate": 1.41e-05,
"loss": 0.7602,
"step": 48
},
{
"epoch": 0.1675929884566054,
"grad_norm": 2.378687443172595,
"learning_rate": 1.41e-05,
"loss": 0.8561,
"step": 49
},
{
"epoch": 0.17101325352714836,
"grad_norm": 2.3489013022006056,
"learning_rate": 1.41e-05,
"loss": 0.5867,
"step": 50
},
{
"epoch": 0.17443351859769132,
"grad_norm": 1.7892416152344746,
"learning_rate": 1.41e-05,
"loss": 0.4853,
"step": 51
},
{
"epoch": 0.1778537836682343,
"grad_norm": 2.077063902741533,
"learning_rate": 1.41e-05,
"loss": 0.6375,
"step": 52
},
{
"epoch": 0.18127404873877725,
"grad_norm": 2.2312628788967412,
"learning_rate": 1.41e-05,
"loss": 0.6439,
"step": 53
},
{
"epoch": 0.18469431380932022,
"grad_norm": 2.996527586876452,
"learning_rate": 1.41e-05,
"loss": 0.5925,
"step": 54
},
{
"epoch": 0.1881145788798632,
"grad_norm": 1.8130852934136357,
"learning_rate": 1.41e-05,
"loss": 0.5575,
"step": 55
},
{
"epoch": 0.19153484395040615,
"grad_norm": 1.5587429488330231,
"learning_rate": 1.41e-05,
"loss": 0.5954,
"step": 56
},
{
"epoch": 0.19495510902094912,
"grad_norm": 1.7839250019187394,
"learning_rate": 1.41e-05,
"loss": 0.6196,
"step": 57
},
{
"epoch": 0.19837537409149208,
"grad_norm": 1.6933241820211493,
"learning_rate": 1.41e-05,
"loss": 0.4969,
"step": 58
},
{
"epoch": 0.20179563916203505,
"grad_norm": 1.731887729628178,
"learning_rate": 1.41e-05,
"loss": 0.6548,
"step": 59
},
{
"epoch": 0.20521590423257802,
"grad_norm": 2.3987513930499995,
"learning_rate": 1.41e-05,
"loss": 0.4595,
"step": 60
},
{
"epoch": 0.20863616930312098,
"grad_norm": 2.639090838839165,
"learning_rate": 1.41e-05,
"loss": 0.5341,
"step": 61
},
{
"epoch": 0.21205643437366395,
"grad_norm": 1.7474798765435617,
"learning_rate": 1.41e-05,
"loss": 0.6015,
"step": 62
},
{
"epoch": 0.21547669944420692,
"grad_norm": 3.571763528033457,
"learning_rate": 1.41e-05,
"loss": 0.4851,
"step": 63
},
{
"epoch": 0.21889696451474988,
"grad_norm": 1.9011170860485487,
"learning_rate": 1.41e-05,
"loss": 0.4911,
"step": 64
},
{
"epoch": 0.22231722958529285,
"grad_norm": 2.625704311611419,
"learning_rate": 1.41e-05,
"loss": 0.5247,
"step": 65
},
{
"epoch": 0.22573749465583584,
"grad_norm": 1.8272075489091404,
"learning_rate": 1.41e-05,
"loss": 0.4123,
"step": 66
},
{
"epoch": 0.2291577597263788,
"grad_norm": 2.005801690308512,
"learning_rate": 1.41e-05,
"loss": 0.437,
"step": 67
},
{
"epoch": 0.23257802479692177,
"grad_norm": 2.009191411617363,
"learning_rate": 1.41e-05,
"loss": 0.4553,
"step": 68
},
{
"epoch": 0.23599828986746474,
"grad_norm": 1.9856248091054856,
"learning_rate": 1.41e-05,
"loss": 0.4911,
"step": 69
},
{
"epoch": 0.2394185549380077,
"grad_norm": 1.9582222549850024,
"learning_rate": 1.41e-05,
"loss": 0.4345,
"step": 70
},
{
"epoch": 0.24283882000855067,
"grad_norm": 1.6336088110301379,
"learning_rate": 1.41e-05,
"loss": 0.4649,
"step": 71
},
{
"epoch": 0.24625908507909364,
"grad_norm": 1.946765368736228,
"learning_rate": 1.41e-05,
"loss": 0.3968,
"step": 72
},
{
"epoch": 0.2496793501496366,
"grad_norm": 1.9515164464490413,
"learning_rate": 1.41e-05,
"loss": 0.405,
"step": 73
},
{
"epoch": 0.25309961522017954,
"grad_norm": 1.5507200892542412,
"learning_rate": 1.41e-05,
"loss": 0.2883,
"step": 74
},
{
"epoch": 0.25651988029072254,
"grad_norm": 1.8953946078798778,
"learning_rate": 1.41e-05,
"loss": 0.3697,
"step": 75
},
{
"epoch": 0.2599401453612655,
"grad_norm": 1.609567252811076,
"learning_rate": 1.41e-05,
"loss": 0.3912,
"step": 76
},
{
"epoch": 0.26336041043180847,
"grad_norm": 1.841628814862342,
"learning_rate": 1.41e-05,
"loss": 0.3753,
"step": 77
},
{
"epoch": 0.2667806755023514,
"grad_norm": 1.3866520663429722,
"learning_rate": 1.41e-05,
"loss": 0.2978,
"step": 78
},
{
"epoch": 0.2702009405728944,
"grad_norm": 1.8662166292908873,
"learning_rate": 1.41e-05,
"loss": 0.325,
"step": 79
},
{
"epoch": 0.2736212056434374,
"grad_norm": 1.86990469712952,
"learning_rate": 1.41e-05,
"loss": 0.3608,
"step": 80
},
{
"epoch": 0.27704147071398033,
"grad_norm": 1.7776490861424512,
"learning_rate": 1.41e-05,
"loss": 0.3164,
"step": 81
},
{
"epoch": 0.2804617357845233,
"grad_norm": 1.5257018693275182,
"learning_rate": 1.41e-05,
"loss": 0.2736,
"step": 82
},
{
"epoch": 0.28388200085506626,
"grad_norm": 1.2107786177311144,
"learning_rate": 1.41e-05,
"loss": 0.2293,
"step": 83
},
{
"epoch": 0.28730226592560926,
"grad_norm": 1.7075082762966491,
"learning_rate": 1.41e-05,
"loss": 0.3312,
"step": 84
},
{
"epoch": 0.2907225309961522,
"grad_norm": 1.4271608707593932,
"learning_rate": 1.41e-05,
"loss": 0.2731,
"step": 85
},
{
"epoch": 0.2941427960666952,
"grad_norm": 2.0611639591955466,
"learning_rate": 1.41e-05,
"loss": 0.2708,
"step": 86
},
{
"epoch": 0.2975630611372381,
"grad_norm": 1.2754227728205574,
"learning_rate": 1.41e-05,
"loss": 0.2301,
"step": 87
},
{
"epoch": 0.3009833262077811,
"grad_norm": 1.312428216280633,
"learning_rate": 1.41e-05,
"loss": 0.2219,
"step": 88
},
{
"epoch": 0.30440359127832406,
"grad_norm": 1.36584246397569,
"learning_rate": 1.41e-05,
"loss": 0.2761,
"step": 89
},
{
"epoch": 0.30782385634886705,
"grad_norm": 1.1768002468324985,
"learning_rate": 1.41e-05,
"loss": 0.1884,
"step": 90
},
{
"epoch": 0.31124412141941,
"grad_norm": 1.402779753210287,
"learning_rate": 1.41e-05,
"loss": 0.24,
"step": 91
},
{
"epoch": 0.314664386489953,
"grad_norm": 1.2917943074724656,
"learning_rate": 1.41e-05,
"loss": 0.2487,
"step": 92
},
{
"epoch": 0.3180846515604959,
"grad_norm": 1.4168579440960474,
"learning_rate": 1.41e-05,
"loss": 0.2429,
"step": 93
},
{
"epoch": 0.3215049166310389,
"grad_norm": 1.1040710078361045,
"learning_rate": 1.41e-05,
"loss": 0.1676,
"step": 94
},
{
"epoch": 0.32492518170158186,
"grad_norm": 1.6692824066389471,
"learning_rate": 1.41e-05,
"loss": 0.2456,
"step": 95
},
{
"epoch": 0.32834544677212485,
"grad_norm": 1.4794418835245668,
"learning_rate": 1.41e-05,
"loss": 0.2346,
"step": 96
},
{
"epoch": 0.3317657118426678,
"grad_norm": 1.2426804107195764,
"learning_rate": 1.41e-05,
"loss": 0.2286,
"step": 97
},
{
"epoch": 0.3351859769132108,
"grad_norm": 1.5976518274152767,
"learning_rate": 1.41e-05,
"loss": 0.2356,
"step": 98
},
{
"epoch": 0.3386062419837537,
"grad_norm": 1.3380817380576575,
"learning_rate": 1.41e-05,
"loss": 0.2361,
"step": 99
},
{
"epoch": 0.3420265070542967,
"grad_norm": 1.8499907335313168,
"learning_rate": 1.41e-05,
"loss": 0.2144,
"step": 100
},
{
"epoch": 0.34544677212483965,
"grad_norm": 1.5519650604852626,
"learning_rate": 1.41e-05,
"loss": 0.2211,
"step": 101
},
{
"epoch": 0.34886703719538265,
"grad_norm": 1.4110088914262258,
"learning_rate": 1.41e-05,
"loss": 0.2422,
"step": 102
},
{
"epoch": 0.3522873022659256,
"grad_norm": 1.4052513404508866,
"learning_rate": 1.41e-05,
"loss": 0.1939,
"step": 103
},
{
"epoch": 0.3557075673364686,
"grad_norm": 1.377026410169766,
"learning_rate": 1.41e-05,
"loss": 0.193,
"step": 104
},
{
"epoch": 0.3591278324070115,
"grad_norm": 1.0901968649772857,
"learning_rate": 1.41e-05,
"loss": 0.1705,
"step": 105
},
{
"epoch": 0.3625480974775545,
"grad_norm": 1.2716891700474584,
"learning_rate": 1.41e-05,
"loss": 0.1998,
"step": 106
},
{
"epoch": 0.3659683625480975,
"grad_norm": 1.269092339983086,
"learning_rate": 1.41e-05,
"loss": 0.1845,
"step": 107
},
{
"epoch": 0.36938862761864044,
"grad_norm": 1.0134148691503204,
"learning_rate": 1.41e-05,
"loss": 0.1457,
"step": 108
},
{
"epoch": 0.37280889268918344,
"grad_norm": 1.0973618617348355,
"learning_rate": 1.41e-05,
"loss": 0.1829,
"step": 109
},
{
"epoch": 0.3762291577597264,
"grad_norm": 1.3476960032405298,
"learning_rate": 1.41e-05,
"loss": 0.1936,
"step": 110
},
{
"epoch": 0.37964942283026937,
"grad_norm": 1.1243588113049168,
"learning_rate": 1.41e-05,
"loss": 0.1763,
"step": 111
},
{
"epoch": 0.3830696879008123,
"grad_norm": 1.756776902549336,
"learning_rate": 1.41e-05,
"loss": 0.1772,
"step": 112
},
{
"epoch": 0.3864899529713553,
"grad_norm": 1.0653616123456815,
"learning_rate": 1.41e-05,
"loss": 0.1485,
"step": 113
},
{
"epoch": 0.38991021804189824,
"grad_norm": 1.322506116172751,
"learning_rate": 1.41e-05,
"loss": 0.1511,
"step": 114
},
{
"epoch": 0.39333048311244123,
"grad_norm": 1.134939568700316,
"learning_rate": 1.41e-05,
"loss": 0.1595,
"step": 115
},
{
"epoch": 0.39675074818298417,
"grad_norm": 1.2351108355732012,
"learning_rate": 1.41e-05,
"loss": 0.1653,
"step": 116
},
{
"epoch": 0.40017101325352716,
"grad_norm": 1.291174811726909,
"learning_rate": 1.41e-05,
"loss": 0.1414,
"step": 117
},
{
"epoch": 0.4035912783240701,
"grad_norm": 1.3110441911359814,
"learning_rate": 1.41e-05,
"loss": 0.1602,
"step": 118
},
{
"epoch": 0.4070115433946131,
"grad_norm": 1.0324677945393839,
"learning_rate": 1.41e-05,
"loss": 0.1589,
"step": 119
},
{
"epoch": 0.41043180846515603,
"grad_norm": 1.1528407686685203,
"learning_rate": 1.41e-05,
"loss": 0.142,
"step": 120
},
{
"epoch": 0.413852073535699,
"grad_norm": 1.259359755456077,
"learning_rate": 1.41e-05,
"loss": 0.1671,
"step": 121
},
{
"epoch": 0.41727233860624197,
"grad_norm": 0.9342347863650876,
"learning_rate": 1.41e-05,
"loss": 0.123,
"step": 122
},
{
"epoch": 0.42069260367678496,
"grad_norm": 1.0262297732751853,
"learning_rate": 1.41e-05,
"loss": 0.1379,
"step": 123
},
{
"epoch": 0.4241128687473279,
"grad_norm": 0.8528122922890459,
"learning_rate": 1.41e-05,
"loss": 0.1115,
"step": 124
},
{
"epoch": 0.4275331338178709,
"grad_norm": 1.0773688633850005,
"learning_rate": 1.41e-05,
"loss": 0.1368,
"step": 125
},
{
"epoch": 0.43095339888841383,
"grad_norm": 1.02543648603311,
"learning_rate": 1.41e-05,
"loss": 0.1116,
"step": 126
},
{
"epoch": 0.4343736639589568,
"grad_norm": 0.8824964139082571,
"learning_rate": 1.41e-05,
"loss": 0.1149,
"step": 127
},
{
"epoch": 0.43779392902949976,
"grad_norm": 1.1649358884512249,
"learning_rate": 1.41e-05,
"loss": 0.127,
"step": 128
},
{
"epoch": 0.44121419410004276,
"grad_norm": 1.0614948335318424,
"learning_rate": 1.41e-05,
"loss": 0.1248,
"step": 129
},
{
"epoch": 0.4446344591705857,
"grad_norm": 0.7776250289623791,
"learning_rate": 1.41e-05,
"loss": 0.1145,
"step": 130
},
{
"epoch": 0.4480547242411287,
"grad_norm": 0.8384984456348001,
"learning_rate": 1.41e-05,
"loss": 0.0921,
"step": 131
},
{
"epoch": 0.4514749893116717,
"grad_norm": 1.0151085870429353,
"learning_rate": 1.41e-05,
"loss": 0.1226,
"step": 132
},
{
"epoch": 0.4548952543822146,
"grad_norm": 0.7615407647681977,
"learning_rate": 1.41e-05,
"loss": 0.0951,
"step": 133
},
{
"epoch": 0.4583155194527576,
"grad_norm": 0.9555465270405811,
"learning_rate": 1.41e-05,
"loss": 0.1186,
"step": 134
},
{
"epoch": 0.46173578452330055,
"grad_norm": 0.9186643738206431,
"learning_rate": 1.41e-05,
"loss": 0.1207,
"step": 135
},
{
"epoch": 0.46515604959384355,
"grad_norm": 0.8860210174403064,
"learning_rate": 1.41e-05,
"loss": 0.1195,
"step": 136
},
{
"epoch": 0.4685763146643865,
"grad_norm": 0.8817179419098865,
"learning_rate": 1.41e-05,
"loss": 0.0897,
"step": 137
},
{
"epoch": 0.4719965797349295,
"grad_norm": 0.8289163487116727,
"learning_rate": 1.41e-05,
"loss": 0.1003,
"step": 138
},
{
"epoch": 0.4754168448054724,
"grad_norm": 0.904736501387183,
"learning_rate": 1.41e-05,
"loss": 0.0924,
"step": 139
},
{
"epoch": 0.4788371098760154,
"grad_norm": 0.943604807746642,
"learning_rate": 1.41e-05,
"loss": 0.1109,
"step": 140
},
{
"epoch": 0.48225737494655835,
"grad_norm": 0.9402997510757618,
"learning_rate": 1.41e-05,
"loss": 0.1145,
"step": 141
},
{
"epoch": 0.48567764001710134,
"grad_norm": 0.7366551780058475,
"learning_rate": 1.41e-05,
"loss": 0.0899,
"step": 142
},
{
"epoch": 0.4890979050876443,
"grad_norm": 0.9208800933128714,
"learning_rate": 1.41e-05,
"loss": 0.1167,
"step": 143
},
{
"epoch": 0.4925181701581873,
"grad_norm": 0.750448679907497,
"learning_rate": 1.41e-05,
"loss": 0.0809,
"step": 144
},
{
"epoch": 0.4959384352287302,
"grad_norm": 0.8092636444718945,
"learning_rate": 1.41e-05,
"loss": 0.0908,
"step": 145
},
{
"epoch": 0.4993587002992732,
"grad_norm": 0.8625942305572842,
"learning_rate": 1.41e-05,
"loss": 0.1031,
"step": 146
},
{
"epoch": 0.5027789653698161,
"grad_norm": 0.9198105373299891,
"learning_rate": 1.41e-05,
"loss": 0.1113,
"step": 147
},
{
"epoch": 0.5061992304403591,
"grad_norm": 0.7848012242788545,
"learning_rate": 1.41e-05,
"loss": 0.0865,
"step": 148
},
{
"epoch": 0.5096194955109021,
"grad_norm": 0.9324979893254519,
"learning_rate": 1.41e-05,
"loss": 0.0971,
"step": 149
},
{
"epoch": 0.5130397605814451,
"grad_norm": 0.7143510040718761,
"learning_rate": 1.41e-05,
"loss": 0.0795,
"step": 150
},
{
"epoch": 0.516460025651988,
"grad_norm": 1.0676584465757966,
"learning_rate": 1.41e-05,
"loss": 0.1135,
"step": 151
},
{
"epoch": 0.519880290722531,
"grad_norm": 0.6822736052229258,
"learning_rate": 1.41e-05,
"loss": 0.0733,
"step": 152
},
{
"epoch": 0.523300555793074,
"grad_norm": 0.8374258007686984,
"learning_rate": 1.41e-05,
"loss": 0.0907,
"step": 153
},
{
"epoch": 0.5267208208636169,
"grad_norm": 0.8341896688419776,
"learning_rate": 1.41e-05,
"loss": 0.0939,
"step": 154
},
{
"epoch": 0.5301410859341599,
"grad_norm": 0.7611257220630606,
"learning_rate": 1.41e-05,
"loss": 0.0869,
"step": 155
},
{
"epoch": 0.5335613510047028,
"grad_norm": 0.740325417856638,
"learning_rate": 1.41e-05,
"loss": 0.0832,
"step": 156
},
{
"epoch": 0.5369816160752459,
"grad_norm": 0.7042263268553958,
"learning_rate": 1.41e-05,
"loss": 0.0716,
"step": 157
},
{
"epoch": 0.5404018811457888,
"grad_norm": 0.7335797206103793,
"learning_rate": 1.41e-05,
"loss": 0.0807,
"step": 158
},
{
"epoch": 0.5438221462163317,
"grad_norm": 0.8019722685081757,
"learning_rate": 1.41e-05,
"loss": 0.0826,
"step": 159
},
{
"epoch": 0.5472424112868748,
"grad_norm": 0.6929468279193534,
"learning_rate": 1.41e-05,
"loss": 0.0821,
"step": 160
},
{
"epoch": 0.5506626763574177,
"grad_norm": 0.6854252563729888,
"learning_rate": 1.41e-05,
"loss": 0.0675,
"step": 161
},
{
"epoch": 0.5540829414279607,
"grad_norm": 0.6741146081895844,
"learning_rate": 1.41e-05,
"loss": 0.0749,
"step": 162
},
{
"epoch": 0.5575032064985036,
"grad_norm": 0.6902694268516201,
"learning_rate": 1.41e-05,
"loss": 0.0707,
"step": 163
},
{
"epoch": 0.5609234715690466,
"grad_norm": 0.6948144741970704,
"learning_rate": 1.41e-05,
"loss": 0.0795,
"step": 164
},
{
"epoch": 0.5643437366395896,
"grad_norm": 0.7169974641783955,
"learning_rate": 1.41e-05,
"loss": 0.0818,
"step": 165
},
{
"epoch": 0.5677640017101325,
"grad_norm": 0.6384211122986987,
"learning_rate": 1.41e-05,
"loss": 0.0735,
"step": 166
},
{
"epoch": 0.5711842667806755,
"grad_norm": 0.7417356609403256,
"learning_rate": 1.41e-05,
"loss": 0.0842,
"step": 167
},
{
"epoch": 0.5746045318512185,
"grad_norm": 0.66647458408933,
"learning_rate": 1.41e-05,
"loss": 0.0814,
"step": 168
},
{
"epoch": 0.5780247969217615,
"grad_norm": 0.7378707726234803,
"learning_rate": 1.41e-05,
"loss": 0.0746,
"step": 169
},
{
"epoch": 0.5814450619923044,
"grad_norm": 0.6812399388437269,
"learning_rate": 1.41e-05,
"loss": 0.0828,
"step": 170
},
{
"epoch": 0.5848653270628473,
"grad_norm": 0.6793042032294047,
"learning_rate": 1.41e-05,
"loss": 0.0724,
"step": 171
},
{
"epoch": 0.5882855921333904,
"grad_norm": 0.6062435025746024,
"learning_rate": 1.41e-05,
"loss": 0.0649,
"step": 172
},
{
"epoch": 0.5917058572039333,
"grad_norm": 0.6813501327595134,
"learning_rate": 1.41e-05,
"loss": 0.0711,
"step": 173
},
{
"epoch": 0.5951261222744763,
"grad_norm": 0.672907340612349,
"learning_rate": 1.41e-05,
"loss": 0.0769,
"step": 174
},
{
"epoch": 0.5985463873450192,
"grad_norm": 0.5410302293555107,
"learning_rate": 1.41e-05,
"loss": 0.0588,
"step": 175
},
{
"epoch": 0.6019666524155622,
"grad_norm": 0.6636988853462809,
"learning_rate": 1.41e-05,
"loss": 0.0702,
"step": 176
},
{
"epoch": 0.6053869174861052,
"grad_norm": 0.5918082791413835,
"learning_rate": 1.41e-05,
"loss": 0.0626,
"step": 177
},
{
"epoch": 0.6088071825566481,
"grad_norm": 0.5816411742116699,
"learning_rate": 1.41e-05,
"loss": 0.059,
"step": 178
},
{
"epoch": 0.6122274476271911,
"grad_norm": 0.6819230663398479,
"learning_rate": 1.41e-05,
"loss": 0.074,
"step": 179
},
{
"epoch": 0.6156477126977341,
"grad_norm": 0.6727834641069249,
"learning_rate": 1.41e-05,
"loss": 0.0739,
"step": 180
},
{
"epoch": 0.619067977768277,
"grad_norm": 0.5791689945437091,
"learning_rate": 1.41e-05,
"loss": 0.0553,
"step": 181
},
{
"epoch": 0.62248824283882,
"grad_norm": 0.6060693467771211,
"learning_rate": 1.41e-05,
"loss": 0.0572,
"step": 182
},
{
"epoch": 0.6259085079093629,
"grad_norm": 0.6611378853926027,
"learning_rate": 1.41e-05,
"loss": 0.0697,
"step": 183
},
{
"epoch": 0.629328772979906,
"grad_norm": 0.5887572335679387,
"learning_rate": 1.41e-05,
"loss": 0.0603,
"step": 184
},
{
"epoch": 0.6327490380504489,
"grad_norm": 0.538930391171475,
"learning_rate": 1.41e-05,
"loss": 0.0462,
"step": 185
},
{
"epoch": 0.6361693031209918,
"grad_norm": 0.5617056384496442,
"learning_rate": 1.41e-05,
"loss": 0.0549,
"step": 186
},
{
"epoch": 0.6395895681915349,
"grad_norm": 0.5912693387471951,
"learning_rate": 1.41e-05,
"loss": 0.0574,
"step": 187
},
{
"epoch": 0.6430098332620778,
"grad_norm": 0.5373216387973052,
"learning_rate": 1.41e-05,
"loss": 0.0574,
"step": 188
},
{
"epoch": 0.6464300983326208,
"grad_norm": 0.6151671164129469,
"learning_rate": 1.41e-05,
"loss": 0.0533,
"step": 189
},
{
"epoch": 0.6498503634031637,
"grad_norm": 0.5394742707247884,
"learning_rate": 1.41e-05,
"loss": 0.0575,
"step": 190
},
{
"epoch": 0.6532706284737068,
"grad_norm": 0.5752514447611141,
"learning_rate": 1.41e-05,
"loss": 0.0574,
"step": 191
},
{
"epoch": 0.6566908935442497,
"grad_norm": 0.5136422669182581,
"learning_rate": 1.41e-05,
"loss": 0.054,
"step": 192
},
{
"epoch": 0.6601111586147926,
"grad_norm": 0.6261951776293332,
"learning_rate": 1.41e-05,
"loss": 0.0612,
"step": 193
},
{
"epoch": 0.6635314236853356,
"grad_norm": 0.5067466055288193,
"learning_rate": 1.41e-05,
"loss": 0.054,
"step": 194
},
{
"epoch": 0.6669516887558786,
"grad_norm": 0.5892942817895197,
"learning_rate": 1.41e-05,
"loss": 0.0572,
"step": 195
},
{
"epoch": 0.6703719538264216,
"grad_norm": 0.5652390088377635,
"learning_rate": 1.41e-05,
"loss": 0.0655,
"step": 196
},
{
"epoch": 0.6737922188969645,
"grad_norm": 0.5285822929717092,
"learning_rate": 1.41e-05,
"loss": 0.0494,
"step": 197
},
{
"epoch": 0.6772124839675074,
"grad_norm": 0.5246191872665474,
"learning_rate": 1.41e-05,
"loss": 0.0614,
"step": 198
},
{
"epoch": 0.6806327490380505,
"grad_norm": 0.5387443642901717,
"learning_rate": 1.41e-05,
"loss": 0.0531,
"step": 199
},
{
"epoch": 0.6840530141085934,
"grad_norm": 0.5238421425157995,
"learning_rate": 1.41e-05,
"loss": 0.0512,
"step": 200
},
{
"epoch": 0.6874732791791364,
"grad_norm": 0.4882221825014573,
"learning_rate": 1.41e-05,
"loss": 0.0489,
"step": 201
},
{
"epoch": 0.6908935442496793,
"grad_norm": 0.5646516308611281,
"learning_rate": 1.41e-05,
"loss": 0.0512,
"step": 202
},
{
"epoch": 0.6943138093202224,
"grad_norm": 0.5778234440206057,
"learning_rate": 1.41e-05,
"loss": 0.0634,
"step": 203
},
{
"epoch": 0.6977340743907653,
"grad_norm": 0.5293741220834561,
"learning_rate": 1.41e-05,
"loss": 0.056,
"step": 204
},
{
"epoch": 0.7011543394613082,
"grad_norm": 0.527641578215631,
"learning_rate": 1.41e-05,
"loss": 0.0561,
"step": 205
},
{
"epoch": 0.7045746045318512,
"grad_norm": 0.4000174523279179,
"learning_rate": 1.41e-05,
"loss": 0.0401,
"step": 206
},
{
"epoch": 0.7079948696023942,
"grad_norm": 0.5298298097123045,
"learning_rate": 1.41e-05,
"loss": 0.0609,
"step": 207
},
{
"epoch": 0.7114151346729372,
"grad_norm": 0.5349126180633413,
"learning_rate": 1.41e-05,
"loss": 0.052,
"step": 208
},
{
"epoch": 0.7148353997434801,
"grad_norm": 0.5748237411918808,
"learning_rate": 1.41e-05,
"loss": 0.0651,
"step": 209
},
{
"epoch": 0.718255664814023,
"grad_norm": 0.4989630175919984,
"learning_rate": 1.41e-05,
"loss": 0.0502,
"step": 210
},
{
"epoch": 0.7216759298845661,
"grad_norm": 0.5604803895822472,
"learning_rate": 1.41e-05,
"loss": 0.054,
"step": 211
},
{
"epoch": 0.725096194955109,
"grad_norm": 0.6062121706000653,
"learning_rate": 1.41e-05,
"loss": 0.0644,
"step": 212
},
{
"epoch": 0.728516460025652,
"grad_norm": 0.49966896896386376,
"learning_rate": 1.41e-05,
"loss": 0.0469,
"step": 213
},
{
"epoch": 0.731936725096195,
"grad_norm": 0.44454274508391683,
"learning_rate": 1.41e-05,
"loss": 0.0428,
"step": 214
},
{
"epoch": 0.735356990166738,
"grad_norm": 0.563352438281472,
"learning_rate": 1.41e-05,
"loss": 0.055,
"step": 215
},
{
"epoch": 0.7387772552372809,
"grad_norm": 0.6191641486840816,
"learning_rate": 1.41e-05,
"loss": 0.063,
"step": 216
},
{
"epoch": 0.7421975203078238,
"grad_norm": 0.4827104401750106,
"learning_rate": 1.41e-05,
"loss": 0.0496,
"step": 217
},
{
"epoch": 0.7456177853783669,
"grad_norm": 0.4656663916077845,
"learning_rate": 1.41e-05,
"loss": 0.0528,
"step": 218
},
{
"epoch": 0.7490380504489098,
"grad_norm": 0.4785719764753492,
"learning_rate": 1.41e-05,
"loss": 0.0438,
"step": 219
},
{
"epoch": 0.7524583155194527,
"grad_norm": 0.5448096052982832,
"learning_rate": 1.41e-05,
"loss": 0.0422,
"step": 220
},
{
"epoch": 0.7558785805899957,
"grad_norm": 0.5065090698982245,
"learning_rate": 1.41e-05,
"loss": 0.0489,
"step": 221
},
{
"epoch": 0.7592988456605387,
"grad_norm": 0.5059637979656728,
"learning_rate": 1.41e-05,
"loss": 0.0485,
"step": 222
},
{
"epoch": 0.7627191107310817,
"grad_norm": 0.49583806553345944,
"learning_rate": 1.41e-05,
"loss": 0.0525,
"step": 223
},
{
"epoch": 0.7661393758016246,
"grad_norm": 0.4012288566084756,
"learning_rate": 1.41e-05,
"loss": 0.0384,
"step": 224
},
{
"epoch": 0.7695596408721675,
"grad_norm": 0.49449469823574593,
"learning_rate": 1.41e-05,
"loss": 0.05,
"step": 225
},
{
"epoch": 0.7729799059427106,
"grad_norm": 0.45672465215152086,
"learning_rate": 1.41e-05,
"loss": 0.0479,
"step": 226
},
{
"epoch": 0.7764001710132535,
"grad_norm": 0.4909318818146214,
"learning_rate": 1.41e-05,
"loss": 0.0499,
"step": 227
},
{
"epoch": 0.7798204360837965,
"grad_norm": 0.40392311014383353,
"learning_rate": 1.41e-05,
"loss": 0.0406,
"step": 228
},
{
"epoch": 0.7832407011543394,
"grad_norm": 0.4131849222227493,
"learning_rate": 1.41e-05,
"loss": 0.0388,
"step": 229
},
{
"epoch": 0.7866609662248825,
"grad_norm": 0.5611423817088044,
"learning_rate": 1.41e-05,
"loss": 0.0594,
"step": 230
},
{
"epoch": 0.7900812312954254,
"grad_norm": 0.47364645404517464,
"learning_rate": 1.41e-05,
"loss": 0.0522,
"step": 231
},
{
"epoch": 0.7935014963659683,
"grad_norm": 0.4545208834696141,
"learning_rate": 1.41e-05,
"loss": 0.048,
"step": 232
},
{
"epoch": 0.7969217614365113,
"grad_norm": 0.4113439560012879,
"learning_rate": 1.41e-05,
"loss": 0.0393,
"step": 233
},
{
"epoch": 0.8003420265070543,
"grad_norm": 0.42334212760511825,
"learning_rate": 1.41e-05,
"loss": 0.0404,
"step": 234
},
{
"epoch": 0.8037622915775973,
"grad_norm": 0.4768224768125407,
"learning_rate": 1.41e-05,
"loss": 0.0479,
"step": 235
},
{
"epoch": 0.8071825566481402,
"grad_norm": 0.43678389875922824,
"learning_rate": 1.41e-05,
"loss": 0.0442,
"step": 236
},
{
"epoch": 0.8106028217186833,
"grad_norm": 0.43277663891476426,
"learning_rate": 1.41e-05,
"loss": 0.0409,
"step": 237
},
{
"epoch": 0.8140230867892262,
"grad_norm": 0.4443462520817696,
"learning_rate": 1.41e-05,
"loss": 0.0465,
"step": 238
},
{
"epoch": 0.8174433518597691,
"grad_norm": 0.43839684525150946,
"learning_rate": 1.41e-05,
"loss": 0.0415,
"step": 239
},
{
"epoch": 0.8208636169303121,
"grad_norm": 0.43297499486580887,
"learning_rate": 1.41e-05,
"loss": 0.0445,
"step": 240
},
{
"epoch": 0.8242838820008551,
"grad_norm": 0.38283138235459124,
"learning_rate": 1.41e-05,
"loss": 0.0404,
"step": 241
},
{
"epoch": 0.827704147071398,
"grad_norm": 0.40333504070789256,
"learning_rate": 1.41e-05,
"loss": 0.0373,
"step": 242
},
{
"epoch": 0.831124412141941,
"grad_norm": 0.40950403615291003,
"learning_rate": 1.41e-05,
"loss": 0.0423,
"step": 243
},
{
"epoch": 0.8345446772124839,
"grad_norm": 0.4314317792707956,
"learning_rate": 1.41e-05,
"loss": 0.0424,
"step": 244
},
{
"epoch": 0.837964942283027,
"grad_norm": 0.4189847423801003,
"learning_rate": 1.41e-05,
"loss": 0.0434,
"step": 245
},
{
"epoch": 0.8413852073535699,
"grad_norm": 0.4593355477826361,
"learning_rate": 1.41e-05,
"loss": 0.0456,
"step": 246
},
{
"epoch": 0.8448054724241129,
"grad_norm": 0.43148788855113257,
"learning_rate": 1.41e-05,
"loss": 0.0429,
"step": 247
},
{
"epoch": 0.8482257374946558,
"grad_norm": 0.41015663281431336,
"learning_rate": 1.41e-05,
"loss": 0.0411,
"step": 248
},
{
"epoch": 0.8516460025651988,
"grad_norm": 0.39479744422344626,
"learning_rate": 1.41e-05,
"loss": 0.0408,
"step": 249
},
{
"epoch": 0.8550662676357418,
"grad_norm": 0.49034951176740266,
"learning_rate": 1.41e-05,
"loss": 0.056,
"step": 250
},
{
"epoch": 0.8584865327062847,
"grad_norm": 0.4241786968876316,
"learning_rate": 1.41e-05,
"loss": 0.0414,
"step": 251
},
{
"epoch": 0.8619067977768277,
"grad_norm": 0.40856481199229994,
"learning_rate": 1.41e-05,
"loss": 0.0384,
"step": 252
},
{
"epoch": 0.8653270628473707,
"grad_norm": 0.44863803613683295,
"learning_rate": 1.41e-05,
"loss": 0.0454,
"step": 253
},
{
"epoch": 0.8687473279179136,
"grad_norm": 0.4315245075802853,
"learning_rate": 1.41e-05,
"loss": 0.0425,
"step": 254
},
{
"epoch": 0.8721675929884566,
"grad_norm": 0.4300931739828693,
"learning_rate": 1.41e-05,
"loss": 0.0386,
"step": 255
},
{
"epoch": 0.8755878580589995,
"grad_norm": 0.41516523943726863,
"learning_rate": 1.41e-05,
"loss": 0.044,
"step": 256
},
{
"epoch": 0.8790081231295426,
"grad_norm": 0.41978138385172276,
"learning_rate": 1.41e-05,
"loss": 0.0376,
"step": 257
},
{
"epoch": 0.8824283882000855,
"grad_norm": 0.38286432398443565,
"learning_rate": 1.41e-05,
"loss": 0.0357,
"step": 258
},
{
"epoch": 0.8858486532706284,
"grad_norm": 1.0190380922512265,
"learning_rate": 1.41e-05,
"loss": 0.0413,
"step": 259
},
{
"epoch": 0.8892689183411714,
"grad_norm": 0.46857802366186146,
"learning_rate": 1.41e-05,
"loss": 0.0468,
"step": 260
},
{
"epoch": 0.8926891834117144,
"grad_norm": 0.38236119116632356,
"learning_rate": 1.41e-05,
"loss": 0.0383,
"step": 261
},
{
"epoch": 0.8961094484822574,
"grad_norm": 0.4551658885616909,
"learning_rate": 1.41e-05,
"loss": 0.0453,
"step": 262
},
{
"epoch": 0.8995297135528003,
"grad_norm": 0.4286695716016865,
"learning_rate": 1.41e-05,
"loss": 0.0408,
"step": 263
},
{
"epoch": 0.9029499786233434,
"grad_norm": 0.43951997404486826,
"learning_rate": 1.41e-05,
"loss": 0.0395,
"step": 264
},
{
"epoch": 0.9063702436938863,
"grad_norm": 0.4395704115846384,
"learning_rate": 1.41e-05,
"loss": 0.0419,
"step": 265
},
{
"epoch": 0.9097905087644292,
"grad_norm": 0.3839941648623069,
"learning_rate": 1.41e-05,
"loss": 0.0374,
"step": 266
},
{
"epoch": 0.9132107738349722,
"grad_norm": 0.922460042226267,
"learning_rate": 1.41e-05,
"loss": 0.0372,
"step": 267
},
{
"epoch": 0.9166310389055152,
"grad_norm": 0.44811463225682846,
"learning_rate": 1.41e-05,
"loss": 0.043,
"step": 268
},
{
"epoch": 0.9200513039760582,
"grad_norm": 0.4172986063615246,
"learning_rate": 1.41e-05,
"loss": 0.0365,
"step": 269
},
{
"epoch": 0.9234715690466011,
"grad_norm": 0.37219225103541853,
"learning_rate": 1.41e-05,
"loss": 0.0377,
"step": 270
},
{
"epoch": 0.926891834117144,
"grad_norm": 0.3447899677572012,
"learning_rate": 1.41e-05,
"loss": 0.0317,
"step": 271
},
{
"epoch": 0.9303120991876871,
"grad_norm": 0.3964175357323583,
"learning_rate": 1.41e-05,
"loss": 0.034,
"step": 272
},
{
"epoch": 0.93373236425823,
"grad_norm": 0.36028858741706155,
"learning_rate": 1.41e-05,
"loss": 0.0351,
"step": 273
},
{
"epoch": 0.937152629328773,
"grad_norm": 0.46256283545817356,
"learning_rate": 1.41e-05,
"loss": 0.041,
"step": 274
},
{
"epoch": 0.9405728943993159,
"grad_norm": 0.3781578237183448,
"learning_rate": 1.41e-05,
"loss": 0.0366,
"step": 275
},
{
"epoch": 0.943993159469859,
"grad_norm": 0.3818392999731623,
"learning_rate": 1.41e-05,
"loss": 0.0321,
"step": 276
},
{
"epoch": 0.9474134245404019,
"grad_norm": 0.38568204387768645,
"learning_rate": 1.41e-05,
"loss": 0.0336,
"step": 277
},
{
"epoch": 0.9508336896109448,
"grad_norm": 0.3905057560704898,
"learning_rate": 1.41e-05,
"loss": 0.0349,
"step": 278
},
{
"epoch": 0.9542539546814878,
"grad_norm": 0.36210340527903007,
"learning_rate": 1.41e-05,
"loss": 0.0351,
"step": 279
},
{
"epoch": 0.9576742197520308,
"grad_norm": 0.3784578196301756,
"learning_rate": 1.41e-05,
"loss": 0.0382,
"step": 280
},
{
"epoch": 0.9610944848225738,
"grad_norm": 0.4415630974468222,
"learning_rate": 1.41e-05,
"loss": 0.0417,
"step": 281
},
{
"epoch": 0.9645147498931167,
"grad_norm": 0.36220744727441767,
"learning_rate": 1.41e-05,
"loss": 0.0336,
"step": 282
},
{
"epoch": 0.9679350149636596,
"grad_norm": 0.4266084587936558,
"learning_rate": 1.41e-05,
"loss": 0.0384,
"step": 283
},
{
"epoch": 0.9713552800342027,
"grad_norm": 0.3599843427563046,
"learning_rate": 1.41e-05,
"loss": 0.035,
"step": 284
},
{
"epoch": 0.9747755451047456,
"grad_norm": 0.38722401771389997,
"learning_rate": 1.41e-05,
"loss": 0.0348,
"step": 285
},
{
"epoch": 0.9781958101752886,
"grad_norm": 0.40946001056055625,
"learning_rate": 1.41e-05,
"loss": 0.0373,
"step": 286
},
{
"epoch": 0.9816160752458315,
"grad_norm": 0.41550323303820474,
"learning_rate": 1.41e-05,
"loss": 0.04,
"step": 287
},
{
"epoch": 0.9850363403163745,
"grad_norm": 0.37048272944847027,
"learning_rate": 1.41e-05,
"loss": 0.0309,
"step": 288
},
{
"epoch": 0.9884566053869175,
"grad_norm": 0.3620011070515116,
"learning_rate": 1.41e-05,
"loss": 0.0341,
"step": 289
},
{
"epoch": 0.9918768704574604,
"grad_norm": 0.4132102658215653,
"learning_rate": 1.41e-05,
"loss": 0.0383,
"step": 290
},
{
"epoch": 0.9952971355280035,
"grad_norm": 0.3852351778306183,
"learning_rate": 1.41e-05,
"loss": 0.0357,
"step": 291
},
{
"epoch": 0.9987174005985464,
"grad_norm": 0.37414254856973184,
"learning_rate": 1.41e-05,
"loss": 0.0348,
"step": 292
},
{
"epoch": 1.0021376656690895,
"grad_norm": 0.31841258787737364,
"learning_rate": 1.41e-05,
"loss": 0.0308,
"step": 293
},
{
"epoch": 1.0055579307396323,
"grad_norm": 0.30923263652050564,
"learning_rate": 1.41e-05,
"loss": 0.0263,
"step": 294
},
{
"epoch": 1.0089781958101753,
"grad_norm": 0.3118052788978316,
"learning_rate": 1.41e-05,
"loss": 0.0243,
"step": 295
},
{
"epoch": 1.0123984608807182,
"grad_norm": 0.2908942931886208,
"learning_rate": 1.41e-05,
"loss": 0.0239,
"step": 296
},
{
"epoch": 1.0158187259512612,
"grad_norm": 0.2904620406307295,
"learning_rate": 1.41e-05,
"loss": 0.0222,
"step": 297
},
{
"epoch": 1.0192389910218043,
"grad_norm": 0.28692807204183246,
"learning_rate": 1.41e-05,
"loss": 0.0225,
"step": 298
},
{
"epoch": 1.022659256092347,
"grad_norm": 0.32077856448530445,
"learning_rate": 1.41e-05,
"loss": 0.025,
"step": 299
},
{
"epoch": 1.0260795211628901,
"grad_norm": 0.28877834524497115,
"learning_rate": 1.41e-05,
"loss": 0.0214,
"step": 300
},
{
"epoch": 1.0294997862334332,
"grad_norm": 0.27974144729840145,
"learning_rate": 1.41e-05,
"loss": 0.0226,
"step": 301
},
{
"epoch": 1.032920051303976,
"grad_norm": 0.2976755981973845,
"learning_rate": 1.41e-05,
"loss": 0.025,
"step": 302
},
{
"epoch": 1.036340316374519,
"grad_norm": 0.2909288296995041,
"learning_rate": 1.41e-05,
"loss": 0.0241,
"step": 303
},
{
"epoch": 1.039760581445062,
"grad_norm": 0.26580362507255084,
"learning_rate": 1.41e-05,
"loss": 0.0227,
"step": 304
},
{
"epoch": 1.043180846515605,
"grad_norm": 0.3375634201266733,
"learning_rate": 1.41e-05,
"loss": 0.0261,
"step": 305
},
{
"epoch": 1.046601111586148,
"grad_norm": 0.2797681439324777,
"learning_rate": 1.41e-05,
"loss": 0.023,
"step": 306
},
{
"epoch": 1.0500213766566908,
"grad_norm": 0.3197672217160271,
"learning_rate": 1.41e-05,
"loss": 0.0258,
"step": 307
},
{
"epoch": 1.0534416417272339,
"grad_norm": 0.2834421342617412,
"learning_rate": 1.41e-05,
"loss": 0.0231,
"step": 308
},
{
"epoch": 1.056861906797777,
"grad_norm": 0.30638464271397325,
"learning_rate": 1.41e-05,
"loss": 0.0237,
"step": 309
},
{
"epoch": 1.0602821718683197,
"grad_norm": 0.2772969145815675,
"learning_rate": 1.41e-05,
"loss": 0.0227,
"step": 310
},
{
"epoch": 1.0637024369388628,
"grad_norm": 0.27940000335528603,
"learning_rate": 1.41e-05,
"loss": 0.0232,
"step": 311
},
{
"epoch": 1.0671227020094056,
"grad_norm": 0.2793001905583242,
"learning_rate": 1.41e-05,
"loss": 0.024,
"step": 312
},
{
"epoch": 1.0705429670799487,
"grad_norm": 0.2615333153159237,
"learning_rate": 1.41e-05,
"loss": 0.0219,
"step": 313
},
{
"epoch": 1.0739632321504917,
"grad_norm": 0.2876480036224866,
"learning_rate": 1.41e-05,
"loss": 0.0253,
"step": 314
},
{
"epoch": 1.0773834972210345,
"grad_norm": 0.30478384773242695,
"learning_rate": 1.41e-05,
"loss": 0.0252,
"step": 315
},
{
"epoch": 1.0808037622915776,
"grad_norm": 0.2528846190302502,
"learning_rate": 1.41e-05,
"loss": 0.0206,
"step": 316
},
{
"epoch": 1.0842240273621206,
"grad_norm": 0.2435449168695854,
"learning_rate": 1.41e-05,
"loss": 0.0224,
"step": 317
},
{
"epoch": 1.0876442924326635,
"grad_norm": 0.3079236522189953,
"learning_rate": 1.41e-05,
"loss": 0.0253,
"step": 318
},
{
"epoch": 1.0910645575032065,
"grad_norm": 0.2675432502865711,
"learning_rate": 1.41e-05,
"loss": 0.0239,
"step": 319
},
{
"epoch": 1.0944848225737496,
"grad_norm": 0.3004587131742902,
"learning_rate": 1.41e-05,
"loss": 0.0239,
"step": 320
},
{
"epoch": 1.0979050876442924,
"grad_norm": 0.29332378399603554,
"learning_rate": 1.41e-05,
"loss": 0.0247,
"step": 321
},
{
"epoch": 1.1013253527148354,
"grad_norm": 0.23762802516720216,
"learning_rate": 1.41e-05,
"loss": 0.0242,
"step": 322
},
{
"epoch": 1.1047456177853783,
"grad_norm": 0.26342688417785715,
"learning_rate": 1.41e-05,
"loss": 0.0244,
"step": 323
},
{
"epoch": 1.1081658828559213,
"grad_norm": 0.3055498089807232,
"learning_rate": 1.41e-05,
"loss": 0.0265,
"step": 324
},
{
"epoch": 1.1115861479264644,
"grad_norm": 0.23787182811525248,
"learning_rate": 1.41e-05,
"loss": 0.0226,
"step": 325
},
{
"epoch": 1.1150064129970072,
"grad_norm": 0.2647940637873088,
"learning_rate": 1.41e-05,
"loss": 0.0242,
"step": 326
},
{
"epoch": 1.1184266780675503,
"grad_norm": 0.2913294201237873,
"learning_rate": 1.41e-05,
"loss": 0.0261,
"step": 327
},
{
"epoch": 1.1218469431380933,
"grad_norm": 0.2581915579915153,
"learning_rate": 1.41e-05,
"loss": 0.0228,
"step": 328
},
{
"epoch": 1.1252672082086361,
"grad_norm": 0.2500171384316944,
"learning_rate": 1.41e-05,
"loss": 0.0206,
"step": 329
},
{
"epoch": 1.1286874732791792,
"grad_norm": 0.2689440388389032,
"learning_rate": 1.41e-05,
"loss": 0.0237,
"step": 330
},
{
"epoch": 1.1321077383497222,
"grad_norm": 0.2551497009606492,
"learning_rate": 1.41e-05,
"loss": 0.0216,
"step": 331
},
{
"epoch": 1.135528003420265,
"grad_norm": 0.25757413049163996,
"learning_rate": 1.41e-05,
"loss": 0.0224,
"step": 332
},
{
"epoch": 1.138948268490808,
"grad_norm": 0.2699506485343775,
"learning_rate": 1.41e-05,
"loss": 0.0218,
"step": 333
},
{
"epoch": 1.142368533561351,
"grad_norm": 0.25767905743807257,
"learning_rate": 1.41e-05,
"loss": 0.0227,
"step": 334
},
{
"epoch": 1.145788798631894,
"grad_norm": 0.2273460686067317,
"learning_rate": 1.41e-05,
"loss": 0.0226,
"step": 335
},
{
"epoch": 1.149209063702437,
"grad_norm": 0.31200680170359746,
"learning_rate": 1.41e-05,
"loss": 0.0196,
"step": 336
},
{
"epoch": 1.1526293287729799,
"grad_norm": 0.23994204361556526,
"learning_rate": 1.41e-05,
"loss": 0.022,
"step": 337
},
{
"epoch": 1.156049593843523,
"grad_norm": 0.2867059420441838,
"learning_rate": 1.41e-05,
"loss": 0.0256,
"step": 338
},
{
"epoch": 1.1594698589140657,
"grad_norm": 0.2523475692407159,
"learning_rate": 1.41e-05,
"loss": 0.0212,
"step": 339
},
{
"epoch": 1.1628901239846088,
"grad_norm": 0.24499307458764866,
"learning_rate": 1.41e-05,
"loss": 0.0198,
"step": 340
},
{
"epoch": 1.1663103890551518,
"grad_norm": 0.2436283905820348,
"learning_rate": 1.41e-05,
"loss": 0.0216,
"step": 341
},
{
"epoch": 1.1697306541256947,
"grad_norm": 0.25602523324809817,
"learning_rate": 1.41e-05,
"loss": 0.0216,
"step": 342
},
{
"epoch": 1.1731509191962377,
"grad_norm": 0.25021177697750296,
"learning_rate": 1.41e-05,
"loss": 0.0233,
"step": 343
},
{
"epoch": 1.1765711842667808,
"grad_norm": 0.28379855080073285,
"learning_rate": 1.41e-05,
"loss": 0.0259,
"step": 344
},
{
"epoch": 1.1799914493373236,
"grad_norm": 0.25642024168179856,
"learning_rate": 1.41e-05,
"loss": 0.0208,
"step": 345
},
{
"epoch": 1.1834117144078666,
"grad_norm": 0.24053576849839317,
"learning_rate": 1.41e-05,
"loss": 0.0206,
"step": 346
},
{
"epoch": 1.1868319794784097,
"grad_norm": 0.26458302891901536,
"learning_rate": 1.41e-05,
"loss": 0.025,
"step": 347
},
{
"epoch": 1.1902522445489525,
"grad_norm": 0.25340954518832737,
"learning_rate": 1.41e-05,
"loss": 0.0232,
"step": 348
},
{
"epoch": 1.1936725096194956,
"grad_norm": 0.2822910057628071,
"learning_rate": 1.41e-05,
"loss": 0.0258,
"step": 349
},
{
"epoch": 1.1970927746900384,
"grad_norm": 0.24401825117116366,
"learning_rate": 1.41e-05,
"loss": 0.0207,
"step": 350
},
{
"epoch": 1.2005130397605814,
"grad_norm": 0.25351547147013626,
"learning_rate": 1.41e-05,
"loss": 0.0232,
"step": 351
},
{
"epoch": 1.2039333048311245,
"grad_norm": 0.2514277079230116,
"learning_rate": 1.41e-05,
"loss": 0.0236,
"step": 352
},
{
"epoch": 1.2073535699016673,
"grad_norm": 0.2499624840027022,
"learning_rate": 1.41e-05,
"loss": 0.0218,
"step": 353
},
{
"epoch": 1.2107738349722104,
"grad_norm": 0.25541034452698597,
"learning_rate": 1.41e-05,
"loss": 0.0226,
"step": 354
},
{
"epoch": 1.2141941000427534,
"grad_norm": 0.26116617349172355,
"learning_rate": 1.41e-05,
"loss": 0.0233,
"step": 355
},
{
"epoch": 1.2176143651132962,
"grad_norm": 0.3144220810007225,
"learning_rate": 1.41e-05,
"loss": 0.0238,
"step": 356
},
{
"epoch": 1.2210346301838393,
"grad_norm": 0.2397648420075889,
"learning_rate": 1.41e-05,
"loss": 0.0218,
"step": 357
},
{
"epoch": 1.2244548952543823,
"grad_norm": 0.23056631938102404,
"learning_rate": 1.41e-05,
"loss": 0.0211,
"step": 358
},
{
"epoch": 1.2278751603249252,
"grad_norm": 0.6359133071495683,
"learning_rate": 1.41e-05,
"loss": 0.0221,
"step": 359
},
{
"epoch": 1.2312954253954682,
"grad_norm": 0.2494306454740628,
"learning_rate": 1.41e-05,
"loss": 0.0212,
"step": 360
},
{
"epoch": 1.234715690466011,
"grad_norm": 0.2661512294789303,
"learning_rate": 1.41e-05,
"loss": 0.0254,
"step": 361
},
{
"epoch": 1.238135955536554,
"grad_norm": 0.7710920857502603,
"learning_rate": 1.41e-05,
"loss": 0.025,
"step": 362
},
{
"epoch": 1.2415562206070971,
"grad_norm": 0.3942637887185519,
"learning_rate": 1.41e-05,
"loss": 0.0236,
"step": 363
},
{
"epoch": 1.24497648567764,
"grad_norm": 0.27087010860427124,
"learning_rate": 1.41e-05,
"loss": 0.0252,
"step": 364
},
{
"epoch": 1.248396750748183,
"grad_norm": 0.24919842057781497,
"learning_rate": 1.41e-05,
"loss": 0.022,
"step": 365
},
{
"epoch": 1.2518170158187258,
"grad_norm": 0.247824008548709,
"learning_rate": 1.41e-05,
"loss": 0.0235,
"step": 366
},
{
"epoch": 1.255237280889269,
"grad_norm": 0.2113113650016047,
"learning_rate": 1.41e-05,
"loss": 0.0202,
"step": 367
},
{
"epoch": 1.258657545959812,
"grad_norm": 0.2259169472785117,
"learning_rate": 1.41e-05,
"loss": 0.0192,
"step": 368
},
{
"epoch": 1.262077811030355,
"grad_norm": 0.2752050707801685,
"learning_rate": 1.41e-05,
"loss": 0.0254,
"step": 369
},
{
"epoch": 1.2654980761008978,
"grad_norm": 0.23830392409535447,
"learning_rate": 1.41e-05,
"loss": 0.0229,
"step": 370
},
{
"epoch": 1.2689183411714409,
"grad_norm": 0.24302497993054667,
"learning_rate": 1.41e-05,
"loss": 0.0219,
"step": 371
},
{
"epoch": 1.2723386062419837,
"grad_norm": 0.24504347526605771,
"learning_rate": 1.41e-05,
"loss": 0.0224,
"step": 372
},
{
"epoch": 1.2757588713125267,
"grad_norm": 0.2346626865644856,
"learning_rate": 1.41e-05,
"loss": 0.0209,
"step": 373
},
{
"epoch": 1.2791791363830698,
"grad_norm": 0.24245589828220748,
"learning_rate": 1.41e-05,
"loss": 0.0218,
"step": 374
},
{
"epoch": 1.2825994014536126,
"grad_norm": 0.282140360657227,
"learning_rate": 1.41e-05,
"loss": 0.0234,
"step": 375
},
{
"epoch": 1.2860196665241557,
"grad_norm": 0.26692130820889887,
"learning_rate": 1.41e-05,
"loss": 0.02,
"step": 376
},
{
"epoch": 1.2894399315946985,
"grad_norm": 0.2691231899902035,
"learning_rate": 1.41e-05,
"loss": 0.0221,
"step": 377
},
{
"epoch": 1.2928601966652415,
"grad_norm": 0.2281687474950334,
"learning_rate": 1.41e-05,
"loss": 0.0222,
"step": 378
},
{
"epoch": 1.2962804617357846,
"grad_norm": 0.373488505328009,
"learning_rate": 1.41e-05,
"loss": 0.0217,
"step": 379
},
{
"epoch": 1.2997007268063274,
"grad_norm": 0.2345889821941022,
"learning_rate": 1.41e-05,
"loss": 0.021,
"step": 380
},
{
"epoch": 1.3031209918768705,
"grad_norm": 0.2105687923684266,
"learning_rate": 1.41e-05,
"loss": 0.0204,
"step": 381
},
{
"epoch": 1.3065412569474133,
"grad_norm": 0.23464935916200427,
"learning_rate": 1.41e-05,
"loss": 0.0213,
"step": 382
},
{
"epoch": 1.3099615220179563,
"grad_norm": 0.23646699592209017,
"learning_rate": 1.41e-05,
"loss": 0.0211,
"step": 383
},
{
"epoch": 1.3133817870884994,
"grad_norm": 0.2631733185003215,
"learning_rate": 1.41e-05,
"loss": 0.0219,
"step": 384
},
{
"epoch": 1.3168020521590424,
"grad_norm": 0.25931322588284866,
"learning_rate": 1.41e-05,
"loss": 0.0229,
"step": 385
},
{
"epoch": 1.3202223172295853,
"grad_norm": 0.28688530125524797,
"learning_rate": 1.41e-05,
"loss": 0.0226,
"step": 386
},
{
"epoch": 1.3236425823001283,
"grad_norm": 0.25729658498109925,
"learning_rate": 1.41e-05,
"loss": 0.0229,
"step": 387
},
{
"epoch": 1.3270628473706712,
"grad_norm": 0.25760420815172724,
"learning_rate": 1.41e-05,
"loss": 0.0207,
"step": 388
},
{
"epoch": 1.3304831124412142,
"grad_norm": 0.32657940169239236,
"learning_rate": 1.41e-05,
"loss": 0.0217,
"step": 389
},
{
"epoch": 1.3339033775117572,
"grad_norm": 0.22557643602886415,
"learning_rate": 1.41e-05,
"loss": 0.0204,
"step": 390
},
{
"epoch": 1.3373236425823,
"grad_norm": 0.2661843657709089,
"learning_rate": 1.41e-05,
"loss": 0.0243,
"step": 391
},
{
"epoch": 1.3407439076528431,
"grad_norm": 0.2974504589558172,
"learning_rate": 1.41e-05,
"loss": 0.0241,
"step": 392
},
{
"epoch": 1.344164172723386,
"grad_norm": 0.26843823846639436,
"learning_rate": 1.41e-05,
"loss": 0.0234,
"step": 393
},
{
"epoch": 1.347584437793929,
"grad_norm": 0.3218845205293079,
"learning_rate": 1.41e-05,
"loss": 0.0221,
"step": 394
},
{
"epoch": 1.351004702864472,
"grad_norm": 0.2428094674581459,
"learning_rate": 1.41e-05,
"loss": 0.0222,
"step": 395
},
{
"epoch": 1.354424967935015,
"grad_norm": 0.2455927243682846,
"learning_rate": 1.41e-05,
"loss": 0.0211,
"step": 396
},
{
"epoch": 1.357845233005558,
"grad_norm": 0.2733123552071639,
"learning_rate": 1.41e-05,
"loss": 0.0212,
"step": 397
},
{
"epoch": 1.361265498076101,
"grad_norm": 0.2659612332359629,
"learning_rate": 1.41e-05,
"loss": 0.0203,
"step": 398
},
{
"epoch": 1.3646857631466438,
"grad_norm": 0.25254417828478626,
"learning_rate": 1.41e-05,
"loss": 0.0201,
"step": 399
},
{
"epoch": 1.3681060282171869,
"grad_norm": 0.2598602588125554,
"learning_rate": 1.41e-05,
"loss": 0.0217,
"step": 400
},
{
"epoch": 1.37152629328773,
"grad_norm": 0.25849620548333013,
"learning_rate": 1.41e-05,
"loss": 0.0209,
"step": 401
},
{
"epoch": 1.3749465583582727,
"grad_norm": 0.3137023430125288,
"learning_rate": 1.41e-05,
"loss": 0.0211,
"step": 402
},
{
"epoch": 1.3783668234288158,
"grad_norm": 0.2113162287336589,
"learning_rate": 1.41e-05,
"loss": 0.0195,
"step": 403
},
{
"epoch": 1.3817870884993586,
"grad_norm": 0.2503171178420045,
"learning_rate": 1.41e-05,
"loss": 0.0234,
"step": 404
},
{
"epoch": 1.3852073535699017,
"grad_norm": 0.23119898543579737,
"learning_rate": 1.41e-05,
"loss": 0.0209,
"step": 405
},
{
"epoch": 1.3886276186404447,
"grad_norm": 0.26039590115583117,
"learning_rate": 1.41e-05,
"loss": 0.0244,
"step": 406
},
{
"epoch": 1.3920478837109875,
"grad_norm": 0.23886269607593336,
"learning_rate": 1.41e-05,
"loss": 0.0223,
"step": 407
},
{
"epoch": 1.3954681487815306,
"grad_norm": 0.2714710828662534,
"learning_rate": 1.41e-05,
"loss": 0.0247,
"step": 408
},
{
"epoch": 1.3988884138520734,
"grad_norm": 0.27319555229782644,
"learning_rate": 1.41e-05,
"loss": 0.0244,
"step": 409
},
{
"epoch": 1.4023086789226165,
"grad_norm": 0.24019773624911636,
"learning_rate": 1.41e-05,
"loss": 0.0218,
"step": 410
},
{
"epoch": 1.4057289439931595,
"grad_norm": 0.2708671308101268,
"learning_rate": 1.41e-05,
"loss": 0.0219,
"step": 411
},
{
"epoch": 1.4091492090637026,
"grad_norm": 0.22702757974948617,
"learning_rate": 1.41e-05,
"loss": 0.0198,
"step": 412
},
{
"epoch": 1.4125694741342454,
"grad_norm": 0.25555770375998416,
"learning_rate": 1.41e-05,
"loss": 0.0224,
"step": 413
},
{
"epoch": 1.4159897392047884,
"grad_norm": 0.23780595691689027,
"learning_rate": 1.41e-05,
"loss": 0.0213,
"step": 414
},
{
"epoch": 1.4194100042753313,
"grad_norm": 0.26451555701259444,
"learning_rate": 1.41e-05,
"loss": 0.0211,
"step": 415
},
{
"epoch": 1.4228302693458743,
"grad_norm": 0.24436230077463678,
"learning_rate": 1.41e-05,
"loss": 0.0226,
"step": 416
},
{
"epoch": 1.4262505344164174,
"grad_norm": 0.2216762882841438,
"learning_rate": 1.41e-05,
"loss": 0.021,
"step": 417
},
{
"epoch": 1.4296707994869602,
"grad_norm": 0.26405628674467824,
"learning_rate": 1.41e-05,
"loss": 0.0219,
"step": 418
},
{
"epoch": 1.4330910645575032,
"grad_norm": 0.23921814427898216,
"learning_rate": 1.41e-05,
"loss": 0.021,
"step": 419
},
{
"epoch": 1.436511329628046,
"grad_norm": 0.265453432302215,
"learning_rate": 1.41e-05,
"loss": 0.0226,
"step": 420
},
{
"epoch": 1.4399315946985891,
"grad_norm": 0.24492268359315303,
"learning_rate": 1.41e-05,
"loss": 0.0227,
"step": 421
},
{
"epoch": 1.4433518597691322,
"grad_norm": 0.25191935351384814,
"learning_rate": 1.41e-05,
"loss": 0.0215,
"step": 422
},
{
"epoch": 1.4467721248396752,
"grad_norm": 0.27197993846345997,
"learning_rate": 1.41e-05,
"loss": 0.0229,
"step": 423
},
{
"epoch": 1.450192389910218,
"grad_norm": 0.2075718815542033,
"learning_rate": 1.41e-05,
"loss": 0.0183,
"step": 424
},
{
"epoch": 1.453612654980761,
"grad_norm": 0.26376818444878,
"learning_rate": 1.41e-05,
"loss": 0.0221,
"step": 425
},
{
"epoch": 1.457032920051304,
"grad_norm": 0.26080564628665626,
"learning_rate": 1.41e-05,
"loss": 0.0229,
"step": 426
},
{
"epoch": 1.460453185121847,
"grad_norm": 0.2434588148667514,
"learning_rate": 1.41e-05,
"loss": 0.022,
"step": 427
},
{
"epoch": 1.46387345019239,
"grad_norm": 0.20943782903685337,
"learning_rate": 1.41e-05,
"loss": 0.0185,
"step": 428
},
{
"epoch": 1.4672937152629328,
"grad_norm": 0.24301387125104526,
"learning_rate": 1.41e-05,
"loss": 0.0208,
"step": 429
},
{
"epoch": 1.470713980333476,
"grad_norm": 0.25766203422027834,
"learning_rate": 1.41e-05,
"loss": 0.0237,
"step": 430
},
{
"epoch": 1.4741342454040187,
"grad_norm": 0.21556702000374744,
"learning_rate": 1.41e-05,
"loss": 0.0213,
"step": 431
},
{
"epoch": 1.4775545104745618,
"grad_norm": 0.23784909822124217,
"learning_rate": 1.41e-05,
"loss": 0.0207,
"step": 432
},
{
"epoch": 1.4809747755451048,
"grad_norm": 0.22414128247781562,
"learning_rate": 1.41e-05,
"loss": 0.0198,
"step": 433
},
{
"epoch": 1.4843950406156476,
"grad_norm": 0.23386987857579358,
"learning_rate": 1.41e-05,
"loss": 0.0201,
"step": 434
},
{
"epoch": 1.4878153056861907,
"grad_norm": 0.2853961947266083,
"learning_rate": 1.41e-05,
"loss": 0.023,
"step": 435
},
{
"epoch": 1.4912355707567335,
"grad_norm": 0.24974092416116495,
"learning_rate": 1.41e-05,
"loss": 0.0209,
"step": 436
},
{
"epoch": 1.4946558358272766,
"grad_norm": 0.23615654630506394,
"learning_rate": 1.41e-05,
"loss": 0.022,
"step": 437
},
{
"epoch": 1.4980761008978196,
"grad_norm": 0.2284456001474283,
"learning_rate": 1.41e-05,
"loss": 0.0195,
"step": 438
},
{
"epoch": 1.5014963659683627,
"grad_norm": 0.26063959990204033,
"learning_rate": 1.41e-05,
"loss": 0.0258,
"step": 439
},
{
"epoch": 1.5049166310389055,
"grad_norm": 0.24423575575966894,
"learning_rate": 1.41e-05,
"loss": 0.0215,
"step": 440
},
{
"epoch": 1.5083368961094483,
"grad_norm": 0.24270403980054853,
"learning_rate": 1.41e-05,
"loss": 0.0198,
"step": 441
},
{
"epoch": 1.5117571611799914,
"grad_norm": 0.24660594248828135,
"learning_rate": 1.41e-05,
"loss": 0.0208,
"step": 442
},
{
"epoch": 1.5151774262505344,
"grad_norm": 0.23764056229245814,
"learning_rate": 1.41e-05,
"loss": 0.0185,
"step": 443
},
{
"epoch": 1.5185976913210775,
"grad_norm": 0.2327212499624357,
"learning_rate": 1.41e-05,
"loss": 0.0205,
"step": 444
},
{
"epoch": 1.5220179563916203,
"grad_norm": 0.2644477025039793,
"learning_rate": 1.41e-05,
"loss": 0.0222,
"step": 445
},
{
"epoch": 1.5254382214621633,
"grad_norm": 0.2385202884230959,
"learning_rate": 1.41e-05,
"loss": 0.023,
"step": 446
},
{
"epoch": 1.5288584865327062,
"grad_norm": 0.24308327876991292,
"learning_rate": 1.41e-05,
"loss": 0.0198,
"step": 447
},
{
"epoch": 1.5322787516032492,
"grad_norm": 0.30634205660699,
"learning_rate": 1.41e-05,
"loss": 0.026,
"step": 448
},
{
"epoch": 1.5356990166737923,
"grad_norm": 0.21217473130594544,
"learning_rate": 1.41e-05,
"loss": 0.0199,
"step": 449
},
{
"epoch": 1.5391192817443353,
"grad_norm": 0.2658497297922587,
"learning_rate": 1.41e-05,
"loss": 0.0221,
"step": 450
},
{
"epoch": 1.5425395468148781,
"grad_norm": 0.21481338352800414,
"learning_rate": 1.41e-05,
"loss": 0.0197,
"step": 451
},
{
"epoch": 1.545959811885421,
"grad_norm": 0.27821761611964035,
"learning_rate": 1.41e-05,
"loss": 0.0246,
"step": 452
},
{
"epoch": 1.549380076955964,
"grad_norm": 0.2416104328737188,
"learning_rate": 1.41e-05,
"loss": 0.0216,
"step": 453
},
{
"epoch": 1.552800342026507,
"grad_norm": 0.2322000952546936,
"learning_rate": 1.41e-05,
"loss": 0.0219,
"step": 454
},
{
"epoch": 1.5562206070970501,
"grad_norm": 0.23752597721176905,
"learning_rate": 1.41e-05,
"loss": 0.0218,
"step": 455
},
{
"epoch": 1.559640872167593,
"grad_norm": 0.24308738119455534,
"learning_rate": 1.41e-05,
"loss": 0.0205,
"step": 456
},
{
"epoch": 1.563061137238136,
"grad_norm": 0.26073824102297316,
"learning_rate": 1.41e-05,
"loss": 0.0247,
"step": 457
},
{
"epoch": 1.5664814023086788,
"grad_norm": 0.24050011455971732,
"learning_rate": 1.41e-05,
"loss": 0.0223,
"step": 458
},
{
"epoch": 1.5699016673792219,
"grad_norm": 0.21610207139710053,
"learning_rate": 1.41e-05,
"loss": 0.0192,
"step": 459
},
{
"epoch": 1.573321932449765,
"grad_norm": 0.25801793315676924,
"learning_rate": 1.41e-05,
"loss": 0.0236,
"step": 460
},
{
"epoch": 1.576742197520308,
"grad_norm": 0.21228131497792052,
"learning_rate": 1.41e-05,
"loss": 0.0191,
"step": 461
},
{
"epoch": 1.5801624625908508,
"grad_norm": 0.25197778439538576,
"learning_rate": 1.41e-05,
"loss": 0.022,
"step": 462
},
{
"epoch": 1.5835827276613936,
"grad_norm": 0.2565258961786049,
"learning_rate": 1.41e-05,
"loss": 0.0219,
"step": 463
},
{
"epoch": 1.5870029927319367,
"grad_norm": 0.2559923791329211,
"learning_rate": 1.41e-05,
"loss": 0.0217,
"step": 464
},
{
"epoch": 1.5904232578024797,
"grad_norm": 0.2535527705987336,
"learning_rate": 1.41e-05,
"loss": 0.0221,
"step": 465
},
{
"epoch": 1.5938435228730228,
"grad_norm": 0.28495146888910583,
"learning_rate": 1.41e-05,
"loss": 0.0272,
"step": 466
},
{
"epoch": 1.5972637879435656,
"grad_norm": 0.26151357261060504,
"learning_rate": 1.41e-05,
"loss": 0.022,
"step": 467
},
{
"epoch": 1.6006840530141084,
"grad_norm": 0.24889528453574525,
"learning_rate": 1.41e-05,
"loss": 0.0218,
"step": 468
},
{
"epoch": 1.6041043180846515,
"grad_norm": 0.21219419933729458,
"learning_rate": 1.41e-05,
"loss": 0.0176,
"step": 469
},
{
"epoch": 1.6075245831551945,
"grad_norm": 0.24995940249087767,
"learning_rate": 1.41e-05,
"loss": 0.0207,
"step": 470
},
{
"epoch": 1.6109448482257376,
"grad_norm": 0.23890689663178338,
"learning_rate": 1.41e-05,
"loss": 0.0204,
"step": 471
},
{
"epoch": 1.6143651132962804,
"grad_norm": 0.23090495982296502,
"learning_rate": 1.41e-05,
"loss": 0.0209,
"step": 472
},
{
"epoch": 1.6177853783668235,
"grad_norm": 0.24984846656221682,
"learning_rate": 1.41e-05,
"loss": 0.0191,
"step": 473
},
{
"epoch": 1.6212056434373663,
"grad_norm": 0.2864149291763989,
"learning_rate": 1.41e-05,
"loss": 0.0243,
"step": 474
},
{
"epoch": 1.6246259085079093,
"grad_norm": 0.23371681508272213,
"learning_rate": 1.41e-05,
"loss": 0.0197,
"step": 475
},
{
"epoch": 1.6280461735784524,
"grad_norm": 0.21807560914421245,
"learning_rate": 1.41e-05,
"loss": 0.019,
"step": 476
},
{
"epoch": 1.6314664386489954,
"grad_norm": 0.24325192491721911,
"learning_rate": 1.41e-05,
"loss": 0.0212,
"step": 477
},
{
"epoch": 1.6348867037195383,
"grad_norm": 0.21802834032775958,
"learning_rate": 1.41e-05,
"loss": 0.0183,
"step": 478
},
{
"epoch": 1.638306968790081,
"grad_norm": 0.25015961868430003,
"learning_rate": 1.41e-05,
"loss": 0.0206,
"step": 479
},
{
"epoch": 1.6417272338606241,
"grad_norm": 0.2047993695964016,
"learning_rate": 1.41e-05,
"loss": 0.0166,
"step": 480
},
{
"epoch": 1.6451474989311672,
"grad_norm": 0.24407449809464615,
"learning_rate": 1.41e-05,
"loss": 0.0209,
"step": 481
},
{
"epoch": 1.6485677640017102,
"grad_norm": 0.22905836500660803,
"learning_rate": 1.41e-05,
"loss": 0.0197,
"step": 482
},
{
"epoch": 1.651988029072253,
"grad_norm": 0.2302659533712275,
"learning_rate": 1.41e-05,
"loss": 0.0184,
"step": 483
},
{
"epoch": 1.655408294142796,
"grad_norm": 0.22469778783055896,
"learning_rate": 1.41e-05,
"loss": 0.022,
"step": 484
},
{
"epoch": 1.658828559213339,
"grad_norm": 0.2403880470710616,
"learning_rate": 1.41e-05,
"loss": 0.0184,
"step": 485
},
{
"epoch": 1.662248824283882,
"grad_norm": 0.21080166415704205,
"learning_rate": 1.41e-05,
"loss": 0.0194,
"step": 486
},
{
"epoch": 1.665669089354425,
"grad_norm": 0.24366180394977457,
"learning_rate": 1.41e-05,
"loss": 0.0212,
"step": 487
},
{
"epoch": 1.669089354424968,
"grad_norm": 0.2445725370169137,
"learning_rate": 1.41e-05,
"loss": 0.0201,
"step": 488
},
{
"epoch": 1.672509619495511,
"grad_norm": 0.2400419596996237,
"learning_rate": 1.41e-05,
"loss": 0.0217,
"step": 489
},
{
"epoch": 1.6759298845660537,
"grad_norm": 0.22845732992118456,
"learning_rate": 1.41e-05,
"loss": 0.0193,
"step": 490
},
{
"epoch": 1.6793501496365968,
"grad_norm": 0.21963101467880122,
"learning_rate": 1.41e-05,
"loss": 0.0205,
"step": 491
},
{
"epoch": 1.6827704147071398,
"grad_norm": 0.23832272298065416,
"learning_rate": 1.41e-05,
"loss": 0.0208,
"step": 492
},
{
"epoch": 1.6861906797776829,
"grad_norm": 0.2701770424720032,
"learning_rate": 1.41e-05,
"loss": 0.0245,
"step": 493
},
{
"epoch": 1.6896109448482257,
"grad_norm": 0.21999352850800136,
"learning_rate": 1.41e-05,
"loss": 0.0191,
"step": 494
},
{
"epoch": 1.6930312099187685,
"grad_norm": 0.2303202011396172,
"learning_rate": 1.41e-05,
"loss": 0.0206,
"step": 495
},
{
"epoch": 1.6964514749893116,
"grad_norm": 0.23207638496444852,
"learning_rate": 1.41e-05,
"loss": 0.0196,
"step": 496
},
{
"epoch": 1.6998717400598546,
"grad_norm": 0.27476502715987833,
"learning_rate": 1.41e-05,
"loss": 0.0243,
"step": 497
},
{
"epoch": 1.7032920051303977,
"grad_norm": 0.25943127784425596,
"learning_rate": 1.41e-05,
"loss": 0.0209,
"step": 498
},
{
"epoch": 1.7067122702009405,
"grad_norm": 0.2582673258172837,
"learning_rate": 1.41e-05,
"loss": 0.0213,
"step": 499
},
{
"epoch": 1.7101325352714836,
"grad_norm": 0.2270336558731971,
"learning_rate": 1.41e-05,
"loss": 0.0207,
"step": 500
},
{
"epoch": 1.7135528003420264,
"grad_norm": 0.21716332073489333,
"learning_rate": 1.41e-05,
"loss": 0.017,
"step": 501
},
{
"epoch": 1.7169730654125694,
"grad_norm": 0.22361076006822656,
"learning_rate": 1.41e-05,
"loss": 0.0198,
"step": 502
},
{
"epoch": 1.7203933304831125,
"grad_norm": 0.25453906115757163,
"learning_rate": 1.41e-05,
"loss": 0.0215,
"step": 503
},
{
"epoch": 1.7238135955536555,
"grad_norm": 0.2639274732447762,
"learning_rate": 1.41e-05,
"loss": 0.0211,
"step": 504
},
{
"epoch": 1.7272338606241984,
"grad_norm": 0.25322752853555974,
"learning_rate": 1.41e-05,
"loss": 0.0228,
"step": 505
},
{
"epoch": 1.7306541256947412,
"grad_norm": 0.23285923741403275,
"learning_rate": 1.41e-05,
"loss": 0.0205,
"step": 506
},
{
"epoch": 1.7340743907652842,
"grad_norm": 0.2323064412599781,
"learning_rate": 1.41e-05,
"loss": 0.0195,
"step": 507
},
{
"epoch": 1.7374946558358273,
"grad_norm": 0.23653121619208753,
"learning_rate": 1.41e-05,
"loss": 0.0211,
"step": 508
},
{
"epoch": 1.7409149209063703,
"grad_norm": 0.2576532470063006,
"learning_rate": 1.41e-05,
"loss": 0.0216,
"step": 509
},
{
"epoch": 1.7443351859769132,
"grad_norm": 0.3031982152602418,
"learning_rate": 1.41e-05,
"loss": 0.0212,
"step": 510
},
{
"epoch": 1.7477554510474562,
"grad_norm": 0.21353603367638405,
"learning_rate": 1.41e-05,
"loss": 0.0187,
"step": 511
},
{
"epoch": 1.751175716117999,
"grad_norm": 0.23897983103928835,
"learning_rate": 1.41e-05,
"loss": 0.0211,
"step": 512
},
{
"epoch": 1.754595981188542,
"grad_norm": 0.26759279857241974,
"learning_rate": 1.41e-05,
"loss": 0.0225,
"step": 513
},
{
"epoch": 1.7580162462590851,
"grad_norm": 0.21204297476027956,
"learning_rate": 1.41e-05,
"loss": 0.0182,
"step": 514
},
{
"epoch": 1.7614365113296282,
"grad_norm": 0.2522156014013123,
"learning_rate": 1.41e-05,
"loss": 0.0224,
"step": 515
},
{
"epoch": 1.764856776400171,
"grad_norm": 0.2483681502861828,
"learning_rate": 1.41e-05,
"loss": 0.0225,
"step": 516
},
{
"epoch": 1.7682770414707139,
"grad_norm": 0.2757743080951287,
"learning_rate": 1.41e-05,
"loss": 0.0252,
"step": 517
},
{
"epoch": 1.771697306541257,
"grad_norm": 0.20783407363518988,
"learning_rate": 1.41e-05,
"loss": 0.0189,
"step": 518
},
{
"epoch": 1.7751175716118,
"grad_norm": 0.25743244499425244,
"learning_rate": 1.41e-05,
"loss": 0.022,
"step": 519
},
{
"epoch": 1.778537836682343,
"grad_norm": 0.23017428673046214,
"learning_rate": 1.41e-05,
"loss": 0.0195,
"step": 520
},
{
"epoch": 1.7819581017528858,
"grad_norm": 0.2315895066049246,
"learning_rate": 1.41e-05,
"loss": 0.0178,
"step": 521
},
{
"epoch": 1.7853783668234287,
"grad_norm": 0.22782043412403868,
"learning_rate": 1.41e-05,
"loss": 0.0215,
"step": 522
},
{
"epoch": 1.7887986318939717,
"grad_norm": 0.2724892881429743,
"learning_rate": 1.41e-05,
"loss": 0.0224,
"step": 523
},
{
"epoch": 1.7922188969645148,
"grad_norm": 0.28330092514107247,
"learning_rate": 1.41e-05,
"loss": 0.02,
"step": 524
},
{
"epoch": 1.7956391620350578,
"grad_norm": 0.23561517254595046,
"learning_rate": 1.41e-05,
"loss": 0.0199,
"step": 525
},
{
"epoch": 1.7990594271056009,
"grad_norm": 0.27080028676550905,
"learning_rate": 1.41e-05,
"loss": 0.0213,
"step": 526
},
{
"epoch": 1.8024796921761437,
"grad_norm": 0.20908394348220563,
"learning_rate": 1.41e-05,
"loss": 0.0181,
"step": 527
},
{
"epoch": 1.8058999572466865,
"grad_norm": 0.24764828335378633,
"learning_rate": 1.41e-05,
"loss": 0.0208,
"step": 528
},
{
"epoch": 1.8093202223172296,
"grad_norm": 0.29733366903532993,
"learning_rate": 1.41e-05,
"loss": 0.0218,
"step": 529
},
{
"epoch": 1.8127404873877726,
"grad_norm": 0.2708827452564214,
"learning_rate": 1.41e-05,
"loss": 0.0226,
"step": 530
},
{
"epoch": 1.8161607524583157,
"grad_norm": 0.2887461777468973,
"learning_rate": 1.41e-05,
"loss": 0.0226,
"step": 531
},
{
"epoch": 1.8195810175288585,
"grad_norm": 0.2774042332926017,
"learning_rate": 1.41e-05,
"loss": 0.0246,
"step": 532
},
{
"epoch": 1.8230012825994013,
"grad_norm": 0.23877829686783267,
"learning_rate": 1.41e-05,
"loss": 0.0202,
"step": 533
},
{
"epoch": 1.8264215476699444,
"grad_norm": 0.2544233995482896,
"learning_rate": 1.41e-05,
"loss": 0.0204,
"step": 534
},
{
"epoch": 1.8298418127404874,
"grad_norm": 0.26365092898440745,
"learning_rate": 1.41e-05,
"loss": 0.0223,
"step": 535
},
{
"epoch": 1.8332620778110305,
"grad_norm": 0.21178009134118125,
"learning_rate": 1.41e-05,
"loss": 0.018,
"step": 536
},
{
"epoch": 1.8366823428815733,
"grad_norm": 0.27115664081831004,
"learning_rate": 1.41e-05,
"loss": 0.0211,
"step": 537
},
{
"epoch": 1.8401026079521163,
"grad_norm": 0.2653048117342068,
"learning_rate": 1.41e-05,
"loss": 0.0212,
"step": 538
},
{
"epoch": 1.8435228730226592,
"grad_norm": 0.2757728122775812,
"learning_rate": 1.41e-05,
"loss": 0.0204,
"step": 539
},
{
"epoch": 1.8469431380932022,
"grad_norm": 0.2664986190346045,
"learning_rate": 1.41e-05,
"loss": 0.0207,
"step": 540
},
{
"epoch": 1.8503634031637453,
"grad_norm": 0.265518302380727,
"learning_rate": 1.41e-05,
"loss": 0.0179,
"step": 541
},
{
"epoch": 1.8537836682342883,
"grad_norm": 0.2895623799588168,
"learning_rate": 1.41e-05,
"loss": 0.0228,
"step": 542
},
{
"epoch": 1.8572039333048311,
"grad_norm": 0.23284728474850006,
"learning_rate": 1.41e-05,
"loss": 0.0192,
"step": 543
},
{
"epoch": 1.860624198375374,
"grad_norm": 0.21864913411739378,
"learning_rate": 1.41e-05,
"loss": 0.0178,
"step": 544
},
{
"epoch": 1.864044463445917,
"grad_norm": 0.25369925309969443,
"learning_rate": 1.41e-05,
"loss": 0.0204,
"step": 545
},
{
"epoch": 1.86746472851646,
"grad_norm": 0.22162789215208448,
"learning_rate": 1.41e-05,
"loss": 0.0214,
"step": 546
},
{
"epoch": 1.870884993587003,
"grad_norm": 0.228575157771765,
"learning_rate": 1.41e-05,
"loss": 0.0215,
"step": 547
},
{
"epoch": 1.874305258657546,
"grad_norm": 0.29177202889082,
"learning_rate": 1.41e-05,
"loss": 0.0231,
"step": 548
},
{
"epoch": 1.8777255237280888,
"grad_norm": 0.2526426700464145,
"learning_rate": 1.41e-05,
"loss": 0.022,
"step": 549
},
{
"epoch": 1.8811457887986318,
"grad_norm": 0.25824463173696005,
"learning_rate": 1.41e-05,
"loss": 0.0238,
"step": 550
},
{
"epoch": 1.8845660538691749,
"grad_norm": 0.22931141277969364,
"learning_rate": 1.41e-05,
"loss": 0.0195,
"step": 551
},
{
"epoch": 1.887986318939718,
"grad_norm": 0.20128540781283585,
"learning_rate": 1.41e-05,
"loss": 0.018,
"step": 552
},
{
"epoch": 1.891406584010261,
"grad_norm": 0.2726453682164885,
"learning_rate": 1.41e-05,
"loss": 0.0215,
"step": 553
},
{
"epoch": 1.8948268490808038,
"grad_norm": 0.2379175413754306,
"learning_rate": 1.41e-05,
"loss": 0.02,
"step": 554
},
{
"epoch": 1.8982471141513466,
"grad_norm": 0.2657340317097815,
"learning_rate": 1.41e-05,
"loss": 0.0215,
"step": 555
},
{
"epoch": 1.9016673792218897,
"grad_norm": 0.2469655254797632,
"learning_rate": 1.41e-05,
"loss": 0.0218,
"step": 556
},
{
"epoch": 1.9050876442924327,
"grad_norm": 0.2306898046638619,
"learning_rate": 1.41e-05,
"loss": 0.0206,
"step": 557
},
{
"epoch": 1.9085079093629758,
"grad_norm": 0.2501966986368517,
"learning_rate": 1.41e-05,
"loss": 0.0208,
"step": 558
},
{
"epoch": 1.9119281744335186,
"grad_norm": 0.26337448659750273,
"learning_rate": 1.41e-05,
"loss": 0.0224,
"step": 559
},
{
"epoch": 1.9153484395040614,
"grad_norm": 0.23788126235745702,
"learning_rate": 1.41e-05,
"loss": 0.0207,
"step": 560
},
{
"epoch": 1.9187687045746045,
"grad_norm": 0.24400460810027635,
"learning_rate": 1.41e-05,
"loss": 0.0199,
"step": 561
},
{
"epoch": 1.9221889696451475,
"grad_norm": 0.2524032823359455,
"learning_rate": 1.41e-05,
"loss": 0.02,
"step": 562
},
{
"epoch": 1.9256092347156906,
"grad_norm": 0.2288025108079512,
"learning_rate": 1.41e-05,
"loss": 0.0201,
"step": 563
},
{
"epoch": 1.9290294997862334,
"grad_norm": 0.269452001004884,
"learning_rate": 1.41e-05,
"loss": 0.0214,
"step": 564
},
{
"epoch": 1.9324497648567764,
"grad_norm": 0.2316242696085615,
"learning_rate": 1.41e-05,
"loss": 0.0196,
"step": 565
},
{
"epoch": 1.9358700299273193,
"grad_norm": 0.221180291707861,
"learning_rate": 1.41e-05,
"loss": 0.0177,
"step": 566
},
{
"epoch": 1.9392902949978623,
"grad_norm": 0.2607363855888593,
"learning_rate": 1.41e-05,
"loss": 0.0192,
"step": 567
},
{
"epoch": 1.9427105600684054,
"grad_norm": 0.24514765767262625,
"learning_rate": 1.41e-05,
"loss": 0.0182,
"step": 568
},
{
"epoch": 1.9461308251389484,
"grad_norm": 0.28916534222959817,
"learning_rate": 1.41e-05,
"loss": 0.0207,
"step": 569
},
{
"epoch": 1.9495510902094912,
"grad_norm": 0.26414654167917073,
"learning_rate": 1.41e-05,
"loss": 0.0204,
"step": 570
},
{
"epoch": 1.952971355280034,
"grad_norm": 0.2482098982197082,
"learning_rate": 1.41e-05,
"loss": 0.0206,
"step": 571
},
{
"epoch": 1.9563916203505771,
"grad_norm": 0.2608562249261069,
"learning_rate": 1.41e-05,
"loss": 0.0208,
"step": 572
},
{
"epoch": 1.9598118854211202,
"grad_norm": 0.25525358687825855,
"learning_rate": 1.41e-05,
"loss": 0.0213,
"step": 573
},
{
"epoch": 1.9632321504916632,
"grad_norm": 0.23125518036524792,
"learning_rate": 1.41e-05,
"loss": 0.0212,
"step": 574
},
{
"epoch": 1.966652415562206,
"grad_norm": 0.2432676298521834,
"learning_rate": 1.41e-05,
"loss": 0.0187,
"step": 575
},
{
"epoch": 1.9700726806327489,
"grad_norm": 0.24502134522738972,
"learning_rate": 1.41e-05,
"loss": 0.0194,
"step": 576
},
{
"epoch": 1.973492945703292,
"grad_norm": 0.26553777081307495,
"learning_rate": 1.41e-05,
"loss": 0.0212,
"step": 577
},
{
"epoch": 1.976913210773835,
"grad_norm": 0.24919535132544937,
"learning_rate": 1.41e-05,
"loss": 0.0208,
"step": 578
},
{
"epoch": 1.980333475844378,
"grad_norm": 0.2828767110434045,
"learning_rate": 1.41e-05,
"loss": 0.0238,
"step": 579
},
{
"epoch": 1.983753740914921,
"grad_norm": 0.26780635861278174,
"learning_rate": 1.41e-05,
"loss": 0.0225,
"step": 580
},
{
"epoch": 1.987174005985464,
"grad_norm": 0.26755832407143076,
"learning_rate": 1.41e-05,
"loss": 0.0197,
"step": 581
},
{
"epoch": 1.9905942710560067,
"grad_norm": 0.23150201894075045,
"learning_rate": 1.41e-05,
"loss": 0.0225,
"step": 582
},
{
"epoch": 1.9940145361265498,
"grad_norm": 0.2556938828383787,
"learning_rate": 1.41e-05,
"loss": 0.0217,
"step": 583
},
{
"epoch": 1.9974348011970928,
"grad_norm": 0.24779362245147543,
"learning_rate": 1.41e-05,
"loss": 0.0209,
"step": 584
},
{
"epoch": 2.000855066267636,
"grad_norm": 0.23220294742842815,
"learning_rate": 1.41e-05,
"loss": 0.0205,
"step": 585
},
{
"epoch": 2.004275331338179,
"grad_norm": 0.19611669457431613,
"learning_rate": 1.41e-05,
"loss": 0.0136,
"step": 586
},
{
"epoch": 2.0076955964087215,
"grad_norm": 0.18201675420540842,
"learning_rate": 1.41e-05,
"loss": 0.0146,
"step": 587
},
{
"epoch": 2.0111158614792646,
"grad_norm": 0.2111608412027396,
"learning_rate": 1.41e-05,
"loss": 0.0165,
"step": 588
},
{
"epoch": 2.0145361265498076,
"grad_norm": 0.19725473414903413,
"learning_rate": 1.41e-05,
"loss": 0.015,
"step": 589
},
{
"epoch": 2.0179563916203507,
"grad_norm": 0.21354984112177863,
"learning_rate": 1.41e-05,
"loss": 0.0161,
"step": 590
},
{
"epoch": 2.0213766566908937,
"grad_norm": 0.22940862450456587,
"learning_rate": 1.41e-05,
"loss": 0.0157,
"step": 591
},
{
"epoch": 2.0247969217614363,
"grad_norm": 0.18767782352642373,
"learning_rate": 1.41e-05,
"loss": 0.0152,
"step": 592
},
{
"epoch": 2.0282171868319794,
"grad_norm": 0.20643159409018694,
"learning_rate": 1.41e-05,
"loss": 0.0148,
"step": 593
},
{
"epoch": 2.0316374519025224,
"grad_norm": 0.2037046870595368,
"learning_rate": 1.41e-05,
"loss": 0.0138,
"step": 594
},
{
"epoch": 2.0350577169730655,
"grad_norm": 0.19705696567608205,
"learning_rate": 1.41e-05,
"loss": 0.0162,
"step": 595
},
{
"epoch": 2.0384779820436085,
"grad_norm": 0.21000255959562664,
"learning_rate": 1.41e-05,
"loss": 0.0156,
"step": 596
},
{
"epoch": 2.041898247114151,
"grad_norm": 0.1909421240987611,
"learning_rate": 1.41e-05,
"loss": 0.0136,
"step": 597
},
{
"epoch": 2.045318512184694,
"grad_norm": 0.22254593762119865,
"learning_rate": 1.41e-05,
"loss": 0.0155,
"step": 598
},
{
"epoch": 2.0487387772552372,
"grad_norm": 0.18039596470753322,
"learning_rate": 1.41e-05,
"loss": 0.0147,
"step": 599
},
{
"epoch": 2.0521590423257803,
"grad_norm": 0.1836910759989102,
"learning_rate": 1.41e-05,
"loss": 0.0139,
"step": 600
},
{
"epoch": 2.0555793073963233,
"grad_norm": 0.2007119695404956,
"learning_rate": 1.41e-05,
"loss": 0.015,
"step": 601
},
{
"epoch": 2.0589995724668664,
"grad_norm": 0.17881220811912174,
"learning_rate": 1.41e-05,
"loss": 0.016,
"step": 602
},
{
"epoch": 2.062419837537409,
"grad_norm": 0.18131752461639056,
"learning_rate": 1.41e-05,
"loss": 0.0152,
"step": 603
},
{
"epoch": 2.065840102607952,
"grad_norm": 0.2091101968035948,
"learning_rate": 1.41e-05,
"loss": 0.0147,
"step": 604
},
{
"epoch": 2.069260367678495,
"grad_norm": 0.1785580403127973,
"learning_rate": 1.41e-05,
"loss": 0.0142,
"step": 605
},
{
"epoch": 2.072680632749038,
"grad_norm": 0.1952402053161801,
"learning_rate": 1.41e-05,
"loss": 0.0143,
"step": 606
},
{
"epoch": 2.076100897819581,
"grad_norm": 0.16849209697986273,
"learning_rate": 1.41e-05,
"loss": 0.0135,
"step": 607
},
{
"epoch": 2.079521162890124,
"grad_norm": 0.18960749528358953,
"learning_rate": 1.41e-05,
"loss": 0.0158,
"step": 608
},
{
"epoch": 2.082941427960667,
"grad_norm": 0.1822523628091137,
"learning_rate": 1.41e-05,
"loss": 0.0157,
"step": 609
},
{
"epoch": 2.08636169303121,
"grad_norm": 0.22244344533378702,
"learning_rate": 1.41e-05,
"loss": 0.0145,
"step": 610
},
{
"epoch": 2.089781958101753,
"grad_norm": 0.16677137506686548,
"learning_rate": 1.41e-05,
"loss": 0.0144,
"step": 611
},
{
"epoch": 2.093202223172296,
"grad_norm": 0.19008011213315038,
"learning_rate": 1.41e-05,
"loss": 0.0149,
"step": 612
},
{
"epoch": 2.096622488242839,
"grad_norm": 0.1965146825157404,
"learning_rate": 1.41e-05,
"loss": 0.0158,
"step": 613
},
{
"epoch": 2.1000427533133816,
"grad_norm": 0.17095730271612053,
"learning_rate": 1.41e-05,
"loss": 0.0148,
"step": 614
},
{
"epoch": 2.1034630183839247,
"grad_norm": 0.18376098918489842,
"learning_rate": 1.41e-05,
"loss": 0.0133,
"step": 615
},
{
"epoch": 2.1068832834544677,
"grad_norm": 0.18859892344363186,
"learning_rate": 1.41e-05,
"loss": 0.0155,
"step": 616
},
{
"epoch": 2.110303548525011,
"grad_norm": 0.17678403587042746,
"learning_rate": 1.41e-05,
"loss": 0.013,
"step": 617
},
{
"epoch": 2.113723813595554,
"grad_norm": 0.18933984169673357,
"learning_rate": 1.41e-05,
"loss": 0.0152,
"step": 618
},
{
"epoch": 2.1171440786660964,
"grad_norm": 0.20060664740874196,
"learning_rate": 1.41e-05,
"loss": 0.0157,
"step": 619
},
{
"epoch": 2.1205643437366395,
"grad_norm": 0.1803038264349401,
"learning_rate": 1.41e-05,
"loss": 0.0141,
"step": 620
},
{
"epoch": 2.1239846088071825,
"grad_norm": 0.1910210164605926,
"learning_rate": 1.41e-05,
"loss": 0.015,
"step": 621
},
{
"epoch": 2.1274048738777256,
"grad_norm": 0.1794365441256405,
"learning_rate": 1.41e-05,
"loss": 0.0137,
"step": 622
},
{
"epoch": 2.1308251389482686,
"grad_norm": 0.15893205685478506,
"learning_rate": 1.41e-05,
"loss": 0.0136,
"step": 623
},
{
"epoch": 2.1342454040188112,
"grad_norm": 0.19724991519919266,
"learning_rate": 1.41e-05,
"loss": 0.0143,
"step": 624
},
{
"epoch": 2.1376656690893543,
"grad_norm": 0.17458605756275755,
"learning_rate": 1.41e-05,
"loss": 0.016,
"step": 625
},
{
"epoch": 2.1410859341598973,
"grad_norm": 0.172521593573168,
"learning_rate": 1.41e-05,
"loss": 0.0141,
"step": 626
},
{
"epoch": 2.1445061992304404,
"grad_norm": 0.14557636073307398,
"learning_rate": 1.41e-05,
"loss": 0.0135,
"step": 627
},
{
"epoch": 2.1479264643009834,
"grad_norm": 0.22843890765431576,
"learning_rate": 1.41e-05,
"loss": 0.016,
"step": 628
},
{
"epoch": 2.1513467293715265,
"grad_norm": 0.18196169286713992,
"learning_rate": 1.41e-05,
"loss": 0.0146,
"step": 629
},
{
"epoch": 2.154766994442069,
"grad_norm": 0.17272140560874683,
"learning_rate": 1.41e-05,
"loss": 0.0135,
"step": 630
},
{
"epoch": 2.158187259512612,
"grad_norm": 0.15595539359263064,
"learning_rate": 1.41e-05,
"loss": 0.014,
"step": 631
},
{
"epoch": 2.161607524583155,
"grad_norm": 0.16465865001486402,
"learning_rate": 1.41e-05,
"loss": 0.0136,
"step": 632
},
{
"epoch": 2.1650277896536982,
"grad_norm": 0.17986632980817158,
"learning_rate": 1.41e-05,
"loss": 0.015,
"step": 633
},
{
"epoch": 2.1684480547242413,
"grad_norm": 0.18929877257592365,
"learning_rate": 1.41e-05,
"loss": 0.0163,
"step": 634
},
{
"epoch": 2.171868319794784,
"grad_norm": 0.19542505698465773,
"learning_rate": 1.41e-05,
"loss": 0.0159,
"step": 635
},
{
"epoch": 2.175288584865327,
"grad_norm": 0.17114565781513597,
"learning_rate": 1.41e-05,
"loss": 0.0145,
"step": 636
},
{
"epoch": 2.17870884993587,
"grad_norm": 0.18331913158118285,
"learning_rate": 1.41e-05,
"loss": 0.0146,
"step": 637
},
{
"epoch": 2.182129115006413,
"grad_norm": 0.16055196305467795,
"learning_rate": 1.41e-05,
"loss": 0.015,
"step": 638
},
{
"epoch": 2.185549380076956,
"grad_norm": 0.15985440322290387,
"learning_rate": 1.41e-05,
"loss": 0.0141,
"step": 639
},
{
"epoch": 2.188969645147499,
"grad_norm": 0.16417247457666104,
"learning_rate": 1.41e-05,
"loss": 0.0145,
"step": 640
},
{
"epoch": 2.1923899102180417,
"grad_norm": 0.16210688538374043,
"learning_rate": 1.41e-05,
"loss": 0.0141,
"step": 641
},
{
"epoch": 2.195810175288585,
"grad_norm": 0.16634144533077172,
"learning_rate": 1.41e-05,
"loss": 0.014,
"step": 642
},
{
"epoch": 2.199230440359128,
"grad_norm": 0.18582606043218244,
"learning_rate": 1.41e-05,
"loss": 0.0156,
"step": 643
},
{
"epoch": 2.202650705429671,
"grad_norm": 0.1747184961523454,
"learning_rate": 1.41e-05,
"loss": 0.0164,
"step": 644
},
{
"epoch": 2.206070970500214,
"grad_norm": 0.15610336933105856,
"learning_rate": 1.41e-05,
"loss": 0.013,
"step": 645
},
{
"epoch": 2.2094912355707566,
"grad_norm": 0.16920938302300378,
"learning_rate": 1.41e-05,
"loss": 0.0139,
"step": 646
},
{
"epoch": 2.2129115006412996,
"grad_norm": 0.17648344233280136,
"learning_rate": 1.41e-05,
"loss": 0.0147,
"step": 647
},
{
"epoch": 2.2163317657118426,
"grad_norm": 0.17020885614859071,
"learning_rate": 1.41e-05,
"loss": 0.0155,
"step": 648
},
{
"epoch": 2.2197520307823857,
"grad_norm": 0.17716196526554184,
"learning_rate": 1.41e-05,
"loss": 0.0142,
"step": 649
},
{
"epoch": 2.2231722958529287,
"grad_norm": 0.17375550168851311,
"learning_rate": 1.41e-05,
"loss": 0.0144,
"step": 650
},
{
"epoch": 2.2265925609234714,
"grad_norm": 0.15660012931864775,
"learning_rate": 1.41e-05,
"loss": 0.0138,
"step": 651
},
{
"epoch": 2.2300128259940144,
"grad_norm": 0.192388171113886,
"learning_rate": 1.41e-05,
"loss": 0.0151,
"step": 652
},
{
"epoch": 2.2334330910645575,
"grad_norm": 0.17112223174304642,
"learning_rate": 1.41e-05,
"loss": 0.0137,
"step": 653
},
{
"epoch": 2.2368533561351005,
"grad_norm": 0.17046414663473125,
"learning_rate": 1.41e-05,
"loss": 0.0163,
"step": 654
},
{
"epoch": 2.2402736212056436,
"grad_norm": 0.14723103129157328,
"learning_rate": 1.41e-05,
"loss": 0.0141,
"step": 655
},
{
"epoch": 2.2436938862761866,
"grad_norm": 0.16260020026720062,
"learning_rate": 1.41e-05,
"loss": 0.0134,
"step": 656
},
{
"epoch": 2.247114151346729,
"grad_norm": 0.1605529186115587,
"learning_rate": 1.41e-05,
"loss": 0.0148,
"step": 657
},
{
"epoch": 2.2505344164172723,
"grad_norm": 0.1649326783596916,
"learning_rate": 1.41e-05,
"loss": 0.0146,
"step": 658
},
{
"epoch": 2.2539546814878153,
"grad_norm": 0.1664480794095166,
"learning_rate": 1.41e-05,
"loss": 0.0144,
"step": 659
},
{
"epoch": 2.2573749465583584,
"grad_norm": 0.16946179512795181,
"learning_rate": 1.41e-05,
"loss": 0.0146,
"step": 660
},
{
"epoch": 2.2607952116289014,
"grad_norm": 0.166434202624617,
"learning_rate": 1.41e-05,
"loss": 0.0135,
"step": 661
},
{
"epoch": 2.2642154766994445,
"grad_norm": 0.1739998754952991,
"learning_rate": 1.41e-05,
"loss": 0.0142,
"step": 662
},
{
"epoch": 2.267635741769987,
"grad_norm": 0.1809093145862042,
"learning_rate": 1.41e-05,
"loss": 0.014,
"step": 663
},
{
"epoch": 2.27105600684053,
"grad_norm": 0.15727492334875873,
"learning_rate": 1.41e-05,
"loss": 0.0137,
"step": 664
},
{
"epoch": 2.274476271911073,
"grad_norm": 0.18448939109027693,
"learning_rate": 1.41e-05,
"loss": 0.0157,
"step": 665
},
{
"epoch": 2.277896536981616,
"grad_norm": 0.18236721526184443,
"learning_rate": 1.41e-05,
"loss": 0.0166,
"step": 666
},
{
"epoch": 2.2813168020521593,
"grad_norm": 0.1695210468590141,
"learning_rate": 1.41e-05,
"loss": 0.0144,
"step": 667
},
{
"epoch": 2.284737067122702,
"grad_norm": 0.17157602170689792,
"learning_rate": 1.41e-05,
"loss": 0.0153,
"step": 668
},
{
"epoch": 2.288157332193245,
"grad_norm": 0.16583571772207567,
"learning_rate": 1.41e-05,
"loss": 0.0133,
"step": 669
},
{
"epoch": 2.291577597263788,
"grad_norm": 0.15309231682278962,
"learning_rate": 1.41e-05,
"loss": 0.0134,
"step": 670
},
{
"epoch": 2.294997862334331,
"grad_norm": 0.18010491103515566,
"learning_rate": 1.41e-05,
"loss": 0.0155,
"step": 671
},
{
"epoch": 2.298418127404874,
"grad_norm": 0.165226793210629,
"learning_rate": 1.41e-05,
"loss": 0.0134,
"step": 672
},
{
"epoch": 2.3018383924754167,
"grad_norm": 0.19017257841407276,
"learning_rate": 1.41e-05,
"loss": 0.016,
"step": 673
},
{
"epoch": 2.3052586575459597,
"grad_norm": 0.16105989698952458,
"learning_rate": 1.41e-05,
"loss": 0.0154,
"step": 674
},
{
"epoch": 2.3086789226165028,
"grad_norm": 0.17435397438208278,
"learning_rate": 1.41e-05,
"loss": 0.0136,
"step": 675
},
{
"epoch": 2.312099187687046,
"grad_norm": 0.15489352805926157,
"learning_rate": 1.41e-05,
"loss": 0.0149,
"step": 676
},
{
"epoch": 2.315519452757589,
"grad_norm": 0.15749574161575872,
"learning_rate": 1.41e-05,
"loss": 0.0136,
"step": 677
},
{
"epoch": 2.3189397178281315,
"grad_norm": 0.17861788971942064,
"learning_rate": 1.41e-05,
"loss": 0.015,
"step": 678
},
{
"epoch": 2.3223599828986745,
"grad_norm": 0.16852344878390402,
"learning_rate": 1.41e-05,
"loss": 0.0152,
"step": 679
},
{
"epoch": 2.3257802479692176,
"grad_norm": 0.1903544466095535,
"learning_rate": 1.41e-05,
"loss": 0.0146,
"step": 680
},
{
"epoch": 2.3292005130397606,
"grad_norm": 0.1938655036641079,
"learning_rate": 1.41e-05,
"loss": 0.0158,
"step": 681
},
{
"epoch": 2.3326207781103037,
"grad_norm": 0.20203642790445486,
"learning_rate": 1.41e-05,
"loss": 0.0143,
"step": 682
},
{
"epoch": 2.3360410431808463,
"grad_norm": 0.16328748761388734,
"learning_rate": 1.41e-05,
"loss": 0.0154,
"step": 683
},
{
"epoch": 2.3394613082513893,
"grad_norm": 0.17437303421082254,
"learning_rate": 1.41e-05,
"loss": 0.0149,
"step": 684
},
{
"epoch": 2.3428815733219324,
"grad_norm": 0.17762818156916235,
"learning_rate": 1.41e-05,
"loss": 0.0149,
"step": 685
},
{
"epoch": 2.3463018383924754,
"grad_norm": 0.18594818733099802,
"learning_rate": 1.41e-05,
"loss": 0.0161,
"step": 686
},
{
"epoch": 2.3497221034630185,
"grad_norm": 0.16139308140544603,
"learning_rate": 1.41e-05,
"loss": 0.0148,
"step": 687
},
{
"epoch": 2.3531423685335615,
"grad_norm": 0.16108681383332696,
"learning_rate": 1.41e-05,
"loss": 0.0139,
"step": 688
},
{
"epoch": 2.3565626336041046,
"grad_norm": 0.15992689822195585,
"learning_rate": 1.41e-05,
"loss": 0.0146,
"step": 689
},
{
"epoch": 2.359982898674647,
"grad_norm": 0.19759610045563245,
"learning_rate": 1.41e-05,
"loss": 0.0158,
"step": 690
},
{
"epoch": 2.36340316374519,
"grad_norm": 0.16149200857684542,
"learning_rate": 1.41e-05,
"loss": 0.014,
"step": 691
},
{
"epoch": 2.3668234288157333,
"grad_norm": 0.15072989325955685,
"learning_rate": 1.41e-05,
"loss": 0.0144,
"step": 692
},
{
"epoch": 2.3702436938862763,
"grad_norm": 0.14913939759012662,
"learning_rate": 1.41e-05,
"loss": 0.014,
"step": 693
},
{
"epoch": 2.3736639589568194,
"grad_norm": 0.15636976839856181,
"learning_rate": 1.41e-05,
"loss": 0.0153,
"step": 694
},
{
"epoch": 2.377084224027362,
"grad_norm": 0.18599274943363073,
"learning_rate": 1.41e-05,
"loss": 0.0141,
"step": 695
},
{
"epoch": 2.380504489097905,
"grad_norm": 0.16568136382204926,
"learning_rate": 1.41e-05,
"loss": 0.013,
"step": 696
},
{
"epoch": 2.383924754168448,
"grad_norm": 0.17332302727716312,
"learning_rate": 1.41e-05,
"loss": 0.0128,
"step": 697
},
{
"epoch": 2.387345019238991,
"grad_norm": 0.1879178833471297,
"learning_rate": 1.41e-05,
"loss": 0.0146,
"step": 698
},
{
"epoch": 2.390765284309534,
"grad_norm": 0.1726315547096704,
"learning_rate": 1.41e-05,
"loss": 0.0136,
"step": 699
},
{
"epoch": 2.3941855493800768,
"grad_norm": 0.1720038600446686,
"learning_rate": 1.41e-05,
"loss": 0.0158,
"step": 700
},
{
"epoch": 2.39760581445062,
"grad_norm": 0.17262771034044544,
"learning_rate": 1.41e-05,
"loss": 0.0146,
"step": 701
},
{
"epoch": 2.401026079521163,
"grad_norm": 0.18381965008540402,
"learning_rate": 1.41e-05,
"loss": 0.0151,
"step": 702
},
{
"epoch": 2.404446344591706,
"grad_norm": 0.1943786246553059,
"learning_rate": 1.41e-05,
"loss": 0.0148,
"step": 703
},
{
"epoch": 2.407866609662249,
"grad_norm": 0.15285642974808178,
"learning_rate": 1.41e-05,
"loss": 0.0133,
"step": 704
},
{
"epoch": 2.4112868747327916,
"grad_norm": 0.19330131469876405,
"learning_rate": 1.41e-05,
"loss": 0.0146,
"step": 705
},
{
"epoch": 2.4147071398033346,
"grad_norm": 0.16935279234922626,
"learning_rate": 1.41e-05,
"loss": 0.0143,
"step": 706
},
{
"epoch": 2.4181274048738777,
"grad_norm": 0.17130297774748693,
"learning_rate": 1.41e-05,
"loss": 0.016,
"step": 707
},
{
"epoch": 2.4215476699444207,
"grad_norm": 0.16649757139294383,
"learning_rate": 1.41e-05,
"loss": 0.0135,
"step": 708
},
{
"epoch": 2.4249679350149638,
"grad_norm": 0.1509762981692361,
"learning_rate": 1.41e-05,
"loss": 0.0131,
"step": 709
},
{
"epoch": 2.428388200085507,
"grad_norm": 0.16903612165580836,
"learning_rate": 1.41e-05,
"loss": 0.015,
"step": 710
},
{
"epoch": 2.4318084651560494,
"grad_norm": 0.16965158831335697,
"learning_rate": 1.41e-05,
"loss": 0.0139,
"step": 711
},
{
"epoch": 2.4352287302265925,
"grad_norm": 0.17489335625033847,
"learning_rate": 1.41e-05,
"loss": 0.0127,
"step": 712
},
{
"epoch": 2.4386489952971355,
"grad_norm": 0.17619733289193135,
"learning_rate": 1.41e-05,
"loss": 0.0144,
"step": 713
},
{
"epoch": 2.4420692603676786,
"grad_norm": 0.17617881507074173,
"learning_rate": 1.41e-05,
"loss": 0.0152,
"step": 714
},
{
"epoch": 2.4454895254382216,
"grad_norm": 0.18048518098716906,
"learning_rate": 1.41e-05,
"loss": 0.0139,
"step": 715
},
{
"epoch": 2.4489097905087647,
"grad_norm": 0.2051952297924243,
"learning_rate": 1.41e-05,
"loss": 0.0146,
"step": 716
},
{
"epoch": 2.4523300555793073,
"grad_norm": 0.15664970563386238,
"learning_rate": 1.41e-05,
"loss": 0.0143,
"step": 717
},
{
"epoch": 2.4557503206498503,
"grad_norm": 0.16743723595706328,
"learning_rate": 1.41e-05,
"loss": 0.0152,
"step": 718
},
{
"epoch": 2.4591705857203934,
"grad_norm": 0.17613442783841668,
"learning_rate": 1.41e-05,
"loss": 0.0162,
"step": 719
},
{
"epoch": 2.4625908507909364,
"grad_norm": 0.1818706957568455,
"learning_rate": 1.41e-05,
"loss": 0.0135,
"step": 720
},
{
"epoch": 2.4660111158614795,
"grad_norm": 0.17047836187664528,
"learning_rate": 1.41e-05,
"loss": 0.0141,
"step": 721
},
{
"epoch": 2.469431380932022,
"grad_norm": 0.17741467330052185,
"learning_rate": 1.41e-05,
"loss": 0.0158,
"step": 722
},
{
"epoch": 2.472851646002565,
"grad_norm": 0.1800167905144302,
"learning_rate": 1.41e-05,
"loss": 0.0156,
"step": 723
},
{
"epoch": 2.476271911073108,
"grad_norm": 0.1520569062216661,
"learning_rate": 1.41e-05,
"loss": 0.014,
"step": 724
},
{
"epoch": 2.4796921761436512,
"grad_norm": 0.19504143457216305,
"learning_rate": 1.41e-05,
"loss": 0.015,
"step": 725
},
{
"epoch": 2.4831124412141943,
"grad_norm": 0.16406995475864966,
"learning_rate": 1.41e-05,
"loss": 0.0152,
"step": 726
},
{
"epoch": 2.486532706284737,
"grad_norm": 0.1640747075139151,
"learning_rate": 1.41e-05,
"loss": 0.0138,
"step": 727
},
{
"epoch": 2.48995297135528,
"grad_norm": 0.17515606314180554,
"learning_rate": 1.41e-05,
"loss": 0.0156,
"step": 728
},
{
"epoch": 2.493373236425823,
"grad_norm": 0.17908072157399882,
"learning_rate": 1.41e-05,
"loss": 0.0155,
"step": 729
},
{
"epoch": 2.496793501496366,
"grad_norm": 0.15664812551985538,
"learning_rate": 1.41e-05,
"loss": 0.0147,
"step": 730
},
{
"epoch": 2.500213766566909,
"grad_norm": 0.165379248202876,
"learning_rate": 1.41e-05,
"loss": 0.0151,
"step": 731
},
{
"epoch": 2.5036340316374517,
"grad_norm": 0.15423115964706519,
"learning_rate": 1.41e-05,
"loss": 0.0148,
"step": 732
},
{
"epoch": 2.5070542967079947,
"grad_norm": 0.16485219288254147,
"learning_rate": 1.41e-05,
"loss": 0.0149,
"step": 733
},
{
"epoch": 2.510474561778538,
"grad_norm": 0.176637422447442,
"learning_rate": 1.41e-05,
"loss": 0.0154,
"step": 734
},
{
"epoch": 2.513894826849081,
"grad_norm": 0.17814754105281233,
"learning_rate": 1.41e-05,
"loss": 0.0154,
"step": 735
},
{
"epoch": 2.517315091919624,
"grad_norm": 0.17118656720070607,
"learning_rate": 1.41e-05,
"loss": 0.0155,
"step": 736
},
{
"epoch": 2.5207353569901665,
"grad_norm": 0.18260703508475823,
"learning_rate": 1.41e-05,
"loss": 0.0154,
"step": 737
},
{
"epoch": 2.52415562206071,
"grad_norm": 0.17080402514684412,
"learning_rate": 1.41e-05,
"loss": 0.0136,
"step": 738
},
{
"epoch": 2.5275758871312526,
"grad_norm": 0.15709135265175292,
"learning_rate": 1.41e-05,
"loss": 0.0134,
"step": 739
},
{
"epoch": 2.5309961522017956,
"grad_norm": 0.18246981305252055,
"learning_rate": 1.41e-05,
"loss": 0.0142,
"step": 740
},
{
"epoch": 2.5344164172723387,
"grad_norm": 0.1808585737198185,
"learning_rate": 1.41e-05,
"loss": 0.0139,
"step": 741
},
{
"epoch": 2.5378366823428817,
"grad_norm": 0.16454795740570335,
"learning_rate": 1.41e-05,
"loss": 0.0153,
"step": 742
},
{
"epoch": 2.541256947413425,
"grad_norm": 0.21032421935647805,
"learning_rate": 1.41e-05,
"loss": 0.0165,
"step": 743
},
{
"epoch": 2.5446772124839674,
"grad_norm": 0.15249763976570654,
"learning_rate": 1.41e-05,
"loss": 0.0139,
"step": 744
},
{
"epoch": 2.5480974775545104,
"grad_norm": 0.1613635941460105,
"learning_rate": 1.41e-05,
"loss": 0.0131,
"step": 745
},
{
"epoch": 2.5515177426250535,
"grad_norm": 0.1700060737892137,
"learning_rate": 1.41e-05,
"loss": 0.0145,
"step": 746
},
{
"epoch": 2.5549380076955965,
"grad_norm": 0.20547449901345968,
"learning_rate": 1.41e-05,
"loss": 0.0168,
"step": 747
},
{
"epoch": 2.5583582727661396,
"grad_norm": 0.16484966537277382,
"learning_rate": 1.41e-05,
"loss": 0.0142,
"step": 748
},
{
"epoch": 2.561778537836682,
"grad_norm": 0.17889219393433903,
"learning_rate": 1.41e-05,
"loss": 0.014,
"step": 749
},
{
"epoch": 2.5651988029072252,
"grad_norm": 0.17679721987917787,
"learning_rate": 1.41e-05,
"loss": 0.0138,
"step": 750
},
{
"epoch": 2.5686190679777683,
"grad_norm": 0.16210612559541202,
"learning_rate": 1.41e-05,
"loss": 0.0145,
"step": 751
},
{
"epoch": 2.5720393330483113,
"grad_norm": 0.17192503019428163,
"learning_rate": 1.41e-05,
"loss": 0.0143,
"step": 752
},
{
"epoch": 2.5754595981188544,
"grad_norm": 0.16114230533462182,
"learning_rate": 1.41e-05,
"loss": 0.0151,
"step": 753
},
{
"epoch": 2.578879863189397,
"grad_norm": 0.16683326324078276,
"learning_rate": 1.41e-05,
"loss": 0.014,
"step": 754
},
{
"epoch": 2.58230012825994,
"grad_norm": 0.15929030309974343,
"learning_rate": 1.41e-05,
"loss": 0.0135,
"step": 755
},
{
"epoch": 2.585720393330483,
"grad_norm": 0.17728404681160545,
"learning_rate": 1.41e-05,
"loss": 0.0134,
"step": 756
},
{
"epoch": 2.589140658401026,
"grad_norm": 0.16727441441267749,
"learning_rate": 1.41e-05,
"loss": 0.0143,
"step": 757
},
{
"epoch": 2.592560923471569,
"grad_norm": 0.21841845784314023,
"learning_rate": 1.41e-05,
"loss": 0.0149,
"step": 758
},
{
"epoch": 2.595981188542112,
"grad_norm": 0.1770465088709539,
"learning_rate": 1.41e-05,
"loss": 0.015,
"step": 759
},
{
"epoch": 2.599401453612655,
"grad_norm": 0.1737371304814951,
"learning_rate": 1.41e-05,
"loss": 0.0156,
"step": 760
},
{
"epoch": 2.602821718683198,
"grad_norm": 0.17069864594223258,
"learning_rate": 1.41e-05,
"loss": 0.0134,
"step": 761
},
{
"epoch": 2.606241983753741,
"grad_norm": 0.16609359338004503,
"learning_rate": 1.41e-05,
"loss": 0.0158,
"step": 762
},
{
"epoch": 2.609662248824284,
"grad_norm": 0.17378348121220705,
"learning_rate": 1.41e-05,
"loss": 0.0141,
"step": 763
},
{
"epoch": 2.6130825138948266,
"grad_norm": 0.17355912661147865,
"learning_rate": 1.41e-05,
"loss": 0.0142,
"step": 764
},
{
"epoch": 2.61650277896537,
"grad_norm": 0.20077889987954006,
"learning_rate": 1.41e-05,
"loss": 0.0154,
"step": 765
},
{
"epoch": 2.6199230440359127,
"grad_norm": 0.16419074893195182,
"learning_rate": 1.41e-05,
"loss": 0.014,
"step": 766
},
{
"epoch": 2.6233433091064557,
"grad_norm": 0.19476203022751112,
"learning_rate": 1.41e-05,
"loss": 0.0154,
"step": 767
},
{
"epoch": 2.626763574176999,
"grad_norm": 0.16568838103054007,
"learning_rate": 1.41e-05,
"loss": 0.0148,
"step": 768
},
{
"epoch": 2.630183839247542,
"grad_norm": 0.21099152906068663,
"learning_rate": 1.41e-05,
"loss": 0.017,
"step": 769
},
{
"epoch": 2.633604104318085,
"grad_norm": 0.17959731227982292,
"learning_rate": 1.41e-05,
"loss": 0.014,
"step": 770
},
{
"epoch": 2.6370243693886275,
"grad_norm": 0.16972780919186436,
"learning_rate": 1.41e-05,
"loss": 0.0139,
"step": 771
},
{
"epoch": 2.6404446344591705,
"grad_norm": 0.1842651161264449,
"learning_rate": 1.41e-05,
"loss": 0.0157,
"step": 772
},
{
"epoch": 2.6438648995297136,
"grad_norm": 0.17834439152890905,
"learning_rate": 1.41e-05,
"loss": 0.0153,
"step": 773
},
{
"epoch": 2.6472851646002566,
"grad_norm": 0.18047152749420897,
"learning_rate": 1.41e-05,
"loss": 0.0156,
"step": 774
},
{
"epoch": 2.6507054296707997,
"grad_norm": 0.17154124492966955,
"learning_rate": 1.41e-05,
"loss": 0.0147,
"step": 775
},
{
"epoch": 2.6541256947413423,
"grad_norm": 0.17549885521671557,
"learning_rate": 1.41e-05,
"loss": 0.0161,
"step": 776
},
{
"epoch": 2.6575459598118853,
"grad_norm": 0.17861344190439962,
"learning_rate": 1.41e-05,
"loss": 0.0151,
"step": 777
},
{
"epoch": 2.6609662248824284,
"grad_norm": 0.16730938712809765,
"learning_rate": 1.41e-05,
"loss": 0.0145,
"step": 778
},
{
"epoch": 2.6643864899529714,
"grad_norm": 0.1775818865748014,
"learning_rate": 1.41e-05,
"loss": 0.0159,
"step": 779
},
{
"epoch": 2.6678067550235145,
"grad_norm": 0.191333871334205,
"learning_rate": 1.41e-05,
"loss": 0.013,
"step": 780
},
{
"epoch": 2.671227020094057,
"grad_norm": 0.16907634204579375,
"learning_rate": 1.41e-05,
"loss": 0.0145,
"step": 781
},
{
"epoch": 2.6746472851646,
"grad_norm": 0.18178434604030544,
"learning_rate": 1.41e-05,
"loss": 0.0135,
"step": 782
},
{
"epoch": 2.678067550235143,
"grad_norm": 0.20503354023831713,
"learning_rate": 1.41e-05,
"loss": 0.0146,
"step": 783
},
{
"epoch": 2.6814878153056863,
"grad_norm": 0.16391814776977276,
"learning_rate": 1.41e-05,
"loss": 0.0156,
"step": 784
},
{
"epoch": 2.6849080803762293,
"grad_norm": 0.1839989787303562,
"learning_rate": 1.41e-05,
"loss": 0.0148,
"step": 785
},
{
"epoch": 2.688328345446772,
"grad_norm": 0.19236560848567033,
"learning_rate": 1.41e-05,
"loss": 0.0154,
"step": 786
},
{
"epoch": 2.691748610517315,
"grad_norm": 0.16811298067940522,
"learning_rate": 1.41e-05,
"loss": 0.0141,
"step": 787
},
{
"epoch": 2.695168875587858,
"grad_norm": 0.19975209346986805,
"learning_rate": 1.41e-05,
"loss": 0.0159,
"step": 788
},
{
"epoch": 2.698589140658401,
"grad_norm": 0.16338526349657181,
"learning_rate": 1.41e-05,
"loss": 0.0133,
"step": 789
},
{
"epoch": 2.702009405728944,
"grad_norm": 0.17351362318047445,
"learning_rate": 1.41e-05,
"loss": 0.0159,
"step": 790
},
{
"epoch": 2.7054296707994867,
"grad_norm": 0.17661074926464723,
"learning_rate": 1.41e-05,
"loss": 0.0154,
"step": 791
},
{
"epoch": 2.70884993587003,
"grad_norm": 0.1762002497021315,
"learning_rate": 1.41e-05,
"loss": 0.016,
"step": 792
},
{
"epoch": 2.712270200940573,
"grad_norm": 0.19710675588819915,
"learning_rate": 1.41e-05,
"loss": 0.0161,
"step": 793
},
{
"epoch": 2.715690466011116,
"grad_norm": 0.1692117872805568,
"learning_rate": 1.41e-05,
"loss": 0.014,
"step": 794
},
{
"epoch": 2.719110731081659,
"grad_norm": 0.20731978857327257,
"learning_rate": 1.41e-05,
"loss": 0.0156,
"step": 795
},
{
"epoch": 2.722530996152202,
"grad_norm": 0.17806201725432017,
"learning_rate": 1.41e-05,
"loss": 0.0147,
"step": 796
},
{
"epoch": 2.725951261222745,
"grad_norm": 0.16617443428776074,
"learning_rate": 1.41e-05,
"loss": 0.0143,
"step": 797
},
{
"epoch": 2.7293715262932876,
"grad_norm": 0.20059647446620907,
"learning_rate": 1.41e-05,
"loss": 0.014,
"step": 798
},
{
"epoch": 2.7327917913638307,
"grad_norm": 0.1773142221832559,
"learning_rate": 1.41e-05,
"loss": 0.0171,
"step": 799
},
{
"epoch": 2.7362120564343737,
"grad_norm": 0.1979846553461278,
"learning_rate": 1.41e-05,
"loss": 0.0139,
"step": 800
},
{
"epoch": 2.7396323215049168,
"grad_norm": 0.18511763263524342,
"learning_rate": 1.41e-05,
"loss": 0.0155,
"step": 801
},
{
"epoch": 2.74305258657546,
"grad_norm": 0.17699099585761474,
"learning_rate": 1.41e-05,
"loss": 0.0118,
"step": 802
},
{
"epoch": 2.7464728516460024,
"grad_norm": 0.19352657405112964,
"learning_rate": 1.41e-05,
"loss": 0.0163,
"step": 803
},
{
"epoch": 2.7498931167165455,
"grad_norm": 0.19135035210758095,
"learning_rate": 1.41e-05,
"loss": 0.0156,
"step": 804
},
{
"epoch": 2.7533133817870885,
"grad_norm": 0.1841148474869883,
"learning_rate": 1.41e-05,
"loss": 0.0155,
"step": 805
},
{
"epoch": 2.7567336468576316,
"grad_norm": 0.19008317628622337,
"learning_rate": 1.41e-05,
"loss": 0.0152,
"step": 806
},
{
"epoch": 2.7601539119281746,
"grad_norm": 0.1949140281276627,
"learning_rate": 1.41e-05,
"loss": 0.0143,
"step": 807
},
{
"epoch": 2.763574176998717,
"grad_norm": 0.1944281661049677,
"learning_rate": 1.41e-05,
"loss": 0.0145,
"step": 808
},
{
"epoch": 2.7669944420692603,
"grad_norm": 0.1674859606182937,
"learning_rate": 1.41e-05,
"loss": 0.0162,
"step": 809
},
{
"epoch": 2.7704147071398033,
"grad_norm": 0.18850185666818006,
"learning_rate": 1.41e-05,
"loss": 0.0148,
"step": 810
},
{
"epoch": 2.7738349722103464,
"grad_norm": 0.18475272615726215,
"learning_rate": 1.41e-05,
"loss": 0.0156,
"step": 811
},
{
"epoch": 2.7772552372808894,
"grad_norm": 0.15365265630124134,
"learning_rate": 1.41e-05,
"loss": 0.0134,
"step": 812
},
{
"epoch": 2.780675502351432,
"grad_norm": 0.16024686744719197,
"learning_rate": 1.41e-05,
"loss": 0.0143,
"step": 813
},
{
"epoch": 2.784095767421975,
"grad_norm": 0.18228181650312342,
"learning_rate": 1.41e-05,
"loss": 0.0142,
"step": 814
},
{
"epoch": 2.787516032492518,
"grad_norm": 0.15232575479299343,
"learning_rate": 1.41e-05,
"loss": 0.0141,
"step": 815
},
{
"epoch": 2.790936297563061,
"grad_norm": 0.19473439330433306,
"learning_rate": 1.41e-05,
"loss": 0.0141,
"step": 816
},
{
"epoch": 2.794356562633604,
"grad_norm": 0.20524717935488615,
"learning_rate": 1.41e-05,
"loss": 0.017,
"step": 817
},
{
"epoch": 2.797776827704147,
"grad_norm": 0.1908817371704934,
"learning_rate": 1.41e-05,
"loss": 0.0147,
"step": 818
},
{
"epoch": 2.8011970927746903,
"grad_norm": 0.1817449147111606,
"learning_rate": 1.41e-05,
"loss": 0.0146,
"step": 819
},
{
"epoch": 2.804617357845233,
"grad_norm": 0.20849626412897376,
"learning_rate": 1.41e-05,
"loss": 0.0144,
"step": 820
},
{
"epoch": 2.808037622915776,
"grad_norm": 0.17439115442992892,
"learning_rate": 1.41e-05,
"loss": 0.0165,
"step": 821
},
{
"epoch": 2.811457887986319,
"grad_norm": 0.184212269622675,
"learning_rate": 1.41e-05,
"loss": 0.0139,
"step": 822
},
{
"epoch": 2.814878153056862,
"grad_norm": 0.1742874984626096,
"learning_rate": 1.41e-05,
"loss": 0.0146,
"step": 823
},
{
"epoch": 2.818298418127405,
"grad_norm": 0.16854467961459083,
"learning_rate": 1.41e-05,
"loss": 0.0145,
"step": 824
},
{
"epoch": 2.8217186831979477,
"grad_norm": 0.17468169666793745,
"learning_rate": 1.41e-05,
"loss": 0.0149,
"step": 825
},
{
"epoch": 2.8251389482684908,
"grad_norm": 0.16936997795252084,
"learning_rate": 1.41e-05,
"loss": 0.0139,
"step": 826
},
{
"epoch": 2.828559213339034,
"grad_norm": 0.16572236927644424,
"learning_rate": 1.41e-05,
"loss": 0.0154,
"step": 827
},
{
"epoch": 2.831979478409577,
"grad_norm": 0.1651962024013233,
"learning_rate": 1.41e-05,
"loss": 0.0127,
"step": 828
},
{
"epoch": 2.83539974348012,
"grad_norm": 0.1657376085150308,
"learning_rate": 1.41e-05,
"loss": 0.0139,
"step": 829
},
{
"epoch": 2.8388200085506625,
"grad_norm": 0.17582133731054114,
"learning_rate": 1.41e-05,
"loss": 0.0131,
"step": 830
},
{
"epoch": 2.8422402736212056,
"grad_norm": 0.20723879555852162,
"learning_rate": 1.41e-05,
"loss": 0.0162,
"step": 831
},
{
"epoch": 2.8456605386917486,
"grad_norm": 0.6025905984134068,
"learning_rate": 1.41e-05,
"loss": 0.0142,
"step": 832
},
{
"epoch": 2.8490808037622917,
"grad_norm": 0.1678623970594971,
"learning_rate": 1.41e-05,
"loss": 0.0143,
"step": 833
},
{
"epoch": 2.8525010688328347,
"grad_norm": 0.21086844483510658,
"learning_rate": 1.41e-05,
"loss": 0.0158,
"step": 834
},
{
"epoch": 2.8559213339033773,
"grad_norm": 0.18493480224421863,
"learning_rate": 1.41e-05,
"loss": 0.0126,
"step": 835
},
{
"epoch": 2.8593415989739204,
"grad_norm": 0.1904246277238966,
"learning_rate": 1.41e-05,
"loss": 0.0155,
"step": 836
},
{
"epoch": 2.8627618640444634,
"grad_norm": 0.18774356330645445,
"learning_rate": 1.41e-05,
"loss": 0.0174,
"step": 837
},
{
"epoch": 2.8661821291150065,
"grad_norm": 0.20481371594917794,
"learning_rate": 1.41e-05,
"loss": 0.0132,
"step": 838
},
{
"epoch": 2.8696023941855495,
"grad_norm": 0.19259669092688833,
"learning_rate": 1.41e-05,
"loss": 0.0144,
"step": 839
},
{
"epoch": 2.873022659256092,
"grad_norm": 0.20161241494039436,
"learning_rate": 1.41e-05,
"loss": 0.0159,
"step": 840
},
{
"epoch": 2.876442924326635,
"grad_norm": 0.19885243582338097,
"learning_rate": 1.41e-05,
"loss": 0.0155,
"step": 841
},
{
"epoch": 2.8798631893971782,
"grad_norm": 0.18089297372993837,
"learning_rate": 1.41e-05,
"loss": 0.015,
"step": 842
},
{
"epoch": 2.8832834544677213,
"grad_norm": 0.19064879861955514,
"learning_rate": 1.41e-05,
"loss": 0.0132,
"step": 843
},
{
"epoch": 2.8867037195382643,
"grad_norm": 0.18907692558323141,
"learning_rate": 1.41e-05,
"loss": 0.0146,
"step": 844
},
{
"epoch": 2.890123984608807,
"grad_norm": 0.19585899032623924,
"learning_rate": 1.41e-05,
"loss": 0.0153,
"step": 845
},
{
"epoch": 2.8935442496793504,
"grad_norm": 0.22806745672139891,
"learning_rate": 1.41e-05,
"loss": 0.0169,
"step": 846
},
{
"epoch": 2.896964514749893,
"grad_norm": 0.19336051551290398,
"learning_rate": 1.41e-05,
"loss": 0.0155,
"step": 847
},
{
"epoch": 2.900384779820436,
"grad_norm": 0.17663309403974298,
"learning_rate": 1.41e-05,
"loss": 0.013,
"step": 848
},
{
"epoch": 2.903805044890979,
"grad_norm": 0.1884422733573296,
"learning_rate": 1.41e-05,
"loss": 0.0153,
"step": 849
},
{
"epoch": 2.907225309961522,
"grad_norm": 0.19574362737358483,
"learning_rate": 1.41e-05,
"loss": 0.016,
"step": 850
},
{
"epoch": 2.9106455750320652,
"grad_norm": 0.16218777133767662,
"learning_rate": 1.41e-05,
"loss": 0.0135,
"step": 851
},
{
"epoch": 2.914065840102608,
"grad_norm": 0.1844053227394077,
"learning_rate": 1.41e-05,
"loss": 0.0147,
"step": 852
},
{
"epoch": 2.917486105173151,
"grad_norm": 0.19292851510616071,
"learning_rate": 1.41e-05,
"loss": 0.0149,
"step": 853
},
{
"epoch": 2.920906370243694,
"grad_norm": 0.2089035835617463,
"learning_rate": 1.41e-05,
"loss": 0.017,
"step": 854
},
{
"epoch": 2.924326635314237,
"grad_norm": 0.7055863227979324,
"learning_rate": 1.41e-05,
"loss": 0.016,
"step": 855
},
{
"epoch": 2.92774690038478,
"grad_norm": 0.18518793131577801,
"learning_rate": 1.41e-05,
"loss": 0.0141,
"step": 856
},
{
"epoch": 2.9311671654553226,
"grad_norm": 0.17688744692401304,
"learning_rate": 1.41e-05,
"loss": 0.0144,
"step": 857
},
{
"epoch": 2.9345874305258657,
"grad_norm": 0.20202970506605689,
"learning_rate": 1.41e-05,
"loss": 0.0144,
"step": 858
},
{
"epoch": 2.9380076955964087,
"grad_norm": 0.19190788298961772,
"learning_rate": 1.41e-05,
"loss": 0.0154,
"step": 859
},
{
"epoch": 2.941427960666952,
"grad_norm": 0.19843345575867563,
"learning_rate": 1.41e-05,
"loss": 0.0167,
"step": 860
},
{
"epoch": 2.944848225737495,
"grad_norm": 0.18043514187233164,
"learning_rate": 1.41e-05,
"loss": 0.0124,
"step": 861
},
{
"epoch": 2.9482684908080374,
"grad_norm": 0.1852741160641503,
"learning_rate": 1.41e-05,
"loss": 0.0153,
"step": 862
},
{
"epoch": 2.9516887558785805,
"grad_norm": 0.19552493492009015,
"learning_rate": 1.41e-05,
"loss": 0.0161,
"step": 863
},
{
"epoch": 2.9551090209491235,
"grad_norm": 0.1972068517674688,
"learning_rate": 1.41e-05,
"loss": 0.0151,
"step": 864
},
{
"epoch": 2.9585292860196666,
"grad_norm": 0.2016593632101422,
"learning_rate": 1.41e-05,
"loss": 0.0161,
"step": 865
},
{
"epoch": 2.9619495510902096,
"grad_norm": 0.18720575761526204,
"learning_rate": 1.41e-05,
"loss": 0.014,
"step": 866
},
{
"epoch": 2.9653698161607522,
"grad_norm": 0.1948235931544798,
"learning_rate": 1.41e-05,
"loss": 0.0142,
"step": 867
},
{
"epoch": 2.9687900812312953,
"grad_norm": 0.18580729685744493,
"learning_rate": 1.41e-05,
"loss": 0.0148,
"step": 868
},
{
"epoch": 2.9722103463018383,
"grad_norm": 0.1747215712068137,
"learning_rate": 1.41e-05,
"loss": 0.0148,
"step": 869
},
{
"epoch": 2.9756306113723814,
"grad_norm": 0.21255162912353637,
"learning_rate": 1.41e-05,
"loss": 0.0149,
"step": 870
},
{
"epoch": 2.9790508764429244,
"grad_norm": 0.18848472150934822,
"learning_rate": 1.41e-05,
"loss": 0.0141,
"step": 871
},
{
"epoch": 2.982471141513467,
"grad_norm": 0.18016163548131123,
"learning_rate": 1.41e-05,
"loss": 0.0147,
"step": 872
},
{
"epoch": 2.9858914065840105,
"grad_norm": 0.1795705522164676,
"learning_rate": 1.41e-05,
"loss": 0.014,
"step": 873
},
{
"epoch": 2.989311671654553,
"grad_norm": 0.17401025878895968,
"learning_rate": 1.41e-05,
"loss": 0.0138,
"step": 874
},
{
"epoch": 2.992731936725096,
"grad_norm": 0.1783730795841426,
"learning_rate": 1.41e-05,
"loss": 0.0132,
"step": 875
},
{
"epoch": 2.9961522017956392,
"grad_norm": 0.20051631261532152,
"learning_rate": 1.41e-05,
"loss": 0.0166,
"step": 876
},
{
"epoch": 2.9995724668661823,
"grad_norm": 0.2056542584608787,
"learning_rate": 1.41e-05,
"loss": 0.0155,
"step": 877
}
],
"logging_steps": 1.0,
"max_steps": 1460,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 367251916062720.0,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}