Yukang Chen
Initial commit for nvila-8b-sft-length32k-epoch5-r1-solution
39a524d
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 5.0,
"eval_steps": 500,
"global_step": 780,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.01,
"grad_norm": 3.157794332025747,
"learning_rate": 8.974358974358974e-08,
"loss": 0.5858,
"step": 1
},
{
"epoch": 0.01,
"grad_norm": 3.0242493468561227,
"learning_rate": 1.7948717948717948e-07,
"loss": 0.6092,
"step": 2
},
{
"epoch": 0.02,
"grad_norm": 2.7745230799899288,
"learning_rate": 2.692307692307692e-07,
"loss": 0.5905,
"step": 3
},
{
"epoch": 0.03,
"grad_norm": 3.0024697741882873,
"learning_rate": 3.5897435897435896e-07,
"loss": 0.6548,
"step": 4
},
{
"epoch": 0.03,
"grad_norm": 3.131554528264668,
"learning_rate": 4.4871794871794865e-07,
"loss": 0.6324,
"step": 5
},
{
"epoch": 0.04,
"grad_norm": 3.6251050835532377,
"learning_rate": 5.384615384615384e-07,
"loss": 0.5575,
"step": 6
},
{
"epoch": 0.04,
"grad_norm": 3.114615782713657,
"learning_rate": 6.282051282051282e-07,
"loss": 0.5568,
"step": 7
},
{
"epoch": 0.05,
"grad_norm": 2.883211828526437,
"learning_rate": 7.179487179487179e-07,
"loss": 0.6242,
"step": 8
},
{
"epoch": 0.06,
"grad_norm": 3.097582117144788,
"learning_rate": 8.076923076923077e-07,
"loss": 0.5944,
"step": 9
},
{
"epoch": 0.06,
"grad_norm": 2.8984916101799536,
"learning_rate": 8.974358974358973e-07,
"loss": 0.6359,
"step": 10
},
{
"epoch": 0.07,
"grad_norm": 2.7500845254975426,
"learning_rate": 9.871794871794872e-07,
"loss": 0.6004,
"step": 11
},
{
"epoch": 0.08,
"grad_norm": 2.9464098867975723,
"learning_rate": 1.0769230769230769e-06,
"loss": 0.6183,
"step": 12
},
{
"epoch": 0.08,
"grad_norm": 2.742713392529882,
"learning_rate": 1.1666666666666666e-06,
"loss": 0.5473,
"step": 13
},
{
"epoch": 0.09,
"grad_norm": 2.554893065802702,
"learning_rate": 1.2564102564102565e-06,
"loss": 0.5389,
"step": 14
},
{
"epoch": 0.1,
"grad_norm": 2.8440244944229423,
"learning_rate": 1.3461538461538462e-06,
"loss": 0.563,
"step": 15
},
{
"epoch": 0.1,
"grad_norm": 2.3051065909946815,
"learning_rate": 1.4358974358974359e-06,
"loss": 0.5832,
"step": 16
},
{
"epoch": 0.11,
"grad_norm": 2.0186256795764828,
"learning_rate": 1.5256410256410255e-06,
"loss": 0.5626,
"step": 17
},
{
"epoch": 0.12,
"grad_norm": 2.3684781031146307,
"learning_rate": 1.6153846153846154e-06,
"loss": 0.5593,
"step": 18
},
{
"epoch": 0.12,
"grad_norm": 2.1296922237578837,
"learning_rate": 1.7051282051282051e-06,
"loss": 0.5083,
"step": 19
},
{
"epoch": 0.13,
"grad_norm": 2.2333623615611975,
"learning_rate": 1.7948717948717946e-06,
"loss": 0.5701,
"step": 20
},
{
"epoch": 0.13,
"grad_norm": 2.3716536064596365,
"learning_rate": 1.8846153846153845e-06,
"loss": 0.7116,
"step": 21
},
{
"epoch": 0.14,
"grad_norm": 2.6249010351941537,
"learning_rate": 1.9743589743589744e-06,
"loss": 0.6667,
"step": 22
},
{
"epoch": 0.15,
"grad_norm": 2.025777944483822,
"learning_rate": 2.064102564102564e-06,
"loss": 0.5018,
"step": 23
},
{
"epoch": 0.15,
"grad_norm": 2.0694837392683163,
"learning_rate": 2.1538461538461538e-06,
"loss": 0.5145,
"step": 24
},
{
"epoch": 0.16,
"grad_norm": 1.7977920012457864,
"learning_rate": 2.243589743589744e-06,
"loss": 0.4688,
"step": 25
},
{
"epoch": 0.17,
"grad_norm": 2.030335865996782,
"learning_rate": 2.333333333333333e-06,
"loss": 0.4719,
"step": 26
},
{
"epoch": 0.17,
"grad_norm": 2.2391443374029514,
"learning_rate": 2.423076923076923e-06,
"loss": 0.5303,
"step": 27
},
{
"epoch": 0.18,
"grad_norm": 1.9912422183097178,
"learning_rate": 2.512820512820513e-06,
"loss": 0.5431,
"step": 28
},
{
"epoch": 0.19,
"grad_norm": 2.0255309552484095,
"learning_rate": 2.6025641025641026e-06,
"loss": 0.5824,
"step": 29
},
{
"epoch": 0.19,
"grad_norm": 1.8770215122908283,
"learning_rate": 2.6923076923076923e-06,
"loss": 0.5082,
"step": 30
},
{
"epoch": 0.2,
"grad_norm": 1.98939906714483,
"learning_rate": 2.782051282051282e-06,
"loss": 0.5154,
"step": 31
},
{
"epoch": 0.21,
"grad_norm": 1.6816545538392116,
"learning_rate": 2.8717948717948717e-06,
"loss": 0.4833,
"step": 32
},
{
"epoch": 0.21,
"grad_norm": 1.8708630053016435,
"learning_rate": 2.9615384615384614e-06,
"loss": 0.4298,
"step": 33
},
{
"epoch": 0.22,
"grad_norm": 1.804255585106581,
"learning_rate": 3.051282051282051e-06,
"loss": 0.4966,
"step": 34
},
{
"epoch": 0.22,
"grad_norm": 2.1939960340576983,
"learning_rate": 3.141025641025641e-06,
"loss": 0.5959,
"step": 35
},
{
"epoch": 0.23,
"grad_norm": 2.251434624865691,
"learning_rate": 3.230769230769231e-06,
"loss": 0.5375,
"step": 36
},
{
"epoch": 0.24,
"grad_norm": 1.729304380380877,
"learning_rate": 3.32051282051282e-06,
"loss": 0.454,
"step": 37
},
{
"epoch": 0.24,
"grad_norm": 1.78866126998266,
"learning_rate": 3.4102564102564103e-06,
"loss": 0.4974,
"step": 38
},
{
"epoch": 0.25,
"grad_norm": 1.8070608359204237,
"learning_rate": 3.5e-06,
"loss": 0.5061,
"step": 39
},
{
"epoch": 0.26,
"grad_norm": 1.7633109832646323,
"learning_rate": 3.5897435897435892e-06,
"loss": 0.5171,
"step": 40
},
{
"epoch": 0.26,
"grad_norm": 1.7502966702710157,
"learning_rate": 3.6794871794871797e-06,
"loss": 0.5277,
"step": 41
},
{
"epoch": 0.27,
"grad_norm": 1.7810880679374266,
"learning_rate": 3.769230769230769e-06,
"loss": 0.486,
"step": 42
},
{
"epoch": 0.28,
"grad_norm": 1.7345248362598074,
"learning_rate": 3.858974358974359e-06,
"loss": 0.4403,
"step": 43
},
{
"epoch": 0.28,
"grad_norm": 1.8544499121276545,
"learning_rate": 3.948717948717949e-06,
"loss": 0.4559,
"step": 44
},
{
"epoch": 0.29,
"grad_norm": 1.9281302680981884,
"learning_rate": 4.038461538461538e-06,
"loss": 0.4892,
"step": 45
},
{
"epoch": 0.29,
"grad_norm": 1.823077284882776,
"learning_rate": 4.128205128205128e-06,
"loss": 0.4578,
"step": 46
},
{
"epoch": 0.3,
"grad_norm": 1.7087373941464228,
"learning_rate": 4.217948717948718e-06,
"loss": 0.4312,
"step": 47
},
{
"epoch": 0.31,
"grad_norm": 1.7936679868143406,
"learning_rate": 4.3076923076923076e-06,
"loss": 0.4268,
"step": 48
},
{
"epoch": 0.31,
"grad_norm": 1.7913865129800892,
"learning_rate": 4.397435897435897e-06,
"loss": 0.4978,
"step": 49
},
{
"epoch": 0.32,
"grad_norm": 1.680234009753019,
"learning_rate": 4.487179487179488e-06,
"loss": 0.3909,
"step": 50
},
{
"epoch": 0.33,
"grad_norm": 1.7084054193024387,
"learning_rate": 4.576923076923077e-06,
"loss": 0.4556,
"step": 51
},
{
"epoch": 0.33,
"grad_norm": 2.8283052735253307,
"learning_rate": 4.666666666666666e-06,
"loss": 0.6066,
"step": 52
},
{
"epoch": 0.34,
"grad_norm": 1.7700007607445167,
"learning_rate": 4.756410256410257e-06,
"loss": 0.4966,
"step": 53
},
{
"epoch": 0.35,
"grad_norm": 1.695503292082949,
"learning_rate": 4.846153846153846e-06,
"loss": 0.5229,
"step": 54
},
{
"epoch": 0.35,
"grad_norm": 1.7249370659175542,
"learning_rate": 4.935897435897436e-06,
"loss": 0.4866,
"step": 55
},
{
"epoch": 0.36,
"grad_norm": 1.6670150319109511,
"learning_rate": 5.025641025641026e-06,
"loss": 0.4372,
"step": 56
},
{
"epoch": 0.37,
"grad_norm": 1.9001530129586224,
"learning_rate": 5.115384615384615e-06,
"loss": 0.4961,
"step": 57
},
{
"epoch": 0.37,
"grad_norm": 1.8138821119614343,
"learning_rate": 5.205128205128205e-06,
"loss": 0.4956,
"step": 58
},
{
"epoch": 0.38,
"grad_norm": 1.7702871650778012,
"learning_rate": 5.294871794871795e-06,
"loss": 0.4819,
"step": 59
},
{
"epoch": 0.38,
"grad_norm": 1.6027880830820047,
"learning_rate": 5.384615384615385e-06,
"loss": 0.4691,
"step": 60
},
{
"epoch": 0.39,
"grad_norm": 1.5406345693116157,
"learning_rate": 5.474358974358974e-06,
"loss": 0.3892,
"step": 61
},
{
"epoch": 0.4,
"grad_norm": 1.4960225687619988,
"learning_rate": 5.564102564102564e-06,
"loss": 0.4675,
"step": 62
},
{
"epoch": 0.4,
"grad_norm": 1.4165399860657466,
"learning_rate": 5.653846153846154e-06,
"loss": 0.3659,
"step": 63
},
{
"epoch": 0.41,
"grad_norm": 1.783351454505062,
"learning_rate": 5.743589743589743e-06,
"loss": 0.5025,
"step": 64
},
{
"epoch": 0.42,
"grad_norm": 1.7461313827694103,
"learning_rate": 5.833333333333333e-06,
"loss": 0.4599,
"step": 65
},
{
"epoch": 0.42,
"grad_norm": 1.7177479188522022,
"learning_rate": 5.923076923076923e-06,
"loss": 0.4658,
"step": 66
},
{
"epoch": 0.43,
"grad_norm": 1.7430319738026099,
"learning_rate": 6.0128205128205125e-06,
"loss": 0.5034,
"step": 67
},
{
"epoch": 0.44,
"grad_norm": 1.8694656181485532,
"learning_rate": 6.102564102564102e-06,
"loss": 0.5335,
"step": 68
},
{
"epoch": 0.44,
"grad_norm": 1.5949394110724386,
"learning_rate": 6.192307692307692e-06,
"loss": 0.4341,
"step": 69
},
{
"epoch": 0.45,
"grad_norm": 2.0737191543993223,
"learning_rate": 6.282051282051282e-06,
"loss": 0.5444,
"step": 70
},
{
"epoch": 0.46,
"grad_norm": 1.701318366396795,
"learning_rate": 6.371794871794871e-06,
"loss": 0.4897,
"step": 71
},
{
"epoch": 0.46,
"grad_norm": 1.5564495848821291,
"learning_rate": 6.461538461538462e-06,
"loss": 0.4486,
"step": 72
},
{
"epoch": 0.47,
"grad_norm": 1.7515969674361473,
"learning_rate": 6.5512820512820515e-06,
"loss": 0.4836,
"step": 73
},
{
"epoch": 0.47,
"grad_norm": 1.723669518863802,
"learning_rate": 6.64102564102564e-06,
"loss": 0.4881,
"step": 74
},
{
"epoch": 0.48,
"grad_norm": 1.6232767387422016,
"learning_rate": 6.730769230769231e-06,
"loss": 0.4517,
"step": 75
},
{
"epoch": 0.49,
"grad_norm": 1.6836950386453475,
"learning_rate": 6.8205128205128205e-06,
"loss": 0.453,
"step": 76
},
{
"epoch": 0.49,
"grad_norm": 1.5970207533021141,
"learning_rate": 6.91025641025641e-06,
"loss": 0.4102,
"step": 77
},
{
"epoch": 0.5,
"grad_norm": 1.806038689668669,
"learning_rate": 7e-06,
"loss": 0.4938,
"step": 78
},
{
"epoch": 0.51,
"grad_norm": 1.682873077514678,
"learning_rate": 6.9999649520318915e-06,
"loss": 0.4656,
"step": 79
},
{
"epoch": 0.51,
"grad_norm": 1.7622694283436713,
"learning_rate": 6.999859808829483e-06,
"loss": 0.3833,
"step": 80
},
{
"epoch": 0.52,
"grad_norm": 1.7028611976071402,
"learning_rate": 6.999684572498523e-06,
"loss": 0.5228,
"step": 81
},
{
"epoch": 0.53,
"grad_norm": 1.8078825168256631,
"learning_rate": 6.999439246548541e-06,
"loss": 0.4219,
"step": 82
},
{
"epoch": 0.53,
"grad_norm": 1.5963387300318541,
"learning_rate": 6.999123835892781e-06,
"loss": 0.3838,
"step": 83
},
{
"epoch": 0.54,
"grad_norm": 1.752439994185435,
"learning_rate": 6.998738346848099e-06,
"loss": 0.5353,
"step": 84
},
{
"epoch": 0.54,
"grad_norm": 1.5765967217454646,
"learning_rate": 6.998282787134845e-06,
"loss": 0.4013,
"step": 85
},
{
"epoch": 0.55,
"grad_norm": 1.64099993127281,
"learning_rate": 6.997757165876698e-06,
"loss": 0.5268,
"step": 86
},
{
"epoch": 0.56,
"grad_norm": 1.6629395270016087,
"learning_rate": 6.9971614936004935e-06,
"loss": 0.4364,
"step": 87
},
{
"epoch": 0.56,
"grad_norm": 1.7737143292943345,
"learning_rate": 6.996495782236003e-06,
"loss": 0.445,
"step": 88
},
{
"epoch": 0.57,
"grad_norm": 1.7709235055203036,
"learning_rate": 6.9957600451157e-06,
"loss": 0.4809,
"step": 89
},
{
"epoch": 0.58,
"grad_norm": 1.6973893875173904,
"learning_rate": 6.9949542969744955e-06,
"loss": 0.475,
"step": 90
},
{
"epoch": 0.58,
"grad_norm": 1.6645616366762754,
"learning_rate": 6.9940785539494385e-06,
"loss": 0.4656,
"step": 91
},
{
"epoch": 0.59,
"grad_norm": 1.6726681899225762,
"learning_rate": 6.9931328335793926e-06,
"loss": 0.436,
"step": 92
},
{
"epoch": 0.6,
"grad_norm": 1.7236949440975784,
"learning_rate": 6.992117154804688e-06,
"loss": 0.474,
"step": 93
},
{
"epoch": 0.6,
"grad_norm": 1.6637387868033124,
"learning_rate": 6.991031537966741e-06,
"loss": 0.421,
"step": 94
},
{
"epoch": 0.61,
"grad_norm": 1.815724145981235,
"learning_rate": 6.989876004807644e-06,
"loss": 0.4889,
"step": 95
},
{
"epoch": 0.62,
"grad_norm": 1.65090673754103,
"learning_rate": 6.9886505784697354e-06,
"loss": 0.4313,
"step": 96
},
{
"epoch": 0.62,
"grad_norm": 1.7879715436757848,
"learning_rate": 6.98735528349513e-06,
"loss": 0.5158,
"step": 97
},
{
"epoch": 0.63,
"grad_norm": 1.6116310457967884,
"learning_rate": 6.985990145825233e-06,
"loss": 0.4152,
"step": 98
},
{
"epoch": 0.63,
"grad_norm": 1.7530216353063874,
"learning_rate": 6.984555192800216e-06,
"loss": 0.5415,
"step": 99
},
{
"epoch": 0.64,
"grad_norm": 1.6028754704316188,
"learning_rate": 6.983050453158471e-06,
"loss": 0.4666,
"step": 100
},
{
"epoch": 0.65,
"grad_norm": 1.6554540478959205,
"learning_rate": 6.981475957036039e-06,
"loss": 0.4338,
"step": 101
},
{
"epoch": 0.65,
"grad_norm": 1.7244082720999658,
"learning_rate": 6.979831735965997e-06,
"loss": 0.3997,
"step": 102
},
{
"epoch": 0.66,
"grad_norm": 1.8613328382698524,
"learning_rate": 6.9781178228778385e-06,
"loss": 0.4822,
"step": 103
},
{
"epoch": 0.67,
"grad_norm": 1.7472563980348217,
"learning_rate": 6.9763342520968e-06,
"loss": 0.531,
"step": 104
},
{
"epoch": 0.67,
"grad_norm": 1.7296385478020857,
"learning_rate": 6.974481059343188e-06,
"loss": 0.4556,
"step": 105
},
{
"epoch": 0.68,
"grad_norm": 1.7421477412265767,
"learning_rate": 6.972558281731655e-06,
"loss": 0.4739,
"step": 106
},
{
"epoch": 0.69,
"grad_norm": 1.8252883807781282,
"learning_rate": 6.970565957770456e-06,
"loss": 0.4603,
"step": 107
},
{
"epoch": 0.69,
"grad_norm": 1.4893874441098909,
"learning_rate": 6.96850412736068e-06,
"loss": 0.4074,
"step": 108
},
{
"epoch": 0.7,
"grad_norm": 1.706254344170796,
"learning_rate": 6.9663728317954505e-06,
"loss": 0.4931,
"step": 109
},
{
"epoch": 0.71,
"grad_norm": 1.7729579454846884,
"learning_rate": 6.9641721137591e-06,
"loss": 0.5236,
"step": 110
},
{
"epoch": 0.71,
"grad_norm": 1.673651236600404,
"learning_rate": 6.961902017326311e-06,
"loss": 0.4678,
"step": 111
},
{
"epoch": 0.72,
"grad_norm": 1.6483821733028223,
"learning_rate": 6.959562587961235e-06,
"loss": 0.4539,
"step": 112
},
{
"epoch": 0.72,
"grad_norm": 1.8334930429778364,
"learning_rate": 6.9571538725165855e-06,
"loss": 0.4598,
"step": 113
},
{
"epoch": 0.73,
"grad_norm": 1.6259182099235296,
"learning_rate": 6.9546759192326944e-06,
"loss": 0.4618,
"step": 114
},
{
"epoch": 0.74,
"grad_norm": 1.586763484875926,
"learning_rate": 6.95212877773655e-06,
"loss": 0.3916,
"step": 115
},
{
"epoch": 0.74,
"grad_norm": 1.729278408278385,
"learning_rate": 6.949512499040799e-06,
"loss": 0.443,
"step": 116
},
{
"epoch": 0.75,
"grad_norm": 1.7287260985135913,
"learning_rate": 6.946827135542729e-06,
"loss": 0.4058,
"step": 117
},
{
"epoch": 0.76,
"grad_norm": 1.5105369132311874,
"learning_rate": 6.944072741023215e-06,
"loss": 0.3816,
"step": 118
},
{
"epoch": 0.76,
"grad_norm": 1.7827477765014352,
"learning_rate": 6.941249370645649e-06,
"loss": 0.4411,
"step": 119
},
{
"epoch": 0.77,
"grad_norm": 1.6865474969699326,
"learning_rate": 6.938357080954826e-06,
"loss": 0.4536,
"step": 120
},
{
"epoch": 0.78,
"grad_norm": 1.6697703500975725,
"learning_rate": 6.935395929875821e-06,
"loss": 0.4773,
"step": 121
},
{
"epoch": 0.78,
"grad_norm": 1.846095347111975,
"learning_rate": 6.93236597671282e-06,
"loss": 0.5273,
"step": 122
},
{
"epoch": 0.79,
"grad_norm": 1.5599554260333042,
"learning_rate": 6.929267282147936e-06,
"loss": 0.4108,
"step": 123
},
{
"epoch": 0.79,
"grad_norm": 1.5737374777092443,
"learning_rate": 6.9260999082400014e-06,
"loss": 0.4233,
"step": 124
},
{
"epoch": 0.8,
"grad_norm": 1.7607156921341576,
"learning_rate": 6.922863918423311e-06,
"loss": 0.4391,
"step": 125
},
{
"epoch": 0.81,
"grad_norm": 1.8081142800994836,
"learning_rate": 6.91955937750636e-06,
"loss": 0.5029,
"step": 126
},
{
"epoch": 0.81,
"grad_norm": 1.6888032055738713,
"learning_rate": 6.916186351670546e-06,
"loss": 0.442,
"step": 127
},
{
"epoch": 0.82,
"grad_norm": 1.59484869751713,
"learning_rate": 6.912744908468841e-06,
"loss": 0.4274,
"step": 128
},
{
"epoch": 0.83,
"grad_norm": 1.5838918315915649,
"learning_rate": 6.909235116824441e-06,
"loss": 0.4862,
"step": 129
},
{
"epoch": 0.83,
"grad_norm": 1.5793657946026944,
"learning_rate": 6.905657047029383e-06,
"loss": 0.4122,
"step": 130
},
{
"epoch": 0.84,
"grad_norm": 1.8119234908825441,
"learning_rate": 6.90201077074314e-06,
"loss": 0.5385,
"step": 131
},
{
"epoch": 0.85,
"grad_norm": 1.7221288311251346,
"learning_rate": 6.898296360991182e-06,
"loss": 0.4986,
"step": 132
},
{
"epoch": 0.85,
"grad_norm": 1.6177348169551482,
"learning_rate": 6.894513892163519e-06,
"loss": 0.4351,
"step": 133
},
{
"epoch": 0.86,
"grad_norm": 1.8266262713227248,
"learning_rate": 6.890663440013204e-06,
"loss": 0.4635,
"step": 134
},
{
"epoch": 0.87,
"grad_norm": 1.5762526247077566,
"learning_rate": 6.886745081654823e-06,
"loss": 0.4404,
"step": 135
},
{
"epoch": 0.87,
"grad_norm": 1.7127197602668496,
"learning_rate": 6.882758895562948e-06,
"loss": 0.4798,
"step": 136
},
{
"epoch": 0.88,
"grad_norm": 1.7272250386199177,
"learning_rate": 6.8787049615705635e-06,
"loss": 0.4491,
"step": 137
},
{
"epoch": 0.88,
"grad_norm": 1.7401938465087832,
"learning_rate": 6.8745833608674685e-06,
"loss": 0.513,
"step": 138
},
{
"epoch": 0.89,
"grad_norm": 1.476494218502608,
"learning_rate": 6.870394175998651e-06,
"loss": 0.4126,
"step": 139
},
{
"epoch": 0.9,
"grad_norm": 1.6735073474825053,
"learning_rate": 6.866137490862636e-06,
"loss": 0.4784,
"step": 140
},
{
"epoch": 0.9,
"grad_norm": 1.61528219577023,
"learning_rate": 6.861813390709803e-06,
"loss": 0.3993,
"step": 141
},
{
"epoch": 0.91,
"grad_norm": 1.7160841750089657,
"learning_rate": 6.857421962140681e-06,
"loss": 0.437,
"step": 142
},
{
"epoch": 0.92,
"grad_norm": 1.6003971415765827,
"learning_rate": 6.852963293104211e-06,
"loss": 0.4234,
"step": 143
},
{
"epoch": 0.92,
"grad_norm": 1.673821864143384,
"learning_rate": 6.848437472895989e-06,
"loss": 0.36,
"step": 144
},
{
"epoch": 0.93,
"grad_norm": 1.5584858029998052,
"learning_rate": 6.84384459215647e-06,
"loss": 0.3831,
"step": 145
},
{
"epoch": 0.94,
"grad_norm": 1.8434118860530437,
"learning_rate": 6.839184742869166e-06,
"loss": 0.481,
"step": 146
},
{
"epoch": 0.94,
"grad_norm": 1.735764854231206,
"learning_rate": 6.8344580183587866e-06,
"loss": 0.4604,
"step": 147
},
{
"epoch": 0.95,
"grad_norm": 1.5746871313601694,
"learning_rate": 6.829664513289387e-06,
"loss": 0.4481,
"step": 148
},
{
"epoch": 0.96,
"grad_norm": 1.554650404317531,
"learning_rate": 6.824804323662456e-06,
"loss": 0.4246,
"step": 149
},
{
"epoch": 0.96,
"grad_norm": 1.776675138837105,
"learning_rate": 6.8198775468150085e-06,
"loss": 0.505,
"step": 150
},
{
"epoch": 0.97,
"grad_norm": 1.6556936569780774,
"learning_rate": 6.814884281417627e-06,
"loss": 0.4684,
"step": 151
},
{
"epoch": 0.97,
"grad_norm": 1.6529466478362556,
"learning_rate": 6.8098246274724835e-06,
"loss": 0.4179,
"step": 152
},
{
"epoch": 0.98,
"grad_norm": 1.5675006143134382,
"learning_rate": 6.8046986863113455e-06,
"loss": 0.3936,
"step": 153
},
{
"epoch": 0.99,
"grad_norm": 1.543693186637755,
"learning_rate": 6.7995065605935405e-06,
"loss": 0.4343,
"step": 154
},
{
"epoch": 0.99,
"grad_norm": 1.5960852724486323,
"learning_rate": 6.7942483543039e-06,
"loss": 0.4028,
"step": 155
},
{
"epoch": 1.0,
"grad_norm": 1.6707113583767172,
"learning_rate": 6.788924172750679e-06,
"loss": 0.4456,
"step": 156
},
{
"epoch": 1.01,
"grad_norm": 1.592042298214689,
"learning_rate": 6.783534122563447e-06,
"loss": 0.3896,
"step": 157
},
{
"epoch": 1.01,
"grad_norm": 1.594380804915327,
"learning_rate": 6.7780783116909495e-06,
"loss": 0.4269,
"step": 158
},
{
"epoch": 1.02,
"grad_norm": 1.425175264752805,
"learning_rate": 6.772556849398952e-06,
"loss": 0.4136,
"step": 159
},
{
"epoch": 1.03,
"grad_norm": 1.580030153001224,
"learning_rate": 6.7669698462680434e-06,
"loss": 0.4534,
"step": 160
},
{
"epoch": 1.03,
"grad_norm": 1.656736559657467,
"learning_rate": 6.761317414191428e-06,
"loss": 0.4262,
"step": 161
},
{
"epoch": 1.04,
"grad_norm": 1.5516325617950246,
"learning_rate": 6.755599666372685e-06,
"loss": 0.3525,
"step": 162
},
{
"epoch": 1.04,
"grad_norm": 1.5964299616457083,
"learning_rate": 6.749816717323493e-06,
"loss": 0.3582,
"step": 163
},
{
"epoch": 1.05,
"grad_norm": 1.5572727726198186,
"learning_rate": 6.743968682861346e-06,
"loss": 0.4277,
"step": 164
},
{
"epoch": 1.06,
"grad_norm": 1.59926615487551,
"learning_rate": 6.738055680107233e-06,
"loss": 0.3878,
"step": 165
},
{
"epoch": 1.06,
"grad_norm": 1.7037332135435532,
"learning_rate": 6.7320778274832836e-06,
"loss": 0.4137,
"step": 166
},
{
"epoch": 1.07,
"grad_norm": 1.5513259799733465,
"learning_rate": 6.726035244710406e-06,
"loss": 0.4037,
"step": 167
},
{
"epoch": 1.08,
"grad_norm": 1.6898460188300195,
"learning_rate": 6.7199280528058844e-06,
"loss": 0.3961,
"step": 168
},
{
"epoch": 1.08,
"grad_norm": 1.5729523035551953,
"learning_rate": 6.713756374080959e-06,
"loss": 0.3434,
"step": 169
},
{
"epoch": 1.09,
"grad_norm": 1.544572572819997,
"learning_rate": 6.70752033213837e-06,
"loss": 0.3416,
"step": 170
},
{
"epoch": 1.1,
"grad_norm": 1.7321648255912703,
"learning_rate": 6.7012200518698904e-06,
"loss": 0.3349,
"step": 171
},
{
"epoch": 1.1,
"grad_norm": 1.5902383819866788,
"learning_rate": 6.6948556594538185e-06,
"loss": 0.3897,
"step": 172
},
{
"epoch": 1.11,
"grad_norm": 1.578037443536385,
"learning_rate": 6.688427282352449e-06,
"loss": 0.374,
"step": 173
},
{
"epoch": 1.12,
"grad_norm": 1.637206599992538,
"learning_rate": 6.681935049309533e-06,
"loss": 0.356,
"step": 174
},
{
"epoch": 1.12,
"grad_norm": 1.6426011449900024,
"learning_rate": 6.6753790903476814e-06,
"loss": 0.3204,
"step": 175
},
{
"epoch": 1.13,
"grad_norm": 1.7335754223028719,
"learning_rate": 6.668759536765778e-06,
"loss": 0.3447,
"step": 176
},
{
"epoch": 1.13,
"grad_norm": 1.8716868794333945,
"learning_rate": 6.6620765211363376e-06,
"loss": 0.4708,
"step": 177
},
{
"epoch": 1.14,
"grad_norm": 2.0355113763091306,
"learning_rate": 6.655330177302857e-06,
"loss": 0.4357,
"step": 178
},
{
"epoch": 1.15,
"grad_norm": 1.6067153937975995,
"learning_rate": 6.64852064037713e-06,
"loss": 0.3237,
"step": 179
},
{
"epoch": 1.15,
"grad_norm": 1.566156260471038,
"learning_rate": 6.6416480467365494e-06,
"loss": 0.3271,
"step": 180
},
{
"epoch": 1.16,
"grad_norm": 1.415525082706651,
"learning_rate": 6.634712534021367e-06,
"loss": 0.3125,
"step": 181
},
{
"epoch": 1.17,
"grad_norm": 1.5322544502916775,
"learning_rate": 6.627714241131943e-06,
"loss": 0.2987,
"step": 182
},
{
"epoch": 1.17,
"grad_norm": 1.6559751396978282,
"learning_rate": 6.62065330822596e-06,
"loss": 0.331,
"step": 183
},
{
"epoch": 1.18,
"grad_norm": 1.6362651282972798,
"learning_rate": 6.613529876715619e-06,
"loss": 0.3522,
"step": 184
},
{
"epoch": 1.19,
"grad_norm": 1.7432266355020243,
"learning_rate": 6.606344089264805e-06,
"loss": 0.3721,
"step": 185
},
{
"epoch": 1.19,
"grad_norm": 1.6595653155086751,
"learning_rate": 6.599096089786234e-06,
"loss": 0.3272,
"step": 186
},
{
"epoch": 1.2,
"grad_norm": 1.7169078900169732,
"learning_rate": 6.591786023438565e-06,
"loss": 0.3205,
"step": 187
},
{
"epoch": 1.21,
"grad_norm": 1.5217863220075283,
"learning_rate": 6.5844140366234956e-06,
"loss": 0.3007,
"step": 188
},
{
"epoch": 1.21,
"grad_norm": 1.5386588460515103,
"learning_rate": 6.576980276982832e-06,
"loss": 0.2683,
"step": 189
},
{
"epoch": 1.22,
"grad_norm": 1.5586980831749289,
"learning_rate": 6.569484893395527e-06,
"loss": 0.3109,
"step": 190
},
{
"epoch": 1.22,
"grad_norm": 1.835334084851163,
"learning_rate": 6.5619280359747045e-06,
"loss": 0.3659,
"step": 191
},
{
"epoch": 1.23,
"grad_norm": 1.7303049780029398,
"learning_rate": 6.55430985606465e-06,
"loss": 0.3169,
"step": 192
},
{
"epoch": 1.24,
"grad_norm": 1.573272600381468,
"learning_rate": 6.546630506237778e-06,
"loss": 0.2737,
"step": 193
},
{
"epoch": 1.24,
"grad_norm": 1.5800960246726725,
"learning_rate": 6.538890140291578e-06,
"loss": 0.2962,
"step": 194
},
{
"epoch": 1.25,
"grad_norm": 1.5942214142753601,
"learning_rate": 6.531088913245536e-06,
"loss": 0.2912,
"step": 195
},
{
"epoch": 1.26,
"grad_norm": 1.6047857651904212,
"learning_rate": 6.5232269813380254e-06,
"loss": 0.3033,
"step": 196
},
{
"epoch": 1.26,
"grad_norm": 1.6641616618876198,
"learning_rate": 6.5153045020231855e-06,
"loss": 0.3071,
"step": 197
},
{
"epoch": 1.27,
"grad_norm": 1.684992093107794,
"learning_rate": 6.507321633967758e-06,
"loss": 0.2783,
"step": 198
},
{
"epoch": 1.28,
"grad_norm": 1.6325774194230926,
"learning_rate": 6.499278537047919e-06,
"loss": 0.2527,
"step": 199
},
{
"epoch": 1.28,
"grad_norm": 1.6121399506538092,
"learning_rate": 6.49117537234607e-06,
"loss": 0.2459,
"step": 200
},
{
"epoch": 1.29,
"grad_norm": 1.7019340316973153,
"learning_rate": 6.483012302147617e-06,
"loss": 0.2639,
"step": 201
},
{
"epoch": 1.29,
"grad_norm": 1.5471416195498167,
"learning_rate": 6.474789489937715e-06,
"loss": 0.2507,
"step": 202
},
{
"epoch": 1.3,
"grad_norm": 1.5639059610283246,
"learning_rate": 6.4665071003979985e-06,
"loss": 0.2227,
"step": 203
},
{
"epoch": 1.31,
"grad_norm": 1.6480830867630272,
"learning_rate": 6.4581652994032816e-06,
"loss": 0.2199,
"step": 204
},
{
"epoch": 1.31,
"grad_norm": 1.6418221792531331,
"learning_rate": 6.449764254018236e-06,
"loss": 0.2676,
"step": 205
},
{
"epoch": 1.32,
"grad_norm": 1.557597222846848,
"learning_rate": 6.441304132494045e-06,
"loss": 0.2057,
"step": 206
},
{
"epoch": 1.33,
"grad_norm": 1.6032321620665173,
"learning_rate": 6.432785104265034e-06,
"loss": 0.2325,
"step": 207
},
{
"epoch": 1.33,
"grad_norm": 2.068457437087488,
"learning_rate": 6.424207339945278e-06,
"loss": 0.3075,
"step": 208
},
{
"epoch": 1.34,
"grad_norm": 1.6387109523391752,
"learning_rate": 6.415571011325181e-06,
"loss": 0.2638,
"step": 209
},
{
"epoch": 1.35,
"grad_norm": 1.6478934557021592,
"learning_rate": 6.406876291368041e-06,
"loss": 0.2829,
"step": 210
},
{
"epoch": 1.35,
"grad_norm": 1.6015206620029365,
"learning_rate": 6.3981233542065824e-06,
"loss": 0.2542,
"step": 211
},
{
"epoch": 1.36,
"grad_norm": 1.6246765840899604,
"learning_rate": 6.3893123751394695e-06,
"loss": 0.2084,
"step": 212
},
{
"epoch": 1.37,
"grad_norm": 1.6287901735275234,
"learning_rate": 6.380443530627797e-06,
"loss": 0.2424,
"step": 213
},
{
"epoch": 1.37,
"grad_norm": 1.6578723488093603,
"learning_rate": 6.371516998291552e-06,
"loss": 0.2458,
"step": 214
},
{
"epoch": 1.38,
"grad_norm": 1.6660041316351297,
"learning_rate": 6.3625329569060595e-06,
"loss": 0.2427,
"step": 215
},
{
"epoch": 1.38,
"grad_norm": 1.578903354700284,
"learning_rate": 6.3534915863984045e-06,
"loss": 0.248,
"step": 216
},
{
"epoch": 1.39,
"grad_norm": 1.480820395340969,
"learning_rate": 6.344393067843825e-06,
"loss": 0.1903,
"step": 217
},
{
"epoch": 1.4,
"grad_norm": 1.5351995236858405,
"learning_rate": 6.335237583462083e-06,
"loss": 0.2444,
"step": 218
},
{
"epoch": 1.4,
"grad_norm": 1.4430941536344153,
"learning_rate": 6.326025316613824e-06,
"loss": 0.1888,
"step": 219
},
{
"epoch": 1.41,
"grad_norm": 1.705699712353775,
"learning_rate": 6.3167564517968944e-06,
"loss": 0.2381,
"step": 220
},
{
"epoch": 1.42,
"grad_norm": 1.5317499613408436,
"learning_rate": 6.307431174642653e-06,
"loss": 0.2012,
"step": 221
},
{
"epoch": 1.42,
"grad_norm": 1.6439194933028012,
"learning_rate": 6.2980496719122544e-06,
"loss": 0.2213,
"step": 222
},
{
"epoch": 1.43,
"grad_norm": 1.5931209129358592,
"learning_rate": 6.288612131492901e-06,
"loss": 0.2418,
"step": 223
},
{
"epoch": 1.44,
"grad_norm": 1.6370343275461072,
"learning_rate": 6.279118742394089e-06,
"loss": 0.2256,
"step": 224
},
{
"epoch": 1.44,
"grad_norm": 1.4744567429276645,
"learning_rate": 6.2695696947438165e-06,
"loss": 0.2009,
"step": 225
},
{
"epoch": 1.45,
"grad_norm": 1.8325833367980933,
"learning_rate": 6.25996517978478e-06,
"loss": 0.2345,
"step": 226
},
{
"epoch": 1.46,
"grad_norm": 1.6355207967953467,
"learning_rate": 6.2503053898705416e-06,
"loss": 0.2232,
"step": 227
},
{
"epoch": 1.46,
"grad_norm": 1.5592250115768092,
"learning_rate": 6.2405905184616776e-06,
"loss": 0.2144,
"step": 228
},
{
"epoch": 1.47,
"grad_norm": 1.5359450179179217,
"learning_rate": 6.230820760121904e-06,
"loss": 0.2025,
"step": 229
},
{
"epoch": 1.47,
"grad_norm": 1.6594758906536755,
"learning_rate": 6.220996310514181e-06,
"loss": 0.2248,
"step": 230
},
{
"epoch": 1.48,
"grad_norm": 1.614953425314903,
"learning_rate": 6.21111736639679e-06,
"loss": 0.2072,
"step": 231
},
{
"epoch": 1.49,
"grad_norm": 1.5868322471592755,
"learning_rate": 6.201184125619403e-06,
"loss": 0.1954,
"step": 232
},
{
"epoch": 1.49,
"grad_norm": 1.5846521300517467,
"learning_rate": 6.191196787119104e-06,
"loss": 0.1872,
"step": 233
},
{
"epoch": 1.5,
"grad_norm": 1.5858305578529914,
"learning_rate": 6.181155550916423e-06,
"loss": 0.2173,
"step": 234
},
{
"epoch": 1.51,
"grad_norm": 1.5685831511245525,
"learning_rate": 6.171060618111317e-06,
"loss": 0.2035,
"step": 235
},
{
"epoch": 1.51,
"grad_norm": 1.416118500599548,
"learning_rate": 6.160912190879146e-06,
"loss": 0.1546,
"step": 236
},
{
"epoch": 1.52,
"grad_norm": 1.6574554300523399,
"learning_rate": 6.15071047246663e-06,
"loss": 0.2404,
"step": 237
},
{
"epoch": 1.53,
"grad_norm": 1.555316260098642,
"learning_rate": 6.140455667187765e-06,
"loss": 0.1578,
"step": 238
},
{
"epoch": 1.53,
"grad_norm": 1.4540496982385553,
"learning_rate": 6.13014798041975e-06,
"loss": 0.1595,
"step": 239
},
{
"epoch": 1.54,
"grad_norm": 1.6687470479114859,
"learning_rate": 6.119787618598854e-06,
"loss": 0.2505,
"step": 240
},
{
"epoch": 1.54,
"grad_norm": 1.5334245760102851,
"learning_rate": 6.109374789216296e-06,
"loss": 0.1838,
"step": 241
},
{
"epoch": 1.55,
"grad_norm": 1.542810949224468,
"learning_rate": 6.098909700814082e-06,
"loss": 0.241,
"step": 242
},
{
"epoch": 1.56,
"grad_norm": 1.72874768735861,
"learning_rate": 6.08839256298083e-06,
"loss": 0.2054,
"step": 243
},
{
"epoch": 1.56,
"grad_norm": 1.5566017473024847,
"learning_rate": 6.077823586347579e-06,
"loss": 0.1827,
"step": 244
},
{
"epoch": 1.57,
"grad_norm": 1.5454330143421287,
"learning_rate": 6.06720298258356e-06,
"loss": 0.1951,
"step": 245
},
{
"epoch": 1.58,
"grad_norm": 1.5699376808386447,
"learning_rate": 6.056530964391961e-06,
"loss": 0.217,
"step": 246
},
{
"epoch": 1.58,
"grad_norm": 1.4790843551806858,
"learning_rate": 6.0458077455056704e-06,
"loss": 0.2034,
"step": 247
},
{
"epoch": 1.59,
"grad_norm": 1.5292597992204815,
"learning_rate": 6.035033540682993e-06,
"loss": 0.1917,
"step": 248
},
{
"epoch": 1.6,
"grad_norm": 1.5536070824489197,
"learning_rate": 6.024208565703351e-06,
"loss": 0.2102,
"step": 249
},
{
"epoch": 1.6,
"grad_norm": 1.519765331872237,
"learning_rate": 6.013333037362959e-06,
"loss": 0.1775,
"step": 250
},
{
"epoch": 1.61,
"grad_norm": 1.664352545830077,
"learning_rate": 6.002407173470486e-06,
"loss": 0.2252,
"step": 251
},
{
"epoch": 1.62,
"grad_norm": 1.6289900634763281,
"learning_rate": 5.991431192842692e-06,
"loss": 0.1929,
"step": 252
},
{
"epoch": 1.62,
"grad_norm": 1.650116318598881,
"learning_rate": 5.980405315300045e-06,
"loss": 0.2363,
"step": 253
},
{
"epoch": 1.63,
"grad_norm": 1.55668794790212,
"learning_rate": 5.969329761662319e-06,
"loss": 0.19,
"step": 254
},
{
"epoch": 1.63,
"grad_norm": 1.6677234649563633,
"learning_rate": 5.9582047537441716e-06,
"loss": 0.2332,
"step": 255
},
{
"epoch": 1.64,
"grad_norm": 1.5661036857828248,
"learning_rate": 5.9470305143507e-06,
"loss": 0.2237,
"step": 256
},
{
"epoch": 1.65,
"grad_norm": 1.4888010043108841,
"learning_rate": 5.9358072672729845e-06,
"loss": 0.2087,
"step": 257
},
{
"epoch": 1.65,
"grad_norm": 1.5244864099257356,
"learning_rate": 5.924535237283598e-06,
"loss": 0.1658,
"step": 258
},
{
"epoch": 1.66,
"grad_norm": 1.6137026161987666,
"learning_rate": 5.913214650132112e-06,
"loss": 0.1901,
"step": 259
},
{
"epoch": 1.67,
"grad_norm": 1.6772522686974038,
"learning_rate": 5.901845732540568e-06,
"loss": 0.258,
"step": 260
},
{
"epoch": 1.67,
"grad_norm": 1.5097022747012565,
"learning_rate": 5.8904287121989455e-06,
"loss": 0.1804,
"step": 261
},
{
"epoch": 1.68,
"grad_norm": 1.695764383503543,
"learning_rate": 5.878963817760597e-06,
"loss": 0.2051,
"step": 262
},
{
"epoch": 1.69,
"grad_norm": 1.6150072622733846,
"learning_rate": 5.867451278837666e-06,
"loss": 0.1778,
"step": 263
},
{
"epoch": 1.69,
"grad_norm": 1.4156360311668714,
"learning_rate": 5.855891325996495e-06,
"loss": 0.1941,
"step": 264
},
{
"epoch": 1.7,
"grad_norm": 1.626327498057916,
"learning_rate": 5.8442841907530035e-06,
"loss": 0.2307,
"step": 265
},
{
"epoch": 1.71,
"grad_norm": 2.0421635487814385,
"learning_rate": 5.83263010556805e-06,
"loss": 0.2468,
"step": 266
},
{
"epoch": 1.71,
"grad_norm": 1.670954185746529,
"learning_rate": 5.820929303842783e-06,
"loss": 0.2244,
"step": 267
},
{
"epoch": 1.72,
"grad_norm": 1.6749192214251005,
"learning_rate": 5.809182019913959e-06,
"loss": 0.2079,
"step": 268
},
{
"epoch": 1.72,
"grad_norm": 1.567908091990771,
"learning_rate": 5.797388489049253e-06,
"loss": 0.2012,
"step": 269
},
{
"epoch": 1.73,
"grad_norm": 1.558652423776024,
"learning_rate": 5.785548947442547e-06,
"loss": 0.2136,
"step": 270
},
{
"epoch": 1.74,
"grad_norm": 1.5035278794955325,
"learning_rate": 5.7736636322092016e-06,
"loss": 0.1752,
"step": 271
},
{
"epoch": 1.74,
"grad_norm": 1.6085107962690712,
"learning_rate": 5.7617327813813e-06,
"loss": 0.1841,
"step": 272
},
{
"epoch": 1.75,
"grad_norm": 1.5748033314990677,
"learning_rate": 5.749756633902887e-06,
"loss": 0.1564,
"step": 273
},
{
"epoch": 1.76,
"grad_norm": 1.4911389890937712,
"learning_rate": 5.7377354296251855e-06,
"loss": 0.1852,
"step": 274
},
{
"epoch": 1.76,
"grad_norm": 1.6116400263763508,
"learning_rate": 5.725669409301782e-06,
"loss": 0.1648,
"step": 275
},
{
"epoch": 1.77,
"grad_norm": 1.6655723963052835,
"learning_rate": 5.71355881458382e-06,
"loss": 0.2038,
"step": 276
},
{
"epoch": 1.78,
"grad_norm": 1.6292883296504315,
"learning_rate": 5.701403888015149e-06,
"loss": 0.2151,
"step": 277
},
{
"epoch": 1.78,
"grad_norm": 1.7652594110330915,
"learning_rate": 5.689204873027471e-06,
"loss": 0.2306,
"step": 278
},
{
"epoch": 1.79,
"grad_norm": 1.5674611629004112,
"learning_rate": 5.676962013935464e-06,
"loss": 0.1986,
"step": 279
},
{
"epoch": 1.79,
"grad_norm": 1.5432658298998765,
"learning_rate": 5.664675555931892e-06,
"loss": 0.1961,
"step": 280
},
{
"epoch": 1.8,
"grad_norm": 1.7003413063407276,
"learning_rate": 5.652345745082691e-06,
"loss": 0.1903,
"step": 281
},
{
"epoch": 1.81,
"grad_norm": 1.7383657441416522,
"learning_rate": 5.639972828322043e-06,
"loss": 0.2251,
"step": 282
},
{
"epoch": 1.81,
"grad_norm": 1.6827488344063495,
"learning_rate": 5.627557053447427e-06,
"loss": 0.1928,
"step": 283
},
{
"epoch": 1.82,
"grad_norm": 1.516962697761087,
"learning_rate": 5.615098669114664e-06,
"loss": 0.1969,
"step": 284
},
{
"epoch": 1.83,
"grad_norm": 1.5820218859618687,
"learning_rate": 5.6025979248329265e-06,
"loss": 0.2507,
"step": 285
},
{
"epoch": 1.83,
"grad_norm": 1.4664214189683091,
"learning_rate": 5.590055070959752e-06,
"loss": 0.1823,
"step": 286
},
{
"epoch": 1.84,
"grad_norm": 1.6412853370022484,
"learning_rate": 5.577470358696021e-06,
"loss": 0.2569,
"step": 287
},
{
"epoch": 1.85,
"grad_norm": 1.5791518887916571,
"learning_rate": 5.564844040080931e-06,
"loss": 0.2248,
"step": 288
},
{
"epoch": 1.85,
"grad_norm": 1.5129411232044963,
"learning_rate": 5.5521763679869445e-06,
"loss": 0.2014,
"step": 289
},
{
"epoch": 1.86,
"grad_norm": 1.6335198834860136,
"learning_rate": 5.53946759611473e-06,
"loss": 0.1881,
"step": 290
},
{
"epoch": 1.87,
"grad_norm": 1.4274726284837593,
"learning_rate": 5.526717978988076e-06,
"loss": 0.2049,
"step": 291
},
{
"epoch": 1.87,
"grad_norm": 1.766839389365021,
"learning_rate": 5.513927771948798e-06,
"loss": 0.2144,
"step": 292
},
{
"epoch": 1.88,
"grad_norm": 1.555081227625382,
"learning_rate": 5.5010972311516184e-06,
"loss": 0.1938,
"step": 293
},
{
"epoch": 1.88,
"grad_norm": 1.657103095692518,
"learning_rate": 5.488226613559045e-06,
"loss": 0.2392,
"step": 294
},
{
"epoch": 1.89,
"grad_norm": 1.4598629682187232,
"learning_rate": 5.475316176936217e-06,
"loss": 0.2095,
"step": 295
},
{
"epoch": 1.9,
"grad_norm": 1.5819596345489928,
"learning_rate": 5.462366179845746e-06,
"loss": 0.2112,
"step": 296
},
{
"epoch": 1.9,
"grad_norm": 1.6175196309898452,
"learning_rate": 5.449376881642537e-06,
"loss": 0.1802,
"step": 297
},
{
"epoch": 1.91,
"grad_norm": 1.5465731746504927,
"learning_rate": 5.436348542468598e-06,
"loss": 0.1841,
"step": 298
},
{
"epoch": 1.92,
"grad_norm": 1.572217945594608,
"learning_rate": 5.423281423247821e-06,
"loss": 0.1845,
"step": 299
},
{
"epoch": 1.92,
"grad_norm": 1.4536353215576343,
"learning_rate": 5.4101757856807655e-06,
"loss": 0.1327,
"step": 300
},
{
"epoch": 1.93,
"grad_norm": 1.4814309250755353,
"learning_rate": 5.397031892239414e-06,
"loss": 0.1659,
"step": 301
},
{
"epoch": 1.94,
"grad_norm": 1.7529052634887694,
"learning_rate": 5.383850006161913e-06,
"loss": 0.1938,
"step": 302
},
{
"epoch": 1.94,
"grad_norm": 1.6512796709873832,
"learning_rate": 5.370630391447303e-06,
"loss": 0.1961,
"step": 303
},
{
"epoch": 1.95,
"grad_norm": 1.7490980362206077,
"learning_rate": 5.357373312850236e-06,
"loss": 0.2206,
"step": 304
},
{
"epoch": 1.96,
"grad_norm": 1.5498728220026698,
"learning_rate": 5.3440790358756615e-06,
"loss": 0.2085,
"step": 305
},
{
"epoch": 1.96,
"grad_norm": 1.7873875096307177,
"learning_rate": 5.330747826773522e-06,
"loss": 0.2229,
"step": 306
},
{
"epoch": 1.97,
"grad_norm": 1.6231900940290174,
"learning_rate": 5.317379952533411e-06,
"loss": 0.2133,
"step": 307
},
{
"epoch": 1.97,
"grad_norm": 1.6076562253029956,
"learning_rate": 5.303975680879232e-06,
"loss": 0.189,
"step": 308
},
{
"epoch": 1.98,
"grad_norm": 1.5127508023725418,
"learning_rate": 5.290535280263835e-06,
"loss": 0.179,
"step": 309
},
{
"epoch": 1.99,
"grad_norm": 1.5724862709076817,
"learning_rate": 5.277059019863637e-06,
"loss": 0.1939,
"step": 310
},
{
"epoch": 1.99,
"grad_norm": 1.4830237205662076,
"learning_rate": 5.263547169573235e-06,
"loss": 0.1816,
"step": 311
},
{
"epoch": 2.0,
"grad_norm": 1.5159324470099858,
"learning_rate": 5.25e-06,
"loss": 0.1886,
"step": 312
},
{
"epoch": 2.01,
"grad_norm": 1.555949366187299,
"learning_rate": 5.236417782458656e-06,
"loss": 0.1648,
"step": 313
},
{
"epoch": 2.01,
"grad_norm": 1.4779915210345689,
"learning_rate": 5.222800788965847e-06,
"loss": 0.1949,
"step": 314
},
{
"epoch": 2.02,
"grad_norm": 1.3981863685885323,
"learning_rate": 5.2091492922346894e-06,
"loss": 0.201,
"step": 315
},
{
"epoch": 2.03,
"grad_norm": 1.4384702865295196,
"learning_rate": 5.195463565669309e-06,
"loss": 0.1997,
"step": 316
},
{
"epoch": 2.03,
"grad_norm": 1.417972528504501,
"learning_rate": 5.18174388335937e-06,
"loss": 0.1696,
"step": 317
},
{
"epoch": 2.04,
"grad_norm": 1.42232835008326,
"learning_rate": 5.167990520074577e-06,
"loss": 0.1399,
"step": 318
},
{
"epoch": 2.04,
"grad_norm": 1.4639848529175286,
"learning_rate": 5.154203751259183e-06,
"loss": 0.1462,
"step": 319
},
{
"epoch": 2.05,
"grad_norm": 1.5626053893272325,
"learning_rate": 5.140383853026463e-06,
"loss": 0.1969,
"step": 320
},
{
"epoch": 2.06,
"grad_norm": 1.4260156641561081,
"learning_rate": 5.12653110215319e-06,
"loss": 0.1604,
"step": 321
},
{
"epoch": 2.06,
"grad_norm": 1.4958379917355866,
"learning_rate": 5.11264577607409e-06,
"loss": 0.1677,
"step": 322
},
{
"epoch": 2.07,
"grad_norm": 1.5033492694128057,
"learning_rate": 5.098728152876287e-06,
"loss": 0.1747,
"step": 323
},
{
"epoch": 2.08,
"grad_norm": 1.603601503409923,
"learning_rate": 5.084778511293731e-06,
"loss": 0.1426,
"step": 324
},
{
"epoch": 2.08,
"grad_norm": 1.4112657690685506,
"learning_rate": 5.070797130701618e-06,
"loss": 0.1251,
"step": 325
},
{
"epoch": 2.09,
"grad_norm": 1.5494589174434166,
"learning_rate": 5.056784291110794e-06,
"loss": 0.1271,
"step": 326
},
{
"epoch": 2.1,
"grad_norm": 1.4343244322587299,
"learning_rate": 5.04274027316215e-06,
"loss": 0.1072,
"step": 327
},
{
"epoch": 2.1,
"grad_norm": 1.5882705358537534,
"learning_rate": 5.028665358120995e-06,
"loss": 0.1503,
"step": 328
},
{
"epoch": 2.11,
"grad_norm": 1.5363658563678113,
"learning_rate": 5.014559827871426e-06,
"loss": 0.1303,
"step": 329
},
{
"epoch": 2.12,
"grad_norm": 1.5525555773630861,
"learning_rate": 5.00042396491069e-06,
"loss": 0.1231,
"step": 330
},
{
"epoch": 2.12,
"grad_norm": 1.5224615240989583,
"learning_rate": 4.9862580523435116e-06,
"loss": 0.0949,
"step": 331
},
{
"epoch": 2.13,
"grad_norm": 1.630845815321759,
"learning_rate": 4.972062373876435e-06,
"loss": 0.0923,
"step": 332
},
{
"epoch": 2.13,
"grad_norm": 1.7508423838612863,
"learning_rate": 4.95783721381214e-06,
"loss": 0.1557,
"step": 333
},
{
"epoch": 2.14,
"grad_norm": 1.9071964194859563,
"learning_rate": 4.943582857043742e-06,
"loss": 0.1315,
"step": 334
},
{
"epoch": 2.15,
"grad_norm": 1.605736114180956,
"learning_rate": 4.9292995890490945e-06,
"loss": 0.1017,
"step": 335
},
{
"epoch": 2.15,
"grad_norm": 1.5908541224269843,
"learning_rate": 4.914987695885067e-06,
"loss": 0.1097,
"step": 336
},
{
"epoch": 2.16,
"grad_norm": 1.4636955088177965,
"learning_rate": 4.900647464181817e-06,
"loss": 0.1237,
"step": 337
},
{
"epoch": 2.17,
"grad_norm": 1.4495226641748151,
"learning_rate": 4.886279181137049e-06,
"loss": 0.0968,
"step": 338
},
{
"epoch": 2.17,
"grad_norm": 1.4675886550946595,
"learning_rate": 4.871883134510263e-06,
"loss": 0.1011,
"step": 339
},
{
"epoch": 2.18,
"grad_norm": 1.6564244575191527,
"learning_rate": 4.8574596126169925e-06,
"loss": 0.1273,
"step": 340
},
{
"epoch": 2.19,
"grad_norm": 1.533227037458805,
"learning_rate": 4.843008904323029e-06,
"loss": 0.1228,
"step": 341
},
{
"epoch": 2.19,
"grad_norm": 1.4820663768576567,
"learning_rate": 4.828531299038638e-06,
"loss": 0.1099,
"step": 342
},
{
"epoch": 2.2,
"grad_norm": 1.6027220866794623,
"learning_rate": 4.81402708671276e-06,
"loss": 0.0969,
"step": 343
},
{
"epoch": 2.21,
"grad_norm": 1.5893012168299308,
"learning_rate": 4.799496557827208e-06,
"loss": 0.1102,
"step": 344
},
{
"epoch": 2.21,
"grad_norm": 1.5207909320236375,
"learning_rate": 4.7849400033908465e-06,
"loss": 0.1002,
"step": 345
},
{
"epoch": 2.22,
"grad_norm": 1.4497038100819872,
"learning_rate": 4.770357714933765e-06,
"loss": 0.1122,
"step": 346
},
{
"epoch": 2.22,
"grad_norm": 1.5568347840805732,
"learning_rate": 4.755749984501437e-06,
"loss": 0.1132,
"step": 347
},
{
"epoch": 2.23,
"grad_norm": 1.4518839890957804,
"learning_rate": 4.741117104648874e-06,
"loss": 0.1007,
"step": 348
},
{
"epoch": 2.24,
"grad_norm": 1.3028020064942243,
"learning_rate": 4.726459368434768e-06,
"loss": 0.0843,
"step": 349
},
{
"epoch": 2.24,
"grad_norm": 1.297097957128824,
"learning_rate": 4.711777069415615e-06,
"loss": 0.0967,
"step": 350
},
{
"epoch": 2.25,
"grad_norm": 1.4228303278540766,
"learning_rate": 4.697070501639841e-06,
"loss": 0.089,
"step": 351
},
{
"epoch": 2.26,
"grad_norm": 1.3898888054417786,
"learning_rate": 4.682339959641915e-06,
"loss": 0.0903,
"step": 352
},
{
"epoch": 2.26,
"grad_norm": 1.5233221859336192,
"learning_rate": 4.667585738436448e-06,
"loss": 0.0964,
"step": 353
},
{
"epoch": 2.27,
"grad_norm": 1.5295654378299237,
"learning_rate": 4.652808133512279e-06,
"loss": 0.0848,
"step": 354
},
{
"epoch": 2.28,
"grad_norm": 1.6437228500672438,
"learning_rate": 4.638007440826568e-06,
"loss": 0.0804,
"step": 355
},
{
"epoch": 2.28,
"grad_norm": 1.5383689907196065,
"learning_rate": 4.62318395679886e-06,
"loss": 0.0709,
"step": 356
},
{
"epoch": 2.29,
"grad_norm": 1.648388694748838,
"learning_rate": 4.6083379783051545e-06,
"loss": 0.0858,
"step": 357
},
{
"epoch": 2.29,
"grad_norm": 1.4723706715778246,
"learning_rate": 4.593469802671951e-06,
"loss": 0.077,
"step": 358
},
{
"epoch": 2.3,
"grad_norm": 1.3850116104824282,
"learning_rate": 4.5785797276703075e-06,
"loss": 0.0559,
"step": 359
},
{
"epoch": 2.31,
"grad_norm": 1.5055414247850987,
"learning_rate": 4.563668051509864e-06,
"loss": 0.058,
"step": 360
},
{
"epoch": 2.31,
"grad_norm": 1.5049490521822901,
"learning_rate": 4.548735072832879e-06,
"loss": 0.0809,
"step": 361
},
{
"epoch": 2.32,
"grad_norm": 1.4324576412777454,
"learning_rate": 4.533781090708244e-06,
"loss": 0.0639,
"step": 362
},
{
"epoch": 2.33,
"grad_norm": 1.3643187351188586,
"learning_rate": 4.518806404625495e-06,
"loss": 0.0694,
"step": 363
},
{
"epoch": 2.33,
"grad_norm": 1.6866178990377743,
"learning_rate": 4.503811314488816e-06,
"loss": 0.0985,
"step": 364
},
{
"epoch": 2.34,
"grad_norm": 1.4991148902916245,
"learning_rate": 4.48879612061103e-06,
"loss": 0.079,
"step": 365
},
{
"epoch": 2.35,
"grad_norm": 1.4720035797123598,
"learning_rate": 4.473761123707584e-06,
"loss": 0.0921,
"step": 366
},
{
"epoch": 2.35,
"grad_norm": 1.44832360985762,
"learning_rate": 4.458706624890534e-06,
"loss": 0.0786,
"step": 367
},
{
"epoch": 2.36,
"grad_norm": 1.4275898527484465,
"learning_rate": 4.443632925662504e-06,
"loss": 0.0626,
"step": 368
},
{
"epoch": 2.37,
"grad_norm": 1.5167475621388269,
"learning_rate": 4.428540327910652e-06,
"loss": 0.0723,
"step": 369
},
{
"epoch": 2.37,
"grad_norm": 1.4931710488796675,
"learning_rate": 4.41342913390063e-06,
"loss": 0.0789,
"step": 370
},
{
"epoch": 2.38,
"grad_norm": 1.6494646753660156,
"learning_rate": 4.398299646270518e-06,
"loss": 0.0779,
"step": 371
},
{
"epoch": 2.38,
"grad_norm": 1.4632009250956346,
"learning_rate": 4.3831521680247765e-06,
"loss": 0.0874,
"step": 372
},
{
"epoch": 2.39,
"grad_norm": 1.3624276184155342,
"learning_rate": 4.3679870025281644e-06,
"loss": 0.0595,
"step": 373
},
{
"epoch": 2.4,
"grad_norm": 1.4387226841227507,
"learning_rate": 4.352804453499677e-06,
"loss": 0.0799,
"step": 374
},
{
"epoch": 2.4,
"grad_norm": 1.4191511182239602,
"learning_rate": 4.3376048250064525e-06,
"loss": 0.0676,
"step": 375
},
{
"epoch": 2.41,
"grad_norm": 1.5148194681863063,
"learning_rate": 4.322388421457687e-06,
"loss": 0.0804,
"step": 376
},
{
"epoch": 2.42,
"grad_norm": 1.4992240453279386,
"learning_rate": 4.30715554759854e-06,
"loss": 0.0663,
"step": 377
},
{
"epoch": 2.42,
"grad_norm": 1.4940181433918933,
"learning_rate": 4.2919065085040285e-06,
"loss": 0.0729,
"step": 378
},
{
"epoch": 2.43,
"grad_norm": 1.4292506822415902,
"learning_rate": 4.276641609572911e-06,
"loss": 0.077,
"step": 379
},
{
"epoch": 2.44,
"grad_norm": 1.3880559130653471,
"learning_rate": 4.261361156521586e-06,
"loss": 0.0621,
"step": 380
},
{
"epoch": 2.44,
"grad_norm": 1.3224698912209505,
"learning_rate": 4.246065455377956e-06,
"loss": 0.0664,
"step": 381
},
{
"epoch": 2.45,
"grad_norm": 1.6220854664403273,
"learning_rate": 4.230754812475306e-06,
"loss": 0.075,
"step": 382
},
{
"epoch": 2.46,
"grad_norm": 1.3904399822037876,
"learning_rate": 4.215429534446161e-06,
"loss": 0.068,
"step": 383
},
{
"epoch": 2.46,
"grad_norm": 1.394224095253479,
"learning_rate": 4.200089928216156e-06,
"loss": 0.0694,
"step": 384
},
{
"epoch": 2.47,
"grad_norm": 1.574441933806699,
"learning_rate": 4.1847363009978776e-06,
"loss": 0.0682,
"step": 385
},
{
"epoch": 2.47,
"grad_norm": 1.5729152887669275,
"learning_rate": 4.169368960284718e-06,
"loss": 0.0737,
"step": 386
},
{
"epoch": 2.48,
"grad_norm": 1.4927966019088243,
"learning_rate": 4.153988213844717e-06,
"loss": 0.0654,
"step": 387
},
{
"epoch": 2.49,
"grad_norm": 1.450195896852466,
"learning_rate": 4.138594369714394e-06,
"loss": 0.0642,
"step": 388
},
{
"epoch": 2.49,
"grad_norm": 1.490138365122018,
"learning_rate": 4.123187736192583e-06,
"loss": 0.0606,
"step": 389
},
{
"epoch": 2.5,
"grad_norm": 1.4999776523444637,
"learning_rate": 4.107768621834257e-06,
"loss": 0.0649,
"step": 390
},
{
"epoch": 2.51,
"grad_norm": 1.4426594028410584,
"learning_rate": 4.092337335444343e-06,
"loss": 0.059,
"step": 391
},
{
"epoch": 2.51,
"grad_norm": 1.2377723587830607,
"learning_rate": 4.076894186071548e-06,
"loss": 0.0486,
"step": 392
},
{
"epoch": 2.52,
"grad_norm": 1.6147108783025788,
"learning_rate": 4.061439483002161e-06,
"loss": 0.0776,
"step": 393
},
{
"epoch": 2.53,
"grad_norm": 1.4665221548831489,
"learning_rate": 4.045973535753863e-06,
"loss": 0.0485,
"step": 394
},
{
"epoch": 2.53,
"grad_norm": 1.3979103729080165,
"learning_rate": 4.030496654069524e-06,
"loss": 0.0534,
"step": 395
},
{
"epoch": 2.54,
"grad_norm": 1.5024378161971237,
"learning_rate": 4.015009147911007e-06,
"loss": 0.0795,
"step": 396
},
{
"epoch": 2.54,
"grad_norm": 1.503457250091824,
"learning_rate": 3.9995113274529506e-06,
"loss": 0.0589,
"step": 397
},
{
"epoch": 2.55,
"grad_norm": 1.3018136990746167,
"learning_rate": 3.984003503076566e-06,
"loss": 0.0751,
"step": 398
},
{
"epoch": 2.56,
"grad_norm": 1.439789169382561,
"learning_rate": 3.968485985363416e-06,
"loss": 0.0671,
"step": 399
},
{
"epoch": 2.56,
"grad_norm": 1.2600931675117872,
"learning_rate": 3.952959085089193e-06,
"loss": 0.0481,
"step": 400
},
{
"epoch": 2.57,
"grad_norm": 1.3902574629104398,
"learning_rate": 3.937423113217505e-06,
"loss": 0.0605,
"step": 401
},
{
"epoch": 2.58,
"grad_norm": 1.3343472976969517,
"learning_rate": 3.92187838089363e-06,
"loss": 0.0709,
"step": 402
},
{
"epoch": 2.58,
"grad_norm": 1.260320342311725,
"learning_rate": 3.9063251994383055e-06,
"loss": 0.0646,
"step": 403
},
{
"epoch": 2.59,
"grad_norm": 1.3456955333019955,
"learning_rate": 3.8907638803414774e-06,
"loss": 0.063,
"step": 404
},
{
"epoch": 2.6,
"grad_norm": 1.3089048629072635,
"learning_rate": 3.875194735256067e-06,
"loss": 0.0663,
"step": 405
},
{
"epoch": 2.6,
"grad_norm": 1.458165867482108,
"learning_rate": 3.859618075991735e-06,
"loss": 0.0592,
"step": 406
},
{
"epoch": 2.61,
"grad_norm": 1.549636534322225,
"learning_rate": 3.844034214508625e-06,
"loss": 0.0773,
"step": 407
},
{
"epoch": 2.62,
"grad_norm": 1.4213891673540038,
"learning_rate": 3.828443462911128e-06,
"loss": 0.0628,
"step": 408
},
{
"epoch": 2.62,
"grad_norm": 1.4478063303906572,
"learning_rate": 3.8128461334416223e-06,
"loss": 0.0746,
"step": 409
},
{
"epoch": 2.63,
"grad_norm": 1.371728736392384,
"learning_rate": 3.7972425384742264e-06,
"loss": 0.0592,
"step": 410
},
{
"epoch": 2.63,
"grad_norm": 1.550899976232104,
"learning_rate": 3.781632990508541e-06,
"loss": 0.0771,
"step": 411
},
{
"epoch": 2.64,
"grad_norm": 1.4139524918953215,
"learning_rate": 3.766017802163386e-06,
"loss": 0.0687,
"step": 412
},
{
"epoch": 2.65,
"grad_norm": 1.3375701401404378,
"learning_rate": 3.7503972861705478e-06,
"loss": 0.0699,
"step": 413
},
{
"epoch": 2.65,
"grad_norm": 1.272959839235587,
"learning_rate": 3.7347717553685084e-06,
"loss": 0.0469,
"step": 414
},
{
"epoch": 2.66,
"grad_norm": 1.423066310160438,
"learning_rate": 3.7191415226961867e-06,
"loss": 0.0557,
"step": 415
},
{
"epoch": 2.67,
"grad_norm": 1.481943036415743,
"learning_rate": 3.703506901186665e-06,
"loss": 0.0861,
"step": 416
},
{
"epoch": 2.67,
"grad_norm": 1.282314760769859,
"learning_rate": 3.6878682039609253e-06,
"loss": 0.0473,
"step": 417
},
{
"epoch": 2.68,
"grad_norm": 1.3737402610119698,
"learning_rate": 3.6722257442215736e-06,
"loss": 0.063,
"step": 418
},
{
"epoch": 2.69,
"grad_norm": 1.297558587148971,
"learning_rate": 3.6565798352465697e-06,
"loss": 0.0479,
"step": 419
},
{
"epoch": 2.69,
"grad_norm": 1.2272283470980672,
"learning_rate": 3.640930790382953e-06,
"loss": 0.0614,
"step": 420
},
{
"epoch": 2.7,
"grad_norm": 1.396522035608687,
"learning_rate": 3.625278923040567e-06,
"loss": 0.0757,
"step": 421
},
{
"epoch": 2.71,
"grad_norm": 1.4743070354454288,
"learning_rate": 3.6096245466857808e-06,
"loss": 0.0835,
"step": 422
},
{
"epoch": 2.71,
"grad_norm": 1.462951272372007,
"learning_rate": 3.5939679748352146e-06,
"loss": 0.0771,
"step": 423
},
{
"epoch": 2.72,
"grad_norm": 1.6323533328578446,
"learning_rate": 3.578309521049456e-06,
"loss": 0.0658,
"step": 424
},
{
"epoch": 2.72,
"grad_norm": 1.4279696560264612,
"learning_rate": 3.562649498926785e-06,
"loss": 0.0603,
"step": 425
},
{
"epoch": 2.73,
"grad_norm": 1.5206763046785612,
"learning_rate": 3.546988222096891e-06,
"loss": 0.0688,
"step": 426
},
{
"epoch": 2.74,
"grad_norm": 1.5407667167480086,
"learning_rate": 3.531326004214592e-06,
"loss": 0.055,
"step": 427
},
{
"epoch": 2.74,
"grad_norm": 1.6834449307372186,
"learning_rate": 3.515663158953552e-06,
"loss": 0.0649,
"step": 428
},
{
"epoch": 2.75,
"grad_norm": 1.46259189275902,
"learning_rate": 3.5e-06,
"loss": 0.0509,
"step": 429
},
{
"epoch": 2.76,
"grad_norm": 1.3866087835219674,
"learning_rate": 3.484336841046448e-06,
"loss": 0.0618,
"step": 430
},
{
"epoch": 2.76,
"grad_norm": 1.3649733562457513,
"learning_rate": 3.468673995785409e-06,
"loss": 0.0479,
"step": 431
},
{
"epoch": 2.77,
"grad_norm": 1.389982768703036,
"learning_rate": 3.4530117779031096e-06,
"loss": 0.0623,
"step": 432
},
{
"epoch": 2.78,
"grad_norm": 1.324763468165176,
"learning_rate": 3.4373505010732152e-06,
"loss": 0.0654,
"step": 433
},
{
"epoch": 2.78,
"grad_norm": 1.4641723601034753,
"learning_rate": 3.4216904789505444e-06,
"loss": 0.0628,
"step": 434
},
{
"epoch": 2.79,
"grad_norm": 1.2958642357583487,
"learning_rate": 3.4060320251647866e-06,
"loss": 0.0621,
"step": 435
},
{
"epoch": 2.79,
"grad_norm": 1.2985540091619692,
"learning_rate": 3.3903754533142195e-06,
"loss": 0.0632,
"step": 436
},
{
"epoch": 2.8,
"grad_norm": 1.5315659150226084,
"learning_rate": 3.374721076959433e-06,
"loss": 0.0577,
"step": 437
},
{
"epoch": 2.81,
"grad_norm": 1.5259417180562567,
"learning_rate": 3.359069209617048e-06,
"loss": 0.0714,
"step": 438
},
{
"epoch": 2.81,
"grad_norm": 1.3901162153354532,
"learning_rate": 3.3434201647534306e-06,
"loss": 0.0552,
"step": 439
},
{
"epoch": 2.82,
"grad_norm": 1.3074942111445218,
"learning_rate": 3.3277742557784263e-06,
"loss": 0.0597,
"step": 440
},
{
"epoch": 2.83,
"grad_norm": 1.4421771168227446,
"learning_rate": 3.312131796039074e-06,
"loss": 0.0888,
"step": 441
},
{
"epoch": 2.83,
"grad_norm": 1.217310142638148,
"learning_rate": 3.296493098813335e-06,
"loss": 0.0533,
"step": 442
},
{
"epoch": 2.84,
"grad_norm": 1.541068920563886,
"learning_rate": 3.280858477303813e-06,
"loss": 0.0866,
"step": 443
},
{
"epoch": 2.85,
"grad_norm": 1.469329265275979,
"learning_rate": 3.265228244631491e-06,
"loss": 0.0746,
"step": 444
},
{
"epoch": 2.85,
"grad_norm": 1.3663935959813323,
"learning_rate": 3.2496027138294534e-06,
"loss": 0.062,
"step": 445
},
{
"epoch": 2.86,
"grad_norm": 1.4233076114764764,
"learning_rate": 3.2339821978366144e-06,
"loss": 0.0549,
"step": 446
},
{
"epoch": 2.87,
"grad_norm": 1.4125981717018385,
"learning_rate": 3.2183670094914596e-06,
"loss": 0.0785,
"step": 447
},
{
"epoch": 2.87,
"grad_norm": 1.2443476277494099,
"learning_rate": 3.2027574615257726e-06,
"loss": 0.0594,
"step": 448
},
{
"epoch": 2.88,
"grad_norm": 1.3649462287414904,
"learning_rate": 3.1871538665583784e-06,
"loss": 0.0643,
"step": 449
},
{
"epoch": 2.88,
"grad_norm": 1.4442414248699849,
"learning_rate": 3.171556537088873e-06,
"loss": 0.0838,
"step": 450
},
{
"epoch": 2.89,
"grad_norm": 1.3162685413703437,
"learning_rate": 3.155965785491375e-06,
"loss": 0.0748,
"step": 451
},
{
"epoch": 2.9,
"grad_norm": 1.3094037910120564,
"learning_rate": 3.140381924008266e-06,
"loss": 0.0634,
"step": 452
},
{
"epoch": 2.9,
"grad_norm": 1.430140211858692,
"learning_rate": 3.1248052647439327e-06,
"loss": 0.0585,
"step": 453
},
{
"epoch": 2.91,
"grad_norm": 1.3430910234317202,
"learning_rate": 3.109236119658523e-06,
"loss": 0.0545,
"step": 454
},
{
"epoch": 2.92,
"grad_norm": 1.32857879641887,
"learning_rate": 3.0936748005616936e-06,
"loss": 0.0548,
"step": 455
},
{
"epoch": 2.92,
"grad_norm": 1.2304110525061114,
"learning_rate": 3.0781216191063695e-06,
"loss": 0.0367,
"step": 456
},
{
"epoch": 2.93,
"grad_norm": 1.2175987934274146,
"learning_rate": 3.0625768867824957e-06,
"loss": 0.0481,
"step": 457
},
{
"epoch": 2.94,
"grad_norm": 1.4251000510540908,
"learning_rate": 3.047040914910806e-06,
"loss": 0.0607,
"step": 458
},
{
"epoch": 2.94,
"grad_norm": 1.2232338433186838,
"learning_rate": 3.0315140146365854e-06,
"loss": 0.0508,
"step": 459
},
{
"epoch": 2.95,
"grad_norm": 1.8361121743484377,
"learning_rate": 3.015996496923435e-06,
"loss": 0.101,
"step": 460
},
{
"epoch": 2.96,
"grad_norm": 1.4585021415736545,
"learning_rate": 3.00048867254705e-06,
"loss": 0.0739,
"step": 461
},
{
"epoch": 2.96,
"grad_norm": 1.5678316587634316,
"learning_rate": 2.9849908520889936e-06,
"loss": 0.0732,
"step": 462
},
{
"epoch": 2.97,
"grad_norm": 1.5229061347267796,
"learning_rate": 2.9695033459304766e-06,
"loss": 0.0728,
"step": 463
},
{
"epoch": 2.97,
"grad_norm": 1.452042874894508,
"learning_rate": 2.954026464246138e-06,
"loss": 0.0566,
"step": 464
},
{
"epoch": 2.98,
"grad_norm": 1.4256233554820992,
"learning_rate": 2.9385605169978387e-06,
"loss": 0.0527,
"step": 465
},
{
"epoch": 2.99,
"grad_norm": 1.3611842237197957,
"learning_rate": 2.923105813928453e-06,
"loss": 0.0513,
"step": 466
},
{
"epoch": 2.99,
"grad_norm": 1.2023007795378646,
"learning_rate": 2.907662664555658e-06,
"loss": 0.0481,
"step": 467
},
{
"epoch": 3.0,
"grad_norm": 1.5067980607363667,
"learning_rate": 2.8922313781657437e-06,
"loss": 0.0554,
"step": 468
},
{
"epoch": 3.01,
"grad_norm": 1.1875974890545655,
"learning_rate": 2.876812263807417e-06,
"loss": 0.0455,
"step": 469
},
{
"epoch": 3.01,
"grad_norm": 1.4615326175716894,
"learning_rate": 2.861405630285606e-06,
"loss": 0.0653,
"step": 470
},
{
"epoch": 3.02,
"grad_norm": 1.2370763939930607,
"learning_rate": 2.8460117861552833e-06,
"loss": 0.0683,
"step": 471
},
{
"epoch": 3.03,
"grad_norm": 1.2836073017492384,
"learning_rate": 2.8306310397152817e-06,
"loss": 0.0638,
"step": 472
},
{
"epoch": 3.03,
"grad_norm": 1.2609833632520322,
"learning_rate": 2.815263699002124e-06,
"loss": 0.0469,
"step": 473
},
{
"epoch": 3.04,
"grad_norm": 1.1616499807754248,
"learning_rate": 2.799910071783845e-06,
"loss": 0.0408,
"step": 474
},
{
"epoch": 3.04,
"grad_norm": 1.1766220146156734,
"learning_rate": 2.7845704655538383e-06,
"loss": 0.0447,
"step": 475
},
{
"epoch": 3.05,
"grad_norm": 1.3126347990175358,
"learning_rate": 2.7692451875246956e-06,
"loss": 0.0644,
"step": 476
},
{
"epoch": 3.06,
"grad_norm": 1.1740389147488146,
"learning_rate": 2.7539345446220444e-06,
"loss": 0.0472,
"step": 477
},
{
"epoch": 3.06,
"grad_norm": 1.16464372281638,
"learning_rate": 2.7386388434784143e-06,
"loss": 0.0537,
"step": 478
},
{
"epoch": 3.07,
"grad_norm": 1.3253535370867997,
"learning_rate": 2.723358390427089e-06,
"loss": 0.0589,
"step": 479
},
{
"epoch": 3.08,
"grad_norm": 1.2251011192581163,
"learning_rate": 2.708093491495973e-06,
"loss": 0.0443,
"step": 480
},
{
"epoch": 3.08,
"grad_norm": 1.0091161261068804,
"learning_rate": 2.6928444524014595e-06,
"loss": 0.0359,
"step": 481
},
{
"epoch": 3.09,
"grad_norm": 1.146439960851428,
"learning_rate": 2.6776115785423123e-06,
"loss": 0.0365,
"step": 482
},
{
"epoch": 3.1,
"grad_norm": 1.0769495975019767,
"learning_rate": 2.6623951749935487e-06,
"loss": 0.0327,
"step": 483
},
{
"epoch": 3.1,
"grad_norm": 1.2241733458914144,
"learning_rate": 2.6471955465003237e-06,
"loss": 0.0478,
"step": 484
},
{
"epoch": 3.11,
"grad_norm": 1.3735061216517803,
"learning_rate": 2.6320129974718355e-06,
"loss": 0.0465,
"step": 485
},
{
"epoch": 3.12,
"grad_norm": 1.2436263847431321,
"learning_rate": 2.616847831975224e-06,
"loss": 0.0372,
"step": 486
},
{
"epoch": 3.12,
"grad_norm": 1.155193565242057,
"learning_rate": 2.601700353729481e-06,
"loss": 0.028,
"step": 487
},
{
"epoch": 3.13,
"grad_norm": 1.388487580103464,
"learning_rate": 2.58657086609937e-06,
"loss": 0.0359,
"step": 488
},
{
"epoch": 3.13,
"grad_norm": 1.3725232569394905,
"learning_rate": 2.5714596720893473e-06,
"loss": 0.0528,
"step": 489
},
{
"epoch": 3.14,
"grad_norm": 1.7184133842279257,
"learning_rate": 2.5563670743374973e-06,
"loss": 0.0514,
"step": 490
},
{
"epoch": 3.15,
"grad_norm": 1.3252001006412628,
"learning_rate": 2.5412933751094662e-06,
"loss": 0.0327,
"step": 491
},
{
"epoch": 3.15,
"grad_norm": 1.2819895992542942,
"learning_rate": 2.5262388762924157e-06,
"loss": 0.036,
"step": 492
},
{
"epoch": 3.16,
"grad_norm": 1.1604966130662278,
"learning_rate": 2.5112038793889706e-06,
"loss": 0.0412,
"step": 493
},
{
"epoch": 3.17,
"grad_norm": 1.3519367739400956,
"learning_rate": 2.496188685511185e-06,
"loss": 0.0302,
"step": 494
},
{
"epoch": 3.17,
"grad_norm": 1.2596321452487949,
"learning_rate": 2.481193595374505e-06,
"loss": 0.0295,
"step": 495
},
{
"epoch": 3.18,
"grad_norm": 1.3630477349522254,
"learning_rate": 2.4662189092917563e-06,
"loss": 0.0471,
"step": 496
},
{
"epoch": 3.19,
"grad_norm": 1.279411209620521,
"learning_rate": 2.4512649271671214e-06,
"loss": 0.0358,
"step": 497
},
{
"epoch": 3.19,
"grad_norm": 1.324111956217256,
"learning_rate": 2.436331948490136e-06,
"loss": 0.0379,
"step": 498
},
{
"epoch": 3.2,
"grad_norm": 1.3781397599646155,
"learning_rate": 2.4214202723296924e-06,
"loss": 0.0312,
"step": 499
},
{
"epoch": 3.21,
"grad_norm": 1.2268909864908872,
"learning_rate": 2.4065301973280486e-06,
"loss": 0.039,
"step": 500
},
{
"epoch": 3.21,
"grad_norm": 1.2617157160234982,
"learning_rate": 2.391662021694847e-06,
"loss": 0.0348,
"step": 501
},
{
"epoch": 3.22,
"grad_norm": 1.354583484533038,
"learning_rate": 2.3768160432011395e-06,
"loss": 0.045,
"step": 502
},
{
"epoch": 3.22,
"grad_norm": 1.4745147513254047,
"learning_rate": 2.3619925591734323e-06,
"loss": 0.0336,
"step": 503
},
{
"epoch": 3.23,
"grad_norm": 1.1138377432924618,
"learning_rate": 2.3471918664877217e-06,
"loss": 0.0327,
"step": 504
},
{
"epoch": 3.24,
"grad_norm": 1.1310391517344176,
"learning_rate": 2.332414261563553e-06,
"loss": 0.0322,
"step": 505
},
{
"epoch": 3.24,
"grad_norm": 1.1441571285331864,
"learning_rate": 2.317660040358085e-06,
"loss": 0.0366,
"step": 506
},
{
"epoch": 3.25,
"grad_norm": 1.1413826309676043,
"learning_rate": 2.3029294983601598e-06,
"loss": 0.0318,
"step": 507
},
{
"epoch": 3.26,
"grad_norm": 1.021725022217316,
"learning_rate": 2.2882229305843866e-06,
"loss": 0.0313,
"step": 508
},
{
"epoch": 3.26,
"grad_norm": 1.0432467561700225,
"learning_rate": 2.2735406315652323e-06,
"loss": 0.0313,
"step": 509
},
{
"epoch": 3.27,
"grad_norm": 1.0667155748112274,
"learning_rate": 2.258882895351125e-06,
"loss": 0.0276,
"step": 510
},
{
"epoch": 3.28,
"grad_norm": 1.076264156121743,
"learning_rate": 2.2442500154985643e-06,
"loss": 0.0314,
"step": 511
},
{
"epoch": 3.28,
"grad_norm": 1.2337389573803583,
"learning_rate": 2.229642285066236e-06,
"loss": 0.0276,
"step": 512
},
{
"epoch": 3.29,
"grad_norm": 1.1832754817783542,
"learning_rate": 2.215059996609154e-06,
"loss": 0.04,
"step": 513
},
{
"epoch": 3.29,
"grad_norm": 1.0855122122885388,
"learning_rate": 2.200503442172792e-06,
"loss": 0.0307,
"step": 514
},
{
"epoch": 3.3,
"grad_norm": 0.9598119147789916,
"learning_rate": 2.185972913287241e-06,
"loss": 0.0188,
"step": 515
},
{
"epoch": 3.31,
"grad_norm": 0.9791085255254713,
"learning_rate": 2.1714687009613628e-06,
"loss": 0.0177,
"step": 516
},
{
"epoch": 3.31,
"grad_norm": 1.0477977537384306,
"learning_rate": 2.156991095676971e-06,
"loss": 0.0306,
"step": 517
},
{
"epoch": 3.32,
"grad_norm": 1.0696773212764674,
"learning_rate": 2.1425403873830083e-06,
"loss": 0.0245,
"step": 518
},
{
"epoch": 3.33,
"grad_norm": 0.9827891858254738,
"learning_rate": 2.1281168654897376e-06,
"loss": 0.0242,
"step": 519
},
{
"epoch": 3.33,
"grad_norm": 1.0998427460091256,
"learning_rate": 2.113720818862951e-06,
"loss": 0.0324,
"step": 520
},
{
"epoch": 3.34,
"grad_norm": 1.1371184865733825,
"learning_rate": 2.099352535818182e-06,
"loss": 0.0292,
"step": 521
},
{
"epoch": 3.35,
"grad_norm": 0.9408745922040473,
"learning_rate": 2.085012304114933e-06,
"loss": 0.0287,
"step": 522
},
{
"epoch": 3.35,
"grad_norm": 0.9922924631524535,
"learning_rate": 2.070700410950906e-06,
"loss": 0.0278,
"step": 523
},
{
"epoch": 3.36,
"grad_norm": 1.2103088874857002,
"learning_rate": 2.0564171429562587e-06,
"loss": 0.0253,
"step": 524
},
{
"epoch": 3.37,
"grad_norm": 1.1353549204106046,
"learning_rate": 2.042162786187862e-06,
"loss": 0.0256,
"step": 525
},
{
"epoch": 3.37,
"grad_norm": 1.1954879688124018,
"learning_rate": 2.027937626123565e-06,
"loss": 0.0281,
"step": 526
},
{
"epoch": 3.38,
"grad_norm": 1.1145609650353256,
"learning_rate": 2.0137419476564896e-06,
"loss": 0.0254,
"step": 527
},
{
"epoch": 3.38,
"grad_norm": 0.9904281919589162,
"learning_rate": 1.9995760350893098e-06,
"loss": 0.0288,
"step": 528
},
{
"epoch": 3.39,
"grad_norm": 0.916410698775379,
"learning_rate": 1.985440172128573e-06,
"loss": 0.0192,
"step": 529
},
{
"epoch": 3.4,
"grad_norm": 0.9135967072547435,
"learning_rate": 1.9713346418790058e-06,
"loss": 0.0243,
"step": 530
},
{
"epoch": 3.4,
"grad_norm": 1.0987171731854852,
"learning_rate": 1.957259726837849e-06,
"loss": 0.0237,
"step": 531
},
{
"epoch": 3.41,
"grad_norm": 1.0651180981741601,
"learning_rate": 1.9432157088892064e-06,
"loss": 0.0231,
"step": 532
},
{
"epoch": 3.42,
"grad_norm": 1.3470623952115612,
"learning_rate": 1.9292028692983824e-06,
"loss": 0.0234,
"step": 533
},
{
"epoch": 3.42,
"grad_norm": 1.1189511726531576,
"learning_rate": 1.91522148870627e-06,
"loss": 0.0223,
"step": 534
},
{
"epoch": 3.43,
"grad_norm": 0.9537039081471808,
"learning_rate": 1.9012718471237144e-06,
"loss": 0.0252,
"step": 535
},
{
"epoch": 3.44,
"grad_norm": 1.3303345811221687,
"learning_rate": 1.887354223925911e-06,
"loss": 0.0227,
"step": 536
},
{
"epoch": 3.44,
"grad_norm": 0.9577549623126181,
"learning_rate": 1.87346889784681e-06,
"loss": 0.0235,
"step": 537
},
{
"epoch": 3.45,
"grad_norm": 1.103057703050364,
"learning_rate": 1.8596161469735374e-06,
"loss": 0.0251,
"step": 538
},
{
"epoch": 3.46,
"grad_norm": 0.9698969241730576,
"learning_rate": 1.8457962487408175e-06,
"loss": 0.021,
"step": 539
},
{
"epoch": 3.46,
"grad_norm": 0.9735505582853365,
"learning_rate": 1.8320094799254222e-06,
"loss": 0.024,
"step": 540
},
{
"epoch": 3.47,
"grad_norm": 1.0579516176062307,
"learning_rate": 1.8182561166406308e-06,
"loss": 0.0252,
"step": 541
},
{
"epoch": 3.47,
"grad_norm": 1.3591760479632482,
"learning_rate": 1.8045364343306915e-06,
"loss": 0.0232,
"step": 542
},
{
"epoch": 3.48,
"grad_norm": 0.9698743028005155,
"learning_rate": 1.7908507077653124e-06,
"loss": 0.0228,
"step": 543
},
{
"epoch": 3.49,
"grad_norm": 1.0221498824263056,
"learning_rate": 1.7771992110341533e-06,
"loss": 0.0203,
"step": 544
},
{
"epoch": 3.49,
"grad_norm": 0.9962335713738956,
"learning_rate": 1.7635822175413446e-06,
"loss": 0.0208,
"step": 545
},
{
"epoch": 3.5,
"grad_norm": 0.9025864381589136,
"learning_rate": 1.7500000000000008e-06,
"loss": 0.0196,
"step": 546
},
{
"epoch": 3.51,
"grad_norm": 1.0627981287441168,
"learning_rate": 1.7364528304267646e-06,
"loss": 0.0216,
"step": 547
},
{
"epoch": 3.51,
"grad_norm": 0.8998289761019697,
"learning_rate": 1.7229409801363635e-06,
"loss": 0.0177,
"step": 548
},
{
"epoch": 3.52,
"grad_norm": 1.1032294275464554,
"learning_rate": 1.7094647197361656e-06,
"loss": 0.0252,
"step": 549
},
{
"epoch": 3.53,
"grad_norm": 1.0706133630954544,
"learning_rate": 1.6960243191207686e-06,
"loss": 0.0169,
"step": 550
},
{
"epoch": 3.53,
"grad_norm": 1.0878049572672386,
"learning_rate": 1.6826200474665891e-06,
"loss": 0.0218,
"step": 551
},
{
"epoch": 3.54,
"grad_norm": 1.0854933564945395,
"learning_rate": 1.669252173226479e-06,
"loss": 0.0276,
"step": 552
},
{
"epoch": 3.54,
"grad_norm": 1.1327704937861105,
"learning_rate": 1.6559209641243388e-06,
"loss": 0.0211,
"step": 553
},
{
"epoch": 3.55,
"grad_norm": 0.9110425525754956,
"learning_rate": 1.642626687149765e-06,
"loss": 0.0279,
"step": 554
},
{
"epoch": 3.56,
"grad_norm": 1.021236405302058,
"learning_rate": 1.629369608552696e-06,
"loss": 0.0186,
"step": 555
},
{
"epoch": 3.56,
"grad_norm": 0.9240338884131678,
"learning_rate": 1.6161499938380873e-06,
"loss": 0.0156,
"step": 556
},
{
"epoch": 3.57,
"grad_norm": 0.890168680945506,
"learning_rate": 1.6029681077605864e-06,
"loss": 0.0205,
"step": 557
},
{
"epoch": 3.58,
"grad_norm": 1.066699979491714,
"learning_rate": 1.5898242143192336e-06,
"loss": 0.0232,
"step": 558
},
{
"epoch": 3.58,
"grad_norm": 0.916678367435508,
"learning_rate": 1.576718576752179e-06,
"loss": 0.0195,
"step": 559
},
{
"epoch": 3.59,
"grad_norm": 0.8932364526829346,
"learning_rate": 1.5636514575314024e-06,
"loss": 0.0183,
"step": 560
},
{
"epoch": 3.6,
"grad_norm": 0.8546476261334689,
"learning_rate": 1.550623118357463e-06,
"loss": 0.0187,
"step": 561
},
{
"epoch": 3.6,
"grad_norm": 0.9062797963432817,
"learning_rate": 1.5376338201542538e-06,
"loss": 0.0189,
"step": 562
},
{
"epoch": 3.61,
"grad_norm": 0.9603196879330562,
"learning_rate": 1.5246838230637831e-06,
"loss": 0.0218,
"step": 563
},
{
"epoch": 3.62,
"grad_norm": 0.8140801991948985,
"learning_rate": 1.511773386440955e-06,
"loss": 0.0154,
"step": 564
},
{
"epoch": 3.62,
"grad_norm": 0.8292223643786333,
"learning_rate": 1.4989027688483808e-06,
"loss": 0.023,
"step": 565
},
{
"epoch": 3.63,
"grad_norm": 1.172191159788987,
"learning_rate": 1.4860722280512022e-06,
"loss": 0.0229,
"step": 566
},
{
"epoch": 3.63,
"grad_norm": 1.0524048010257814,
"learning_rate": 1.473282021011924e-06,
"loss": 0.0237,
"step": 567
},
{
"epoch": 3.64,
"grad_norm": 0.8884909829231797,
"learning_rate": 1.4605324038852707e-06,
"loss": 0.0184,
"step": 568
},
{
"epoch": 3.65,
"grad_norm": 0.9152406178203973,
"learning_rate": 1.4478236320130554e-06,
"loss": 0.0215,
"step": 569
},
{
"epoch": 3.65,
"grad_norm": 0.8117379173139481,
"learning_rate": 1.4351559599190708e-06,
"loss": 0.0133,
"step": 570
},
{
"epoch": 3.66,
"grad_norm": 1.4449523252498588,
"learning_rate": 1.4225296413039794e-06,
"loss": 0.0154,
"step": 571
},
{
"epoch": 3.67,
"grad_norm": 0.965819888705113,
"learning_rate": 1.4099449290402492e-06,
"loss": 0.0242,
"step": 572
},
{
"epoch": 3.67,
"grad_norm": 0.753534692636084,
"learning_rate": 1.3974020751670734e-06,
"loss": 0.0147,
"step": 573
},
{
"epoch": 3.68,
"grad_norm": 1.04736497011196,
"learning_rate": 1.3849013308853369e-06,
"loss": 0.0254,
"step": 574
},
{
"epoch": 3.69,
"grad_norm": 0.8503127295336432,
"learning_rate": 1.3724429465525733e-06,
"loss": 0.0125,
"step": 575
},
{
"epoch": 3.69,
"grad_norm": 0.8142674930830482,
"learning_rate": 1.360027171677957e-06,
"loss": 0.0169,
"step": 576
},
{
"epoch": 3.7,
"grad_norm": 0.8572587753256264,
"learning_rate": 1.3476542549173097e-06,
"loss": 0.0238,
"step": 577
},
{
"epoch": 3.71,
"grad_norm": 0.9520723281018718,
"learning_rate": 1.335324444068108e-06,
"loss": 0.0261,
"step": 578
},
{
"epoch": 3.71,
"grad_norm": 0.9394209110837798,
"learning_rate": 1.3230379860645363e-06,
"loss": 0.0219,
"step": 579
},
{
"epoch": 3.72,
"grad_norm": 0.9794515331971883,
"learning_rate": 1.3107951269725286e-06,
"loss": 0.0167,
"step": 580
},
{
"epoch": 3.72,
"grad_norm": 0.839665023620838,
"learning_rate": 1.2985961119848508e-06,
"loss": 0.0154,
"step": 581
},
{
"epoch": 3.73,
"grad_norm": 0.8596538810285997,
"learning_rate": 1.28644118541618e-06,
"loss": 0.0217,
"step": 582
},
{
"epoch": 3.74,
"grad_norm": 0.9648126172460845,
"learning_rate": 1.2743305906982184e-06,
"loss": 0.0185,
"step": 583
},
{
"epoch": 3.74,
"grad_norm": 0.9764156247033144,
"learning_rate": 1.2622645703748163e-06,
"loss": 0.018,
"step": 584
},
{
"epoch": 3.75,
"grad_norm": 0.8055243584840867,
"learning_rate": 1.2502433660971122e-06,
"loss": 0.012,
"step": 585
},
{
"epoch": 3.76,
"grad_norm": 0.7744512728259326,
"learning_rate": 1.2382672186187003e-06,
"loss": 0.0161,
"step": 586
},
{
"epoch": 3.76,
"grad_norm": 0.7743427992830083,
"learning_rate": 1.2263363677907975e-06,
"loss": 0.0137,
"step": 587
},
{
"epoch": 3.77,
"grad_norm": 0.7684669655309241,
"learning_rate": 1.214451052557453e-06,
"loss": 0.0151,
"step": 588
},
{
"epoch": 3.78,
"grad_norm": 0.7858334522617288,
"learning_rate": 1.202611510950747e-06,
"loss": 0.0174,
"step": 589
},
{
"epoch": 3.78,
"grad_norm": 0.9474020449844386,
"learning_rate": 1.1908179800860415e-06,
"loss": 0.0179,
"step": 590
},
{
"epoch": 3.79,
"grad_norm": 0.7981094796922926,
"learning_rate": 1.1790706961572176e-06,
"loss": 0.0156,
"step": 591
},
{
"epoch": 3.79,
"grad_norm": 0.8306780864601689,
"learning_rate": 1.167369894431949e-06,
"loss": 0.0218,
"step": 592
},
{
"epoch": 3.8,
"grad_norm": 0.8809681256095874,
"learning_rate": 1.1557158092469968e-06,
"loss": 0.015,
"step": 593
},
{
"epoch": 3.81,
"grad_norm": 0.805530152973458,
"learning_rate": 1.1441086740035036e-06,
"loss": 0.017,
"step": 594
},
{
"epoch": 3.81,
"grad_norm": 0.8917842406873597,
"learning_rate": 1.1325487211623343e-06,
"loss": 0.0162,
"step": 595
},
{
"epoch": 3.82,
"grad_norm": 1.0306467922307088,
"learning_rate": 1.121036182239403e-06,
"loss": 0.0136,
"step": 596
},
{
"epoch": 3.83,
"grad_norm": 0.9249608665631512,
"learning_rate": 1.1095712878010542e-06,
"loss": 0.0245,
"step": 597
},
{
"epoch": 3.83,
"grad_norm": 0.8900342673768057,
"learning_rate": 1.0981542674594327e-06,
"loss": 0.0188,
"step": 598
},
{
"epoch": 3.84,
"grad_norm": 1.0239966323535892,
"learning_rate": 1.08678534986789e-06,
"loss": 0.0247,
"step": 599
},
{
"epoch": 3.85,
"grad_norm": 0.895563506776096,
"learning_rate": 1.0754647627164022e-06,
"loss": 0.0245,
"step": 600
},
{
"epoch": 3.85,
"grad_norm": 0.92433246398148,
"learning_rate": 1.064192732727016e-06,
"loss": 0.0161,
"step": 601
},
{
"epoch": 3.86,
"grad_norm": 0.9108465392412978,
"learning_rate": 1.0529694856493002e-06,
"loss": 0.0166,
"step": 602
},
{
"epoch": 3.87,
"grad_norm": 0.8310135807305727,
"learning_rate": 1.0417952462558286e-06,
"loss": 0.0223,
"step": 603
},
{
"epoch": 3.87,
"grad_norm": 0.8255760949970268,
"learning_rate": 1.0306702383376813e-06,
"loss": 0.0195,
"step": 604
},
{
"epoch": 3.88,
"grad_norm": 0.7324462030084228,
"learning_rate": 1.0195946846999551e-06,
"loss": 0.0169,
"step": 605
},
{
"epoch": 3.88,
"grad_norm": 0.7994877527959611,
"learning_rate": 1.0085688071573086e-06,
"loss": 0.0225,
"step": 606
},
{
"epoch": 3.89,
"grad_norm": 0.931125141975873,
"learning_rate": 9.97592826529514e-07,
"loss": 0.0252,
"step": 607
},
{
"epoch": 3.9,
"grad_norm": 0.9471536735997285,
"learning_rate": 9.866669626370412e-07,
"loss": 0.0205,
"step": 608
},
{
"epoch": 3.9,
"grad_norm": 0.9266292677740381,
"learning_rate": 9.757914342966495e-07,
"loss": 0.015,
"step": 609
},
{
"epoch": 3.91,
"grad_norm": 0.7337228958953126,
"learning_rate": 9.649664593170062e-07,
"loss": 0.0146,
"step": 610
},
{
"epoch": 3.92,
"grad_norm": 0.8563548759407368,
"learning_rate": 9.541922544943295e-07,
"loss": 0.0181,
"step": 611
},
{
"epoch": 3.92,
"grad_norm": 0.8954454947897389,
"learning_rate": 9.434690356080394e-07,
"loss": 0.0131,
"step": 612
},
{
"epoch": 3.93,
"grad_norm": 0.7629698438514336,
"learning_rate": 9.327970174164409e-07,
"loss": 0.0147,
"step": 613
},
{
"epoch": 3.94,
"grad_norm": 0.7699002072470454,
"learning_rate": 9.221764136524202e-07,
"loss": 0.0171,
"step": 614
},
{
"epoch": 3.94,
"grad_norm": 0.7224366715433516,
"learning_rate": 9.116074370191705e-07,
"loss": 0.0141,
"step": 615
},
{
"epoch": 3.95,
"grad_norm": 0.7827602493253287,
"learning_rate": 9.010902991859196e-07,
"loss": 0.0333,
"step": 616
},
{
"epoch": 3.96,
"grad_norm": 0.8407830538126995,
"learning_rate": 8.906252107837054e-07,
"loss": 0.0229,
"step": 617
},
{
"epoch": 3.96,
"grad_norm": 0.9271058679893376,
"learning_rate": 8.802123814011458e-07,
"loss": 0.0212,
"step": 618
},
{
"epoch": 3.97,
"grad_norm": 0.8249118909539904,
"learning_rate": 8.698520195802499e-07,
"loss": 0.0178,
"step": 619
},
{
"epoch": 3.97,
"grad_norm": 0.9062535578279738,
"learning_rate": 8.595443328122345e-07,
"loss": 0.0151,
"step": 620
},
{
"epoch": 3.98,
"grad_norm": 0.9007400301918379,
"learning_rate": 8.492895275333705e-07,
"loss": 0.0149,
"step": 621
},
{
"epoch": 3.99,
"grad_norm": 0.6906468788571227,
"learning_rate": 8.390878091208544e-07,
"loss": 0.0121,
"step": 622
},
{
"epoch": 3.99,
"grad_norm": 0.8191320475767269,
"learning_rate": 8.289393818886837e-07,
"loss": 0.0159,
"step": 623
},
{
"epoch": 4.0,
"grad_norm": 1.0171382420120563,
"learning_rate": 8.188444490835774e-07,
"loss": 0.0182,
"step": 624
},
{
"epoch": 4.01,
"grad_norm": 0.6877168878883905,
"learning_rate": 8.088032128808952e-07,
"loss": 0.0115,
"step": 625
},
{
"epoch": 4.01,
"grad_norm": 0.8301412318523347,
"learning_rate": 7.988158743805973e-07,
"loss": 0.0192,
"step": 626
},
{
"epoch": 4.02,
"grad_norm": 0.7473939831637764,
"learning_rate": 7.888826336032093e-07,
"loss": 0.02,
"step": 627
},
{
"epoch": 4.03,
"grad_norm": 0.6891353595867798,
"learning_rate": 7.790036894858198e-07,
"loss": 0.0197,
"step": 628
},
{
"epoch": 4.03,
"grad_norm": 0.6608127967094858,
"learning_rate": 7.691792398780962e-07,
"loss": 0.0122,
"step": 629
},
{
"epoch": 4.04,
"grad_norm": 0.6801259718598915,
"learning_rate": 7.594094815383223e-07,
"loss": 0.0102,
"step": 630
},
{
"epoch": 4.04,
"grad_norm": 0.6607722356860742,
"learning_rate": 7.496946101294585e-07,
"loss": 0.0124,
"step": 631
},
{
"epoch": 4.05,
"grad_norm": 0.6764859766385257,
"learning_rate": 7.400348202152192e-07,
"loss": 0.0186,
"step": 632
},
{
"epoch": 4.06,
"grad_norm": 0.5890171229802623,
"learning_rate": 7.304303052561841e-07,
"loss": 0.0121,
"step": 633
},
{
"epoch": 4.06,
"grad_norm": 0.7586985486515609,
"learning_rate": 7.208812576059113e-07,
"loss": 0.0164,
"step": 634
},
{
"epoch": 4.07,
"grad_norm": 0.7284236571177901,
"learning_rate": 7.113878685070994e-07,
"loss": 0.0165,
"step": 635
},
{
"epoch": 4.08,
"grad_norm": 0.72155125556459,
"learning_rate": 7.019503280877466e-07,
"loss": 0.0105,
"step": 636
},
{
"epoch": 4.08,
"grad_norm": 0.7008419280912078,
"learning_rate": 6.925688253573465e-07,
"loss": 0.0125,
"step": 637
},
{
"epoch": 4.09,
"grad_norm": 0.5900754345006012,
"learning_rate": 6.832435482031064e-07,
"loss": 0.01,
"step": 638
},
{
"epoch": 4.1,
"grad_norm": 0.5236568988968674,
"learning_rate": 6.73974683386176e-07,
"loss": 0.0093,
"step": 639
},
{
"epoch": 4.1,
"grad_norm": 0.5402196308668088,
"learning_rate": 6.647624165379173e-07,
"loss": 0.012,
"step": 640
},
{
"epoch": 4.11,
"grad_norm": 0.7153834377519408,
"learning_rate": 6.55606932156175e-07,
"loss": 0.0125,
"step": 641
},
{
"epoch": 4.12,
"grad_norm": 0.7035478419818489,
"learning_rate": 6.465084136015951e-07,
"loss": 0.0095,
"step": 642
},
{
"epoch": 4.12,
"grad_norm": 0.450441289716298,
"learning_rate": 6.374670430939404e-07,
"loss": 0.0063,
"step": 643
},
{
"epoch": 4.13,
"grad_norm": 0.6704283408089174,
"learning_rate": 6.284830017084488e-07,
"loss": 0.0074,
"step": 644
},
{
"epoch": 4.13,
"grad_norm": 0.5795149303845601,
"learning_rate": 6.195564693722027e-07,
"loss": 0.014,
"step": 645
},
{
"epoch": 4.14,
"grad_norm": 0.6404441767465421,
"learning_rate": 6.106876248605299e-07,
"loss": 0.0126,
"step": 646
},
{
"epoch": 4.15,
"grad_norm": 0.49124988946928483,
"learning_rate": 6.018766457934177e-07,
"loss": 0.0054,
"step": 647
},
{
"epoch": 4.15,
"grad_norm": 0.6895422647706185,
"learning_rate": 5.931237086319592e-07,
"loss": 0.009,
"step": 648
},
{
"epoch": 4.16,
"grad_norm": 0.7133260869582704,
"learning_rate": 5.844289886748196e-07,
"loss": 0.0136,
"step": 649
},
{
"epoch": 4.17,
"grad_norm": 0.6485862251152251,
"learning_rate": 5.757926600547231e-07,
"loss": 0.0076,
"step": 650
},
{
"epoch": 4.17,
"grad_norm": 0.7007270882813788,
"learning_rate": 5.672148957349661e-07,
"loss": 0.0095,
"step": 651
},
{
"epoch": 4.18,
"grad_norm": 0.5453225501286796,
"learning_rate": 5.586958675059548e-07,
"loss": 0.0137,
"step": 652
},
{
"epoch": 4.19,
"grad_norm": 0.46759765618604715,
"learning_rate": 5.502357459817639e-07,
"loss": 0.01,
"step": 653
},
{
"epoch": 4.19,
"grad_norm": 0.40298536871030216,
"learning_rate": 5.418347005967189e-07,
"loss": 0.0074,
"step": 654
},
{
"epoch": 4.2,
"grad_norm": 0.7139987824205981,
"learning_rate": 5.334928996020013e-07,
"loss": 0.0109,
"step": 655
},
{
"epoch": 4.21,
"grad_norm": 0.4688415136809408,
"learning_rate": 5.252105100622848e-07,
"loss": 0.0077,
"step": 656
},
{
"epoch": 4.21,
"grad_norm": 0.5717898546159872,
"learning_rate": 5.169876978523828e-07,
"loss": 0.0105,
"step": 657
},
{
"epoch": 4.22,
"grad_norm": 0.5863304725088486,
"learning_rate": 5.088246276539292e-07,
"loss": 0.0127,
"step": 658
},
{
"epoch": 4.22,
"grad_norm": 0.748734639494948,
"learning_rate": 5.0072146295208e-07,
"loss": 0.0083,
"step": 659
},
{
"epoch": 4.23,
"grad_norm": 0.5106657629410736,
"learning_rate": 4.926783660322411e-07,
"loss": 0.0093,
"step": 660
},
{
"epoch": 4.24,
"grad_norm": 0.5249256431071327,
"learning_rate": 4.846954979768149e-07,
"loss": 0.009,
"step": 661
},
{
"epoch": 4.24,
"grad_norm": 0.5837936537112237,
"learning_rate": 4.7677301866197455e-07,
"loss": 0.0111,
"step": 662
},
{
"epoch": 4.25,
"grad_norm": 0.5436208344015698,
"learning_rate": 4.6891108675446453e-07,
"loss": 0.0081,
"step": 663
},
{
"epoch": 4.26,
"grad_norm": 0.45668925332803584,
"learning_rate": 4.611098597084226e-07,
"loss": 0.0086,
"step": 664
},
{
"epoch": 4.26,
"grad_norm": 0.5234369553141341,
"learning_rate": 4.533694937622227e-07,
"loss": 0.0098,
"step": 665
},
{
"epoch": 4.27,
"grad_norm": 0.43090051877682023,
"learning_rate": 4.456901439353499e-07,
"loss": 0.0069,
"step": 666
},
{
"epoch": 4.28,
"grad_norm": 0.5065114847080426,
"learning_rate": 4.3807196402529535e-07,
"loss": 0.0078,
"step": 667
},
{
"epoch": 4.28,
"grad_norm": 0.3681825929403806,
"learning_rate": 4.3051510660447336e-07,
"loss": 0.0046,
"step": 668
},
{
"epoch": 4.29,
"grad_norm": 0.5198397329641357,
"learning_rate": 4.2301972301716934e-07,
"loss": 0.0091,
"step": 669
},
{
"epoch": 4.29,
"grad_norm": 0.9754162271751636,
"learning_rate": 4.155859633765044e-07,
"loss": 0.0081,
"step": 670
},
{
"epoch": 4.3,
"grad_norm": 0.3500513380364396,
"learning_rate": 4.0821397656143503e-07,
"loss": 0.005,
"step": 671
},
{
"epoch": 4.31,
"grad_norm": 0.5822401487852554,
"learning_rate": 4.009039102137657e-07,
"loss": 0.0043,
"step": 672
},
{
"epoch": 4.31,
"grad_norm": 0.41873061343417606,
"learning_rate": 3.9365591073519387e-07,
"loss": 0.0078,
"step": 673
},
{
"epoch": 4.32,
"grad_norm": 0.49737057792780415,
"learning_rate": 3.8647012328438085e-07,
"loss": 0.0063,
"step": 674
},
{
"epoch": 4.33,
"grad_norm": 0.5416348114415329,
"learning_rate": 3.793466917740402e-07,
"loss": 0.0069,
"step": 675
},
{
"epoch": 4.33,
"grad_norm": 0.5946803237802584,
"learning_rate": 3.7228575886805744e-07,
"loss": 0.0101,
"step": 676
},
{
"epoch": 4.34,
"grad_norm": 0.5792466423496542,
"learning_rate": 3.6528746597863283e-07,
"loss": 0.0091,
"step": 677
},
{
"epoch": 4.35,
"grad_norm": 0.5149239013215101,
"learning_rate": 3.583519532634516e-07,
"loss": 0.0094,
"step": 678
},
{
"epoch": 4.35,
"grad_norm": 0.5108129659609797,
"learning_rate": 3.514793596228702e-07,
"loss": 0.0079,
"step": 679
},
{
"epoch": 4.36,
"grad_norm": 0.5884548387452648,
"learning_rate": 3.44669822697144e-07,
"loss": 0.0065,
"step": 680
},
{
"epoch": 4.37,
"grad_norm": 0.45883688525931454,
"learning_rate": 3.3792347886366265e-07,
"loss": 0.0086,
"step": 681
},
{
"epoch": 4.37,
"grad_norm": 0.5151424364772744,
"learning_rate": 3.31240463234221e-07,
"loss": 0.0075,
"step": 682
},
{
"epoch": 4.38,
"grad_norm": 0.5951741308945779,
"learning_rate": 3.2462090965231767e-07,
"loss": 0.0055,
"step": 683
},
{
"epoch": 4.38,
"grad_norm": 0.4910893940517418,
"learning_rate": 3.180649506904667e-07,
"loss": 0.0094,
"step": 684
},
{
"epoch": 4.39,
"grad_norm": 0.3032369959343678,
"learning_rate": 3.1157271764755085e-07,
"loss": 0.0056,
"step": 685
},
{
"epoch": 4.4,
"grad_norm": 0.46113289164382765,
"learning_rate": 3.0514434054618216e-07,
"loss": 0.0063,
"step": 686
},
{
"epoch": 4.4,
"grad_norm": 0.649482604619634,
"learning_rate": 2.987799481301091e-07,
"loss": 0.0062,
"step": 687
},
{
"epoch": 4.41,
"grad_norm": 0.56089990875009,
"learning_rate": 2.924796678616297e-07,
"loss": 0.0069,
"step": 688
},
{
"epoch": 4.42,
"grad_norm": 0.4098789353635756,
"learning_rate": 2.862436259190414e-07,
"loss": 0.0072,
"step": 689
},
{
"epoch": 4.42,
"grad_norm": 0.42421611582010255,
"learning_rate": 2.800719471941152e-07,
"loss": 0.0055,
"step": 690
},
{
"epoch": 4.43,
"grad_norm": 0.46243164776878504,
"learning_rate": 2.739647552895949e-07,
"loss": 0.0101,
"step": 691
},
{
"epoch": 4.44,
"grad_norm": 0.506707466448847,
"learning_rate": 2.6792217251671744e-07,
"loss": 0.0049,
"step": 692
},
{
"epoch": 4.44,
"grad_norm": 0.3925167921918842,
"learning_rate": 2.619443198927677e-07,
"loss": 0.0072,
"step": 693
},
{
"epoch": 4.45,
"grad_norm": 0.3420002956507376,
"learning_rate": 2.5603131713865374e-07,
"loss": 0.0076,
"step": 694
},
{
"epoch": 4.46,
"grad_norm": 0.29143440480895755,
"learning_rate": 2.50183282676508e-07,
"loss": 0.0053,
"step": 695
},
{
"epoch": 4.46,
"grad_norm": 0.35335225088438177,
"learning_rate": 2.444003336273163e-07,
"loss": 0.0075,
"step": 696
},
{
"epoch": 4.47,
"grad_norm": 0.7377895380554025,
"learning_rate": 2.3868258580857164e-07,
"loss": 0.0079,
"step": 697
},
{
"epoch": 4.47,
"grad_norm": 0.3732383566357089,
"learning_rate": 2.3303015373195713e-07,
"loss": 0.0079,
"step": 698
},
{
"epoch": 4.48,
"grad_norm": 0.30406904337344,
"learning_rate": 2.2744315060104846e-07,
"loss": 0.0058,
"step": 699
},
{
"epoch": 4.49,
"grad_norm": 0.2979113383173518,
"learning_rate": 2.2192168830904963e-07,
"loss": 0.0046,
"step": 700
},
{
"epoch": 4.49,
"grad_norm": 0.26022505864795753,
"learning_rate": 2.1646587743655287e-07,
"loss": 0.0045,
"step": 701
},
{
"epoch": 4.5,
"grad_norm": 0.38203960053911995,
"learning_rate": 2.1107582724932088e-07,
"loss": 0.0067,
"step": 702
},
{
"epoch": 4.51,
"grad_norm": 0.4611969431592941,
"learning_rate": 2.0575164569610016e-07,
"loss": 0.0058,
"step": 703
},
{
"epoch": 4.51,
"grad_norm": 0.48964538570578786,
"learning_rate": 2.0049343940645935e-07,
"loss": 0.0058,
"step": 704
},
{
"epoch": 4.52,
"grad_norm": 0.3748977896181229,
"learning_rate": 1.953013136886541e-07,
"loss": 0.0079,
"step": 705
},
{
"epoch": 4.53,
"grad_norm": 0.3684735348004529,
"learning_rate": 1.901753725275166e-07,
"loss": 0.0034,
"step": 706
},
{
"epoch": 4.53,
"grad_norm": 0.45813437926806905,
"learning_rate": 1.8511571858237357e-07,
"loss": 0.0049,
"step": 707
},
{
"epoch": 4.54,
"grad_norm": 0.38090807447648417,
"learning_rate": 1.801224531849908e-07,
"loss": 0.0081,
"step": 708
},
{
"epoch": 4.54,
"grad_norm": 0.7300382830511903,
"learning_rate": 1.7519567633754352e-07,
"loss": 0.0067,
"step": 709
},
{
"epoch": 4.55,
"grad_norm": 0.4521252420340857,
"learning_rate": 1.70335486710614e-07,
"loss": 0.0107,
"step": 710
},
{
"epoch": 4.56,
"grad_norm": 0.3290046038461478,
"learning_rate": 1.6554198164121265e-07,
"loss": 0.0056,
"step": 711
},
{
"epoch": 4.56,
"grad_norm": 0.23389758856912887,
"learning_rate": 1.6081525713083428e-07,
"loss": 0.004,
"step": 712
},
{
"epoch": 4.57,
"grad_norm": 0.4187101993019587,
"learning_rate": 1.561554078435296e-07,
"loss": 0.0062,
"step": 713
},
{
"epoch": 4.58,
"grad_norm": 0.43647005515014636,
"learning_rate": 1.5156252710401207e-07,
"loss": 0.0076,
"step": 714
},
{
"epoch": 4.58,
"grad_norm": 0.3727002889542965,
"learning_rate": 1.4703670689578884e-07,
"loss": 0.0066,
"step": 715
},
{
"epoch": 4.59,
"grad_norm": 0.39326688282125816,
"learning_rate": 1.4257803785931926e-07,
"loss": 0.0069,
"step": 716
},
{
"epoch": 4.6,
"grad_norm": 0.31355113796472983,
"learning_rate": 1.3818660929019717e-07,
"loss": 0.0058,
"step": 717
},
{
"epoch": 4.6,
"grad_norm": 0.3582549090050492,
"learning_rate": 1.3386250913736408e-07,
"loss": 0.0055,
"step": 718
},
{
"epoch": 4.61,
"grad_norm": 0.3557691839899451,
"learning_rate": 1.296058240013491e-07,
"loss": 0.007,
"step": 719
},
{
"epoch": 4.62,
"grad_norm": 0.27232509097291063,
"learning_rate": 1.2541663913253191e-07,
"loss": 0.0054,
"step": 720
},
{
"epoch": 4.62,
"grad_norm": 0.29919984285618584,
"learning_rate": 1.2129503842943645e-07,
"loss": 0.0073,
"step": 721
},
{
"epoch": 4.63,
"grad_norm": 0.32031918250676045,
"learning_rate": 1.1724110443705115e-07,
"loss": 0.0062,
"step": 722
},
{
"epoch": 4.63,
"grad_norm": 0.32462046027184965,
"learning_rate": 1.1325491834517676e-07,
"loss": 0.0069,
"step": 723
},
{
"epoch": 4.64,
"grad_norm": 0.3703162253466964,
"learning_rate": 1.0933655998679653e-07,
"loss": 0.0049,
"step": 724
},
{
"epoch": 4.65,
"grad_norm": 0.3830835431373469,
"learning_rate": 1.0548610783648199e-07,
"loss": 0.0077,
"step": 725
},
{
"epoch": 4.65,
"grad_norm": 0.3042237914246762,
"learning_rate": 1.0170363900881795e-07,
"loss": 0.0039,
"step": 726
},
{
"epoch": 4.66,
"grad_norm": 0.2615931115864544,
"learning_rate": 9.798922925685994e-08,
"loss": 0.0042,
"step": 727
},
{
"epoch": 4.67,
"grad_norm": 0.35403759124068945,
"learning_rate": 9.434295297061668e-08,
"loss": 0.0083,
"step": 728
},
{
"epoch": 4.67,
"grad_norm": 0.44091888309449573,
"learning_rate": 9.076488317555886e-08,
"loss": 0.0058,
"step": 729
},
{
"epoch": 4.68,
"grad_norm": 0.405216970933536,
"learning_rate": 8.725509153115918e-08,
"loss": 0.0092,
"step": 730
},
{
"epoch": 4.69,
"grad_norm": 0.28509534429106465,
"learning_rate": 8.38136483294546e-08,
"loss": 0.0037,
"step": 731
},
{
"epoch": 4.69,
"grad_norm": 0.2940024574327348,
"learning_rate": 8.044062249364048e-08,
"loss": 0.0054,
"step": 732
},
{
"epoch": 4.7,
"grad_norm": 0.36361039302741816,
"learning_rate": 7.713608157668921e-08,
"loss": 0.0081,
"step": 733
},
{
"epoch": 4.71,
"grad_norm": 0.39672091027135914,
"learning_rate": 7.390009175999835e-08,
"loss": 0.0094,
"step": 734
},
{
"epoch": 4.71,
"grad_norm": 0.40191807872522295,
"learning_rate": 7.073271785206314e-08,
"loss": 0.0078,
"step": 735
},
{
"epoch": 4.72,
"grad_norm": 0.2901495174113024,
"learning_rate": 6.763402328718116e-08,
"loss": 0.0052,
"step": 736
},
{
"epoch": 4.72,
"grad_norm": 0.2532854567420693,
"learning_rate": 6.460407012417918e-08,
"loss": 0.004,
"step": 737
},
{
"epoch": 4.73,
"grad_norm": 0.3225727935088736,
"learning_rate": 6.164291904517333e-08,
"loss": 0.0083,
"step": 738
},
{
"epoch": 4.74,
"grad_norm": 0.3069170665460628,
"learning_rate": 5.875062935435121e-08,
"loss": 0.0052,
"step": 739
},
{
"epoch": 4.74,
"grad_norm": 0.2612538982055557,
"learning_rate": 5.592725897678446e-08,
"loss": 0.0057,
"step": 740
},
{
"epoch": 4.75,
"grad_norm": 0.24090688702155097,
"learning_rate": 5.3172864457271926e-08,
"loss": 0.0035,
"step": 741
},
{
"epoch": 4.76,
"grad_norm": 0.36351641612749136,
"learning_rate": 5.048750095920151e-08,
"loss": 0.0062,
"step": 742
},
{
"epoch": 4.76,
"grad_norm": 0.43612520351048434,
"learning_rate": 4.787122226345014e-08,
"loss": 0.005,
"step": 743
},
{
"epoch": 4.77,
"grad_norm": 0.29747843114335953,
"learning_rate": 4.532408076730504e-08,
"loss": 0.0057,
"step": 744
},
{
"epoch": 4.78,
"grad_norm": 0.3032920639088825,
"learning_rate": 4.2846127483414206e-08,
"loss": 0.0061,
"step": 745
},
{
"epoch": 4.78,
"grad_norm": 0.33781142844587425,
"learning_rate": 4.043741203876483e-08,
"loss": 0.0048,
"step": 746
},
{
"epoch": 4.79,
"grad_norm": 0.2791313716105004,
"learning_rate": 3.80979826736893e-08,
"loss": 0.0054,
"step": 747
},
{
"epoch": 4.79,
"grad_norm": 0.424062860990937,
"learning_rate": 3.58278862409e-08,
"loss": 0.0076,
"step": 748
},
{
"epoch": 4.8,
"grad_norm": 0.35602806799069403,
"learning_rate": 3.3627168204549306e-08,
"loss": 0.0044,
"step": 749
},
{
"epoch": 4.81,
"grad_norm": 0.2967466134268081,
"learning_rate": 3.1495872639320357e-08,
"loss": 0.0059,
"step": 750
},
{
"epoch": 4.81,
"grad_norm": 0.29343066100283227,
"learning_rate": 2.9434042229544543e-08,
"loss": 0.0067,
"step": 751
},
{
"epoch": 4.82,
"grad_norm": 0.25925792441623674,
"learning_rate": 2.7441718268344737e-08,
"loss": 0.0055,
"step": 752
},
{
"epoch": 4.83,
"grad_norm": 0.33458705494389285,
"learning_rate": 2.5518940656811095e-08,
"loss": 0.0088,
"step": 753
},
{
"epoch": 4.83,
"grad_norm": 0.32973000192844754,
"learning_rate": 2.3665747903199418e-08,
"loss": 0.0076,
"step": 754
},
{
"epoch": 4.84,
"grad_norm": 0.3978241736214976,
"learning_rate": 2.1882177122162173e-08,
"loss": 0.0083,
"step": 755
},
{
"epoch": 4.85,
"grad_norm": 0.3846684615361078,
"learning_rate": 2.0168264034002404e-08,
"loss": 0.0107,
"step": 756
},
{
"epoch": 4.85,
"grad_norm": 0.3281597086771846,
"learning_rate": 1.8524042963961095e-08,
"loss": 0.0057,
"step": 757
},
{
"epoch": 4.86,
"grad_norm": 0.3098674134494085,
"learning_rate": 1.6949546841528607e-08,
"loss": 0.0055,
"step": 758
},
{
"epoch": 4.87,
"grad_norm": 0.3827204692079133,
"learning_rate": 1.544480719978447e-08,
"loss": 0.0086,
"step": 759
},
{
"epoch": 4.87,
"grad_norm": 0.332400995275947,
"learning_rate": 1.4009854174767521e-08,
"loss": 0.0093,
"step": 760
},
{
"epoch": 4.88,
"grad_norm": 0.29676359757111426,
"learning_rate": 1.2644716504870091e-08,
"loss": 0.0073,
"step": 761
},
{
"epoch": 4.88,
"grad_norm": 0.4005292260275929,
"learning_rate": 1.1349421530265246e-08,
"loss": 0.0094,
"step": 762
},
{
"epoch": 4.89,
"grad_norm": 0.3811448231039689,
"learning_rate": 1.0123995192356183e-08,
"loss": 0.0099,
"step": 763
},
{
"epoch": 4.9,
"grad_norm": 0.3593225301938938,
"learning_rate": 8.968462033259405e-09,
"loss": 0.0086,
"step": 764
},
{
"epoch": 4.9,
"grad_norm": 0.3738117335803207,
"learning_rate": 7.882845195312016e-09,
"loss": 0.0052,
"step": 765
},
{
"epoch": 4.91,
"grad_norm": 0.2792515305128813,
"learning_rate": 6.8671664206073625e-09,
"loss": 0.0048,
"step": 766
},
{
"epoch": 4.92,
"grad_norm": 0.325832261539929,
"learning_rate": 5.921446050561386e-09,
"loss": 0.0083,
"step": 767
},
{
"epoch": 4.92,
"grad_norm": 0.3060573357618961,
"learning_rate": 5.0457030255038334e-09,
"loss": 0.0043,
"step": 768
},
{
"epoch": 4.93,
"grad_norm": 0.29897144167044903,
"learning_rate": 4.239954884299401e-09,
"loss": 0.0068,
"step": 769
},
{
"epoch": 4.94,
"grad_norm": 0.33960368756040765,
"learning_rate": 3.5042177639972304e-09,
"loss": 0.0086,
"step": 770
},
{
"epoch": 4.94,
"grad_norm": 0.31535432780159894,
"learning_rate": 2.838506399506446e-09,
"loss": 0.007,
"step": 771
},
{
"epoch": 4.95,
"grad_norm": 0.4802190663449603,
"learning_rate": 2.2428341233012294e-09,
"loss": 0.0199,
"step": 772
},
{
"epoch": 4.96,
"grad_norm": 0.42217488555827554,
"learning_rate": 1.7172128651554152e-09,
"loss": 0.0108,
"step": 773
},
{
"epoch": 4.96,
"grad_norm": 0.39472643041620875,
"learning_rate": 1.2616531519011874e-09,
"loss": 0.0081,
"step": 774
},
{
"epoch": 4.97,
"grad_norm": 0.29963966680365745,
"learning_rate": 8.761641072196346e-10,
"loss": 0.0072,
"step": 775
},
{
"epoch": 4.97,
"grad_norm": 0.3767724810022492,
"learning_rate": 5.607534514585066e-10,
"loss": 0.0057,
"step": 776
},
{
"epoch": 4.98,
"grad_norm": 0.2953142117019799,
"learning_rate": 3.1542750147639517e-10,
"loss": 0.0055,
"step": 777
},
{
"epoch": 4.99,
"grad_norm": 0.28636138519269905,
"learning_rate": 1.401911705168346e-10,
"loss": 0.0056,
"step": 778
},
{
"epoch": 4.99,
"grad_norm": 0.2993184966019565,
"learning_rate": 3.5047968109214176e-11,
"loss": 0.0063,
"step": 779
},
{
"epoch": 5.0,
"grad_norm": 0.3531925803337977,
"learning_rate": 0.0,
"loss": 0.0076,
"step": 780
},
{
"epoch": 5.0,
"step": 780,
"total_flos": 0.0,
"train_loss": 0.17006511208034145,
"train_runtime": 1554.1821,
"train_samples_per_second": 16.086,
"train_steps_per_second": 0.502
}
],
"logging_steps": 1.0,
"max_steps": 780,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 1000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": false,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 0.0,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}