diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,12152 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.9985559566787003, + "eval_steps": 500, + "global_step": 1730, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0011552346570397113, + "grad_norm": 0.717101514339447, + "learning_rate": 0.0, + "loss": 1.5901, + "step": 1 + }, + { + "epoch": 0.0023104693140794225, + "grad_norm": 0.7195026874542236, + "learning_rate": 5.017166594399687e-07, + "loss": 1.583, + "step": 2 + }, + { + "epoch": 0.0034657039711191336, + "grad_norm": 0.7274531722068787, + "learning_rate": 7.952020911994375e-07, + "loss": 1.6153, + "step": 3 + }, + { + "epoch": 0.004620938628158845, + "grad_norm": 0.7189211249351501, + "learning_rate": 1.0034333188799374e-06, + "loss": 1.5939, + "step": 4 + }, + { + "epoch": 0.005776173285198556, + "grad_norm": 0.7074140906333923, + "learning_rate": 1.164950007226698e-06, + "loss": 1.6126, + "step": 5 + }, + { + "epoch": 0.006931407942238267, + "grad_norm": 0.7060695886611938, + "learning_rate": 1.2969187506394062e-06, + "loss": 1.597, + "step": 6 + }, + { + "epoch": 0.008086642599277978, + "grad_norm": 0.693950891494751, + "learning_rate": 1.4084967333570947e-06, + "loss": 1.584, + "step": 7 + }, + { + "epoch": 0.00924187725631769, + "grad_norm": 0.6204413771629333, + "learning_rate": 1.5051499783199062e-06, + "loss": 1.5333, + "step": 8 + }, + { + "epoch": 0.0103971119133574, + "grad_norm": 0.5834112763404846, + "learning_rate": 1.590404182398875e-06, + "loss": 1.5173, + "step": 9 + }, + { + "epoch": 0.011552346570397111, + "grad_norm": 0.5479596257209778, + "learning_rate": 1.666666666666667e-06, + "loss": 1.4693, + "step": 10 + }, + { + "epoch": 0.012707581227436824, + "grad_norm": 0.3792664110660553, + "learning_rate": 1.7356544752637086e-06, + "loss": 1.3613, + "step": 11 + }, + { + "epoch": 0.013862815884476534, + "grad_norm": 0.346934974193573, + "learning_rate": 1.798635410079375e-06, + "loss": 1.3096, + "step": 12 + }, + { + "epoch": 0.015018050541516245, + "grad_norm": 0.30829983949661255, + "learning_rate": 1.8565722538447281e-06, + "loss": 1.3096, + "step": 13 + }, + { + "epoch": 0.016173285198555955, + "grad_norm": 0.29539182782173157, + "learning_rate": 1.9102133927970633e-06, + "loss": 1.3067, + "step": 14 + }, + { + "epoch": 0.017328519855595668, + "grad_norm": 0.17869696021080017, + "learning_rate": 1.960152098426136e-06, + "loss": 1.2191, + "step": 15 + }, + { + "epoch": 0.01848375451263538, + "grad_norm": 0.3124949634075165, + "learning_rate": 2.0068666377598747e-06, + "loss": 1.2029, + "step": 16 + }, + { + "epoch": 0.01963898916967509, + "grad_norm": 0.33391040563583374, + "learning_rate": 2.0507482022971233e-06, + "loss": 1.2076, + "step": 17 + }, + { + "epoch": 0.0207942238267148, + "grad_norm": 0.3112436830997467, + "learning_rate": 2.0921208418388438e-06, + "loss": 1.1862, + "step": 18 + }, + { + "epoch": 0.021949458483754514, + "grad_norm": 0.29002705216407776, + "learning_rate": 2.1312560015880486e-06, + "loss": 1.1844, + "step": 19 + }, + { + "epoch": 0.023104693140794223, + "grad_norm": 0.2618213891983032, + "learning_rate": 2.1683833261066357e-06, + "loss": 1.1763, + "step": 20 + }, + { + "epoch": 0.024259927797833935, + "grad_norm": 0.20633897185325623, + "learning_rate": 2.2036988245565326e-06, + "loss": 1.1683, + "step": 21 + }, + { + "epoch": 0.025415162454873647, + "grad_norm": 0.17382092773914337, + "learning_rate": 2.2373711347036773e-06, + "loss": 1.1119, + "step": 22 + }, + { + "epoch": 0.026570397111913356, + "grad_norm": 0.12656044960021973, + "learning_rate": 2.269546393362655e-06, + "loss": 1.1001, + "step": 23 + }, + { + "epoch": 0.02772563176895307, + "grad_norm": 0.1064457818865776, + "learning_rate": 2.3003520695193436e-06, + "loss": 1.1154, + "step": 24 + }, + { + "epoch": 0.02888086642599278, + "grad_norm": 0.09630957245826721, + "learning_rate": 2.329900014453396e-06, + "loss": 1.0793, + "step": 25 + }, + { + "epoch": 0.03003610108303249, + "grad_norm": 0.10916559398174286, + "learning_rate": 2.358288913284697e-06, + "loss": 1.0672, + "step": 26 + }, + { + "epoch": 0.031191335740072202, + "grad_norm": 0.11219058930873871, + "learning_rate": 2.385606273598312e-06, + "loss": 1.0833, + "step": 27 + }, + { + "epoch": 0.03234657039711191, + "grad_norm": 0.11489821970462799, + "learning_rate": 2.4119300522370323e-06, + "loss": 1.0795, + "step": 28 + }, + { + "epoch": 0.03350180505415162, + "grad_norm": 0.11093555390834808, + "learning_rate": 2.4373299964982607e-06, + "loss": 1.0599, + "step": 29 + }, + { + "epoch": 0.034657039711191336, + "grad_norm": 0.10726247727870941, + "learning_rate": 2.4618687578661045e-06, + "loss": 1.066, + "step": 30 + }, + { + "epoch": 0.03581227436823105, + "grad_norm": 0.0960482507944107, + "learning_rate": 2.4856028230571215e-06, + "loss": 1.0332, + "step": 31 + }, + { + "epoch": 0.03696750902527076, + "grad_norm": 0.09282615780830383, + "learning_rate": 2.5085832971998437e-06, + "loss": 1.0566, + "step": 32 + }, + { + "epoch": 0.03812274368231047, + "grad_norm": 0.0818951278924942, + "learning_rate": 2.530856566463146e-06, + "loss": 1.0351, + "step": 33 + }, + { + "epoch": 0.03927797833935018, + "grad_norm": 0.07113130390644073, + "learning_rate": 2.5524648617370923e-06, + "loss": 1.0164, + "step": 34 + }, + { + "epoch": 0.04043321299638989, + "grad_norm": 0.062365710735321045, + "learning_rate": 2.5734467405837933e-06, + "loss": 1.0125, + "step": 35 + }, + { + "epoch": 0.0415884476534296, + "grad_norm": 0.057508960366249084, + "learning_rate": 2.5938375012788124e-06, + "loss": 1.0392, + "step": 36 + }, + { + "epoch": 0.042743682310469315, + "grad_norm": 0.0539795346558094, + "learning_rate": 2.6136695401116585e-06, + "loss": 1.0244, + "step": 37 + }, + { + "epoch": 0.04389891696750903, + "grad_norm": 0.053800784051418304, + "learning_rate": 2.632972661028017e-06, + "loss": 0.9972, + "step": 38 + }, + { + "epoch": 0.04505415162454874, + "grad_norm": 0.05475891754031181, + "learning_rate": 2.6517743450441657e-06, + "loss": 1.0142, + "step": 39 + }, + { + "epoch": 0.046209386281588445, + "grad_norm": 0.0571669340133667, + "learning_rate": 2.6700999855466042e-06, + "loss": 0.9995, + "step": 40 + }, + { + "epoch": 0.04736462093862816, + "grad_norm": 0.05862262472510338, + "learning_rate": 2.687973094532893e-06, + "loss": 0.998, + "step": 41 + }, + { + "epoch": 0.04851985559566787, + "grad_norm": 0.0590708963572979, + "learning_rate": 2.705415483996501e-06, + "loss": 0.9798, + "step": 42 + }, + { + "epoch": 0.04967509025270758, + "grad_norm": 0.056892745196819305, + "learning_rate": 2.722447425965978e-06, + "loss": 0.9879, + "step": 43 + }, + { + "epoch": 0.050830324909747295, + "grad_norm": 0.0541134737432003, + "learning_rate": 2.739087794143646e-06, + "loss": 0.9722, + "step": 44 + }, + { + "epoch": 0.05198555956678701, + "grad_norm": 0.048609230667352676, + "learning_rate": 2.7553541896255733e-06, + "loss": 0.9491, + "step": 45 + }, + { + "epoch": 0.05314079422382671, + "grad_norm": 0.04965611547231674, + "learning_rate": 2.771263052802624e-06, + "loss": 1.0037, + "step": 46 + }, + { + "epoch": 0.054296028880866425, + "grad_norm": 0.043069027364254, + "learning_rate": 2.7868297632261957e-06, + "loss": 0.971, + "step": 47 + }, + { + "epoch": 0.05545126353790614, + "grad_norm": 0.03820377215743065, + "learning_rate": 2.8020687289593126e-06, + "loss": 1.0084, + "step": 48 + }, + { + "epoch": 0.05660649819494585, + "grad_norm": 0.03588235378265381, + "learning_rate": 2.8169934667141895e-06, + "loss": 0.9507, + "step": 49 + }, + { + "epoch": 0.05776173285198556, + "grad_norm": 0.03598296642303467, + "learning_rate": 2.8316166738933647e-06, + "loss": 0.973, + "step": 50 + }, + { + "epoch": 0.058916967509025274, + "grad_norm": 0.0384756475687027, + "learning_rate": 2.845950293496561e-06, + "loss": 0.9758, + "step": 51 + }, + { + "epoch": 0.06007220216606498, + "grad_norm": 0.03798473998904228, + "learning_rate": 2.8600055727246655e-06, + "loss": 0.9991, + "step": 52 + }, + { + "epoch": 0.06122743682310469, + "grad_norm": 0.039415981620550156, + "learning_rate": 2.8737931160013154e-06, + "loss": 0.9661, + "step": 53 + }, + { + "epoch": 0.062382671480144404, + "grad_norm": 0.04056290537118912, + "learning_rate": 2.887322933038281e-06, + "loss": 0.9773, + "step": 54 + }, + { + "epoch": 0.06353790613718412, + "grad_norm": 0.03856893256306648, + "learning_rate": 2.900604482490407e-06, + "loss": 0.9644, + "step": 55 + }, + { + "epoch": 0.06469314079422382, + "grad_norm": 0.03817951679229736, + "learning_rate": 2.9136467116770013e-06, + "loss": 0.9701, + "step": 56 + }, + { + "epoch": 0.06584837545126354, + "grad_norm": 0.03559865057468414, + "learning_rate": 2.926458092787486e-06, + "loss": 0.9314, + "step": 57 + }, + { + "epoch": 0.06700361010830325, + "grad_norm": 0.03424916788935661, + "learning_rate": 2.9390466559382293e-06, + "loss": 0.9489, + "step": 58 + }, + { + "epoch": 0.06815884476534297, + "grad_norm": 0.03320642188191414, + "learning_rate": 2.951420019403574e-06, + "loss": 0.9697, + "step": 59 + }, + { + "epoch": 0.06931407942238267, + "grad_norm": 0.031773000955581665, + "learning_rate": 2.963585417306073e-06, + "loss": 0.923, + "step": 60 + }, + { + "epoch": 0.07046931407942238, + "grad_norm": 0.0303028617054224, + "learning_rate": 2.9755497250179457e-06, + "loss": 0.9448, + "step": 61 + }, + { + "epoch": 0.0716245487364621, + "grad_norm": 0.029582129791378975, + "learning_rate": 2.98731948249709e-06, + "loss": 0.9521, + "step": 62 + }, + { + "epoch": 0.0727797833935018, + "grad_norm": 0.029476916417479515, + "learning_rate": 2.9989009157559695e-06, + "loss": 0.9429, + "step": 63 + }, + { + "epoch": 0.07393501805054152, + "grad_norm": 0.02916543185710907, + "learning_rate": 3.0102999566398123e-06, + "loss": 0.9641, + "step": 64 + }, + { + "epoch": 0.07509025270758123, + "grad_norm": 0.029549594968557358, + "learning_rate": 3.021522261071426e-06, + "loss": 0.9555, + "step": 65 + }, + { + "epoch": 0.07624548736462095, + "grad_norm": 0.03075719065964222, + "learning_rate": 3.0325732259031143e-06, + "loss": 0.9531, + "step": 66 + }, + { + "epoch": 0.07740072202166065, + "grad_norm": 0.0289381705224514, + "learning_rate": 3.0434580045013773e-06, + "loss": 0.9358, + "step": 67 + }, + { + "epoch": 0.07855595667870036, + "grad_norm": 0.03085014969110489, + "learning_rate": 3.054181521177061e-06, + "loss": 0.9496, + "step": 68 + }, + { + "epoch": 0.07971119133574008, + "grad_norm": 0.02944289892911911, + "learning_rate": 3.064748484562093e-06, + "loss": 0.931, + "step": 69 + }, + { + "epoch": 0.08086642599277978, + "grad_norm": 0.02958507277071476, + "learning_rate": 3.075163400023762e-06, + "loss": 0.9674, + "step": 70 + }, + { + "epoch": 0.0820216606498195, + "grad_norm": 0.029631255194544792, + "learning_rate": 3.085430581198459e-06, + "loss": 0.9504, + "step": 71 + }, + { + "epoch": 0.0831768953068592, + "grad_norm": 0.028514275327324867, + "learning_rate": 3.095554160718781e-06, + "loss": 0.9513, + "step": 72 + }, + { + "epoch": 0.08433212996389891, + "grad_norm": 0.02768518030643463, + "learning_rate": 3.1055381002007602e-06, + "loss": 0.9152, + "step": 73 + }, + { + "epoch": 0.08548736462093863, + "grad_norm": 0.028020448982715607, + "learning_rate": 3.1153861995516275e-06, + "loss": 0.9247, + "step": 74 + }, + { + "epoch": 0.08664259927797834, + "grad_norm": 0.02758488804101944, + "learning_rate": 3.1251021056528336e-06, + "loss": 0.9285, + "step": 75 + }, + { + "epoch": 0.08779783393501805, + "grad_norm": 0.02713642828166485, + "learning_rate": 3.1346893204679857e-06, + "loss": 0.928, + "step": 76 + }, + { + "epoch": 0.08895306859205776, + "grad_norm": 0.026601964607834816, + "learning_rate": 3.1441512086208035e-06, + "loss": 0.9135, + "step": 77 + }, + { + "epoch": 0.09010830324909748, + "grad_norm": 0.02707557938992977, + "learning_rate": 3.1534910044841343e-06, + "loss": 0.9306, + "step": 78 + }, + { + "epoch": 0.09126353790613718, + "grad_norm": 0.027205491438508034, + "learning_rate": 3.1627118188174026e-06, + "loss": 0.9475, + "step": 79 + }, + { + "epoch": 0.09241877256317689, + "grad_norm": 0.02767670899629593, + "learning_rate": 3.171816644986573e-06, + "loss": 0.9445, + "step": 80 + }, + { + "epoch": 0.09357400722021661, + "grad_norm": 0.027132879942655563, + "learning_rate": 3.18080836479775e-06, + "loss": 0.9388, + "step": 81 + }, + { + "epoch": 0.09472924187725631, + "grad_norm": 0.07792849093675613, + "learning_rate": 3.1896897539728615e-06, + "loss": 0.9097, + "step": 82 + }, + { + "epoch": 0.09588447653429603, + "grad_norm": 0.026650305837392807, + "learning_rate": 3.1984634872934573e-06, + "loss": 0.9216, + "step": 83 + }, + { + "epoch": 0.09703971119133574, + "grad_norm": 0.02636835351586342, + "learning_rate": 3.2071321434364693e-06, + "loss": 0.9397, + "step": 84 + }, + { + "epoch": 0.09819494584837545, + "grad_norm": 0.025988014414906502, + "learning_rate": 3.2156982095238214e-06, + "loss": 0.9166, + "step": 85 + }, + { + "epoch": 0.09935018050541516, + "grad_norm": 0.02622906304895878, + "learning_rate": 3.2241640854059465e-06, + "loss": 0.9213, + "step": 86 + }, + { + "epoch": 0.10050541516245487, + "grad_norm": 0.02663242444396019, + "learning_rate": 3.232532087697698e-06, + "loss": 0.9292, + "step": 87 + }, + { + "epoch": 0.10166064981949459, + "grad_norm": 0.02621094323694706, + "learning_rate": 3.2408044535836154e-06, + "loss": 0.923, + "step": 88 + }, + { + "epoch": 0.1028158844765343, + "grad_norm": 0.02623113803565502, + "learning_rate": 3.248983344408188e-06, + "loss": 0.9252, + "step": 89 + }, + { + "epoch": 0.10397111913357401, + "grad_norm": 0.02616356685757637, + "learning_rate": 3.257070849065542e-06, + "loss": 0.9423, + "step": 90 + }, + { + "epoch": 0.10512635379061372, + "grad_norm": 0.025917142629623413, + "learning_rate": 3.2650689872018227e-06, + "loss": 0.9337, + "step": 91 + }, + { + "epoch": 0.10628158844765342, + "grad_norm": 0.02607305720448494, + "learning_rate": 3.2729797122425927e-06, + "loss": 0.9134, + "step": 92 + }, + { + "epoch": 0.10743682310469314, + "grad_norm": 0.02566445618867874, + "learning_rate": 3.280804914256559e-06, + "loss": 0.9182, + "step": 93 + }, + { + "epoch": 0.10859205776173285, + "grad_norm": 0.025618452578783035, + "learning_rate": 3.2885464226661647e-06, + "loss": 0.883, + "step": 94 + }, + { + "epoch": 0.10974729241877257, + "grad_norm": 0.02570994757115841, + "learning_rate": 3.2962060088147467e-06, + "loss": 0.9009, + "step": 95 + }, + { + "epoch": 0.11090252707581227, + "grad_norm": 0.02652270346879959, + "learning_rate": 3.303785388399281e-06, + "loss": 0.9264, + "step": 96 + }, + { + "epoch": 0.11205776173285198, + "grad_norm": 0.026324449107050896, + "learning_rate": 3.3112862237770753e-06, + "loss": 0.9072, + "step": 97 + }, + { + "epoch": 0.1132129963898917, + "grad_norm": 0.026178548112511635, + "learning_rate": 3.318710126154159e-06, + "loss": 0.9226, + "step": 98 + }, + { + "epoch": 0.1143682310469314, + "grad_norm": 0.025392455980181694, + "learning_rate": 3.3260586576625835e-06, + "loss": 0.9002, + "step": 99 + }, + { + "epoch": 0.11552346570397112, + "grad_norm": 0.02630504220724106, + "learning_rate": 3.333333333333334e-06, + "loss": 0.9335, + "step": 100 + }, + { + "epoch": 0.11667870036101083, + "grad_norm": 0.026046302169561386, + "learning_rate": 3.340535622971072e-06, + "loss": 0.9079, + "step": 101 + }, + { + "epoch": 0.11783393501805055, + "grad_norm": 0.025761395692825317, + "learning_rate": 3.3476669529365297e-06, + "loss": 0.9188, + "step": 102 + }, + { + "epoch": 0.11898916967509025, + "grad_norm": 0.025253284722566605, + "learning_rate": 3.3547287078419544e-06, + "loss": 0.9207, + "step": 103 + }, + { + "epoch": 0.12014440433212996, + "grad_norm": 0.025493199005723, + "learning_rate": 3.361722232164634e-06, + "loss": 0.8937, + "step": 104 + }, + { + "epoch": 0.12129963898916968, + "grad_norm": 0.025204647332429886, + "learning_rate": 3.3686488317832306e-06, + "loss": 0.9189, + "step": 105 + }, + { + "epoch": 0.12245487364620938, + "grad_norm": 0.025601711124181747, + "learning_rate": 3.375509775441284e-06, + "loss": 0.9034, + "step": 106 + }, + { + "epoch": 0.1236101083032491, + "grad_norm": 0.025162257254123688, + "learning_rate": 3.3823062961420163e-06, + "loss": 0.9008, + "step": 107 + }, + { + "epoch": 0.12476534296028881, + "grad_norm": 0.025047749280929565, + "learning_rate": 3.3890395924782498e-06, + "loss": 0.8788, + "step": 108 + }, + { + "epoch": 0.12592057761732853, + "grad_norm": 0.026021016761660576, + "learning_rate": 3.3957108299010395e-06, + "loss": 0.9252, + "step": 109 + }, + { + "epoch": 0.12707581227436823, + "grad_norm": 0.025183433666825294, + "learning_rate": 3.402321141930376e-06, + "loss": 0.9221, + "step": 110 + }, + { + "epoch": 0.12823104693140794, + "grad_norm": 0.02558548003435135, + "learning_rate": 3.408871631311096e-06, + "loss": 0.8957, + "step": 111 + }, + { + "epoch": 0.12938628158844764, + "grad_norm": 0.025612782686948776, + "learning_rate": 3.415363371116969e-06, + "loss": 0.8895, + "step": 112 + }, + { + "epoch": 0.13054151624548738, + "grad_norm": 0.024990661069750786, + "learning_rate": 3.4217974058057e-06, + "loss": 0.9094, + "step": 113 + }, + { + "epoch": 0.13169675090252708, + "grad_norm": 0.026293708011507988, + "learning_rate": 3.428174752227455e-06, + "loss": 0.9063, + "step": 114 + }, + { + "epoch": 0.1328519855595668, + "grad_norm": 0.02521086297929287, + "learning_rate": 3.434496400589353e-06, + "loss": 0.9164, + "step": 115 + }, + { + "epoch": 0.1340072202166065, + "grad_norm": 0.026099544018507004, + "learning_rate": 3.440763315378198e-06, + "loss": 0.9125, + "step": 116 + }, + { + "epoch": 0.1351624548736462, + "grad_norm": 0.02469206601381302, + "learning_rate": 3.446976436243603e-06, + "loss": 0.892, + "step": 117 + }, + { + "epoch": 0.13631768953068593, + "grad_norm": 0.025795504450798035, + "learning_rate": 3.4531366788435426e-06, + "loss": 0.896, + "step": 118 + }, + { + "epoch": 0.13747292418772564, + "grad_norm": 0.024762745946645737, + "learning_rate": 3.4592449356542185e-06, + "loss": 0.8718, + "step": 119 + }, + { + "epoch": 0.13862815884476534, + "grad_norm": 0.02494460716843605, + "learning_rate": 3.4653020767460416e-06, + "loss": 0.9059, + "step": 120 + }, + { + "epoch": 0.13978339350180505, + "grad_norm": 0.02601913921535015, + "learning_rate": 3.471308950527417e-06, + "loss": 0.9188, + "step": 121 + }, + { + "epoch": 0.14093862815884475, + "grad_norm": 0.025598011910915375, + "learning_rate": 3.4772663844579142e-06, + "loss": 0.9289, + "step": 122 + }, + { + "epoch": 0.1420938628158845, + "grad_norm": 0.025548091158270836, + "learning_rate": 3.48317518573233e-06, + "loss": 0.9007, + "step": 123 + }, + { + "epoch": 0.1432490974729242, + "grad_norm": 0.02472161501646042, + "learning_rate": 3.4890361419370587e-06, + "loss": 0.8807, + "step": 124 + }, + { + "epoch": 0.1444043321299639, + "grad_norm": 0.025738514959812164, + "learning_rate": 3.4948500216800947e-06, + "loss": 0.8969, + "step": 125 + }, + { + "epoch": 0.1455595667870036, + "grad_norm": 0.025138631463050842, + "learning_rate": 3.5006175751959385e-06, + "loss": 0.9001, + "step": 126 + }, + { + "epoch": 0.1467148014440433, + "grad_norm": 0.025624800473451614, + "learning_rate": 3.506339534926595e-06, + "loss": 0.9093, + "step": 127 + }, + { + "epoch": 0.14787003610108304, + "grad_norm": 0.02503197453916073, + "learning_rate": 3.512016616079781e-06, + "loss": 0.8793, + "step": 128 + }, + { + "epoch": 0.14902527075812275, + "grad_norm": 0.02515401318669319, + "learning_rate": 3.5176495171654153e-06, + "loss": 0.8855, + "step": 129 + }, + { + "epoch": 0.15018050541516245, + "grad_norm": 0.024870432913303375, + "learning_rate": 3.523238920511395e-06, + "loss": 0.8642, + "step": 130 + }, + { + "epoch": 0.15133574007220216, + "grad_norm": 0.02646622247993946, + "learning_rate": 3.528785492759607e-06, + "loss": 0.8966, + "step": 131 + }, + { + "epoch": 0.1524909747292419, + "grad_norm": 0.025933699682354927, + "learning_rate": 3.5342898853430833e-06, + "loss": 0.9036, + "step": 132 + }, + { + "epoch": 0.1536462093862816, + "grad_norm": 0.025004137307405472, + "learning_rate": 3.5397527349451433e-06, + "loss": 0.8796, + "step": 133 + }, + { + "epoch": 0.1548014440433213, + "grad_norm": 0.027062473818659782, + "learning_rate": 3.5451746639413463e-06, + "loss": 0.8993, + "step": 134 + }, + { + "epoch": 0.155956678700361, + "grad_norm": 0.025223543867468834, + "learning_rate": 3.550556280825011e-06, + "loss": 0.8887, + "step": 135 + }, + { + "epoch": 0.1571119133574007, + "grad_norm": 0.02600521221756935, + "learning_rate": 3.55589818061703e-06, + "loss": 0.8799, + "step": 136 + }, + { + "epoch": 0.15826714801444045, + "grad_norm": 0.026399623602628708, + "learning_rate": 3.5612009452606784e-06, + "loss": 0.9006, + "step": 137 + }, + { + "epoch": 0.15942238267148015, + "grad_norm": 0.025787660852074623, + "learning_rate": 3.5664651440020615e-06, + "loss": 0.9254, + "step": 138 + }, + { + "epoch": 0.16057761732851986, + "grad_norm": 0.025667186826467514, + "learning_rate": 3.5716913337568255e-06, + "loss": 0.8843, + "step": 139 + }, + { + "epoch": 0.16173285198555956, + "grad_norm": 0.025686215609312057, + "learning_rate": 3.5768800594637304e-06, + "loss": 0.8904, + "step": 140 + }, + { + "epoch": 0.16288808664259927, + "grad_norm": 0.027956154197454453, + "learning_rate": 3.582031854425634e-06, + "loss": 0.9226, + "step": 141 + }, + { + "epoch": 0.164043321299639, + "grad_norm": 0.025962915271520615, + "learning_rate": 3.587147240638428e-06, + "loss": 0.9165, + "step": 142 + }, + { + "epoch": 0.1651985559566787, + "grad_norm": 0.026672059670090675, + "learning_rate": 3.5922267291084367e-06, + "loss": 0.8984, + "step": 143 + }, + { + "epoch": 0.1663537906137184, + "grad_norm": 0.025454359129071236, + "learning_rate": 3.59727082015875e-06, + "loss": 0.9018, + "step": 144 + }, + { + "epoch": 0.16750902527075812, + "grad_norm": 0.02756405621767044, + "learning_rate": 3.6022800037249583e-06, + "loss": 0.8928, + "step": 145 + }, + { + "epoch": 0.16866425992779782, + "grad_norm": 0.025225916877388954, + "learning_rate": 3.607254759640729e-06, + "loss": 0.887, + "step": 146 + }, + { + "epoch": 0.16981949458483755, + "grad_norm": 0.026238933205604553, + "learning_rate": 3.612195557913627e-06, + "loss": 0.8906, + "step": 147 + }, + { + "epoch": 0.17097472924187726, + "grad_norm": 0.026623785495758057, + "learning_rate": 3.6171028589915957e-06, + "loss": 0.8866, + "step": 148 + }, + { + "epoch": 0.17212996389891697, + "grad_norm": 0.02516297437250614, + "learning_rate": 3.6219771140204575e-06, + "loss": 0.875, + "step": 149 + }, + { + "epoch": 0.17328519855595667, + "grad_norm": 0.026302075013518333, + "learning_rate": 3.626818765092802e-06, + "loss": 0.9158, + "step": 150 + }, + { + "epoch": 0.17444043321299638, + "grad_norm": 0.02607031911611557, + "learning_rate": 3.631628245488616e-06, + "loss": 0.8595, + "step": 151 + }, + { + "epoch": 0.1755956678700361, + "grad_norm": 0.025099189952015877, + "learning_rate": 3.6364059799079547e-06, + "loss": 0.8762, + "step": 152 + }, + { + "epoch": 0.17675090252707581, + "grad_norm": 0.024900630116462708, + "learning_rate": 3.6411523846959985e-06, + "loss": 0.8769, + "step": 153 + }, + { + "epoch": 0.17790613718411552, + "grad_norm": 0.02502143569290638, + "learning_rate": 3.6458678680607725e-06, + "loss": 0.9109, + "step": 154 + }, + { + "epoch": 0.17906137184115523, + "grad_norm": 0.025041181594133377, + "learning_rate": 3.6505528302838196e-06, + "loss": 0.8624, + "step": 155 + }, + { + "epoch": 0.18021660649819496, + "grad_norm": 0.02584444358944893, + "learning_rate": 3.655207663924103e-06, + "loss": 0.8581, + "step": 156 + }, + { + "epoch": 0.18137184115523466, + "grad_norm": 0.024667399004101753, + "learning_rate": 3.65983275401539e-06, + "loss": 0.8699, + "step": 157 + }, + { + "epoch": 0.18252707581227437, + "grad_norm": 0.025528263300657272, + "learning_rate": 3.664428478257371e-06, + "loss": 0.8896, + "step": 158 + }, + { + "epoch": 0.18368231046931408, + "grad_norm": 0.025437112897634506, + "learning_rate": 3.6689952072007528e-06, + "loss": 0.8954, + "step": 159 + }, + { + "epoch": 0.18483754512635378, + "grad_norm": 0.025140732526779175, + "learning_rate": 3.6735333044265414e-06, + "loss": 0.8794, + "step": 160 + }, + { + "epoch": 0.1859927797833935, + "grad_norm": 0.026126103475689888, + "learning_rate": 3.6780431267197503e-06, + "loss": 0.9103, + "step": 161 + }, + { + "epoch": 0.18714801444043322, + "grad_norm": 0.025869259610772133, + "learning_rate": 3.6825250242377186e-06, + "loss": 0.8945, + "step": 162 + }, + { + "epoch": 0.18830324909747292, + "grad_norm": 0.02550615929067135, + "learning_rate": 3.6869793406732633e-06, + "loss": 0.8884, + "step": 163 + }, + { + "epoch": 0.18945848375451263, + "grad_norm": 0.02724931389093399, + "learning_rate": 3.69140641341283e-06, + "loss": 0.893, + "step": 164 + }, + { + "epoch": 0.19061371841155234, + "grad_norm": 0.02519422210752964, + "learning_rate": 3.6958065736898442e-06, + "loss": 0.8739, + "step": 165 + }, + { + "epoch": 0.19176895306859207, + "grad_norm": 0.025711793452501297, + "learning_rate": 3.700180146733426e-06, + "loss": 0.8832, + "step": 166 + }, + { + "epoch": 0.19292418772563177, + "grad_norm": 0.02538181096315384, + "learning_rate": 3.7045274519126395e-06, + "loss": 0.8785, + "step": 167 + }, + { + "epoch": 0.19407942238267148, + "grad_norm": 0.0263076052069664, + "learning_rate": 3.7088488028764387e-06, + "loss": 0.8722, + "step": 168 + }, + { + "epoch": 0.19523465703971118, + "grad_norm": 0.02613968588411808, + "learning_rate": 3.7131445076894563e-06, + "loss": 0.8833, + "step": 169 + }, + { + "epoch": 0.1963898916967509, + "grad_norm": 0.024628346785902977, + "learning_rate": 3.717414868963791e-06, + "loss": 0.8795, + "step": 170 + }, + { + "epoch": 0.19754512635379062, + "grad_norm": 0.026414738968014717, + "learning_rate": 3.721660183986924e-06, + "loss": 0.9025, + "step": 171 + }, + { + "epoch": 0.19870036101083033, + "grad_norm": 0.025311259552836418, + "learning_rate": 3.725880744845915e-06, + "loss": 0.9047, + "step": 172 + }, + { + "epoch": 0.19985559566787003, + "grad_norm": 0.025876285508275032, + "learning_rate": 3.7300768385479928e-06, + "loss": 0.8571, + "step": 173 + }, + { + "epoch": 0.20101083032490974, + "grad_norm": 0.02702619880437851, + "learning_rate": 3.7342487471376667e-06, + "loss": 0.8776, + "step": 174 + }, + { + "epoch": 0.20216606498194944, + "grad_norm": 0.026685267686843872, + "learning_rate": 3.7383967478104918e-06, + "loss": 0.9018, + "step": 175 + }, + { + "epoch": 0.20332129963898918, + "grad_norm": 0.026544688269495964, + "learning_rate": 3.7425211130235835e-06, + "loss": 0.8861, + "step": 176 + }, + { + "epoch": 0.20447653429602888, + "grad_norm": 0.026582278311252594, + "learning_rate": 3.7466221106030114e-06, + "loss": 0.8974, + "step": 177 + }, + { + "epoch": 0.2056317689530686, + "grad_norm": 0.02530672959983349, + "learning_rate": 3.7507000038481574e-06, + "loss": 0.8904, + "step": 178 + }, + { + "epoch": 0.2067870036101083, + "grad_norm": 0.02869655378162861, + "learning_rate": 3.7547550516331556e-06, + "loss": 0.8637, + "step": 179 + }, + { + "epoch": 0.20794223826714803, + "grad_norm": 0.0269013699144125, + "learning_rate": 3.7587875085055104e-06, + "loss": 0.8718, + "step": 180 + }, + { + "epoch": 0.20909747292418773, + "grad_norm": 0.027417806908488274, + "learning_rate": 3.762797624781975e-06, + "loss": 0.8697, + "step": 181 + }, + { + "epoch": 0.21025270758122744, + "grad_norm": 0.025989564135670662, + "learning_rate": 3.7667856466417917e-06, + "loss": 0.8838, + "step": 182 + }, + { + "epoch": 0.21140794223826714, + "grad_norm": 0.026576591655611992, + "learning_rate": 3.7707518162173835e-06, + "loss": 0.8704, + "step": 183 + }, + { + "epoch": 0.21256317689530685, + "grad_norm": 0.0275897067040205, + "learning_rate": 3.7746963716825613e-06, + "loss": 0.855, + "step": 184 + }, + { + "epoch": 0.21371841155234658, + "grad_norm": 0.0267089381814003, + "learning_rate": 3.778619547338356e-06, + "loss": 0.8712, + "step": 185 + }, + { + "epoch": 0.2148736462093863, + "grad_norm": 0.02677794173359871, + "learning_rate": 3.7825215736965283e-06, + "loss": 0.8729, + "step": 186 + }, + { + "epoch": 0.216028880866426, + "grad_norm": 0.0281060878187418, + "learning_rate": 3.786402677560832e-06, + "loss": 0.8845, + "step": 187 + }, + { + "epoch": 0.2171841155234657, + "grad_norm": 0.025514084845781326, + "learning_rate": 3.7902630821061337e-06, + "loss": 0.8669, + "step": 188 + }, + { + "epoch": 0.2183393501805054, + "grad_norm": 0.02548467181622982, + "learning_rate": 3.7941030069554073e-06, + "loss": 0.8897, + "step": 189 + }, + { + "epoch": 0.21949458483754514, + "grad_norm": 0.026287022978067398, + "learning_rate": 3.7979226682547152e-06, + "loss": 0.8731, + "step": 190 + }, + { + "epoch": 0.22064981949458484, + "grad_norm": 0.025090090930461884, + "learning_rate": 3.8017222787462132e-06, + "loss": 0.8946, + "step": 191 + }, + { + "epoch": 0.22180505415162455, + "grad_norm": 0.02495192363858223, + "learning_rate": 3.8055020478392497e-06, + "loss": 0.8567, + "step": 192 + }, + { + "epoch": 0.22296028880866425, + "grad_norm": 0.025306979194283485, + "learning_rate": 3.8092621816796233e-06, + "loss": 0.8819, + "step": 193 + }, + { + "epoch": 0.22411552346570396, + "grad_norm": 0.027035973966121674, + "learning_rate": 3.813002883217044e-06, + "loss": 0.8733, + "step": 194 + }, + { + "epoch": 0.2252707581227437, + "grad_norm": 0.025134000927209854, + "learning_rate": 3.816724352270864e-06, + "loss": 0.8764, + "step": 195 + }, + { + "epoch": 0.2264259927797834, + "grad_norm": 0.025370297953486443, + "learning_rate": 3.820426785594127e-06, + "loss": 0.8607, + "step": 196 + }, + { + "epoch": 0.2275812274368231, + "grad_norm": 0.02679363079369068, + "learning_rate": 3.824110376935989e-06, + "loss": 0.8726, + "step": 197 + }, + { + "epoch": 0.2287364620938628, + "grad_norm": 0.024456890299916267, + "learning_rate": 3.827775317102552e-06, + "loss": 0.858, + "step": 198 + }, + { + "epoch": 0.2298916967509025, + "grad_norm": 0.026032116264104843, + "learning_rate": 3.831421794016178e-06, + "loss": 0.8606, + "step": 199 + }, + { + "epoch": 0.23104693140794225, + "grad_norm": 0.026059836149215698, + "learning_rate": 3.835049992773302e-06, + "loss": 0.8613, + "step": 200 + }, + { + "epoch": 0.23220216606498195, + "grad_norm": 0.02430492639541626, + "learning_rate": 3.8386600957008155e-06, + "loss": 0.866, + "step": 201 + }, + { + "epoch": 0.23335740072202166, + "grad_norm": 0.026345418766140938, + "learning_rate": 3.8422522824110405e-06, + "loss": 0.8814, + "step": 202 + }, + { + "epoch": 0.23451263537906136, + "grad_norm": 0.025827715173363686, + "learning_rate": 3.8458267298553554e-06, + "loss": 0.8791, + "step": 203 + }, + { + "epoch": 0.2356678700361011, + "grad_norm": 0.025157129392027855, + "learning_rate": 3.849383612376498e-06, + "loss": 0.8797, + "step": 204 + }, + { + "epoch": 0.2368231046931408, + "grad_norm": 0.02571243792772293, + "learning_rate": 3.852923101759591e-06, + "loss": 0.8866, + "step": 205 + }, + { + "epoch": 0.2379783393501805, + "grad_norm": 0.025041643530130386, + "learning_rate": 3.856445367281923e-06, + "loss": 0.8666, + "step": 206 + }, + { + "epoch": 0.2391335740072202, + "grad_norm": 0.0259910486638546, + "learning_rate": 3.8599505757615295e-06, + "loss": 0.8727, + "step": 207 + }, + { + "epoch": 0.24028880866425992, + "grad_norm": 0.02511589229106903, + "learning_rate": 3.863438891604603e-06, + "loss": 0.8801, + "step": 208 + }, + { + "epoch": 0.24144404332129965, + "grad_norm": 0.024072881788015366, + "learning_rate": 3.866910476851757e-06, + "loss": 0.8203, + "step": 209 + }, + { + "epoch": 0.24259927797833936, + "grad_norm": 0.025221582502126694, + "learning_rate": 3.870365491223199e-06, + "loss": 0.8384, + "step": 210 + }, + { + "epoch": 0.24375451263537906, + "grad_norm": 0.02576667070388794, + "learning_rate": 3.873804092162822e-06, + "loss": 0.8423, + "step": 211 + }, + { + "epoch": 0.24490974729241877, + "grad_norm": 0.02616226114332676, + "learning_rate": 3.877226434881253e-06, + "loss": 0.8521, + "step": 212 + }, + { + "epoch": 0.24606498194945847, + "grad_norm": 0.024735651910305023, + "learning_rate": 3.880632672397897e-06, + "loss": 0.8847, + "step": 213 + }, + { + "epoch": 0.2472202166064982, + "grad_norm": 0.02537655271589756, + "learning_rate": 3.884022955581984e-06, + "loss": 0.86, + "step": 214 + }, + { + "epoch": 0.2483754512635379, + "grad_norm": 0.02478659339249134, + "learning_rate": 3.887397433192676e-06, + "loss": 0.8489, + "step": 215 + }, + { + "epoch": 0.24953068592057762, + "grad_norm": 0.025592437013983727, + "learning_rate": 3.890756251918219e-06, + "loss": 0.845, + "step": 216 + }, + { + "epoch": 0.25068592057761735, + "grad_norm": 0.02625833824276924, + "learning_rate": 3.894099556414216e-06, + "loss": 0.8584, + "step": 217 + }, + { + "epoch": 0.25184115523465705, + "grad_norm": 0.02509414032101631, + "learning_rate": 3.897427489341009e-06, + "loss": 0.8608, + "step": 218 + }, + { + "epoch": 0.25299638989169676, + "grad_norm": 0.02498655766248703, + "learning_rate": 3.900740191400198e-06, + "loss": 0.8509, + "step": 219 + }, + { + "epoch": 0.25415162454873647, + "grad_norm": 0.025928007438778877, + "learning_rate": 3.9040378013703444e-06, + "loss": 0.8556, + "step": 220 + }, + { + "epoch": 0.25530685920577617, + "grad_norm": 0.026060784235596657, + "learning_rate": 3.907320456141851e-06, + "loss": 0.8495, + "step": 221 + }, + { + "epoch": 0.2564620938628159, + "grad_norm": 0.02592143975198269, + "learning_rate": 3.910588290751064e-06, + "loss": 0.8527, + "step": 222 + }, + { + "epoch": 0.2576173285198556, + "grad_norm": 0.025333942845463753, + "learning_rate": 3.913841438413601e-06, + "loss": 0.8398, + "step": 223 + }, + { + "epoch": 0.2587725631768953, + "grad_norm": 0.026303421705961227, + "learning_rate": 3.9170800305569385e-06, + "loss": 0.8575, + "step": 224 + }, + { + "epoch": 0.259927797833935, + "grad_norm": 0.025790376588702202, + "learning_rate": 3.920304196852272e-06, + "loss": 0.8804, + "step": 225 + }, + { + "epoch": 0.26108303249097475, + "grad_norm": 0.02600492723286152, + "learning_rate": 3.923514065245669e-06, + "loss": 0.8532, + "step": 226 + }, + { + "epoch": 0.26223826714801446, + "grad_norm": 0.025459101423621178, + "learning_rate": 3.9267097619885385e-06, + "loss": 0.8597, + "step": 227 + }, + { + "epoch": 0.26339350180505416, + "grad_norm": 0.025850312784314156, + "learning_rate": 3.9298914116674236e-06, + "loss": 0.8605, + "step": 228 + }, + { + "epoch": 0.26454873646209387, + "grad_norm": 0.0261206217110157, + "learning_rate": 3.9330591372331475e-06, + "loss": 0.8463, + "step": 229 + }, + { + "epoch": 0.2657039711191336, + "grad_norm": 0.026850640773773193, + "learning_rate": 3.936213060029322e-06, + "loss": 0.8655, + "step": 230 + }, + { + "epoch": 0.2668592057761733, + "grad_norm": 0.02615203708410263, + "learning_rate": 3.9393532998202405e-06, + "loss": 0.8729, + "step": 231 + }, + { + "epoch": 0.268014440433213, + "grad_norm": 0.026082858443260193, + "learning_rate": 3.942479974818166e-06, + "loss": 0.8626, + "step": 232 + }, + { + "epoch": 0.2691696750902527, + "grad_norm": 0.02707120031118393, + "learning_rate": 3.9455932017100315e-06, + "loss": 0.8665, + "step": 233 + }, + { + "epoch": 0.2703249097472924, + "grad_norm": 0.02531982958316803, + "learning_rate": 3.9486930956835726e-06, + "loss": 0.841, + "step": 234 + }, + { + "epoch": 0.27148014440433216, + "grad_norm": 0.02577449567615986, + "learning_rate": 3.951779770452894e-06, + "loss": 0.8528, + "step": 235 + }, + { + "epoch": 0.27263537906137186, + "grad_norm": 0.025777166709303856, + "learning_rate": 3.954853338283512e-06, + "loss": 0.8636, + "step": 236 + }, + { + "epoch": 0.27379061371841157, + "grad_norm": 0.027185678482055664, + "learning_rate": 3.95791391001684e-06, + "loss": 0.884, + "step": 237 + }, + { + "epoch": 0.2749458483754513, + "grad_norm": 0.025784213095903397, + "learning_rate": 3.960961595094187e-06, + "loss": 0.8532, + "step": 238 + }, + { + "epoch": 0.276101083032491, + "grad_norm": 0.025824446231126785, + "learning_rate": 3.96399650158023e-06, + "loss": 0.8555, + "step": 239 + }, + { + "epoch": 0.2772563176895307, + "grad_norm": 0.026414718478918076, + "learning_rate": 3.96701873618601e-06, + "loss": 0.8634, + "step": 240 + }, + { + "epoch": 0.2784115523465704, + "grad_norm": 0.02620028331875801, + "learning_rate": 3.970028404291448e-06, + "loss": 0.8574, + "step": 241 + }, + { + "epoch": 0.2795667870036101, + "grad_norm": 0.025768935680389404, + "learning_rate": 3.973025609967386e-06, + "loss": 0.8741, + "step": 242 + }, + { + "epoch": 0.2807220216606498, + "grad_norm": 0.027099501341581345, + "learning_rate": 3.976010455997187e-06, + "loss": 0.8705, + "step": 243 + }, + { + "epoch": 0.2818772563176895, + "grad_norm": 0.02535291016101837, + "learning_rate": 3.978983043897884e-06, + "loss": 0.853, + "step": 244 + }, + { + "epoch": 0.28303249097472927, + "grad_norm": 0.02731909230351448, + "learning_rate": 3.981943473940888e-06, + "loss": 0.8776, + "step": 245 + }, + { + "epoch": 0.284187725631769, + "grad_norm": 0.02630050666630268, + "learning_rate": 3.984891845172299e-06, + "loss": 0.86, + "step": 246 + }, + { + "epoch": 0.2853429602888087, + "grad_norm": 0.025558151304721832, + "learning_rate": 3.987828255432777e-06, + "loss": 0.8758, + "step": 247 + }, + { + "epoch": 0.2864981949458484, + "grad_norm": 0.026159491389989853, + "learning_rate": 3.990752801377028e-06, + "loss": 0.8626, + "step": 248 + }, + { + "epoch": 0.2876534296028881, + "grad_norm": 0.02615942806005478, + "learning_rate": 3.993665578492894e-06, + "loss": 0.858, + "step": 249 + }, + { + "epoch": 0.2888086642599278, + "grad_norm": 0.025161130353808403, + "learning_rate": 3.996566681120062e-06, + "loss": 0.8429, + "step": 250 + }, + { + "epoch": 0.2899638989169675, + "grad_norm": 0.02635674737393856, + "learning_rate": 3.999456202468397e-06, + "loss": 0.8686, + "step": 251 + }, + { + "epoch": 0.2911191335740072, + "grad_norm": 0.025895683094859123, + "learning_rate": 4.0023342346359075e-06, + "loss": 0.8422, + "step": 252 + }, + { + "epoch": 0.2922743682310469, + "grad_norm": 0.027750222012400627, + "learning_rate": 4.0052008686263635e-06, + "loss": 0.8868, + "step": 253 + }, + { + "epoch": 0.2934296028880866, + "grad_norm": 0.02600325271487236, + "learning_rate": 4.0080561943665644e-06, + "loss": 0.8639, + "step": 254 + }, + { + "epoch": 0.2945848375451264, + "grad_norm": 0.02592737227678299, + "learning_rate": 4.010900300723259e-06, + "loss": 0.8453, + "step": 255 + }, + { + "epoch": 0.2957400722021661, + "grad_norm": 0.025766368955373764, + "learning_rate": 4.0137332755197495e-06, + "loss": 0.8771, + "step": 256 + }, + { + "epoch": 0.2968953068592058, + "grad_norm": 0.026158476248383522, + "learning_rate": 4.016555205552159e-06, + "loss": 0.8524, + "step": 257 + }, + { + "epoch": 0.2980505415162455, + "grad_norm": 0.024570690467953682, + "learning_rate": 4.019366176605384e-06, + "loss": 0.8345, + "step": 258 + }, + { + "epoch": 0.2992057761732852, + "grad_norm": 0.025776837021112442, + "learning_rate": 4.022166273468753e-06, + "loss": 0.8441, + "step": 259 + }, + { + "epoch": 0.3003610108303249, + "grad_norm": 0.02488660253584385, + "learning_rate": 4.024955579951363e-06, + "loss": 0.8639, + "step": 260 + }, + { + "epoch": 0.3015162454873646, + "grad_norm": 0.025739185512065887, + "learning_rate": 4.0277341788971355e-06, + "loss": 0.8465, + "step": 261 + }, + { + "epoch": 0.3026714801444043, + "grad_norm": 0.02594810351729393, + "learning_rate": 4.030502152199576e-06, + "loss": 0.8752, + "step": 262 + }, + { + "epoch": 0.303826714801444, + "grad_norm": 0.025780096650123596, + "learning_rate": 4.033259580816264e-06, + "loss": 0.8433, + "step": 263 + }, + { + "epoch": 0.3049819494584838, + "grad_norm": 0.025661000981926918, + "learning_rate": 4.036006544783052e-06, + "loss": 0.8769, + "step": 264 + }, + { + "epoch": 0.3061371841155235, + "grad_norm": 0.025767603889107704, + "learning_rate": 4.0387431232280135e-06, + "loss": 0.8705, + "step": 265 + }, + { + "epoch": 0.3072924187725632, + "grad_norm": 0.0245257169008255, + "learning_rate": 4.041469394385113e-06, + "loss": 0.8497, + "step": 266 + }, + { + "epoch": 0.3084476534296029, + "grad_norm": 0.026073751971125603, + "learning_rate": 4.044185435607626e-06, + "loss": 0.8503, + "step": 267 + }, + { + "epoch": 0.3096028880866426, + "grad_norm": 0.025399446487426758, + "learning_rate": 4.046891323381315e-06, + "loss": 0.8594, + "step": 268 + }, + { + "epoch": 0.3107581227436823, + "grad_norm": 0.02536724880337715, + "learning_rate": 4.049587133337347e-06, + "loss": 0.8513, + "step": 269 + }, + { + "epoch": 0.311913357400722, + "grad_norm": 0.024877896532416344, + "learning_rate": 4.05227294026498e-06, + "loss": 0.8302, + "step": 270 + }, + { + "epoch": 0.3130685920577617, + "grad_norm": 0.026714155450463295, + "learning_rate": 4.05494881812401e-06, + "loss": 0.8787, + "step": 271 + }, + { + "epoch": 0.3142238267148014, + "grad_norm": 0.02595258131623268, + "learning_rate": 4.057614840056999e-06, + "loss": 0.8572, + "step": 272 + }, + { + "epoch": 0.31537906137184113, + "grad_norm": 0.02663499116897583, + "learning_rate": 4.060271078401261e-06, + "loss": 0.8656, + "step": 273 + }, + { + "epoch": 0.3165342960288809, + "grad_norm": 0.02580447867512703, + "learning_rate": 4.0629176047006474e-06, + "loss": 0.8574, + "step": 274 + }, + { + "epoch": 0.3176895306859206, + "grad_norm": 0.02633056230843067, + "learning_rate": 4.065554489717105e-06, + "loss": 0.864, + "step": 275 + }, + { + "epoch": 0.3188447653429603, + "grad_norm": 0.026197999715805054, + "learning_rate": 4.06818180344203e-06, + "loss": 0.8492, + "step": 276 + }, + { + "epoch": 0.32, + "grad_norm": 0.025230281054973602, + "learning_rate": 4.070799615107415e-06, + "loss": 0.8255, + "step": 277 + }, + { + "epoch": 0.3211552346570397, + "grad_norm": 0.02663682959973812, + "learning_rate": 4.073407993196794e-06, + "loss": 0.8557, + "step": 278 + }, + { + "epoch": 0.3223104693140794, + "grad_norm": 0.025046920403838158, + "learning_rate": 4.076007005455996e-06, + "loss": 0.8433, + "step": 279 + }, + { + "epoch": 0.3234657039711191, + "grad_norm": 0.025554031133651733, + "learning_rate": 4.078596718903699e-06, + "loss": 0.8639, + "step": 280 + }, + { + "epoch": 0.32462093862815883, + "grad_norm": 0.025266235694289207, + "learning_rate": 4.0811771998418e-06, + "loss": 0.8617, + "step": 281 + }, + { + "epoch": 0.32577617328519853, + "grad_norm": 0.026751738041639328, + "learning_rate": 4.083748513865602e-06, + "loss": 0.8472, + "step": 282 + }, + { + "epoch": 0.3269314079422383, + "grad_norm": 0.026282917708158493, + "learning_rate": 4.086310725873818e-06, + "loss": 0.8518, + "step": 283 + }, + { + "epoch": 0.328086642599278, + "grad_norm": 0.024683093652129173, + "learning_rate": 4.088863900078397e-06, + "loss": 0.856, + "step": 284 + }, + { + "epoch": 0.3292418772563177, + "grad_norm": 0.025772644206881523, + "learning_rate": 4.091408100014184e-06, + "loss": 0.849, + "step": 285 + }, + { + "epoch": 0.3303971119133574, + "grad_norm": 0.02520694211125374, + "learning_rate": 4.093943388548406e-06, + "loss": 0.8363, + "step": 286 + }, + { + "epoch": 0.3315523465703971, + "grad_norm": 0.026992499828338623, + "learning_rate": 4.096469827889988e-06, + "loss": 0.859, + "step": 287 + }, + { + "epoch": 0.3327075812274368, + "grad_norm": 0.02580207772552967, + "learning_rate": 4.0989874795987185e-06, + "loss": 0.873, + "step": 288 + }, + { + "epoch": 0.3338628158844765, + "grad_norm": 0.025334378704428673, + "learning_rate": 4.101496404594247e-06, + "loss": 0.848, + "step": 289 + }, + { + "epoch": 0.33501805054151623, + "grad_norm": 0.027060015127062798, + "learning_rate": 4.103996663164927e-06, + "loss": 0.834, + "step": 290 + }, + { + "epoch": 0.33617328519855594, + "grad_norm": 0.027299024164676666, + "learning_rate": 4.106488314976513e-06, + "loss": 0.8667, + "step": 291 + }, + { + "epoch": 0.33732851985559564, + "grad_norm": 0.027205299586057663, + "learning_rate": 4.108971419080698e-06, + "loss": 0.8548, + "step": 292 + }, + { + "epoch": 0.3384837545126354, + "grad_norm": 0.025082072243094444, + "learning_rate": 4.111446033923516e-06, + "loss": 0.8418, + "step": 293 + }, + { + "epoch": 0.3396389891696751, + "grad_norm": 0.025831829756498337, + "learning_rate": 4.113912217353596e-06, + "loss": 0.8419, + "step": 294 + }, + { + "epoch": 0.3407942238267148, + "grad_norm": 0.02547086775302887, + "learning_rate": 4.1163700266302726e-06, + "loss": 0.8708, + "step": 295 + }, + { + "epoch": 0.3419494584837545, + "grad_norm": 0.02637997455894947, + "learning_rate": 4.118819518431564e-06, + "loss": 0.8492, + "step": 296 + }, + { + "epoch": 0.3431046931407942, + "grad_norm": 0.02541852556169033, + "learning_rate": 4.121260748862021e-06, + "loss": 0.8607, + "step": 297 + }, + { + "epoch": 0.34425992779783393, + "grad_norm": 0.026499446481466293, + "learning_rate": 4.123693773460426e-06, + "loss": 0.8543, + "step": 298 + }, + { + "epoch": 0.34541516245487364, + "grad_norm": 0.025943726301193237, + "learning_rate": 4.126118647207383e-06, + "loss": 0.852, + "step": 299 + }, + { + "epoch": 0.34657039711191334, + "grad_norm": 0.024514541029930115, + "learning_rate": 4.128535424532771e-06, + "loss": 0.8335, + "step": 300 + }, + { + "epoch": 0.34772563176895305, + "grad_norm": 0.026036018505692482, + "learning_rate": 4.130944159323072e-06, + "loss": 0.8373, + "step": 301 + }, + { + "epoch": 0.34888086642599275, + "grad_norm": 0.0261723380535841, + "learning_rate": 4.133344904928585e-06, + "loss": 0.839, + "step": 302 + }, + { + "epoch": 0.3500361010830325, + "grad_norm": 0.025829095393419266, + "learning_rate": 4.135737714170509e-06, + "loss": 0.8691, + "step": 303 + }, + { + "epoch": 0.3511913357400722, + "grad_norm": 0.026276499032974243, + "learning_rate": 4.138122639347924e-06, + "loss": 0.839, + "step": 304 + }, + { + "epoch": 0.3523465703971119, + "grad_norm": 0.026492850854992867, + "learning_rate": 4.140499732244644e-06, + "loss": 0.8585, + "step": 305 + }, + { + "epoch": 0.35350180505415163, + "grad_norm": 0.027172986418008804, + "learning_rate": 4.142869044135967e-06, + "loss": 0.8369, + "step": 306 + }, + { + "epoch": 0.35465703971119134, + "grad_norm": 0.026907015591859818, + "learning_rate": 4.145230625795312e-06, + "loss": 0.8578, + "step": 307 + }, + { + "epoch": 0.35581227436823104, + "grad_norm": 0.02687431313097477, + "learning_rate": 4.147584527500741e-06, + "loss": 0.8727, + "step": 308 + }, + { + "epoch": 0.35696750902527075, + "grad_norm": 0.026860354468226433, + "learning_rate": 4.149930799041391e-06, + "loss": 0.8026, + "step": 309 + }, + { + "epoch": 0.35812274368231045, + "grad_norm": 0.02500147931277752, + "learning_rate": 4.152269489723789e-06, + "loss": 0.8309, + "step": 310 + }, + { + "epoch": 0.35927797833935016, + "grad_norm": 0.026802562177181244, + "learning_rate": 4.154600648378063e-06, + "loss": 0.8524, + "step": 311 + }, + { + "epoch": 0.3604332129963899, + "grad_norm": 0.026412485167384148, + "learning_rate": 4.156924323364072e-06, + "loss": 0.8485, + "step": 312 + }, + { + "epoch": 0.3615884476534296, + "grad_norm": 0.027881808578968048, + "learning_rate": 4.159240562577414e-06, + "loss": 0.8654, + "step": 313 + }, + { + "epoch": 0.36274368231046933, + "grad_norm": 0.026209615170955658, + "learning_rate": 4.1615494134553584e-06, + "loss": 0.8492, + "step": 314 + }, + { + "epoch": 0.36389891696750903, + "grad_norm": 0.026174476370215416, + "learning_rate": 4.163850922982668e-06, + "loss": 0.818, + "step": 315 + }, + { + "epoch": 0.36505415162454874, + "grad_norm": 0.02707611583173275, + "learning_rate": 4.166145137697341e-06, + "loss": 0.8586, + "step": 316 + }, + { + "epoch": 0.36620938628158844, + "grad_norm": 0.02789073996245861, + "learning_rate": 4.1684321036962525e-06, + "loss": 0.8468, + "step": 317 + }, + { + "epoch": 0.36736462093862815, + "grad_norm": 0.02748328261077404, + "learning_rate": 4.170711866640721e-06, + "loss": 0.8688, + "step": 318 + }, + { + "epoch": 0.36851985559566786, + "grad_norm": 0.029765894636511803, + "learning_rate": 4.172984471761969e-06, + "loss": 0.8521, + "step": 319 + }, + { + "epoch": 0.36967509025270756, + "grad_norm": 0.025119414553046227, + "learning_rate": 4.17524996386651e-06, + "loss": 0.829, + "step": 320 + }, + { + "epoch": 0.37083032490974727, + "grad_norm": 0.028743257746100426, + "learning_rate": 4.177508387341454e-06, + "loss": 0.8676, + "step": 321 + }, + { + "epoch": 0.371985559566787, + "grad_norm": 0.02732737548649311, + "learning_rate": 4.179759786159719e-06, + "loss": 0.8516, + "step": 322 + }, + { + "epoch": 0.37314079422382673, + "grad_norm": 0.026275500655174255, + "learning_rate": 4.182004203885172e-06, + "loss": 0.8606, + "step": 323 + }, + { + "epoch": 0.37429602888086644, + "grad_norm": 0.028016693890094757, + "learning_rate": 4.1842416836776876e-06, + "loss": 0.8272, + "step": 324 + }, + { + "epoch": 0.37545126353790614, + "grad_norm": 0.026512511074543, + "learning_rate": 4.1864722682981245e-06, + "loss": 0.8438, + "step": 325 + }, + { + "epoch": 0.37660649819494585, + "grad_norm": 0.026299525052309036, + "learning_rate": 4.188696000113232e-06, + "loss": 0.8485, + "step": 326 + }, + { + "epoch": 0.37776173285198555, + "grad_norm": 0.02613169699907303, + "learning_rate": 4.190912921100477e-06, + "loss": 0.8184, + "step": 327 + }, + { + "epoch": 0.37891696750902526, + "grad_norm": 0.02610696479678154, + "learning_rate": 4.193123072852799e-06, + "loss": 0.8609, + "step": 328 + }, + { + "epoch": 0.38007220216606497, + "grad_norm": 0.026195967569947243, + "learning_rate": 4.1953264965832905e-06, + "loss": 0.8302, + "step": 329 + }, + { + "epoch": 0.38122743682310467, + "grad_norm": 0.026481349021196365, + "learning_rate": 4.197523233129813e-06, + "loss": 0.8491, + "step": 330 + }, + { + "epoch": 0.38238267148014443, + "grad_norm": 0.024858448654413223, + "learning_rate": 4.199713322959531e-06, + "loss": 0.8404, + "step": 331 + }, + { + "epoch": 0.38353790613718414, + "grad_norm": 0.026140978559851646, + "learning_rate": 4.201896806173394e-06, + "loss": 0.8558, + "step": 332 + }, + { + "epoch": 0.38469314079422384, + "grad_norm": 0.026811236515641212, + "learning_rate": 4.204073722510534e-06, + "loss": 0.8747, + "step": 333 + }, + { + "epoch": 0.38584837545126355, + "grad_norm": 0.0256299190223217, + "learning_rate": 4.206244111352608e-06, + "loss": 0.8521, + "step": 334 + }, + { + "epoch": 0.38700361010830325, + "grad_norm": 0.026677099987864494, + "learning_rate": 4.208408011728075e-06, + "loss": 0.8548, + "step": 335 + }, + { + "epoch": 0.38815884476534296, + "grad_norm": 0.026028303429484367, + "learning_rate": 4.210565462316407e-06, + "loss": 0.8474, + "step": 336 + }, + { + "epoch": 0.38931407942238266, + "grad_norm": 0.02715076506137848, + "learning_rate": 4.212716501452232e-06, + "loss": 0.8318, + "step": 337 + }, + { + "epoch": 0.39046931407942237, + "grad_norm": 0.02765047550201416, + "learning_rate": 4.214861167129425e-06, + "loss": 0.8558, + "step": 338 + }, + { + "epoch": 0.3916245487364621, + "grad_norm": 0.025798538699746132, + "learning_rate": 4.216999497005137e-06, + "loss": 0.8436, + "step": 339 + }, + { + "epoch": 0.3927797833935018, + "grad_norm": 0.028038429096341133, + "learning_rate": 4.219131528403759e-06, + "loss": 0.845, + "step": 340 + }, + { + "epoch": 0.39393501805054154, + "grad_norm": 0.026474367827177048, + "learning_rate": 4.22125729832083e-06, + "loss": 0.8411, + "step": 341 + }, + { + "epoch": 0.39509025270758125, + "grad_norm": 0.025686634704470634, + "learning_rate": 4.223376843426892e-06, + "loss": 0.8485, + "step": 342 + }, + { + "epoch": 0.39624548736462095, + "grad_norm": 0.026812469586730003, + "learning_rate": 4.225490200071284e-06, + "loss": 0.8571, + "step": 343 + }, + { + "epoch": 0.39740072202166066, + "grad_norm": 0.02608347311615944, + "learning_rate": 4.227597404285883e-06, + "loss": 0.8426, + "step": 344 + }, + { + "epoch": 0.39855595667870036, + "grad_norm": 0.02631702460348606, + "learning_rate": 4.229698491788791e-06, + "loss": 0.876, + "step": 345 + }, + { + "epoch": 0.39971119133574007, + "grad_norm": 0.0272465068846941, + "learning_rate": 4.231793497987962e-06, + "loss": 0.861, + "step": 346 + }, + { + "epoch": 0.4008664259927798, + "grad_norm": 0.02596902847290039, + "learning_rate": 4.233882457984791e-06, + "loss": 0.8645, + "step": 347 + }, + { + "epoch": 0.4020216606498195, + "grad_norm": 0.02580340951681137, + "learning_rate": 4.235965406577636e-06, + "loss": 0.8447, + "step": 348 + }, + { + "epoch": 0.4031768953068592, + "grad_norm": 0.027888286858797073, + "learning_rate": 4.2380423782653e-06, + "loss": 0.8425, + "step": 349 + }, + { + "epoch": 0.4043321299638989, + "grad_norm": 0.025501729920506477, + "learning_rate": 4.2401134072504595e-06, + "loss": 0.8307, + "step": 350 + }, + { + "epoch": 0.40548736462093865, + "grad_norm": 0.026491057127714157, + "learning_rate": 4.24217852744304e-06, + "loss": 0.8237, + "step": 351 + }, + { + "epoch": 0.40664259927797836, + "grad_norm": 0.027327535673975945, + "learning_rate": 4.244237772463552e-06, + "loss": 0.8605, + "step": 352 + }, + { + "epoch": 0.40779783393501806, + "grad_norm": 0.026969779282808304, + "learning_rate": 4.246291175646372e-06, + "loss": 0.8357, + "step": 353 + }, + { + "epoch": 0.40895306859205777, + "grad_norm": 0.02584182471036911, + "learning_rate": 4.2483387700429804e-06, + "loss": 0.8348, + "step": 354 + }, + { + "epoch": 0.41010830324909747, + "grad_norm": 0.025704173371195793, + "learning_rate": 4.2503805884251575e-06, + "loss": 0.8392, + "step": 355 + }, + { + "epoch": 0.4112635379061372, + "grad_norm": 0.026744280010461807, + "learning_rate": 4.252416663288126e-06, + "loss": 0.8405, + "step": 356 + }, + { + "epoch": 0.4124187725631769, + "grad_norm": 0.02655099704861641, + "learning_rate": 4.2544470268536555e-06, + "loss": 0.8374, + "step": 357 + }, + { + "epoch": 0.4135740072202166, + "grad_norm": 0.02521936595439911, + "learning_rate": 4.2564717110731246e-06, + "loss": 0.8442, + "step": 358 + }, + { + "epoch": 0.4147292418772563, + "grad_norm": 0.02607394941151142, + "learning_rate": 4.258490747630532e-06, + "loss": 0.8273, + "step": 359 + }, + { + "epoch": 0.41588447653429605, + "grad_norm": 0.027000917121767998, + "learning_rate": 4.26050416794548e-06, + "loss": 0.8664, + "step": 360 + }, + { + "epoch": 0.41703971119133576, + "grad_norm": 0.026261812075972557, + "learning_rate": 4.262512003176097e-06, + "loss": 0.8351, + "step": 361 + }, + { + "epoch": 0.41819494584837547, + "grad_norm": 0.025927625596523285, + "learning_rate": 4.264514284221944e-06, + "loss": 0.8419, + "step": 362 + }, + { + "epoch": 0.41935018050541517, + "grad_norm": 0.02672671154141426, + "learning_rate": 4.2665110417268545e-06, + "loss": 0.84, + "step": 363 + }, + { + "epoch": 0.4205054151624549, + "grad_norm": 0.02695775218307972, + "learning_rate": 4.26850230608176e-06, + "loss": 0.8483, + "step": 364 + }, + { + "epoch": 0.4216606498194946, + "grad_norm": 0.026013897731900215, + "learning_rate": 4.270488107427459e-06, + "loss": 0.8371, + "step": 365 + }, + { + "epoch": 0.4228158844765343, + "grad_norm": 0.026113096624612808, + "learning_rate": 4.272468475657351e-06, + "loss": 0.8448, + "step": 366 + }, + { + "epoch": 0.423971119133574, + "grad_norm": 0.02611883170902729, + "learning_rate": 4.27444344042015e-06, + "loss": 0.8522, + "step": 367 + }, + { + "epoch": 0.4251263537906137, + "grad_norm": 0.02588835172355175, + "learning_rate": 4.27641303112253e-06, + "loss": 0.8348, + "step": 368 + }, + { + "epoch": 0.4262815884476534, + "grad_norm": 0.026020022109150887, + "learning_rate": 4.278377276931768e-06, + "loss": 0.8376, + "step": 369 + }, + { + "epoch": 0.42743682310469316, + "grad_norm": 0.026923442259430885, + "learning_rate": 4.280336206778326e-06, + "loss": 0.8435, + "step": 370 + }, + { + "epoch": 0.42859205776173287, + "grad_norm": 0.026913795620203018, + "learning_rate": 4.2822898493584105e-06, + "loss": 0.8374, + "step": 371 + }, + { + "epoch": 0.4297472924187726, + "grad_norm": 0.025143541395664215, + "learning_rate": 4.2842382331364965e-06, + "loss": 0.8169, + "step": 372 + }, + { + "epoch": 0.4309025270758123, + "grad_norm": 0.027908481657505035, + "learning_rate": 4.2861813863478135e-06, + "loss": 0.8561, + "step": 373 + }, + { + "epoch": 0.432057761732852, + "grad_norm": 0.026535294950008392, + "learning_rate": 4.288119337000801e-06, + "loss": 0.8299, + "step": 374 + }, + { + "epoch": 0.4332129963898917, + "grad_norm": 0.027421629056334496, + "learning_rate": 4.290052112879532e-06, + "loss": 0.8273, + "step": 375 + }, + { + "epoch": 0.4343682310469314, + "grad_norm": 0.02671075239777565, + "learning_rate": 4.291979741546102e-06, + "loss": 0.8215, + "step": 376 + }, + { + "epoch": 0.4355234657039711, + "grad_norm": 0.025190116837620735, + "learning_rate": 4.293902250342989e-06, + "loss": 0.871, + "step": 377 + }, + { + "epoch": 0.4366787003610108, + "grad_norm": 0.026498563587665558, + "learning_rate": 4.295819666395376e-06, + "loss": 0.8582, + "step": 378 + }, + { + "epoch": 0.43783393501805057, + "grad_norm": 0.02722996659576893, + "learning_rate": 4.297732016613455e-06, + "loss": 0.8288, + "step": 379 + }, + { + "epoch": 0.4389891696750903, + "grad_norm": 0.028074050322175026, + "learning_rate": 4.299639327694684e-06, + "loss": 0.8574, + "step": 380 + }, + { + "epoch": 0.44014440433213, + "grad_norm": 0.026562221348285675, + "learning_rate": 4.301541626126033e-06, + "loss": 0.8337, + "step": 381 + }, + { + "epoch": 0.4412996389891697, + "grad_norm": 0.025497550144791603, + "learning_rate": 4.303438938186182e-06, + "loss": 0.8388, + "step": 382 + }, + { + "epoch": 0.4424548736462094, + "grad_norm": 0.028150785714387894, + "learning_rate": 4.305331289947705e-06, + "loss": 0.8381, + "step": 383 + }, + { + "epoch": 0.4436101083032491, + "grad_norm": 0.024889659136533737, + "learning_rate": 4.307218707279219e-06, + "loss": 0.8297, + "step": 384 + }, + { + "epoch": 0.4447653429602888, + "grad_norm": 0.02572575956583023, + "learning_rate": 4.309101215847502e-06, + "loss": 0.8464, + "step": 385 + }, + { + "epoch": 0.4459205776173285, + "grad_norm": 0.028156662359833717, + "learning_rate": 4.310978841119592e-06, + "loss": 0.8307, + "step": 386 + }, + { + "epoch": 0.4470758122743682, + "grad_norm": 0.02584012597799301, + "learning_rate": 4.312851608364853e-06, + "loss": 0.8342, + "step": 387 + }, + { + "epoch": 0.4482310469314079, + "grad_norm": 0.02681022137403488, + "learning_rate": 4.3147195426570124e-06, + "loss": 0.839, + "step": 388 + }, + { + "epoch": 0.4493862815884477, + "grad_norm": 0.028267353773117065, + "learning_rate": 4.31658266887618e-06, + "loss": 0.8602, + "step": 389 + }, + { + "epoch": 0.4505415162454874, + "grad_norm": 0.026196565479040146, + "learning_rate": 4.318441011710832e-06, + "loss": 0.8374, + "step": 390 + }, + { + "epoch": 0.4516967509025271, + "grad_norm": 0.028164999559521675, + "learning_rate": 4.3202945956597785e-06, + "loss": 0.8458, + "step": 391 + }, + { + "epoch": 0.4528519855595668, + "grad_norm": 0.026657233014702797, + "learning_rate": 4.322143445034095e-06, + "loss": 0.8238, + "step": 392 + }, + { + "epoch": 0.4540072202166065, + "grad_norm": 0.025035889819264412, + "learning_rate": 4.323987583959045e-06, + "loss": 0.8371, + "step": 393 + }, + { + "epoch": 0.4551624548736462, + "grad_norm": 0.02667805179953575, + "learning_rate": 4.325827036375958e-06, + "loss": 0.8201, + "step": 394 + }, + { + "epoch": 0.4563176895306859, + "grad_norm": 0.028383539989590645, + "learning_rate": 4.327661826044101e-06, + "loss": 0.8385, + "step": 395 + }, + { + "epoch": 0.4574729241877256, + "grad_norm": 0.025015531107783318, + "learning_rate": 4.329491976542521e-06, + "loss": 0.8013, + "step": 396 + }, + { + "epoch": 0.4586281588447653, + "grad_norm": 0.0283355675637722, + "learning_rate": 4.3313175112718595e-06, + "loss": 0.8228, + "step": 397 + }, + { + "epoch": 0.459783393501805, + "grad_norm": 0.026209138333797455, + "learning_rate": 4.333138453456147e-06, + "loss": 0.8336, + "step": 398 + }, + { + "epoch": 0.4609386281588448, + "grad_norm": 0.024707714095711708, + "learning_rate": 4.334954826144581e-06, + "loss": 0.8417, + "step": 399 + }, + { + "epoch": 0.4620938628158845, + "grad_norm": 0.02748985029757023, + "learning_rate": 4.336766652213271e-06, + "loss": 0.843, + "step": 400 + }, + { + "epoch": 0.4632490974729242, + "grad_norm": 0.02513727732002735, + "learning_rate": 4.3385739543669715e-06, + "loss": 0.8301, + "step": 401 + }, + { + "epoch": 0.4644043321299639, + "grad_norm": 0.028213754296302795, + "learning_rate": 4.340376755140784e-06, + "loss": 0.8497, + "step": 402 + }, + { + "epoch": 0.4655595667870036, + "grad_norm": 0.02582547254860401, + "learning_rate": 4.34217507690185e-06, + "loss": 0.8252, + "step": 403 + }, + { + "epoch": 0.4667148014440433, + "grad_norm": 0.028116989880800247, + "learning_rate": 4.343968941851009e-06, + "loss": 0.8312, + "step": 404 + }, + { + "epoch": 0.467870036101083, + "grad_norm": 0.026428736746311188, + "learning_rate": 4.345758372024448e-06, + "loss": 0.8177, + "step": 405 + }, + { + "epoch": 0.4690252707581227, + "grad_norm": 0.025875557214021683, + "learning_rate": 4.347543389295324e-06, + "loss": 0.8258, + "step": 406 + }, + { + "epoch": 0.47018050541516243, + "grad_norm": 0.025924712419509888, + "learning_rate": 4.3493240153753665e-06, + "loss": 0.8379, + "step": 407 + }, + { + "epoch": 0.4713357400722022, + "grad_norm": 0.025729961693286896, + "learning_rate": 4.3511002718164665e-06, + "loss": 0.8396, + "step": 408 + }, + { + "epoch": 0.4724909747292419, + "grad_norm": 0.026374271139502525, + "learning_rate": 4.352872180012237e-06, + "loss": 0.8381, + "step": 409 + }, + { + "epoch": 0.4736462093862816, + "grad_norm": 0.028471384197473526, + "learning_rate": 4.35463976119956e-06, + "loss": 0.8333, + "step": 410 + }, + { + "epoch": 0.4748014440433213, + "grad_norm": 0.025666672736406326, + "learning_rate": 4.356403036460116e-06, + "loss": 0.8191, + "step": 411 + }, + { + "epoch": 0.475956678700361, + "grad_norm": 0.026298929005861282, + "learning_rate": 4.358162026721892e-06, + "loss": 0.8225, + "step": 412 + }, + { + "epoch": 0.4771119133574007, + "grad_norm": 0.027312815189361572, + "learning_rate": 4.359916752760669e-06, + "loss": 0.804, + "step": 413 + }, + { + "epoch": 0.4782671480144404, + "grad_norm": 0.02659662254154682, + "learning_rate": 4.361667235201499e-06, + "loss": 0.8286, + "step": 414 + }, + { + "epoch": 0.47942238267148013, + "grad_norm": 0.02585594914853573, + "learning_rate": 4.363413494520154e-06, + "loss": 0.8229, + "step": 415 + }, + { + "epoch": 0.48057761732851983, + "grad_norm": 0.027580194175243378, + "learning_rate": 4.365155551044572e-06, + "loss": 0.8397, + "step": 416 + }, + { + "epoch": 0.48173285198555954, + "grad_norm": 0.025151105597615242, + "learning_rate": 4.366893424956263e-06, + "loss": 0.8356, + "step": 417 + }, + { + "epoch": 0.4828880866425993, + "grad_norm": 0.025705143809318542, + "learning_rate": 4.368627136291726e-06, + "loss": 0.8342, + "step": 418 + }, + { + "epoch": 0.484043321299639, + "grad_norm": 0.027614813297986984, + "learning_rate": 4.370356704943825e-06, + "loss": 0.8481, + "step": 419 + }, + { + "epoch": 0.4851985559566787, + "grad_norm": 0.026671946048736572, + "learning_rate": 4.372082150663167e-06, + "loss": 0.8375, + "step": 420 + }, + { + "epoch": 0.4863537906137184, + "grad_norm": 0.026453910395503044, + "learning_rate": 4.3738034930594475e-06, + "loss": 0.8259, + "step": 421 + }, + { + "epoch": 0.4875090252707581, + "grad_norm": 0.027380600571632385, + "learning_rate": 4.3755207516027904e-06, + "loss": 0.8394, + "step": 422 + }, + { + "epoch": 0.48866425992779783, + "grad_norm": 0.02569480426609516, + "learning_rate": 4.3772339456250705e-06, + "loss": 0.82, + "step": 423 + }, + { + "epoch": 0.48981949458483753, + "grad_norm": 0.028223766013979912, + "learning_rate": 4.3789430943212215e-06, + "loss": 0.8223, + "step": 424 + }, + { + "epoch": 0.49097472924187724, + "grad_norm": 0.025967473164200783, + "learning_rate": 4.38064821675052e-06, + "loss": 0.8316, + "step": 425 + }, + { + "epoch": 0.49212996389891694, + "grad_norm": 0.026815980672836304, + "learning_rate": 4.382349331837865e-06, + "loss": 0.8497, + "step": 426 + }, + { + "epoch": 0.4932851985559567, + "grad_norm": 0.026486823335289955, + "learning_rate": 4.38404645837504e-06, + "loss": 0.8224, + "step": 427 + }, + { + "epoch": 0.4944404332129964, + "grad_norm": 0.026307394728064537, + "learning_rate": 4.385739615021954e-06, + "loss": 0.8055, + "step": 428 + }, + { + "epoch": 0.4955956678700361, + "grad_norm": 0.02565447799861431, + "learning_rate": 4.387428820307874e-06, + "loss": 0.85, + "step": 429 + }, + { + "epoch": 0.4967509025270758, + "grad_norm": 0.026436088606715202, + "learning_rate": 4.389114092632645e-06, + "loss": 0.8343, + "step": 430 + }, + { + "epoch": 0.4979061371841155, + "grad_norm": 0.028431309387087822, + "learning_rate": 4.390795450267887e-06, + "loss": 0.8457, + "step": 431 + }, + { + "epoch": 0.49906137184115523, + "grad_norm": 0.026016119867563248, + "learning_rate": 4.392472911358188e-06, + "loss": 0.8307, + "step": 432 + }, + { + "epoch": 0.500216606498195, + "grad_norm": 0.02676302008330822, + "learning_rate": 4.394146493922276e-06, + "loss": 0.8565, + "step": 433 + }, + { + "epoch": 0.5013718411552347, + "grad_norm": 0.026580410078167915, + "learning_rate": 4.395816215854185e-06, + "loss": 0.8364, + "step": 434 + }, + { + "epoch": 0.5025270758122744, + "grad_norm": 0.02553938329219818, + "learning_rate": 4.397482094924396e-06, + "loss": 0.8266, + "step": 435 + }, + { + "epoch": 0.5036823104693141, + "grad_norm": 0.02656836435198784, + "learning_rate": 4.3991441487809775e-06, + "loss": 0.8181, + "step": 436 + }, + { + "epoch": 0.5048375451263538, + "grad_norm": 0.025670040398836136, + "learning_rate": 4.400802394950703e-06, + "loss": 0.7895, + "step": 437 + }, + { + "epoch": 0.5059927797833935, + "grad_norm": 0.02619147300720215, + "learning_rate": 4.402456850840167e-06, + "loss": 0.8364, + "step": 438 + }, + { + "epoch": 0.5071480144404332, + "grad_norm": 0.025901198387145996, + "learning_rate": 4.4041075337368695e-06, + "loss": 0.8449, + "step": 439 + }, + { + "epoch": 0.5083032490974729, + "grad_norm": 0.02623150125145912, + "learning_rate": 4.405754460810312e-06, + "loss": 0.8422, + "step": 440 + }, + { + "epoch": 0.5094584837545126, + "grad_norm": 0.025403983891010284, + "learning_rate": 4.407397649113065e-06, + "loss": 0.851, + "step": 441 + }, + { + "epoch": 0.5106137184115523, + "grad_norm": 0.027808185666799545, + "learning_rate": 4.40903711558182e-06, + "loss": 0.8238, + "step": 442 + }, + { + "epoch": 0.511768953068592, + "grad_norm": 0.025581343099474907, + "learning_rate": 4.41067287703845e-06, + "loss": 0.848, + "step": 443 + }, + { + "epoch": 0.5129241877256318, + "grad_norm": 0.026919787749648094, + "learning_rate": 4.4123049501910335e-06, + "loss": 0.8171, + "step": 444 + }, + { + "epoch": 0.5140794223826715, + "grad_norm": 0.025836626067757607, + "learning_rate": 4.4139333516348865e-06, + "loss": 0.8338, + "step": 445 + }, + { + "epoch": 0.5152346570397112, + "grad_norm": 0.026238281279802322, + "learning_rate": 4.415558097853571e-06, + "loss": 0.8164, + "step": 446 + }, + { + "epoch": 0.5163898916967509, + "grad_norm": 0.02682877704501152, + "learning_rate": 4.4171792052198945e-06, + "loss": 0.8414, + "step": 447 + }, + { + "epoch": 0.5175451263537906, + "grad_norm": 0.02541991136968136, + "learning_rate": 4.418796689996907e-06, + "loss": 0.8043, + "step": 448 + }, + { + "epoch": 0.5187003610108303, + "grad_norm": 0.026446549221873283, + "learning_rate": 4.420410568338873e-06, + "loss": 0.8499, + "step": 449 + }, + { + "epoch": 0.51985559566787, + "grad_norm": 0.02652100659906864, + "learning_rate": 4.42202085629224e-06, + "loss": 0.8331, + "step": 450 + }, + { + "epoch": 0.5210108303249098, + "grad_norm": 0.025452058762311935, + "learning_rate": 4.423627569796601e-06, + "loss": 0.8265, + "step": 451 + }, + { + "epoch": 0.5221660649819495, + "grad_norm": 0.025869838893413544, + "learning_rate": 4.425230724685638e-06, + "loss": 0.8276, + "step": 452 + }, + { + "epoch": 0.5233212996389892, + "grad_norm": 0.026051780208945274, + "learning_rate": 4.426830336688054e-06, + "loss": 0.8575, + "step": 453 + }, + { + "epoch": 0.5244765342960289, + "grad_norm": 0.02754388563334942, + "learning_rate": 4.428426421428507e-06, + "loss": 0.8255, + "step": 454 + }, + { + "epoch": 0.5256317689530686, + "grad_norm": 0.02607525885105133, + "learning_rate": 4.430018994428521e-06, + "loss": 0.8397, + "step": 455 + }, + { + "epoch": 0.5267870036101083, + "grad_norm": 0.026636334136128426, + "learning_rate": 4.431608071107392e-06, + "loss": 0.8377, + "step": 456 + }, + { + "epoch": 0.527942238267148, + "grad_norm": 0.026849418878555298, + "learning_rate": 4.433193666783084e-06, + "loss": 0.8257, + "step": 457 + }, + { + "epoch": 0.5290974729241877, + "grad_norm": 0.02537579834461212, + "learning_rate": 4.434775796673116e-06, + "loss": 0.8455, + "step": 458 + }, + { + "epoch": 0.5302527075812274, + "grad_norm": 0.025962024927139282, + "learning_rate": 4.4363544758954355e-06, + "loss": 0.8263, + "step": 459 + }, + { + "epoch": 0.5314079422382672, + "grad_norm": 0.027943387627601624, + "learning_rate": 4.437929719469291e-06, + "loss": 0.8546, + "step": 460 + }, + { + "epoch": 0.5325631768953069, + "grad_norm": 0.02550574764609337, + "learning_rate": 4.439501542316081e-06, + "loss": 0.8367, + "step": 461 + }, + { + "epoch": 0.5337184115523466, + "grad_norm": 0.030560094863176346, + "learning_rate": 4.44106995926021e-06, + "loss": 0.8443, + "step": 462 + }, + { + "epoch": 0.5348736462093863, + "grad_norm": 0.025380682200193405, + "learning_rate": 4.442634985029922e-06, + "loss": 0.8341, + "step": 463 + }, + { + "epoch": 0.536028880866426, + "grad_norm": 0.027356211096048355, + "learning_rate": 4.444196634258136e-06, + "loss": 0.8427, + "step": 464 + }, + { + "epoch": 0.5371841155234657, + "grad_norm": 0.027884816750884056, + "learning_rate": 4.445754921483257e-06, + "loss": 0.8441, + "step": 465 + }, + { + "epoch": 0.5383393501805054, + "grad_norm": 0.025398530066013336, + "learning_rate": 4.44730986115e-06, + "loss": 0.8342, + "step": 466 + }, + { + "epoch": 0.5394945848375451, + "grad_norm": 0.027301618829369545, + "learning_rate": 4.448861467610187e-06, + "loss": 0.8453, + "step": 467 + }, + { + "epoch": 0.5406498194945848, + "grad_norm": 0.025283178314566612, + "learning_rate": 4.450409755123541e-06, + "loss": 0.8468, + "step": 468 + }, + { + "epoch": 0.5418050541516245, + "grad_norm": 0.025396686047315598, + "learning_rate": 4.4519547378584725e-06, + "loss": 0.8352, + "step": 469 + }, + { + "epoch": 0.5429602888086643, + "grad_norm": 0.026122555136680603, + "learning_rate": 4.453496429892863e-06, + "loss": 0.8255, + "step": 470 + }, + { + "epoch": 0.544115523465704, + "grad_norm": 0.026710988953709602, + "learning_rate": 4.455034845214828e-06, + "loss": 0.8033, + "step": 471 + }, + { + "epoch": 0.5452707581227437, + "grad_norm": 0.027530129998922348, + "learning_rate": 4.45656999772348e-06, + "loss": 0.827, + "step": 472 + }, + { + "epoch": 0.5464259927797834, + "grad_norm": 0.02589859999716282, + "learning_rate": 4.458101901229686e-06, + "loss": 0.8585, + "step": 473 + }, + { + "epoch": 0.5475812274368231, + "grad_norm": 0.026990870013833046, + "learning_rate": 4.459630569456809e-06, + "loss": 0.8228, + "step": 474 + }, + { + "epoch": 0.5487364620938628, + "grad_norm": 0.026618408039212227, + "learning_rate": 4.461156016041445e-06, + "loss": 0.8308, + "step": 475 + }, + { + "epoch": 0.5498916967509025, + "grad_norm": 0.026187585666775703, + "learning_rate": 4.4626782545341565e-06, + "loss": 0.841, + "step": 476 + }, + { + "epoch": 0.5510469314079423, + "grad_norm": 0.024839429184794426, + "learning_rate": 4.4641972984001906e-06, + "loss": 0.8265, + "step": 477 + }, + { + "epoch": 0.552202166064982, + "grad_norm": 0.02818799950182438, + "learning_rate": 4.465713161020199e-06, + "loss": 0.8231, + "step": 478 + }, + { + "epoch": 0.5533574007220217, + "grad_norm": 0.02686317451298237, + "learning_rate": 4.46722585569094e-06, + "loss": 0.8266, + "step": 479 + }, + { + "epoch": 0.5545126353790614, + "grad_norm": 0.02528228797018528, + "learning_rate": 4.468735395625979e-06, + "loss": 0.8296, + "step": 480 + }, + { + "epoch": 0.5556678700361011, + "grad_norm": 0.02689657174050808, + "learning_rate": 4.470241793956387e-06, + "loss": 0.8345, + "step": 481 + }, + { + "epoch": 0.5568231046931408, + "grad_norm": 0.028286360204219818, + "learning_rate": 4.471745063731417e-06, + "loss": 0.8371, + "step": 482 + }, + { + "epoch": 0.5579783393501805, + "grad_norm": 0.0263553187251091, + "learning_rate": 4.473245217919187e-06, + "loss": 0.8183, + "step": 483 + }, + { + "epoch": 0.5591335740072202, + "grad_norm": 0.027919506654143333, + "learning_rate": 4.474742269407355e-06, + "loss": 0.8317, + "step": 484 + }, + { + "epoch": 0.5602888086642599, + "grad_norm": 0.02882731519639492, + "learning_rate": 4.476236231003773e-06, + "loss": 0.8414, + "step": 485 + }, + { + "epoch": 0.5614440433212996, + "grad_norm": 0.02561136521399021, + "learning_rate": 4.477727115437156e-06, + "loss": 0.8004, + "step": 486 + }, + { + "epoch": 0.5625992779783393, + "grad_norm": 0.02664412185549736, + "learning_rate": 4.479214935357724e-06, + "loss": 0.7967, + "step": 487 + }, + { + "epoch": 0.563754512635379, + "grad_norm": 0.028001073747873306, + "learning_rate": 4.480699703337852e-06, + "loss": 0.841, + "step": 488 + }, + { + "epoch": 0.5649097472924187, + "grad_norm": 0.026740211993455887, + "learning_rate": 4.4821814318727016e-06, + "loss": 0.8369, + "step": 489 + }, + { + "epoch": 0.5660649819494585, + "grad_norm": 0.024850796908140182, + "learning_rate": 4.4836601333808566e-06, + "loss": 0.8405, + "step": 490 + }, + { + "epoch": 0.5672202166064982, + "grad_norm": 0.025452926754951477, + "learning_rate": 4.485135820204948e-06, + "loss": 0.8106, + "step": 491 + }, + { + "epoch": 0.568375451263538, + "grad_norm": 0.026447713375091553, + "learning_rate": 4.4866085046122675e-06, + "loss": 0.7992, + "step": 492 + }, + { + "epoch": 0.5695306859205777, + "grad_norm": 0.027519937604665756, + "learning_rate": 4.488078198795384e-06, + "loss": 0.8227, + "step": 493 + }, + { + "epoch": 0.5706859205776174, + "grad_norm": 0.026447774842381477, + "learning_rate": 4.4895449148727455e-06, + "loss": 0.8377, + "step": 494 + }, + { + "epoch": 0.5718411552346571, + "grad_norm": 0.02749788761138916, + "learning_rate": 4.491008664889282e-06, + "loss": 0.789, + "step": 495 + }, + { + "epoch": 0.5729963898916968, + "grad_norm": 0.025579964742064476, + "learning_rate": 4.492469460816997e-06, + "loss": 0.8288, + "step": 496 + }, + { + "epoch": 0.5741516245487365, + "grad_norm": 0.027195453643798828, + "learning_rate": 4.493927314555554e-06, + "loss": 0.8194, + "step": 497 + }, + { + "epoch": 0.5753068592057762, + "grad_norm": 0.027325714007019997, + "learning_rate": 4.495382237932863e-06, + "loss": 0.8136, + "step": 498 + }, + { + "epoch": 0.5764620938628159, + "grad_norm": 0.026480497792363167, + "learning_rate": 4.496834242705651e-06, + "loss": 0.8239, + "step": 499 + }, + { + "epoch": 0.5776173285198556, + "grad_norm": 0.026893138885498047, + "learning_rate": 4.498283340560032e-06, + "loss": 0.8234, + "step": 500 + }, + { + "epoch": 0.5787725631768953, + "grad_norm": 0.02725539728999138, + "learning_rate": 4.4997295431120764e-06, + "loss": 0.8167, + "step": 501 + }, + { + "epoch": 0.579927797833935, + "grad_norm": 0.027873387560248375, + "learning_rate": 4.5011728619083665e-06, + "loss": 0.8012, + "step": 502 + }, + { + "epoch": 0.5810830324909747, + "grad_norm": 0.027902130037546158, + "learning_rate": 4.502613308426546e-06, + "loss": 0.8248, + "step": 503 + }, + { + "epoch": 0.5822382671480144, + "grad_norm": 0.027641048654913902, + "learning_rate": 4.504050894075876e-06, + "loss": 0.8298, + "step": 504 + }, + { + "epoch": 0.5833935018050541, + "grad_norm": 0.026641003787517548, + "learning_rate": 4.5054856301977696e-06, + "loss": 0.8228, + "step": 505 + }, + { + "epoch": 0.5845487364620938, + "grad_norm": 0.028835156932473183, + "learning_rate": 4.506917528066332e-06, + "loss": 0.8222, + "step": 506 + }, + { + "epoch": 0.5857039711191335, + "grad_norm": 0.02770557627081871, + "learning_rate": 4.5083465988888945e-06, + "loss": 0.8455, + "step": 507 + }, + { + "epoch": 0.5868592057761732, + "grad_norm": 0.0266465675085783, + "learning_rate": 4.509772853806533e-06, + "loss": 0.8269, + "step": 508 + }, + { + "epoch": 0.588014440433213, + "grad_norm": 0.029550906270742416, + "learning_rate": 4.511196303894598e-06, + "loss": 0.817, + "step": 509 + }, + { + "epoch": 0.5891696750902528, + "grad_norm": 0.025802530348300934, + "learning_rate": 4.512616960163228e-06, + "loss": 0.8352, + "step": 510 + }, + { + "epoch": 0.5903249097472925, + "grad_norm": 0.02622455172240734, + "learning_rate": 4.514034833557855e-06, + "loss": 0.8209, + "step": 511 + }, + { + "epoch": 0.5914801444043322, + "grad_norm": 0.028669588267803192, + "learning_rate": 4.515449934959719e-06, + "loss": 0.8199, + "step": 512 + }, + { + "epoch": 0.5926353790613719, + "grad_norm": 0.02638787031173706, + "learning_rate": 4.516862275186361e-06, + "loss": 0.8396, + "step": 513 + }, + { + "epoch": 0.5937906137184116, + "grad_norm": 0.025254135951399803, + "learning_rate": 4.518271864992127e-06, + "loss": 0.8314, + "step": 514 + }, + { + "epoch": 0.5949458483754513, + "grad_norm": 0.026823000982403755, + "learning_rate": 4.519678715068652e-06, + "loss": 0.8063, + "step": 515 + }, + { + "epoch": 0.596101083032491, + "grad_norm": 0.029466571286320686, + "learning_rate": 4.521082836045353e-06, + "loss": 0.8279, + "step": 516 + }, + { + "epoch": 0.5972563176895307, + "grad_norm": 0.02610476315021515, + "learning_rate": 4.5224842384899045e-06, + "loss": 0.8217, + "step": 517 + }, + { + "epoch": 0.5984115523465704, + "grad_norm": 0.027522264048457146, + "learning_rate": 4.523882932908722e-06, + "loss": 0.8465, + "step": 518 + }, + { + "epoch": 0.5995667870036101, + "grad_norm": 0.025692617520689964, + "learning_rate": 4.52527892974743e-06, + "loss": 0.8448, + "step": 519 + }, + { + "epoch": 0.6007220216606498, + "grad_norm": 0.027035709470510483, + "learning_rate": 4.526672239391333e-06, + "loss": 0.8216, + "step": 520 + }, + { + "epoch": 0.6018772563176895, + "grad_norm": 0.026100315153598785, + "learning_rate": 4.528062872165875e-06, + "loss": 0.8007, + "step": 521 + }, + { + "epoch": 0.6030324909747292, + "grad_norm": 0.025961345061659813, + "learning_rate": 4.529450838337104e-06, + "loss": 0.8185, + "step": 522 + }, + { + "epoch": 0.6041877256317689, + "grad_norm": 0.02683369629085064, + "learning_rate": 4.5308361481121244e-06, + "loss": 0.8377, + "step": 523 + }, + { + "epoch": 0.6053429602888086, + "grad_norm": 0.02755955420434475, + "learning_rate": 4.532218811639545e-06, + "loss": 0.8197, + "step": 524 + }, + { + "epoch": 0.6064981949458483, + "grad_norm": 0.02744685672223568, + "learning_rate": 4.533598839009929e-06, + "loss": 0.8146, + "step": 525 + }, + { + "epoch": 0.607653429602888, + "grad_norm": 0.026420993730425835, + "learning_rate": 4.5349762402562325e-06, + "loss": 0.8324, + "step": 526 + }, + { + "epoch": 0.6088086642599277, + "grad_norm": 0.026225924491882324, + "learning_rate": 4.5363510253542444e-06, + "loss": 0.8299, + "step": 527 + }, + { + "epoch": 0.6099638989169676, + "grad_norm": 0.025415120646357536, + "learning_rate": 4.537723204223021e-06, + "loss": 0.8373, + "step": 528 + }, + { + "epoch": 0.6111191335740073, + "grad_norm": 0.026079954579472542, + "learning_rate": 4.53909278672531e-06, + "loss": 0.8064, + "step": 529 + }, + { + "epoch": 0.612274368231047, + "grad_norm": 0.027824481949210167, + "learning_rate": 4.540459782667983e-06, + "loss": 0.834, + "step": 530 + }, + { + "epoch": 0.6134296028880867, + "grad_norm": 0.02630164660513401, + "learning_rate": 4.541824201802449e-06, + "loss": 0.8341, + "step": 531 + }, + { + "epoch": 0.6145848375451264, + "grad_norm": 0.025940794497728348, + "learning_rate": 4.543186053825081e-06, + "loss": 0.8356, + "step": 532 + }, + { + "epoch": 0.6157400722021661, + "grad_norm": 0.025478770956397057, + "learning_rate": 4.544545348377621e-06, + "loss": 0.8118, + "step": 533 + }, + { + "epoch": 0.6168953068592058, + "grad_norm": 0.02738497406244278, + "learning_rate": 4.545902095047594e-06, + "loss": 0.8397, + "step": 534 + }, + { + "epoch": 0.6180505415162455, + "grad_norm": 0.025361735373735428, + "learning_rate": 4.547256303368714e-06, + "loss": 0.8422, + "step": 535 + }, + { + "epoch": 0.6192057761732852, + "grad_norm": 0.02575971558690071, + "learning_rate": 4.5486079828212835e-06, + "loss": 0.8224, + "step": 536 + }, + { + "epoch": 0.6203610108303249, + "grad_norm": 0.026920614764094353, + "learning_rate": 4.549957142832593e-06, + "loss": 0.7988, + "step": 537 + }, + { + "epoch": 0.6215162454873646, + "grad_norm": 0.026717036962509155, + "learning_rate": 4.5513037927773155e-06, + "loss": 0.8397, + "step": 538 + }, + { + "epoch": 0.6226714801444043, + "grad_norm": 0.028026578947901726, + "learning_rate": 4.552647941977898e-06, + "loss": 0.8421, + "step": 539 + }, + { + "epoch": 0.623826714801444, + "grad_norm": 0.02701553702354431, + "learning_rate": 4.553989599704948e-06, + "loss": 0.8129, + "step": 540 + }, + { + "epoch": 0.6249819494584837, + "grad_norm": 0.028897127136588097, + "learning_rate": 4.555328775177616e-06, + "loss": 0.8049, + "step": 541 + }, + { + "epoch": 0.6261371841155234, + "grad_norm": 0.027024636045098305, + "learning_rate": 4.556665477563979e-06, + "loss": 0.8255, + "step": 542 + }, + { + "epoch": 0.6272924187725631, + "grad_norm": 0.025836756452918053, + "learning_rate": 4.557999715981412e-06, + "loss": 0.8172, + "step": 543 + }, + { + "epoch": 0.6284476534296028, + "grad_norm": 0.027048436924815178, + "learning_rate": 4.559331499496967e-06, + "loss": 0.8336, + "step": 544 + }, + { + "epoch": 0.6296028880866426, + "grad_norm": 0.02669835276901722, + "learning_rate": 4.560660837127738e-06, + "loss": 0.7958, + "step": 545 + }, + { + "epoch": 0.6307581227436823, + "grad_norm": 0.02850779891014099, + "learning_rate": 4.5619877378412295e-06, + "loss": 0.8313, + "step": 546 + }, + { + "epoch": 0.6319133574007221, + "grad_norm": 0.027087301015853882, + "learning_rate": 4.563312210555719e-06, + "loss": 0.8257, + "step": 547 + }, + { + "epoch": 0.6330685920577618, + "grad_norm": 0.029034119099378586, + "learning_rate": 4.564634264140616e-06, + "loss": 0.8292, + "step": 548 + }, + { + "epoch": 0.6342238267148015, + "grad_norm": 0.02700372040271759, + "learning_rate": 4.5659539074168204e-06, + "loss": 0.828, + "step": 549 + }, + { + "epoch": 0.6353790613718412, + "grad_norm": 0.027077239006757736, + "learning_rate": 4.5672711491570735e-06, + "loss": 0.829, + "step": 550 + }, + { + "epoch": 0.6365342960288809, + "grad_norm": 0.027820030227303505, + "learning_rate": 4.568585998086309e-06, + "loss": 0.8323, + "step": 551 + }, + { + "epoch": 0.6376895306859206, + "grad_norm": 0.02544805407524109, + "learning_rate": 4.569898462881998e-06, + "loss": 0.8424, + "step": 552 + }, + { + "epoch": 0.6388447653429603, + "grad_norm": 0.025698795914649963, + "learning_rate": 4.571208552174497e-06, + "loss": 0.8471, + "step": 553 + }, + { + "epoch": 0.64, + "grad_norm": 0.026754576712846756, + "learning_rate": 4.572516274547383e-06, + "loss": 0.8277, + "step": 554 + }, + { + "epoch": 0.6411552346570397, + "grad_norm": 0.026160219684243202, + "learning_rate": 4.573821638537794e-06, + "loss": 0.8458, + "step": 555 + }, + { + "epoch": 0.6423104693140794, + "grad_norm": 0.02805655263364315, + "learning_rate": 4.5751246526367635e-06, + "loss": 0.8196, + "step": 556 + }, + { + "epoch": 0.6434657039711191, + "grad_norm": 0.02578035369515419, + "learning_rate": 4.576425325289549e-06, + "loss": 0.8235, + "step": 557 + }, + { + "epoch": 0.6446209386281588, + "grad_norm": 0.02676086314022541, + "learning_rate": 4.577723664895965e-06, + "loss": 0.7819, + "step": 558 + }, + { + "epoch": 0.6457761732851985, + "grad_norm": 0.02770860493183136, + "learning_rate": 4.579019679810706e-06, + "loss": 0.8063, + "step": 559 + }, + { + "epoch": 0.6469314079422382, + "grad_norm": 0.025705819949507713, + "learning_rate": 4.5803133783436676e-06, + "loss": 0.8145, + "step": 560 + }, + { + "epoch": 0.648086642599278, + "grad_norm": 0.025856079533696175, + "learning_rate": 4.5816047687602695e-06, + "loss": 0.7849, + "step": 561 + }, + { + "epoch": 0.6492418772563177, + "grad_norm": 0.026201602071523666, + "learning_rate": 4.582893859281769e-06, + "loss": 0.8262, + "step": 562 + }, + { + "epoch": 0.6503971119133574, + "grad_norm": 0.026211684569716454, + "learning_rate": 4.584180658085578e-06, + "loss": 0.8371, + "step": 563 + }, + { + "epoch": 0.6515523465703971, + "grad_norm": 0.02780756726861, + "learning_rate": 4.585465173305571e-06, + "loss": 0.8258, + "step": 564 + }, + { + "epoch": 0.6527075812274368, + "grad_norm": 0.02710585482418537, + "learning_rate": 4.586747413032399e-06, + "loss": 0.8289, + "step": 565 + }, + { + "epoch": 0.6538628158844766, + "grad_norm": 0.02654946967959404, + "learning_rate": 4.588027385313786e-06, + "loss": 0.8208, + "step": 566 + }, + { + "epoch": 0.6550180505415163, + "grad_norm": 0.025670049712061882, + "learning_rate": 4.589305098154845e-06, + "loss": 0.8236, + "step": 567 + }, + { + "epoch": 0.656173285198556, + "grad_norm": 0.025593513622879982, + "learning_rate": 4.5905805595183656e-06, + "loss": 0.8397, + "step": 568 + }, + { + "epoch": 0.6573285198555957, + "grad_norm": 0.02713758684694767, + "learning_rate": 4.59185377732512e-06, + "loss": 0.8077, + "step": 569 + }, + { + "epoch": 0.6584837545126354, + "grad_norm": 0.025683598592877388, + "learning_rate": 4.5931247594541535e-06, + "loss": 0.828, + "step": 570 + }, + { + "epoch": 0.6596389891696751, + "grad_norm": 0.027867255732417107, + "learning_rate": 4.594393513743081e-06, + "loss": 0.8053, + "step": 571 + }, + { + "epoch": 0.6607942238267148, + "grad_norm": 0.02871118299663067, + "learning_rate": 4.595660047988374e-06, + "loss": 0.8308, + "step": 572 + }, + { + "epoch": 0.6619494584837545, + "grad_norm": 0.02591397613286972, + "learning_rate": 4.59692436994565e-06, + "loss": 0.8376, + "step": 573 + }, + { + "epoch": 0.6631046931407942, + "grad_norm": 0.02970442920923233, + "learning_rate": 4.598186487329957e-06, + "loss": 0.8343, + "step": 574 + }, + { + "epoch": 0.6642599277978339, + "grad_norm": 0.027048716321587563, + "learning_rate": 4.599446407816052e-06, + "loss": 0.8169, + "step": 575 + }, + { + "epoch": 0.6654151624548736, + "grad_norm": 0.027099261060357094, + "learning_rate": 4.600704139038687e-06, + "loss": 0.8414, + "step": 576 + }, + { + "epoch": 0.6665703971119133, + "grad_norm": 0.02764103189110756, + "learning_rate": 4.601959688592886e-06, + "loss": 0.8144, + "step": 577 + }, + { + "epoch": 0.667725631768953, + "grad_norm": 0.025973627343773842, + "learning_rate": 4.603213064034216e-06, + "loss": 0.8086, + "step": 578 + }, + { + "epoch": 0.6688808664259928, + "grad_norm": 0.02816241979598999, + "learning_rate": 4.6044642728790615e-06, + "loss": 0.8266, + "step": 579 + }, + { + "epoch": 0.6700361010830325, + "grad_norm": 0.02614632435142994, + "learning_rate": 4.605713322604896e-06, + "loss": 0.8365, + "step": 580 + }, + { + "epoch": 0.6711913357400722, + "grad_norm": 0.02658606320619583, + "learning_rate": 4.606960220650551e-06, + "loss": 0.8273, + "step": 581 + }, + { + "epoch": 0.6723465703971119, + "grad_norm": 0.02635134756565094, + "learning_rate": 4.608204974416482e-06, + "loss": 0.8169, + "step": 582 + }, + { + "epoch": 0.6735018050541516, + "grad_norm": 0.025622064247727394, + "learning_rate": 4.609447591265024e-06, + "loss": 0.8256, + "step": 583 + }, + { + "epoch": 0.6746570397111913, + "grad_norm": 0.02630884386599064, + "learning_rate": 4.610688078520667e-06, + "loss": 0.8546, + "step": 584 + }, + { + "epoch": 0.675812274368231, + "grad_norm": 0.026515616104006767, + "learning_rate": 4.611926443470302e-06, + "loss": 0.829, + "step": 585 + }, + { + "epoch": 0.6769675090252708, + "grad_norm": 0.025966297835111618, + "learning_rate": 4.613162693363485e-06, + "loss": 0.8206, + "step": 586 + }, + { + "epoch": 0.6781227436823105, + "grad_norm": 0.025645708665251732, + "learning_rate": 4.614396835412691e-06, + "loss": 0.8214, + "step": 587 + }, + { + "epoch": 0.6792779783393502, + "grad_norm": 0.026266956701874733, + "learning_rate": 4.6156288767935645e-06, + "loss": 0.8128, + "step": 588 + }, + { + "epoch": 0.6804332129963899, + "grad_norm": 0.027987468987703323, + "learning_rate": 4.61685882464517e-06, + "loss": 0.8269, + "step": 589 + }, + { + "epoch": 0.6815884476534296, + "grad_norm": 0.02621353790163994, + "learning_rate": 4.61808668607024e-06, + "loss": 0.817, + "step": 590 + }, + { + "epoch": 0.6827436823104693, + "grad_norm": 0.026081636548042297, + "learning_rate": 4.619312468135426e-06, + "loss": 0.8168, + "step": 591 + }, + { + "epoch": 0.683898916967509, + "grad_norm": 0.026946187019348145, + "learning_rate": 4.620536177871534e-06, + "loss": 0.8208, + "step": 592 + }, + { + "epoch": 0.6850541516245487, + "grad_norm": 0.02429202012717724, + "learning_rate": 4.621757822273772e-06, + "loss": 0.8032, + "step": 593 + }, + { + "epoch": 0.6862093862815885, + "grad_norm": 0.027351096272468567, + "learning_rate": 4.62297740830199e-06, + "loss": 0.8234, + "step": 594 + }, + { + "epoch": 0.6873646209386282, + "grad_norm": 0.026615051552653313, + "learning_rate": 4.624194942880917e-06, + "loss": 0.8205, + "step": 595 + }, + { + "epoch": 0.6885198555956679, + "grad_norm": 0.025758862495422363, + "learning_rate": 4.625410432900395e-06, + "loss": 0.8372, + "step": 596 + }, + { + "epoch": 0.6896750902527076, + "grad_norm": 0.026305649429559708, + "learning_rate": 4.626623885215616e-06, + "loss": 0.813, + "step": 597 + }, + { + "epoch": 0.6908303249097473, + "grad_norm": 0.029083121567964554, + "learning_rate": 4.627835306647352e-06, + "loss": 0.801, + "step": 598 + }, + { + "epoch": 0.691985559566787, + "grad_norm": 0.02733645774424076, + "learning_rate": 4.6290447039821865e-06, + "loss": 0.818, + "step": 599 + }, + { + "epoch": 0.6931407942238267, + "grad_norm": 0.026958592236042023, + "learning_rate": 4.63025208397274e-06, + "loss": 0.8189, + "step": 600 + }, + { + "epoch": 0.6942960288808664, + "grad_norm": 0.02624497003853321, + "learning_rate": 4.631457453337899e-06, + "loss": 0.8217, + "step": 601 + }, + { + "epoch": 0.6954512635379061, + "grad_norm": 0.026590224355459213, + "learning_rate": 4.632660818763041e-06, + "loss": 0.8413, + "step": 602 + }, + { + "epoch": 0.6966064981949458, + "grad_norm": 0.029455525800585747, + "learning_rate": 4.633862186900253e-06, + "loss": 0.8437, + "step": 603 + }, + { + "epoch": 0.6977617328519855, + "grad_norm": 0.026593301445245743, + "learning_rate": 4.635061564368554e-06, + "loss": 0.8396, + "step": 604 + }, + { + "epoch": 0.6989169675090253, + "grad_norm": 0.026324933394789696, + "learning_rate": 4.636258957754115e-06, + "loss": 0.828, + "step": 605 + }, + { + "epoch": 0.700072202166065, + "grad_norm": 0.027765916660428047, + "learning_rate": 4.6374543736104774e-06, + "loss": 0.8403, + "step": 606 + }, + { + "epoch": 0.7012274368231047, + "grad_norm": 0.026490546762943268, + "learning_rate": 4.638647818458763e-06, + "loss": 0.8206, + "step": 607 + }, + { + "epoch": 0.7023826714801444, + "grad_norm": 0.028683004900813103, + "learning_rate": 4.639839298787892e-06, + "loss": 0.8203, + "step": 608 + }, + { + "epoch": 0.7035379061371841, + "grad_norm": 0.026209495961666107, + "learning_rate": 4.641028821054792e-06, + "loss": 0.8127, + "step": 609 + }, + { + "epoch": 0.7046931407942238, + "grad_norm": 0.02771477773785591, + "learning_rate": 4.642216391684613e-06, + "loss": 0.8218, + "step": 610 + }, + { + "epoch": 0.7058483754512636, + "grad_norm": 0.02859386056661606, + "learning_rate": 4.643402017070924e-06, + "loss": 0.7877, + "step": 611 + }, + { + "epoch": 0.7070036101083033, + "grad_norm": 0.026568232104182243, + "learning_rate": 4.644585703575936e-06, + "loss": 0.8062, + "step": 612 + }, + { + "epoch": 0.708158844765343, + "grad_norm": 0.030504100024700165, + "learning_rate": 4.645767457530693e-06, + "loss": 0.8068, + "step": 613 + }, + { + "epoch": 0.7093140794223827, + "grad_norm": 0.026128064841032028, + "learning_rate": 4.64694728523528e-06, + "loss": 0.8185, + "step": 614 + }, + { + "epoch": 0.7104693140794224, + "grad_norm": 0.026173869147896767, + "learning_rate": 4.648125192959028e-06, + "loss": 0.8305, + "step": 615 + }, + { + "epoch": 0.7116245487364621, + "grad_norm": 0.02751108445227146, + "learning_rate": 4.649301186940709e-06, + "loss": 0.8176, + "step": 616 + }, + { + "epoch": 0.7127797833935018, + "grad_norm": 0.026463057845830917, + "learning_rate": 4.650475273388737e-06, + "loss": 0.8011, + "step": 617 + }, + { + "epoch": 0.7139350180505415, + "grad_norm": 0.026889406144618988, + "learning_rate": 4.65164745848136e-06, + "loss": 0.8319, + "step": 618 + }, + { + "epoch": 0.7150902527075812, + "grad_norm": 0.02862033247947693, + "learning_rate": 4.652817748366864e-06, + "loss": 0.8394, + "step": 619 + }, + { + "epoch": 0.7162454873646209, + "grad_norm": 0.025286676362156868, + "learning_rate": 4.653986149163757e-06, + "loss": 0.825, + "step": 620 + }, + { + "epoch": 0.7174007220216606, + "grad_norm": 0.026653330773115158, + "learning_rate": 4.655152666960967e-06, + "loss": 0.7891, + "step": 621 + }, + { + "epoch": 0.7185559566787003, + "grad_norm": 0.027324773371219635, + "learning_rate": 4.656317307818032e-06, + "loss": 0.8105, + "step": 622 + }, + { + "epoch": 0.71971119133574, + "grad_norm": 0.026749735698103905, + "learning_rate": 4.657480077765284e-06, + "loss": 0.8125, + "step": 623 + }, + { + "epoch": 0.7208664259927798, + "grad_norm": 0.027720510959625244, + "learning_rate": 4.6586409828040405e-06, + "loss": 0.8385, + "step": 624 + }, + { + "epoch": 0.7220216606498195, + "grad_norm": 0.02868843637406826, + "learning_rate": 4.659800028906792e-06, + "loss": 0.8221, + "step": 625 + }, + { + "epoch": 0.7231768953068592, + "grad_norm": 0.026634184643626213, + "learning_rate": 4.660957222017384e-06, + "loss": 0.824, + "step": 626 + }, + { + "epoch": 0.724332129963899, + "grad_norm": 0.02601156197488308, + "learning_rate": 4.662112568051194e-06, + "loss": 0.813, + "step": 627 + }, + { + "epoch": 0.7254873646209387, + "grad_norm": 0.028468729928135872, + "learning_rate": 4.663266072895327e-06, + "loss": 0.8229, + "step": 628 + }, + { + "epoch": 0.7266425992779784, + "grad_norm": 0.026884114369750023, + "learning_rate": 4.664417742408782e-06, + "loss": 0.8201, + "step": 629 + }, + { + "epoch": 0.7277978339350181, + "grad_norm": 0.02819686383008957, + "learning_rate": 4.6655675824226375e-06, + "loss": 0.8048, + "step": 630 + }, + { + "epoch": 0.7289530685920578, + "grad_norm": 0.02609233930706978, + "learning_rate": 4.666715598740224e-06, + "loss": 0.8145, + "step": 631 + }, + { + "epoch": 0.7301083032490975, + "grad_norm": 0.025876475498080254, + "learning_rate": 4.667861797137309e-06, + "loss": 0.8236, + "step": 632 + }, + { + "epoch": 0.7312635379061372, + "grad_norm": 0.025990156456828117, + "learning_rate": 4.669006183362259e-06, + "loss": 0.8416, + "step": 633 + }, + { + "epoch": 0.7324187725631769, + "grad_norm": 0.026048380881547928, + "learning_rate": 4.670148763136221e-06, + "loss": 0.8004, + "step": 634 + }, + { + "epoch": 0.7335740072202166, + "grad_norm": 0.0267790500074625, + "learning_rate": 4.6712895421532935e-06, + "loss": 0.8249, + "step": 635 + }, + { + "epoch": 0.7347292418772563, + "grad_norm": 0.027045415714383125, + "learning_rate": 4.672428526080691e-06, + "loss": 0.8346, + "step": 636 + }, + { + "epoch": 0.735884476534296, + "grad_norm": 0.026687374338507652, + "learning_rate": 4.673565720558918e-06, + "loss": 0.8206, + "step": 637 + }, + { + "epoch": 0.7370397111913357, + "grad_norm": 0.025013815611600876, + "learning_rate": 4.674701131201937e-06, + "loss": 0.8278, + "step": 638 + }, + { + "epoch": 0.7381949458483754, + "grad_norm": 0.02536069042980671, + "learning_rate": 4.675834763597334e-06, + "loss": 0.8051, + "step": 639 + }, + { + "epoch": 0.7393501805054151, + "grad_norm": 0.025835523381829262, + "learning_rate": 4.676966623306479e-06, + "loss": 0.8131, + "step": 640 + }, + { + "epoch": 0.7405054151624548, + "grad_norm": 0.026328476145863533, + "learning_rate": 4.678096715864696e-06, + "loss": 0.834, + "step": 641 + }, + { + "epoch": 0.7416606498194945, + "grad_norm": 0.027017908170819283, + "learning_rate": 4.679225046781422e-06, + "loss": 0.808, + "step": 642 + }, + { + "epoch": 0.7428158844765343, + "grad_norm": 0.026868514716625214, + "learning_rate": 4.6803516215403704e-06, + "loss": 0.8236, + "step": 643 + }, + { + "epoch": 0.743971119133574, + "grad_norm": 0.026412272825837135, + "learning_rate": 4.6814764455996875e-06, + "loss": 0.8343, + "step": 644 + }, + { + "epoch": 0.7451263537906138, + "grad_norm": 0.025585437193512917, + "learning_rate": 4.682599524392114e-06, + "loss": 0.8165, + "step": 645 + }, + { + "epoch": 0.7462815884476535, + "grad_norm": 0.026194840669631958, + "learning_rate": 4.683720863325141e-06, + "loss": 0.8284, + "step": 646 + }, + { + "epoch": 0.7474368231046932, + "grad_norm": 0.025945279747247696, + "learning_rate": 4.6848404677811685e-06, + "loss": 0.8213, + "step": 647 + }, + { + "epoch": 0.7485920577617329, + "grad_norm": 0.02698802575469017, + "learning_rate": 4.685958343117656e-06, + "loss": 0.8399, + "step": 648 + }, + { + "epoch": 0.7497472924187726, + "grad_norm": 0.02705751173198223, + "learning_rate": 4.687074494667283e-06, + "loss": 0.8254, + "step": 649 + }, + { + "epoch": 0.7509025270758123, + "grad_norm": 0.02601143904030323, + "learning_rate": 4.688188927738093e-06, + "loss": 0.8283, + "step": 650 + }, + { + "epoch": 0.752057761732852, + "grad_norm": 0.026112712919712067, + "learning_rate": 4.689301647613653e-06, + "loss": 0.8334, + "step": 651 + }, + { + "epoch": 0.7532129963898917, + "grad_norm": 0.02763291820883751, + "learning_rate": 4.690412659553201e-06, + "loss": 0.8195, + "step": 652 + }, + { + "epoch": 0.7543682310469314, + "grad_norm": 0.02655601128935814, + "learning_rate": 4.69152196879179e-06, + "loss": 0.8164, + "step": 653 + }, + { + "epoch": 0.7555234657039711, + "grad_norm": 0.024583589285612106, + "learning_rate": 4.692629580540446e-06, + "loss": 0.8112, + "step": 654 + }, + { + "epoch": 0.7566787003610108, + "grad_norm": 0.02777348831295967, + "learning_rate": 4.693735499986305e-06, + "loss": 0.8228, + "step": 655 + }, + { + "epoch": 0.7578339350180505, + "grad_norm": 0.025867123156785965, + "learning_rate": 4.694839732292768e-06, + "loss": 0.8169, + "step": 656 + }, + { + "epoch": 0.7589891696750902, + "grad_norm": 0.02569735422730446, + "learning_rate": 4.695942282599635e-06, + "loss": 0.8274, + "step": 657 + }, + { + "epoch": 0.7601444043321299, + "grad_norm": 0.026997772976756096, + "learning_rate": 4.69704315602326e-06, + "loss": 0.8127, + "step": 658 + }, + { + "epoch": 0.7612996389891696, + "grad_norm": 0.026205215603113174, + "learning_rate": 4.698142357656684e-06, + "loss": 0.8053, + "step": 659 + }, + { + "epoch": 0.7624548736462093, + "grad_norm": 0.026929359883069992, + "learning_rate": 4.699239892569782e-06, + "loss": 0.8049, + "step": 660 + }, + { + "epoch": 0.763610108303249, + "grad_norm": 0.02786414325237274, + "learning_rate": 4.700335765809401e-06, + "loss": 0.8098, + "step": 661 + }, + { + "epoch": 0.7647653429602889, + "grad_norm": 0.02785409241914749, + "learning_rate": 4.701429982399501e-06, + "loss": 0.8203, + "step": 662 + }, + { + "epoch": 0.7659205776173286, + "grad_norm": 0.02913859114050865, + "learning_rate": 4.702522547341289e-06, + "loss": 0.8174, + "step": 663 + }, + { + "epoch": 0.7670758122743683, + "grad_norm": 0.025960449129343033, + "learning_rate": 4.703613465613363e-06, + "loss": 0.8111, + "step": 664 + }, + { + "epoch": 0.768231046931408, + "grad_norm": 0.027514390647411346, + "learning_rate": 4.704702742171842e-06, + "loss": 0.7999, + "step": 665 + }, + { + "epoch": 0.7693862815884477, + "grad_norm": 0.026732420548796654, + "learning_rate": 4.705790381950503e-06, + "loss": 0.8332, + "step": 666 + }, + { + "epoch": 0.7705415162454874, + "grad_norm": 0.025522086769342422, + "learning_rate": 4.706876389860915e-06, + "loss": 0.8179, + "step": 667 + }, + { + "epoch": 0.7716967509025271, + "grad_norm": 0.026885733008384705, + "learning_rate": 4.707960770792576e-06, + "loss": 0.7875, + "step": 668 + }, + { + "epoch": 0.7728519855595668, + "grad_norm": 0.026839058846235275, + "learning_rate": 4.709043529613039e-06, + "loss": 0.8029, + "step": 669 + }, + { + "epoch": 0.7740072202166065, + "grad_norm": 0.02562631480395794, + "learning_rate": 4.710124671168044e-06, + "loss": 0.8184, + "step": 670 + }, + { + "epoch": 0.7751624548736462, + "grad_norm": 0.026263324543833733, + "learning_rate": 4.7112042002816544e-06, + "loss": 0.8192, + "step": 671 + }, + { + "epoch": 0.7763176895306859, + "grad_norm": 0.026700858026742935, + "learning_rate": 4.7122821217563755e-06, + "loss": 0.827, + "step": 672 + }, + { + "epoch": 0.7774729241877256, + "grad_norm": 0.02566208317875862, + "learning_rate": 4.7133584403732955e-06, + "loss": 0.8072, + "step": 673 + }, + { + "epoch": 0.7786281588447653, + "grad_norm": 0.02590363658964634, + "learning_rate": 4.7144331608922e-06, + "loss": 0.8111, + "step": 674 + }, + { + "epoch": 0.779783393501805, + "grad_norm": 0.025736572220921516, + "learning_rate": 4.715506288051709e-06, + "loss": 0.8077, + "step": 675 + }, + { + "epoch": 0.7809386281588447, + "grad_norm": 0.027111530303955078, + "learning_rate": 4.716577826569394e-06, + "loss": 0.8429, + "step": 676 + }, + { + "epoch": 0.7820938628158844, + "grad_norm": 0.02599494345486164, + "learning_rate": 4.717647781141908e-06, + "loss": 0.8125, + "step": 677 + }, + { + "epoch": 0.7832490974729241, + "grad_norm": 0.02675134316086769, + "learning_rate": 4.7187161564451065e-06, + "loss": 0.8131, + "step": 678 + }, + { + "epoch": 0.7844043321299639, + "grad_norm": 0.026113273575901985, + "learning_rate": 4.71978295713417e-06, + "loss": 0.8292, + "step": 679 + }, + { + "epoch": 0.7855595667870036, + "grad_norm": 0.024617226794362068, + "learning_rate": 4.720848187843727e-06, + "loss": 0.8032, + "step": 680 + }, + { + "epoch": 0.7867148014440434, + "grad_norm": 0.02645159512758255, + "learning_rate": 4.721911853187975e-06, + "loss": 0.8287, + "step": 681 + }, + { + "epoch": 0.7878700361010831, + "grad_norm": 0.026287630200386047, + "learning_rate": 4.722973957760799e-06, + "loss": 0.8098, + "step": 682 + }, + { + "epoch": 0.7890252707581228, + "grad_norm": 0.02640565298497677, + "learning_rate": 4.724034506135888e-06, + "loss": 0.8303, + "step": 683 + }, + { + "epoch": 0.7901805054151625, + "grad_norm": 0.027293162420392036, + "learning_rate": 4.725093502866861e-06, + "loss": 0.7875, + "step": 684 + }, + { + "epoch": 0.7913357400722022, + "grad_norm": 0.02671169303357601, + "learning_rate": 4.7261509524873765e-06, + "loss": 0.7986, + "step": 685 + }, + { + "epoch": 0.7924909747292419, + "grad_norm": 0.025389740243554115, + "learning_rate": 4.727206859511253e-06, + "loss": 0.8168, + "step": 686 + }, + { + "epoch": 0.7936462093862816, + "grad_norm": 0.02669006958603859, + "learning_rate": 4.7282612284325845e-06, + "loss": 0.8061, + "step": 687 + }, + { + "epoch": 0.7948014440433213, + "grad_norm": 0.027196478098630905, + "learning_rate": 4.729314063725853e-06, + "loss": 0.8077, + "step": 688 + }, + { + "epoch": 0.795956678700361, + "grad_norm": 0.026497790589928627, + "learning_rate": 4.730365369846044e-06, + "loss": 0.8161, + "step": 689 + }, + { + "epoch": 0.7971119133574007, + "grad_norm": 0.025808461010456085, + "learning_rate": 4.73141515122876e-06, + "loss": 0.8042, + "step": 690 + }, + { + "epoch": 0.7982671480144404, + "grad_norm": 0.02556486800312996, + "learning_rate": 4.732463412290331e-06, + "loss": 0.7871, + "step": 691 + }, + { + "epoch": 0.7994223826714801, + "grad_norm": 0.02653394266963005, + "learning_rate": 4.73351015742793e-06, + "loss": 0.8163, + "step": 692 + }, + { + "epoch": 0.8005776173285198, + "grad_norm": 0.027373354882001877, + "learning_rate": 4.734555391019678e-06, + "loss": 0.8073, + "step": 693 + }, + { + "epoch": 0.8017328519855595, + "grad_norm": 0.02534923329949379, + "learning_rate": 4.735599117424759e-06, + "loss": 0.8186, + "step": 694 + }, + { + "epoch": 0.8028880866425993, + "grad_norm": 0.025402942672371864, + "learning_rate": 4.736641340983524e-06, + "loss": 0.8198, + "step": 695 + }, + { + "epoch": 0.804043321299639, + "grad_norm": 0.025877540931105614, + "learning_rate": 4.737682066017604e-06, + "loss": 0.7954, + "step": 696 + }, + { + "epoch": 0.8051985559566787, + "grad_norm": 0.027829859405755997, + "learning_rate": 4.738721296830016e-06, + "loss": 0.8197, + "step": 697 + }, + { + "epoch": 0.8063537906137184, + "grad_norm": 0.02695806324481964, + "learning_rate": 4.739759037705269e-06, + "loss": 0.8361, + "step": 698 + }, + { + "epoch": 0.8075090252707581, + "grad_norm": 0.02655518613755703, + "learning_rate": 4.740795292909469e-06, + "loss": 0.8055, + "step": 699 + }, + { + "epoch": 0.8086642599277978, + "grad_norm": 0.02709355391561985, + "learning_rate": 4.741830066690428e-06, + "loss": 0.7907, + "step": 700 + }, + { + "epoch": 0.8098194945848376, + "grad_norm": 0.026433579623699188, + "learning_rate": 4.742863363277765e-06, + "loss": 0.8235, + "step": 701 + }, + { + "epoch": 0.8109747292418773, + "grad_norm": 0.025645030662417412, + "learning_rate": 4.743895186883009e-06, + "loss": 0.8076, + "step": 702 + }, + { + "epoch": 0.812129963898917, + "grad_norm": 0.027833297848701477, + "learning_rate": 4.744925541699707e-06, + "loss": 0.8316, + "step": 703 + }, + { + "epoch": 0.8132851985559567, + "grad_norm": 0.025668229907751083, + "learning_rate": 4.74595443190352e-06, + "loss": 0.8075, + "step": 704 + }, + { + "epoch": 0.8144404332129964, + "grad_norm": 0.026288649067282677, + "learning_rate": 4.746981861652332e-06, + "loss": 0.8421, + "step": 705 + }, + { + "epoch": 0.8155956678700361, + "grad_norm": 0.024897335097193718, + "learning_rate": 4.7480078350863404e-06, + "loss": 0.8132, + "step": 706 + }, + { + "epoch": 0.8167509025270758, + "grad_norm": 0.0259522944688797, + "learning_rate": 4.749032356328167e-06, + "loss": 0.8326, + "step": 707 + }, + { + "epoch": 0.8179061371841155, + "grad_norm": 0.026102952659130096, + "learning_rate": 4.750055429482949e-06, + "loss": 0.8118, + "step": 708 + }, + { + "epoch": 0.8190613718411552, + "grad_norm": 0.025872284546494484, + "learning_rate": 4.751077058638445e-06, + "loss": 0.816, + "step": 709 + }, + { + "epoch": 0.8202166064981949, + "grad_norm": 0.025316176936030388, + "learning_rate": 4.752097247865126e-06, + "loss": 0.8061, + "step": 710 + }, + { + "epoch": 0.8213718411552346, + "grad_norm": 0.02691754698753357, + "learning_rate": 4.753116001216277e-06, + "loss": 0.7946, + "step": 711 + }, + { + "epoch": 0.8225270758122744, + "grad_norm": 0.027329301461577415, + "learning_rate": 4.754133322728095e-06, + "loss": 0.8379, + "step": 712 + }, + { + "epoch": 0.8236823104693141, + "grad_norm": 0.026258062571287155, + "learning_rate": 4.755149216419776e-06, + "loss": 0.8101, + "step": 713 + }, + { + "epoch": 0.8248375451263538, + "grad_norm": 0.02672073245048523, + "learning_rate": 4.756163686293624e-06, + "loss": 0.8225, + "step": 714 + }, + { + "epoch": 0.8259927797833935, + "grad_norm": 0.028196029365062714, + "learning_rate": 4.757176736335135e-06, + "loss": 0.848, + "step": 715 + }, + { + "epoch": 0.8271480144404332, + "grad_norm": 0.02594706416130066, + "learning_rate": 4.758188370513093e-06, + "loss": 0.8426, + "step": 716 + }, + { + "epoch": 0.8283032490974729, + "grad_norm": 0.028120605275034904, + "learning_rate": 4.759198592779668e-06, + "loss": 0.8306, + "step": 717 + }, + { + "epoch": 0.8294584837545126, + "grad_norm": 0.027138734236359596, + "learning_rate": 4.760207407070501e-06, + "loss": 0.825, + "step": 718 + }, + { + "epoch": 0.8306137184115523, + "grad_norm": 0.027197662740945816, + "learning_rate": 4.761214817304805e-06, + "loss": 0.8071, + "step": 719 + }, + { + "epoch": 0.8317689530685921, + "grad_norm": 0.026295404881238937, + "learning_rate": 4.7622208273854484e-06, + "loss": 0.8158, + "step": 720 + }, + { + "epoch": 0.8329241877256318, + "grad_norm": 0.02767680026590824, + "learning_rate": 4.763225441199049e-06, + "loss": 0.8226, + "step": 721 + }, + { + "epoch": 0.8340794223826715, + "grad_norm": 0.025539277121424675, + "learning_rate": 4.764228662616066e-06, + "loss": 0.8025, + "step": 722 + }, + { + "epoch": 0.8352346570397112, + "grad_norm": 0.02579374797642231, + "learning_rate": 4.765230495490885e-06, + "loss": 0.842, + "step": 723 + }, + { + "epoch": 0.8363898916967509, + "grad_norm": 0.025891058146953583, + "learning_rate": 4.766230943661912e-06, + "loss": 0.8135, + "step": 724 + }, + { + "epoch": 0.8375451263537906, + "grad_norm": 0.02527502365410328, + "learning_rate": 4.767230010951657e-06, + "loss": 0.8087, + "step": 725 + }, + { + "epoch": 0.8387003610108303, + "grad_norm": 0.02618500031530857, + "learning_rate": 4.768227701166823e-06, + "loss": 0.8242, + "step": 726 + }, + { + "epoch": 0.83985559566787, + "grad_norm": 0.026595618575811386, + "learning_rate": 4.769224018098397e-06, + "loss": 0.8418, + "step": 727 + }, + { + "epoch": 0.8410108303249098, + "grad_norm": 0.02652081847190857, + "learning_rate": 4.770218965521729e-06, + "loss": 0.8013, + "step": 728 + }, + { + "epoch": 0.8421660649819495, + "grad_norm": 0.026961620897054672, + "learning_rate": 4.771212547196624e-06, + "loss": 0.7984, + "step": 729 + }, + { + "epoch": 0.8433212996389892, + "grad_norm": 0.024961085990071297, + "learning_rate": 4.772204766867427e-06, + "loss": 0.809, + "step": 730 + }, + { + "epoch": 0.8444765342960289, + "grad_norm": 0.026465538889169693, + "learning_rate": 4.773195628263101e-06, + "loss": 0.8186, + "step": 731 + }, + { + "epoch": 0.8456317689530686, + "grad_norm": 0.025633295997977257, + "learning_rate": 4.77418513509732e-06, + "loss": 0.8111, + "step": 732 + }, + { + "epoch": 0.8467870036101083, + "grad_norm": 0.026612253859639168, + "learning_rate": 4.775173291068547e-06, + "loss": 0.8218, + "step": 733 + }, + { + "epoch": 0.847942238267148, + "grad_norm": 0.026292981579899788, + "learning_rate": 4.776160099860117e-06, + "loss": 0.8107, + "step": 734 + }, + { + "epoch": 0.8490974729241877, + "grad_norm": 0.026987800374627113, + "learning_rate": 4.777145565140325e-06, + "loss": 0.8088, + "step": 735 + }, + { + "epoch": 0.8502527075812274, + "grad_norm": 0.027124376967549324, + "learning_rate": 4.778129690562499e-06, + "loss": 0.7977, + "step": 736 + }, + { + "epoch": 0.8514079422382671, + "grad_norm": 0.025853095576167107, + "learning_rate": 4.7791124797650865e-06, + "loss": 0.8105, + "step": 737 + }, + { + "epoch": 0.8525631768953068, + "grad_norm": 0.027579553425312042, + "learning_rate": 4.780093936371737e-06, + "loss": 0.825, + "step": 738 + }, + { + "epoch": 0.8537184115523466, + "grad_norm": 0.027602966874837875, + "learning_rate": 4.781074063991377e-06, + "loss": 0.8088, + "step": 739 + }, + { + "epoch": 0.8548736462093863, + "grad_norm": 0.025124140083789825, + "learning_rate": 4.782052866218294e-06, + "loss": 0.7909, + "step": 740 + }, + { + "epoch": 0.856028880866426, + "grad_norm": 0.028426257893443108, + "learning_rate": 4.783030346632214e-06, + "loss": 0.8036, + "step": 741 + }, + { + "epoch": 0.8571841155234657, + "grad_norm": 0.026561887934803963, + "learning_rate": 4.784006508798379e-06, + "loss": 0.8118, + "step": 742 + }, + { + "epoch": 0.8583393501805054, + "grad_norm": 0.027730902656912804, + "learning_rate": 4.7849813562676265e-06, + "loss": 0.8319, + "step": 743 + }, + { + "epoch": 0.8594945848375452, + "grad_norm": 0.02669842354953289, + "learning_rate": 4.785954892576465e-06, + "loss": 0.8153, + "step": 744 + }, + { + "epoch": 0.8606498194945849, + "grad_norm": 0.02621045894920826, + "learning_rate": 4.786927121247155e-06, + "loss": 0.8164, + "step": 745 + }, + { + "epoch": 0.8618050541516246, + "grad_norm": 0.02661043219268322, + "learning_rate": 4.787898045787781e-06, + "loss": 0.7877, + "step": 746 + }, + { + "epoch": 0.8629602888086643, + "grad_norm": 0.02501833625137806, + "learning_rate": 4.788867669692332e-06, + "loss": 0.8096, + "step": 747 + }, + { + "epoch": 0.864115523465704, + "grad_norm": 0.025194406509399414, + "learning_rate": 4.78983599644077e-06, + "loss": 0.791, + "step": 748 + }, + { + "epoch": 0.8652707581227437, + "grad_norm": 0.026089461520314217, + "learning_rate": 4.790803029499111e-06, + "loss": 0.7906, + "step": 749 + }, + { + "epoch": 0.8664259927797834, + "grad_norm": 0.025562454015016556, + "learning_rate": 4.7917687723195e-06, + "loss": 0.7749, + "step": 750 + }, + { + "epoch": 0.8675812274368231, + "grad_norm": 0.025910962373018265, + "learning_rate": 4.792733228340281e-06, + "loss": 0.7991, + "step": 751 + }, + { + "epoch": 0.8687364620938628, + "grad_norm": 0.026650816202163696, + "learning_rate": 4.793696400986071e-06, + "loss": 0.7999, + "step": 752 + }, + { + "epoch": 0.8698916967509025, + "grad_norm": 0.025604519993066788, + "learning_rate": 4.794658293667835e-06, + "loss": 0.8337, + "step": 753 + }, + { + "epoch": 0.8710469314079422, + "grad_norm": 0.02709670551121235, + "learning_rate": 4.795618909782958e-06, + "loss": 0.8128, + "step": 754 + }, + { + "epoch": 0.8722021660649819, + "grad_norm": 0.027338657528162003, + "learning_rate": 4.7965782527153145e-06, + "loss": 0.8088, + "step": 755 + }, + { + "epoch": 0.8733574007220216, + "grad_norm": 0.028811268508434296, + "learning_rate": 4.797536325835345e-06, + "loss": 0.8195, + "step": 756 + }, + { + "epoch": 0.8745126353790613, + "grad_norm": 0.02585168555378914, + "learning_rate": 4.798493132500121e-06, + "loss": 0.8067, + "step": 757 + }, + { + "epoch": 0.8756678700361011, + "grad_norm": 0.024853166192770004, + "learning_rate": 4.799448676053423e-06, + "loss": 0.8032, + "step": 758 + }, + { + "epoch": 0.8768231046931408, + "grad_norm": 0.030403772369027138, + "learning_rate": 4.800402959825802e-06, + "loss": 0.8369, + "step": 759 + }, + { + "epoch": 0.8779783393501805, + "grad_norm": 0.02667844668030739, + "learning_rate": 4.801355987134653e-06, + "loss": 0.8112, + "step": 760 + }, + { + "epoch": 0.8791335740072203, + "grad_norm": 0.026796355843544006, + "learning_rate": 4.802307761284289e-06, + "loss": 0.819, + "step": 761 + }, + { + "epoch": 0.88028880866426, + "grad_norm": 0.02874310314655304, + "learning_rate": 4.803258285566001e-06, + "loss": 0.8149, + "step": 762 + }, + { + "epoch": 0.8814440433212997, + "grad_norm": 0.027145931497216225, + "learning_rate": 4.804207563258135e-06, + "loss": 0.8095, + "step": 763 + }, + { + "epoch": 0.8825992779783394, + "grad_norm": 0.02635267935693264, + "learning_rate": 4.80515559762615e-06, + "loss": 0.8188, + "step": 764 + }, + { + "epoch": 0.8837545126353791, + "grad_norm": 0.028352461755275726, + "learning_rate": 4.806102391922697e-06, + "loss": 0.8248, + "step": 765 + }, + { + "epoch": 0.8849097472924188, + "grad_norm": 0.026468653231859207, + "learning_rate": 4.8070479493876735e-06, + "loss": 0.8083, + "step": 766 + }, + { + "epoch": 0.8860649819494585, + "grad_norm": 0.02515551634132862, + "learning_rate": 4.8079922732483016e-06, + "loss": 0.8104, + "step": 767 + }, + { + "epoch": 0.8872202166064982, + "grad_norm": 0.02771628648042679, + "learning_rate": 4.808935366719187e-06, + "loss": 0.8083, + "step": 768 + }, + { + "epoch": 0.8883754512635379, + "grad_norm": 0.0274192001670599, + "learning_rate": 4.8098772330023855e-06, + "loss": 0.8181, + "step": 769 + }, + { + "epoch": 0.8895306859205776, + "grad_norm": 0.025738628581166267, + "learning_rate": 4.81081787528747e-06, + "loss": 0.8166, + "step": 770 + }, + { + "epoch": 0.8906859205776173, + "grad_norm": 0.026444412767887115, + "learning_rate": 4.8117572967515955e-06, + "loss": 0.8061, + "step": 771 + }, + { + "epoch": 0.891841155234657, + "grad_norm": 0.027295244857668877, + "learning_rate": 4.812695500559561e-06, + "loss": 0.796, + "step": 772 + }, + { + "epoch": 0.8929963898916967, + "grad_norm": 0.029825204983353615, + "learning_rate": 4.813632489863876e-06, + "loss": 0.8269, + "step": 773 + }, + { + "epoch": 0.8941516245487364, + "grad_norm": 0.027049973607063293, + "learning_rate": 4.814568267804822e-06, + "loss": 0.8166, + "step": 774 + }, + { + "epoch": 0.8953068592057761, + "grad_norm": 0.027095604687929153, + "learning_rate": 4.815502837510518e-06, + "loss": 0.8006, + "step": 775 + }, + { + "epoch": 0.8964620938628158, + "grad_norm": 0.02874009497463703, + "learning_rate": 4.816436202096981e-06, + "loss": 0.7987, + "step": 776 + }, + { + "epoch": 0.8976173285198557, + "grad_norm": 0.027292495593428612, + "learning_rate": 4.817368364668191e-06, + "loss": 0.8148, + "step": 777 + }, + { + "epoch": 0.8987725631768954, + "grad_norm": 0.025094101205468178, + "learning_rate": 4.8182993283161485e-06, + "loss": 0.833, + "step": 778 + }, + { + "epoch": 0.8999277978339351, + "grad_norm": 0.026829306036233902, + "learning_rate": 4.819229096120941e-06, + "loss": 0.8245, + "step": 779 + }, + { + "epoch": 0.9010830324909748, + "grad_norm": 0.0263043362647295, + "learning_rate": 4.820157671150801e-06, + "loss": 0.826, + "step": 780 + }, + { + "epoch": 0.9022382671480145, + "grad_norm": 0.029344236478209496, + "learning_rate": 4.821085056462168e-06, + "loss": 0.8322, + "step": 781 + }, + { + "epoch": 0.9033935018050542, + "grad_norm": 0.02591596357524395, + "learning_rate": 4.822011255099747e-06, + "loss": 0.8305, + "step": 782 + }, + { + "epoch": 0.9045487364620939, + "grad_norm": 0.02538052201271057, + "learning_rate": 4.8229362700965724e-06, + "loss": 0.8119, + "step": 783 + }, + { + "epoch": 0.9057039711191336, + "grad_norm": 0.028109556064009666, + "learning_rate": 4.823860104474065e-06, + "loss": 0.8334, + "step": 784 + }, + { + "epoch": 0.9068592057761733, + "grad_norm": 0.027156807482242584, + "learning_rate": 4.8247827612420875e-06, + "loss": 0.8148, + "step": 785 + }, + { + "epoch": 0.908014440433213, + "grad_norm": 0.02679731696844101, + "learning_rate": 4.825704243399014e-06, + "loss": 0.8228, + "step": 786 + }, + { + "epoch": 0.9091696750902527, + "grad_norm": 0.026814332231879234, + "learning_rate": 4.8266245539317745e-06, + "loss": 0.8027, + "step": 787 + }, + { + "epoch": 0.9103249097472924, + "grad_norm": 0.02507951483130455, + "learning_rate": 4.827543695815926e-06, + "loss": 0.796, + "step": 788 + }, + { + "epoch": 0.9114801444043321, + "grad_norm": 0.026251574978232384, + "learning_rate": 4.828461672015701e-06, + "loss": 0.8125, + "step": 789 + }, + { + "epoch": 0.9126353790613718, + "grad_norm": 0.02636023238301277, + "learning_rate": 4.82937848548407e-06, + "loss": 0.7954, + "step": 790 + }, + { + "epoch": 0.9137906137184115, + "grad_norm": 0.026472898200154305, + "learning_rate": 4.830294139162795e-06, + "loss": 0.8045, + "step": 791 + }, + { + "epoch": 0.9149458483754512, + "grad_norm": 0.025792265310883522, + "learning_rate": 4.83120863598249e-06, + "loss": 0.8072, + "step": 792 + }, + { + "epoch": 0.9161010830324909, + "grad_norm": 0.028488213196396828, + "learning_rate": 4.832121978862673e-06, + "loss": 0.7935, + "step": 793 + }, + { + "epoch": 0.9172563176895306, + "grad_norm": 0.02507089264690876, + "learning_rate": 4.833034170711828e-06, + "loss": 0.7944, + "step": 794 + }, + { + "epoch": 0.9184115523465703, + "grad_norm": 0.026674294844269753, + "learning_rate": 4.833945214427451e-06, + "loss": 0.8029, + "step": 795 + }, + { + "epoch": 0.91956678700361, + "grad_norm": 0.02713027596473694, + "learning_rate": 4.834855112896116e-06, + "loss": 0.8342, + "step": 796 + }, + { + "epoch": 0.9207220216606499, + "grad_norm": 0.026090459898114204, + "learning_rate": 4.835763868993521e-06, + "loss": 0.7986, + "step": 797 + }, + { + "epoch": 0.9218772563176896, + "grad_norm": 0.02631318010389805, + "learning_rate": 4.83667148558455e-06, + "loss": 0.8021, + "step": 798 + }, + { + "epoch": 0.9230324909747293, + "grad_norm": 0.027506891638040543, + "learning_rate": 4.8375779655233195e-06, + "loss": 0.8305, + "step": 799 + }, + { + "epoch": 0.924187725631769, + "grad_norm": 0.02776806242763996, + "learning_rate": 4.83848331165324e-06, + "loss": 0.8141, + "step": 800 + }, + { + "epoch": 0.9253429602888087, + "grad_norm": 0.027539484202861786, + "learning_rate": 4.839387526807064e-06, + "loss": 0.809, + "step": 801 + }, + { + "epoch": 0.9264981949458484, + "grad_norm": 0.02546604722738266, + "learning_rate": 4.84029061380694e-06, + "loss": 0.8075, + "step": 802 + }, + { + "epoch": 0.9276534296028881, + "grad_norm": 0.02793939970433712, + "learning_rate": 4.841192575464469e-06, + "loss": 0.8116, + "step": 803 + }, + { + "epoch": 0.9288086642599278, + "grad_norm": 0.026220016181468964, + "learning_rate": 4.842093414580753e-06, + "loss": 0.8085, + "step": 804 + }, + { + "epoch": 0.9299638989169675, + "grad_norm": 0.0267406664788723, + "learning_rate": 4.842993133946448e-06, + "loss": 0.8202, + "step": 805 + }, + { + "epoch": 0.9311191335740072, + "grad_norm": 0.028274521231651306, + "learning_rate": 4.8438917363418184e-06, + "loss": 0.8222, + "step": 806 + }, + { + "epoch": 0.9322743682310469, + "grad_norm": 0.025681311264634132, + "learning_rate": 4.844789224536785e-06, + "loss": 0.8009, + "step": 807 + }, + { + "epoch": 0.9334296028880866, + "grad_norm": 0.02876345068216324, + "learning_rate": 4.845685601290977e-06, + "loss": 0.8129, + "step": 808 + }, + { + "epoch": 0.9345848375451263, + "grad_norm": 0.027744382619857788, + "learning_rate": 4.846580869353787e-06, + "loss": 0.8084, + "step": 809 + }, + { + "epoch": 0.935740072202166, + "grad_norm": 0.026011621579527855, + "learning_rate": 4.847475031464417e-06, + "loss": 0.8058, + "step": 810 + }, + { + "epoch": 0.9368953068592057, + "grad_norm": 0.028499187901616096, + "learning_rate": 4.848368090351928e-06, + "loss": 0.8, + "step": 811 + }, + { + "epoch": 0.9380505415162455, + "grad_norm": 0.026256656274199486, + "learning_rate": 4.849260048735293e-06, + "loss": 0.8071, + "step": 812 + }, + { + "epoch": 0.9392057761732852, + "grad_norm": 0.0278424471616745, + "learning_rate": 4.8501509093234474e-06, + "loss": 0.8173, + "step": 813 + }, + { + "epoch": 0.9403610108303249, + "grad_norm": 0.028124138712882996, + "learning_rate": 4.851040674815336e-06, + "loss": 0.8057, + "step": 814 + }, + { + "epoch": 0.9415162454873646, + "grad_norm": 0.02599998749792576, + "learning_rate": 4.851929347899962e-06, + "loss": 0.7952, + "step": 815 + }, + { + "epoch": 0.9426714801444044, + "grad_norm": 0.03042020834982395, + "learning_rate": 4.852816931256436e-06, + "loss": 0.815, + "step": 816 + }, + { + "epoch": 0.9438267148014441, + "grad_norm": 0.02526993304491043, + "learning_rate": 4.853703427554027e-06, + "loss": 0.8169, + "step": 817 + }, + { + "epoch": 0.9449819494584838, + "grad_norm": 0.029504787176847458, + "learning_rate": 4.854588839452205e-06, + "loss": 0.8134, + "step": 818 + }, + { + "epoch": 0.9461371841155235, + "grad_norm": 0.02730252593755722, + "learning_rate": 4.855473169600699e-06, + "loss": 0.8198, + "step": 819 + }, + { + "epoch": 0.9472924187725632, + "grad_norm": 0.02669229544699192, + "learning_rate": 4.856356420639528e-06, + "loss": 0.8426, + "step": 820 + }, + { + "epoch": 0.9484476534296029, + "grad_norm": 0.029394106939435005, + "learning_rate": 4.857238595199068e-06, + "loss": 0.8025, + "step": 821 + }, + { + "epoch": 0.9496028880866426, + "grad_norm": 0.026510139927268028, + "learning_rate": 4.858119695900084e-06, + "loss": 0.8267, + "step": 822 + }, + { + "epoch": 0.9507581227436823, + "grad_norm": 0.025655683130025864, + "learning_rate": 4.858999725353784e-06, + "loss": 0.7936, + "step": 823 + }, + { + "epoch": 0.951913357400722, + "grad_norm": 0.029900815337896347, + "learning_rate": 4.859878686161861e-06, + "loss": 0.8245, + "step": 824 + }, + { + "epoch": 0.9530685920577617, + "grad_norm": 0.02642049826681614, + "learning_rate": 4.860756580916543e-06, + "loss": 0.8036, + "step": 825 + }, + { + "epoch": 0.9542238267148014, + "grad_norm": 0.027764586731791496, + "learning_rate": 4.861633412200637e-06, + "loss": 0.8132, + "step": 826 + }, + { + "epoch": 0.9553790613718411, + "grad_norm": 0.02786201983690262, + "learning_rate": 4.862509182587578e-06, + "loss": 0.8189, + "step": 827 + }, + { + "epoch": 0.9565342960288808, + "grad_norm": 0.02661985345184803, + "learning_rate": 4.8633838946414675e-06, + "loss": 0.8056, + "step": 828 + }, + { + "epoch": 0.9576895306859206, + "grad_norm": 0.029807835817337036, + "learning_rate": 4.864257550917123e-06, + "loss": 0.8108, + "step": 829 + }, + { + "epoch": 0.9588447653429603, + "grad_norm": 0.025031132623553276, + "learning_rate": 4.8651301539601235e-06, + "loss": 0.8031, + "step": 830 + }, + { + "epoch": 0.96, + "grad_norm": 0.028132742270827293, + "learning_rate": 4.866001706306853e-06, + "loss": 0.8199, + "step": 831 + }, + { + "epoch": 0.9611552346570397, + "grad_norm": 0.02886558324098587, + "learning_rate": 4.866872210484541e-06, + "loss": 0.8175, + "step": 832 + }, + { + "epoch": 0.9623104693140794, + "grad_norm": 0.026310063898563385, + "learning_rate": 4.867741669011313e-06, + "loss": 0.7998, + "step": 833 + }, + { + "epoch": 0.9634657039711191, + "grad_norm": 0.02729332074522972, + "learning_rate": 4.868610084396232e-06, + "loss": 0.7809, + "step": 834 + }, + { + "epoch": 0.9646209386281589, + "grad_norm": 0.027129124850034714, + "learning_rate": 4.869477459139337e-06, + "loss": 0.8219, + "step": 835 + }, + { + "epoch": 0.9657761732851986, + "grad_norm": 0.027403220534324646, + "learning_rate": 4.8703437957316945e-06, + "loss": 0.807, + "step": 836 + }, + { + "epoch": 0.9669314079422383, + "grad_norm": 0.028556736186146736, + "learning_rate": 4.871209096655434e-06, + "loss": 0.8213, + "step": 837 + }, + { + "epoch": 0.968086642599278, + "grad_norm": 0.027089878916740417, + "learning_rate": 4.872073364383795e-06, + "loss": 0.8167, + "step": 838 + }, + { + "epoch": 0.9692418772563177, + "grad_norm": 0.02591611072421074, + "learning_rate": 4.872936601381167e-06, + "loss": 0.7926, + "step": 839 + }, + { + "epoch": 0.9703971119133574, + "grad_norm": 0.028653450310230255, + "learning_rate": 4.873798810103137e-06, + "loss": 0.769, + "step": 840 + }, + { + "epoch": 0.9715523465703971, + "grad_norm": 0.025429286062717438, + "learning_rate": 4.874659992996521e-06, + "loss": 0.8033, + "step": 841 + }, + { + "epoch": 0.9727075812274368, + "grad_norm": 0.027507685124874115, + "learning_rate": 4.875520152499416e-06, + "loss": 0.7991, + "step": 842 + }, + { + "epoch": 0.9738628158844765, + "grad_norm": 0.02781766653060913, + "learning_rate": 4.876379291041238e-06, + "loss": 0.811, + "step": 843 + }, + { + "epoch": 0.9750180505415162, + "grad_norm": 0.02732367254793644, + "learning_rate": 4.87723741104276e-06, + "loss": 0.7795, + "step": 844 + }, + { + "epoch": 0.976173285198556, + "grad_norm": 0.027977894991636276, + "learning_rate": 4.878094514916154e-06, + "loss": 0.8024, + "step": 845 + }, + { + "epoch": 0.9773285198555957, + "grad_norm": 0.02706928178668022, + "learning_rate": 4.87895060506504e-06, + "loss": 0.8273, + "step": 846 + }, + { + "epoch": 0.9784837545126354, + "grad_norm": 0.027973534539341927, + "learning_rate": 4.879805683884512e-06, + "loss": 0.8249, + "step": 847 + }, + { + "epoch": 0.9796389891696751, + "grad_norm": 0.028234517201781273, + "learning_rate": 4.88065975376119e-06, + "loss": 0.8009, + "step": 848 + }, + { + "epoch": 0.9807942238267148, + "grad_norm": 0.026143617928028107, + "learning_rate": 4.881512817073255e-06, + "loss": 0.8114, + "step": 849 + }, + { + "epoch": 0.9819494584837545, + "grad_norm": 0.027196498587727547, + "learning_rate": 4.882364876190489e-06, + "loss": 0.7919, + "step": 850 + }, + { + "epoch": 0.9831046931407942, + "grad_norm": 0.027922434732317924, + "learning_rate": 4.883215933474314e-06, + "loss": 0.7986, + "step": 851 + }, + { + "epoch": 0.9842599277978339, + "grad_norm": 0.0268290676176548, + "learning_rate": 4.884065991277834e-06, + "loss": 0.8142, + "step": 852 + }, + { + "epoch": 0.9854151624548736, + "grad_norm": 0.026363542303442955, + "learning_rate": 4.8849150519458725e-06, + "loss": 0.7958, + "step": 853 + }, + { + "epoch": 0.9865703971119134, + "grad_norm": 0.026863981038331985, + "learning_rate": 4.88576311781501e-06, + "loss": 0.8132, + "step": 854 + }, + { + "epoch": 0.9877256317689531, + "grad_norm": 0.026298578828573227, + "learning_rate": 4.886610191213622e-06, + "loss": 0.8156, + "step": 855 + }, + { + "epoch": 0.9888808664259928, + "grad_norm": 0.026729200035333633, + "learning_rate": 4.8874562744619224e-06, + "loss": 0.8221, + "step": 856 + }, + { + "epoch": 0.9900361010830325, + "grad_norm": 0.025892335921525955, + "learning_rate": 4.888301369871998e-06, + "loss": 0.8267, + "step": 857 + }, + { + "epoch": 0.9911913357400722, + "grad_norm": 0.026062045246362686, + "learning_rate": 4.8891454797478435e-06, + "loss": 0.8206, + "step": 858 + }, + { + "epoch": 0.9923465703971119, + "grad_norm": 0.025573009625077248, + "learning_rate": 4.889988606385404e-06, + "loss": 0.8466, + "step": 859 + }, + { + "epoch": 0.9935018050541516, + "grad_norm": 0.026895977556705475, + "learning_rate": 4.890830752072613e-06, + "loss": 0.801, + "step": 860 + }, + { + "epoch": 0.9946570397111913, + "grad_norm": 0.026373952627182007, + "learning_rate": 4.891671919089425e-06, + "loss": 0.7935, + "step": 861 + }, + { + "epoch": 0.995812274368231, + "grad_norm": 0.025112492963671684, + "learning_rate": 4.892512109707855e-06, + "loss": 0.8104, + "step": 862 + }, + { + "epoch": 0.9969675090252708, + "grad_norm": 0.025861382484436035, + "learning_rate": 4.893351326192016e-06, + "loss": 0.7755, + "step": 863 + }, + { + "epoch": 0.9981227436823105, + "grad_norm": 0.025520671159029007, + "learning_rate": 4.8941895707981555e-06, + "loss": 0.783, + "step": 864 + }, + { + "epoch": 0.9992779783393502, + "grad_norm": 0.02590048499405384, + "learning_rate": 4.895026845774691e-06, + "loss": 0.7952, + "step": 865 + }, + { + "epoch": 1.00043321299639, + "grad_norm": 0.037370167672634125, + "learning_rate": 4.895863153362244e-06, + "loss": 1.0888, + "step": 866 + }, + { + "epoch": 1.0015884476534296, + "grad_norm": 0.028089042752981186, + "learning_rate": 4.8966984957936845e-06, + "loss": 0.8026, + "step": 867 + }, + { + "epoch": 1.0027436823104694, + "grad_norm": 0.02722124569118023, + "learning_rate": 4.897532875294154e-06, + "loss": 0.7892, + "step": 868 + }, + { + "epoch": 1.003898916967509, + "grad_norm": 0.02784401923418045, + "learning_rate": 4.898366294081111e-06, + "loss": 0.7766, + "step": 869 + }, + { + "epoch": 1.0050541516245488, + "grad_norm": 0.028576616197824478, + "learning_rate": 4.899198754364365e-06, + "loss": 0.794, + "step": 870 + }, + { + "epoch": 1.0062093862815884, + "grad_norm": 0.02609376236796379, + "learning_rate": 4.900030258346106e-06, + "loss": 0.787, + "step": 871 + }, + { + "epoch": 1.0073646209386282, + "grad_norm": 0.026441458612680435, + "learning_rate": 4.900860808220946e-06, + "loss": 0.7726, + "step": 872 + }, + { + "epoch": 1.0085198555956678, + "grad_norm": 0.026604607701301575, + "learning_rate": 4.90169040617595e-06, + "loss": 0.7672, + "step": 873 + }, + { + "epoch": 1.0096750902527076, + "grad_norm": 0.025642454624176025, + "learning_rate": 4.902519054390672e-06, + "loss": 0.7804, + "step": 874 + }, + { + "epoch": 1.0108303249097472, + "grad_norm": 0.027134709060192108, + "learning_rate": 4.9033467550371886e-06, + "loss": 0.7895, + "step": 875 + }, + { + "epoch": 1.011985559566787, + "grad_norm": 0.02571208029985428, + "learning_rate": 4.904173510280135e-06, + "loss": 0.7916, + "step": 876 + }, + { + "epoch": 1.0131407942238266, + "grad_norm": 0.026592353358864784, + "learning_rate": 4.904999322276735e-06, + "loss": 0.7975, + "step": 877 + }, + { + "epoch": 1.0142960288808665, + "grad_norm": 0.02615254372358322, + "learning_rate": 4.905824193176839e-06, + "loss": 0.7881, + "step": 878 + }, + { + "epoch": 1.015451263537906, + "grad_norm": 0.027889380231499672, + "learning_rate": 4.906648125122953e-06, + "loss": 0.7897, + "step": 879 + }, + { + "epoch": 1.0166064981949459, + "grad_norm": 0.02607213519513607, + "learning_rate": 4.907471120250281e-06, + "loss": 0.7735, + "step": 880 + }, + { + "epoch": 1.0177617328519855, + "grad_norm": 0.025891225785017014, + "learning_rate": 4.9082931806867475e-06, + "loss": 0.788, + "step": 881 + }, + { + "epoch": 1.0189169675090253, + "grad_norm": 0.02776011824607849, + "learning_rate": 4.909114308553034e-06, + "loss": 0.7902, + "step": 882 + }, + { + "epoch": 1.020072202166065, + "grad_norm": 0.026471663266420364, + "learning_rate": 4.909934505962614e-06, + "loss": 0.8, + "step": 883 + }, + { + "epoch": 1.0212274368231047, + "grad_norm": 0.026362843811511993, + "learning_rate": 4.9107537750217884e-06, + "loss": 0.7957, + "step": 884 + }, + { + "epoch": 1.0223826714801445, + "grad_norm": 0.027810046449303627, + "learning_rate": 4.9115721178297095e-06, + "loss": 0.7986, + "step": 885 + }, + { + "epoch": 1.023537906137184, + "grad_norm": 0.02654910273849964, + "learning_rate": 4.9123895364784185e-06, + "loss": 0.7611, + "step": 886 + }, + { + "epoch": 1.024693140794224, + "grad_norm": 0.026264095678925514, + "learning_rate": 4.913206033052878e-06, + "loss": 0.808, + "step": 887 + }, + { + "epoch": 1.0258483754512635, + "grad_norm": 0.02679434046149254, + "learning_rate": 4.914021609631002e-06, + "loss": 0.7432, + "step": 888 + }, + { + "epoch": 1.0270036101083033, + "grad_norm": 0.026940811425447464, + "learning_rate": 4.914836268283691e-06, + "loss": 0.7872, + "step": 889 + }, + { + "epoch": 1.028158844765343, + "grad_norm": 0.025167059153318405, + "learning_rate": 4.915650011074855e-06, + "loss": 0.7876, + "step": 890 + }, + { + "epoch": 1.0293140794223827, + "grad_norm": 0.026694072410464287, + "learning_rate": 4.916462840061458e-06, + "loss": 0.7919, + "step": 891 + }, + { + "epoch": 1.0304693140794223, + "grad_norm": 0.02730594575405121, + "learning_rate": 4.917274757293539e-06, + "loss": 0.7961, + "step": 892 + }, + { + "epoch": 1.0316245487364621, + "grad_norm": 0.026232946664094925, + "learning_rate": 4.918085764814244e-06, + "loss": 0.7908, + "step": 893 + }, + { + "epoch": 1.0327797833935017, + "grad_norm": 0.027884062379598618, + "learning_rate": 4.918895864659863e-06, + "loss": 0.7752, + "step": 894 + }, + { + "epoch": 1.0339350180505416, + "grad_norm": 0.027524368837475777, + "learning_rate": 4.919705058859854e-06, + "loss": 0.8114, + "step": 895 + }, + { + "epoch": 1.0350902527075811, + "grad_norm": 0.02609703689813614, + "learning_rate": 4.920513349436876e-06, + "loss": 0.801, + "step": 896 + }, + { + "epoch": 1.036245487364621, + "grad_norm": 0.027548260986804962, + "learning_rate": 4.921320738406821e-06, + "loss": 0.7736, + "step": 897 + }, + { + "epoch": 1.0374007220216606, + "grad_norm": 0.028695981949567795, + "learning_rate": 4.922127227778841e-06, + "loss": 0.7968, + "step": 898 + }, + { + "epoch": 1.0385559566787004, + "grad_norm": 0.02490762434899807, + "learning_rate": 4.922932819555381e-06, + "loss": 0.7695, + "step": 899 + }, + { + "epoch": 1.03971119133574, + "grad_norm": 0.030678994953632355, + "learning_rate": 4.923737515732209e-06, + "loss": 0.7764, + "step": 900 + }, + { + "epoch": 1.0408664259927798, + "grad_norm": 0.02829810604453087, + "learning_rate": 4.924541318298439e-06, + "loss": 0.7881, + "step": 901 + }, + { + "epoch": 1.0420216606498194, + "grad_norm": 0.028521962463855743, + "learning_rate": 4.925344229236571e-06, + "loss": 0.791, + "step": 902 + }, + { + "epoch": 1.0431768953068592, + "grad_norm": 0.028740065172314644, + "learning_rate": 4.926146250522511e-06, + "loss": 0.778, + "step": 903 + }, + { + "epoch": 1.044332129963899, + "grad_norm": 0.02634969726204872, + "learning_rate": 4.926947384125606e-06, + "loss": 0.7784, + "step": 904 + }, + { + "epoch": 1.0454873646209386, + "grad_norm": 0.02609422616660595, + "learning_rate": 4.927747632008672e-06, + "loss": 0.7729, + "step": 905 + }, + { + "epoch": 1.0466425992779784, + "grad_norm": 0.025464767590165138, + "learning_rate": 4.928546996128022e-06, + "loss": 0.7846, + "step": 906 + }, + { + "epoch": 1.047797833935018, + "grad_norm": 0.026383783668279648, + "learning_rate": 4.929345478433492e-06, + "loss": 0.7672, + "step": 907 + }, + { + "epoch": 1.0489530685920578, + "grad_norm": 0.02692020870745182, + "learning_rate": 4.930143080868476e-06, + "loss": 0.777, + "step": 908 + }, + { + "epoch": 1.0501083032490974, + "grad_norm": 0.026965700089931488, + "learning_rate": 4.930939805369947e-06, + "loss": 0.8104, + "step": 909 + }, + { + "epoch": 1.0512635379061372, + "grad_norm": 0.0277117807418108, + "learning_rate": 4.931735653868489e-06, + "loss": 0.7729, + "step": 910 + }, + { + "epoch": 1.0524187725631768, + "grad_norm": 0.027171596884727478, + "learning_rate": 4.932530628288331e-06, + "loss": 0.7957, + "step": 911 + }, + { + "epoch": 1.0535740072202167, + "grad_norm": 0.0277806855738163, + "learning_rate": 4.933324730547361e-06, + "loss": 0.7988, + "step": 912 + }, + { + "epoch": 1.0547292418772563, + "grad_norm": 0.02666284516453743, + "learning_rate": 4.934117962557165e-06, + "loss": 0.7637, + "step": 913 + }, + { + "epoch": 1.055884476534296, + "grad_norm": 0.028444141149520874, + "learning_rate": 4.934910326223052e-06, + "loss": 0.77, + "step": 914 + }, + { + "epoch": 1.0570397111913357, + "grad_norm": 0.028864728286862373, + "learning_rate": 4.93570182344408e-06, + "loss": 0.7778, + "step": 915 + }, + { + "epoch": 1.0581949458483755, + "grad_norm": 0.026793263852596283, + "learning_rate": 4.936492456113085e-06, + "loss": 0.7879, + "step": 916 + }, + { + "epoch": 1.059350180505415, + "grad_norm": 0.027911733835935593, + "learning_rate": 4.937282226116702e-06, + "loss": 0.7698, + "step": 917 + }, + { + "epoch": 1.060505415162455, + "grad_norm": 0.026382414624094963, + "learning_rate": 4.938071135335405e-06, + "loss": 0.7637, + "step": 918 + }, + { + "epoch": 1.0616606498194945, + "grad_norm": 0.027525540441274643, + "learning_rate": 4.938859185643519e-06, + "loss": 0.8057, + "step": 919 + }, + { + "epoch": 1.0628158844765343, + "grad_norm": 0.028536858037114143, + "learning_rate": 4.93964637890926e-06, + "loss": 0.7773, + "step": 920 + }, + { + "epoch": 1.0639711191335741, + "grad_norm": 0.026540234684944153, + "learning_rate": 4.9404327169947486e-06, + "loss": 0.7912, + "step": 921 + }, + { + "epoch": 1.0651263537906137, + "grad_norm": 0.026590172201395035, + "learning_rate": 4.941218201756049e-06, + "loss": 0.7873, + "step": 922 + }, + { + "epoch": 1.0662815884476535, + "grad_norm": 0.026739628985524178, + "learning_rate": 4.942002835043187e-06, + "loss": 0.7853, + "step": 923 + }, + { + "epoch": 1.0674368231046931, + "grad_norm": 0.027753658592700958, + "learning_rate": 4.9427866187001785e-06, + "loss": 0.7853, + "step": 924 + }, + { + "epoch": 1.068592057761733, + "grad_norm": 0.0267436932772398, + "learning_rate": 4.943569554565055e-06, + "loss": 0.7851, + "step": 925 + }, + { + "epoch": 1.0697472924187725, + "grad_norm": 0.02634557895362377, + "learning_rate": 4.9443516444698915e-06, + "loss": 0.7881, + "step": 926 + }, + { + "epoch": 1.0709025270758123, + "grad_norm": 0.027419744059443474, + "learning_rate": 4.945132890240829e-06, + "loss": 0.813, + "step": 927 + }, + { + "epoch": 1.072057761732852, + "grad_norm": 0.026219435036182404, + "learning_rate": 4.9459132936981044e-06, + "loss": 0.7998, + "step": 928 + }, + { + "epoch": 1.0732129963898918, + "grad_norm": 0.026673052459955215, + "learning_rate": 4.94669285665607e-06, + "loss": 0.764, + "step": 929 + }, + { + "epoch": 1.0743682310469314, + "grad_norm": 0.027227727696299553, + "learning_rate": 4.9474715809232256e-06, + "loss": 0.7734, + "step": 930 + }, + { + "epoch": 1.0755234657039712, + "grad_norm": 0.025797106325626373, + "learning_rate": 4.948249468302239e-06, + "loss": 0.7959, + "step": 931 + }, + { + "epoch": 1.0766787003610108, + "grad_norm": 0.026355307549238205, + "learning_rate": 4.9490265205899695e-06, + "loss": 0.7987, + "step": 932 + }, + { + "epoch": 1.0778339350180506, + "grad_norm": 0.02700088918209076, + "learning_rate": 4.949802739577501e-06, + "loss": 0.8067, + "step": 933 + }, + { + "epoch": 1.0789891696750902, + "grad_norm": 0.02693251334130764, + "learning_rate": 4.950578127050157e-06, + "loss": 0.7855, + "step": 934 + }, + { + "epoch": 1.08014440433213, + "grad_norm": 0.02637103572487831, + "learning_rate": 4.95135268478753e-06, + "loss": 0.7821, + "step": 935 + }, + { + "epoch": 1.0812996389891696, + "grad_norm": 0.02696199156343937, + "learning_rate": 4.952126414563509e-06, + "loss": 0.7954, + "step": 936 + }, + { + "epoch": 1.0824548736462094, + "grad_norm": 0.025795504450798035, + "learning_rate": 4.952899318146298e-06, + "loss": 0.771, + "step": 937 + }, + { + "epoch": 1.083610108303249, + "grad_norm": 0.0262451134622097, + "learning_rate": 4.953671397298441e-06, + "loss": 0.7737, + "step": 938 + }, + { + "epoch": 1.0847653429602888, + "grad_norm": 0.027646278962492943, + "learning_rate": 4.954442653776852e-06, + "loss": 0.7702, + "step": 939 + }, + { + "epoch": 1.0859205776173284, + "grad_norm": 0.026158776134252548, + "learning_rate": 4.955213089332832e-06, + "loss": 0.772, + "step": 940 + }, + { + "epoch": 1.0870758122743682, + "grad_norm": 0.02821963280439377, + "learning_rate": 4.955982705712095e-06, + "loss": 0.7828, + "step": 941 + }, + { + "epoch": 1.088231046931408, + "grad_norm": 0.026813939213752747, + "learning_rate": 4.956751504654796e-06, + "loss": 0.7848, + "step": 942 + }, + { + "epoch": 1.0893862815884476, + "grad_norm": 0.025459513068199158, + "learning_rate": 4.957519487895548e-06, + "loss": 0.7655, + "step": 943 + }, + { + "epoch": 1.0905415162454875, + "grad_norm": 0.026770230382680893, + "learning_rate": 4.958286657163448e-06, + "loss": 0.773, + "step": 944 + }, + { + "epoch": 1.091696750902527, + "grad_norm": 0.02704155258834362, + "learning_rate": 4.959053014182106e-06, + "loss": 0.7755, + "step": 945 + }, + { + "epoch": 1.0928519855595669, + "grad_norm": 0.02700476534664631, + "learning_rate": 4.959818560669656e-06, + "loss": 0.8009, + "step": 946 + }, + { + "epoch": 1.0940072202166065, + "grad_norm": 0.028186574578285217, + "learning_rate": 4.96058329833879e-06, + "loss": 0.7705, + "step": 947 + }, + { + "epoch": 1.0951624548736463, + "grad_norm": 0.027806002646684647, + "learning_rate": 4.9613472288967775e-06, + "loss": 0.7727, + "step": 948 + }, + { + "epoch": 1.0963176895306859, + "grad_norm": 0.028479190543293953, + "learning_rate": 4.962110354045489e-06, + "loss": 0.7921, + "step": 949 + }, + { + "epoch": 1.0974729241877257, + "grad_norm": 0.027802351862192154, + "learning_rate": 4.962872675481414e-06, + "loss": 0.7827, + "step": 950 + }, + { + "epoch": 1.0986281588447653, + "grad_norm": 0.026106078177690506, + "learning_rate": 4.96363419489569e-06, + "loss": 0.7663, + "step": 951 + }, + { + "epoch": 1.099783393501805, + "grad_norm": 0.027961455285549164, + "learning_rate": 4.964394913974124e-06, + "loss": 0.7735, + "step": 952 + }, + { + "epoch": 1.1009386281588447, + "grad_norm": 0.02648748643696308, + "learning_rate": 4.965154834397211e-06, + "loss": 0.8092, + "step": 953 + }, + { + "epoch": 1.1020938628158845, + "grad_norm": 0.026987139135599136, + "learning_rate": 4.965913957840159e-06, + "loss": 0.791, + "step": 954 + }, + { + "epoch": 1.103249097472924, + "grad_norm": 0.026114538311958313, + "learning_rate": 4.966672285972911e-06, + "loss": 0.7682, + "step": 955 + }, + { + "epoch": 1.104404332129964, + "grad_norm": 0.027255605906248093, + "learning_rate": 4.967429820460167e-06, + "loss": 0.7934, + "step": 956 + }, + { + "epoch": 1.1055595667870035, + "grad_norm": 0.026975391432642937, + "learning_rate": 4.968186562961406e-06, + "loss": 0.7758, + "step": 957 + }, + { + "epoch": 1.1067148014440433, + "grad_norm": 0.026317190378904343, + "learning_rate": 4.968942515130908e-06, + "loss": 0.7949, + "step": 958 + }, + { + "epoch": 1.1078700361010831, + "grad_norm": 0.027674200013279915, + "learning_rate": 4.969697678617774e-06, + "loss": 0.801, + "step": 959 + }, + { + "epoch": 1.1090252707581227, + "grad_norm": 0.026397835463285446, + "learning_rate": 4.970452055065948e-06, + "loss": 0.7872, + "step": 960 + }, + { + "epoch": 1.1101805054151626, + "grad_norm": 0.025970855727791786, + "learning_rate": 4.971205646114243e-06, + "loss": 0.8076, + "step": 961 + }, + { + "epoch": 1.1113357400722021, + "grad_norm": 0.026946989819407463, + "learning_rate": 4.9719584533963555e-06, + "loss": 0.7646, + "step": 962 + }, + { + "epoch": 1.112490974729242, + "grad_norm": 0.026165366172790527, + "learning_rate": 4.9727104785408915e-06, + "loss": 0.7704, + "step": 963 + }, + { + "epoch": 1.1136462093862816, + "grad_norm": 0.026523860171437263, + "learning_rate": 4.973461723171385e-06, + "loss": 0.7787, + "step": 964 + }, + { + "epoch": 1.1148014440433214, + "grad_norm": 0.02800065465271473, + "learning_rate": 4.974212188906321e-06, + "loss": 0.7824, + "step": 965 + }, + { + "epoch": 1.115956678700361, + "grad_norm": 0.02660745568573475, + "learning_rate": 4.974961877359156e-06, + "loss": 0.8069, + "step": 966 + }, + { + "epoch": 1.1171119133574008, + "grad_norm": 0.026590686291456223, + "learning_rate": 4.975710790138337e-06, + "loss": 0.7603, + "step": 967 + }, + { + "epoch": 1.1182671480144404, + "grad_norm": 0.028299605473876, + "learning_rate": 4.976458928847323e-06, + "loss": 0.7884, + "step": 968 + }, + { + "epoch": 1.1194223826714802, + "grad_norm": 0.02655804343521595, + "learning_rate": 4.977206295084609e-06, + "loss": 0.7902, + "step": 969 + }, + { + "epoch": 1.1205776173285198, + "grad_norm": 0.02828875742852688, + "learning_rate": 4.977952890443742e-06, + "loss": 0.7927, + "step": 970 + }, + { + "epoch": 1.1217328519855596, + "grad_norm": 0.03140028193593025, + "learning_rate": 4.978698716513342e-06, + "loss": 0.8075, + "step": 971 + }, + { + "epoch": 1.1228880866425992, + "grad_norm": 0.028589541092514992, + "learning_rate": 4.9794437748771245e-06, + "loss": 0.767, + "step": 972 + }, + { + "epoch": 1.124043321299639, + "grad_norm": 0.028249988332390785, + "learning_rate": 4.98018806711392e-06, + "loss": 0.8103, + "step": 973 + }, + { + "epoch": 1.1251985559566786, + "grad_norm": 0.027847470715641975, + "learning_rate": 4.980931594797693e-06, + "loss": 0.7939, + "step": 974 + }, + { + "epoch": 1.1263537906137184, + "grad_norm": 0.02874613367021084, + "learning_rate": 4.981674359497562e-06, + "loss": 0.7986, + "step": 975 + }, + { + "epoch": 1.1275090252707582, + "grad_norm": 0.02685857191681862, + "learning_rate": 4.98241636277782e-06, + "loss": 0.7765, + "step": 976 + }, + { + "epoch": 1.1286642599277978, + "grad_norm": 0.026808686554431915, + "learning_rate": 4.9831576061979556e-06, + "loss": 0.7786, + "step": 977 + }, + { + "epoch": 1.1298194945848374, + "grad_norm": 0.02733052521944046, + "learning_rate": 4.98389809131267e-06, + "loss": 0.7663, + "step": 978 + }, + { + "epoch": 1.1309747292418773, + "grad_norm": 0.0273138340562582, + "learning_rate": 4.984637819671897e-06, + "loss": 0.8018, + "step": 979 + }, + { + "epoch": 1.132129963898917, + "grad_norm": 0.02772151120007038, + "learning_rate": 4.985376792820825e-06, + "loss": 0.8052, + "step": 980 + }, + { + "epoch": 1.1332851985559567, + "grad_norm": 0.0273757204413414, + "learning_rate": 4.986115012299915e-06, + "loss": 0.7608, + "step": 981 + }, + { + "epoch": 1.1344404332129965, + "grad_norm": 0.026899464428424835, + "learning_rate": 4.986852479644916e-06, + "loss": 0.8069, + "step": 982 + }, + { + "epoch": 1.135595667870036, + "grad_norm": 0.026683615520596504, + "learning_rate": 4.987589196386893e-06, + "loss": 0.7694, + "step": 983 + }, + { + "epoch": 1.136750902527076, + "grad_norm": 0.028294900432229042, + "learning_rate": 4.988325164052236e-06, + "loss": 0.7738, + "step": 984 + }, + { + "epoch": 1.1379061371841155, + "grad_norm": 0.026210768148303032, + "learning_rate": 4.989060384162687e-06, + "loss": 0.7624, + "step": 985 + }, + { + "epoch": 1.1390613718411553, + "grad_norm": 0.026641108095645905, + "learning_rate": 4.989794858235352e-06, + "loss": 0.7827, + "step": 986 + }, + { + "epoch": 1.140216606498195, + "grad_norm": 0.027551405131816864, + "learning_rate": 4.990528587782728e-06, + "loss": 0.7849, + "step": 987 + }, + { + "epoch": 1.1413718411552347, + "grad_norm": 0.026633942499756813, + "learning_rate": 4.991261574312715e-06, + "loss": 0.8019, + "step": 988 + }, + { + "epoch": 1.1425270758122743, + "grad_norm": 0.027740631252527237, + "learning_rate": 4.991993819328633e-06, + "loss": 0.8075, + "step": 989 + }, + { + "epoch": 1.1436823104693141, + "grad_norm": 0.027059046551585197, + "learning_rate": 4.992725324329251e-06, + "loss": 0.7916, + "step": 990 + }, + { + "epoch": 1.1448375451263537, + "grad_norm": 0.027994418516755104, + "learning_rate": 4.993456090808793e-06, + "loss": 0.7943, + "step": 991 + }, + { + "epoch": 1.1459927797833935, + "grad_norm": 0.026348162442445755, + "learning_rate": 4.994186120256964e-06, + "loss": 0.7993, + "step": 992 + }, + { + "epoch": 1.1471480144404331, + "grad_norm": 0.027533039450645447, + "learning_rate": 4.99491541415897e-06, + "loss": 0.7806, + "step": 993 + }, + { + "epoch": 1.148303249097473, + "grad_norm": 0.02699647657573223, + "learning_rate": 4.995643973995523e-06, + "loss": 0.7753, + "step": 994 + }, + { + "epoch": 1.1494584837545125, + "grad_norm": 0.025739947333931923, + "learning_rate": 4.996371801242876e-06, + "loss": 0.7487, + "step": 995 + }, + { + "epoch": 1.1506137184115524, + "grad_norm": 0.026720624417066574, + "learning_rate": 4.997098897372832e-06, + "loss": 0.7776, + "step": 996 + }, + { + "epoch": 1.1517689530685922, + "grad_norm": 0.027732260525226593, + "learning_rate": 4.99782526385276e-06, + "loss": 0.7718, + "step": 997 + }, + { + "epoch": 1.1529241877256318, + "grad_norm": 0.026213763281702995, + "learning_rate": 4.998550902145619e-06, + "loss": 0.7849, + "step": 998 + }, + { + "epoch": 1.1540794223826714, + "grad_norm": 0.0285005122423172, + "learning_rate": 4.999275813709971e-06, + "loss": 0.7944, + "step": 999 + }, + { + "epoch": 1.1552346570397112, + "grad_norm": 0.025959184393286705, + "learning_rate": 5e-06, + "loss": 0.8039, + "step": 1000 + }, + { + "epoch": 1.156389891696751, + "grad_norm": 0.026393594220280647, + "learning_rate": 4.999976851655541e-06, + "loss": 0.792, + "step": 1001 + }, + { + "epoch": 1.1575451263537906, + "grad_norm": 0.02741307206451893, + "learning_rate": 4.999907407050885e-06, + "loss": 0.7816, + "step": 1002 + }, + { + "epoch": 1.1587003610108304, + "grad_norm": 0.02743341587483883, + "learning_rate": 4.999791667472181e-06, + "loss": 0.7473, + "step": 1003 + }, + { + "epoch": 1.15985559566787, + "grad_norm": 0.025454385206103325, + "learning_rate": 4.999629635062988e-06, + "loss": 0.7964, + "step": 1004 + }, + { + "epoch": 1.1610108303249098, + "grad_norm": 0.029390884563326836, + "learning_rate": 4.999421312824233e-06, + "loss": 0.7885, + "step": 1005 + }, + { + "epoch": 1.1621660649819494, + "grad_norm": 0.026846162974834442, + "learning_rate": 4.99916670461415e-06, + "loss": 0.7553, + "step": 1006 + }, + { + "epoch": 1.1633212996389892, + "grad_norm": 0.027146577835083008, + "learning_rate": 4.998865815148222e-06, + "loss": 0.757, + "step": 1007 + }, + { + "epoch": 1.1644765342960288, + "grad_norm": 0.028116164728999138, + "learning_rate": 4.998518649999077e-06, + "loss": 0.793, + "step": 1008 + }, + { + "epoch": 1.1656317689530686, + "grad_norm": 0.027374397963285446, + "learning_rate": 4.998125215596399e-06, + "loss": 0.7684, + "step": 1009 + }, + { + "epoch": 1.1667870036101082, + "grad_norm": 0.026328133419156075, + "learning_rate": 4.9976855192268e-06, + "loss": 0.794, + "step": 1010 + }, + { + "epoch": 1.167942238267148, + "grad_norm": 0.028758594766259193, + "learning_rate": 4.997199569033688e-06, + "loss": 0.7872, + "step": 1011 + }, + { + "epoch": 1.1690974729241876, + "grad_norm": 0.026665737852454185, + "learning_rate": 4.996667374017118e-06, + "loss": 0.797, + "step": 1012 + }, + { + "epoch": 1.1702527075812275, + "grad_norm": 0.02736995927989483, + "learning_rate": 4.996088944033622e-06, + "loss": 0.7912, + "step": 1013 + }, + { + "epoch": 1.171407942238267, + "grad_norm": 0.027897171676158905, + "learning_rate": 4.995464289796029e-06, + "loss": 0.7714, + "step": 1014 + }, + { + "epoch": 1.1725631768953069, + "grad_norm": 0.027566097676753998, + "learning_rate": 4.994793422873265e-06, + "loss": 0.7755, + "step": 1015 + }, + { + "epoch": 1.1737184115523465, + "grad_norm": 0.026376303285360336, + "learning_rate": 4.994076355690139e-06, + "loss": 0.8052, + "step": 1016 + }, + { + "epoch": 1.1748736462093863, + "grad_norm": 0.027220716699957848, + "learning_rate": 4.993313101527112e-06, + "loss": 0.7955, + "step": 1017 + }, + { + "epoch": 1.176028880866426, + "grad_norm": 0.02771547995507717, + "learning_rate": 4.992503674520057e-06, + "loss": 0.7595, + "step": 1018 + }, + { + "epoch": 1.1771841155234657, + "grad_norm": 0.026792939752340317, + "learning_rate": 4.991648089659987e-06, + "loss": 0.7812, + "step": 1019 + }, + { + "epoch": 1.1783393501805055, + "grad_norm": 0.026456760242581367, + "learning_rate": 4.990746362792786e-06, + "loss": 0.7884, + "step": 1020 + }, + { + "epoch": 1.179494584837545, + "grad_norm": 0.028625672683119774, + "learning_rate": 4.989798510618911e-06, + "loss": 0.7999, + "step": 1021 + }, + { + "epoch": 1.180649819494585, + "grad_norm": 0.026891304180026054, + "learning_rate": 4.988804550693086e-06, + "loss": 0.7949, + "step": 1022 + }, + { + "epoch": 1.1818050541516245, + "grad_norm": 0.025918442755937576, + "learning_rate": 4.9877645014239686e-06, + "loss": 0.793, + "step": 1023 + }, + { + "epoch": 1.1829602888086643, + "grad_norm": 0.026719292625784874, + "learning_rate": 4.9866783820738255e-06, + "loss": 0.7915, + "step": 1024 + }, + { + "epoch": 1.184115523465704, + "grad_norm": 0.027593158185482025, + "learning_rate": 4.985546212758157e-06, + "loss": 0.7678, + "step": 1025 + }, + { + "epoch": 1.1852707581227437, + "grad_norm": 0.02738940343260765, + "learning_rate": 4.984368014445336e-06, + "loss": 0.79, + "step": 1026 + }, + { + "epoch": 1.1864259927797833, + "grad_norm": 0.02584446780383587, + "learning_rate": 4.983143808956219e-06, + "loss": 0.7638, + "step": 1027 + }, + { + "epoch": 1.1875812274368232, + "grad_norm": 0.03016565926373005, + "learning_rate": 4.981873618963737e-06, + "loss": 0.7906, + "step": 1028 + }, + { + "epoch": 1.1887364620938627, + "grad_norm": 0.026769449934363365, + "learning_rate": 4.980557467992476e-06, + "loss": 0.7743, + "step": 1029 + }, + { + "epoch": 1.1898916967509026, + "grad_norm": 0.02779136225581169, + "learning_rate": 4.97919538041825e-06, + "loss": 0.7917, + "step": 1030 + }, + { + "epoch": 1.1910469314079422, + "grad_norm": 0.028257260099053383, + "learning_rate": 4.977787381467638e-06, + "loss": 0.774, + "step": 1031 + }, + { + "epoch": 1.192202166064982, + "grad_norm": 0.026401590555906296, + "learning_rate": 4.976333497217524e-06, + "loss": 0.7989, + "step": 1032 + }, + { + "epoch": 1.1933574007220216, + "grad_norm": 0.028177831321954727, + "learning_rate": 4.974833754594611e-06, + "loss": 0.7832, + "step": 1033 + }, + { + "epoch": 1.1945126353790614, + "grad_norm": 0.028338629752397537, + "learning_rate": 4.9732881813749236e-06, + "loss": 0.7991, + "step": 1034 + }, + { + "epoch": 1.1956678700361012, + "grad_norm": 0.027354028075933456, + "learning_rate": 4.9716968061832935e-06, + "loss": 0.7917, + "step": 1035 + }, + { + "epoch": 1.1968231046931408, + "grad_norm": 0.02721875160932541, + "learning_rate": 4.9700596584928275e-06, + "loss": 0.7765, + "step": 1036 + }, + { + "epoch": 1.1979783393501804, + "grad_norm": 0.026294628158211708, + "learning_rate": 4.968376768624368e-06, + "loss": 0.7871, + "step": 1037 + }, + { + "epoch": 1.1991335740072202, + "grad_norm": 0.027313537895679474, + "learning_rate": 4.96664816774592e-06, + "loss": 0.7858, + "step": 1038 + }, + { + "epoch": 1.20028880866426, + "grad_norm": 0.02785896323621273, + "learning_rate": 4.964873887872085e-06, + "loss": 0.7938, + "step": 1039 + }, + { + "epoch": 1.2014440433212996, + "grad_norm": 0.02687663771212101, + "learning_rate": 4.963053961863463e-06, + "loss": 0.7776, + "step": 1040 + }, + { + "epoch": 1.2025992779783394, + "grad_norm": 0.026713555678725243, + "learning_rate": 4.961188423426042e-06, + "loss": 0.7741, + "step": 1041 + }, + { + "epoch": 1.203754512635379, + "grad_norm": 0.02718759886920452, + "learning_rate": 4.959277307110582e-06, + "loss": 0.7667, + "step": 1042 + }, + { + "epoch": 1.2049097472924188, + "grad_norm": 0.027160272002220154, + "learning_rate": 4.957320648311961e-06, + "loss": 0.7894, + "step": 1043 + }, + { + "epoch": 1.2060649819494584, + "grad_norm": 0.02789607271552086, + "learning_rate": 4.955318483268535e-06, + "loss": 0.7812, + "step": 1044 + }, + { + "epoch": 1.2072202166064983, + "grad_norm": 0.02740464359521866, + "learning_rate": 4.953270849061456e-06, + "loss": 0.7585, + "step": 1045 + }, + { + "epoch": 1.2083754512635378, + "grad_norm": 0.028817661106586456, + "learning_rate": 4.951177783613991e-06, + "loss": 0.8091, + "step": 1046 + }, + { + "epoch": 1.2095306859205777, + "grad_norm": 0.026185913011431694, + "learning_rate": 4.9490393256908144e-06, + "loss": 0.7956, + "step": 1047 + }, + { + "epoch": 1.2106859205776173, + "grad_norm": 0.029244182631373405, + "learning_rate": 4.946855514897297e-06, + "loss": 0.7779, + "step": 1048 + }, + { + "epoch": 1.211841155234657, + "grad_norm": 0.027704982087016106, + "learning_rate": 4.944626391678766e-06, + "loss": 0.7993, + "step": 1049 + }, + { + "epoch": 1.2129963898916967, + "grad_norm": 0.02663758210837841, + "learning_rate": 4.942351997319761e-06, + "loss": 0.812, + "step": 1050 + }, + { + "epoch": 1.2141516245487365, + "grad_norm": 0.026222404092550278, + "learning_rate": 4.940032373943262e-06, + "loss": 0.7558, + "step": 1051 + }, + { + "epoch": 1.215306859205776, + "grad_norm": 0.02578769624233246, + "learning_rate": 4.937667564509923e-06, + "loss": 0.7781, + "step": 1052 + }, + { + "epoch": 1.216462093862816, + "grad_norm": 0.02732442319393158, + "learning_rate": 4.935257612817259e-06, + "loss": 0.7916, + "step": 1053 + }, + { + "epoch": 1.2176173285198555, + "grad_norm": 0.029004095122218132, + "learning_rate": 4.932802563498847e-06, + "loss": 0.7979, + "step": 1054 + }, + { + "epoch": 1.2187725631768953, + "grad_norm": 0.02736952155828476, + "learning_rate": 4.9303024620234966e-06, + "loss": 0.7984, + "step": 1055 + }, + { + "epoch": 1.2199277978339351, + "grad_norm": 0.026861999183893204, + "learning_rate": 4.927757354694406e-06, + "loss": 0.7986, + "step": 1056 + }, + { + "epoch": 1.2210830324909747, + "grad_norm": 0.0265817791223526, + "learning_rate": 4.925167288648305e-06, + "loss": 0.7763, + "step": 1057 + }, + { + "epoch": 1.2222382671480145, + "grad_norm": 0.028195342049002647, + "learning_rate": 4.922532311854585e-06, + "loss": 0.7967, + "step": 1058 + }, + { + "epoch": 1.2233935018050541, + "grad_norm": 0.028484554961323738, + "learning_rate": 4.919852473114405e-06, + "loss": 0.7926, + "step": 1059 + }, + { + "epoch": 1.224548736462094, + "grad_norm": 0.025893505662679672, + "learning_rate": 4.917127822059792e-06, + "loss": 0.783, + "step": 1060 + }, + { + "epoch": 1.2257039711191335, + "grad_norm": 0.028708558529615402, + "learning_rate": 4.9143584091527236e-06, + "loss": 0.7954, + "step": 1061 + }, + { + "epoch": 1.2268592057761734, + "grad_norm": 0.025731157511472702, + "learning_rate": 4.911544285684186e-06, + "loss": 0.7875, + "step": 1062 + }, + { + "epoch": 1.228014440433213, + "grad_norm": 0.02737441286444664, + "learning_rate": 4.9086855037732315e-06, + "loss": 0.7618, + "step": 1063 + }, + { + "epoch": 1.2291696750902528, + "grad_norm": 0.026509573683142662, + "learning_rate": 4.9057821163660085e-06, + "loss": 0.7513, + "step": 1064 + }, + { + "epoch": 1.2303249097472924, + "grad_norm": 0.027171408757567406, + "learning_rate": 4.9028341772347845e-06, + "loss": 0.7825, + "step": 1065 + }, + { + "epoch": 1.2314801444043322, + "grad_norm": 0.02830067090690136, + "learning_rate": 4.899841740976947e-06, + "loss": 0.7784, + "step": 1066 + }, + { + "epoch": 1.2326353790613718, + "grad_norm": 0.02830381691455841, + "learning_rate": 4.8968048630139945e-06, + "loss": 0.8057, + "step": 1067 + }, + { + "epoch": 1.2337906137184116, + "grad_norm": 0.027096690610051155, + "learning_rate": 4.89372359959051e-06, + "loss": 0.778, + "step": 1068 + }, + { + "epoch": 1.2349458483754512, + "grad_norm": 0.026616832241415977, + "learning_rate": 4.890598007773116e-06, + "loss": 0.7719, + "step": 1069 + }, + { + "epoch": 1.236101083032491, + "grad_norm": 0.02842404507100582, + "learning_rate": 4.887428145449425e-06, + "loss": 0.7898, + "step": 1070 + }, + { + "epoch": 1.2372563176895306, + "grad_norm": 0.029439568519592285, + "learning_rate": 4.884214071326957e-06, + "loss": 0.7586, + "step": 1071 + }, + { + "epoch": 1.2384115523465704, + "grad_norm": 0.027274589985609055, + "learning_rate": 4.880955844932061e-06, + "loss": 0.7779, + "step": 1072 + }, + { + "epoch": 1.2395667870036102, + "grad_norm": 0.030161483213305473, + "learning_rate": 4.877653526608811e-06, + "loss": 0.7774, + "step": 1073 + }, + { + "epoch": 1.2407220216606498, + "grad_norm": 0.026620803400874138, + "learning_rate": 4.874307177517882e-06, + "loss": 0.7714, + "step": 1074 + }, + { + "epoch": 1.2418772563176894, + "grad_norm": 0.031550221145153046, + "learning_rate": 4.870916859635426e-06, + "loss": 0.7951, + "step": 1075 + }, + { + "epoch": 1.2430324909747292, + "grad_norm": 0.02677975781261921, + "learning_rate": 4.86748263575192e-06, + "loss": 0.8094, + "step": 1076 + }, + { + "epoch": 1.244187725631769, + "grad_norm": 0.02815908007323742, + "learning_rate": 4.864004569471e-06, + "loss": 0.7828, + "step": 1077 + }, + { + "epoch": 1.2453429602888086, + "grad_norm": 0.028923533856868744, + "learning_rate": 4.860482725208291e-06, + "loss": 0.8106, + "step": 1078 + }, + { + "epoch": 1.2464981949458485, + "grad_norm": 0.026107341051101685, + "learning_rate": 4.856917168190205e-06, + "loss": 0.7893, + "step": 1079 + }, + { + "epoch": 1.247653429602888, + "grad_norm": 0.029115671291947365, + "learning_rate": 4.853307964452739e-06, + "loss": 0.749, + "step": 1080 + }, + { + "epoch": 1.2488086642599279, + "grad_norm": 0.02833879366517067, + "learning_rate": 4.849655180840253e-06, + "loss": 0.8041, + "step": 1081 + }, + { + "epoch": 1.2499638989169675, + "grad_norm": 0.027344336733222008, + "learning_rate": 4.8459588850042255e-06, + "loss": 0.7455, + "step": 1082 + }, + { + "epoch": 1.2511191335740073, + "grad_norm": 0.02860669605433941, + "learning_rate": 4.8422191454020055e-06, + "loss": 0.7995, + "step": 1083 + }, + { + "epoch": 1.2522743682310469, + "grad_norm": 0.026610156521201134, + "learning_rate": 4.838436031295543e-06, + "loss": 0.7776, + "step": 1084 + }, + { + "epoch": 1.2534296028880867, + "grad_norm": 0.029180224984884262, + "learning_rate": 4.834609612750108e-06, + "loss": 0.7755, + "step": 1085 + }, + { + "epoch": 1.2545848375451263, + "grad_norm": 0.025666316971182823, + "learning_rate": 4.8307399606329906e-06, + "loss": 0.7815, + "step": 1086 + }, + { + "epoch": 1.255740072202166, + "grad_norm": 0.028815045952796936, + "learning_rate": 4.82682714661219e-06, + "loss": 0.7618, + "step": 1087 + }, + { + "epoch": 1.2568953068592057, + "grad_norm": 0.02659596875309944, + "learning_rate": 4.822871243155085e-06, + "loss": 0.7906, + "step": 1088 + }, + { + "epoch": 1.2580505415162455, + "grad_norm": 0.027543647214770317, + "learning_rate": 4.818872323527097e-06, + "loss": 0.7871, + "step": 1089 + }, + { + "epoch": 1.2592057761732853, + "grad_norm": 0.027373237535357475, + "learning_rate": 4.814830461790327e-06, + "loss": 0.763, + "step": 1090 + }, + { + "epoch": 1.260361010830325, + "grad_norm": 0.025914175435900688, + "learning_rate": 4.810745732802186e-06, + "loss": 0.7744, + "step": 1091 + }, + { + "epoch": 1.2615162454873645, + "grad_norm": 0.026502788066864014, + "learning_rate": 4.806618212214009e-06, + "loss": 0.7901, + "step": 1092 + }, + { + "epoch": 1.2626714801444043, + "grad_norm": 0.026811202988028526, + "learning_rate": 4.802447976469658e-06, + "loss": 0.797, + "step": 1093 + }, + { + "epoch": 1.2638267148014442, + "grad_norm": 0.027904195711016655, + "learning_rate": 4.7982351028040966e-06, + "loss": 0.7956, + "step": 1094 + }, + { + "epoch": 1.2649819494584837, + "grad_norm": 0.025807814672589302, + "learning_rate": 4.79397966924197e-06, + "loss": 0.7785, + "step": 1095 + }, + { + "epoch": 1.2661371841155233, + "grad_norm": 0.02669427916407585, + "learning_rate": 4.789681754596152e-06, + "loss": 0.7882, + "step": 1096 + }, + { + "epoch": 1.2672924187725632, + "grad_norm": 0.026608362793922424, + "learning_rate": 4.785341438466289e-06, + "loss": 0.7716, + "step": 1097 + }, + { + "epoch": 1.268447653429603, + "grad_norm": 0.025953490287065506, + "learning_rate": 4.7809588012373255e-06, + "loss": 0.7702, + "step": 1098 + }, + { + "epoch": 1.2696028880866426, + "grad_norm": 0.02710854634642601, + "learning_rate": 4.776533924078018e-06, + "loss": 0.7734, + "step": 1099 + }, + { + "epoch": 1.2707581227436824, + "grad_norm": 0.027277441695332527, + "learning_rate": 4.772066888939423e-06, + "loss": 0.7894, + "step": 1100 + }, + { + "epoch": 1.271913357400722, + "grad_norm": 0.025714052841067314, + "learning_rate": 4.767557778553389e-06, + "loss": 0.7658, + "step": 1101 + }, + { + "epoch": 1.2730685920577618, + "grad_norm": 0.02643239125609398, + "learning_rate": 4.763006676431021e-06, + "loss": 0.775, + "step": 1102 + }, + { + "epoch": 1.2742238267148014, + "grad_norm": 0.028819050639867783, + "learning_rate": 4.75841366686113e-06, + "loss": 0.7823, + "step": 1103 + }, + { + "epoch": 1.2753790613718412, + "grad_norm": 0.026137549430131912, + "learning_rate": 4.753778834908678e-06, + "loss": 0.7607, + "step": 1104 + }, + { + "epoch": 1.2765342960288808, + "grad_norm": 0.027080422267317772, + "learning_rate": 4.749102266413195e-06, + "loss": 0.7739, + "step": 1105 + }, + { + "epoch": 1.2776895306859206, + "grad_norm": 0.026252347975969315, + "learning_rate": 4.744384047987202e-06, + "loss": 0.7914, + "step": 1106 + }, + { + "epoch": 1.2788447653429602, + "grad_norm": 0.02620067074894905, + "learning_rate": 4.739624267014589e-06, + "loss": 0.754, + "step": 1107 + }, + { + "epoch": 1.28, + "grad_norm": 0.026302233338356018, + "learning_rate": 4.734823011649013e-06, + "loss": 0.7964, + "step": 1108 + }, + { + "epoch": 1.2811552346570396, + "grad_norm": 0.02616049535572529, + "learning_rate": 4.7299803708122565e-06, + "loss": 0.7831, + "step": 1109 + }, + { + "epoch": 1.2823104693140794, + "grad_norm": 0.02667761594057083, + "learning_rate": 4.725096434192583e-06, + "loss": 0.7651, + "step": 1110 + }, + { + "epoch": 1.2834657039711193, + "grad_norm": 0.026840372011065483, + "learning_rate": 4.720171292243076e-06, + "loss": 0.7776, + "step": 1111 + }, + { + "epoch": 1.2846209386281588, + "grad_norm": 0.027080677449703217, + "learning_rate": 4.7152050361799626e-06, + "loss": 0.7804, + "step": 1112 + }, + { + "epoch": 1.2857761732851984, + "grad_norm": 0.026866400614380836, + "learning_rate": 4.710197757980923e-06, + "loss": 0.798, + "step": 1113 + }, + { + "epoch": 1.2869314079422383, + "grad_norm": 0.025592336431145668, + "learning_rate": 4.7051495503833955e-06, + "loss": 0.7717, + "step": 1114 + }, + { + "epoch": 1.288086642599278, + "grad_norm": 0.02509910985827446, + "learning_rate": 4.700060506882846e-06, + "loss": 0.7494, + "step": 1115 + }, + { + "epoch": 1.2892418772563177, + "grad_norm": 0.02745307981967926, + "learning_rate": 4.694930721731046e-06, + "loss": 0.7955, + "step": 1116 + }, + { + "epoch": 1.2903971119133573, + "grad_norm": 0.025139886885881424, + "learning_rate": 4.689760289934322e-06, + "loss": 0.7752, + "step": 1117 + }, + { + "epoch": 1.291552346570397, + "grad_norm": 0.025314299389719963, + "learning_rate": 4.684549307251799e-06, + "loss": 0.7631, + "step": 1118 + }, + { + "epoch": 1.292707581227437, + "grad_norm": 0.027501968666911125, + "learning_rate": 4.679297870193628e-06, + "loss": 0.7844, + "step": 1119 + }, + { + "epoch": 1.2938628158844765, + "grad_norm": 0.026297206059098244, + "learning_rate": 4.6740060760191905e-06, + "loss": 0.74, + "step": 1120 + }, + { + "epoch": 1.2950180505415163, + "grad_norm": 0.025890007615089417, + "learning_rate": 4.668674022735311e-06, + "loss": 0.7923, + "step": 1121 + }, + { + "epoch": 1.296173285198556, + "grad_norm": 0.026110520586371422, + "learning_rate": 4.663301809094425e-06, + "loss": 0.7899, + "step": 1122 + }, + { + "epoch": 1.2973285198555957, + "grad_norm": 0.026979975402355194, + "learning_rate": 4.657889534592767e-06, + "loss": 0.7831, + "step": 1123 + }, + { + "epoch": 1.2984837545126353, + "grad_norm": 0.026237450540065765, + "learning_rate": 4.652437299468514e-06, + "loss": 0.7728, + "step": 1124 + }, + { + "epoch": 1.2996389891696751, + "grad_norm": 0.02676897682249546, + "learning_rate": 4.646945204699939e-06, + "loss": 0.7736, + "step": 1125 + }, + { + "epoch": 1.3007942238267147, + "grad_norm": 0.02639893814921379, + "learning_rate": 4.641413352003534e-06, + "loss": 0.7738, + "step": 1126 + }, + { + "epoch": 1.3019494584837545, + "grad_norm": 0.026181381195783615, + "learning_rate": 4.63584184383213e-06, + "loss": 0.7766, + "step": 1127 + }, + { + "epoch": 1.3031046931407944, + "grad_norm": 0.029100872576236725, + "learning_rate": 4.630230783372998e-06, + "loss": 0.7939, + "step": 1128 + }, + { + "epoch": 1.304259927797834, + "grad_norm": 0.027207808569073677, + "learning_rate": 4.624580274545938e-06, + "loss": 0.7701, + "step": 1129 + }, + { + "epoch": 1.3054151624548735, + "grad_norm": 0.027454031631350517, + "learning_rate": 4.618890422001355e-06, + "loss": 0.7979, + "step": 1130 + }, + { + "epoch": 1.3065703971119134, + "grad_norm": 0.02654033899307251, + "learning_rate": 4.61316133111832e-06, + "loss": 0.7473, + "step": 1131 + }, + { + "epoch": 1.3077256317689532, + "grad_norm": 0.02540537528693676, + "learning_rate": 4.607393108002618e-06, + "loss": 0.7663, + "step": 1132 + }, + { + "epoch": 1.3088808664259928, + "grad_norm": 0.026679422706365585, + "learning_rate": 4.601585859484786e-06, + "loss": 0.7811, + "step": 1133 + }, + { + "epoch": 1.3100361010830324, + "grad_norm": 0.02628672868013382, + "learning_rate": 4.595739693118132e-06, + "loss": 0.7836, + "step": 1134 + }, + { + "epoch": 1.3111913357400722, + "grad_norm": 0.026694132015109062, + "learning_rate": 4.5898547171767375e-06, + "loss": 0.7708, + "step": 1135 + }, + { + "epoch": 1.312346570397112, + "grad_norm": 0.02691902220249176, + "learning_rate": 4.583931040653466e-06, + "loss": 0.772, + "step": 1136 + }, + { + "epoch": 1.3135018050541516, + "grad_norm": 0.02754952199757099, + "learning_rate": 4.5779687732579285e-06, + "loss": 0.783, + "step": 1137 + }, + { + "epoch": 1.3146570397111914, + "grad_norm": 0.025518443435430527, + "learning_rate": 4.571968025414466e-06, + "loss": 0.7736, + "step": 1138 + }, + { + "epoch": 1.315812274368231, + "grad_norm": 0.025681914761662483, + "learning_rate": 4.565928908260093e-06, + "loss": 0.7823, + "step": 1139 + }, + { + "epoch": 1.3169675090252708, + "grad_norm": 0.028058268129825592, + "learning_rate": 4.559851533642446e-06, + "loss": 0.7966, + "step": 1140 + }, + { + "epoch": 1.3181227436823104, + "grad_norm": 0.027302585542201996, + "learning_rate": 4.55373601411771e-06, + "loss": 0.8043, + "step": 1141 + }, + { + "epoch": 1.3192779783393502, + "grad_norm": 0.027836551889777184, + "learning_rate": 4.547582462948533e-06, + "loss": 0.7903, + "step": 1142 + }, + { + "epoch": 1.3204332129963898, + "grad_norm": 0.027944888919591904, + "learning_rate": 4.541390994101928e-06, + "loss": 0.7834, + "step": 1143 + }, + { + "epoch": 1.3215884476534296, + "grad_norm": 0.026407793164253235, + "learning_rate": 4.535161722247168e-06, + "loss": 0.7774, + "step": 1144 + }, + { + "epoch": 1.3227436823104692, + "grad_norm": 0.026148339733481407, + "learning_rate": 4.528894762753651e-06, + "loss": 0.7909, + "step": 1145 + }, + { + "epoch": 1.323898916967509, + "grad_norm": 0.027284881100058556, + "learning_rate": 4.522590231688775e-06, + "loss": 0.7885, + "step": 1146 + }, + { + "epoch": 1.3250541516245486, + "grad_norm": 0.026616675779223442, + "learning_rate": 4.516248245815782e-06, + "loss": 0.7924, + "step": 1147 + }, + { + "epoch": 1.3262093862815885, + "grad_norm": 0.026135679334402084, + "learning_rate": 4.509868922591595e-06, + "loss": 0.7845, + "step": 1148 + }, + { + "epoch": 1.3273646209386283, + "grad_norm": 0.027222778648138046, + "learning_rate": 4.503452380164648e-06, + "loss": 0.7983, + "step": 1149 + }, + { + "epoch": 1.3285198555956679, + "grad_norm": 0.02566412277519703, + "learning_rate": 4.496998737372691e-06, + "loss": 0.7561, + "step": 1150 + }, + { + "epoch": 1.3296750902527075, + "grad_norm": 0.026768745854496956, + "learning_rate": 4.490508113740594e-06, + "loss": 0.7753, + "step": 1151 + }, + { + "epoch": 1.3308303249097473, + "grad_norm": 0.027144471183419228, + "learning_rate": 4.483980629478131e-06, + "loss": 0.769, + "step": 1152 + }, + { + "epoch": 1.331985559566787, + "grad_norm": 0.0260478463023901, + "learning_rate": 4.477416405477754e-06, + "loss": 0.7879, + "step": 1153 + }, + { + "epoch": 1.3331407942238267, + "grad_norm": 0.025150155648589134, + "learning_rate": 4.470815563312356e-06, + "loss": 0.7795, + "step": 1154 + }, + { + "epoch": 1.3342960288808663, + "grad_norm": 0.026815764605998993, + "learning_rate": 4.464178225233016e-06, + "loss": 0.7577, + "step": 1155 + }, + { + "epoch": 1.335451263537906, + "grad_norm": 0.02605642005801201, + "learning_rate": 4.457504514166738e-06, + "loss": 0.7617, + "step": 1156 + }, + { + "epoch": 1.336606498194946, + "grad_norm": 0.02583806961774826, + "learning_rate": 4.4507945537141714e-06, + "loss": 0.7924, + "step": 1157 + }, + { + "epoch": 1.3377617328519855, + "grad_norm": 0.027298036962747574, + "learning_rate": 4.444048468147324e-06, + "loss": 0.7756, + "step": 1158 + }, + { + "epoch": 1.3389169675090253, + "grad_norm": 0.02701791748404503, + "learning_rate": 4.4372663824072595e-06, + "loss": 0.785, + "step": 1159 + }, + { + "epoch": 1.340072202166065, + "grad_norm": 0.02676539123058319, + "learning_rate": 4.430448422101786e-06, + "loss": 0.7914, + "step": 1160 + }, + { + "epoch": 1.3412274368231047, + "grad_norm": 0.026288233697414398, + "learning_rate": 4.423594713503124e-06, + "loss": 0.7948, + "step": 1161 + }, + { + "epoch": 1.3423826714801443, + "grad_norm": 0.02532966062426567, + "learning_rate": 4.416705383545574e-06, + "loss": 0.7571, + "step": 1162 + }, + { + "epoch": 1.3435379061371842, + "grad_norm": 0.025268996134400368, + "learning_rate": 4.409780559823159e-06, + "loss": 0.7544, + "step": 1163 + }, + { + "epoch": 1.3446931407942238, + "grad_norm": 0.025901716202497482, + "learning_rate": 4.402820370587272e-06, + "loss": 0.7928, + "step": 1164 + }, + { + "epoch": 1.3458483754512636, + "grad_norm": 0.026023518294095993, + "learning_rate": 4.395824944744287e-06, + "loss": 0.7821, + "step": 1165 + }, + { + "epoch": 1.3470036101083034, + "grad_norm": 0.025839975103735924, + "learning_rate": 4.388794411853183e-06, + "loss": 0.7563, + "step": 1166 + }, + { + "epoch": 1.348158844765343, + "grad_norm": 0.02656048908829689, + "learning_rate": 4.381728902123138e-06, + "loss": 0.7675, + "step": 1167 + }, + { + "epoch": 1.3493140794223826, + "grad_norm": 0.025841468945145607, + "learning_rate": 4.374628546411121e-06, + "loss": 0.7839, + "step": 1168 + }, + { + "epoch": 1.3504693140794224, + "grad_norm": 0.02622513473033905, + "learning_rate": 4.367493476219464e-06, + "loss": 0.7764, + "step": 1169 + }, + { + "epoch": 1.3516245487364622, + "grad_norm": 0.0271604061126709, + "learning_rate": 4.360323823693434e-06, + "loss": 0.7873, + "step": 1170 + }, + { + "epoch": 1.3527797833935018, + "grad_norm": 0.025581173598766327, + "learning_rate": 4.353119721618778e-06, + "loss": 0.7719, + "step": 1171 + }, + { + "epoch": 1.3539350180505414, + "grad_norm": 0.027338258922100067, + "learning_rate": 4.345881303419266e-06, + "loss": 0.8023, + "step": 1172 + }, + { + "epoch": 1.3550902527075812, + "grad_norm": 0.024792378768324852, + "learning_rate": 4.338608703154224e-06, + "loss": 0.7612, + "step": 1173 + }, + { + "epoch": 1.356245487364621, + "grad_norm": 0.02713456004858017, + "learning_rate": 4.3313020555160445e-06, + "loss": 0.7735, + "step": 1174 + }, + { + "epoch": 1.3574007220216606, + "grad_norm": 0.026577133685350418, + "learning_rate": 4.323961495827698e-06, + "loss": 0.7942, + "step": 1175 + }, + { + "epoch": 1.3585559566787004, + "grad_norm": 0.02608906850218773, + "learning_rate": 4.316587160040222e-06, + "loss": 0.7675, + "step": 1176 + }, + { + "epoch": 1.35971119133574, + "grad_norm": 0.026449469849467278, + "learning_rate": 4.309179184730207e-06, + "loss": 0.7586, + "step": 1177 + }, + { + "epoch": 1.3608664259927798, + "grad_norm": 0.02639465406537056, + "learning_rate": 4.301737707097264e-06, + "loss": 0.7792, + "step": 1178 + }, + { + "epoch": 1.3620216606498194, + "grad_norm": 0.026730941608548164, + "learning_rate": 4.294262864961486e-06, + "loss": 0.77, + "step": 1179 + }, + { + "epoch": 1.3631768953068593, + "grad_norm": 0.02690793387591839, + "learning_rate": 4.286754796760892e-06, + "loss": 0.7898, + "step": 1180 + }, + { + "epoch": 1.3643321299638989, + "grad_norm": 0.02708018571138382, + "learning_rate": 4.279213641548867e-06, + "loss": 0.7644, + "step": 1181 + }, + { + "epoch": 1.3654873646209387, + "grad_norm": 0.02672554738819599, + "learning_rate": 4.271639538991584e-06, + "loss": 0.7484, + "step": 1182 + }, + { + "epoch": 1.3666425992779783, + "grad_norm": 0.028112929314374924, + "learning_rate": 4.264032629365418e-06, + "loss": 0.8274, + "step": 1183 + }, + { + "epoch": 1.367797833935018, + "grad_norm": 0.026468411087989807, + "learning_rate": 4.256393053554351e-06, + "loss": 0.7438, + "step": 1184 + }, + { + "epoch": 1.3689530685920577, + "grad_norm": 0.02600760944187641, + "learning_rate": 4.248720953047355e-06, + "loss": 0.764, + "step": 1185 + }, + { + "epoch": 1.3701083032490975, + "grad_norm": 0.026687582954764366, + "learning_rate": 4.241016469935782e-06, + "loss": 0.764, + "step": 1186 + }, + { + "epoch": 1.3712635379061373, + "grad_norm": 0.028025714680552483, + "learning_rate": 4.233279746910722e-06, + "loss": 0.7729, + "step": 1187 + }, + { + "epoch": 1.372418772563177, + "grad_norm": 0.028168534860014915, + "learning_rate": 4.225510927260369e-06, + "loss": 0.7901, + "step": 1188 + }, + { + "epoch": 1.3735740072202165, + "grad_norm": 0.027783645316958427, + "learning_rate": 4.217710154867362e-06, + "loss": 0.8064, + "step": 1189 + }, + { + "epoch": 1.3747292418772563, + "grad_norm": 0.025505684316158295, + "learning_rate": 4.209877574206121e-06, + "loss": 0.7248, + "step": 1190 + }, + { + "epoch": 1.3758844765342961, + "grad_norm": 0.02887566015124321, + "learning_rate": 4.2020133303401714e-06, + "loss": 0.8044, + "step": 1191 + }, + { + "epoch": 1.3770397111913357, + "grad_norm": 0.02724134922027588, + "learning_rate": 4.1941175689194615e-06, + "loss": 0.771, + "step": 1192 + }, + { + "epoch": 1.3781949458483753, + "grad_norm": 0.026623480021953583, + "learning_rate": 4.186190436177655e-06, + "loss": 0.7926, + "step": 1193 + }, + { + "epoch": 1.3793501805054151, + "grad_norm": 0.026381775736808777, + "learning_rate": 4.178232078929437e-06, + "loss": 0.7658, + "step": 1194 + }, + { + "epoch": 1.380505415162455, + "grad_norm": 0.027352899312973022, + "learning_rate": 4.170242644567781e-06, + "loss": 0.7598, + "step": 1195 + }, + { + "epoch": 1.3816606498194945, + "grad_norm": 0.027212858200073242, + "learning_rate": 4.162222281061226e-06, + "loss": 0.8047, + "step": 1196 + }, + { + "epoch": 1.3828158844765344, + "grad_norm": 0.02757844887673855, + "learning_rate": 4.154171136951136e-06, + "loss": 0.7707, + "step": 1197 + }, + { + "epoch": 1.383971119133574, + "grad_norm": 0.02665034867823124, + "learning_rate": 4.146089361348948e-06, + "loss": 0.7659, + "step": 1198 + }, + { + "epoch": 1.3851263537906138, + "grad_norm": 0.02670939266681671, + "learning_rate": 4.137977103933411e-06, + "loss": 0.7898, + "step": 1199 + }, + { + "epoch": 1.3862815884476534, + "grad_norm": 0.02680130861699581, + "learning_rate": 4.1298345149478115e-06, + "loss": 0.8019, + "step": 1200 + }, + { + "epoch": 1.3874368231046932, + "grad_norm": 0.02670557238161564, + "learning_rate": 4.121661745197195e-06, + "loss": 0.7749, + "step": 1201 + }, + { + "epoch": 1.3885920577617328, + "grad_norm": 0.02557177096605301, + "learning_rate": 4.113458946045569e-06, + "loss": 0.7807, + "step": 1202 + }, + { + "epoch": 1.3897472924187726, + "grad_norm": 0.027189351618289948, + "learning_rate": 4.1052262694131e-06, + "loss": 0.8127, + "step": 1203 + }, + { + "epoch": 1.3909025270758124, + "grad_norm": 0.026801222935318947, + "learning_rate": 4.096963867773306e-06, + "loss": 0.7887, + "step": 1204 + }, + { + "epoch": 1.392057761732852, + "grad_norm": 0.027125921100378036, + "learning_rate": 4.0886718941502215e-06, + "loss": 0.7705, + "step": 1205 + }, + { + "epoch": 1.3932129963898916, + "grad_norm": 0.027330022305250168, + "learning_rate": 4.080350502115575e-06, + "loss": 0.7655, + "step": 1206 + }, + { + "epoch": 1.3943682310469314, + "grad_norm": 0.025932665914297104, + "learning_rate": 4.071999845785938e-06, + "loss": 0.7839, + "step": 1207 + }, + { + "epoch": 1.3955234657039712, + "grad_norm": 0.02584053948521614, + "learning_rate": 4.06362007981987e-06, + "loss": 0.7676, + "step": 1208 + }, + { + "epoch": 1.3966787003610108, + "grad_norm": 0.025614172220230103, + "learning_rate": 4.055211359415058e-06, + "loss": 0.7685, + "step": 1209 + }, + { + "epoch": 1.3978339350180504, + "grad_norm": 0.026585347950458527, + "learning_rate": 4.0467738403054425e-06, + "loss": 0.7716, + "step": 1210 + }, + { + "epoch": 1.3989891696750902, + "grad_norm": 0.02616872265934944, + "learning_rate": 4.038307678758328e-06, + "loss": 0.8029, + "step": 1211 + }, + { + "epoch": 1.40014440433213, + "grad_norm": 0.02624761499464512, + "learning_rate": 4.0298130315714934e-06, + "loss": 0.766, + "step": 1212 + }, + { + "epoch": 1.4012996389891696, + "grad_norm": 0.025757934898138046, + "learning_rate": 4.021290056070287e-06, + "loss": 0.7724, + "step": 1213 + }, + { + "epoch": 1.4024548736462095, + "grad_norm": 0.027395786717534065, + "learning_rate": 4.012738910104711e-06, + "loss": 0.7987, + "step": 1214 + }, + { + "epoch": 1.403610108303249, + "grad_norm": 0.0256296806037426, + "learning_rate": 4.004159752046501e-06, + "loss": 0.7516, + "step": 1215 + }, + { + "epoch": 1.4047653429602889, + "grad_norm": 0.027848385274410248, + "learning_rate": 3.99555274078619e-06, + "loss": 0.7778, + "step": 1216 + }, + { + "epoch": 1.4059205776173285, + "grad_norm": 0.02629752829670906, + "learning_rate": 3.986918035730169e-06, + "loss": 0.7689, + "step": 1217 + }, + { + "epoch": 1.4070758122743683, + "grad_norm": 0.027596496045589447, + "learning_rate": 3.9782557967977295e-06, + "loss": 0.7976, + "step": 1218 + }, + { + "epoch": 1.4082310469314079, + "grad_norm": 0.02730746567249298, + "learning_rate": 3.969566184418109e-06, + "loss": 0.7775, + "step": 1219 + }, + { + "epoch": 1.4093862815884477, + "grad_norm": 0.025179557502269745, + "learning_rate": 3.960849359527515e-06, + "loss": 0.7639, + "step": 1220 + }, + { + "epoch": 1.4105415162454873, + "grad_norm": 0.026616785675287247, + "learning_rate": 3.952105483566139e-06, + "loss": 0.7583, + "step": 1221 + }, + { + "epoch": 1.411696750902527, + "grad_norm": 0.027852557599544525, + "learning_rate": 3.94333471847518e-06, + "loss": 0.7936, + "step": 1222 + }, + { + "epoch": 1.4128519855595667, + "grad_norm": 0.026742927730083466, + "learning_rate": 3.934537226693836e-06, + "loss": 0.761, + "step": 1223 + }, + { + "epoch": 1.4140072202166065, + "grad_norm": 0.025545494630932808, + "learning_rate": 3.925713171156295e-06, + "loss": 0.7952, + "step": 1224 + }, + { + "epoch": 1.4151624548736463, + "grad_norm": 0.02676277793943882, + "learning_rate": 3.9168627152887214e-06, + "loss": 0.7623, + "step": 1225 + }, + { + "epoch": 1.416317689530686, + "grad_norm": 0.024539409205317497, + "learning_rate": 3.9079860230062286e-06, + "loss": 0.7451, + "step": 1226 + }, + { + "epoch": 1.4174729241877255, + "grad_norm": 0.026942811906337738, + "learning_rate": 3.8990832587098396e-06, + "loss": 0.7825, + "step": 1227 + }, + { + "epoch": 1.4186281588447653, + "grad_norm": 0.026073571294546127, + "learning_rate": 3.890154587283447e-06, + "loss": 0.7788, + "step": 1228 + }, + { + "epoch": 1.4197833935018052, + "grad_norm": 0.02608587220311165, + "learning_rate": 3.881200174090758e-06, + "loss": 0.7752, + "step": 1229 + }, + { + "epoch": 1.4209386281588448, + "grad_norm": 0.02780032530426979, + "learning_rate": 3.8722201849722264e-06, + "loss": 0.7834, + "step": 1230 + }, + { + "epoch": 1.4220938628158843, + "grad_norm": 0.025982137769460678, + "learning_rate": 3.863214786241993e-06, + "loss": 0.7633, + "step": 1231 + }, + { + "epoch": 1.4232490974729242, + "grad_norm": 0.026244191452860832, + "learning_rate": 3.85418414468479e-06, + "loss": 0.7901, + "step": 1232 + }, + { + "epoch": 1.424404332129964, + "grad_norm": 0.028074799105525017, + "learning_rate": 3.845128427552865e-06, + "loss": 0.7579, + "step": 1233 + }, + { + "epoch": 1.4255595667870036, + "grad_norm": 0.027848297730088234, + "learning_rate": 3.836047802562878e-06, + "loss": 0.7633, + "step": 1234 + }, + { + "epoch": 1.4267148014440434, + "grad_norm": 0.026308046653866768, + "learning_rate": 3.8269424378927925e-06, + "loss": 0.7959, + "step": 1235 + }, + { + "epoch": 1.427870036101083, + "grad_norm": 0.02657618559896946, + "learning_rate": 3.81781250217877e-06, + "loss": 0.782, + "step": 1236 + }, + { + "epoch": 1.4290252707581228, + "grad_norm": 0.02592160366475582, + "learning_rate": 3.8086581645120347e-06, + "loss": 0.7875, + "step": 1237 + }, + { + "epoch": 1.4301805054151624, + "grad_norm": 0.02750803716480732, + "learning_rate": 3.799479594435751e-06, + "loss": 0.7712, + "step": 1238 + }, + { + "epoch": 1.4313357400722022, + "grad_norm": 0.02752981334924698, + "learning_rate": 3.7902769619418783e-06, + "loss": 0.8094, + "step": 1239 + }, + { + "epoch": 1.4324909747292418, + "grad_norm": 0.025965960696339607, + "learning_rate": 3.7810504374680274e-06, + "loss": 0.7687, + "step": 1240 + }, + { + "epoch": 1.4336462093862816, + "grad_norm": 0.025762738659977913, + "learning_rate": 3.7718001918942982e-06, + "loss": 0.7943, + "step": 1241 + }, + { + "epoch": 1.4348014440433212, + "grad_norm": 0.026879120618104935, + "learning_rate": 3.762526396540121e-06, + "loss": 0.7735, + "step": 1242 + }, + { + "epoch": 1.435956678700361, + "grad_norm": 0.02714950405061245, + "learning_rate": 3.7532292231610774e-06, + "loss": 0.7641, + "step": 1243 + }, + { + "epoch": 1.4371119133574006, + "grad_norm": 0.02679312974214554, + "learning_rate": 3.7439088439457257e-06, + "loss": 0.7627, + "step": 1244 + }, + { + "epoch": 1.4382671480144404, + "grad_norm": 0.027805205434560776, + "learning_rate": 3.7345654315124035e-06, + "loss": 0.7811, + "step": 1245 + }, + { + "epoch": 1.4394223826714803, + "grad_norm": 0.02763986401259899, + "learning_rate": 3.725199158906041e-06, + "loss": 0.7924, + "step": 1246 + }, + { + "epoch": 1.4405776173285199, + "grad_norm": 0.026331216096878052, + "learning_rate": 3.7158101995949486e-06, + "loss": 0.7601, + "step": 1247 + }, + { + "epoch": 1.4417328519855594, + "grad_norm": 0.025142505764961243, + "learning_rate": 3.706398727467606e-06, + "loss": 0.7624, + "step": 1248 + }, + { + "epoch": 1.4428880866425993, + "grad_norm": 0.02704186551272869, + "learning_rate": 3.696964916829443e-06, + "loss": 0.7867, + "step": 1249 + }, + { + "epoch": 1.444043321299639, + "grad_norm": 0.027327412739396095, + "learning_rate": 3.6875089423996097e-06, + "loss": 0.7861, + "step": 1250 + }, + { + "epoch": 1.4451985559566787, + "grad_norm": 0.027434786781668663, + "learning_rate": 3.678030979307741e-06, + "loss": 0.7869, + "step": 1251 + }, + { + "epoch": 1.4463537906137185, + "grad_norm": 0.02410939894616604, + "learning_rate": 3.6685312030907167e-06, + "loss": 0.7568, + "step": 1252 + }, + { + "epoch": 1.447509025270758, + "grad_norm": 0.02866499125957489, + "learning_rate": 3.6590097896894017e-06, + "loss": 0.8089, + "step": 1253 + }, + { + "epoch": 1.448664259927798, + "grad_norm": 0.025909971445798874, + "learning_rate": 3.6494669154453978e-06, + "loss": 0.794, + "step": 1254 + }, + { + "epoch": 1.4498194945848375, + "grad_norm": 0.026149652898311615, + "learning_rate": 3.6399027570977703e-06, + "loss": 0.7657, + "step": 1255 + }, + { + "epoch": 1.4509747292418773, + "grad_norm": 0.027486305683851242, + "learning_rate": 3.6303174917797783e-06, + "loss": 0.7573, + "step": 1256 + }, + { + "epoch": 1.452129963898917, + "grad_norm": 0.026965491473674774, + "learning_rate": 3.6207112970155915e-06, + "loss": 0.7739, + "step": 1257 + }, + { + "epoch": 1.4532851985559567, + "grad_norm": 0.026813145726919174, + "learning_rate": 3.611084350717008e-06, + "loss": 0.7612, + "step": 1258 + }, + { + "epoch": 1.4544404332129963, + "grad_norm": 0.027455372735857964, + "learning_rate": 3.60143683118015e-06, + "loss": 0.7892, + "step": 1259 + }, + { + "epoch": 1.4555956678700361, + "grad_norm": 0.027145925909280777, + "learning_rate": 3.5917689170821714e-06, + "loss": 0.7806, + "step": 1260 + }, + { + "epoch": 1.4567509025270757, + "grad_norm": 0.02541586011648178, + "learning_rate": 3.582080787477941e-06, + "loss": 0.7772, + "step": 1261 + }, + { + "epoch": 1.4579061371841155, + "grad_norm": 0.028484217822551727, + "learning_rate": 3.5723726217967315e-06, + "loss": 0.7926, + "step": 1262 + }, + { + "epoch": 1.4590613718411554, + "grad_norm": 0.027112239971756935, + "learning_rate": 3.5626445998388926e-06, + "loss": 0.7819, + "step": 1263 + }, + { + "epoch": 1.460216606498195, + "grad_norm": 0.027995990589261055, + "learning_rate": 3.552896901772525e-06, + "loss": 0.7845, + "step": 1264 + }, + { + "epoch": 1.4613718411552346, + "grad_norm": 0.026287071406841278, + "learning_rate": 3.543129708130139e-06, + "loss": 0.7979, + "step": 1265 + }, + { + "epoch": 1.4625270758122744, + "grad_norm": 0.026646843180060387, + "learning_rate": 3.533343199805315e-06, + "loss": 0.7869, + "step": 1266 + }, + { + "epoch": 1.4636823104693142, + "grad_norm": 0.025323543697595596, + "learning_rate": 3.52353755804935e-06, + "loss": 0.7765, + "step": 1267 + }, + { + "epoch": 1.4648375451263538, + "grad_norm": 0.025138625875115395, + "learning_rate": 3.5137129644679035e-06, + "loss": 0.7606, + "step": 1268 + }, + { + "epoch": 1.4659927797833934, + "grad_norm": 0.026612093672156334, + "learning_rate": 3.5038696010176316e-06, + "loss": 0.7624, + "step": 1269 + }, + { + "epoch": 1.4671480144404332, + "grad_norm": 0.025180159136652946, + "learning_rate": 3.4940076500028193e-06, + "loss": 0.7704, + "step": 1270 + }, + { + "epoch": 1.468303249097473, + "grad_norm": 0.0247277170419693, + "learning_rate": 3.484127294072003e-06, + "loss": 0.7799, + "step": 1271 + }, + { + "epoch": 1.4694584837545126, + "grad_norm": 0.025542214512825012, + "learning_rate": 3.474228716214588e-06, + "loss": 0.7859, + "step": 1272 + }, + { + "epoch": 1.4706137184115524, + "grad_norm": 0.02490636333823204, + "learning_rate": 3.464312099757457e-06, + "loss": 0.7553, + "step": 1273 + }, + { + "epoch": 1.471768953068592, + "grad_norm": 0.027338897809386253, + "learning_rate": 3.4543776283615815e-06, + "loss": 0.7869, + "step": 1274 + }, + { + "epoch": 1.4729241877256318, + "grad_norm": 0.0249380711466074, + "learning_rate": 3.444425486018611e-06, + "loss": 0.7734, + "step": 1275 + }, + { + "epoch": 1.4740794223826714, + "grad_norm": 0.02633252553641796, + "learning_rate": 3.434455857047475e-06, + "loss": 0.7714, + "step": 1276 + }, + { + "epoch": 1.4752346570397112, + "grad_norm": 0.026209495961666107, + "learning_rate": 3.424468926090961e-06, + "loss": 0.7912, + "step": 1277 + }, + { + "epoch": 1.4763898916967508, + "grad_norm": 0.02743818424642086, + "learning_rate": 3.4144648781122987e-06, + "loss": 0.7725, + "step": 1278 + }, + { + "epoch": 1.4775451263537906, + "grad_norm": 0.02479531429708004, + "learning_rate": 3.404443898391737e-06, + "loss": 0.7548, + "step": 1279 + }, + { + "epoch": 1.4787003610108302, + "grad_norm": 0.026155853644013405, + "learning_rate": 3.3944061725231055e-06, + "loss": 0.7882, + "step": 1280 + }, + { + "epoch": 1.47985559566787, + "grad_norm": 0.02570260874927044, + "learning_rate": 3.3843518864103857e-06, + "loss": 0.7617, + "step": 1281 + }, + { + "epoch": 1.4810108303249097, + "grad_norm": 0.026218149811029434, + "learning_rate": 3.374281226264259e-06, + "loss": 0.7798, + "step": 1282 + }, + { + "epoch": 1.4821660649819495, + "grad_norm": 0.025268124416470528, + "learning_rate": 3.364194378598666e-06, + "loss": 0.7647, + "step": 1283 + }, + { + "epoch": 1.4833212996389893, + "grad_norm": 0.028134865686297417, + "learning_rate": 3.3540915302273476e-06, + "loss": 0.777, + "step": 1284 + }, + { + "epoch": 1.4844765342960289, + "grad_norm": 0.026391401886940002, + "learning_rate": 3.343972868260385e-06, + "loss": 0.7704, + "step": 1285 + }, + { + "epoch": 1.4856317689530685, + "grad_norm": 0.025334736332297325, + "learning_rate": 3.333838580100737e-06, + "loss": 0.7465, + "step": 1286 + }, + { + "epoch": 1.4867870036101083, + "grad_norm": 0.025972386822104454, + "learning_rate": 3.32368885344077e-06, + "loss": 0.798, + "step": 1287 + }, + { + "epoch": 1.487942238267148, + "grad_norm": 0.02676592580974102, + "learning_rate": 3.3135238762587723e-06, + "loss": 0.7946, + "step": 1288 + }, + { + "epoch": 1.4890974729241877, + "grad_norm": 0.026947803795337677, + "learning_rate": 3.3033438368154886e-06, + "loss": 0.7762, + "step": 1289 + }, + { + "epoch": 1.4902527075812273, + "grad_norm": 0.02568644843995571, + "learning_rate": 3.2931489236506183e-06, + "loss": 0.7977, + "step": 1290 + }, + { + "epoch": 1.4914079422382671, + "grad_norm": 0.02489590272307396, + "learning_rate": 3.282939325579333e-06, + "loss": 0.7634, + "step": 1291 + }, + { + "epoch": 1.492563176895307, + "grad_norm": 0.02679099142551422, + "learning_rate": 3.272715231688775e-06, + "loss": 0.7917, + "step": 1292 + }, + { + "epoch": 1.4937184115523465, + "grad_norm": 0.025316089391708374, + "learning_rate": 3.2624768313345594e-06, + "loss": 0.7507, + "step": 1293 + }, + { + "epoch": 1.4948736462093863, + "grad_norm": 0.024965766817331314, + "learning_rate": 3.25222431413726e-06, + "loss": 0.7751, + "step": 1294 + }, + { + "epoch": 1.496028880866426, + "grad_norm": 0.026891566812992096, + "learning_rate": 3.241957869978907e-06, + "loss": 0.7697, + "step": 1295 + }, + { + "epoch": 1.4971841155234658, + "grad_norm": 0.02724074199795723, + "learning_rate": 3.2316776889994627e-06, + "loss": 0.8087, + "step": 1296 + }, + { + "epoch": 1.4983393501805053, + "grad_norm": 0.025544699281454086, + "learning_rate": 3.221383961593303e-06, + "loss": 0.7511, + "step": 1297 + }, + { + "epoch": 1.4994945848375452, + "grad_norm": 0.02615639939904213, + "learning_rate": 3.2110768784056884e-06, + "loss": 0.7741, + "step": 1298 + }, + { + "epoch": 1.5006498194945848, + "grad_norm": 0.026993228122591972, + "learning_rate": 3.2007566303292416e-06, + "loss": 0.7631, + "step": 1299 + }, + { + "epoch": 1.5018050541516246, + "grad_norm": 0.02476736344397068, + "learning_rate": 3.1904234085003976e-06, + "loss": 0.7574, + "step": 1300 + }, + { + "epoch": 1.5029602888086644, + "grad_norm": 0.025930294767022133, + "learning_rate": 3.180077404295881e-06, + "loss": 0.8047, + "step": 1301 + }, + { + "epoch": 1.504115523465704, + "grad_norm": 0.025386760011315346, + "learning_rate": 3.1697188093291454e-06, + "loss": 0.7663, + "step": 1302 + }, + { + "epoch": 1.5052707581227436, + "grad_norm": 0.025852493941783905, + "learning_rate": 3.1593478154468364e-06, + "loss": 0.8022, + "step": 1303 + }, + { + "epoch": 1.5064259927797834, + "grad_norm": 0.024930665269494057, + "learning_rate": 3.148964614725232e-06, + "loss": 0.779, + "step": 1304 + }, + { + "epoch": 1.5075812274368232, + "grad_norm": 0.025383364409208298, + "learning_rate": 3.138569399466689e-06, + "loss": 0.7833, + "step": 1305 + }, + { + "epoch": 1.5087364620938628, + "grad_norm": 0.026384184136986732, + "learning_rate": 3.128162362196077e-06, + "loss": 0.7753, + "step": 1306 + }, + { + "epoch": 1.5098916967509024, + "grad_norm": 0.02558097243309021, + "learning_rate": 3.117743695657219e-06, + "loss": 0.7648, + "step": 1307 + }, + { + "epoch": 1.5110469314079422, + "grad_norm": 0.02592851221561432, + "learning_rate": 3.107313592809315e-06, + "loss": 0.7819, + "step": 1308 + }, + { + "epoch": 1.512202166064982, + "grad_norm": 0.02552459016442299, + "learning_rate": 3.0968722468233743e-06, + "loss": 0.7831, + "step": 1309 + }, + { + "epoch": 1.5133574007220216, + "grad_norm": 0.026385486125946045, + "learning_rate": 3.086419851078632e-06, + "loss": 0.7874, + "step": 1310 + }, + { + "epoch": 1.5145126353790612, + "grad_norm": 0.025948703289031982, + "learning_rate": 3.0759565991589733e-06, + "loss": 0.7687, + "step": 1311 + }, + { + "epoch": 1.515667870036101, + "grad_norm": 0.02731943503022194, + "learning_rate": 3.0654826848493436e-06, + "loss": 0.7722, + "step": 1312 + }, + { + "epoch": 1.5168231046931409, + "grad_norm": 0.02646474353969097, + "learning_rate": 3.0549983021321634e-06, + "loss": 0.7731, + "step": 1313 + }, + { + "epoch": 1.5179783393501805, + "grad_norm": 0.0249773059040308, + "learning_rate": 3.044503645183731e-06, + "loss": 0.7773, + "step": 1314 + }, + { + "epoch": 1.5191335740072203, + "grad_norm": 0.0276559516787529, + "learning_rate": 3.0339989083706306e-06, + "loss": 0.7911, + "step": 1315 + }, + { + "epoch": 1.5202888086642599, + "grad_norm": 0.026679757982492447, + "learning_rate": 3.0234842862461307e-06, + "loss": 0.7383, + "step": 1316 + }, + { + "epoch": 1.5214440433212997, + "grad_norm": 0.02656986005604267, + "learning_rate": 3.0129599735465816e-06, + "loss": 0.766, + "step": 1317 + }, + { + "epoch": 1.5225992779783395, + "grad_norm": 0.024834152311086655, + "learning_rate": 3.0024261651878056e-06, + "loss": 0.7661, + "step": 1318 + }, + { + "epoch": 1.523754512635379, + "grad_norm": 0.026497021317481995, + "learning_rate": 2.9918830562614927e-06, + "loss": 0.785, + "step": 1319 + }, + { + "epoch": 1.5249097472924187, + "grad_norm": 0.02679789625108242, + "learning_rate": 2.9813308420315825e-06, + "loss": 0.7814, + "step": 1320 + }, + { + "epoch": 1.5260649819494585, + "grad_norm": 0.02663526125252247, + "learning_rate": 2.9707697179306505e-06, + "loss": 0.7875, + "step": 1321 + }, + { + "epoch": 1.5272202166064983, + "grad_norm": 0.02751866728067398, + "learning_rate": 2.9601998795562865e-06, + "loss": 0.7683, + "step": 1322 + }, + { + "epoch": 1.528375451263538, + "grad_norm": 0.026420842856168747, + "learning_rate": 2.9496215226674747e-06, + "loss": 0.7671, + "step": 1323 + }, + { + "epoch": 1.5295306859205775, + "grad_norm": 0.026854515075683594, + "learning_rate": 2.9390348431809656e-06, + "loss": 0.7705, + "step": 1324 + }, + { + "epoch": 1.5306859205776173, + "grad_norm": 0.025556296110153198, + "learning_rate": 2.9284400371676486e-06, + "loss": 0.7771, + "step": 1325 + }, + { + "epoch": 1.5318411552346571, + "grad_norm": 0.025579258799552917, + "learning_rate": 2.917837300848923e-06, + "loss": 0.7683, + "step": 1326 + }, + { + "epoch": 1.5329963898916967, + "grad_norm": 0.027229176834225655, + "learning_rate": 2.9072268305930578e-06, + "loss": 0.789, + "step": 1327 + }, + { + "epoch": 1.5341516245487363, + "grad_norm": 0.025784116238355637, + "learning_rate": 2.896608822911561e-06, + "loss": 0.7699, + "step": 1328 + }, + { + "epoch": 1.5353068592057761, + "grad_norm": 0.02648346498608589, + "learning_rate": 2.8859834744555377e-06, + "loss": 0.773, + "step": 1329 + }, + { + "epoch": 1.536462093862816, + "grad_norm": 0.025939345359802246, + "learning_rate": 2.8753509820120464e-06, + "loss": 0.791, + "step": 1330 + }, + { + "epoch": 1.5376173285198556, + "grad_norm": 0.025402942672371864, + "learning_rate": 2.864711542500457e-06, + "loss": 0.7643, + "step": 1331 + }, + { + "epoch": 1.5387725631768951, + "grad_norm": 0.026020778343081474, + "learning_rate": 2.8540653529688033e-06, + "loss": 0.7813, + "step": 1332 + }, + { + "epoch": 1.539927797833935, + "grad_norm": 0.026241673156619072, + "learning_rate": 2.8434126105901333e-06, + "loss": 0.7905, + "step": 1333 + }, + { + "epoch": 1.5410830324909748, + "grad_norm": 0.02648642659187317, + "learning_rate": 2.8327535126588563e-06, + "loss": 0.771, + "step": 1334 + }, + { + "epoch": 1.5422382671480146, + "grad_norm": 0.02610144577920437, + "learning_rate": 2.8220882565870893e-06, + "loss": 0.792, + "step": 1335 + }, + { + "epoch": 1.5433935018050542, + "grad_norm": 0.025254933163523674, + "learning_rate": 2.811417039901001e-06, + "loss": 0.7674, + "step": 1336 + }, + { + "epoch": 1.5445487364620938, + "grad_norm": 0.025538668036460876, + "learning_rate": 2.800740060237157e-06, + "loss": 0.7745, + "step": 1337 + }, + { + "epoch": 1.5457039711191336, + "grad_norm": 0.025685755535960197, + "learning_rate": 2.7900575153388525e-06, + "loss": 0.7645, + "step": 1338 + }, + { + "epoch": 1.5468592057761734, + "grad_norm": 0.026221225038170815, + "learning_rate": 2.779369603052456e-06, + "loss": 0.8149, + "step": 1339 + }, + { + "epoch": 1.548014440433213, + "grad_norm": 0.02525060623884201, + "learning_rate": 2.768676521323742e-06, + "loss": 0.7552, + "step": 1340 + }, + { + "epoch": 1.5491696750902526, + "grad_norm": 0.026158245280385017, + "learning_rate": 2.757978468194226e-06, + "loss": 0.7712, + "step": 1341 + }, + { + "epoch": 1.5503249097472924, + "grad_norm": 0.02666587382555008, + "learning_rate": 2.747275641797496e-06, + "loss": 0.7511, + "step": 1342 + }, + { + "epoch": 1.5514801444043322, + "grad_norm": 0.026437923312187195, + "learning_rate": 2.736568240355544e-06, + "loss": 0.7659, + "step": 1343 + }, + { + "epoch": 1.5526353790613718, + "grad_norm": 0.0248862411826849, + "learning_rate": 2.725856462175095e-06, + "loss": 0.7547, + "step": 1344 + }, + { + "epoch": 1.5537906137184114, + "grad_norm": 0.026547906920313835, + "learning_rate": 2.715140505643932e-06, + "loss": 0.7996, + "step": 1345 + }, + { + "epoch": 1.5549458483754512, + "grad_norm": 0.02599526010453701, + "learning_rate": 2.704420569227224e-06, + "loss": 0.7641, + "step": 1346 + }, + { + "epoch": 1.556101083032491, + "grad_norm": 0.02453439123928547, + "learning_rate": 2.6936968514638495e-06, + "loss": 0.7466, + "step": 1347 + }, + { + "epoch": 1.5572563176895307, + "grad_norm": 0.027833128347992897, + "learning_rate": 2.68296955096272e-06, + "loss": 0.7826, + "step": 1348 + }, + { + "epoch": 1.5584115523465703, + "grad_norm": 0.026869479566812515, + "learning_rate": 2.672238866399099e-06, + "loss": 0.789, + "step": 1349 + }, + { + "epoch": 1.55956678700361, + "grad_norm": 0.025666510686278343, + "learning_rate": 2.6615049965109296e-06, + "loss": 0.784, + "step": 1350 + }, + { + "epoch": 1.5607220216606499, + "grad_norm": 0.02734527923166752, + "learning_rate": 2.6507681400951414e-06, + "loss": 0.7934, + "step": 1351 + }, + { + "epoch": 1.5618772563176895, + "grad_norm": 0.025378704071044922, + "learning_rate": 2.640028496003984e-06, + "loss": 0.7554, + "step": 1352 + }, + { + "epoch": 1.5630324909747293, + "grad_norm": 0.02528882399201393, + "learning_rate": 2.629286263141329e-06, + "loss": 0.7652, + "step": 1353 + }, + { + "epoch": 1.5641877256317689, + "grad_norm": 0.0248930174857378, + "learning_rate": 2.6185416404589985e-06, + "loss": 0.7515, + "step": 1354 + }, + { + "epoch": 1.5653429602888087, + "grad_norm": 0.025928398594260216, + "learning_rate": 2.607794826953075e-06, + "loss": 0.7839, + "step": 1355 + }, + { + "epoch": 1.5664981949458485, + "grad_norm": 0.025345591828227043, + "learning_rate": 2.5970460216602133e-06, + "loss": 0.773, + "step": 1356 + }, + { + "epoch": 1.5676534296028881, + "grad_norm": 0.02638527937233448, + "learning_rate": 2.5862954236539587e-06, + "loss": 0.7641, + "step": 1357 + }, + { + "epoch": 1.5688086642599277, + "grad_norm": 0.025557557120919228, + "learning_rate": 2.5755432320410594e-06, + "loss": 0.7869, + "step": 1358 + }, + { + "epoch": 1.5699638989169675, + "grad_norm": 0.025332162156701088, + "learning_rate": 2.5647896459577757e-06, + "loss": 0.7873, + "step": 1359 + }, + { + "epoch": 1.5711191335740073, + "grad_norm": 0.025012118741869926, + "learning_rate": 2.5540348645661965e-06, + "loss": 0.7805, + "step": 1360 + }, + { + "epoch": 1.572274368231047, + "grad_norm": 0.02651703916490078, + "learning_rate": 2.5432790870505487e-06, + "loss": 0.791, + "step": 1361 + }, + { + "epoch": 1.5734296028880865, + "grad_norm": 0.025666911154985428, + "learning_rate": 2.5325225126135045e-06, + "loss": 0.7814, + "step": 1362 + }, + { + "epoch": 1.5745848375451263, + "grad_norm": 0.02593865618109703, + "learning_rate": 2.521765340472499e-06, + "loss": 0.8046, + "step": 1363 + }, + { + "epoch": 1.5757400722021662, + "grad_norm": 0.025667283684015274, + "learning_rate": 2.5110077698560364e-06, + "loss": 0.757, + "step": 1364 + }, + { + "epoch": 1.5768953068592058, + "grad_norm": 0.025065291672945023, + "learning_rate": 2.5002500000000003e-06, + "loss": 0.7729, + "step": 1365 + }, + { + "epoch": 1.5780505415162454, + "grad_norm": 0.025896169245243073, + "learning_rate": 2.4894922301439643e-06, + "loss": 0.7825, + "step": 1366 + }, + { + "epoch": 1.5792057761732852, + "grad_norm": 0.02581976167857647, + "learning_rate": 2.478734659527501e-06, + "loss": 0.7783, + "step": 1367 + }, + { + "epoch": 1.580361010830325, + "grad_norm": 0.025588102638721466, + "learning_rate": 2.467977487386497e-06, + "loss": 0.7617, + "step": 1368 + }, + { + "epoch": 1.5815162454873646, + "grad_norm": 0.02477039210498333, + "learning_rate": 2.4572209129494524e-06, + "loss": 0.7494, + "step": 1369 + }, + { + "epoch": 1.5826714801444042, + "grad_norm": 0.027211980894207954, + "learning_rate": 2.4464651354338037e-06, + "loss": 0.8013, + "step": 1370 + }, + { + "epoch": 1.583826714801444, + "grad_norm": 0.02574116736650467, + "learning_rate": 2.435710354042225e-06, + "loss": 0.7706, + "step": 1371 + }, + { + "epoch": 1.5849819494584838, + "grad_norm": 0.025494717061519623, + "learning_rate": 2.4249567679589417e-06, + "loss": 0.8006, + "step": 1372 + }, + { + "epoch": 1.5861371841155236, + "grad_norm": 0.024281838908791542, + "learning_rate": 2.4142045763460415e-06, + "loss": 0.7609, + "step": 1373 + }, + { + "epoch": 1.5872924187725632, + "grad_norm": 0.025773445144295692, + "learning_rate": 2.403453978339787e-06, + "loss": 0.8048, + "step": 1374 + }, + { + "epoch": 1.5884476534296028, + "grad_norm": 0.02534416690468788, + "learning_rate": 2.392705173046925e-06, + "loss": 0.7731, + "step": 1375 + }, + { + "epoch": 1.5896028880866426, + "grad_norm": 0.02497626096010208, + "learning_rate": 2.3819583595410013e-06, + "loss": 0.7858, + "step": 1376 + }, + { + "epoch": 1.5907581227436824, + "grad_norm": 0.02502652071416378, + "learning_rate": 2.3712137368586717e-06, + "loss": 0.7438, + "step": 1377 + }, + { + "epoch": 1.591913357400722, + "grad_norm": 0.024966862052679062, + "learning_rate": 2.3604715039960173e-06, + "loss": 0.7835, + "step": 1378 + }, + { + "epoch": 1.5930685920577616, + "grad_norm": 0.025360535830259323, + "learning_rate": 2.3497318599048592e-06, + "loss": 0.7716, + "step": 1379 + }, + { + "epoch": 1.5942238267148015, + "grad_norm": 0.026140306144952774, + "learning_rate": 2.338995003489072e-06, + "loss": 0.7556, + "step": 1380 + }, + { + "epoch": 1.5953790613718413, + "grad_norm": 0.02430606074631214, + "learning_rate": 2.328261133600902e-06, + "loss": 0.7708, + "step": 1381 + }, + { + "epoch": 1.5965342960288809, + "grad_norm": 0.02544374018907547, + "learning_rate": 2.317530449037281e-06, + "loss": 0.7837, + "step": 1382 + }, + { + "epoch": 1.5976895306859205, + "grad_norm": 0.025840749964118004, + "learning_rate": 2.3068031485361516e-06, + "loss": 0.7575, + "step": 1383 + }, + { + "epoch": 1.5988447653429603, + "grad_norm": 0.027307111769914627, + "learning_rate": 2.296079430772777e-06, + "loss": 0.8002, + "step": 1384 + }, + { + "epoch": 1.6, + "grad_norm": 0.026851478964090347, + "learning_rate": 2.285359494356068e-06, + "loss": 0.8076, + "step": 1385 + }, + { + "epoch": 1.6011552346570397, + "grad_norm": 0.025269558653235435, + "learning_rate": 2.274643537824905e-06, + "loss": 0.7601, + "step": 1386 + }, + { + "epoch": 1.6023104693140793, + "grad_norm": 0.024502936750650406, + "learning_rate": 2.263931759644456e-06, + "loss": 0.7777, + "step": 1387 + }, + { + "epoch": 1.603465703971119, + "grad_norm": 0.025695212185382843, + "learning_rate": 2.2532243582025048e-06, + "loss": 0.7779, + "step": 1388 + }, + { + "epoch": 1.604620938628159, + "grad_norm": 0.025510141626000404, + "learning_rate": 2.242521531805775e-06, + "loss": 0.7623, + "step": 1389 + }, + { + "epoch": 1.6057761732851985, + "grad_norm": 0.025397466495633125, + "learning_rate": 2.2318234786762586e-06, + "loss": 0.769, + "step": 1390 + }, + { + "epoch": 1.6069314079422383, + "grad_norm": 0.02594076469540596, + "learning_rate": 2.2211303969475444e-06, + "loss": 0.8051, + "step": 1391 + }, + { + "epoch": 1.608086642599278, + "grad_norm": 0.024837691336870193, + "learning_rate": 2.210442484661148e-06, + "loss": 0.8117, + "step": 1392 + }, + { + "epoch": 1.6092418772563177, + "grad_norm": 0.026649747043848038, + "learning_rate": 2.199759939762843e-06, + "loss": 0.7768, + "step": 1393 + }, + { + "epoch": 1.6103971119133575, + "grad_norm": 0.026813820004463196, + "learning_rate": 2.1890829600989995e-06, + "loss": 0.7575, + "step": 1394 + }, + { + "epoch": 1.6115523465703971, + "grad_norm": 0.026133766397833824, + "learning_rate": 2.1784117434129113e-06, + "loss": 0.7599, + "step": 1395 + }, + { + "epoch": 1.6127075812274367, + "grad_norm": 0.026183927431702614, + "learning_rate": 2.167746487341144e-06, + "loss": 0.7661, + "step": 1396 + }, + { + "epoch": 1.6138628158844766, + "grad_norm": 0.02729174494743347, + "learning_rate": 2.157087389409867e-06, + "loss": 0.7698, + "step": 1397 + }, + { + "epoch": 1.6150180505415164, + "grad_norm": 0.026295259594917297, + "learning_rate": 2.1464346470311965e-06, + "loss": 0.7695, + "step": 1398 + }, + { + "epoch": 1.616173285198556, + "grad_norm": 0.026081614196300507, + "learning_rate": 2.135788457499544e-06, + "loss": 0.7622, + "step": 1399 + }, + { + "epoch": 1.6173285198555956, + "grad_norm": 0.025280749425292015, + "learning_rate": 2.1251490179879547e-06, + "loss": 0.7639, + "step": 1400 + }, + { + "epoch": 1.6184837545126354, + "grad_norm": 0.02538752183318138, + "learning_rate": 2.1145165255444643e-06, + "loss": 0.7893, + "step": 1401 + }, + { + "epoch": 1.6196389891696752, + "grad_norm": 0.026337897405028343, + "learning_rate": 2.1038911770884395e-06, + "loss": 0.7936, + "step": 1402 + }, + { + "epoch": 1.6207942238267148, + "grad_norm": 0.027165520936250687, + "learning_rate": 2.0932731694069438e-06, + "loss": 0.7506, + "step": 1403 + }, + { + "epoch": 1.6219494584837544, + "grad_norm": 0.02547120675444603, + "learning_rate": 2.0826626991510775e-06, + "loss": 0.7623, + "step": 1404 + }, + { + "epoch": 1.6231046931407942, + "grad_norm": 0.023889468982815742, + "learning_rate": 2.072059962832352e-06, + "loss": 0.7444, + "step": 1405 + }, + { + "epoch": 1.624259927797834, + "grad_norm": 0.025629183277487755, + "learning_rate": 2.0614651568190354e-06, + "loss": 0.7869, + "step": 1406 + }, + { + "epoch": 1.6254151624548736, + "grad_norm": 0.02670186199247837, + "learning_rate": 2.050878477332526e-06, + "loss": 0.8075, + "step": 1407 + }, + { + "epoch": 1.6265703971119132, + "grad_norm": 0.025403102859854698, + "learning_rate": 2.040300120443714e-06, + "loss": 0.7647, + "step": 1408 + }, + { + "epoch": 1.627725631768953, + "grad_norm": 0.025784939527511597, + "learning_rate": 2.02973028206935e-06, + "loss": 0.7594, + "step": 1409 + }, + { + "epoch": 1.6288808664259928, + "grad_norm": 0.024970732629299164, + "learning_rate": 2.019169157968418e-06, + "loss": 0.7664, + "step": 1410 + }, + { + "epoch": 1.6300361010830327, + "grad_norm": 0.025724420323967934, + "learning_rate": 2.008616943738508e-06, + "loss": 0.7637, + "step": 1411 + }, + { + "epoch": 1.6311913357400722, + "grad_norm": 0.02508534863591194, + "learning_rate": 1.9980738348121955e-06, + "loss": 0.7674, + "step": 1412 + }, + { + "epoch": 1.6323465703971118, + "grad_norm": 0.024077199399471283, + "learning_rate": 1.9875400264534194e-06, + "loss": 0.7703, + "step": 1413 + }, + { + "epoch": 1.6335018050541517, + "grad_norm": 0.02483062632381916, + "learning_rate": 1.97701571375387e-06, + "loss": 0.7825, + "step": 1414 + }, + { + "epoch": 1.6346570397111915, + "grad_norm": 0.025150980800390244, + "learning_rate": 1.9665010916293704e-06, + "loss": 0.7875, + "step": 1415 + }, + { + "epoch": 1.635812274368231, + "grad_norm": 0.02459767460823059, + "learning_rate": 1.95599635481627e-06, + "loss": 0.7745, + "step": 1416 + }, + { + "epoch": 1.6369675090252707, + "grad_norm": 0.024894608184695244, + "learning_rate": 1.9455016978678377e-06, + "loss": 0.7825, + "step": 1417 + }, + { + "epoch": 1.6381227436823105, + "grad_norm": 0.02483339235186577, + "learning_rate": 1.935017315150656e-06, + "loss": 0.7654, + "step": 1418 + }, + { + "epoch": 1.6392779783393503, + "grad_norm": 0.025194387882947922, + "learning_rate": 1.9245434008410273e-06, + "loss": 0.7513, + "step": 1419 + }, + { + "epoch": 1.6404332129963899, + "grad_norm": 0.025890490040183067, + "learning_rate": 1.914080148921368e-06, + "loss": 0.7756, + "step": 1420 + }, + { + "epoch": 1.6415884476534295, + "grad_norm": 0.025699496269226074, + "learning_rate": 1.903627753176627e-06, + "loss": 0.7702, + "step": 1421 + }, + { + "epoch": 1.6427436823104693, + "grad_norm": 0.02435440570116043, + "learning_rate": 1.893186407190685e-06, + "loss": 0.7727, + "step": 1422 + }, + { + "epoch": 1.6438989169675091, + "grad_norm": 0.024738416075706482, + "learning_rate": 1.8827563043427818e-06, + "loss": 0.7856, + "step": 1423 + }, + { + "epoch": 1.6450541516245487, + "grad_norm": 0.025477735325694084, + "learning_rate": 1.8723376378039229e-06, + "loss": 0.775, + "step": 1424 + }, + { + "epoch": 1.6462093862815883, + "grad_norm": 0.02498653531074524, + "learning_rate": 1.8619306005333118e-06, + "loss": 0.7464, + "step": 1425 + }, + { + "epoch": 1.6473646209386281, + "grad_norm": 0.025513045489788055, + "learning_rate": 1.851535385274768e-06, + "loss": 0.7944, + "step": 1426 + }, + { + "epoch": 1.648519855595668, + "grad_norm": 0.024236127734184265, + "learning_rate": 1.8411521845531638e-06, + "loss": 0.7646, + "step": 1427 + }, + { + "epoch": 1.6496750902527075, + "grad_norm": 0.025221243500709534, + "learning_rate": 1.8307811906708552e-06, + "loss": 0.7802, + "step": 1428 + }, + { + "epoch": 1.6508303249097471, + "grad_norm": 0.024581240490078926, + "learning_rate": 1.8204225957041197e-06, + "loss": 0.7686, + "step": 1429 + }, + { + "epoch": 1.651985559566787, + "grad_norm": 0.025168737396597862, + "learning_rate": 1.8100765914996026e-06, + "loss": 0.79, + "step": 1430 + }, + { + "epoch": 1.6531407942238268, + "grad_norm": 0.024541551247239113, + "learning_rate": 1.7997433696707597e-06, + "loss": 0.766, + "step": 1431 + }, + { + "epoch": 1.6542960288808666, + "grad_norm": 0.02476644702255726, + "learning_rate": 1.789423121594312e-06, + "loss": 0.7718, + "step": 1432 + }, + { + "epoch": 1.6554512635379062, + "grad_norm": 0.027230584993958473, + "learning_rate": 1.779116038406698e-06, + "loss": 0.7813, + "step": 1433 + }, + { + "epoch": 1.6566064981949458, + "grad_norm": 0.025931481271982193, + "learning_rate": 1.768822311000538e-06, + "loss": 0.7772, + "step": 1434 + }, + { + "epoch": 1.6577617328519856, + "grad_norm": 0.025082925334572792, + "learning_rate": 1.7585421300210932e-06, + "loss": 0.7958, + "step": 1435 + }, + { + "epoch": 1.6589169675090254, + "grad_norm": 0.02672106772661209, + "learning_rate": 1.7482756858627406e-06, + "loss": 0.774, + "step": 1436 + }, + { + "epoch": 1.660072202166065, + "grad_norm": 0.026696957647800446, + "learning_rate": 1.7380231686654419e-06, + "loss": 0.7763, + "step": 1437 + }, + { + "epoch": 1.6612274368231046, + "grad_norm": 0.025145625695586205, + "learning_rate": 1.7277847683112251e-06, + "loss": 0.7602, + "step": 1438 + }, + { + "epoch": 1.6623826714801444, + "grad_norm": 0.025679778307676315, + "learning_rate": 1.7175606744206683e-06, + "loss": 0.7772, + "step": 1439 + }, + { + "epoch": 1.6635379061371842, + "grad_norm": 0.02643396146595478, + "learning_rate": 1.707351076349382e-06, + "loss": 0.7776, + "step": 1440 + }, + { + "epoch": 1.6646931407942238, + "grad_norm": 0.025024591013789177, + "learning_rate": 1.697156163184512e-06, + "loss": 0.769, + "step": 1441 + }, + { + "epoch": 1.6658483754512634, + "grad_norm": 0.02451016753911972, + "learning_rate": 1.6869761237412275e-06, + "loss": 0.7614, + "step": 1442 + }, + { + "epoch": 1.6670036101083032, + "grad_norm": 0.02593870460987091, + "learning_rate": 1.6768111465592312e-06, + "loss": 0.7828, + "step": 1443 + }, + { + "epoch": 1.668158844765343, + "grad_norm": 0.024905387312173843, + "learning_rate": 1.6666614198992625e-06, + "loss": 0.7847, + "step": 1444 + }, + { + "epoch": 1.6693140794223826, + "grad_norm": 0.025363383814692497, + "learning_rate": 1.6565271317396156e-06, + "loss": 0.7757, + "step": 1445 + }, + { + "epoch": 1.6704693140794222, + "grad_norm": 0.02425825409591198, + "learning_rate": 1.6464084697726532e-06, + "loss": 0.756, + "step": 1446 + }, + { + "epoch": 1.671624548736462, + "grad_norm": 0.025201529264450073, + "learning_rate": 1.636305621401335e-06, + "loss": 0.7701, + "step": 1447 + }, + { + "epoch": 1.6727797833935019, + "grad_norm": 0.02498772367835045, + "learning_rate": 1.626218773735742e-06, + "loss": 0.8012, + "step": 1448 + }, + { + "epoch": 1.6739350180505417, + "grad_norm": 0.02500982955098152, + "learning_rate": 1.6161481135896154e-06, + "loss": 0.7796, + "step": 1449 + }, + { + "epoch": 1.6750902527075813, + "grad_norm": 0.025785304605960846, + "learning_rate": 1.606093827476895e-06, + "loss": 0.7584, + "step": 1450 + }, + { + "epoch": 1.6762454873646209, + "grad_norm": 0.024516083300113678, + "learning_rate": 1.5960561016082638e-06, + "loss": 0.7812, + "step": 1451 + }, + { + "epoch": 1.6774007220216607, + "grad_norm": 0.025357436388731003, + "learning_rate": 1.5860351218877021e-06, + "loss": 0.7642, + "step": 1452 + }, + { + "epoch": 1.6785559566787005, + "grad_norm": 0.025423161685466766, + "learning_rate": 1.5760310739090402e-06, + "loss": 0.7838, + "step": 1453 + }, + { + "epoch": 1.67971119133574, + "grad_norm": 0.025020305067300797, + "learning_rate": 1.5660441429525258e-06, + "loss": 0.7713, + "step": 1454 + }, + { + "epoch": 1.6808664259927797, + "grad_norm": 0.02521519362926483, + "learning_rate": 1.556074513981389e-06, + "loss": 0.8022, + "step": 1455 + }, + { + "epoch": 1.6820216606498195, + "grad_norm": 0.02515571005642414, + "learning_rate": 1.5461223716384196e-06, + "loss": 0.7794, + "step": 1456 + }, + { + "epoch": 1.6831768953068593, + "grad_norm": 0.025751549750566483, + "learning_rate": 1.5361879002425432e-06, + "loss": 0.8067, + "step": 1457 + }, + { + "epoch": 1.684332129963899, + "grad_norm": 0.024167675524950027, + "learning_rate": 1.5262712837854139e-06, + "loss": 0.7497, + "step": 1458 + }, + { + "epoch": 1.6854873646209385, + "grad_norm": 0.025504915043711662, + "learning_rate": 1.5163727059279972e-06, + "loss": 0.7645, + "step": 1459 + }, + { + "epoch": 1.6866425992779783, + "grad_norm": 0.026490593329072, + "learning_rate": 1.506492349997181e-06, + "loss": 0.7698, + "step": 1460 + }, + { + "epoch": 1.6877978339350181, + "grad_norm": 0.024468280375003815, + "learning_rate": 1.4966303989823688e-06, + "loss": 0.7541, + "step": 1461 + }, + { + "epoch": 1.6889530685920577, + "grad_norm": 0.02487323433160782, + "learning_rate": 1.4867870355320972e-06, + "loss": 0.7611, + "step": 1462 + }, + { + "epoch": 1.6901083032490973, + "grad_norm": 0.024844063445925713, + "learning_rate": 1.4769624419506507e-06, + "loss": 0.7598, + "step": 1463 + }, + { + "epoch": 1.6912635379061371, + "grad_norm": 0.024609016254544258, + "learning_rate": 1.4671568001946851e-06, + "loss": 0.7863, + "step": 1464 + }, + { + "epoch": 1.692418772563177, + "grad_norm": 0.025440316647291183, + "learning_rate": 1.4573702918698615e-06, + "loss": 0.7955, + "step": 1465 + }, + { + "epoch": 1.6935740072202166, + "grad_norm": 0.024768877774477005, + "learning_rate": 1.4476030982274746e-06, + "loss": 0.7666, + "step": 1466 + }, + { + "epoch": 1.6947292418772562, + "grad_norm": 0.024406442418694496, + "learning_rate": 1.4378554001611087e-06, + "loss": 0.747, + "step": 1467 + }, + { + "epoch": 1.695884476534296, + "grad_norm": 0.02518375776708126, + "learning_rate": 1.4281273782032696e-06, + "loss": 0.7914, + "step": 1468 + }, + { + "epoch": 1.6970397111913358, + "grad_norm": 0.025701193138957024, + "learning_rate": 1.41841921252206e-06, + "loss": 0.7879, + "step": 1469 + }, + { + "epoch": 1.6981949458483756, + "grad_norm": 0.025167284533381462, + "learning_rate": 1.4087310829178295e-06, + "loss": 0.7707, + "step": 1470 + }, + { + "epoch": 1.6993501805054152, + "grad_norm": 0.024784136563539505, + "learning_rate": 1.3990631688198505e-06, + "loss": 0.7753, + "step": 1471 + }, + { + "epoch": 1.7005054151624548, + "grad_norm": 0.024584993720054626, + "learning_rate": 1.3894156492829932e-06, + "loss": 0.7579, + "step": 1472 + }, + { + "epoch": 1.7016606498194946, + "grad_norm": 0.025017768144607544, + "learning_rate": 1.3797887029844083e-06, + "loss": 0.7512, + "step": 1473 + }, + { + "epoch": 1.7028158844765344, + "grad_norm": 0.02527061477303505, + "learning_rate": 1.3701825082202227e-06, + "loss": 0.772, + "step": 1474 + }, + { + "epoch": 1.703971119133574, + "grad_norm": 0.02521699294447899, + "learning_rate": 1.3605972429022308e-06, + "loss": 0.766, + "step": 1475 + }, + { + "epoch": 1.7051263537906136, + "grad_norm": 0.024668460711836815, + "learning_rate": 1.3510330845546029e-06, + "loss": 0.7638, + "step": 1476 + }, + { + "epoch": 1.7062815884476534, + "grad_norm": 0.02491624839603901, + "learning_rate": 1.3414902103105984e-06, + "loss": 0.7572, + "step": 1477 + }, + { + "epoch": 1.7074368231046932, + "grad_norm": 0.02490483969449997, + "learning_rate": 1.3319687969092844e-06, + "loss": 0.7763, + "step": 1478 + }, + { + "epoch": 1.7085920577617328, + "grad_norm": 0.02502390742301941, + "learning_rate": 1.3224690206922594e-06, + "loss": 0.7859, + "step": 1479 + }, + { + "epoch": 1.7097472924187724, + "grad_norm": 0.024547290056943893, + "learning_rate": 1.3129910576003918e-06, + "loss": 0.76, + "step": 1480 + }, + { + "epoch": 1.7109025270758123, + "grad_norm": 0.024496039375662804, + "learning_rate": 1.3035350831705581e-06, + "loss": 0.7541, + "step": 1481 + }, + { + "epoch": 1.712057761732852, + "grad_norm": 0.0256884153932333, + "learning_rate": 1.2941012725323943e-06, + "loss": 0.7794, + "step": 1482 + }, + { + "epoch": 1.7132129963898917, + "grad_norm": 0.025446278974413872, + "learning_rate": 1.2846898004050527e-06, + "loss": 0.789, + "step": 1483 + }, + { + "epoch": 1.7143682310469313, + "grad_norm": 0.02467747963964939, + "learning_rate": 1.2753008410939593e-06, + "loss": 0.7704, + "step": 1484 + }, + { + "epoch": 1.715523465703971, + "grad_norm": 0.024466682225465775, + "learning_rate": 1.2659345684875971e-06, + "loss": 0.7751, + "step": 1485 + }, + { + "epoch": 1.716678700361011, + "grad_norm": 0.02474912814795971, + "learning_rate": 1.2565911560542752e-06, + "loss": 0.7721, + "step": 1486 + }, + { + "epoch": 1.7178339350180507, + "grad_norm": 0.024523675441741943, + "learning_rate": 1.2472707768389226e-06, + "loss": 0.7638, + "step": 1487 + }, + { + "epoch": 1.7189891696750903, + "grad_norm": 0.02448326162993908, + "learning_rate": 1.2379736034598793e-06, + "loss": 0.7898, + "step": 1488 + }, + { + "epoch": 1.72014440433213, + "grad_norm": 0.024665791541337967, + "learning_rate": 1.2286998081057024e-06, + "loss": 0.757, + "step": 1489 + }, + { + "epoch": 1.7212996389891697, + "grad_norm": 0.024392489343881607, + "learning_rate": 1.219449562531973e-06, + "loss": 0.7756, + "step": 1490 + }, + { + "epoch": 1.7224548736462095, + "grad_norm": 0.024781666696071625, + "learning_rate": 1.2102230380581221e-06, + "loss": 0.7851, + "step": 1491 + }, + { + "epoch": 1.7236101083032491, + "grad_norm": 0.024531826376914978, + "learning_rate": 1.2010204055642504e-06, + "loss": 0.7786, + "step": 1492 + }, + { + "epoch": 1.7247653429602887, + "grad_norm": 0.024737264961004257, + "learning_rate": 1.1918418354879655e-06, + "loss": 0.7918, + "step": 1493 + }, + { + "epoch": 1.7259205776173285, + "grad_norm": 0.024697288870811462, + "learning_rate": 1.1826874978212304e-06, + "loss": 0.7797, + "step": 1494 + }, + { + "epoch": 1.7270758122743683, + "grad_norm": 0.023727286607027054, + "learning_rate": 1.1735575621072076e-06, + "loss": 0.7622, + "step": 1495 + }, + { + "epoch": 1.728231046931408, + "grad_norm": 0.024227799847722054, + "learning_rate": 1.1644521974371236e-06, + "loss": 0.7538, + "step": 1496 + }, + { + "epoch": 1.7293862815884475, + "grad_norm": 0.02443331852555275, + "learning_rate": 1.1553715724471356e-06, + "loss": 0.7567, + "step": 1497 + }, + { + "epoch": 1.7305415162454874, + "grad_norm": 0.02441231720149517, + "learning_rate": 1.146315855315211e-06, + "loss": 0.763, + "step": 1498 + }, + { + "epoch": 1.7316967509025272, + "grad_norm": 0.02491583675146103, + "learning_rate": 1.1372852137580082e-06, + "loss": 0.7519, + "step": 1499 + }, + { + "epoch": 1.7328519855595668, + "grad_norm": 0.024958504363894463, + "learning_rate": 1.1282798150277738e-06, + "loss": 0.7704, + "step": 1500 + }, + { + "epoch": 1.7340072202166064, + "grad_norm": 0.02489922009408474, + "learning_rate": 1.1192998259092428e-06, + "loss": 0.8001, + "step": 1501 + }, + { + "epoch": 1.7351624548736462, + "grad_norm": 0.024233028292655945, + "learning_rate": 1.1103454127165526e-06, + "loss": 0.7725, + "step": 1502 + }, + { + "epoch": 1.736317689530686, + "grad_norm": 0.02405642904341221, + "learning_rate": 1.1014167412901609e-06, + "loss": 0.7227, + "step": 1503 + }, + { + "epoch": 1.7374729241877256, + "grad_norm": 0.025087906047701836, + "learning_rate": 1.092513976993772e-06, + "loss": 0.7852, + "step": 1504 + }, + { + "epoch": 1.7386281588447652, + "grad_norm": 0.025061985477805138, + "learning_rate": 1.083637284711279e-06, + "loss": 0.7648, + "step": 1505 + }, + { + "epoch": 1.739783393501805, + "grad_norm": 0.024631349369883537, + "learning_rate": 1.074786828843705e-06, + "loss": 0.741, + "step": 1506 + }, + { + "epoch": 1.7409386281588448, + "grad_norm": 0.024879056960344315, + "learning_rate": 1.0659627733061643e-06, + "loss": 0.7355, + "step": 1507 + }, + { + "epoch": 1.7420938628158846, + "grad_norm": 0.024683699011802673, + "learning_rate": 1.0571652815248199e-06, + "loss": 0.8, + "step": 1508 + }, + { + "epoch": 1.7432490974729242, + "grad_norm": 0.02492067962884903, + "learning_rate": 1.0483945164338617e-06, + "loss": 0.7701, + "step": 1509 + }, + { + "epoch": 1.7444043321299638, + "grad_norm": 0.02489222027361393, + "learning_rate": 1.039650640472486e-06, + "loss": 0.7831, + "step": 1510 + }, + { + "epoch": 1.7455595667870036, + "grad_norm": 0.024434035643935204, + "learning_rate": 1.0309338155818906e-06, + "loss": 0.7896, + "step": 1511 + }, + { + "epoch": 1.7467148014440435, + "grad_norm": 0.025025788694620132, + "learning_rate": 1.0222442032022705e-06, + "loss": 0.7806, + "step": 1512 + }, + { + "epoch": 1.747870036101083, + "grad_norm": 0.024875549599528313, + "learning_rate": 1.0135819642698318e-06, + "loss": 0.7844, + "step": 1513 + }, + { + "epoch": 1.7490252707581226, + "grad_norm": 0.024595865979790688, + "learning_rate": 1.0049472592138105e-06, + "loss": 0.7766, + "step": 1514 + }, + { + "epoch": 1.7501805054151625, + "grad_norm": 0.02502664178609848, + "learning_rate": 9.963402479535004e-07, + "loss": 0.7813, + "step": 1515 + }, + { + "epoch": 1.7513357400722023, + "grad_norm": 0.024631381034851074, + "learning_rate": 9.877610898952902e-07, + "loss": 0.79, + "step": 1516 + }, + { + "epoch": 1.7524909747292419, + "grad_norm": 0.023990454152226448, + "learning_rate": 9.792099439297134e-07, + "loss": 0.7875, + "step": 1517 + }, + { + "epoch": 1.7536462093862815, + "grad_norm": 0.0241263285279274, + "learning_rate": 9.706869684285066e-07, + "loss": 0.7605, + "step": 1518 + }, + { + "epoch": 1.7548014440433213, + "grad_norm": 0.02398371510207653, + "learning_rate": 9.621923212416713e-07, + "loss": 0.7778, + "step": 1519 + }, + { + "epoch": 1.755956678700361, + "grad_norm": 0.025420811027288437, + "learning_rate": 9.53726159694558e-07, + "loss": 0.7881, + "step": 1520 + }, + { + "epoch": 1.7571119133574007, + "grad_norm": 0.024340439587831497, + "learning_rate": 9.452886405849418e-07, + "loss": 0.7725, + "step": 1521 + }, + { + "epoch": 1.7582671480144403, + "grad_norm": 0.02469663880765438, + "learning_rate": 9.36879920180131e-07, + "loss": 0.771, + "step": 1522 + }, + { + "epoch": 1.75942238267148, + "grad_norm": 0.024449417367577553, + "learning_rate": 9.285001542140625e-07, + "loss": 0.7624, + "step": 1523 + }, + { + "epoch": 1.76057761732852, + "grad_norm": 0.024617070332169533, + "learning_rate": 9.201494978844251e-07, + "loss": 0.7763, + "step": 1524 + }, + { + "epoch": 1.7617328519855595, + "grad_norm": 0.02444697730243206, + "learning_rate": 9.118281058497792e-07, + "loss": 0.7634, + "step": 1525 + }, + { + "epoch": 1.7628880866425993, + "grad_norm": 0.02412438951432705, + "learning_rate": 9.03536132226695e-07, + "loss": 0.7501, + "step": 1526 + }, + { + "epoch": 1.764043321299639, + "grad_norm": 0.024980876594781876, + "learning_rate": 8.952737305869002e-07, + "loss": 0.802, + "step": 1527 + }, + { + "epoch": 1.7651985559566787, + "grad_norm": 0.02424163930118084, + "learning_rate": 8.870410539544322e-07, + "loss": 0.7846, + "step": 1528 + }, + { + "epoch": 1.7663537906137186, + "grad_norm": 0.025128301233053207, + "learning_rate": 8.788382548028056e-07, + "loss": 0.7551, + "step": 1529 + }, + { + "epoch": 1.7675090252707581, + "grad_norm": 0.02520165592432022, + "learning_rate": 8.706654850521883e-07, + "loss": 0.7736, + "step": 1530 + }, + { + "epoch": 1.7686642599277977, + "grad_norm": 0.02502501755952835, + "learning_rate": 8.625228960665893e-07, + "loss": 0.7632, + "step": 1531 + }, + { + "epoch": 1.7698194945848376, + "grad_norm": 0.02456534653902054, + "learning_rate": 8.544106386510526e-07, + "loss": 0.7514, + "step": 1532 + }, + { + "epoch": 1.7709747292418774, + "grad_norm": 0.02477426454424858, + "learning_rate": 8.463288630488653e-07, + "loss": 0.7741, + "step": 1533 + }, + { + "epoch": 1.772129963898917, + "grad_norm": 0.02434214949607849, + "learning_rate": 8.382777189387751e-07, + "loss": 0.7721, + "step": 1534 + }, + { + "epoch": 1.7732851985559566, + "grad_norm": 0.023917306214571, + "learning_rate": 8.302573554322192e-07, + "loss": 0.7507, + "step": 1535 + }, + { + "epoch": 1.7744404332129964, + "grad_norm": 0.024669747799634933, + "learning_rate": 8.222679210705636e-07, + "loss": 0.7993, + "step": 1536 + }, + { + "epoch": 1.7755956678700362, + "grad_norm": 0.024916965514421463, + "learning_rate": 8.143095638223446e-07, + "loss": 0.7536, + "step": 1537 + }, + { + "epoch": 1.7767509025270758, + "grad_norm": 0.02413412742316723, + "learning_rate": 8.063824310805396e-07, + "loss": 0.7552, + "step": 1538 + }, + { + "epoch": 1.7779061371841154, + "grad_norm": 0.025234825909137726, + "learning_rate": 7.984866696598284e-07, + "loss": 0.7845, + "step": 1539 + }, + { + "epoch": 1.7790613718411552, + "grad_norm": 0.024565840139985085, + "learning_rate": 7.906224257938796e-07, + "loss": 0.7674, + "step": 1540 + }, + { + "epoch": 1.780216606498195, + "grad_norm": 0.024029076099395752, + "learning_rate": 7.827898451326386e-07, + "loss": 0.7637, + "step": 1541 + }, + { + "epoch": 1.7813718411552346, + "grad_norm": 0.0241931714117527, + "learning_rate": 7.749890727396312e-07, + "loss": 0.7754, + "step": 1542 + }, + { + "epoch": 1.7825270758122742, + "grad_norm": 0.0242807324975729, + "learning_rate": 7.672202530892778e-07, + "loss": 0.7751, + "step": 1543 + }, + { + "epoch": 1.783682310469314, + "grad_norm": 0.024539347738027573, + "learning_rate": 7.594835300642188e-07, + "loss": 0.7955, + "step": 1544 + }, + { + "epoch": 1.7848375451263538, + "grad_norm": 0.024527210742235184, + "learning_rate": 7.517790469526456e-07, + "loss": 0.7713, + "step": 1545 + }, + { + "epoch": 1.7859927797833937, + "grad_norm": 0.024129964411258698, + "learning_rate": 7.441069464456497e-07, + "loss": 0.7888, + "step": 1546 + }, + { + "epoch": 1.7871480144404333, + "grad_norm": 0.024345900863409042, + "learning_rate": 7.364673706345822e-07, + "loss": 0.7548, + "step": 1547 + }, + { + "epoch": 1.7883032490974728, + "grad_norm": 0.024231646209955215, + "learning_rate": 7.288604610084172e-07, + "loss": 0.7645, + "step": 1548 + }, + { + "epoch": 1.7894584837545127, + "grad_norm": 0.023612959310412407, + "learning_rate": 7.212863584511341e-07, + "loss": 0.7565, + "step": 1549 + }, + { + "epoch": 1.7906137184115525, + "grad_norm": 0.02500954270362854, + "learning_rate": 7.137452032391084e-07, + "loss": 0.7646, + "step": 1550 + }, + { + "epoch": 1.791768953068592, + "grad_norm": 0.024092212319374084, + "learning_rate": 7.062371350385143e-07, + "loss": 0.7667, + "step": 1551 + }, + { + "epoch": 1.7929241877256317, + "grad_norm": 0.023889616131782532, + "learning_rate": 6.987622929027362e-07, + "loss": 0.7848, + "step": 1552 + }, + { + "epoch": 1.7940794223826715, + "grad_norm": 0.024128958582878113, + "learning_rate": 6.913208152697935e-07, + "loss": 0.7667, + "step": 1553 + }, + { + "epoch": 1.7952346570397113, + "grad_norm": 0.024358974769711494, + "learning_rate": 6.839128399597784e-07, + "loss": 0.7742, + "step": 1554 + }, + { + "epoch": 1.796389891696751, + "grad_norm": 0.024110984057188034, + "learning_rate": 6.765385041723024e-07, + "loss": 0.7783, + "step": 1555 + }, + { + "epoch": 1.7975451263537905, + "grad_norm": 0.02420348860323429, + "learning_rate": 6.691979444839561e-07, + "loss": 0.7737, + "step": 1556 + }, + { + "epoch": 1.7987003610108303, + "grad_norm": 0.024441605433821678, + "learning_rate": 6.618912968457771e-07, + "loss": 0.7767, + "step": 1557 + }, + { + "epoch": 1.7998555956678701, + "grad_norm": 0.024723242968320847, + "learning_rate": 6.546186965807347e-07, + "loss": 0.7607, + "step": 1558 + }, + { + "epoch": 1.8010108303249097, + "grad_norm": 0.025666577741503716, + "learning_rate": 6.473802783812225e-07, + "loss": 0.8028, + "step": 1559 + }, + { + "epoch": 1.8021660649819493, + "grad_norm": 0.024900216609239578, + "learning_rate": 6.401761763065661e-07, + "loss": 0.7452, + "step": 1560 + }, + { + "epoch": 1.8033212996389891, + "grad_norm": 0.024593856185674667, + "learning_rate": 6.330065237805361e-07, + "loss": 0.7879, + "step": 1561 + }, + { + "epoch": 1.804476534296029, + "grad_norm": 0.024126332253217697, + "learning_rate": 6.258714535888803e-07, + "loss": 0.7687, + "step": 1562 + }, + { + "epoch": 1.8056317689530685, + "grad_norm": 0.02449883334338665, + "learning_rate": 6.187710978768624e-07, + "loss": 0.7622, + "step": 1563 + }, + { + "epoch": 1.8067870036101084, + "grad_norm": 0.024287080392241478, + "learning_rate": 6.117055881468175e-07, + "loss": 0.7863, + "step": 1564 + }, + { + "epoch": 1.807942238267148, + "grad_norm": 0.024112451821565628, + "learning_rate": 6.046750552557139e-07, + "loss": 0.7887, + "step": 1565 + }, + { + "epoch": 1.8090974729241878, + "grad_norm": 0.02391749620437622, + "learning_rate": 5.976796294127286e-07, + "loss": 0.7687, + "step": 1566 + }, + { + "epoch": 1.8102527075812276, + "grad_norm": 0.023689934983849525, + "learning_rate": 5.907194401768411e-07, + "loss": 0.7705, + "step": 1567 + }, + { + "epoch": 1.8114079422382672, + "grad_norm": 0.025095578283071518, + "learning_rate": 5.837946164544276e-07, + "loss": 0.805, + "step": 1568 + }, + { + "epoch": 1.8125631768953068, + "grad_norm": 0.02462933026254177, + "learning_rate": 5.769052864968768e-07, + "loss": 0.7699, + "step": 1569 + }, + { + "epoch": 1.8137184115523466, + "grad_norm": 0.023868851363658905, + "learning_rate": 5.700515778982142e-07, + "loss": 0.7506, + "step": 1570 + }, + { + "epoch": 1.8148736462093864, + "grad_norm": 0.024339932948350906, + "learning_rate": 5.632336175927404e-07, + "loss": 0.7663, + "step": 1571 + }, + { + "epoch": 1.816028880866426, + "grad_norm": 0.023983048275113106, + "learning_rate": 5.564515318526759e-07, + "loss": 0.7555, + "step": 1572 + }, + { + "epoch": 1.8171841155234656, + "grad_norm": 0.024354344233870506, + "learning_rate": 5.497054462858296e-07, + "loss": 0.7856, + "step": 1573 + }, + { + "epoch": 1.8183393501805054, + "grad_norm": 0.02329828403890133, + "learning_rate": 5.429954858332623e-07, + "loss": 0.7755, + "step": 1574 + }, + { + "epoch": 1.8194945848375452, + "grad_norm": 0.024324821308255196, + "learning_rate": 5.363217747669843e-07, + "loss": 0.7871, + "step": 1575 + }, + { + "epoch": 1.8206498194945848, + "grad_norm": 0.02310972288250923, + "learning_rate": 5.296844366876441e-07, + "loss": 0.7499, + "step": 1576 + }, + { + "epoch": 1.8218050541516244, + "grad_norm": 0.02391042560338974, + "learning_rate": 5.230835945222463e-07, + "loss": 0.7941, + "step": 1577 + }, + { + "epoch": 1.8229602888086642, + "grad_norm": 0.024979131296277046, + "learning_rate": 5.165193705218699e-07, + "loss": 0.7719, + "step": 1578 + }, + { + "epoch": 1.824115523465704, + "grad_norm": 0.024278726428747177, + "learning_rate": 5.099918862594065e-07, + "loss": 0.7815, + "step": 1579 + }, + { + "epoch": 1.8252707581227436, + "grad_norm": 0.024197446182370186, + "learning_rate": 5.035012626273094e-07, + "loss": 0.7538, + "step": 1580 + }, + { + "epoch": 1.8264259927797832, + "grad_norm": 0.02464877814054489, + "learning_rate": 4.970476198353524e-07, + "loss": 0.764, + "step": 1581 + }, + { + "epoch": 1.827581227436823, + "grad_norm": 0.02464544028043747, + "learning_rate": 4.906310774084055e-07, + "loss": 0.787, + "step": 1582 + }, + { + "epoch": 1.8287364620938629, + "grad_norm": 0.023831835016608238, + "learning_rate": 4.842517541842186e-07, + "loss": 0.7527, + "step": 1583 + }, + { + "epoch": 1.8298916967509027, + "grad_norm": 0.02423412911593914, + "learning_rate": 4.779097683112254e-07, + "loss": 0.7639, + "step": 1584 + }, + { + "epoch": 1.8310469314079423, + "grad_norm": 0.023963099345564842, + "learning_rate": 4.7160523724634964e-07, + "loss": 0.7676, + "step": 1585 + }, + { + "epoch": 1.8322021660649819, + "grad_norm": 0.023802466690540314, + "learning_rate": 4.653382777528332e-07, + "loss": 0.7805, + "step": 1586 + }, + { + "epoch": 1.8333574007220217, + "grad_norm": 0.024113498628139496, + "learning_rate": 4.5910900589807164e-07, + "loss": 0.7578, + "step": 1587 + }, + { + "epoch": 1.8345126353790615, + "grad_norm": 0.024253182113170624, + "learning_rate": 4.5291753705146685e-07, + "loss": 0.7816, + "step": 1588 + }, + { + "epoch": 1.835667870036101, + "grad_norm": 0.024671630933880806, + "learning_rate": 4.4676398588229044e-07, + "loss": 0.7712, + "step": 1589 + }, + { + "epoch": 1.8368231046931407, + "grad_norm": 0.02365119196474552, + "learning_rate": 4.40648466357554e-07, + "loss": 0.7673, + "step": 1590 + }, + { + "epoch": 1.8379783393501805, + "grad_norm": 0.024656254798173904, + "learning_rate": 4.3457109173990736e-07, + "loss": 0.7627, + "step": 1591 + }, + { + "epoch": 1.8391335740072203, + "grad_norm": 0.02385905385017395, + "learning_rate": 4.285319745855341e-07, + "loss": 0.7758, + "step": 1592 + }, + { + "epoch": 1.84028880866426, + "grad_norm": 0.02395222708582878, + "learning_rate": 4.2253122674207165e-07, + "loss": 0.7763, + "step": 1593 + }, + { + "epoch": 1.8414440433212995, + "grad_norm": 0.02432720549404621, + "learning_rate": 4.165689593465352e-07, + "loss": 0.7873, + "step": 1594 + }, + { + "epoch": 1.8425992779783393, + "grad_norm": 0.02468527853488922, + "learning_rate": 4.1064528282326263e-07, + "loss": 0.7517, + "step": 1595 + }, + { + "epoch": 1.8437545126353792, + "grad_norm": 0.02376159094274044, + "learning_rate": 4.047603068818685e-07, + "loss": 0.7715, + "step": 1596 + }, + { + "epoch": 1.8449097472924187, + "grad_norm": 0.02394728921353817, + "learning_rate": 3.9891414051521345e-07, + "loss": 0.7697, + "step": 1597 + }, + { + "epoch": 1.8460649819494583, + "grad_norm": 0.024393731728196144, + "learning_rate": 3.93106891997382e-07, + "loss": 0.7651, + "step": 1598 + }, + { + "epoch": 1.8472202166064982, + "grad_norm": 0.02447488158941269, + "learning_rate": 3.8733866888168026e-07, + "loss": 0.7862, + "step": 1599 + }, + { + "epoch": 1.848375451263538, + "grad_norm": 0.023694733157753944, + "learning_rate": 3.816095779986455e-07, + "loss": 0.7704, + "step": 1600 + }, + { + "epoch": 1.8495306859205776, + "grad_norm": 0.023735985159873962, + "learning_rate": 3.759197254540624e-07, + "loss": 0.7712, + "step": 1601 + }, + { + "epoch": 1.8506859205776174, + "grad_norm": 0.02489444985985756, + "learning_rate": 3.702692166270027e-07, + "loss": 0.7918, + "step": 1602 + }, + { + "epoch": 1.851841155234657, + "grad_norm": 0.02421208843588829, + "learning_rate": 3.6465815616787e-07, + "loss": 0.7888, + "step": 1603 + }, + { + "epoch": 1.8529963898916968, + "grad_norm": 0.0240098275244236, + "learning_rate": 3.590866479964664e-07, + "loss": 0.747, + "step": 1604 + }, + { + "epoch": 1.8541516245487366, + "grad_norm": 0.025053462013602257, + "learning_rate": 3.535547953000618e-07, + "loss": 0.7872, + "step": 1605 + }, + { + "epoch": 1.8553068592057762, + "grad_norm": 0.023561321198940277, + "learning_rate": 3.4806270053148703e-07, + "loss": 0.7748, + "step": 1606 + }, + { + "epoch": 1.8564620938628158, + "grad_norm": 0.023926684632897377, + "learning_rate": 3.4261046540723424e-07, + "loss": 0.7576, + "step": 1607 + }, + { + "epoch": 1.8576173285198556, + "grad_norm": 0.023767825216054916, + "learning_rate": 3.371981909055759e-07, + "loss": 0.7577, + "step": 1608 + }, + { + "epoch": 1.8587725631768954, + "grad_norm": 0.0232445877045393, + "learning_rate": 3.318259772646901e-07, + "loss": 0.7442, + "step": 1609 + }, + { + "epoch": 1.859927797833935, + "grad_norm": 0.024164149537682533, + "learning_rate": 3.264939239808092e-07, + "loss": 0.7589, + "step": 1610 + }, + { + "epoch": 1.8610830324909746, + "grad_norm": 0.023966316133737564, + "learning_rate": 3.2120212980637286e-07, + "loss": 0.768, + "step": 1611 + }, + { + "epoch": 1.8622382671480144, + "grad_norm": 0.023847166448831558, + "learning_rate": 3.1595069274820076e-07, + "loss": 0.7603, + "step": 1612 + }, + { + "epoch": 1.8633935018050543, + "grad_norm": 0.02418961189687252, + "learning_rate": 3.1073971006567867e-07, + "loss": 0.7814, + "step": 1613 + }, + { + "epoch": 1.8645487364620938, + "grad_norm": 0.023753009736537933, + "learning_rate": 3.055692782689551e-07, + "loss": 0.7667, + "step": 1614 + }, + { + "epoch": 1.8657039711191334, + "grad_norm": 0.02413536049425602, + "learning_rate": 3.0043949311715463e-07, + "loss": 0.7826, + "step": 1615 + }, + { + "epoch": 1.8668592057761733, + "grad_norm": 0.024381062015891075, + "learning_rate": 2.953504496166049e-07, + "loss": 0.7678, + "step": 1616 + }, + { + "epoch": 1.868014440433213, + "grad_norm": 0.02448740415275097, + "learning_rate": 2.903022420190769e-07, + "loss": 0.7751, + "step": 1617 + }, + { + "epoch": 1.8691696750902527, + "grad_norm": 0.024086831137537956, + "learning_rate": 2.8529496382003895e-07, + "loss": 0.7799, + "step": 1618 + }, + { + "epoch": 1.8703249097472923, + "grad_norm": 0.02341514825820923, + "learning_rate": 2.8032870775692436e-07, + "loss": 0.7723, + "step": 1619 + }, + { + "epoch": 1.871480144404332, + "grad_norm": 0.023932676762342453, + "learning_rate": 2.75403565807417e-07, + "loss": 0.7635, + "step": 1620 + }, + { + "epoch": 1.872635379061372, + "grad_norm": 0.023613903671503067, + "learning_rate": 2.705196291877439e-07, + "loss": 0.7605, + "step": 1621 + }, + { + "epoch": 1.8737906137184117, + "grad_norm": 0.024287506937980652, + "learning_rate": 2.6567698835098765e-07, + "loss": 0.7815, + "step": 1622 + }, + { + "epoch": 1.8749458483754513, + "grad_norm": 0.023454533889889717, + "learning_rate": 2.608757329854116e-07, + "loss": 0.759, + "step": 1623 + }, + { + "epoch": 1.876101083032491, + "grad_norm": 0.024026375263929367, + "learning_rate": 2.561159520127993e-07, + "loss": 0.7847, + "step": 1624 + }, + { + "epoch": 1.8772563176895307, + "grad_norm": 0.024074561893939972, + "learning_rate": 2.5139773358680497e-07, + "loss": 0.7734, + "step": 1625 + }, + { + "epoch": 1.8784115523465705, + "grad_norm": 0.024127675220370293, + "learning_rate": 2.467211650913235e-07, + "loss": 0.7627, + "step": 1626 + }, + { + "epoch": 1.8795667870036101, + "grad_norm": 0.024527592584490776, + "learning_rate": 2.4208633313887026e-07, + "loss": 0.8107, + "step": 1627 + }, + { + "epoch": 1.8807220216606497, + "grad_norm": 0.02324732393026352, + "learning_rate": 2.3749332356897935e-07, + "loss": 0.7739, + "step": 1628 + }, + { + "epoch": 1.8818772563176895, + "grad_norm": 0.02475915476679802, + "learning_rate": 2.3294222144661086e-07, + "loss": 0.778, + "step": 1629 + }, + { + "epoch": 1.8830324909747294, + "grad_norm": 0.024105004966259003, + "learning_rate": 2.2843311106057757e-07, + "loss": 0.7772, + "step": 1630 + }, + { + "epoch": 1.884187725631769, + "grad_norm": 0.02386721596121788, + "learning_rate": 2.2396607592198304e-07, + "loss": 0.7631, + "step": 1631 + }, + { + "epoch": 1.8853429602888085, + "grad_norm": 0.024259351193904877, + "learning_rate": 2.1954119876267445e-07, + "loss": 0.7966, + "step": 1632 + }, + { + "epoch": 1.8864981949458484, + "grad_norm": 0.023345720022916794, + "learning_rate": 2.1515856153371186e-07, + "loss": 0.7685, + "step": 1633 + }, + { + "epoch": 1.8876534296028882, + "grad_norm": 0.024387696757912636, + "learning_rate": 2.108182454038492e-07, + "loss": 0.7807, + "step": 1634 + }, + { + "epoch": 1.8888086642599278, + "grad_norm": 0.024348227307200432, + "learning_rate": 2.065203307580309e-07, + "loss": 0.7761, + "step": 1635 + }, + { + "epoch": 1.8899638989169674, + "grad_norm": 0.023634381592273712, + "learning_rate": 2.0226489719590362e-07, + "loss": 0.7619, + "step": 1636 + }, + { + "epoch": 1.8911191335740072, + "grad_norm": 0.023854384198784828, + "learning_rate": 1.9805202353034296e-07, + "loss": 0.7587, + "step": 1637 + }, + { + "epoch": 1.892274368231047, + "grad_norm": 0.023444773629307747, + "learning_rate": 1.9388178778599145e-07, + "loss": 0.7739, + "step": 1638 + }, + { + "epoch": 1.8934296028880866, + "grad_norm": 0.023775937035679817, + "learning_rate": 1.8975426719781569e-07, + "loss": 0.7671, + "step": 1639 + }, + { + "epoch": 1.8945848375451262, + "grad_norm": 0.023820441216230392, + "learning_rate": 1.85669538209674e-07, + "loss": 0.7873, + "step": 1640 + }, + { + "epoch": 1.895740072202166, + "grad_norm": 0.023935571312904358, + "learning_rate": 1.816276764729035e-07, + "loss": 0.7698, + "step": 1641 + }, + { + "epoch": 1.8968953068592058, + "grad_norm": 0.02413167431950569, + "learning_rate": 1.7762875684491536e-07, + "loss": 0.7851, + "step": 1642 + }, + { + "epoch": 1.8980505415162456, + "grad_norm": 0.02435237169265747, + "learning_rate": 1.7367285338781114e-07, + "loss": 0.7699, + "step": 1643 + }, + { + "epoch": 1.8992057761732852, + "grad_norm": 0.024776320904493332, + "learning_rate": 1.6976003936701042e-07, + "loss": 0.7768, + "step": 1644 + }, + { + "epoch": 1.9003610108303248, + "grad_norm": 0.02364964410662651, + "learning_rate": 1.6589038724989278e-07, + "loss": 0.7931, + "step": 1645 + }, + { + "epoch": 1.9015162454873646, + "grad_norm": 0.023821519687771797, + "learning_rate": 1.6206396870445796e-07, + "loss": 0.7533, + "step": 1646 + }, + { + "epoch": 1.9026714801444045, + "grad_norm": 0.02380574308335781, + "learning_rate": 1.582808545979954e-07, + "loss": 0.7875, + "step": 1647 + }, + { + "epoch": 1.903826714801444, + "grad_norm": 0.024249248206615448, + "learning_rate": 1.5454111499577497e-07, + "loss": 0.7561, + "step": 1648 + }, + { + "epoch": 1.9049819494584836, + "grad_norm": 0.02454947866499424, + "learning_rate": 1.5084481915974666e-07, + "loss": 0.7695, + "step": 1649 + }, + { + "epoch": 1.9061371841155235, + "grad_norm": 0.023830397054553032, + "learning_rate": 1.471920355472605e-07, + "loss": 0.7765, + "step": 1650 + }, + { + "epoch": 1.9072924187725633, + "grad_norm": 0.023539626970887184, + "learning_rate": 1.4358283180979607e-07, + "loss": 0.8193, + "step": 1651 + }, + { + "epoch": 1.9084476534296029, + "grad_norm": 0.02411411888897419, + "learning_rate": 1.4001727479171022e-07, + "loss": 0.8001, + "step": 1652 + }, + { + "epoch": 1.9096028880866425, + "grad_norm": 0.024499407038092613, + "learning_rate": 1.3649543052900035e-07, + "loss": 0.7714, + "step": 1653 + }, + { + "epoch": 1.9107581227436823, + "grad_norm": 0.023902500048279762, + "learning_rate": 1.3301736424808105e-07, + "loss": 0.7906, + "step": 1654 + }, + { + "epoch": 1.911913357400722, + "grad_norm": 0.024184612557291985, + "learning_rate": 1.295831403645745e-07, + "loss": 0.7643, + "step": 1655 + }, + { + "epoch": 1.9130685920577617, + "grad_norm": 0.024717319756746292, + "learning_rate": 1.2619282248211896e-07, + "loss": 0.8194, + "step": 1656 + }, + { + "epoch": 1.9142238267148013, + "grad_norm": 0.023884067311882973, + "learning_rate": 1.2284647339118997e-07, + "loss": 0.753, + "step": 1657 + }, + { + "epoch": 1.915379061371841, + "grad_norm": 0.023947538807988167, + "learning_rate": 1.1954415506793914e-07, + "loss": 0.7659, + "step": 1658 + }, + { + "epoch": 1.916534296028881, + "grad_norm": 0.02345140650868416, + "learning_rate": 1.1628592867304393e-07, + "loss": 0.7673, + "step": 1659 + }, + { + "epoch": 1.9176895306859207, + "grad_norm": 0.02360740303993225, + "learning_rate": 1.1307185455057616e-07, + "loss": 0.7703, + "step": 1660 + }, + { + "epoch": 1.9188447653429603, + "grad_norm": 0.023865753784775734, + "learning_rate": 1.0990199222688366e-07, + "loss": 0.7839, + "step": 1661 + }, + { + "epoch": 1.92, + "grad_norm": 0.02310926280915737, + "learning_rate": 1.0677640040949089e-07, + "loss": 0.7723, + "step": 1662 + }, + { + "epoch": 1.9211552346570397, + "grad_norm": 0.023065300658345222, + "learning_rate": 1.0369513698600607e-07, + "loss": 0.7699, + "step": 1663 + }, + { + "epoch": 1.9223104693140796, + "grad_norm": 0.023747362196445465, + "learning_rate": 1.0065825902305331e-07, + "loss": 0.7736, + "step": 1664 + }, + { + "epoch": 1.9234657039711192, + "grad_norm": 0.02400401420891285, + "learning_rate": 9.766582276521622e-08, + "loss": 0.7661, + "step": 1665 + }, + { + "epoch": 1.9246209386281588, + "grad_norm": 0.02401837147772312, + "learning_rate": 9.471788363399199e-08, + "loss": 0.7723, + "step": 1666 + }, + { + "epoch": 1.9257761732851986, + "grad_norm": 0.024011146277189255, + "learning_rate": 9.181449622676956e-08, + "loss": 0.7702, + "step": 1667 + }, + { + "epoch": 1.9269314079422384, + "grad_norm": 0.023694701492786407, + "learning_rate": 8.8955714315814e-08, + "loss": 0.7538, + "step": 1668 + }, + { + "epoch": 1.928086642599278, + "grad_norm": 0.02426433376967907, + "learning_rate": 8.614159084727651e-08, + "loss": 0.7792, + "step": 1669 + }, + { + "epoch": 1.9292418772563176, + "grad_norm": 0.024202091619372368, + "learning_rate": 8.337217794020758e-08, + "loss": 0.7423, + "step": 1670 + }, + { + "epoch": 1.9303971119133574, + "grad_norm": 0.024061063304543495, + "learning_rate": 8.064752688559553e-08, + "loss": 0.7549, + "step": 1671 + }, + { + "epoch": 1.9315523465703972, + "grad_norm": 0.024105606600642204, + "learning_rate": 7.796768814541562e-08, + "loss": 0.7845, + "step": 1672 + }, + { + "epoch": 1.9327075812274368, + "grad_norm": 0.024117371067404747, + "learning_rate": 7.533271135169531e-08, + "loss": 0.7706, + "step": 1673 + }, + { + "epoch": 1.9338628158844764, + "grad_norm": 0.023541245609521866, + "learning_rate": 7.274264530559488e-08, + "loss": 0.7794, + "step": 1674 + }, + { + "epoch": 1.9350180505415162, + "grad_norm": 0.023918094113469124, + "learning_rate": 7.019753797650377e-08, + "loss": 0.7515, + "step": 1675 + }, + { + "epoch": 1.936173285198556, + "grad_norm": 0.02431383728981018, + "learning_rate": 6.769743650115355e-08, + "loss": 0.7852, + "step": 1676 + }, + { + "epoch": 1.9373285198555956, + "grad_norm": 0.022908180952072144, + "learning_rate": 6.524238718274186e-08, + "loss": 0.7451, + "step": 1677 + }, + { + "epoch": 1.9384837545126352, + "grad_norm": 0.023673677816987038, + "learning_rate": 6.283243549007756e-08, + "loss": 0.7679, + "step": 1678 + }, + { + "epoch": 1.939638989169675, + "grad_norm": 0.02336922660470009, + "learning_rate": 6.046762605673734e-08, + "loss": 0.7619, + "step": 1679 + }, + { + "epoch": 1.9407942238267148, + "grad_norm": 0.023723188787698746, + "learning_rate": 5.814800268024005e-08, + "loss": 0.7792, + "step": 1680 + }, + { + "epoch": 1.9419494584837547, + "grad_norm": 0.02384623885154724, + "learning_rate": 5.58736083212344e-08, + "loss": 0.7648, + "step": 1681 + }, + { + "epoch": 1.9431046931407943, + "grad_norm": 0.024481408298015594, + "learning_rate": 5.364448510270387e-08, + "loss": 0.7907, + "step": 1682 + }, + { + "epoch": 1.9442599277978339, + "grad_norm": 0.023579951375722885, + "learning_rate": 5.146067430918623e-08, + "loss": 0.7737, + "step": 1683 + }, + { + "epoch": 1.9454151624548737, + "grad_norm": 0.02421787939965725, + "learning_rate": 4.932221638601039e-08, + "loss": 0.7746, + "step": 1684 + }, + { + "epoch": 1.9465703971119135, + "grad_norm": 0.02382575534284115, + "learning_rate": 4.722915093854457e-08, + "loss": 0.7547, + "step": 1685 + }, + { + "epoch": 1.947725631768953, + "grad_norm": 0.024289904162287712, + "learning_rate": 4.518151673146557e-08, + "loss": 0.7786, + "step": 1686 + }, + { + "epoch": 1.9488808664259927, + "grad_norm": 0.0236971415579319, + "learning_rate": 4.3179351688039386e-08, + "loss": 0.7785, + "step": 1687 + }, + { + "epoch": 1.9500361010830325, + "grad_norm": 0.023959195241332054, + "learning_rate": 4.122269288941911e-08, + "loss": 0.7665, + "step": 1688 + }, + { + "epoch": 1.9511913357400723, + "grad_norm": 0.02382025495171547, + "learning_rate": 3.931157657395747e-08, + "loss": 0.7539, + "step": 1689 + }, + { + "epoch": 1.952346570397112, + "grad_norm": 0.024375753477215767, + "learning_rate": 3.744603813653741e-08, + "loss": 0.7962, + "step": 1690 + }, + { + "epoch": 1.9535018050541515, + "grad_norm": 0.023500755429267883, + "learning_rate": 3.562611212791524e-08, + "loss": 0.7373, + "step": 1691 + }, + { + "epoch": 1.9546570397111913, + "grad_norm": 0.024235745891928673, + "learning_rate": 3.3851832254080314e-08, + "loss": 0.7971, + "step": 1692 + }, + { + "epoch": 1.9558122743682311, + "grad_norm": 0.02372434362769127, + "learning_rate": 3.212323137563259e-08, + "loss": 0.746, + "step": 1693 + }, + { + "epoch": 1.9569675090252707, + "grad_norm": 0.02388044260442257, + "learning_rate": 3.044034150717202e-08, + "loss": 0.7903, + "step": 1694 + }, + { + "epoch": 1.9581227436823103, + "grad_norm": 0.02452683448791504, + "learning_rate": 2.880319381670718e-08, + "loss": 0.7752, + "step": 1695 + }, + { + "epoch": 1.9592779783393501, + "grad_norm": 0.023390717804431915, + "learning_rate": 2.721181862507687e-08, + "loss": 0.7453, + "step": 1696 + }, + { + "epoch": 1.96043321299639, + "grad_norm": 0.023405345156788826, + "learning_rate": 2.566624540538925e-08, + "loss": 0.7601, + "step": 1697 + }, + { + "epoch": 1.9615884476534298, + "grad_norm": 0.023698341101408005, + "learning_rate": 2.4166502782475928e-08, + "loss": 0.7577, + "step": 1698 + }, + { + "epoch": 1.9627436823104694, + "grad_norm": 0.024021916091442108, + "learning_rate": 2.2712618532361902e-08, + "loss": 0.7723, + "step": 1699 + }, + { + "epoch": 1.963898916967509, + "grad_norm": 0.024296529591083527, + "learning_rate": 2.130461958174988e-08, + "loss": 0.7896, + "step": 1700 + }, + { + "epoch": 1.9650541516245488, + "grad_norm": 0.02386784926056862, + "learning_rate": 1.9942532007523544e-08, + "loss": 0.7587, + "step": 1701 + }, + { + "epoch": 1.9662093862815886, + "grad_norm": 0.023428186774253845, + "learning_rate": 1.8626381036264058e-08, + "loss": 0.7714, + "step": 1702 + }, + { + "epoch": 1.9673646209386282, + "grad_norm": 0.024159930646419525, + "learning_rate": 1.735619104378108e-08, + "loss": 0.7937, + "step": 1703 + }, + { + "epoch": 1.9685198555956678, + "grad_norm": 0.024223441258072853, + "learning_rate": 1.6131985554663975e-08, + "loss": 0.761, + "step": 1704 + }, + { + "epoch": 1.9696750902527076, + "grad_norm": 0.02402389794588089, + "learning_rate": 1.4953787241843607e-08, + "loss": 0.7742, + "step": 1705 + }, + { + "epoch": 1.9708303249097474, + "grad_norm": 0.023779474198818207, + "learning_rate": 1.382161792617522e-08, + "loss": 0.7489, + "step": 1706 + }, + { + "epoch": 1.971985559566787, + "grad_norm": 0.02408597618341446, + "learning_rate": 1.273549857603129e-08, + "loss": 0.7927, + "step": 1707 + }, + { + "epoch": 1.9731407942238266, + "grad_norm": 0.023695794865489006, + "learning_rate": 1.16954493069155e-08, + "loss": 0.7603, + "step": 1708 + }, + { + "epoch": 1.9742960288808664, + "grad_norm": 0.025053711608052254, + "learning_rate": 1.0701489381089175e-08, + "loss": 0.7886, + "step": 1709 + }, + { + "epoch": 1.9754512635379062, + "grad_norm": 0.023919204249978065, + "learning_rate": 9.753637207214111e-09, + "loss": 0.7837, + "step": 1710 + }, + { + "epoch": 1.9766064981949458, + "grad_norm": 0.023150555789470673, + "learning_rate": 8.85191034001343e-09, + "loss": 0.7449, + "step": 1711 + }, + { + "epoch": 1.9777617328519854, + "grad_norm": 0.023474812507629395, + "learning_rate": 7.996325479943542e-09, + "loss": 0.7825, + "step": 1712 + }, + { + "epoch": 1.9789169675090252, + "grad_norm": 0.02393323741853237, + "learning_rate": 7.186898472888319e-09, + "loss": 0.7702, + "step": 1713 + }, + { + "epoch": 1.980072202166065, + "grad_norm": 0.02350715547800064, + "learning_rate": 6.423644309862407e-09, + "loss": 0.7662, + "step": 1714 + }, + { + "epoch": 1.9812274368231046, + "grad_norm": 0.023993050679564476, + "learning_rate": 5.706577126735919e-09, + "loss": 0.794, + "step": 1715 + }, + { + "epoch": 1.9823826714801442, + "grad_norm": 0.024228619411587715, + "learning_rate": 5.035710203971343e-09, + "loss": 0.7874, + "step": 1716 + }, + { + "epoch": 1.983537906137184, + "grad_norm": 0.024353953078389168, + "learning_rate": 4.411055966378209e-09, + "loss": 0.7666, + "step": 1717 + }, + { + "epoch": 1.9846931407942239, + "grad_norm": 0.023848267272114754, + "learning_rate": 3.832625982882452e-09, + "loss": 0.7824, + "step": 1718 + }, + { + "epoch": 1.9858483754512637, + "grad_norm": 0.02303747460246086, + "learning_rate": 3.300430966312174e-09, + "loss": 0.7451, + "step": 1719 + }, + { + "epoch": 1.9870036101083033, + "grad_norm": 0.02363884262740612, + "learning_rate": 2.8144807732005904e-09, + "loss": 0.7673, + "step": 1720 + }, + { + "epoch": 1.9881588447653429, + "grad_norm": 0.024231605231761932, + "learning_rate": 2.374784403601199e-09, + "loss": 0.7823, + "step": 1721 + }, + { + "epoch": 1.9893140794223827, + "grad_norm": 0.02352130599319935, + "learning_rate": 1.9813500009229292e-09, + "loss": 0.7676, + "step": 1722 + }, + { + "epoch": 1.9904693140794225, + "grad_norm": 0.02459130994975567, + "learning_rate": 1.634184851778611e-09, + "loss": 0.7937, + "step": 1723 + }, + { + "epoch": 1.991624548736462, + "grad_norm": 0.023892242461442947, + "learning_rate": 1.3332953858495395e-09, + "loss": 0.7666, + "step": 1724 + }, + { + "epoch": 1.9927797833935017, + "grad_norm": 0.0234362855553627, + "learning_rate": 1.0786871757678059e-09, + "loss": 0.7796, + "step": 1725 + }, + { + "epoch": 1.9939350180505415, + "grad_norm": 0.024267012253403664, + "learning_rate": 8.703649370116684e-10, + "loss": 0.738, + "step": 1726 + }, + { + "epoch": 1.9950902527075813, + "grad_norm": 0.023687295615673065, + "learning_rate": 7.083325278189627e-10, + "loss": 0.7633, + "step": 1727 + }, + { + "epoch": 1.996245487364621, + "grad_norm": 0.02364824153482914, + "learning_rate": 5.925929491152231e-10, + "loss": 0.769, + "step": 1728 + }, + { + "epoch": 1.9974007220216605, + "grad_norm": 0.024108268320560455, + "learning_rate": 5.231483444587321e-10, + "loss": 0.7676, + "step": 1729 + }, + { + "epoch": 1.9985559566787003, + "grad_norm": 0.025233233347535133, + "learning_rate": 5e-10, + "loss": 0.7958, + "step": 1730 + }, + { + "epoch": 1.9985559566787003, + "step": 1730, + "total_flos": 2.356928687040889e+19, + "train_loss": 0.8199663722446199, + "train_runtime": 64534.4414, + "train_samples_per_second": 10.301, + "train_steps_per_second": 0.027 + } + ], + "logging_steps": 1.0, + "max_steps": 1730, + "num_input_tokens_seen": 0, + "num_train_epochs": 2, + "save_steps": 250, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 2.356928687040889e+19, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}