{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 6250, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0008, "grad_norm": 0.7526811361312866, "learning_rate": 4.999992104320636e-05, "loss": 1.0428, "num_input_tokens_seen": 26624, "step": 5 }, { "epoch": 0.0016, "grad_norm": 0.5957724452018738, "learning_rate": 4.999968417332415e-05, "loss": 1.1061, "num_input_tokens_seen": 51408, "step": 10 }, { "epoch": 0.0024, "grad_norm": 0.8826403021812439, "learning_rate": 4.999928939184958e-05, "loss": 1.0426, "num_input_tokens_seen": 75040, "step": 15 }, { "epoch": 0.0032, "grad_norm": 0.6897421479225159, "learning_rate": 4.9998736701276295e-05, "loss": 1.1472, "num_input_tokens_seen": 98848, "step": 20 }, { "epoch": 0.004, "grad_norm": 0.7432862520217896, "learning_rate": 4.9998026105095405e-05, "loss": 0.9781, "num_input_tokens_seen": 123104, "step": 25 }, { "epoch": 0.0048, "grad_norm": 0.9830231666564941, "learning_rate": 4.999715760779541e-05, "loss": 0.9889, "num_input_tokens_seen": 152144, "step": 30 }, { "epoch": 0.0056, "grad_norm": 0.9177331924438477, "learning_rate": 4.999613121486222e-05, "loss": 0.9345, "num_input_tokens_seen": 177216, "step": 35 }, { "epoch": 0.0064, "grad_norm": 0.6646199822425842, "learning_rate": 4.999494693277907e-05, "loss": 0.8539, "num_input_tokens_seen": 203152, "step": 40 }, { "epoch": 0.0072, "grad_norm": 0.5822590589523315, "learning_rate": 4.999360476902656e-05, "loss": 0.9183, "num_input_tokens_seen": 233568, "step": 45 }, { "epoch": 0.008, "grad_norm": 0.7686595916748047, "learning_rate": 4.99921047320825e-05, "loss": 0.873, "num_input_tokens_seen": 264752, "step": 50 }, { "epoch": 0.0088, "grad_norm": 0.729837954044342, "learning_rate": 4.9990446831421955e-05, "loss": 0.8676, "num_input_tokens_seen": 291040, "step": 55 }, { "epoch": 0.0096, "grad_norm": 0.9523835778236389, "learning_rate": 4.998863107751711e-05, "loss": 0.9004, "num_input_tokens_seen": 321760, "step": 60 }, { "epoch": 0.0104, "grad_norm": 0.6720367670059204, "learning_rate": 4.9986657481837277e-05, "loss": 0.8536, "num_input_tokens_seen": 347168, "step": 65 }, { "epoch": 0.0112, "grad_norm": 0.4336840808391571, "learning_rate": 4.998452605684874e-05, "loss": 0.8027, "num_input_tokens_seen": 373888, "step": 70 }, { "epoch": 0.012, "grad_norm": 0.808559238910675, "learning_rate": 4.998223681601473e-05, "loss": 0.8075, "num_input_tokens_seen": 398752, "step": 75 }, { "epoch": 0.0128, "grad_norm": 0.5663979053497314, "learning_rate": 4.997978977379536e-05, "loss": 0.7919, "num_input_tokens_seen": 421344, "step": 80 }, { "epoch": 0.0136, "grad_norm": 0.5677878260612488, "learning_rate": 4.9977184945647473e-05, "loss": 0.7512, "num_input_tokens_seen": 451296, "step": 85 }, { "epoch": 0.0144, "grad_norm": 0.674132227897644, "learning_rate": 4.997442234802456e-05, "loss": 0.7713, "num_input_tokens_seen": 482416, "step": 90 }, { "epoch": 0.0152, "grad_norm": 0.5088427662849426, "learning_rate": 4.997150199837671e-05, "loss": 0.7965, "num_input_tokens_seen": 513008, "step": 95 }, { "epoch": 0.016, "grad_norm": 0.6657032370567322, "learning_rate": 4.996842391515044e-05, "loss": 0.8623, "num_input_tokens_seen": 537984, "step": 100 }, { "epoch": 0.0168, "grad_norm": 0.6862130761146545, "learning_rate": 4.996518811778858e-05, "loss": 0.7797, "num_input_tokens_seen": 564528, "step": 105 }, { "epoch": 0.0176, "grad_norm": 0.6449868083000183, "learning_rate": 4.99617946267302e-05, "loss": 0.7732, "num_input_tokens_seen": 588608, "step": 110 }, { "epoch": 0.0184, "grad_norm": 0.5512914657592773, "learning_rate": 4.9958243463410414e-05, "loss": 0.7478, "num_input_tokens_seen": 620752, "step": 115 }, { "epoch": 0.0192, "grad_norm": 0.7411808371543884, "learning_rate": 4.995453465026032e-05, "loss": 0.7194, "num_input_tokens_seen": 649200, "step": 120 }, { "epoch": 0.02, "grad_norm": 0.9926447868347168, "learning_rate": 4.995066821070679e-05, "loss": 0.7506, "num_input_tokens_seen": 679200, "step": 125 }, { "epoch": 0.0208, "grad_norm": 0.7455246448516846, "learning_rate": 4.9946644169172355e-05, "loss": 0.6886, "num_input_tokens_seen": 702144, "step": 130 }, { "epoch": 0.0216, "grad_norm": 0.5429201126098633, "learning_rate": 4.9942462551075056e-05, "loss": 0.8481, "num_input_tokens_seen": 730128, "step": 135 }, { "epoch": 0.0224, "grad_norm": 0.49708107113838196, "learning_rate": 4.993812338282826e-05, "loss": 0.7999, "num_input_tokens_seen": 757248, "step": 140 }, { "epoch": 0.0232, "grad_norm": 0.6150819063186646, "learning_rate": 4.993362669184051e-05, "loss": 0.7877, "num_input_tokens_seen": 786096, "step": 145 }, { "epoch": 0.024, "grad_norm": 0.6751016974449158, "learning_rate": 4.992897250651535e-05, "loss": 0.9312, "num_input_tokens_seen": 814192, "step": 150 }, { "epoch": 0.0248, "grad_norm": 0.6245042085647583, "learning_rate": 4.992416085625115e-05, "loss": 0.8767, "num_input_tokens_seen": 840144, "step": 155 }, { "epoch": 0.0256, "grad_norm": 0.5385093688964844, "learning_rate": 4.9919191771440905e-05, "loss": 0.8646, "num_input_tokens_seen": 870368, "step": 160 }, { "epoch": 0.0264, "grad_norm": 0.5674800276756287, "learning_rate": 4.991406528347206e-05, "loss": 0.7159, "num_input_tokens_seen": 897296, "step": 165 }, { "epoch": 0.0272, "grad_norm": 0.3683065176010132, "learning_rate": 4.990878142472628e-05, "loss": 0.7573, "num_input_tokens_seen": 924576, "step": 170 }, { "epoch": 0.028, "grad_norm": 0.5186157822608948, "learning_rate": 4.990334022857932e-05, "loss": 0.8083, "num_input_tokens_seen": 945856, "step": 175 }, { "epoch": 0.0288, "grad_norm": 0.5749205350875854, "learning_rate": 4.9897741729400705e-05, "loss": 0.7074, "num_input_tokens_seen": 970416, "step": 180 }, { "epoch": 0.0296, "grad_norm": 0.5518808960914612, "learning_rate": 4.9891985962553606e-05, "loss": 0.7709, "num_input_tokens_seen": 994288, "step": 185 }, { "epoch": 0.0304, "grad_norm": 0.5436009764671326, "learning_rate": 4.988607296439458e-05, "loss": 0.7, "num_input_tokens_seen": 1029872, "step": 190 }, { "epoch": 0.0312, "grad_norm": 0.4599573612213135, "learning_rate": 4.988000277227334e-05, "loss": 0.8251, "num_input_tokens_seen": 1061072, "step": 195 }, { "epoch": 0.032, "grad_norm": 0.4540678858757019, "learning_rate": 4.987377542453251e-05, "loss": 0.6707, "num_input_tokens_seen": 1089312, "step": 200 }, { "epoch": 0.0328, "grad_norm": 0.5874660611152649, "learning_rate": 4.98673909605074e-05, "loss": 0.7264, "num_input_tokens_seen": 1114272, "step": 205 }, { "epoch": 0.0336, "grad_norm": 0.6792047619819641, "learning_rate": 4.9860849420525766e-05, "loss": 0.7906, "num_input_tokens_seen": 1139808, "step": 210 }, { "epoch": 0.0344, "grad_norm": 0.5740874409675598, "learning_rate": 4.985415084590752e-05, "loss": 0.8062, "num_input_tokens_seen": 1163072, "step": 215 }, { "epoch": 0.0352, "grad_norm": 0.5089894533157349, "learning_rate": 4.9847295278964514e-05, "loss": 0.7432, "num_input_tokens_seen": 1193936, "step": 220 }, { "epoch": 0.036, "grad_norm": 0.7231270670890808, "learning_rate": 4.984028276300021e-05, "loss": 0.7586, "num_input_tokens_seen": 1219696, "step": 225 }, { "epoch": 0.0368, "grad_norm": 0.6494696140289307, "learning_rate": 4.98331133423095e-05, "loss": 0.7532, "num_input_tokens_seen": 1248096, "step": 230 }, { "epoch": 0.0376, "grad_norm": 0.6063010692596436, "learning_rate": 4.9825787062178315e-05, "loss": 0.786, "num_input_tokens_seen": 1276624, "step": 235 }, { "epoch": 0.0384, "grad_norm": 0.8775933384895325, "learning_rate": 4.981830396888344e-05, "loss": 0.7947, "num_input_tokens_seen": 1303472, "step": 240 }, { "epoch": 0.0392, "grad_norm": 0.7558068633079529, "learning_rate": 4.981066410969215e-05, "loss": 0.6988, "num_input_tokens_seen": 1326816, "step": 245 }, { "epoch": 0.04, "grad_norm": 0.8880596160888672, "learning_rate": 4.980286753286195e-05, "loss": 0.7078, "num_input_tokens_seen": 1351008, "step": 250 }, { "epoch": 0.0408, "grad_norm": 0.661428689956665, "learning_rate": 4.979491428764026e-05, "loss": 0.7491, "num_input_tokens_seen": 1374656, "step": 255 }, { "epoch": 0.0416, "grad_norm": 0.7624719738960266, "learning_rate": 4.9786804424264085e-05, "loss": 0.75, "num_input_tokens_seen": 1399264, "step": 260 }, { "epoch": 0.0424, "grad_norm": 0.6995192170143127, "learning_rate": 4.977853799395976e-05, "loss": 0.798, "num_input_tokens_seen": 1422304, "step": 265 }, { "epoch": 0.0432, "grad_norm": 0.5227561593055725, "learning_rate": 4.977011504894252e-05, "loss": 0.8814, "num_input_tokens_seen": 1447184, "step": 270 }, { "epoch": 0.044, "grad_norm": 0.7046292424201965, "learning_rate": 4.976153564241628e-05, "loss": 0.7203, "num_input_tokens_seen": 1474304, "step": 275 }, { "epoch": 0.0448, "grad_norm": 0.7567644119262695, "learning_rate": 4.975279982857324e-05, "loss": 0.6936, "num_input_tokens_seen": 1500896, "step": 280 }, { "epoch": 0.0456, "grad_norm": 0.6787880063056946, "learning_rate": 4.9743907662593524e-05, "loss": 0.7872, "num_input_tokens_seen": 1528688, "step": 285 }, { "epoch": 0.0464, "grad_norm": 0.5113949775695801, "learning_rate": 4.9734859200644905e-05, "loss": 0.7517, "num_input_tokens_seen": 1561328, "step": 290 }, { "epoch": 0.0472, "grad_norm": 0.7206217050552368, "learning_rate": 4.972565449988239e-05, "loss": 0.6726, "num_input_tokens_seen": 1589088, "step": 295 }, { "epoch": 0.048, "grad_norm": 0.602922797203064, "learning_rate": 4.971629361844785e-05, "loss": 0.7259, "num_input_tokens_seen": 1615712, "step": 300 }, { "epoch": 0.0488, "grad_norm": 0.7673738598823547, "learning_rate": 4.9706776615469716e-05, "loss": 0.8337, "num_input_tokens_seen": 1638640, "step": 305 }, { "epoch": 0.0496, "grad_norm": 0.7302682995796204, "learning_rate": 4.9697103551062556e-05, "loss": 0.731, "num_input_tokens_seen": 1664304, "step": 310 }, { "epoch": 0.0504, "grad_norm": 0.45416679978370667, "learning_rate": 4.968727448632669e-05, "loss": 0.7285, "num_input_tokens_seen": 1697648, "step": 315 }, { "epoch": 0.0512, "grad_norm": 0.5968911051750183, "learning_rate": 4.967728948334784e-05, "loss": 0.723, "num_input_tokens_seen": 1726608, "step": 320 }, { "epoch": 0.052, "grad_norm": 0.6134063601493835, "learning_rate": 4.96671486051967e-05, "loss": 0.7918, "num_input_tokens_seen": 1750912, "step": 325 }, { "epoch": 0.0528, "grad_norm": 0.5388225317001343, "learning_rate": 4.965685191592859e-05, "loss": 0.6448, "num_input_tokens_seen": 1782912, "step": 330 }, { "epoch": 0.0536, "grad_norm": 0.6615162491798401, "learning_rate": 4.964639948058297e-05, "loss": 0.7874, "num_input_tokens_seen": 1804704, "step": 335 }, { "epoch": 0.0544, "grad_norm": 0.8656606078147888, "learning_rate": 4.963579136518312e-05, "loss": 0.7025, "num_input_tokens_seen": 1827248, "step": 340 }, { "epoch": 0.0552, "grad_norm": 0.7784980535507202, "learning_rate": 4.962502763673565e-05, "loss": 0.6676, "num_input_tokens_seen": 1854304, "step": 345 }, { "epoch": 0.056, "grad_norm": 0.847607433795929, "learning_rate": 4.9614108363230135e-05, "loss": 0.7774, "num_input_tokens_seen": 1878768, "step": 350 }, { "epoch": 0.0568, "grad_norm": 0.5412896275520325, "learning_rate": 4.9603033613638626e-05, "loss": 0.7641, "num_input_tokens_seen": 1905744, "step": 355 }, { "epoch": 0.0576, "grad_norm": 0.5192331671714783, "learning_rate": 4.959180345791528e-05, "loss": 0.7169, "num_input_tokens_seen": 1931392, "step": 360 }, { "epoch": 0.0584, "grad_norm": 0.7992143630981445, "learning_rate": 4.958041796699583e-05, "loss": 0.7033, "num_input_tokens_seen": 1954304, "step": 365 }, { "epoch": 0.0592, "grad_norm": 0.49692437052726746, "learning_rate": 4.956887721279726e-05, "loss": 0.6569, "num_input_tokens_seen": 1987264, "step": 370 }, { "epoch": 0.06, "grad_norm": 0.7032391428947449, "learning_rate": 4.9557181268217227e-05, "loss": 0.7809, "num_input_tokens_seen": 2010160, "step": 375 }, { "epoch": 0.0608, "grad_norm": 0.780989944934845, "learning_rate": 4.9545330207133664e-05, "loss": 0.811, "num_input_tokens_seen": 2038880, "step": 380 }, { "epoch": 0.0616, "grad_norm": 0.819433867931366, "learning_rate": 4.953332410440435e-05, "loss": 0.825, "num_input_tokens_seen": 2065344, "step": 385 }, { "epoch": 0.0624, "grad_norm": 0.7076752781867981, "learning_rate": 4.952116303586631e-05, "loss": 0.7479, "num_input_tokens_seen": 2092064, "step": 390 }, { "epoch": 0.0632, "grad_norm": 0.6264218688011169, "learning_rate": 4.9508847078335495e-05, "loss": 0.7246, "num_input_tokens_seen": 2119360, "step": 395 }, { "epoch": 0.064, "grad_norm": 0.5829480290412903, "learning_rate": 4.949637630960617e-05, "loss": 0.6956, "num_input_tokens_seen": 2146560, "step": 400 }, { "epoch": 0.0648, "grad_norm": 0.5653419494628906, "learning_rate": 4.94837508084505e-05, "loss": 0.7315, "num_input_tokens_seen": 2169232, "step": 405 }, { "epoch": 0.0656, "grad_norm": 1.0192047357559204, "learning_rate": 4.947097065461801e-05, "loss": 0.7075, "num_input_tokens_seen": 2192224, "step": 410 }, { "epoch": 0.0664, "grad_norm": 0.7392141819000244, "learning_rate": 4.945803592883509e-05, "loss": 0.811, "num_input_tokens_seen": 2216784, "step": 415 }, { "epoch": 0.0672, "grad_norm": 0.6470807194709778, "learning_rate": 4.9444946712804494e-05, "loss": 0.7835, "num_input_tokens_seen": 2243120, "step": 420 }, { "epoch": 0.068, "grad_norm": 0.5305742025375366, "learning_rate": 4.943170308920484e-05, "loss": 0.7211, "num_input_tokens_seen": 2270896, "step": 425 }, { "epoch": 0.0688, "grad_norm": 0.8647666573524475, "learning_rate": 4.941830514169004e-05, "loss": 0.72, "num_input_tokens_seen": 2298528, "step": 430 }, { "epoch": 0.0696, "grad_norm": 0.6244668364524841, "learning_rate": 4.9404752954888824e-05, "loss": 0.7206, "num_input_tokens_seen": 2328080, "step": 435 }, { "epoch": 0.0704, "grad_norm": 0.6552883386611938, "learning_rate": 4.939104661440415e-05, "loss": 0.8018, "num_input_tokens_seen": 2355776, "step": 440 }, { "epoch": 0.0712, "grad_norm": 0.8276055455207825, "learning_rate": 4.937718620681273e-05, "loss": 0.8267, "num_input_tokens_seen": 2379056, "step": 445 }, { "epoch": 0.072, "grad_norm": 0.6930189728736877, "learning_rate": 4.9363171819664434e-05, "loss": 0.8961, "num_input_tokens_seen": 2401664, "step": 450 }, { "epoch": 0.0728, "grad_norm": 0.7441433668136597, "learning_rate": 4.934900354148173e-05, "loss": 0.6942, "num_input_tokens_seen": 2427456, "step": 455 }, { "epoch": 0.0736, "grad_norm": 0.5929616093635559, "learning_rate": 4.933468146175918e-05, "loss": 0.7874, "num_input_tokens_seen": 2450752, "step": 460 }, { "epoch": 0.0744, "grad_norm": 0.5789006948471069, "learning_rate": 4.9320205670962814e-05, "loss": 0.7162, "num_input_tokens_seen": 2473856, "step": 465 }, { "epoch": 0.0752, "grad_norm": 0.6359069347381592, "learning_rate": 4.9305576260529607e-05, "loss": 0.7434, "num_input_tokens_seen": 2502928, "step": 470 }, { "epoch": 0.076, "grad_norm": 0.6155191659927368, "learning_rate": 4.929079332286685e-05, "loss": 0.6932, "num_input_tokens_seen": 2536144, "step": 475 }, { "epoch": 0.0768, "grad_norm": 0.6511387228965759, "learning_rate": 4.927585695135162e-05, "loss": 0.8053, "num_input_tokens_seen": 2562688, "step": 480 }, { "epoch": 0.0776, "grad_norm": 0.5791414976119995, "learning_rate": 4.926076724033016e-05, "loss": 0.7482, "num_input_tokens_seen": 2594480, "step": 485 }, { "epoch": 0.0784, "grad_norm": 0.5258495807647705, "learning_rate": 4.9245524285117274e-05, "loss": 0.7075, "num_input_tokens_seen": 2624736, "step": 490 }, { "epoch": 0.0792, "grad_norm": 0.5191717743873596, "learning_rate": 4.923012818199576e-05, "loss": 0.6132, "num_input_tokens_seen": 2648880, "step": 495 }, { "epoch": 0.08, "grad_norm": 0.8281647562980652, "learning_rate": 4.9214579028215776e-05, "loss": 0.6679, "num_input_tokens_seen": 2675888, "step": 500 }, { "epoch": 0.0808, "grad_norm": 0.588010847568512, "learning_rate": 4.919887692199423e-05, "loss": 0.7016, "num_input_tokens_seen": 2699392, "step": 505 }, { "epoch": 0.0816, "grad_norm": 0.8409311771392822, "learning_rate": 4.918302196251415e-05, "loss": 0.7216, "num_input_tokens_seen": 2726432, "step": 510 }, { "epoch": 0.0824, "grad_norm": 0.6029579639434814, "learning_rate": 4.9167014249924075e-05, "loss": 0.6602, "num_input_tokens_seen": 2756336, "step": 515 }, { "epoch": 0.0832, "grad_norm": 0.7269614934921265, "learning_rate": 4.9150853885337426e-05, "loss": 0.6956, "num_input_tokens_seen": 2781648, "step": 520 }, { "epoch": 0.084, "grad_norm": 0.5419861674308777, "learning_rate": 4.913454097083185e-05, "loss": 0.6427, "num_input_tokens_seen": 2810336, "step": 525 }, { "epoch": 0.0848, "grad_norm": 0.9006750583648682, "learning_rate": 4.911807560944858e-05, "loss": 0.8328, "num_input_tokens_seen": 2836432, "step": 530 }, { "epoch": 0.0856, "grad_norm": 0.7180121541023254, "learning_rate": 4.9101457905191774e-05, "loss": 0.8104, "num_input_tokens_seen": 2863616, "step": 535 }, { "epoch": 0.0864, "grad_norm": 0.6724479794502258, "learning_rate": 4.9084687963027894e-05, "loss": 0.6858, "num_input_tokens_seen": 2891264, "step": 540 }, { "epoch": 0.0872, "grad_norm": 0.7073305249214172, "learning_rate": 4.906776588888502e-05, "loss": 0.7271, "num_input_tokens_seen": 2916256, "step": 545 }, { "epoch": 0.088, "grad_norm": 0.7945154309272766, "learning_rate": 4.905069178965215e-05, "loss": 0.7527, "num_input_tokens_seen": 2944112, "step": 550 }, { "epoch": 0.0888, "grad_norm": 0.5791934728622437, "learning_rate": 4.903346577317859e-05, "loss": 0.7341, "num_input_tokens_seen": 2972512, "step": 555 }, { "epoch": 0.0896, "grad_norm": 0.8222031593322754, "learning_rate": 4.90160879482732e-05, "loss": 0.7339, "num_input_tokens_seen": 2997168, "step": 560 }, { "epoch": 0.0904, "grad_norm": 0.6719418168067932, "learning_rate": 4.89985584247038e-05, "loss": 0.768, "num_input_tokens_seen": 3020880, "step": 565 }, { "epoch": 0.0912, "grad_norm": 0.7740746140480042, "learning_rate": 4.898087731319636e-05, "loss": 0.7014, "num_input_tokens_seen": 3044224, "step": 570 }, { "epoch": 0.092, "grad_norm": 0.5642164945602417, "learning_rate": 4.89630447254344e-05, "loss": 0.6203, "num_input_tokens_seen": 3071680, "step": 575 }, { "epoch": 0.0928, "grad_norm": 1.4719825983047485, "learning_rate": 4.894506077405824e-05, "loss": 0.7461, "num_input_tokens_seen": 3099088, "step": 580 }, { "epoch": 0.0936, "grad_norm": 0.6961272954940796, "learning_rate": 4.892692557266429e-05, "loss": 0.7357, "num_input_tokens_seen": 3127728, "step": 585 }, { "epoch": 0.0944, "grad_norm": 0.686820924282074, "learning_rate": 4.8908639235804324e-05, "loss": 0.7819, "num_input_tokens_seen": 3154336, "step": 590 }, { "epoch": 0.0952, "grad_norm": 0.7145109176635742, "learning_rate": 4.8890201878984796e-05, "loss": 0.7121, "num_input_tokens_seen": 3178768, "step": 595 }, { "epoch": 0.096, "grad_norm": 0.6159213781356812, "learning_rate": 4.887161361866608e-05, "loss": 0.6698, "num_input_tokens_seen": 3211968, "step": 600 }, { "epoch": 0.0968, "grad_norm": 0.8054212927818298, "learning_rate": 4.885287457226172e-05, "loss": 0.7606, "num_input_tokens_seen": 3238272, "step": 605 }, { "epoch": 0.0976, "grad_norm": 1.1526386737823486, "learning_rate": 4.8833984858137715e-05, "loss": 0.7694, "num_input_tokens_seen": 3270208, "step": 610 }, { "epoch": 0.0984, "grad_norm": 0.5728780031204224, "learning_rate": 4.8814944595611776e-05, "loss": 0.7227, "num_input_tokens_seen": 3296192, "step": 615 }, { "epoch": 0.0992, "grad_norm": 0.6360820531845093, "learning_rate": 4.8795753904952534e-05, "loss": 0.7275, "num_input_tokens_seen": 3321232, "step": 620 }, { "epoch": 0.1, "grad_norm": 0.6169213056564331, "learning_rate": 4.877641290737884e-05, "loss": 0.6746, "num_input_tokens_seen": 3343984, "step": 625 }, { "epoch": 0.1008, "grad_norm": 0.8000876307487488, "learning_rate": 4.8756921725058934e-05, "loss": 0.8223, "num_input_tokens_seen": 3367824, "step": 630 }, { "epoch": 0.1016, "grad_norm": 0.5983218550682068, "learning_rate": 4.8737280481109724e-05, "loss": 0.8487, "num_input_tokens_seen": 3394800, "step": 635 }, { "epoch": 0.1024, "grad_norm": 0.9402346014976501, "learning_rate": 4.871748929959598e-05, "loss": 0.7441, "num_input_tokens_seen": 3421360, "step": 640 }, { "epoch": 0.1032, "grad_norm": 0.6266387104988098, "learning_rate": 4.869754830552956e-05, "loss": 0.7631, "num_input_tokens_seen": 3449584, "step": 645 }, { "epoch": 0.104, "grad_norm": 0.7829232215881348, "learning_rate": 4.867745762486861e-05, "loss": 0.7793, "num_input_tokens_seen": 3477168, "step": 650 }, { "epoch": 0.1048, "grad_norm": 0.7125943303108215, "learning_rate": 4.86572173845168e-05, "loss": 0.7415, "num_input_tokens_seen": 3505056, "step": 655 }, { "epoch": 0.1056, "grad_norm": 0.6520003080368042, "learning_rate": 4.863682771232248e-05, "loss": 0.7157, "num_input_tokens_seen": 3534576, "step": 660 }, { "epoch": 0.1064, "grad_norm": 0.5907071828842163, "learning_rate": 4.861628873707792e-05, "loss": 0.7287, "num_input_tokens_seen": 3560688, "step": 665 }, { "epoch": 0.1072, "grad_norm": 0.8829016089439392, "learning_rate": 4.859560058851844e-05, "loss": 0.7351, "num_input_tokens_seen": 3586176, "step": 670 }, { "epoch": 0.108, "grad_norm": 0.917322039604187, "learning_rate": 4.8574763397321614e-05, "loss": 0.6213, "num_input_tokens_seen": 3615472, "step": 675 }, { "epoch": 0.1088, "grad_norm": 0.6344768404960632, "learning_rate": 4.855377729510648e-05, "loss": 0.729, "num_input_tokens_seen": 3638256, "step": 680 }, { "epoch": 0.1096, "grad_norm": 0.7305799722671509, "learning_rate": 4.8532642414432674e-05, "loss": 0.7242, "num_input_tokens_seen": 3667824, "step": 685 }, { "epoch": 0.1104, "grad_norm": 0.7569555044174194, "learning_rate": 4.851135888879958e-05, "loss": 0.7831, "num_input_tokens_seen": 3695408, "step": 690 }, { "epoch": 0.1112, "grad_norm": 0.7566932439804077, "learning_rate": 4.8489926852645505e-05, "loss": 0.7181, "num_input_tokens_seen": 3719888, "step": 695 }, { "epoch": 0.112, "grad_norm": 0.7932357788085938, "learning_rate": 4.846834644134686e-05, "loss": 0.7961, "num_input_tokens_seen": 3744512, "step": 700 }, { "epoch": 0.1128, "grad_norm": 0.708210825920105, "learning_rate": 4.844661779121722e-05, "loss": 0.8362, "num_input_tokens_seen": 3771968, "step": 705 }, { "epoch": 0.1136, "grad_norm": 0.7361094951629639, "learning_rate": 4.8424741039506575e-05, "loss": 0.7645, "num_input_tokens_seen": 3801680, "step": 710 }, { "epoch": 0.1144, "grad_norm": 0.48908814787864685, "learning_rate": 4.840271632440038e-05, "loss": 0.7042, "num_input_tokens_seen": 3833952, "step": 715 }, { "epoch": 0.1152, "grad_norm": 0.6167788505554199, "learning_rate": 4.8380543785018677e-05, "loss": 0.7476, "num_input_tokens_seen": 3860144, "step": 720 }, { "epoch": 0.116, "grad_norm": 0.68650883436203, "learning_rate": 4.8358223561415304e-05, "loss": 0.7415, "num_input_tokens_seen": 3890304, "step": 725 }, { "epoch": 0.1168, "grad_norm": 0.7059746384620667, "learning_rate": 4.833575579457691e-05, "loss": 0.6717, "num_input_tokens_seen": 3914560, "step": 730 }, { "epoch": 0.1176, "grad_norm": 0.8362336158752441, "learning_rate": 4.8313140626422125e-05, "loss": 0.7545, "num_input_tokens_seen": 3940128, "step": 735 }, { "epoch": 0.1184, "grad_norm": 0.5400592684745789, "learning_rate": 4.829037819980065e-05, "loss": 0.7809, "num_input_tokens_seen": 3970608, "step": 740 }, { "epoch": 0.1192, "grad_norm": 1.0431326627731323, "learning_rate": 4.8267468658492335e-05, "loss": 0.7904, "num_input_tokens_seen": 3996960, "step": 745 }, { "epoch": 0.12, "grad_norm": 0.5358662605285645, "learning_rate": 4.8244412147206284e-05, "loss": 0.7688, "num_input_tokens_seen": 4021488, "step": 750 }, { "epoch": 0.1208, "grad_norm": 0.8147661685943604, "learning_rate": 4.822120881157998e-05, "loss": 0.7819, "num_input_tokens_seen": 4047136, "step": 755 }, { "epoch": 0.1216, "grad_norm": 0.6247139573097229, "learning_rate": 4.819785879817827e-05, "loss": 0.6757, "num_input_tokens_seen": 4072256, "step": 760 }, { "epoch": 0.1224, "grad_norm": 0.8849884271621704, "learning_rate": 4.817436225449255e-05, "loss": 0.8952, "num_input_tokens_seen": 4095328, "step": 765 }, { "epoch": 0.1232, "grad_norm": 0.8693557977676392, "learning_rate": 4.8150719328939755e-05, "loss": 0.6998, "num_input_tokens_seen": 4118896, "step": 770 }, { "epoch": 0.124, "grad_norm": 0.9492819905281067, "learning_rate": 4.812693017086145e-05, "loss": 0.7675, "num_input_tokens_seen": 4144576, "step": 775 }, { "epoch": 0.1248, "grad_norm": 0.8479375243186951, "learning_rate": 4.810299493052289e-05, "loss": 0.7332, "num_input_tokens_seen": 4172448, "step": 780 }, { "epoch": 0.1256, "grad_norm": 0.7956748008728027, "learning_rate": 4.8078913759112066e-05, "loss": 0.6942, "num_input_tokens_seen": 4196032, "step": 785 }, { "epoch": 0.1264, "grad_norm": 0.6426162123680115, "learning_rate": 4.805468680873874e-05, "loss": 0.7536, "num_input_tokens_seen": 4224320, "step": 790 }, { "epoch": 0.1272, "grad_norm": 0.6501713991165161, "learning_rate": 4.803031423243349e-05, "loss": 0.6722, "num_input_tokens_seen": 4252752, "step": 795 }, { "epoch": 0.128, "grad_norm": 0.773551881313324, "learning_rate": 4.800579618414676e-05, "loss": 0.7651, "num_input_tokens_seen": 4278480, "step": 800 }, { "epoch": 0.1288, "grad_norm": 0.6473078727722168, "learning_rate": 4.7981132818747876e-05, "loss": 0.6626, "num_input_tokens_seen": 4305920, "step": 805 }, { "epoch": 0.1296, "grad_norm": 0.5944277048110962, "learning_rate": 4.795632429202405e-05, "loss": 0.8511, "num_input_tokens_seen": 4330448, "step": 810 }, { "epoch": 0.1304, "grad_norm": 0.6878964900970459, "learning_rate": 4.793137076067942e-05, "loss": 0.7524, "num_input_tokens_seen": 4356880, "step": 815 }, { "epoch": 0.1312, "grad_norm": 0.9247101545333862, "learning_rate": 4.790627238233405e-05, "loss": 0.8498, "num_input_tokens_seen": 4383744, "step": 820 }, { "epoch": 0.132, "grad_norm": 0.9401747584342957, "learning_rate": 4.788102931552294e-05, "loss": 0.647, "num_input_tokens_seen": 4411120, "step": 825 }, { "epoch": 0.1328, "grad_norm": 0.765521764755249, "learning_rate": 4.7855641719695023e-05, "loss": 0.7766, "num_input_tokens_seen": 4435920, "step": 830 }, { "epoch": 0.1336, "grad_norm": 0.8985924124717712, "learning_rate": 4.783010975521216e-05, "loss": 0.7426, "num_input_tokens_seen": 4462768, "step": 835 }, { "epoch": 0.1344, "grad_norm": 0.8223104476928711, "learning_rate": 4.78044335833481e-05, "loss": 0.6919, "num_input_tokens_seen": 4493232, "step": 840 }, { "epoch": 0.1352, "grad_norm": 0.6721159219741821, "learning_rate": 4.7778613366287505e-05, "loss": 0.7221, "num_input_tokens_seen": 4520048, "step": 845 }, { "epoch": 0.136, "grad_norm": 0.6136463284492493, "learning_rate": 4.775264926712489e-05, "loss": 0.7344, "num_input_tokens_seen": 4545984, "step": 850 }, { "epoch": 0.1368, "grad_norm": 0.7662776708602905, "learning_rate": 4.772654144986364e-05, "loss": 0.6648, "num_input_tokens_seen": 4577296, "step": 855 }, { "epoch": 0.1376, "grad_norm": 0.8455452919006348, "learning_rate": 4.7700290079414896e-05, "loss": 0.7513, "num_input_tokens_seen": 4602272, "step": 860 }, { "epoch": 0.1384, "grad_norm": 0.8613116145133972, "learning_rate": 4.767389532159659e-05, "loss": 0.7792, "num_input_tokens_seen": 4631008, "step": 865 }, { "epoch": 0.1392, "grad_norm": 0.5791791677474976, "learning_rate": 4.764735734313236e-05, "loss": 0.7529, "num_input_tokens_seen": 4660112, "step": 870 }, { "epoch": 0.14, "grad_norm": 0.8197824954986572, "learning_rate": 4.762067631165049e-05, "loss": 0.6728, "num_input_tokens_seen": 4689504, "step": 875 }, { "epoch": 0.1408, "grad_norm": 0.9617213606834412, "learning_rate": 4.759385239568289e-05, "loss": 0.6935, "num_input_tokens_seen": 4715312, "step": 880 }, { "epoch": 0.1416, "grad_norm": 0.7933773398399353, "learning_rate": 4.756688576466398e-05, "loss": 0.8062, "num_input_tokens_seen": 4735936, "step": 885 }, { "epoch": 0.1424, "grad_norm": 0.975724458694458, "learning_rate": 4.753977658892967e-05, "loss": 0.7149, "num_input_tokens_seen": 4760256, "step": 890 }, { "epoch": 0.1432, "grad_norm": 0.755167543888092, "learning_rate": 4.751252503971624e-05, "loss": 0.7062, "num_input_tokens_seen": 4789264, "step": 895 }, { "epoch": 0.144, "grad_norm": 0.84686279296875, "learning_rate": 4.7485131289159276e-05, "loss": 0.837, "num_input_tokens_seen": 4815344, "step": 900 }, { "epoch": 0.1448, "grad_norm": 0.9440627098083496, "learning_rate": 4.745759551029261e-05, "loss": 0.6907, "num_input_tokens_seen": 4840528, "step": 905 }, { "epoch": 0.1456, "grad_norm": 0.7293935418128967, "learning_rate": 4.742991787704719e-05, "loss": 0.7192, "num_input_tokens_seen": 4868032, "step": 910 }, { "epoch": 0.1464, "grad_norm": 0.6401370763778687, "learning_rate": 4.7402098564249974e-05, "loss": 0.7223, "num_input_tokens_seen": 4893376, "step": 915 }, { "epoch": 0.1472, "grad_norm": 0.8882667422294617, "learning_rate": 4.737413774762287e-05, "loss": 0.6847, "num_input_tokens_seen": 4918288, "step": 920 }, { "epoch": 0.148, "grad_norm": 0.7663973569869995, "learning_rate": 4.73460356037816e-05, "loss": 0.7273, "num_input_tokens_seen": 4944688, "step": 925 }, { "epoch": 0.1488, "grad_norm": 0.7966660857200623, "learning_rate": 4.731779231023456e-05, "loss": 0.7087, "num_input_tokens_seen": 4969744, "step": 930 }, { "epoch": 0.1496, "grad_norm": 1.3590271472930908, "learning_rate": 4.728940804538176e-05, "loss": 0.7771, "num_input_tokens_seen": 4997072, "step": 935 }, { "epoch": 0.1504, "grad_norm": 0.8245935440063477, "learning_rate": 4.7260882988513624e-05, "loss": 0.7598, "num_input_tokens_seen": 5024672, "step": 940 }, { "epoch": 0.1512, "grad_norm": 0.6972777247428894, "learning_rate": 4.723221731980993e-05, "loss": 0.7961, "num_input_tokens_seen": 5051952, "step": 945 }, { "epoch": 0.152, "grad_norm": 0.7691385746002197, "learning_rate": 4.720341122033862e-05, "loss": 0.773, "num_input_tokens_seen": 5074528, "step": 950 }, { "epoch": 0.1528, "grad_norm": 1.4361584186553955, "learning_rate": 4.717446487205466e-05, "loss": 0.7216, "num_input_tokens_seen": 5099840, "step": 955 }, { "epoch": 0.1536, "grad_norm": 0.9640448689460754, "learning_rate": 4.714537845779894e-05, "loss": 0.6569, "num_input_tokens_seen": 5122848, "step": 960 }, { "epoch": 0.1544, "grad_norm": 0.8036720156669617, "learning_rate": 4.7116152161297045e-05, "loss": 0.7994, "num_input_tokens_seen": 5152320, "step": 965 }, { "epoch": 0.1552, "grad_norm": 0.7904760241508484, "learning_rate": 4.708678616715815e-05, "loss": 0.7259, "num_input_tokens_seen": 5178816, "step": 970 }, { "epoch": 0.156, "grad_norm": 0.7007213830947876, "learning_rate": 4.7057280660873835e-05, "loss": 0.747, "num_input_tokens_seen": 5208112, "step": 975 }, { "epoch": 0.1568, "grad_norm": 0.8959905505180359, "learning_rate": 4.702763582881692e-05, "loss": 0.8487, "num_input_tokens_seen": 5231200, "step": 980 }, { "epoch": 0.1576, "grad_norm": 0.8828222155570984, "learning_rate": 4.699785185824026e-05, "loss": 0.7654, "num_input_tokens_seen": 5257312, "step": 985 }, { "epoch": 0.1584, "grad_norm": 1.0120501518249512, "learning_rate": 4.696792893727562e-05, "loss": 0.7748, "num_input_tokens_seen": 5280288, "step": 990 }, { "epoch": 0.1592, "grad_norm": 0.8728295564651489, "learning_rate": 4.693786725493242e-05, "loss": 0.6957, "num_input_tokens_seen": 5308272, "step": 995 }, { "epoch": 0.16, "grad_norm": 1.252682089805603, "learning_rate": 4.690766700109659e-05, "loss": 0.7418, "num_input_tokens_seen": 5335568, "step": 1000 }, { "epoch": 0.1608, "grad_norm": 1.0933836698532104, "learning_rate": 4.6877328366529346e-05, "loss": 0.8225, "num_input_tokens_seen": 5361872, "step": 1005 }, { "epoch": 0.1616, "grad_norm": 0.8167702555656433, "learning_rate": 4.684685154286599e-05, "loss": 0.8552, "num_input_tokens_seen": 5387456, "step": 1010 }, { "epoch": 0.1624, "grad_norm": 0.6640987396240234, "learning_rate": 4.681623672261469e-05, "loss": 0.6654, "num_input_tokens_seen": 5411472, "step": 1015 }, { "epoch": 0.1632, "grad_norm": 0.9311304688453674, "learning_rate": 4.678548409915532e-05, "loss": 0.7339, "num_input_tokens_seen": 5439648, "step": 1020 }, { "epoch": 0.164, "grad_norm": 0.9733570218086243, "learning_rate": 4.675459386673815e-05, "loss": 0.7324, "num_input_tokens_seen": 5468416, "step": 1025 }, { "epoch": 0.1648, "grad_norm": 0.6521669030189514, "learning_rate": 4.6723566220482664e-05, "loss": 0.7065, "num_input_tokens_seen": 5498800, "step": 1030 }, { "epoch": 0.1656, "grad_norm": 0.6702024340629578, "learning_rate": 4.669240135637635e-05, "loss": 0.6822, "num_input_tokens_seen": 5527856, "step": 1035 }, { "epoch": 0.1664, "grad_norm": 1.3757870197296143, "learning_rate": 4.666109947127343e-05, "loss": 0.7554, "num_input_tokens_seen": 5550848, "step": 1040 }, { "epoch": 0.1672, "grad_norm": 0.7441242933273315, "learning_rate": 4.662966076289362e-05, "loss": 0.6784, "num_input_tokens_seen": 5581552, "step": 1045 }, { "epoch": 0.168, "grad_norm": 0.7709234356880188, "learning_rate": 4.659808542982088e-05, "loss": 0.8294, "num_input_tokens_seen": 5604288, "step": 1050 }, { "epoch": 0.1688, "grad_norm": 0.5358073115348816, "learning_rate": 4.6566373671502196e-05, "loss": 0.6633, "num_input_tokens_seen": 5630336, "step": 1055 }, { "epoch": 0.1696, "grad_norm": 0.5856006741523743, "learning_rate": 4.653452568824625e-05, "loss": 0.6684, "num_input_tokens_seen": 5662480, "step": 1060 }, { "epoch": 0.1704, "grad_norm": 0.7003797292709351, "learning_rate": 4.650254168122222e-05, "loss": 0.7109, "num_input_tokens_seen": 5687376, "step": 1065 }, { "epoch": 0.1712, "grad_norm": 0.7874431014060974, "learning_rate": 4.647042185245847e-05, "loss": 0.8036, "num_input_tokens_seen": 5714896, "step": 1070 }, { "epoch": 0.172, "grad_norm": 0.6988087296485901, "learning_rate": 4.643816640484131e-05, "loss": 0.6575, "num_input_tokens_seen": 5740192, "step": 1075 }, { "epoch": 0.1728, "grad_norm": 0.982477068901062, "learning_rate": 4.640577554211366e-05, "loss": 0.7477, "num_input_tokens_seen": 5768656, "step": 1080 }, { "epoch": 0.1736, "grad_norm": 1.1265698671340942, "learning_rate": 4.6373249468873833e-05, "loss": 0.7555, "num_input_tokens_seen": 5794576, "step": 1085 }, { "epoch": 0.1744, "grad_norm": 0.6747913360595703, "learning_rate": 4.634058839057417e-05, "loss": 0.6695, "num_input_tokens_seen": 5823296, "step": 1090 }, { "epoch": 0.1752, "grad_norm": 0.8027223348617554, "learning_rate": 4.63077925135198e-05, "loss": 0.6948, "num_input_tokens_seen": 5846928, "step": 1095 }, { "epoch": 0.176, "grad_norm": 0.7862293720245361, "learning_rate": 4.6274862044867304e-05, "loss": 0.7728, "num_input_tokens_seen": 5871968, "step": 1100 }, { "epoch": 0.1768, "grad_norm": 0.7790197134017944, "learning_rate": 4.624179719262342e-05, "loss": 0.765, "num_input_tokens_seen": 5900304, "step": 1105 }, { "epoch": 0.1776, "grad_norm": 0.8996221423149109, "learning_rate": 4.6208598165643715e-05, "loss": 0.6515, "num_input_tokens_seen": 5925792, "step": 1110 }, { "epoch": 0.1784, "grad_norm": 0.7972677946090698, "learning_rate": 4.61752651736313e-05, "loss": 0.75, "num_input_tokens_seen": 5950672, "step": 1115 }, { "epoch": 0.1792, "grad_norm": 0.6896753907203674, "learning_rate": 4.614179842713547e-05, "loss": 0.6592, "num_input_tokens_seen": 5985552, "step": 1120 }, { "epoch": 0.18, "grad_norm": 0.823128342628479, "learning_rate": 4.610819813755038e-05, "loss": 0.8463, "num_input_tokens_seen": 6009904, "step": 1125 }, { "epoch": 0.1808, "grad_norm": 0.8550837635993958, "learning_rate": 4.607446451711372e-05, "loss": 0.7349, "num_input_tokens_seen": 6034160, "step": 1130 }, { "epoch": 0.1816, "grad_norm": 0.8120406270027161, "learning_rate": 4.604059777890537e-05, "loss": 0.6396, "num_input_tokens_seen": 6056544, "step": 1135 }, { "epoch": 0.1824, "grad_norm": 0.6196752786636353, "learning_rate": 4.6006598136846056e-05, "loss": 0.6164, "num_input_tokens_seen": 6083920, "step": 1140 }, { "epoch": 0.1832, "grad_norm": 0.6641353368759155, "learning_rate": 4.5972465805695996e-05, "loss": 0.6775, "num_input_tokens_seen": 6111520, "step": 1145 }, { "epoch": 0.184, "grad_norm": 0.7323867082595825, "learning_rate": 4.593820100105355e-05, "loss": 0.6295, "num_input_tokens_seen": 6141056, "step": 1150 }, { "epoch": 0.1848, "grad_norm": 0.6919586658477783, "learning_rate": 4.590380393935383e-05, "loss": 0.7429, "num_input_tokens_seen": 6163408, "step": 1155 }, { "epoch": 0.1856, "grad_norm": 0.9530206322669983, "learning_rate": 4.5869274837867394e-05, "loss": 0.7516, "num_input_tokens_seen": 6188816, "step": 1160 }, { "epoch": 0.1864, "grad_norm": 0.9966915845870972, "learning_rate": 4.583461391469879e-05, "loss": 0.7524, "num_input_tokens_seen": 6216800, "step": 1165 }, { "epoch": 0.1872, "grad_norm": 1.096708059310913, "learning_rate": 4.579982138878527e-05, "loss": 0.7337, "num_input_tokens_seen": 6245888, "step": 1170 }, { "epoch": 0.188, "grad_norm": 0.8707526326179504, "learning_rate": 4.5764897479895317e-05, "loss": 0.7891, "num_input_tokens_seen": 6275120, "step": 1175 }, { "epoch": 0.1888, "grad_norm": 0.7489879727363586, "learning_rate": 4.5729842408627334e-05, "loss": 0.79, "num_input_tokens_seen": 6299760, "step": 1180 }, { "epoch": 0.1896, "grad_norm": 0.7835171222686768, "learning_rate": 4.5694656396408195e-05, "loss": 0.7506, "num_input_tokens_seen": 6326720, "step": 1185 }, { "epoch": 0.1904, "grad_norm": 0.7588552832603455, "learning_rate": 4.565933966549189e-05, "loss": 0.6294, "num_input_tokens_seen": 6353728, "step": 1190 }, { "epoch": 0.1912, "grad_norm": 0.6706573367118835, "learning_rate": 4.5623892438958074e-05, "loss": 0.7564, "num_input_tokens_seen": 6379536, "step": 1195 }, { "epoch": 0.192, "grad_norm": 0.7340586185455322, "learning_rate": 4.558831494071069e-05, "loss": 0.7683, "num_input_tokens_seen": 6407152, "step": 1200 }, { "epoch": 0.1928, "grad_norm": 0.735789954662323, "learning_rate": 4.555260739547657e-05, "loss": 0.7701, "num_input_tokens_seen": 6434480, "step": 1205 }, { "epoch": 0.1936, "grad_norm": 0.8325262069702148, "learning_rate": 4.5516770028803954e-05, "loss": 0.694, "num_input_tokens_seen": 6463424, "step": 1210 }, { "epoch": 0.1944, "grad_norm": 0.7930346727371216, "learning_rate": 4.548080306706114e-05, "loss": 0.7322, "num_input_tokens_seen": 6487136, "step": 1215 }, { "epoch": 0.1952, "grad_norm": 0.7683930397033691, "learning_rate": 4.5444706737435014e-05, "loss": 0.7616, "num_input_tokens_seen": 6513120, "step": 1220 }, { "epoch": 0.196, "grad_norm": 0.600136399269104, "learning_rate": 4.5408481267929605e-05, "loss": 0.6743, "num_input_tokens_seen": 6543040, "step": 1225 }, { "epoch": 0.1968, "grad_norm": 0.9069085121154785, "learning_rate": 4.5372126887364655e-05, "loss": 0.7377, "num_input_tokens_seen": 6572432, "step": 1230 }, { "epoch": 0.1976, "grad_norm": 0.9226580262184143, "learning_rate": 4.533564382537421e-05, "loss": 0.7766, "num_input_tokens_seen": 6593136, "step": 1235 }, { "epoch": 0.1984, "grad_norm": 0.7376300096511841, "learning_rate": 4.529903231240511e-05, "loss": 0.7873, "num_input_tokens_seen": 6621024, "step": 1240 }, { "epoch": 0.1992, "grad_norm": 0.6371731162071228, "learning_rate": 4.5262292579715556e-05, "loss": 0.7096, "num_input_tokens_seen": 6646480, "step": 1245 }, { "epoch": 0.2, "grad_norm": 0.8643271327018738, "learning_rate": 4.522542485937369e-05, "loss": 0.8187, "num_input_tokens_seen": 6674032, "step": 1250 }, { "epoch": 0.2008, "grad_norm": 0.8012109398841858, "learning_rate": 4.518842938425605e-05, "loss": 0.772, "num_input_tokens_seen": 6700112, "step": 1255 }, { "epoch": 0.2016, "grad_norm": 0.7719143033027649, "learning_rate": 4.5151306388046175e-05, "loss": 0.6796, "num_input_tokens_seen": 6727008, "step": 1260 }, { "epoch": 0.2024, "grad_norm": 0.8668113946914673, "learning_rate": 4.511405610523309e-05, "loss": 0.7177, "num_input_tokens_seen": 6752768, "step": 1265 }, { "epoch": 0.2032, "grad_norm": 0.8964220285415649, "learning_rate": 4.5076678771109815e-05, "loss": 0.7078, "num_input_tokens_seen": 6778112, "step": 1270 }, { "epoch": 0.204, "grad_norm": 0.7097613215446472, "learning_rate": 4.503917462177192e-05, "loss": 0.6496, "num_input_tokens_seen": 6804432, "step": 1275 }, { "epoch": 0.2048, "grad_norm": 0.842675507068634, "learning_rate": 4.5001543894115975e-05, "loss": 0.6802, "num_input_tokens_seen": 6829824, "step": 1280 }, { "epoch": 0.2056, "grad_norm": 0.7390193343162537, "learning_rate": 4.496378682583813e-05, "loss": 0.7187, "num_input_tokens_seen": 6858480, "step": 1285 }, { "epoch": 0.2064, "grad_norm": 0.5758505463600159, "learning_rate": 4.492590365543253e-05, "loss": 0.6198, "num_input_tokens_seen": 6886960, "step": 1290 }, { "epoch": 0.2072, "grad_norm": 0.9554662108421326, "learning_rate": 4.488789462218987e-05, "loss": 0.6105, "num_input_tokens_seen": 6912560, "step": 1295 }, { "epoch": 0.208, "grad_norm": 0.9423254728317261, "learning_rate": 4.484975996619589e-05, "loss": 0.7671, "num_input_tokens_seen": 6938912, "step": 1300 }, { "epoch": 0.2088, "grad_norm": 0.7120509743690491, "learning_rate": 4.481149992832977e-05, "loss": 0.6833, "num_input_tokens_seen": 6967616, "step": 1305 }, { "epoch": 0.2096, "grad_norm": 0.9409400224685669, "learning_rate": 4.477311475026271e-05, "loss": 0.7547, "num_input_tokens_seen": 6993872, "step": 1310 }, { "epoch": 0.2104, "grad_norm": 0.8102442026138306, "learning_rate": 4.473460467445637e-05, "loss": 0.7479, "num_input_tokens_seen": 7020784, "step": 1315 }, { "epoch": 0.2112, "grad_norm": 0.787486732006073, "learning_rate": 4.46959699441613e-05, "loss": 0.761, "num_input_tokens_seen": 7045024, "step": 1320 }, { "epoch": 0.212, "grad_norm": 0.8877683877944946, "learning_rate": 4.465721080341547e-05, "loss": 0.7612, "num_input_tokens_seen": 7072448, "step": 1325 }, { "epoch": 0.2128, "grad_norm": 0.7483372688293457, "learning_rate": 4.461832749704268e-05, "loss": 0.6792, "num_input_tokens_seen": 7097776, "step": 1330 }, { "epoch": 0.2136, "grad_norm": 0.7852973341941833, "learning_rate": 4.457932027065102e-05, "loss": 0.7357, "num_input_tokens_seen": 7123568, "step": 1335 }, { "epoch": 0.2144, "grad_norm": 0.7306565642356873, "learning_rate": 4.4540189370631315e-05, "loss": 0.6676, "num_input_tokens_seen": 7151728, "step": 1340 }, { "epoch": 0.2152, "grad_norm": 0.7990534901618958, "learning_rate": 4.4500935044155626e-05, "loss": 0.7394, "num_input_tokens_seen": 7181664, "step": 1345 }, { "epoch": 0.216, "grad_norm": 1.287644863128662, "learning_rate": 4.4461557539175594e-05, "loss": 0.8017, "num_input_tokens_seen": 7210336, "step": 1350 }, { "epoch": 0.2168, "grad_norm": 0.7476962208747864, "learning_rate": 4.4422057104420946e-05, "loss": 0.6533, "num_input_tokens_seen": 7240992, "step": 1355 }, { "epoch": 0.2176, "grad_norm": 0.8233410120010376, "learning_rate": 4.4382433989397895e-05, "loss": 0.7029, "num_input_tokens_seen": 7268048, "step": 1360 }, { "epoch": 0.2184, "grad_norm": 0.609846293926239, "learning_rate": 4.434268844438758e-05, "loss": 0.7096, "num_input_tokens_seen": 7297616, "step": 1365 }, { "epoch": 0.2192, "grad_norm": 1.010886549949646, "learning_rate": 4.4302820720444456e-05, "loss": 0.8103, "num_input_tokens_seen": 7326912, "step": 1370 }, { "epoch": 0.22, "grad_norm": 0.7681688070297241, "learning_rate": 4.426283106939474e-05, "loss": 0.6238, "num_input_tokens_seen": 7355136, "step": 1375 }, { "epoch": 0.2208, "grad_norm": 0.7759270071983337, "learning_rate": 4.422271974383479e-05, "loss": 0.6625, "num_input_tokens_seen": 7377584, "step": 1380 }, { "epoch": 0.2216, "grad_norm": 0.831362783908844, "learning_rate": 4.418248699712955e-05, "loss": 0.6831, "num_input_tokens_seen": 7405552, "step": 1385 }, { "epoch": 0.2224, "grad_norm": 0.7530121207237244, "learning_rate": 4.414213308341092e-05, "loss": 0.7664, "num_input_tokens_seen": 7430960, "step": 1390 }, { "epoch": 0.2232, "grad_norm": 0.8572810292243958, "learning_rate": 4.410165825757613e-05, "loss": 0.7273, "num_input_tokens_seen": 7457136, "step": 1395 }, { "epoch": 0.224, "grad_norm": 0.7553160190582275, "learning_rate": 4.40610627752862e-05, "loss": 0.6607, "num_input_tokens_seen": 7482208, "step": 1400 }, { "epoch": 0.2248, "grad_norm": 0.6897515058517456, "learning_rate": 4.4020346892964246e-05, "loss": 0.731, "num_input_tokens_seen": 7515760, "step": 1405 }, { "epoch": 0.2256, "grad_norm": 0.7974178791046143, "learning_rate": 4.3979510867793917e-05, "loss": 0.7258, "num_input_tokens_seen": 7542944, "step": 1410 }, { "epoch": 0.2264, "grad_norm": 0.8745766282081604, "learning_rate": 4.393855495771774e-05, "loss": 0.6566, "num_input_tokens_seen": 7573760, "step": 1415 }, { "epoch": 0.2272, "grad_norm": 0.749857485294342, "learning_rate": 4.38974794214355e-05, "loss": 0.7433, "num_input_tokens_seen": 7606592, "step": 1420 }, { "epoch": 0.228, "grad_norm": 0.7722298502922058, "learning_rate": 4.3856284518402594e-05, "loss": 0.7452, "num_input_tokens_seen": 7628672, "step": 1425 }, { "epoch": 0.2288, "grad_norm": 0.8768362998962402, "learning_rate": 4.381497050882845e-05, "loss": 0.7077, "num_input_tokens_seen": 7658528, "step": 1430 }, { "epoch": 0.2296, "grad_norm": 0.7979273796081543, "learning_rate": 4.377353765367479e-05, "loss": 0.6274, "num_input_tokens_seen": 7685248, "step": 1435 }, { "epoch": 0.2304, "grad_norm": 0.988314151763916, "learning_rate": 4.3731986214654035e-05, "loss": 0.6845, "num_input_tokens_seen": 7713616, "step": 1440 }, { "epoch": 0.2312, "grad_norm": 0.7991346120834351, "learning_rate": 4.3690316454227674e-05, "loss": 0.7115, "num_input_tokens_seen": 7740304, "step": 1445 }, { "epoch": 0.232, "grad_norm": 1.072383999824524, "learning_rate": 4.3648528635604556e-05, "loss": 0.7209, "num_input_tokens_seen": 7766848, "step": 1450 }, { "epoch": 0.2328, "grad_norm": 1.357325792312622, "learning_rate": 4.360662302273924e-05, "loss": 0.8239, "num_input_tokens_seen": 7791888, "step": 1455 }, { "epoch": 0.2336, "grad_norm": 0.6083495020866394, "learning_rate": 4.3564599880330385e-05, "loss": 0.6199, "num_input_tokens_seen": 7822448, "step": 1460 }, { "epoch": 0.2344, "grad_norm": 0.7359746098518372, "learning_rate": 4.352245947381898e-05, "loss": 0.7481, "num_input_tokens_seen": 7848464, "step": 1465 }, { "epoch": 0.2352, "grad_norm": 0.9160847067832947, "learning_rate": 4.348020206938672e-05, "loss": 0.7235, "num_input_tokens_seen": 7877216, "step": 1470 }, { "epoch": 0.236, "grad_norm": 0.7445215582847595, "learning_rate": 4.343782793395435e-05, "loss": 0.7345, "num_input_tokens_seen": 7904368, "step": 1475 }, { "epoch": 0.2368, "grad_norm": 0.8324536681175232, "learning_rate": 4.3395337335179945e-05, "loss": 0.7532, "num_input_tokens_seen": 7931520, "step": 1480 }, { "epoch": 0.2376, "grad_norm": 1.0249683856964111, "learning_rate": 4.335273054145722e-05, "loss": 0.6902, "num_input_tokens_seen": 7953296, "step": 1485 }, { "epoch": 0.2384, "grad_norm": 0.6565669775009155, "learning_rate": 4.3310007821913836e-05, "loss": 0.7329, "num_input_tokens_seen": 7978832, "step": 1490 }, { "epoch": 0.2392, "grad_norm": 0.8256237506866455, "learning_rate": 4.32671694464097e-05, "loss": 0.6693, "num_input_tokens_seen": 8004992, "step": 1495 }, { "epoch": 0.24, "grad_norm": 0.9722650051116943, "learning_rate": 4.3224215685535294e-05, "loss": 0.7418, "num_input_tokens_seen": 8027824, "step": 1500 }, { "epoch": 0.2408, "grad_norm": 0.599818229675293, "learning_rate": 4.31811468106099e-05, "loss": 0.6157, "num_input_tokens_seen": 8058528, "step": 1505 }, { "epoch": 0.2416, "grad_norm": 1.0976861715316772, "learning_rate": 4.3137963093679945e-05, "loss": 0.6302, "num_input_tokens_seen": 8081984, "step": 1510 }, { "epoch": 0.2424, "grad_norm": 0.5699600577354431, "learning_rate": 4.309466480751726e-05, "loss": 0.628, "num_input_tokens_seen": 8113216, "step": 1515 }, { "epoch": 0.2432, "grad_norm": 0.8899049758911133, "learning_rate": 4.305125222561736e-05, "loss": 0.635, "num_input_tokens_seen": 8142080, "step": 1520 }, { "epoch": 0.244, "grad_norm": 0.9494242072105408, "learning_rate": 4.3007725622197674e-05, "loss": 0.8114, "num_input_tokens_seen": 8171008, "step": 1525 }, { "epoch": 0.2448, "grad_norm": 0.9237959384918213, "learning_rate": 4.296408527219592e-05, "loss": 0.6678, "num_input_tokens_seen": 8197696, "step": 1530 }, { "epoch": 0.2456, "grad_norm": 0.8756378889083862, "learning_rate": 4.292033145126825e-05, "loss": 0.8364, "num_input_tokens_seen": 8225552, "step": 1535 }, { "epoch": 0.2464, "grad_norm": 0.9631836414337158, "learning_rate": 4.287646443578758e-05, "loss": 0.7312, "num_input_tokens_seen": 8257120, "step": 1540 }, { "epoch": 0.2472, "grad_norm": 0.920713484287262, "learning_rate": 4.283248450284182e-05, "loss": 0.8067, "num_input_tokens_seen": 8282400, "step": 1545 }, { "epoch": 0.248, "grad_norm": 1.0773414373397827, "learning_rate": 4.2788391930232136e-05, "loss": 0.7109, "num_input_tokens_seen": 8309568, "step": 1550 }, { "epoch": 0.2488, "grad_norm": 0.5621623396873474, "learning_rate": 4.2744186996471174e-05, "loss": 0.6543, "num_input_tokens_seen": 8338864, "step": 1555 }, { "epoch": 0.2496, "grad_norm": 0.8737258315086365, "learning_rate": 4.269986998078132e-05, "loss": 0.7401, "num_input_tokens_seen": 8364592, "step": 1560 }, { "epoch": 0.2504, "grad_norm": 0.8454060554504395, "learning_rate": 4.265544116309294e-05, "loss": 0.7538, "num_input_tokens_seen": 8391120, "step": 1565 }, { "epoch": 0.2512, "grad_norm": 0.8107228875160217, "learning_rate": 4.261090082404258e-05, "loss": 0.7705, "num_input_tokens_seen": 8418320, "step": 1570 }, { "epoch": 0.252, "grad_norm": 0.7339603304862976, "learning_rate": 4.256624924497123e-05, "loss": 0.6846, "num_input_tokens_seen": 8446640, "step": 1575 }, { "epoch": 0.2528, "grad_norm": 1.0036543607711792, "learning_rate": 4.252148670792254e-05, "loss": 0.8502, "num_input_tokens_seen": 8470416, "step": 1580 }, { "epoch": 0.2536, "grad_norm": 0.8186982870101929, "learning_rate": 4.2476613495641026e-05, "loss": 0.6987, "num_input_tokens_seen": 8498160, "step": 1585 }, { "epoch": 0.2544, "grad_norm": 0.9724487066268921, "learning_rate": 4.2431629891570266e-05, "loss": 0.6461, "num_input_tokens_seen": 8525904, "step": 1590 }, { "epoch": 0.2552, "grad_norm": 0.5958553552627563, "learning_rate": 4.238653617985118e-05, "loss": 0.7143, "num_input_tokens_seen": 8551872, "step": 1595 }, { "epoch": 0.256, "grad_norm": 1.0192784070968628, "learning_rate": 4.234133264532012e-05, "loss": 0.7215, "num_input_tokens_seen": 8583440, "step": 1600 }, { "epoch": 0.2568, "grad_norm": 0.7806874513626099, "learning_rate": 4.229601957350722e-05, "loss": 0.8008, "num_input_tokens_seen": 8609632, "step": 1605 }, { "epoch": 0.2576, "grad_norm": 1.086475133895874, "learning_rate": 4.225059725063444e-05, "loss": 0.6612, "num_input_tokens_seen": 8633888, "step": 1610 }, { "epoch": 0.2584, "grad_norm": 0.6213988065719604, "learning_rate": 4.2205065963613864e-05, "loss": 0.7544, "num_input_tokens_seen": 8660288, "step": 1615 }, { "epoch": 0.2592, "grad_norm": 1.0608100891113281, "learning_rate": 4.2159426000045854e-05, "loss": 0.7569, "num_input_tokens_seen": 8689184, "step": 1620 }, { "epoch": 0.26, "grad_norm": 0.7601464986801147, "learning_rate": 4.211367764821722e-05, "loss": 0.8161, "num_input_tokens_seen": 8713504, "step": 1625 }, { "epoch": 0.2608, "grad_norm": 0.9310168623924255, "learning_rate": 4.206782119709942e-05, "loss": 0.8283, "num_input_tokens_seen": 8741088, "step": 1630 }, { "epoch": 0.2616, "grad_norm": 0.6408126354217529, "learning_rate": 4.20218569363467e-05, "loss": 0.5745, "num_input_tokens_seen": 8767456, "step": 1635 }, { "epoch": 0.2624, "grad_norm": 1.1697090864181519, "learning_rate": 4.197578515629435e-05, "loss": 0.7525, "num_input_tokens_seen": 8791952, "step": 1640 }, { "epoch": 0.2632, "grad_norm": 0.9160236716270447, "learning_rate": 4.192960614795675e-05, "loss": 0.7991, "num_input_tokens_seen": 8816080, "step": 1645 }, { "epoch": 0.264, "grad_norm": 1.0530091524124146, "learning_rate": 4.188332020302561e-05, "loss": 0.7297, "num_input_tokens_seen": 8841536, "step": 1650 }, { "epoch": 0.2648, "grad_norm": 0.8888834118843079, "learning_rate": 4.183692761386813e-05, "loss": 0.6276, "num_input_tokens_seen": 8869872, "step": 1655 }, { "epoch": 0.2656, "grad_norm": 0.6144154667854309, "learning_rate": 4.179042867352511e-05, "loss": 0.7127, "num_input_tokens_seen": 8893152, "step": 1660 }, { "epoch": 0.2664, "grad_norm": 0.814166247844696, "learning_rate": 4.174382367570912e-05, "loss": 0.7712, "num_input_tokens_seen": 8923040, "step": 1665 }, { "epoch": 0.2672, "grad_norm": 0.8960988521575928, "learning_rate": 4.169711291480266e-05, "loss": 0.8388, "num_input_tokens_seen": 8945856, "step": 1670 }, { "epoch": 0.268, "grad_norm": 0.8164514303207397, "learning_rate": 4.165029668585629e-05, "loss": 0.7538, "num_input_tokens_seen": 8971664, "step": 1675 }, { "epoch": 0.2688, "grad_norm": 0.8044324517250061, "learning_rate": 4.160337528458676e-05, "loss": 0.708, "num_input_tokens_seen": 8996064, "step": 1680 }, { "epoch": 0.2696, "grad_norm": 0.7704948782920837, "learning_rate": 4.155634900737513e-05, "loss": 0.668, "num_input_tokens_seen": 9022416, "step": 1685 }, { "epoch": 0.2704, "grad_norm": 0.8315603137016296, "learning_rate": 4.150921815126493e-05, "loss": 0.752, "num_input_tokens_seen": 9052480, "step": 1690 }, { "epoch": 0.2712, "grad_norm": 0.7516652345657349, "learning_rate": 4.1461983013960245e-05, "loss": 0.6534, "num_input_tokens_seen": 9079760, "step": 1695 }, { "epoch": 0.272, "grad_norm": 0.7449467182159424, "learning_rate": 4.1414643893823914e-05, "loss": 0.6808, "num_input_tokens_seen": 9109424, "step": 1700 }, { "epoch": 0.2728, "grad_norm": 0.6889111995697021, "learning_rate": 4.136720108987552e-05, "loss": 0.7627, "num_input_tokens_seen": 9132128, "step": 1705 }, { "epoch": 0.2736, "grad_norm": 0.9195050597190857, "learning_rate": 4.131965490178959e-05, "loss": 0.6527, "num_input_tokens_seen": 9160960, "step": 1710 }, { "epoch": 0.2744, "grad_norm": 0.9934877157211304, "learning_rate": 4.1272005629893714e-05, "loss": 0.7102, "num_input_tokens_seen": 9190992, "step": 1715 }, { "epoch": 0.2752, "grad_norm": 0.8816946148872375, "learning_rate": 4.122425357516658e-05, "loss": 0.67, "num_input_tokens_seen": 9218320, "step": 1720 }, { "epoch": 0.276, "grad_norm": 0.7904371619224548, "learning_rate": 4.1176399039236116e-05, "loss": 0.7159, "num_input_tokens_seen": 9246304, "step": 1725 }, { "epoch": 0.2768, "grad_norm": 0.795921266078949, "learning_rate": 4.112844232437757e-05, "loss": 0.8248, "num_input_tokens_seen": 9271856, "step": 1730 }, { "epoch": 0.2776, "grad_norm": 0.8109453320503235, "learning_rate": 4.108038373351163e-05, "loss": 0.7264, "num_input_tokens_seen": 9297152, "step": 1735 }, { "epoch": 0.2784, "grad_norm": 0.8012672066688538, "learning_rate": 4.1032223570202474e-05, "loss": 0.7368, "num_input_tokens_seen": 9326896, "step": 1740 }, { "epoch": 0.2792, "grad_norm": 0.8711723685264587, "learning_rate": 4.0983962138655873e-05, "loss": 0.6245, "num_input_tokens_seen": 9351680, "step": 1745 }, { "epoch": 0.28, "grad_norm": 1.034636378288269, "learning_rate": 4.093559974371725e-05, "loss": 0.8033, "num_input_tokens_seen": 9374896, "step": 1750 }, { "epoch": 0.2808, "grad_norm": 0.8999419808387756, "learning_rate": 4.088713669086977e-05, "loss": 0.6803, "num_input_tokens_seen": 9400592, "step": 1755 }, { "epoch": 0.2816, "grad_norm": 0.5961094498634338, "learning_rate": 4.083857328623243e-05, "loss": 0.7384, "num_input_tokens_seen": 9429280, "step": 1760 }, { "epoch": 0.2824, "grad_norm": 1.194028377532959, "learning_rate": 4.078990983655807e-05, "loss": 0.8149, "num_input_tokens_seen": 9454736, "step": 1765 }, { "epoch": 0.2832, "grad_norm": 0.904292643070221, "learning_rate": 4.0741146649231504e-05, "loss": 0.7243, "num_input_tokens_seen": 9479648, "step": 1770 }, { "epoch": 0.284, "grad_norm": 0.8501243591308594, "learning_rate": 4.0692284032267516e-05, "loss": 0.7639, "num_input_tokens_seen": 9504432, "step": 1775 }, { "epoch": 0.2848, "grad_norm": 1.0718458890914917, "learning_rate": 4.064332229430895e-05, "loss": 0.6857, "num_input_tokens_seen": 9528880, "step": 1780 }, { "epoch": 0.2856, "grad_norm": 0.7065584063529968, "learning_rate": 4.059426174462476e-05, "loss": 0.69, "num_input_tokens_seen": 9557360, "step": 1785 }, { "epoch": 0.2864, "grad_norm": 1.0800750255584717, "learning_rate": 4.054510269310803e-05, "loss": 0.704, "num_input_tokens_seen": 9580608, "step": 1790 }, { "epoch": 0.2872, "grad_norm": 0.5907096862792969, "learning_rate": 4.0495845450274064e-05, "loss": 0.8015, "num_input_tokens_seen": 9611376, "step": 1795 }, { "epoch": 0.288, "grad_norm": 0.9455146789550781, "learning_rate": 4.044649032725836e-05, "loss": 0.7382, "num_input_tokens_seen": 9640784, "step": 1800 }, { "epoch": 0.2888, "grad_norm": 1.0408939123153687, "learning_rate": 4.039703763581472e-05, "loss": 0.7299, "num_input_tokens_seen": 9667120, "step": 1805 }, { "epoch": 0.2896, "grad_norm": 0.8098856806755066, "learning_rate": 4.0347487688313194e-05, "loss": 0.6402, "num_input_tokens_seen": 9696832, "step": 1810 }, { "epoch": 0.2904, "grad_norm": 0.695599377155304, "learning_rate": 4.02978407977382e-05, "loss": 0.711, "num_input_tokens_seen": 9722080, "step": 1815 }, { "epoch": 0.2912, "grad_norm": 0.6605217456817627, "learning_rate": 4.024809727768648e-05, "loss": 0.6587, "num_input_tokens_seen": 9748096, "step": 1820 }, { "epoch": 0.292, "grad_norm": 0.9249849915504456, "learning_rate": 4.019825744236514e-05, "loss": 0.6656, "num_input_tokens_seen": 9774128, "step": 1825 }, { "epoch": 0.2928, "grad_norm": 0.8226694464683533, "learning_rate": 4.0148321606589656e-05, "loss": 0.7143, "num_input_tokens_seen": 9805488, "step": 1830 }, { "epoch": 0.2936, "grad_norm": 1.0425550937652588, "learning_rate": 4.009829008578192e-05, "loss": 0.6735, "num_input_tokens_seen": 9828480, "step": 1835 }, { "epoch": 0.2944, "grad_norm": 0.6911535263061523, "learning_rate": 4.0048163195968214e-05, "loss": 0.7395, "num_input_tokens_seen": 9863648, "step": 1840 }, { "epoch": 0.2952, "grad_norm": 0.8600900769233704, "learning_rate": 3.999794125377721e-05, "loss": 0.729, "num_input_tokens_seen": 9893184, "step": 1845 }, { "epoch": 0.296, "grad_norm": 1.009696364402771, "learning_rate": 3.9947624576437975e-05, "loss": 0.6565, "num_input_tokens_seen": 9922464, "step": 1850 }, { "epoch": 0.2968, "grad_norm": 0.916327178478241, "learning_rate": 3.9897213481778006e-05, "loss": 0.691, "num_input_tokens_seen": 9948384, "step": 1855 }, { "epoch": 0.2976, "grad_norm": 0.9392701387405396, "learning_rate": 3.984670828822118e-05, "loss": 0.7408, "num_input_tokens_seen": 9973760, "step": 1860 }, { "epoch": 0.2984, "grad_norm": 0.9044517278671265, "learning_rate": 3.979610931478574e-05, "loss": 0.761, "num_input_tokens_seen": 10001648, "step": 1865 }, { "epoch": 0.2992, "grad_norm": 0.9471223950386047, "learning_rate": 3.97454168810823e-05, "loss": 0.8524, "num_input_tokens_seen": 10024912, "step": 1870 }, { "epoch": 0.3, "grad_norm": 1.0985262393951416, "learning_rate": 3.969463130731183e-05, "loss": 0.7221, "num_input_tokens_seen": 10049872, "step": 1875 }, { "epoch": 0.3008, "grad_norm": 0.8284273147583008, "learning_rate": 3.964375291426361e-05, "loss": 0.7708, "num_input_tokens_seen": 10073568, "step": 1880 }, { "epoch": 0.3016, "grad_norm": 0.7012784481048584, "learning_rate": 3.959278202331322e-05, "loss": 0.6842, "num_input_tokens_seen": 10098448, "step": 1885 }, { "epoch": 0.3024, "grad_norm": 1.1056398153305054, "learning_rate": 3.954171895642052e-05, "loss": 0.772, "num_input_tokens_seen": 10123168, "step": 1890 }, { "epoch": 0.3032, "grad_norm": 1.0128076076507568, "learning_rate": 3.949056403612758e-05, "loss": 0.6993, "num_input_tokens_seen": 10149440, "step": 1895 }, { "epoch": 0.304, "grad_norm": 0.7793564796447754, "learning_rate": 3.943931758555669e-05, "loss": 0.7672, "num_input_tokens_seen": 10174496, "step": 1900 }, { "epoch": 0.3048, "grad_norm": 0.909677267074585, "learning_rate": 3.938797992840828e-05, "loss": 0.6716, "num_input_tokens_seen": 10199648, "step": 1905 }, { "epoch": 0.3056, "grad_norm": 0.8851680159568787, "learning_rate": 3.933655138895889e-05, "loss": 0.7062, "num_input_tokens_seen": 10221840, "step": 1910 }, { "epoch": 0.3064, "grad_norm": 0.9452556371688843, "learning_rate": 3.928503229205913e-05, "loss": 0.6748, "num_input_tokens_seen": 10247504, "step": 1915 }, { "epoch": 0.3072, "grad_norm": 0.8891339302062988, "learning_rate": 3.9233422963131616e-05, "loss": 0.6331, "num_input_tokens_seen": 10277984, "step": 1920 }, { "epoch": 0.308, "grad_norm": 0.9662081599235535, "learning_rate": 3.9181723728168916e-05, "loss": 0.779, "num_input_tokens_seen": 10300400, "step": 1925 }, { "epoch": 0.3088, "grad_norm": 0.9517924785614014, "learning_rate": 3.91299349137315e-05, "loss": 0.722, "num_input_tokens_seen": 10326672, "step": 1930 }, { "epoch": 0.3096, "grad_norm": 0.755901575088501, "learning_rate": 3.907805684694566e-05, "loss": 0.6321, "num_input_tokens_seen": 10356864, "step": 1935 }, { "epoch": 0.3104, "grad_norm": 0.8272456526756287, "learning_rate": 3.902608985550147e-05, "loss": 0.6077, "num_input_tokens_seen": 10388032, "step": 1940 }, { "epoch": 0.3112, "grad_norm": 1.138036847114563, "learning_rate": 3.897403426765069e-05, "loss": 0.6726, "num_input_tokens_seen": 10417152, "step": 1945 }, { "epoch": 0.312, "grad_norm": 0.8155280351638794, "learning_rate": 3.8921890412204705e-05, "loss": 0.741, "num_input_tokens_seen": 10448128, "step": 1950 }, { "epoch": 0.3128, "grad_norm": 0.7004032135009766, "learning_rate": 3.886965861853244e-05, "loss": 0.6555, "num_input_tokens_seen": 10474960, "step": 1955 }, { "epoch": 0.3136, "grad_norm": 0.9554680585861206, "learning_rate": 3.881733921655829e-05, "loss": 0.75, "num_input_tokens_seen": 10502848, "step": 1960 }, { "epoch": 0.3144, "grad_norm": 0.8525771498680115, "learning_rate": 3.876493253676004e-05, "loss": 0.7042, "num_input_tokens_seen": 10532640, "step": 1965 }, { "epoch": 0.3152, "grad_norm": 0.8739621043205261, "learning_rate": 3.871243891016676e-05, "loss": 0.6188, "num_input_tokens_seen": 10560096, "step": 1970 }, { "epoch": 0.316, "grad_norm": 0.9146223068237305, "learning_rate": 3.865985866835673e-05, "loss": 0.8165, "num_input_tokens_seen": 10585520, "step": 1975 }, { "epoch": 0.3168, "grad_norm": 1.1149648427963257, "learning_rate": 3.8607192143455326e-05, "loss": 0.7437, "num_input_tokens_seen": 10614560, "step": 1980 }, { "epoch": 0.3176, "grad_norm": 0.9382626414299011, "learning_rate": 3.8554439668132946e-05, "loss": 0.7758, "num_input_tokens_seen": 10637344, "step": 1985 }, { "epoch": 0.3184, "grad_norm": 0.9469596743583679, "learning_rate": 3.85016015756029e-05, "loss": 0.7593, "num_input_tokens_seen": 10663440, "step": 1990 }, { "epoch": 0.3192, "grad_norm": 0.8701977133750916, "learning_rate": 3.844867819961928e-05, "loss": 0.6535, "num_input_tokens_seen": 10693392, "step": 1995 }, { "epoch": 0.32, "grad_norm": 0.7110251188278198, "learning_rate": 3.8395669874474915e-05, "loss": 0.8263, "num_input_tokens_seen": 10719232, "step": 2000 }, { "epoch": 0.3208, "grad_norm": 0.8518005609512329, "learning_rate": 3.8342576934999184e-05, "loss": 0.7992, "num_input_tokens_seen": 10746560, "step": 2005 }, { "epoch": 0.3216, "grad_norm": 0.9604689478874207, "learning_rate": 3.828939971655595e-05, "loss": 0.7513, "num_input_tokens_seen": 10768512, "step": 2010 }, { "epoch": 0.3224, "grad_norm": 0.8639784455299377, "learning_rate": 3.8236138555041434e-05, "loss": 0.6775, "num_input_tokens_seen": 10803648, "step": 2015 }, { "epoch": 0.3232, "grad_norm": 0.8527281880378723, "learning_rate": 3.8182793786882065e-05, "loss": 0.7856, "num_input_tokens_seen": 10830640, "step": 2020 }, { "epoch": 0.324, "grad_norm": 0.7717742919921875, "learning_rate": 3.81293657490324e-05, "loss": 0.6793, "num_input_tokens_seen": 10860272, "step": 2025 }, { "epoch": 0.3248, "grad_norm": 0.6685821413993835, "learning_rate": 3.8075854778972955e-05, "loss": 0.6546, "num_input_tokens_seen": 10887664, "step": 2030 }, { "epoch": 0.3256, "grad_norm": 0.9813340306282043, "learning_rate": 3.802226121470811e-05, "loss": 0.6673, "num_input_tokens_seen": 10912000, "step": 2035 }, { "epoch": 0.3264, "grad_norm": 1.0419212579727173, "learning_rate": 3.796858539476394e-05, "loss": 0.6933, "num_input_tokens_seen": 10936704, "step": 2040 }, { "epoch": 0.3272, "grad_norm": 0.851434588432312, "learning_rate": 3.7914827658186103e-05, "loss": 0.6593, "num_input_tokens_seen": 10960464, "step": 2045 }, { "epoch": 0.328, "grad_norm": 0.7272098660469055, "learning_rate": 3.786098834453766e-05, "loss": 0.6246, "num_input_tokens_seen": 10989680, "step": 2050 }, { "epoch": 0.3288, "grad_norm": 0.7740225791931152, "learning_rate": 3.780706779389701e-05, "loss": 0.7029, "num_input_tokens_seen": 11014928, "step": 2055 }, { "epoch": 0.3296, "grad_norm": 0.963455080986023, "learning_rate": 3.775306634685562e-05, "loss": 0.7331, "num_input_tokens_seen": 11041920, "step": 2060 }, { "epoch": 0.3304, "grad_norm": 0.7765479683876038, "learning_rate": 3.7698984344515997e-05, "loss": 0.6624, "num_input_tokens_seen": 11070304, "step": 2065 }, { "epoch": 0.3312, "grad_norm": 0.8283601999282837, "learning_rate": 3.764482212848948e-05, "loss": 0.7505, "num_input_tokens_seen": 11099520, "step": 2070 }, { "epoch": 0.332, "grad_norm": 0.5610854029655457, "learning_rate": 3.759058004089402e-05, "loss": 0.6908, "num_input_tokens_seen": 11129008, "step": 2075 }, { "epoch": 0.3328, "grad_norm": 0.8462053537368774, "learning_rate": 3.753625842435216e-05, "loss": 0.7062, "num_input_tokens_seen": 11151600, "step": 2080 }, { "epoch": 0.3336, "grad_norm": 0.8926122188568115, "learning_rate": 3.748185762198873e-05, "loss": 0.7177, "num_input_tokens_seen": 11176784, "step": 2085 }, { "epoch": 0.3344, "grad_norm": 0.6711943745613098, "learning_rate": 3.742737797742878e-05, "loss": 0.7504, "num_input_tokens_seen": 11205008, "step": 2090 }, { "epoch": 0.3352, "grad_norm": 1.014253854751587, "learning_rate": 3.7372819834795335e-05, "loss": 0.7144, "num_input_tokens_seen": 11229872, "step": 2095 }, { "epoch": 0.336, "grad_norm": 0.7249706983566284, "learning_rate": 3.731818353870729e-05, "loss": 0.6876, "num_input_tokens_seen": 11253296, "step": 2100 }, { "epoch": 0.3368, "grad_norm": 0.8249915838241577, "learning_rate": 3.726346943427719e-05, "loss": 0.7102, "num_input_tokens_seen": 11279408, "step": 2105 }, { "epoch": 0.3376, "grad_norm": 1.027541995048523, "learning_rate": 3.720867786710904e-05, "loss": 0.7708, "num_input_tokens_seen": 11304176, "step": 2110 }, { "epoch": 0.3384, "grad_norm": 0.7004812955856323, "learning_rate": 3.7153809183296176e-05, "loss": 0.5882, "num_input_tokens_seen": 11330944, "step": 2115 }, { "epoch": 0.3392, "grad_norm": 1.1122636795043945, "learning_rate": 3.7098863729419e-05, "loss": 0.6127, "num_input_tokens_seen": 11354064, "step": 2120 }, { "epoch": 0.34, "grad_norm": 0.925553560256958, "learning_rate": 3.704384185254288e-05, "loss": 0.7732, "num_input_tokens_seen": 11376288, "step": 2125 }, { "epoch": 0.3408, "grad_norm": 0.6940233707427979, "learning_rate": 3.6988743900215894e-05, "loss": 0.7334, "num_input_tokens_seen": 11405472, "step": 2130 }, { "epoch": 0.3416, "grad_norm": 0.7634669542312622, "learning_rate": 3.693357022046665e-05, "loss": 0.8137, "num_input_tokens_seen": 11431552, "step": 2135 }, { "epoch": 0.3424, "grad_norm": 0.804530680179596, "learning_rate": 3.68783211618021e-05, "loss": 0.6987, "num_input_tokens_seen": 11459152, "step": 2140 }, { "epoch": 0.3432, "grad_norm": 1.1058536767959595, "learning_rate": 3.682299707320532e-05, "loss": 0.6614, "num_input_tokens_seen": 11487552, "step": 2145 }, { "epoch": 0.344, "grad_norm": 0.6808910369873047, "learning_rate": 3.6767598304133324e-05, "loss": 0.688, "num_input_tokens_seen": 11515792, "step": 2150 }, { "epoch": 0.3448, "grad_norm": 1.0619826316833496, "learning_rate": 3.671212520451484e-05, "loss": 0.7897, "num_input_tokens_seen": 11541280, "step": 2155 }, { "epoch": 0.3456, "grad_norm": 0.8404290080070496, "learning_rate": 3.665657812474812e-05, "loss": 0.7086, "num_input_tokens_seen": 11569440, "step": 2160 }, { "epoch": 0.3464, "grad_norm": 1.316372036933899, "learning_rate": 3.660095741569871e-05, "loss": 0.7421, "num_input_tokens_seen": 11597792, "step": 2165 }, { "epoch": 0.3472, "grad_norm": 0.7798539400100708, "learning_rate": 3.654526342869724e-05, "loss": 0.6954, "num_input_tokens_seen": 11622864, "step": 2170 }, { "epoch": 0.348, "grad_norm": 0.7101672887802124, "learning_rate": 3.6489496515537204e-05, "loss": 0.6764, "num_input_tokens_seen": 11651280, "step": 2175 }, { "epoch": 0.3488, "grad_norm": 0.8456715941429138, "learning_rate": 3.643365702847272e-05, "loss": 0.705, "num_input_tokens_seen": 11680048, "step": 2180 }, { "epoch": 0.3496, "grad_norm": 0.9790185689926147, "learning_rate": 3.6377745320216346e-05, "loss": 0.7433, "num_input_tokens_seen": 11702144, "step": 2185 }, { "epoch": 0.3504, "grad_norm": 0.9205552935600281, "learning_rate": 3.632176174393682e-05, "loss": 0.653, "num_input_tokens_seen": 11728816, "step": 2190 }, { "epoch": 0.3512, "grad_norm": 0.8499376177787781, "learning_rate": 3.626570665325684e-05, "loss": 0.6381, "num_input_tokens_seen": 11756688, "step": 2195 }, { "epoch": 0.352, "grad_norm": 0.7778225541114807, "learning_rate": 3.6209580402250815e-05, "loss": 0.7347, "num_input_tokens_seen": 11781664, "step": 2200 }, { "epoch": 0.3528, "grad_norm": 0.8913766145706177, "learning_rate": 3.615338334544265e-05, "loss": 0.8036, "num_input_tokens_seen": 11808352, "step": 2205 }, { "epoch": 0.3536, "grad_norm": 1.0191758871078491, "learning_rate": 3.6097115837803505e-05, "loss": 0.7486, "num_input_tokens_seen": 11836400, "step": 2210 }, { "epoch": 0.3544, "grad_norm": 0.7858436703681946, "learning_rate": 3.604077823474954e-05, "loss": 0.7885, "num_input_tokens_seen": 11862608, "step": 2215 }, { "epoch": 0.3552, "grad_norm": 0.6349871158599854, "learning_rate": 3.5984370892139666e-05, "loss": 0.7005, "num_input_tokens_seen": 11886528, "step": 2220 }, { "epoch": 0.356, "grad_norm": 0.8877844214439392, "learning_rate": 3.592789416627332e-05, "loss": 0.607, "num_input_tokens_seen": 11915040, "step": 2225 }, { "epoch": 0.3568, "grad_norm": 1.1504970788955688, "learning_rate": 3.5871348413888204e-05, "loss": 0.6723, "num_input_tokens_seen": 11942768, "step": 2230 }, { "epoch": 0.3576, "grad_norm": 0.7394466400146484, "learning_rate": 3.581473399215802e-05, "loss": 0.7302, "num_input_tokens_seen": 11978464, "step": 2235 }, { "epoch": 0.3584, "grad_norm": 1.1570250988006592, "learning_rate": 3.575805125869022e-05, "loss": 0.6867, "num_input_tokens_seen": 12001392, "step": 2240 }, { "epoch": 0.3592, "grad_norm": 0.8141620755195618, "learning_rate": 3.5701300571523755e-05, "loss": 0.7346, "num_input_tokens_seen": 12030352, "step": 2245 }, { "epoch": 0.36, "grad_norm": 0.8653257489204407, "learning_rate": 3.564448228912682e-05, "loss": 0.6381, "num_input_tokens_seen": 12062384, "step": 2250 }, { "epoch": 0.3608, "grad_norm": 0.8065868020057678, "learning_rate": 3.558759677039455e-05, "loss": 0.7679, "num_input_tokens_seen": 12089408, "step": 2255 }, { "epoch": 0.3616, "grad_norm": 0.7610428929328918, "learning_rate": 3.5530644374646815e-05, "loss": 0.668, "num_input_tokens_seen": 12114656, "step": 2260 }, { "epoch": 0.3624, "grad_norm": 0.8063391447067261, "learning_rate": 3.547362546162588e-05, "loss": 0.7454, "num_input_tokens_seen": 12144832, "step": 2265 }, { "epoch": 0.3632, "grad_norm": 1.0300970077514648, "learning_rate": 3.54165403914942e-05, "loss": 0.7513, "num_input_tokens_seen": 12170096, "step": 2270 }, { "epoch": 0.364, "grad_norm": 1.1293412446975708, "learning_rate": 3.535938952483211e-05, "loss": 0.7881, "num_input_tokens_seen": 12191104, "step": 2275 }, { "epoch": 0.3648, "grad_norm": 0.8911874294281006, "learning_rate": 3.5302173222635524e-05, "loss": 0.7253, "num_input_tokens_seen": 12214416, "step": 2280 }, { "epoch": 0.3656, "grad_norm": 1.0665303468704224, "learning_rate": 3.5244891846313736e-05, "loss": 0.8122, "num_input_tokens_seen": 12241344, "step": 2285 }, { "epoch": 0.3664, "grad_norm": 0.6204916834831238, "learning_rate": 3.5187545757687015e-05, "loss": 0.6188, "num_input_tokens_seen": 12269376, "step": 2290 }, { "epoch": 0.3672, "grad_norm": 0.7871102094650269, "learning_rate": 3.5130135318984456e-05, "loss": 0.7138, "num_input_tokens_seen": 12294960, "step": 2295 }, { "epoch": 0.368, "grad_norm": 0.7584692239761353, "learning_rate": 3.507266089284157e-05, "loss": 0.7425, "num_input_tokens_seen": 12318864, "step": 2300 }, { "epoch": 0.3688, "grad_norm": 0.6678550839424133, "learning_rate": 3.501512284229807e-05, "loss": 0.7238, "num_input_tokens_seen": 12345520, "step": 2305 }, { "epoch": 0.3696, "grad_norm": 0.9825206398963928, "learning_rate": 3.495752153079557e-05, "loss": 0.684, "num_input_tokens_seen": 12369600, "step": 2310 }, { "epoch": 0.3704, "grad_norm": 0.8038123250007629, "learning_rate": 3.489985732217525e-05, "loss": 0.707, "num_input_tokens_seen": 12394400, "step": 2315 }, { "epoch": 0.3712, "grad_norm": 1.158873438835144, "learning_rate": 3.484213058067559e-05, "loss": 0.5843, "num_input_tokens_seen": 12420848, "step": 2320 }, { "epoch": 0.372, "grad_norm": 0.8114385604858398, "learning_rate": 3.4784341670930065e-05, "loss": 0.7014, "num_input_tokens_seen": 12446192, "step": 2325 }, { "epoch": 0.3728, "grad_norm": 0.8132364749908447, "learning_rate": 3.4726490957964834e-05, "loss": 0.777, "num_input_tokens_seen": 12472960, "step": 2330 }, { "epoch": 0.3736, "grad_norm": 0.7918152213096619, "learning_rate": 3.466857880719645e-05, "loss": 0.6856, "num_input_tokens_seen": 12504256, "step": 2335 }, { "epoch": 0.3744, "grad_norm": 0.8399984240531921, "learning_rate": 3.461060558442952e-05, "loss": 0.7742, "num_input_tokens_seen": 12529872, "step": 2340 }, { "epoch": 0.3752, "grad_norm": 1.0398231744766235, "learning_rate": 3.455257165585444e-05, "loss": 0.6815, "num_input_tokens_seen": 12552368, "step": 2345 }, { "epoch": 0.376, "grad_norm": 0.9708042144775391, "learning_rate": 3.4494477388045035e-05, "loss": 0.677, "num_input_tokens_seen": 12576720, "step": 2350 }, { "epoch": 0.3768, "grad_norm": 0.8928380012512207, "learning_rate": 3.443632314795627e-05, "loss": 0.6239, "num_input_tokens_seen": 12606096, "step": 2355 }, { "epoch": 0.3776, "grad_norm": 1.3437156677246094, "learning_rate": 3.437810930292195e-05, "loss": 0.7379, "num_input_tokens_seen": 12631376, "step": 2360 }, { "epoch": 0.3784, "grad_norm": 0.9309334754943848, "learning_rate": 3.4319836220652335e-05, "loss": 0.7315, "num_input_tokens_seen": 12662096, "step": 2365 }, { "epoch": 0.3792, "grad_norm": 1.4636520147323608, "learning_rate": 3.4261504269231904e-05, "loss": 0.7738, "num_input_tokens_seen": 12691696, "step": 2370 }, { "epoch": 0.38, "grad_norm": 0.8436228632926941, "learning_rate": 3.4203113817116957e-05, "loss": 0.7307, "num_input_tokens_seen": 12718368, "step": 2375 }, { "epoch": 0.3808, "grad_norm": 0.877709150314331, "learning_rate": 3.414466523313332e-05, "loss": 0.7119, "num_input_tokens_seen": 12743664, "step": 2380 }, { "epoch": 0.3816, "grad_norm": 1.2288016080856323, "learning_rate": 3.408615888647402e-05, "loss": 0.781, "num_input_tokens_seen": 12775088, "step": 2385 }, { "epoch": 0.3824, "grad_norm": 0.8335594534873962, "learning_rate": 3.402759514669694e-05, "loss": 0.6256, "num_input_tokens_seen": 12802576, "step": 2390 }, { "epoch": 0.3832, "grad_norm": 1.0417710542678833, "learning_rate": 3.3968974383722495e-05, "loss": 0.7672, "num_input_tokens_seen": 12831280, "step": 2395 }, { "epoch": 0.384, "grad_norm": 1.1079373359680176, "learning_rate": 3.3910296967831266e-05, "loss": 0.7665, "num_input_tokens_seen": 12853744, "step": 2400 }, { "epoch": 0.3848, "grad_norm": 0.870614230632782, "learning_rate": 3.3851563269661726e-05, "loss": 0.6321, "num_input_tokens_seen": 12883408, "step": 2405 }, { "epoch": 0.3856, "grad_norm": 1.090280294418335, "learning_rate": 3.379277366020782e-05, "loss": 0.7086, "num_input_tokens_seen": 12914592, "step": 2410 }, { "epoch": 0.3864, "grad_norm": 0.8816367983818054, "learning_rate": 3.373392851081668e-05, "loss": 0.7712, "num_input_tokens_seen": 12936832, "step": 2415 }, { "epoch": 0.3872, "grad_norm": 0.8722823858261108, "learning_rate": 3.367502819318624e-05, "loss": 0.6844, "num_input_tokens_seen": 12962864, "step": 2420 }, { "epoch": 0.388, "grad_norm": 0.9704541563987732, "learning_rate": 3.3616073079362926e-05, "loss": 0.6877, "num_input_tokens_seen": 12992560, "step": 2425 }, { "epoch": 0.3888, "grad_norm": 0.8094004988670349, "learning_rate": 3.355706354173928e-05, "loss": 0.8139, "num_input_tokens_seen": 13015440, "step": 2430 }, { "epoch": 0.3896, "grad_norm": 0.8286037445068359, "learning_rate": 3.349799995305162e-05, "loss": 0.6696, "num_input_tokens_seen": 13039008, "step": 2435 }, { "epoch": 0.3904, "grad_norm": 0.985637366771698, "learning_rate": 3.343888268637765e-05, "loss": 0.7001, "num_input_tokens_seen": 13067648, "step": 2440 }, { "epoch": 0.3912, "grad_norm": 0.8938013911247253, "learning_rate": 3.337971211513417e-05, "loss": 0.8036, "num_input_tokens_seen": 13090064, "step": 2445 }, { "epoch": 0.392, "grad_norm": 0.7293727397918701, "learning_rate": 3.332048861307467e-05, "loss": 0.7405, "num_input_tokens_seen": 13119856, "step": 2450 }, { "epoch": 0.3928, "grad_norm": 0.5999038219451904, "learning_rate": 3.3261212554286975e-05, "loss": 0.6975, "num_input_tokens_seen": 13148288, "step": 2455 }, { "epoch": 0.3936, "grad_norm": 0.8091318607330322, "learning_rate": 3.320188431319088e-05, "loss": 0.6809, "num_input_tokens_seen": 13175616, "step": 2460 }, { "epoch": 0.3944, "grad_norm": 1.0293824672698975, "learning_rate": 3.3142504264535804e-05, "loss": 0.7749, "num_input_tokens_seen": 13199280, "step": 2465 }, { "epoch": 0.3952, "grad_norm": 0.793485701084137, "learning_rate": 3.3083072783398416e-05, "loss": 0.6873, "num_input_tokens_seen": 13224640, "step": 2470 }, { "epoch": 0.396, "grad_norm": 0.8636240363121033, "learning_rate": 3.302359024518024e-05, "loss": 0.7554, "num_input_tokens_seen": 13250448, "step": 2475 }, { "epoch": 0.3968, "grad_norm": 0.9471914172172546, "learning_rate": 3.296405702560532e-05, "loss": 0.7112, "num_input_tokens_seen": 13273472, "step": 2480 }, { "epoch": 0.3976, "grad_norm": 1.1579172611236572, "learning_rate": 3.2904473500717824e-05, "loss": 0.8207, "num_input_tokens_seen": 13300608, "step": 2485 }, { "epoch": 0.3984, "grad_norm": 1.022197961807251, "learning_rate": 3.2844840046879686e-05, "loss": 0.693, "num_input_tokens_seen": 13326976, "step": 2490 }, { "epoch": 0.3992, "grad_norm": 0.7574387788772583, "learning_rate": 3.278515704076821e-05, "loss": 0.6826, "num_input_tokens_seen": 13358528, "step": 2495 }, { "epoch": 0.4, "grad_norm": 0.7097072005271912, "learning_rate": 3.272542485937369e-05, "loss": 0.6714, "num_input_tokens_seen": 13384096, "step": 2500 }, { "epoch": 0.4008, "grad_norm": 0.8780053853988647, "learning_rate": 3.2665643879997056e-05, "loss": 0.7387, "num_input_tokens_seen": 13417120, "step": 2505 }, { "epoch": 0.4016, "grad_norm": 0.8968010544776917, "learning_rate": 3.260581448024745e-05, "loss": 0.6875, "num_input_tokens_seen": 13444832, "step": 2510 }, { "epoch": 0.4024, "grad_norm": 0.9647771716117859, "learning_rate": 3.25459370380399e-05, "loss": 0.834, "num_input_tokens_seen": 13472304, "step": 2515 }, { "epoch": 0.4032, "grad_norm": 0.9738301038742065, "learning_rate": 3.248601193159287e-05, "loss": 0.7144, "num_input_tokens_seen": 13495984, "step": 2520 }, { "epoch": 0.404, "grad_norm": 1.03775155544281, "learning_rate": 3.2426039539425876e-05, "loss": 0.7171, "num_input_tokens_seen": 13523360, "step": 2525 }, { "epoch": 0.4048, "grad_norm": 1.3964909315109253, "learning_rate": 3.236602024035716e-05, "loss": 0.7197, "num_input_tokens_seen": 13550016, "step": 2530 }, { "epoch": 0.4056, "grad_norm": 1.0805152654647827, "learning_rate": 3.230595441350125e-05, "loss": 0.7997, "num_input_tokens_seen": 13575088, "step": 2535 }, { "epoch": 0.4064, "grad_norm": 0.9613687992095947, "learning_rate": 3.2245842438266526e-05, "loss": 0.7847, "num_input_tokens_seen": 13600832, "step": 2540 }, { "epoch": 0.4072, "grad_norm": 0.9843304753303528, "learning_rate": 3.2185684694352916e-05, "loss": 0.7213, "num_input_tokens_seen": 13627328, "step": 2545 }, { "epoch": 0.408, "grad_norm": 0.7906083464622498, "learning_rate": 3.21254815617494e-05, "loss": 0.633, "num_input_tokens_seen": 13651664, "step": 2550 }, { "epoch": 0.4088, "grad_norm": 0.788149893283844, "learning_rate": 3.206523342073172e-05, "loss": 0.7512, "num_input_tokens_seen": 13677248, "step": 2555 }, { "epoch": 0.4096, "grad_norm": 0.7680060863494873, "learning_rate": 3.2004940651859844e-05, "loss": 0.703, "num_input_tokens_seen": 13705904, "step": 2560 }, { "epoch": 0.4104, "grad_norm": 0.8078610301017761, "learning_rate": 3.194460363597569e-05, "loss": 0.7212, "num_input_tokens_seen": 13731520, "step": 2565 }, { "epoch": 0.4112, "grad_norm": 1.2152231931686401, "learning_rate": 3.1884222754200625e-05, "loss": 0.7009, "num_input_tokens_seen": 13753840, "step": 2570 }, { "epoch": 0.412, "grad_norm": 0.8687548637390137, "learning_rate": 3.1823798387933134e-05, "loss": 0.718, "num_input_tokens_seen": 13777504, "step": 2575 }, { "epoch": 0.4128, "grad_norm": 1.1128169298171997, "learning_rate": 3.176333091884635e-05, "loss": 0.6796, "num_input_tokens_seen": 13805392, "step": 2580 }, { "epoch": 0.4136, "grad_norm": 0.6620244383811951, "learning_rate": 3.170282072888566e-05, "loss": 0.6632, "num_input_tokens_seen": 13835600, "step": 2585 }, { "epoch": 0.4144, "grad_norm": 1.0803226232528687, "learning_rate": 3.1642268200266317e-05, "loss": 0.743, "num_input_tokens_seen": 13862528, "step": 2590 }, { "epoch": 0.4152, "grad_norm": 0.8314620852470398, "learning_rate": 3.1581673715471006e-05, "loss": 0.7091, "num_input_tokens_seen": 13890272, "step": 2595 }, { "epoch": 0.416, "grad_norm": 1.0047166347503662, "learning_rate": 3.152103765724743e-05, "loss": 0.8011, "num_input_tokens_seen": 13913328, "step": 2600 }, { "epoch": 0.4168, "grad_norm": 0.9856431484222412, "learning_rate": 3.1460360408605866e-05, "loss": 0.7569, "num_input_tokens_seen": 13943040, "step": 2605 }, { "epoch": 0.4176, "grad_norm": 0.8467027544975281, "learning_rate": 3.139964235281682e-05, "loss": 0.6976, "num_input_tokens_seen": 13971872, "step": 2610 }, { "epoch": 0.4184, "grad_norm": 1.2195795774459839, "learning_rate": 3.1338883873408516e-05, "loss": 0.7039, "num_input_tokens_seen": 13997456, "step": 2615 }, { "epoch": 0.4192, "grad_norm": 0.832929253578186, "learning_rate": 3.127808535416454e-05, "loss": 0.7153, "num_input_tokens_seen": 14024656, "step": 2620 }, { "epoch": 0.42, "grad_norm": 0.8261767625808716, "learning_rate": 3.121724717912138e-05, "loss": 0.7317, "num_input_tokens_seen": 14053680, "step": 2625 }, { "epoch": 0.4208, "grad_norm": 0.8690986633300781, "learning_rate": 3.1156369732566006e-05, "loss": 0.6991, "num_input_tokens_seen": 14080096, "step": 2630 }, { "epoch": 0.4216, "grad_norm": 1.041561484336853, "learning_rate": 3.1095453399033466e-05, "loss": 0.7442, "num_input_tokens_seen": 14108080, "step": 2635 }, { "epoch": 0.4224, "grad_norm": 1.1139183044433594, "learning_rate": 3.103449856330443e-05, "loss": 0.7026, "num_input_tokens_seen": 14132448, "step": 2640 }, { "epoch": 0.4232, "grad_norm": 0.9388411045074463, "learning_rate": 3.0973505610402765e-05, "loss": 0.6425, "num_input_tokens_seen": 14157312, "step": 2645 }, { "epoch": 0.424, "grad_norm": 0.8923696279525757, "learning_rate": 3.091247492559312e-05, "loss": 0.7421, "num_input_tokens_seen": 14184288, "step": 2650 }, { "epoch": 0.4248, "grad_norm": 0.9683478474617004, "learning_rate": 3.085140689437846e-05, "loss": 0.7044, "num_input_tokens_seen": 14207920, "step": 2655 }, { "epoch": 0.4256, "grad_norm": 0.7942652106285095, "learning_rate": 3.0790301902497666e-05, "loss": 0.6892, "num_input_tokens_seen": 14235504, "step": 2660 }, { "epoch": 0.4264, "grad_norm": 0.9955897331237793, "learning_rate": 3.072916033592307e-05, "loss": 0.6595, "num_input_tokens_seen": 14259280, "step": 2665 }, { "epoch": 0.4272, "grad_norm": 0.9912785291671753, "learning_rate": 3.0667982580858044e-05, "loss": 0.6948, "num_input_tokens_seen": 14286592, "step": 2670 }, { "epoch": 0.428, "grad_norm": 1.352742314338684, "learning_rate": 3.0606769023734536e-05, "loss": 0.7009, "num_input_tokens_seen": 14309280, "step": 2675 }, { "epoch": 0.4288, "grad_norm": 1.183185338973999, "learning_rate": 3.054552005121064e-05, "loss": 0.6814, "num_input_tokens_seen": 14335984, "step": 2680 }, { "epoch": 0.4296, "grad_norm": 1.2679824829101562, "learning_rate": 3.0484236050168153e-05, "loss": 0.7468, "num_input_tokens_seen": 14361024, "step": 2685 }, { "epoch": 0.4304, "grad_norm": 1.1353107690811157, "learning_rate": 3.0422917407710137e-05, "loss": 0.629, "num_input_tokens_seen": 14391440, "step": 2690 }, { "epoch": 0.4312, "grad_norm": 1.1603094339370728, "learning_rate": 3.0361564511158457e-05, "loss": 0.7106, "num_input_tokens_seen": 14417952, "step": 2695 }, { "epoch": 0.432, "grad_norm": 0.9477285146713257, "learning_rate": 3.0300177748051373e-05, "loss": 0.7136, "num_input_tokens_seen": 14446752, "step": 2700 }, { "epoch": 0.4328, "grad_norm": 0.9295204281806946, "learning_rate": 3.0238757506141012e-05, "loss": 0.6269, "num_input_tokens_seen": 14475280, "step": 2705 }, { "epoch": 0.4336, "grad_norm": 0.8617603182792664, "learning_rate": 3.0177304173391037e-05, "loss": 0.6517, "num_input_tokens_seen": 14498112, "step": 2710 }, { "epoch": 0.4344, "grad_norm": 0.962295413017273, "learning_rate": 3.0115818137974067e-05, "loss": 0.6903, "num_input_tokens_seen": 14525664, "step": 2715 }, { "epoch": 0.4352, "grad_norm": 0.7317754030227661, "learning_rate": 3.005429978826934e-05, "loss": 0.7302, "num_input_tokens_seen": 14551536, "step": 2720 }, { "epoch": 0.436, "grad_norm": 0.9604383111000061, "learning_rate": 2.9992749512860173e-05, "loss": 0.7126, "num_input_tokens_seen": 14574560, "step": 2725 }, { "epoch": 0.4368, "grad_norm": 0.9363977313041687, "learning_rate": 2.9931167700531578e-05, "loss": 0.6701, "num_input_tokens_seen": 14602384, "step": 2730 }, { "epoch": 0.4376, "grad_norm": 1.0513427257537842, "learning_rate": 2.9869554740267724e-05, "loss": 0.5816, "num_input_tokens_seen": 14633728, "step": 2735 }, { "epoch": 0.4384, "grad_norm": 1.0142287015914917, "learning_rate": 2.9807911021249573e-05, "loss": 0.7965, "num_input_tokens_seen": 14662752, "step": 2740 }, { "epoch": 0.4392, "grad_norm": 0.8593106269836426, "learning_rate": 2.9746236932852355e-05, "loss": 0.6396, "num_input_tokens_seen": 14690896, "step": 2745 }, { "epoch": 0.44, "grad_norm": 0.912413477897644, "learning_rate": 2.9684532864643122e-05, "loss": 0.6914, "num_input_tokens_seen": 14717680, "step": 2750 }, { "epoch": 0.4408, "grad_norm": 1.1753630638122559, "learning_rate": 2.9622799206378305e-05, "loss": 0.7188, "num_input_tokens_seen": 14744176, "step": 2755 }, { "epoch": 0.4416, "grad_norm": 1.0383411645889282, "learning_rate": 2.956103634800126e-05, "loss": 0.6936, "num_input_tokens_seen": 14772464, "step": 2760 }, { "epoch": 0.4424, "grad_norm": 0.8875827789306641, "learning_rate": 2.949924467963975e-05, "loss": 0.709, "num_input_tokens_seen": 14800896, "step": 2765 }, { "epoch": 0.4432, "grad_norm": 1.0359493494033813, "learning_rate": 2.943742459160354e-05, "loss": 0.6361, "num_input_tokens_seen": 14826624, "step": 2770 }, { "epoch": 0.444, "grad_norm": 0.7070389986038208, "learning_rate": 2.9375576474381905e-05, "loss": 0.6062, "num_input_tokens_seen": 14859392, "step": 2775 }, { "epoch": 0.4448, "grad_norm": 1.0716419219970703, "learning_rate": 2.9313700718641167e-05, "loss": 0.7882, "num_input_tokens_seen": 14882336, "step": 2780 }, { "epoch": 0.4456, "grad_norm": 0.8054667115211487, "learning_rate": 2.925179771522223e-05, "loss": 0.7978, "num_input_tokens_seen": 14911312, "step": 2785 }, { "epoch": 0.4464, "grad_norm": 0.9502078294754028, "learning_rate": 2.9189867855138103e-05, "loss": 0.6835, "num_input_tokens_seen": 14938400, "step": 2790 }, { "epoch": 0.4472, "grad_norm": 0.8377355933189392, "learning_rate": 2.912791152957145e-05, "loss": 0.6566, "num_input_tokens_seen": 14965424, "step": 2795 }, { "epoch": 0.448, "grad_norm": 0.8674115538597107, "learning_rate": 2.9065929129872094e-05, "loss": 0.6616, "num_input_tokens_seen": 14994368, "step": 2800 }, { "epoch": 0.4488, "grad_norm": 0.9967759251594543, "learning_rate": 2.900392104755455e-05, "loss": 0.8051, "num_input_tokens_seen": 15018480, "step": 2805 }, { "epoch": 0.4496, "grad_norm": 1.04585862159729, "learning_rate": 2.894188767429557e-05, "loss": 0.6961, "num_input_tokens_seen": 15045840, "step": 2810 }, { "epoch": 0.4504, "grad_norm": 0.8065064549446106, "learning_rate": 2.8879829401931652e-05, "loss": 0.6898, "num_input_tokens_seen": 15070832, "step": 2815 }, { "epoch": 0.4512, "grad_norm": 0.7077392935752869, "learning_rate": 2.881774662245658e-05, "loss": 0.6789, "num_input_tokens_seen": 15097008, "step": 2820 }, { "epoch": 0.452, "grad_norm": 1.068467378616333, "learning_rate": 2.875563972801893e-05, "loss": 0.6684, "num_input_tokens_seen": 15120080, "step": 2825 }, { "epoch": 0.4528, "grad_norm": 0.7860395312309265, "learning_rate": 2.8693509110919598e-05, "loss": 0.6577, "num_input_tokens_seen": 15144976, "step": 2830 }, { "epoch": 0.4536, "grad_norm": 0.86238032579422, "learning_rate": 2.863135516360932e-05, "loss": 0.7893, "num_input_tokens_seen": 15174640, "step": 2835 }, { "epoch": 0.4544, "grad_norm": 0.8910583257675171, "learning_rate": 2.856917827868622e-05, "loss": 0.7377, "num_input_tokens_seen": 15198128, "step": 2840 }, { "epoch": 0.4552, "grad_norm": 0.9576541781425476, "learning_rate": 2.8506978848893302e-05, "loss": 0.821, "num_input_tokens_seen": 15222224, "step": 2845 }, { "epoch": 0.456, "grad_norm": 1.114388108253479, "learning_rate": 2.844475726711595e-05, "loss": 0.695, "num_input_tokens_seen": 15246640, "step": 2850 }, { "epoch": 0.4568, "grad_norm": 0.8997412323951721, "learning_rate": 2.8382513926379504e-05, "loss": 0.7175, "num_input_tokens_seen": 15277728, "step": 2855 }, { "epoch": 0.4576, "grad_norm": 1.1595414876937866, "learning_rate": 2.832024921984674e-05, "loss": 0.6505, "num_input_tokens_seen": 15307040, "step": 2860 }, { "epoch": 0.4584, "grad_norm": 0.7592776417732239, "learning_rate": 2.825796354081537e-05, "loss": 0.6686, "num_input_tokens_seen": 15334176, "step": 2865 }, { "epoch": 0.4592, "grad_norm": 1.0087366104125977, "learning_rate": 2.8195657282715594e-05, "loss": 0.6365, "num_input_tokens_seen": 15360496, "step": 2870 }, { "epoch": 0.46, "grad_norm": 0.9191427826881409, "learning_rate": 2.8133330839107608e-05, "loss": 0.6518, "num_input_tokens_seen": 15381328, "step": 2875 }, { "epoch": 0.4608, "grad_norm": 1.0468344688415527, "learning_rate": 2.8070984603679107e-05, "loss": 0.6262, "num_input_tokens_seen": 15409936, "step": 2880 }, { "epoch": 0.4616, "grad_norm": 1.1070493459701538, "learning_rate": 2.800861897024279e-05, "loss": 0.6684, "num_input_tokens_seen": 15436848, "step": 2885 }, { "epoch": 0.4624, "grad_norm": 1.3349978923797607, "learning_rate": 2.79462343327339e-05, "loss": 0.6978, "num_input_tokens_seen": 15463328, "step": 2890 }, { "epoch": 0.4632, "grad_norm": 0.7566163539886475, "learning_rate": 2.7883831085207707e-05, "loss": 0.7062, "num_input_tokens_seen": 15489232, "step": 2895 }, { "epoch": 0.464, "grad_norm": 0.7610609531402588, "learning_rate": 2.782140962183704e-05, "loss": 0.6642, "num_input_tokens_seen": 15516224, "step": 2900 }, { "epoch": 0.4648, "grad_norm": 0.7749585509300232, "learning_rate": 2.7758970336909795e-05, "loss": 0.6287, "num_input_tokens_seen": 15545584, "step": 2905 }, { "epoch": 0.4656, "grad_norm": 1.0202007293701172, "learning_rate": 2.769651362482642e-05, "loss": 0.6672, "num_input_tokens_seen": 15571216, "step": 2910 }, { "epoch": 0.4664, "grad_norm": 0.7980359792709351, "learning_rate": 2.763403988009746e-05, "loss": 0.737, "num_input_tokens_seen": 15597744, "step": 2915 }, { "epoch": 0.4672, "grad_norm": 0.942456841468811, "learning_rate": 2.7571549497341042e-05, "loss": 0.7715, "num_input_tokens_seen": 15622496, "step": 2920 }, { "epoch": 0.468, "grad_norm": 0.7782229781150818, "learning_rate": 2.7509042871280372e-05, "loss": 0.7435, "num_input_tokens_seen": 15647344, "step": 2925 }, { "epoch": 0.4688, "grad_norm": 1.0889509916305542, "learning_rate": 2.744652039674129e-05, "loss": 0.6946, "num_input_tokens_seen": 15672672, "step": 2930 }, { "epoch": 0.4696, "grad_norm": 1.0606461763381958, "learning_rate": 2.7383982468649714e-05, "loss": 0.7523, "num_input_tokens_seen": 15696144, "step": 2935 }, { "epoch": 0.4704, "grad_norm": 0.942613959312439, "learning_rate": 2.73214294820292e-05, "loss": 0.7218, "num_input_tokens_seen": 15723600, "step": 2940 }, { "epoch": 0.4712, "grad_norm": 0.8650354743003845, "learning_rate": 2.7258861831998388e-05, "loss": 0.6736, "num_input_tokens_seen": 15749680, "step": 2945 }, { "epoch": 0.472, "grad_norm": 0.8573226928710938, "learning_rate": 2.7196279913768584e-05, "loss": 0.6314, "num_input_tokens_seen": 15776768, "step": 2950 }, { "epoch": 0.4728, "grad_norm": 0.9692303538322449, "learning_rate": 2.713368412264118e-05, "loss": 0.7035, "num_input_tokens_seen": 15801376, "step": 2955 }, { "epoch": 0.4736, "grad_norm": 1.2111790180206299, "learning_rate": 2.707107485400521e-05, "loss": 0.6785, "num_input_tokens_seen": 15828416, "step": 2960 }, { "epoch": 0.4744, "grad_norm": 1.0816082954406738, "learning_rate": 2.7008452503334858e-05, "loss": 0.7672, "num_input_tokens_seen": 15852720, "step": 2965 }, { "epoch": 0.4752, "grad_norm": 0.980903148651123, "learning_rate": 2.6945817466186912e-05, "loss": 0.7723, "num_input_tokens_seen": 15880624, "step": 2970 }, { "epoch": 0.476, "grad_norm": 1.012623906135559, "learning_rate": 2.6883170138198323e-05, "loss": 0.6105, "num_input_tokens_seen": 15912176, "step": 2975 }, { "epoch": 0.4768, "grad_norm": 1.069486141204834, "learning_rate": 2.6820510915083648e-05, "loss": 0.6941, "num_input_tokens_seen": 15944384, "step": 2980 }, { "epoch": 0.4776, "grad_norm": 0.8411433100700378, "learning_rate": 2.6757840192632598e-05, "loss": 0.6669, "num_input_tokens_seen": 15969680, "step": 2985 }, { "epoch": 0.4784, "grad_norm": 0.6902319192886353, "learning_rate": 2.6695158366707522e-05, "loss": 0.6814, "num_input_tokens_seen": 15997264, "step": 2990 }, { "epoch": 0.4792, "grad_norm": 0.7844128012657166, "learning_rate": 2.6632465833240893e-05, "loss": 0.5641, "num_input_tokens_seen": 16029664, "step": 2995 }, { "epoch": 0.48, "grad_norm": 1.1021162271499634, "learning_rate": 2.656976298823284e-05, "loss": 0.7797, "num_input_tokens_seen": 16054864, "step": 3000 }, { "epoch": 0.4808, "grad_norm": 1.2080223560333252, "learning_rate": 2.650705022774859e-05, "loss": 0.6778, "num_input_tokens_seen": 16079552, "step": 3005 }, { "epoch": 0.4816, "grad_norm": 1.1015046834945679, "learning_rate": 2.6444327947916036e-05, "loss": 0.6806, "num_input_tokens_seen": 16105632, "step": 3010 }, { "epoch": 0.4824, "grad_norm": 1.0269590616226196, "learning_rate": 2.638159654492318e-05, "loss": 0.7589, "num_input_tokens_seen": 16134688, "step": 3015 }, { "epoch": 0.4832, "grad_norm": 0.8565163612365723, "learning_rate": 2.6318856415015664e-05, "loss": 0.677, "num_input_tokens_seen": 16163152, "step": 3020 }, { "epoch": 0.484, "grad_norm": 0.8519279956817627, "learning_rate": 2.6256107954494242e-05, "loss": 0.6136, "num_input_tokens_seen": 16189248, "step": 3025 }, { "epoch": 0.4848, "grad_norm": 0.9466794729232788, "learning_rate": 2.6193351559712292e-05, "loss": 0.6369, "num_input_tokens_seen": 16214832, "step": 3030 }, { "epoch": 0.4856, "grad_norm": 0.9867402911186218, "learning_rate": 2.6130587627073315e-05, "loss": 0.7202, "num_input_tokens_seen": 16244736, "step": 3035 }, { "epoch": 0.4864, "grad_norm": 0.8893384337425232, "learning_rate": 2.606781655302843e-05, "loss": 0.7057, "num_input_tokens_seen": 16272064, "step": 3040 }, { "epoch": 0.4872, "grad_norm": 1.2341115474700928, "learning_rate": 2.6005038734073833e-05, "loss": 0.6715, "num_input_tokens_seen": 16301344, "step": 3045 }, { "epoch": 0.488, "grad_norm": 1.0158292055130005, "learning_rate": 2.594225456674837e-05, "loss": 0.7479, "num_input_tokens_seen": 16325872, "step": 3050 }, { "epoch": 0.4888, "grad_norm": 0.9316710233688354, "learning_rate": 2.5879464447630946e-05, "loss": 0.6581, "num_input_tokens_seen": 16352272, "step": 3055 }, { "epoch": 0.4896, "grad_norm": 0.9104299545288086, "learning_rate": 2.5816668773338098e-05, "loss": 0.691, "num_input_tokens_seen": 16380464, "step": 3060 }, { "epoch": 0.4904, "grad_norm": 0.8835129737854004, "learning_rate": 2.575386794052142e-05, "loss": 0.6606, "num_input_tokens_seen": 16408736, "step": 3065 }, { "epoch": 0.4912, "grad_norm": 0.869504451751709, "learning_rate": 2.569106234586511e-05, "loss": 0.729, "num_input_tokens_seen": 16436352, "step": 3070 }, { "epoch": 0.492, "grad_norm": 1.0879722833633423, "learning_rate": 2.562825238608344e-05, "loss": 0.7137, "num_input_tokens_seen": 16464624, "step": 3075 }, { "epoch": 0.4928, "grad_norm": 0.9328833818435669, "learning_rate": 2.5565438457918244e-05, "loss": 0.7238, "num_input_tokens_seen": 16496720, "step": 3080 }, { "epoch": 0.4936, "grad_norm": 0.7433749437332153, "learning_rate": 2.5502620958136443e-05, "loss": 0.7019, "num_input_tokens_seen": 16524208, "step": 3085 }, { "epoch": 0.4944, "grad_norm": 0.7768478989601135, "learning_rate": 2.5439800283527494e-05, "loss": 0.5851, "num_input_tokens_seen": 16552192, "step": 3090 }, { "epoch": 0.4952, "grad_norm": 1.139993667602539, "learning_rate": 2.537697683090093e-05, "loss": 0.7357, "num_input_tokens_seen": 16578144, "step": 3095 }, { "epoch": 0.496, "grad_norm": 0.9104892611503601, "learning_rate": 2.531415099708382e-05, "loss": 0.6254, "num_input_tokens_seen": 16608288, "step": 3100 }, { "epoch": 0.4968, "grad_norm": 0.6912931799888611, "learning_rate": 2.5251323178918268e-05, "loss": 0.7284, "num_input_tokens_seen": 16636176, "step": 3105 }, { "epoch": 0.4976, "grad_norm": 0.8370018601417542, "learning_rate": 2.518849377325893e-05, "loss": 0.8136, "num_input_tokens_seen": 16659168, "step": 3110 }, { "epoch": 0.4984, "grad_norm": 1.049914836883545, "learning_rate": 2.5125663176970476e-05, "loss": 0.7334, "num_input_tokens_seen": 16687344, "step": 3115 }, { "epoch": 0.4992, "grad_norm": 1.0298138856887817, "learning_rate": 2.5062831786925102e-05, "loss": 0.7599, "num_input_tokens_seen": 16714496, "step": 3120 }, { "epoch": 0.5, "grad_norm": 0.7912611961364746, "learning_rate": 2.5e-05, "loss": 0.6517, "num_input_tokens_seen": 16742528, "step": 3125 }, { "epoch": 0.5008, "grad_norm": 1.0854965448379517, "learning_rate": 2.4937168213074907e-05, "loss": 0.6778, "num_input_tokens_seen": 16771248, "step": 3130 }, { "epoch": 0.5016, "grad_norm": 0.9674849510192871, "learning_rate": 2.4874336823029526e-05, "loss": 0.6847, "num_input_tokens_seen": 16799136, "step": 3135 }, { "epoch": 0.5024, "grad_norm": 0.8434900641441345, "learning_rate": 2.481150622674108e-05, "loss": 0.6638, "num_input_tokens_seen": 16825648, "step": 3140 }, { "epoch": 0.5032, "grad_norm": 0.8714620471000671, "learning_rate": 2.4748676821081738e-05, "loss": 0.7139, "num_input_tokens_seen": 16852240, "step": 3145 }, { "epoch": 0.504, "grad_norm": 0.8312164545059204, "learning_rate": 2.4685849002916183e-05, "loss": 0.7507, "num_input_tokens_seen": 16878624, "step": 3150 }, { "epoch": 0.5048, "grad_norm": 1.1353472471237183, "learning_rate": 2.4623023169099073e-05, "loss": 0.6951, "num_input_tokens_seen": 16906864, "step": 3155 }, { "epoch": 0.5056, "grad_norm": 1.1486365795135498, "learning_rate": 2.4560199716472508e-05, "loss": 0.733, "num_input_tokens_seen": 16930080, "step": 3160 }, { "epoch": 0.5064, "grad_norm": 0.9651095867156982, "learning_rate": 2.449737904186357e-05, "loss": 0.7517, "num_input_tokens_seen": 16952240, "step": 3165 }, { "epoch": 0.5072, "grad_norm": 0.8921483755111694, "learning_rate": 2.4434561542081762e-05, "loss": 0.7472, "num_input_tokens_seen": 16985408, "step": 3170 }, { "epoch": 0.508, "grad_norm": 1.0625066757202148, "learning_rate": 2.4371747613916566e-05, "loss": 0.7514, "num_input_tokens_seen": 17013776, "step": 3175 }, { "epoch": 0.5088, "grad_norm": 1.10313081741333, "learning_rate": 2.4308937654134893e-05, "loss": 0.7633, "num_input_tokens_seen": 17039120, "step": 3180 }, { "epoch": 0.5096, "grad_norm": 1.115670084953308, "learning_rate": 2.4246132059478578e-05, "loss": 0.6606, "num_input_tokens_seen": 17065296, "step": 3185 }, { "epoch": 0.5104, "grad_norm": 1.0417555570602417, "learning_rate": 2.418333122666191e-05, "loss": 0.764, "num_input_tokens_seen": 17089264, "step": 3190 }, { "epoch": 0.5112, "grad_norm": 0.9926926493644714, "learning_rate": 2.412053555236906e-05, "loss": 0.751, "num_input_tokens_seen": 17117488, "step": 3195 }, { "epoch": 0.512, "grad_norm": 1.1716359853744507, "learning_rate": 2.4057745433251635e-05, "loss": 0.7067, "num_input_tokens_seen": 17141232, "step": 3200 }, { "epoch": 0.5128, "grad_norm": 1.0248827934265137, "learning_rate": 2.3994961265926166e-05, "loss": 0.6432, "num_input_tokens_seen": 17171632, "step": 3205 }, { "epoch": 0.5136, "grad_norm": 0.8832619190216064, "learning_rate": 2.3932183446971583e-05, "loss": 0.6373, "num_input_tokens_seen": 17198640, "step": 3210 }, { "epoch": 0.5144, "grad_norm": 0.8581608533859253, "learning_rate": 2.3869412372926687e-05, "loss": 0.7347, "num_input_tokens_seen": 17228240, "step": 3215 }, { "epoch": 0.5152, "grad_norm": 1.0683339834213257, "learning_rate": 2.3806648440287714e-05, "loss": 0.6789, "num_input_tokens_seen": 17259392, "step": 3220 }, { "epoch": 0.516, "grad_norm": 0.9491643309593201, "learning_rate": 2.3743892045505764e-05, "loss": 0.7548, "num_input_tokens_seen": 17287808, "step": 3225 }, { "epoch": 0.5168, "grad_norm": 0.8620545864105225, "learning_rate": 2.368114358498434e-05, "loss": 0.7792, "num_input_tokens_seen": 17311520, "step": 3230 }, { "epoch": 0.5176, "grad_norm": 1.0956013202667236, "learning_rate": 2.361840345507683e-05, "loss": 0.6575, "num_input_tokens_seen": 17340816, "step": 3235 }, { "epoch": 0.5184, "grad_norm": 1.0029551982879639, "learning_rate": 2.355567205208397e-05, "loss": 0.6414, "num_input_tokens_seen": 17363408, "step": 3240 }, { "epoch": 0.5192, "grad_norm": 1.0472480058670044, "learning_rate": 2.3492949772251414e-05, "loss": 0.7161, "num_input_tokens_seen": 17393248, "step": 3245 }, { "epoch": 0.52, "grad_norm": 0.8757247924804688, "learning_rate": 2.3430237011767167e-05, "loss": 0.6957, "num_input_tokens_seen": 17425232, "step": 3250 }, { "epoch": 0.5208, "grad_norm": 1.0374081134796143, "learning_rate": 2.3367534166759102e-05, "loss": 0.7615, "num_input_tokens_seen": 17446864, "step": 3255 }, { "epoch": 0.5216, "grad_norm": 1.0572500228881836, "learning_rate": 2.3304841633292487e-05, "loss": 0.6994, "num_input_tokens_seen": 17470896, "step": 3260 }, { "epoch": 0.5224, "grad_norm": 1.0209540128707886, "learning_rate": 2.3242159807367408e-05, "loss": 0.7116, "num_input_tokens_seen": 17501488, "step": 3265 }, { "epoch": 0.5232, "grad_norm": 1.0854222774505615, "learning_rate": 2.3179489084916358e-05, "loss": 0.7583, "num_input_tokens_seen": 17526032, "step": 3270 }, { "epoch": 0.524, "grad_norm": 1.0327874422073364, "learning_rate": 2.3116829861801686e-05, "loss": 0.7302, "num_input_tokens_seen": 17550144, "step": 3275 }, { "epoch": 0.5248, "grad_norm": 1.186990737915039, "learning_rate": 2.3054182533813087e-05, "loss": 0.6794, "num_input_tokens_seen": 17575600, "step": 3280 }, { "epoch": 0.5256, "grad_norm": 1.000475287437439, "learning_rate": 2.2991547496665148e-05, "loss": 0.7294, "num_input_tokens_seen": 17601408, "step": 3285 }, { "epoch": 0.5264, "grad_norm": 1.1120193004608154, "learning_rate": 2.2928925145994794e-05, "loss": 0.6333, "num_input_tokens_seen": 17624752, "step": 3290 }, { "epoch": 0.5272, "grad_norm": 0.7765217423439026, "learning_rate": 2.286631587735883e-05, "loss": 0.7779, "num_input_tokens_seen": 17651040, "step": 3295 }, { "epoch": 0.528, "grad_norm": 0.9403998255729675, "learning_rate": 2.280372008623142e-05, "loss": 0.7035, "num_input_tokens_seen": 17678288, "step": 3300 }, { "epoch": 0.5288, "grad_norm": 1.019305944442749, "learning_rate": 2.2741138168001608e-05, "loss": 0.719, "num_input_tokens_seen": 17702816, "step": 3305 }, { "epoch": 0.5296, "grad_norm": 1.0804177522659302, "learning_rate": 2.267857051797081e-05, "loss": 0.7134, "num_input_tokens_seen": 17728848, "step": 3310 }, { "epoch": 0.5304, "grad_norm": 0.7340876460075378, "learning_rate": 2.2616017531350288e-05, "loss": 0.6916, "num_input_tokens_seen": 17756240, "step": 3315 }, { "epoch": 0.5312, "grad_norm": 0.9618902802467346, "learning_rate": 2.255347960325871e-05, "loss": 0.6389, "num_input_tokens_seen": 17781104, "step": 3320 }, { "epoch": 0.532, "grad_norm": 0.9528229832649231, "learning_rate": 2.2490957128719624e-05, "loss": 0.6648, "num_input_tokens_seen": 17808816, "step": 3325 }, { "epoch": 0.5328, "grad_norm": 1.043328881263733, "learning_rate": 2.2428450502658967e-05, "loss": 0.6683, "num_input_tokens_seen": 17834496, "step": 3330 }, { "epoch": 0.5336, "grad_norm": 0.8162310719490051, "learning_rate": 2.2365960119902545e-05, "loss": 0.7686, "num_input_tokens_seen": 17862880, "step": 3335 }, { "epoch": 0.5344, "grad_norm": 0.8925397396087646, "learning_rate": 2.2303486375173585e-05, "loss": 0.7073, "num_input_tokens_seen": 17890064, "step": 3340 }, { "epoch": 0.5352, "grad_norm": 1.0610705614089966, "learning_rate": 2.224102966309021e-05, "loss": 0.642, "num_input_tokens_seen": 17918144, "step": 3345 }, { "epoch": 0.536, "grad_norm": 1.1452162265777588, "learning_rate": 2.217859037816296e-05, "loss": 0.7078, "num_input_tokens_seen": 17945344, "step": 3350 }, { "epoch": 0.5368, "grad_norm": 0.8698946833610535, "learning_rate": 2.2116168914792292e-05, "loss": 0.7437, "num_input_tokens_seen": 17970096, "step": 3355 }, { "epoch": 0.5376, "grad_norm": 1.1551156044006348, "learning_rate": 2.205376566726611e-05, "loss": 0.7606, "num_input_tokens_seen": 17997328, "step": 3360 }, { "epoch": 0.5384, "grad_norm": 1.3479046821594238, "learning_rate": 2.1991381029757215e-05, "loss": 0.6824, "num_input_tokens_seen": 18022464, "step": 3365 }, { "epoch": 0.5392, "grad_norm": 0.9218052625656128, "learning_rate": 2.19290153963209e-05, "loss": 0.7262, "num_input_tokens_seen": 18052176, "step": 3370 }, { "epoch": 0.54, "grad_norm": 1.302252173423767, "learning_rate": 2.186666916089239e-05, "loss": 0.7491, "num_input_tokens_seen": 18079008, "step": 3375 }, { "epoch": 0.5408, "grad_norm": 1.4532941579818726, "learning_rate": 2.1804342717284415e-05, "loss": 0.6246, "num_input_tokens_seen": 18102784, "step": 3380 }, { "epoch": 0.5416, "grad_norm": 0.7572783827781677, "learning_rate": 2.174203645918464e-05, "loss": 0.6712, "num_input_tokens_seen": 18130688, "step": 3385 }, { "epoch": 0.5424, "grad_norm": 1.0954492092132568, "learning_rate": 2.1679750780153267e-05, "loss": 0.7238, "num_input_tokens_seen": 18159200, "step": 3390 }, { "epoch": 0.5432, "grad_norm": 1.1352787017822266, "learning_rate": 2.1617486073620498e-05, "loss": 0.663, "num_input_tokens_seen": 18188736, "step": 3395 }, { "epoch": 0.544, "grad_norm": 1.012987732887268, "learning_rate": 2.155524273288405e-05, "loss": 0.6928, "num_input_tokens_seen": 18217856, "step": 3400 }, { "epoch": 0.5448, "grad_norm": 0.8638446927070618, "learning_rate": 2.1493021151106703e-05, "loss": 0.7373, "num_input_tokens_seen": 18247616, "step": 3405 }, { "epoch": 0.5456, "grad_norm": 1.2647075653076172, "learning_rate": 2.1430821721313782e-05, "loss": 0.7593, "num_input_tokens_seen": 18274416, "step": 3410 }, { "epoch": 0.5464, "grad_norm": 0.8533580899238586, "learning_rate": 2.1368644836390684e-05, "loss": 0.6718, "num_input_tokens_seen": 18298720, "step": 3415 }, { "epoch": 0.5472, "grad_norm": 0.8091197609901428, "learning_rate": 2.130649088908041e-05, "loss": 0.7303, "num_input_tokens_seen": 18326160, "step": 3420 }, { "epoch": 0.548, "grad_norm": 0.886374294757843, "learning_rate": 2.1244360271981073e-05, "loss": 0.74, "num_input_tokens_seen": 18351344, "step": 3425 }, { "epoch": 0.5488, "grad_norm": 0.8284346461296082, "learning_rate": 2.1182253377543425e-05, "loss": 0.6448, "num_input_tokens_seen": 18374752, "step": 3430 }, { "epoch": 0.5496, "grad_norm": 0.9252715706825256, "learning_rate": 2.112017059806835e-05, "loss": 0.6759, "num_input_tokens_seen": 18402432, "step": 3435 }, { "epoch": 0.5504, "grad_norm": 1.0463306903839111, "learning_rate": 2.1058112325704436e-05, "loss": 0.7327, "num_input_tokens_seen": 18428656, "step": 3440 }, { "epoch": 0.5512, "grad_norm": 0.999754011631012, "learning_rate": 2.0996078952445452e-05, "loss": 0.7214, "num_input_tokens_seen": 18451744, "step": 3445 }, { "epoch": 0.552, "grad_norm": 0.759445071220398, "learning_rate": 2.0934070870127912e-05, "loss": 0.7192, "num_input_tokens_seen": 18476960, "step": 3450 }, { "epoch": 0.5528, "grad_norm": 0.757986843585968, "learning_rate": 2.0872088470428553e-05, "loss": 0.6481, "num_input_tokens_seen": 18507280, "step": 3455 }, { "epoch": 0.5536, "grad_norm": 1.555863857269287, "learning_rate": 2.08101321448619e-05, "loss": 0.6906, "num_input_tokens_seen": 18531264, "step": 3460 }, { "epoch": 0.5544, "grad_norm": 0.9648059606552124, "learning_rate": 2.0748202284777777e-05, "loss": 0.6691, "num_input_tokens_seen": 18559552, "step": 3465 }, { "epoch": 0.5552, "grad_norm": 0.7556829452514648, "learning_rate": 2.0686299281358835e-05, "loss": 0.743, "num_input_tokens_seen": 18587408, "step": 3470 }, { "epoch": 0.556, "grad_norm": 1.2063502073287964, "learning_rate": 2.0624423525618098e-05, "loss": 0.6896, "num_input_tokens_seen": 18616384, "step": 3475 }, { "epoch": 0.5568, "grad_norm": 1.0923407077789307, "learning_rate": 2.056257540839647e-05, "loss": 0.7799, "num_input_tokens_seen": 18640432, "step": 3480 }, { "epoch": 0.5576, "grad_norm": 0.9768779873847961, "learning_rate": 2.050075532036026e-05, "loss": 0.6796, "num_input_tokens_seen": 18661696, "step": 3485 }, { "epoch": 0.5584, "grad_norm": 1.0903396606445312, "learning_rate": 2.0438963651998747e-05, "loss": 0.6601, "num_input_tokens_seen": 18689280, "step": 3490 }, { "epoch": 0.5592, "grad_norm": 1.0859960317611694, "learning_rate": 2.037720079362169e-05, "loss": 0.7247, "num_input_tokens_seen": 18713776, "step": 3495 }, { "epoch": 0.56, "grad_norm": 0.979040801525116, "learning_rate": 2.031546713535688e-05, "loss": 0.7113, "num_input_tokens_seen": 18739632, "step": 3500 }, { "epoch": 0.5608, "grad_norm": 0.9411745667457581, "learning_rate": 2.0253763067147657e-05, "loss": 0.604, "num_input_tokens_seen": 18767504, "step": 3505 }, { "epoch": 0.5616, "grad_norm": 1.0369850397109985, "learning_rate": 2.0192088978750433e-05, "loss": 0.6292, "num_input_tokens_seen": 18794976, "step": 3510 }, { "epoch": 0.5624, "grad_norm": 1.033288598060608, "learning_rate": 2.0130445259732285e-05, "loss": 0.7227, "num_input_tokens_seen": 18823456, "step": 3515 }, { "epoch": 0.5632, "grad_norm": 1.0590053796768188, "learning_rate": 2.0068832299468428e-05, "loss": 0.6536, "num_input_tokens_seen": 18851104, "step": 3520 }, { "epoch": 0.564, "grad_norm": 0.8934491276741028, "learning_rate": 2.000725048713983e-05, "loss": 0.6729, "num_input_tokens_seen": 18882096, "step": 3525 }, { "epoch": 0.5648, "grad_norm": 1.3166091442108154, "learning_rate": 1.994570021173067e-05, "loss": 0.8178, "num_input_tokens_seen": 18903520, "step": 3530 }, { "epoch": 0.5656, "grad_norm": 0.7813395857810974, "learning_rate": 1.988418186202594e-05, "loss": 0.6575, "num_input_tokens_seen": 18937200, "step": 3535 }, { "epoch": 0.5664, "grad_norm": 1.1987985372543335, "learning_rate": 1.9822695826608972e-05, "loss": 0.7789, "num_input_tokens_seen": 18965424, "step": 3540 }, { "epoch": 0.5672, "grad_norm": 0.9326161742210388, "learning_rate": 1.9761242493858987e-05, "loss": 0.6699, "num_input_tokens_seen": 18989456, "step": 3545 }, { "epoch": 0.568, "grad_norm": 0.736609160900116, "learning_rate": 1.969982225194864e-05, "loss": 0.7095, "num_input_tokens_seen": 19017808, "step": 3550 }, { "epoch": 0.5688, "grad_norm": 1.0842061042785645, "learning_rate": 1.9638435488841546e-05, "loss": 0.7191, "num_input_tokens_seen": 19046496, "step": 3555 }, { "epoch": 0.5696, "grad_norm": 0.9748028516769409, "learning_rate": 1.957708259228987e-05, "loss": 0.7016, "num_input_tokens_seen": 19072128, "step": 3560 }, { "epoch": 0.5704, "grad_norm": 1.0534452199935913, "learning_rate": 1.951576394983185e-05, "loss": 0.6903, "num_input_tokens_seen": 19096528, "step": 3565 }, { "epoch": 0.5712, "grad_norm": 0.860016405582428, "learning_rate": 1.945447994878937e-05, "loss": 0.6144, "num_input_tokens_seen": 19126240, "step": 3570 }, { "epoch": 0.572, "grad_norm": 0.9540638327598572, "learning_rate": 1.9393230976265473e-05, "loss": 0.6755, "num_input_tokens_seen": 19152752, "step": 3575 }, { "epoch": 0.5728, "grad_norm": 0.8391373157501221, "learning_rate": 1.9332017419141962e-05, "loss": 0.748, "num_input_tokens_seen": 19179296, "step": 3580 }, { "epoch": 0.5736, "grad_norm": 1.1639857292175293, "learning_rate": 1.9270839664076936e-05, "loss": 0.7011, "num_input_tokens_seen": 19205616, "step": 3585 }, { "epoch": 0.5744, "grad_norm": 0.8739202618598938, "learning_rate": 1.920969809750234e-05, "loss": 0.6672, "num_input_tokens_seen": 19231440, "step": 3590 }, { "epoch": 0.5752, "grad_norm": 0.8280954360961914, "learning_rate": 1.914859310562154e-05, "loss": 0.7261, "num_input_tokens_seen": 19258288, "step": 3595 }, { "epoch": 0.576, "grad_norm": 0.671859622001648, "learning_rate": 1.908752507440689e-05, "loss": 0.7838, "num_input_tokens_seen": 19284464, "step": 3600 }, { "epoch": 0.5768, "grad_norm": 0.7985883951187134, "learning_rate": 1.9026494389597238e-05, "loss": 0.6574, "num_input_tokens_seen": 19312272, "step": 3605 }, { "epoch": 0.5776, "grad_norm": 1.1776115894317627, "learning_rate": 1.8965501436695577e-05, "loss": 0.7692, "num_input_tokens_seen": 19335408, "step": 3610 }, { "epoch": 0.5784, "grad_norm": 0.7614251971244812, "learning_rate": 1.890454660096654e-05, "loss": 0.7165, "num_input_tokens_seen": 19360768, "step": 3615 }, { "epoch": 0.5792, "grad_norm": 1.0146969556808472, "learning_rate": 1.8843630267434e-05, "loss": 0.8187, "num_input_tokens_seen": 19386016, "step": 3620 }, { "epoch": 0.58, "grad_norm": 0.8127625584602356, "learning_rate": 1.8782752820878634e-05, "loss": 0.6587, "num_input_tokens_seen": 19408560, "step": 3625 }, { "epoch": 0.5808, "grad_norm": 1.102415680885315, "learning_rate": 1.872191464583547e-05, "loss": 0.6268, "num_input_tokens_seen": 19436592, "step": 3630 }, { "epoch": 0.5816, "grad_norm": 0.8009477853775024, "learning_rate": 1.866111612659149e-05, "loss": 0.6977, "num_input_tokens_seen": 19463440, "step": 3635 }, { "epoch": 0.5824, "grad_norm": 0.9613442420959473, "learning_rate": 1.8600357647183185e-05, "loss": 0.6292, "num_input_tokens_seen": 19493360, "step": 3640 }, { "epoch": 0.5832, "grad_norm": 1.1276973485946655, "learning_rate": 1.8539639591394133e-05, "loss": 0.6547, "num_input_tokens_seen": 19521392, "step": 3645 }, { "epoch": 0.584, "grad_norm": 1.2128732204437256, "learning_rate": 1.8478962342752583e-05, "loss": 0.6717, "num_input_tokens_seen": 19550336, "step": 3650 }, { "epoch": 0.5848, "grad_norm": 1.1931806802749634, "learning_rate": 1.8418326284528996e-05, "loss": 0.7065, "num_input_tokens_seen": 19575776, "step": 3655 }, { "epoch": 0.5856, "grad_norm": 0.921335756778717, "learning_rate": 1.8357731799733686e-05, "loss": 0.7029, "num_input_tokens_seen": 19598128, "step": 3660 }, { "epoch": 0.5864, "grad_norm": 0.8000009655952454, "learning_rate": 1.8297179271114346e-05, "loss": 0.7311, "num_input_tokens_seen": 19625648, "step": 3665 }, { "epoch": 0.5872, "grad_norm": 1.0933367013931274, "learning_rate": 1.8236669081153657e-05, "loss": 0.7296, "num_input_tokens_seen": 19649952, "step": 3670 }, { "epoch": 0.588, "grad_norm": 0.8331469297409058, "learning_rate": 1.817620161206687e-05, "loss": 0.7534, "num_input_tokens_seen": 19677680, "step": 3675 }, { "epoch": 0.5888, "grad_norm": 1.3450491428375244, "learning_rate": 1.811577724579938e-05, "loss": 0.6995, "num_input_tokens_seen": 19711904, "step": 3680 }, { "epoch": 0.5896, "grad_norm": 1.0697826147079468, "learning_rate": 1.8055396364024317e-05, "loss": 0.7517, "num_input_tokens_seen": 19734272, "step": 3685 }, { "epoch": 0.5904, "grad_norm": 0.9218750596046448, "learning_rate": 1.7995059348140165e-05, "loss": 0.7048, "num_input_tokens_seen": 19761136, "step": 3690 }, { "epoch": 0.5912, "grad_norm": 0.7037175297737122, "learning_rate": 1.7934766579268292e-05, "loss": 0.6385, "num_input_tokens_seen": 19784880, "step": 3695 }, { "epoch": 0.592, "grad_norm": 0.9812880754470825, "learning_rate": 1.7874518438250597e-05, "loss": 0.8177, "num_input_tokens_seen": 19811456, "step": 3700 }, { "epoch": 0.5928, "grad_norm": 1.0128806829452515, "learning_rate": 1.7814315305647093e-05, "loss": 0.7373, "num_input_tokens_seen": 19839168, "step": 3705 }, { "epoch": 0.5936, "grad_norm": 0.8506542444229126, "learning_rate": 1.7754157561733476e-05, "loss": 0.723, "num_input_tokens_seen": 19865584, "step": 3710 }, { "epoch": 0.5944, "grad_norm": 0.8591898679733276, "learning_rate": 1.7694045586498752e-05, "loss": 0.6315, "num_input_tokens_seen": 19893232, "step": 3715 }, { "epoch": 0.5952, "grad_norm": 0.9761216640472412, "learning_rate": 1.7633979759642844e-05, "loss": 0.6184, "num_input_tokens_seen": 19918512, "step": 3720 }, { "epoch": 0.596, "grad_norm": 0.9515823721885681, "learning_rate": 1.7573960460574133e-05, "loss": 0.682, "num_input_tokens_seen": 19944544, "step": 3725 }, { "epoch": 0.5968, "grad_norm": 1.2393804788589478, "learning_rate": 1.7513988068407146e-05, "loss": 0.6738, "num_input_tokens_seen": 19971104, "step": 3730 }, { "epoch": 0.5976, "grad_norm": 1.2483285665512085, "learning_rate": 1.74540629619601e-05, "loss": 0.6895, "num_input_tokens_seen": 19996352, "step": 3735 }, { "epoch": 0.5984, "grad_norm": 1.162599802017212, "learning_rate": 1.7394185519752545e-05, "loss": 0.7436, "num_input_tokens_seen": 20021744, "step": 3740 }, { "epoch": 0.5992, "grad_norm": 0.8526731133460999, "learning_rate": 1.7334356120002957e-05, "loss": 0.7612, "num_input_tokens_seen": 20046560, "step": 3745 }, { "epoch": 0.6, "grad_norm": 1.1033904552459717, "learning_rate": 1.7274575140626318e-05, "loss": 0.7139, "num_input_tokens_seen": 20072560, "step": 3750 }, { "epoch": 0.6008, "grad_norm": 0.9515017867088318, "learning_rate": 1.7214842959231794e-05, "loss": 0.6556, "num_input_tokens_seen": 20103488, "step": 3755 }, { "epoch": 0.6016, "grad_norm": 1.1830626726150513, "learning_rate": 1.7155159953120313e-05, "loss": 0.6884, "num_input_tokens_seen": 20130784, "step": 3760 }, { "epoch": 0.6024, "grad_norm": 1.1456624269485474, "learning_rate": 1.7095526499282172e-05, "loss": 0.7729, "num_input_tokens_seen": 20158720, "step": 3765 }, { "epoch": 0.6032, "grad_norm": 0.8993046879768372, "learning_rate": 1.703594297439469e-05, "loss": 0.7427, "num_input_tokens_seen": 20180736, "step": 3770 }, { "epoch": 0.604, "grad_norm": 1.0378142595291138, "learning_rate": 1.6976409754819767e-05, "loss": 0.6831, "num_input_tokens_seen": 20203744, "step": 3775 }, { "epoch": 0.6048, "grad_norm": 1.0006003379821777, "learning_rate": 1.6916927216601593e-05, "loss": 0.6098, "num_input_tokens_seen": 20232784, "step": 3780 }, { "epoch": 0.6056, "grad_norm": 0.9714456796646118, "learning_rate": 1.6857495735464195e-05, "loss": 0.812, "num_input_tokens_seen": 20262256, "step": 3785 }, { "epoch": 0.6064, "grad_norm": 1.226090431213379, "learning_rate": 1.6798115686809125e-05, "loss": 0.6337, "num_input_tokens_seen": 20290720, "step": 3790 }, { "epoch": 0.6072, "grad_norm": 1.2579602003097534, "learning_rate": 1.6738787445713037e-05, "loss": 0.7105, "num_input_tokens_seen": 20314368, "step": 3795 }, { "epoch": 0.608, "grad_norm": 0.9636203050613403, "learning_rate": 1.6679511386925337e-05, "loss": 0.7776, "num_input_tokens_seen": 20337648, "step": 3800 }, { "epoch": 0.6088, "grad_norm": 0.8365712761878967, "learning_rate": 1.662028788486583e-05, "loss": 0.6626, "num_input_tokens_seen": 20367344, "step": 3805 }, { "epoch": 0.6096, "grad_norm": 1.340468168258667, "learning_rate": 1.656111731362236e-05, "loss": 0.6983, "num_input_tokens_seen": 20391616, "step": 3810 }, { "epoch": 0.6104, "grad_norm": 0.878955066204071, "learning_rate": 1.650200004694839e-05, "loss": 0.6948, "num_input_tokens_seen": 20419520, "step": 3815 }, { "epoch": 0.6112, "grad_norm": 0.9543726444244385, "learning_rate": 1.644293645826072e-05, "loss": 0.8154, "num_input_tokens_seen": 20446048, "step": 3820 }, { "epoch": 0.612, "grad_norm": 1.2340530157089233, "learning_rate": 1.6383926920637077e-05, "loss": 0.7234, "num_input_tokens_seen": 20472960, "step": 3825 }, { "epoch": 0.6128, "grad_norm": 0.8794097900390625, "learning_rate": 1.6324971806813767e-05, "loss": 0.668, "num_input_tokens_seen": 20502816, "step": 3830 }, { "epoch": 0.6136, "grad_norm": 1.0433683395385742, "learning_rate": 1.6266071489183327e-05, "loss": 0.6936, "num_input_tokens_seen": 20529056, "step": 3835 }, { "epoch": 0.6144, "grad_norm": 1.3372730016708374, "learning_rate": 1.620722633979219e-05, "loss": 0.7988, "num_input_tokens_seen": 20555392, "step": 3840 }, { "epoch": 0.6152, "grad_norm": 1.0201383829116821, "learning_rate": 1.614843673033828e-05, "loss": 0.7752, "num_input_tokens_seen": 20583888, "step": 3845 }, { "epoch": 0.616, "grad_norm": 0.7360509634017944, "learning_rate": 1.6089703032168733e-05, "loss": 0.6784, "num_input_tokens_seen": 20612112, "step": 3850 }, { "epoch": 0.6168, "grad_norm": 0.8650053143501282, "learning_rate": 1.603102561627751e-05, "loss": 0.6483, "num_input_tokens_seen": 20639296, "step": 3855 }, { "epoch": 0.6176, "grad_norm": 0.9596717953681946, "learning_rate": 1.5972404853303062e-05, "loss": 0.6876, "num_input_tokens_seen": 20663680, "step": 3860 }, { "epoch": 0.6184, "grad_norm": 0.9903700947761536, "learning_rate": 1.5913841113525992e-05, "loss": 0.7651, "num_input_tokens_seen": 20690592, "step": 3865 }, { "epoch": 0.6192, "grad_norm": 1.0361056327819824, "learning_rate": 1.585533476686669e-05, "loss": 0.6692, "num_input_tokens_seen": 20716944, "step": 3870 }, { "epoch": 0.62, "grad_norm": 1.3285175561904907, "learning_rate": 1.5796886182883053e-05, "loss": 0.708, "num_input_tokens_seen": 20742128, "step": 3875 }, { "epoch": 0.6208, "grad_norm": 0.9742456078529358, "learning_rate": 1.5738495730768105e-05, "loss": 0.6734, "num_input_tokens_seen": 20769344, "step": 3880 }, { "epoch": 0.6216, "grad_norm": 0.9866467118263245, "learning_rate": 1.5680163779347667e-05, "loss": 0.7442, "num_input_tokens_seen": 20793920, "step": 3885 }, { "epoch": 0.6224, "grad_norm": 1.2530503273010254, "learning_rate": 1.562189069707807e-05, "loss": 0.786, "num_input_tokens_seen": 20819616, "step": 3890 }, { "epoch": 0.6232, "grad_norm": 1.0180388689041138, "learning_rate": 1.556367685204374e-05, "loss": 0.6565, "num_input_tokens_seen": 20843056, "step": 3895 }, { "epoch": 0.624, "grad_norm": 1.1570924520492554, "learning_rate": 1.5505522611954975e-05, "loss": 0.8403, "num_input_tokens_seen": 20870320, "step": 3900 }, { "epoch": 0.6248, "grad_norm": 0.9555189609527588, "learning_rate": 1.5447428344145563e-05, "loss": 0.717, "num_input_tokens_seen": 20894448, "step": 3905 }, { "epoch": 0.6256, "grad_norm": 0.7047298550605774, "learning_rate": 1.538939441557048e-05, "loss": 0.6563, "num_input_tokens_seen": 20926800, "step": 3910 }, { "epoch": 0.6264, "grad_norm": 1.1212091445922852, "learning_rate": 1.5331421192803565e-05, "loss": 0.7742, "num_input_tokens_seen": 20954016, "step": 3915 }, { "epoch": 0.6272, "grad_norm": 0.9030645489692688, "learning_rate": 1.5273509042035172e-05, "loss": 0.6654, "num_input_tokens_seen": 20982512, "step": 3920 }, { "epoch": 0.628, "grad_norm": 0.9414677619934082, "learning_rate": 1.521565832906994e-05, "loss": 0.6737, "num_input_tokens_seen": 21008768, "step": 3925 }, { "epoch": 0.6288, "grad_norm": 1.1415228843688965, "learning_rate": 1.515786941932441e-05, "loss": 0.7259, "num_input_tokens_seen": 21038144, "step": 3930 }, { "epoch": 0.6296, "grad_norm": 1.0087826251983643, "learning_rate": 1.5100142677824753e-05, "loss": 0.6793, "num_input_tokens_seen": 21065120, "step": 3935 }, { "epoch": 0.6304, "grad_norm": 1.3329883813858032, "learning_rate": 1.5042478469204435e-05, "loss": 0.6934, "num_input_tokens_seen": 21091296, "step": 3940 }, { "epoch": 0.6312, "grad_norm": 0.9850925803184509, "learning_rate": 1.4984877157701932e-05, "loss": 0.7746, "num_input_tokens_seen": 21117568, "step": 3945 }, { "epoch": 0.632, "grad_norm": 0.8925058245658875, "learning_rate": 1.4927339107158437e-05, "loss": 0.6311, "num_input_tokens_seen": 21146640, "step": 3950 }, { "epoch": 0.6328, "grad_norm": 1.2707431316375732, "learning_rate": 1.486986468101555e-05, "loss": 0.7614, "num_input_tokens_seen": 21169680, "step": 3955 }, { "epoch": 0.6336, "grad_norm": 1.0344791412353516, "learning_rate": 1.4812454242312979e-05, "loss": 0.7291, "num_input_tokens_seen": 21195360, "step": 3960 }, { "epoch": 0.6344, "grad_norm": 0.8999541997909546, "learning_rate": 1.4755108153686275e-05, "loss": 0.7421, "num_input_tokens_seen": 21218896, "step": 3965 }, { "epoch": 0.6352, "grad_norm": 1.3539083003997803, "learning_rate": 1.4697826777364477e-05, "loss": 0.7831, "num_input_tokens_seen": 21244080, "step": 3970 }, { "epoch": 0.636, "grad_norm": 0.9629884362220764, "learning_rate": 1.4640610475167898e-05, "loss": 0.6907, "num_input_tokens_seen": 21271024, "step": 3975 }, { "epoch": 0.6368, "grad_norm": 0.9040243625640869, "learning_rate": 1.4583459608505801e-05, "loss": 0.7001, "num_input_tokens_seen": 21298992, "step": 3980 }, { "epoch": 0.6376, "grad_norm": 1.122290849685669, "learning_rate": 1.4526374538374132e-05, "loss": 0.6729, "num_input_tokens_seen": 21324032, "step": 3985 }, { "epoch": 0.6384, "grad_norm": 0.8082273602485657, "learning_rate": 1.4469355625353198e-05, "loss": 0.6636, "num_input_tokens_seen": 21354256, "step": 3990 }, { "epoch": 0.6392, "grad_norm": 0.7639278173446655, "learning_rate": 1.4412403229605454e-05, "loss": 0.6349, "num_input_tokens_seen": 21382144, "step": 3995 }, { "epoch": 0.64, "grad_norm": 1.0013383626937866, "learning_rate": 1.4355517710873184e-05, "loss": 0.6892, "num_input_tokens_seen": 21408112, "step": 4000 }, { "epoch": 0.6408, "grad_norm": 0.918889582157135, "learning_rate": 1.4298699428476236e-05, "loss": 0.6602, "num_input_tokens_seen": 21438800, "step": 4005 }, { "epoch": 0.6416, "grad_norm": 0.9448719620704651, "learning_rate": 1.4241948741309782e-05, "loss": 0.6613, "num_input_tokens_seen": 21466464, "step": 4010 }, { "epoch": 0.6424, "grad_norm": 1.1950000524520874, "learning_rate": 1.418526600784198e-05, "loss": 0.6821, "num_input_tokens_seen": 21496864, "step": 4015 }, { "epoch": 0.6432, "grad_norm": 1.0359631776809692, "learning_rate": 1.412865158611179e-05, "loss": 0.6698, "num_input_tokens_seen": 21523456, "step": 4020 }, { "epoch": 0.644, "grad_norm": 0.9636697769165039, "learning_rate": 1.4072105833726684e-05, "loss": 0.5917, "num_input_tokens_seen": 21554320, "step": 4025 }, { "epoch": 0.6448, "grad_norm": 0.8568287491798401, "learning_rate": 1.401562910786034e-05, "loss": 0.7332, "num_input_tokens_seen": 21584496, "step": 4030 }, { "epoch": 0.6456, "grad_norm": 0.7950714230537415, "learning_rate": 1.3959221765250469e-05, "loss": 0.6826, "num_input_tokens_seen": 21615104, "step": 4035 }, { "epoch": 0.6464, "grad_norm": 0.9343571662902832, "learning_rate": 1.3902884162196508e-05, "loss": 0.7349, "num_input_tokens_seen": 21642144, "step": 4040 }, { "epoch": 0.6472, "grad_norm": 0.8434100151062012, "learning_rate": 1.3846616654557362e-05, "loss": 0.6341, "num_input_tokens_seen": 21671408, "step": 4045 }, { "epoch": 0.648, "grad_norm": 1.1461455821990967, "learning_rate": 1.3790419597749199e-05, "loss": 0.7531, "num_input_tokens_seen": 21698880, "step": 4050 }, { "epoch": 0.6488, "grad_norm": 1.261234164237976, "learning_rate": 1.3734293346743168e-05, "loss": 0.6284, "num_input_tokens_seen": 21727280, "step": 4055 }, { "epoch": 0.6496, "grad_norm": 1.0802944898605347, "learning_rate": 1.367823825606319e-05, "loss": 0.7148, "num_input_tokens_seen": 21751824, "step": 4060 }, { "epoch": 0.6504, "grad_norm": 1.1353379487991333, "learning_rate": 1.3622254679783663e-05, "loss": 0.7022, "num_input_tokens_seen": 21782080, "step": 4065 }, { "epoch": 0.6512, "grad_norm": 1.0912383794784546, "learning_rate": 1.3566342971527291e-05, "loss": 0.7308, "num_input_tokens_seen": 21809376, "step": 4070 }, { "epoch": 0.652, "grad_norm": 1.0142539739608765, "learning_rate": 1.3510503484462805e-05, "loss": 0.7338, "num_input_tokens_seen": 21836240, "step": 4075 }, { "epoch": 0.6528, "grad_norm": 1.0957950353622437, "learning_rate": 1.3454736571302763e-05, "loss": 0.6941, "num_input_tokens_seen": 21862768, "step": 4080 }, { "epoch": 0.6536, "grad_norm": 0.9035006761550903, "learning_rate": 1.3399042584301298e-05, "loss": 0.7218, "num_input_tokens_seen": 21890304, "step": 4085 }, { "epoch": 0.6544, "grad_norm": 1.0284723043441772, "learning_rate": 1.3343421875251888e-05, "loss": 0.8144, "num_input_tokens_seen": 21912192, "step": 4090 }, { "epoch": 0.6552, "grad_norm": 1.0489941835403442, "learning_rate": 1.3287874795485167e-05, "loss": 0.8236, "num_input_tokens_seen": 21939984, "step": 4095 }, { "epoch": 0.656, "grad_norm": 0.9491598606109619, "learning_rate": 1.3232401695866687e-05, "loss": 0.6512, "num_input_tokens_seen": 21967168, "step": 4100 }, { "epoch": 0.6568, "grad_norm": 1.0019705295562744, "learning_rate": 1.3177002926794685e-05, "loss": 0.7271, "num_input_tokens_seen": 21999904, "step": 4105 }, { "epoch": 0.6576, "grad_norm": 1.0153288841247559, "learning_rate": 1.3121678838197909e-05, "loss": 0.635, "num_input_tokens_seen": 22028272, "step": 4110 }, { "epoch": 0.6584, "grad_norm": 0.8823714852333069, "learning_rate": 1.3066429779533351e-05, "loss": 0.6708, "num_input_tokens_seen": 22052224, "step": 4115 }, { "epoch": 0.6592, "grad_norm": 0.8675245642662048, "learning_rate": 1.3011256099784103e-05, "loss": 0.5916, "num_input_tokens_seen": 22081360, "step": 4120 }, { "epoch": 0.66, "grad_norm": 1.1484395265579224, "learning_rate": 1.2956158147457115e-05, "loss": 0.7399, "num_input_tokens_seen": 22112080, "step": 4125 }, { "epoch": 0.6608, "grad_norm": 1.2655909061431885, "learning_rate": 1.2901136270580993e-05, "loss": 0.6543, "num_input_tokens_seen": 22139792, "step": 4130 }, { "epoch": 0.6616, "grad_norm": 1.2049787044525146, "learning_rate": 1.2846190816703835e-05, "loss": 0.6808, "num_input_tokens_seen": 22163136, "step": 4135 }, { "epoch": 0.6624, "grad_norm": 0.7781268358230591, "learning_rate": 1.279132213289096e-05, "loss": 0.6939, "num_input_tokens_seen": 22192464, "step": 4140 }, { "epoch": 0.6632, "grad_norm": 1.1952673196792603, "learning_rate": 1.273653056572282e-05, "loss": 0.6628, "num_input_tokens_seen": 22219424, "step": 4145 }, { "epoch": 0.664, "grad_norm": 1.0534050464630127, "learning_rate": 1.2681816461292715e-05, "loss": 0.686, "num_input_tokens_seen": 22244496, "step": 4150 }, { "epoch": 0.6648, "grad_norm": 1.18624746799469, "learning_rate": 1.2627180165204671e-05, "loss": 0.7278, "num_input_tokens_seen": 22271600, "step": 4155 }, { "epoch": 0.6656, "grad_norm": 0.8680349588394165, "learning_rate": 1.257262202257124e-05, "loss": 0.6531, "num_input_tokens_seen": 22298080, "step": 4160 }, { "epoch": 0.6664, "grad_norm": 0.90425705909729, "learning_rate": 1.251814237801128e-05, "loss": 0.5756, "num_input_tokens_seen": 22324832, "step": 4165 }, { "epoch": 0.6672, "grad_norm": 1.0510259866714478, "learning_rate": 1.246374157564785e-05, "loss": 0.647, "num_input_tokens_seen": 22353728, "step": 4170 }, { "epoch": 0.668, "grad_norm": 1.1430630683898926, "learning_rate": 1.2409419959105981e-05, "loss": 0.7024, "num_input_tokens_seen": 22374880, "step": 4175 }, { "epoch": 0.6688, "grad_norm": 0.8265404105186462, "learning_rate": 1.2355177871510538e-05, "loss": 0.7661, "num_input_tokens_seen": 22402288, "step": 4180 }, { "epoch": 0.6696, "grad_norm": 0.8584622144699097, "learning_rate": 1.2301015655484006e-05, "loss": 0.6462, "num_input_tokens_seen": 22430240, "step": 4185 }, { "epoch": 0.6704, "grad_norm": 1.0526131391525269, "learning_rate": 1.2246933653144385e-05, "loss": 0.6487, "num_input_tokens_seen": 22454800, "step": 4190 }, { "epoch": 0.6712, "grad_norm": 1.1912094354629517, "learning_rate": 1.2192932206103e-05, "loss": 0.7369, "num_input_tokens_seen": 22482528, "step": 4195 }, { "epoch": 0.672, "grad_norm": 0.8804354071617126, "learning_rate": 1.2139011655462337e-05, "loss": 0.6942, "num_input_tokens_seen": 22508976, "step": 4200 }, { "epoch": 0.6728, "grad_norm": 0.9333446025848389, "learning_rate": 1.2085172341813911e-05, "loss": 0.7691, "num_input_tokens_seen": 22538976, "step": 4205 }, { "epoch": 0.6736, "grad_norm": 0.8501102328300476, "learning_rate": 1.2031414605236066e-05, "loss": 0.5865, "num_input_tokens_seen": 22566368, "step": 4210 }, { "epoch": 0.6744, "grad_norm": 0.89410001039505, "learning_rate": 1.1977738785291895e-05, "loss": 0.6916, "num_input_tokens_seen": 22592656, "step": 4215 }, { "epoch": 0.6752, "grad_norm": 0.9672756195068359, "learning_rate": 1.1924145221027047e-05, "loss": 0.7436, "num_input_tokens_seen": 22619872, "step": 4220 }, { "epoch": 0.676, "grad_norm": 0.8680210113525391, "learning_rate": 1.1870634250967605e-05, "loss": 0.6728, "num_input_tokens_seen": 22650320, "step": 4225 }, { "epoch": 0.6768, "grad_norm": 1.0170356035232544, "learning_rate": 1.1817206213117946e-05, "loss": 0.728, "num_input_tokens_seen": 22676896, "step": 4230 }, { "epoch": 0.6776, "grad_norm": 1.0950289964675903, "learning_rate": 1.1763861444958573e-05, "loss": 0.6696, "num_input_tokens_seen": 22702352, "step": 4235 }, { "epoch": 0.6784, "grad_norm": 1.2183908224105835, "learning_rate": 1.1710600283444047e-05, "loss": 0.7827, "num_input_tokens_seen": 22728288, "step": 4240 }, { "epoch": 0.6792, "grad_norm": 0.9134287238121033, "learning_rate": 1.1657423065000811e-05, "loss": 0.7166, "num_input_tokens_seen": 22757632, "step": 4245 }, { "epoch": 0.68, "grad_norm": 1.0772439241409302, "learning_rate": 1.1604330125525079e-05, "loss": 0.7143, "num_input_tokens_seen": 22783440, "step": 4250 }, { "epoch": 0.6808, "grad_norm": 1.003915786743164, "learning_rate": 1.155132180038072e-05, "loss": 0.82, "num_input_tokens_seen": 22809616, "step": 4255 }, { "epoch": 0.6816, "grad_norm": 0.9822829961776733, "learning_rate": 1.1498398424397106e-05, "loss": 0.7297, "num_input_tokens_seen": 22835792, "step": 4260 }, { "epoch": 0.6824, "grad_norm": 0.855888307094574, "learning_rate": 1.1445560331867053e-05, "loss": 0.6956, "num_input_tokens_seen": 22864560, "step": 4265 }, { "epoch": 0.6832, "grad_norm": 0.850237250328064, "learning_rate": 1.1392807856544683e-05, "loss": 0.7157, "num_input_tokens_seen": 22892912, "step": 4270 }, { "epoch": 0.684, "grad_norm": 1.4631861448287964, "learning_rate": 1.1340141331643276e-05, "loss": 0.753, "num_input_tokens_seen": 22912640, "step": 4275 }, { "epoch": 0.6848, "grad_norm": 1.1579556465148926, "learning_rate": 1.1287561089833248e-05, "loss": 0.7247, "num_input_tokens_seen": 22937072, "step": 4280 }, { "epoch": 0.6856, "grad_norm": 1.3090944290161133, "learning_rate": 1.1235067463239967e-05, "loss": 0.7671, "num_input_tokens_seen": 22961104, "step": 4285 }, { "epoch": 0.6864, "grad_norm": 0.9593985676765442, "learning_rate": 1.1182660783441718e-05, "loss": 0.6771, "num_input_tokens_seen": 22983744, "step": 4290 }, { "epoch": 0.6872, "grad_norm": 0.6641414165496826, "learning_rate": 1.1130341381467569e-05, "loss": 0.7179, "num_input_tokens_seen": 23010208, "step": 4295 }, { "epoch": 0.688, "grad_norm": 0.6973831057548523, "learning_rate": 1.107810958779531e-05, "loss": 0.5792, "num_input_tokens_seen": 23043392, "step": 4300 }, { "epoch": 0.6888, "grad_norm": 0.9109097123146057, "learning_rate": 1.1025965732349316e-05, "loss": 0.6574, "num_input_tokens_seen": 23074928, "step": 4305 }, { "epoch": 0.6896, "grad_norm": 0.8845970034599304, "learning_rate": 1.0973910144498534e-05, "loss": 0.6544, "num_input_tokens_seen": 23105728, "step": 4310 }, { "epoch": 0.6904, "grad_norm": 1.181096076965332, "learning_rate": 1.0921943153054343e-05, "loss": 0.6503, "num_input_tokens_seen": 23132768, "step": 4315 }, { "epoch": 0.6912, "grad_norm": 0.785658061504364, "learning_rate": 1.0870065086268505e-05, "loss": 0.6502, "num_input_tokens_seen": 23160080, "step": 4320 }, { "epoch": 0.692, "grad_norm": 1.1847856044769287, "learning_rate": 1.0818276271831093e-05, "loss": 0.7114, "num_input_tokens_seen": 23187696, "step": 4325 }, { "epoch": 0.6928, "grad_norm": 1.1500554084777832, "learning_rate": 1.0766577036868395e-05, "loss": 0.6546, "num_input_tokens_seen": 23211904, "step": 4330 }, { "epoch": 0.6936, "grad_norm": 0.9399601221084595, "learning_rate": 1.0714967707940875e-05, "loss": 0.6965, "num_input_tokens_seen": 23238144, "step": 4335 }, { "epoch": 0.6944, "grad_norm": 0.928415060043335, "learning_rate": 1.0663448611041113e-05, "loss": 0.6232, "num_input_tokens_seen": 23267104, "step": 4340 }, { "epoch": 0.6952, "grad_norm": 1.2702407836914062, "learning_rate": 1.0612020071591722e-05, "loss": 0.6686, "num_input_tokens_seen": 23298976, "step": 4345 }, { "epoch": 0.696, "grad_norm": 1.1251682043075562, "learning_rate": 1.0560682414443315e-05, "loss": 0.6975, "num_input_tokens_seen": 23329552, "step": 4350 }, { "epoch": 0.6968, "grad_norm": 0.8569183945655823, "learning_rate": 1.0509435963872422e-05, "loss": 0.7017, "num_input_tokens_seen": 23359664, "step": 4355 }, { "epoch": 0.6976, "grad_norm": 0.9474520683288574, "learning_rate": 1.0458281043579482e-05, "loss": 0.6856, "num_input_tokens_seen": 23386320, "step": 4360 }, { "epoch": 0.6984, "grad_norm": 1.3578598499298096, "learning_rate": 1.0407217976686775e-05, "loss": 0.6739, "num_input_tokens_seen": 23416512, "step": 4365 }, { "epoch": 0.6992, "grad_norm": 1.1748965978622437, "learning_rate": 1.0356247085736386e-05, "loss": 0.6803, "num_input_tokens_seen": 23439904, "step": 4370 }, { "epoch": 0.7, "grad_norm": 0.7888918519020081, "learning_rate": 1.0305368692688174e-05, "loss": 0.7111, "num_input_tokens_seen": 23466256, "step": 4375 }, { "epoch": 0.7008, "grad_norm": 1.137215256690979, "learning_rate": 1.0254583118917698e-05, "loss": 0.762, "num_input_tokens_seen": 23493536, "step": 4380 }, { "epoch": 0.7016, "grad_norm": 1.337811827659607, "learning_rate": 1.020389068521426e-05, "loss": 0.6206, "num_input_tokens_seen": 23515632, "step": 4385 }, { "epoch": 0.7024, "grad_norm": 1.0022634267807007, "learning_rate": 1.0153291711778826e-05, "loss": 0.6711, "num_input_tokens_seen": 23541152, "step": 4390 }, { "epoch": 0.7032, "grad_norm": 1.339572548866272, "learning_rate": 1.0102786518221997e-05, "loss": 0.7606, "num_input_tokens_seen": 23565424, "step": 4395 }, { "epoch": 0.704, "grad_norm": 0.9340786337852478, "learning_rate": 1.0052375423562038e-05, "loss": 0.7306, "num_input_tokens_seen": 23592256, "step": 4400 }, { "epoch": 0.7048, "grad_norm": 1.1283642053604126, "learning_rate": 1.0002058746222806e-05, "loss": 0.6121, "num_input_tokens_seen": 23622240, "step": 4405 }, { "epoch": 0.7056, "grad_norm": 0.7319700717926025, "learning_rate": 9.951836804031794e-06, "loss": 0.6273, "num_input_tokens_seen": 23650544, "step": 4410 }, { "epoch": 0.7064, "grad_norm": 1.1631896495819092, "learning_rate": 9.90170991421808e-06, "loss": 0.7261, "num_input_tokens_seen": 23676016, "step": 4415 }, { "epoch": 0.7072, "grad_norm": 0.7371265292167664, "learning_rate": 9.851678393410343e-06, "loss": 0.8013, "num_input_tokens_seen": 23701232, "step": 4420 }, { "epoch": 0.708, "grad_norm": 0.8485360741615295, "learning_rate": 9.801742557634872e-06, "loss": 0.7189, "num_input_tokens_seen": 23731984, "step": 4425 }, { "epoch": 0.7088, "grad_norm": 0.958996057510376, "learning_rate": 9.751902722313527e-06, "loss": 0.7397, "num_input_tokens_seen": 23756560, "step": 4430 }, { "epoch": 0.7096, "grad_norm": 0.9431763887405396, "learning_rate": 9.702159202261801e-06, "loss": 0.625, "num_input_tokens_seen": 23785504, "step": 4435 }, { "epoch": 0.7104, "grad_norm": 1.2089407444000244, "learning_rate": 9.652512311686809e-06, "loss": 0.7958, "num_input_tokens_seen": 23811840, "step": 4440 }, { "epoch": 0.7112, "grad_norm": 1.212649941444397, "learning_rate": 9.602962364185286e-06, "loss": 0.7092, "num_input_tokens_seen": 23834688, "step": 4445 }, { "epoch": 0.712, "grad_norm": 0.8656122088432312, "learning_rate": 9.553509672741645e-06, "loss": 0.6516, "num_input_tokens_seen": 23858736, "step": 4450 }, { "epoch": 0.7128, "grad_norm": 0.8871904611587524, "learning_rate": 9.504154549725943e-06, "loss": 0.6276, "num_input_tokens_seen": 23883696, "step": 4455 }, { "epoch": 0.7136, "grad_norm": 0.8539274334907532, "learning_rate": 9.454897306891972e-06, "loss": 0.6741, "num_input_tokens_seen": 23909904, "step": 4460 }, { "epoch": 0.7144, "grad_norm": 1.0730023384094238, "learning_rate": 9.405738255375244e-06, "loss": 0.7054, "num_input_tokens_seen": 23933056, "step": 4465 }, { "epoch": 0.7152, "grad_norm": 1.2047233581542969, "learning_rate": 9.356677705691058e-06, "loss": 0.7593, "num_input_tokens_seen": 23957440, "step": 4470 }, { "epoch": 0.716, "grad_norm": 0.8580662608146667, "learning_rate": 9.307715967732491e-06, "loss": 0.6264, "num_input_tokens_seen": 23985088, "step": 4475 }, { "epoch": 0.7168, "grad_norm": 0.884903073310852, "learning_rate": 9.258853350768499e-06, "loss": 0.6596, "num_input_tokens_seen": 24010448, "step": 4480 }, { "epoch": 0.7176, "grad_norm": 0.9185977578163147, "learning_rate": 9.210090163441929e-06, "loss": 0.7053, "num_input_tokens_seen": 24035040, "step": 4485 }, { "epoch": 0.7184, "grad_norm": 1.3897747993469238, "learning_rate": 9.161426713767574e-06, "loss": 0.6769, "num_input_tokens_seen": 24057872, "step": 4490 }, { "epoch": 0.7192, "grad_norm": 1.220688819885254, "learning_rate": 9.112863309130235e-06, "loss": 0.7486, "num_input_tokens_seen": 24077920, "step": 4495 }, { "epoch": 0.72, "grad_norm": 0.9036649465560913, "learning_rate": 9.064400256282757e-06, "loss": 0.765, "num_input_tokens_seen": 24104320, "step": 4500 }, { "epoch": 0.7208, "grad_norm": 0.7980133891105652, "learning_rate": 9.016037861344129e-06, "loss": 0.653, "num_input_tokens_seen": 24134144, "step": 4505 }, { "epoch": 0.7216, "grad_norm": 0.7849147915840149, "learning_rate": 8.967776429797528e-06, "loss": 0.6412, "num_input_tokens_seen": 24164576, "step": 4510 }, { "epoch": 0.7224, "grad_norm": 0.8543937802314758, "learning_rate": 8.919616266488373e-06, "loss": 0.7113, "num_input_tokens_seen": 24192736, "step": 4515 }, { "epoch": 0.7232, "grad_norm": 0.9191213250160217, "learning_rate": 8.871557675622441e-06, "loss": 0.8171, "num_input_tokens_seen": 24218064, "step": 4520 }, { "epoch": 0.724, "grad_norm": 1.1177440881729126, "learning_rate": 8.8236009607639e-06, "loss": 0.7845, "num_input_tokens_seen": 24244832, "step": 4525 }, { "epoch": 0.7248, "grad_norm": 0.899111807346344, "learning_rate": 8.775746424833427e-06, "loss": 0.7025, "num_input_tokens_seen": 24272848, "step": 4530 }, { "epoch": 0.7256, "grad_norm": 1.1424217224121094, "learning_rate": 8.727994370106288e-06, "loss": 0.868, "num_input_tokens_seen": 24298240, "step": 4535 }, { "epoch": 0.7264, "grad_norm": 0.9559049010276794, "learning_rate": 8.680345098210408e-06, "loss": 0.6285, "num_input_tokens_seen": 24327776, "step": 4540 }, { "epoch": 0.7272, "grad_norm": 0.9032924771308899, "learning_rate": 8.632798910124492e-06, "loss": 0.6583, "num_input_tokens_seen": 24355424, "step": 4545 }, { "epoch": 0.728, "grad_norm": 1.056780457496643, "learning_rate": 8.585356106176094e-06, "loss": 0.756, "num_input_tokens_seen": 24381104, "step": 4550 }, { "epoch": 0.7288, "grad_norm": 1.2001997232437134, "learning_rate": 8.538016986039754e-06, "loss": 0.7739, "num_input_tokens_seen": 24403760, "step": 4555 }, { "epoch": 0.7296, "grad_norm": 1.1103582382202148, "learning_rate": 8.49078184873508e-06, "loss": 0.6998, "num_input_tokens_seen": 24431280, "step": 4560 }, { "epoch": 0.7304, "grad_norm": 1.0271393060684204, "learning_rate": 8.443650992624877e-06, "loss": 0.723, "num_input_tokens_seen": 24459120, "step": 4565 }, { "epoch": 0.7312, "grad_norm": 0.7871257066726685, "learning_rate": 8.39662471541325e-06, "loss": 0.7225, "num_input_tokens_seen": 24485152, "step": 4570 }, { "epoch": 0.732, "grad_norm": 0.8319628238677979, "learning_rate": 8.34970331414371e-06, "loss": 0.5801, "num_input_tokens_seen": 24512416, "step": 4575 }, { "epoch": 0.7328, "grad_norm": 0.7009981274604797, "learning_rate": 8.302887085197341e-06, "loss": 0.6043, "num_input_tokens_seen": 24543328, "step": 4580 }, { "epoch": 0.7336, "grad_norm": 1.0223398208618164, "learning_rate": 8.256176324290885e-06, "loss": 0.6533, "num_input_tokens_seen": 24566000, "step": 4585 }, { "epoch": 0.7344, "grad_norm": 1.127424955368042, "learning_rate": 8.209571326474896e-06, "loss": 0.6906, "num_input_tokens_seen": 24594080, "step": 4590 }, { "epoch": 0.7352, "grad_norm": 0.9771124124526978, "learning_rate": 8.163072386131876e-06, "loss": 0.6661, "num_input_tokens_seen": 24621424, "step": 4595 }, { "epoch": 0.736, "grad_norm": 0.859312117099762, "learning_rate": 8.116679796974392e-06, "loss": 0.6663, "num_input_tokens_seen": 24644288, "step": 4600 }, { "epoch": 0.7368, "grad_norm": 1.3819899559020996, "learning_rate": 8.070393852043251e-06, "loss": 0.7064, "num_input_tokens_seen": 24674048, "step": 4605 }, { "epoch": 0.7376, "grad_norm": 1.034734845161438, "learning_rate": 8.024214843705646e-06, "loss": 0.6837, "num_input_tokens_seen": 24696320, "step": 4610 }, { "epoch": 0.7384, "grad_norm": 0.9610295295715332, "learning_rate": 7.978143063653298e-06, "loss": 0.5342, "num_input_tokens_seen": 24729280, "step": 4615 }, { "epoch": 0.7392, "grad_norm": 1.166585922241211, "learning_rate": 7.93217880290059e-06, "loss": 0.6907, "num_input_tokens_seen": 24758080, "step": 4620 }, { "epoch": 0.74, "grad_norm": 1.1341148614883423, "learning_rate": 7.886322351782783e-06, "loss": 0.6856, "num_input_tokens_seen": 24787472, "step": 4625 }, { "epoch": 0.7408, "grad_norm": 0.9481520056724548, "learning_rate": 7.840573999954153e-06, "loss": 0.713, "num_input_tokens_seen": 24815936, "step": 4630 }, { "epoch": 0.7416, "grad_norm": 1.2403899431228638, "learning_rate": 7.79493403638614e-06, "loss": 0.7692, "num_input_tokens_seen": 24840096, "step": 4635 }, { "epoch": 0.7424, "grad_norm": 0.9576728343963623, "learning_rate": 7.749402749365572e-06, "loss": 0.7177, "num_input_tokens_seen": 24866480, "step": 4640 }, { "epoch": 0.7432, "grad_norm": 1.0239994525909424, "learning_rate": 7.703980426492791e-06, "loss": 0.7124, "num_input_tokens_seen": 24889456, "step": 4645 }, { "epoch": 0.744, "grad_norm": 1.0492584705352783, "learning_rate": 7.658667354679883e-06, "loss": 0.7038, "num_input_tokens_seen": 24913824, "step": 4650 }, { "epoch": 0.7448, "grad_norm": 1.1247596740722656, "learning_rate": 7.613463820148831e-06, "loss": 0.6662, "num_input_tokens_seen": 24940880, "step": 4655 }, { "epoch": 0.7456, "grad_norm": 1.2390748262405396, "learning_rate": 7.568370108429732e-06, "loss": 0.7949, "num_input_tokens_seen": 24965696, "step": 4660 }, { "epoch": 0.7464, "grad_norm": 0.7792567610740662, "learning_rate": 7.523386504358984e-06, "loss": 0.7146, "num_input_tokens_seen": 24992096, "step": 4665 }, { "epoch": 0.7472, "grad_norm": 0.9417341351509094, "learning_rate": 7.478513292077463e-06, "loss": 0.669, "num_input_tokens_seen": 25024320, "step": 4670 }, { "epoch": 0.748, "grad_norm": 1.226563572883606, "learning_rate": 7.433750755028773e-06, "loss": 0.7789, "num_input_tokens_seen": 25049152, "step": 4675 }, { "epoch": 0.7488, "grad_norm": 0.8685075640678406, "learning_rate": 7.389099175957429e-06, "loss": 0.6992, "num_input_tokens_seen": 25077328, "step": 4680 }, { "epoch": 0.7496, "grad_norm": 0.7221574187278748, "learning_rate": 7.344558836907067e-06, "loss": 0.6421, "num_input_tokens_seen": 25105008, "step": 4685 }, { "epoch": 0.7504, "grad_norm": 0.8624604940414429, "learning_rate": 7.300130019218687e-06, "loss": 0.7656, "num_input_tokens_seen": 25131392, "step": 4690 }, { "epoch": 0.7512, "grad_norm": 1.0061527490615845, "learning_rate": 7.255813003528833e-06, "loss": 0.6506, "num_input_tokens_seen": 25159984, "step": 4695 }, { "epoch": 0.752, "grad_norm": 1.0879433155059814, "learning_rate": 7.211608069767867e-06, "loss": 0.6253, "num_input_tokens_seen": 25188192, "step": 4700 }, { "epoch": 0.7528, "grad_norm": 1.2521827220916748, "learning_rate": 7.1675154971581785e-06, "loss": 0.6776, "num_input_tokens_seen": 25215360, "step": 4705 }, { "epoch": 0.7536, "grad_norm": 1.4772545099258423, "learning_rate": 7.123535564212422e-06, "loss": 0.8286, "num_input_tokens_seen": 25240384, "step": 4710 }, { "epoch": 0.7544, "grad_norm": 1.0587224960327148, "learning_rate": 7.079668548731758e-06, "loss": 0.6152, "num_input_tokens_seen": 25263104, "step": 4715 }, { "epoch": 0.7552, "grad_norm": 1.562467336654663, "learning_rate": 7.035914727804085e-06, "loss": 0.7227, "num_input_tokens_seen": 25288176, "step": 4720 }, { "epoch": 0.756, "grad_norm": 1.3081474304199219, "learning_rate": 6.992274377802327e-06, "loss": 0.6808, "num_input_tokens_seen": 25313536, "step": 4725 }, { "epoch": 0.7568, "grad_norm": 1.1339465379714966, "learning_rate": 6.94874777438265e-06, "loss": 0.7039, "num_input_tokens_seen": 25339744, "step": 4730 }, { "epoch": 0.7576, "grad_norm": 0.8575751185417175, "learning_rate": 6.905335192482737e-06, "loss": 0.7081, "num_input_tokens_seen": 25367440, "step": 4735 }, { "epoch": 0.7584, "grad_norm": 0.9493206143379211, "learning_rate": 6.862036906320058e-06, "loss": 0.6139, "num_input_tokens_seen": 25395952, "step": 4740 }, { "epoch": 0.7592, "grad_norm": 1.0198074579238892, "learning_rate": 6.818853189390104e-06, "loss": 0.8142, "num_input_tokens_seen": 25421744, "step": 4745 }, { "epoch": 0.76, "grad_norm": 0.8722714185714722, "learning_rate": 6.775784314464717e-06, "loss": 0.6538, "num_input_tokens_seen": 25448944, "step": 4750 }, { "epoch": 0.7608, "grad_norm": 0.750995397567749, "learning_rate": 6.732830553590305e-06, "loss": 0.6409, "num_input_tokens_seen": 25476640, "step": 4755 }, { "epoch": 0.7616, "grad_norm": 1.152645230293274, "learning_rate": 6.689992178086174e-06, "loss": 0.6814, "num_input_tokens_seen": 25503328, "step": 4760 }, { "epoch": 0.7624, "grad_norm": 0.9856323599815369, "learning_rate": 6.647269458542793e-06, "loss": 0.739, "num_input_tokens_seen": 25530384, "step": 4765 }, { "epoch": 0.7632, "grad_norm": 1.0248849391937256, "learning_rate": 6.604662664820063e-06, "loss": 0.6775, "num_input_tokens_seen": 25558880, "step": 4770 }, { "epoch": 0.764, "grad_norm": 1.1603997945785522, "learning_rate": 6.562172066045655e-06, "loss": 0.8087, "num_input_tokens_seen": 25584016, "step": 4775 }, { "epoch": 0.7648, "grad_norm": 1.0015392303466797, "learning_rate": 6.519797930613289e-06, "loss": 0.6836, "num_input_tokens_seen": 25611712, "step": 4780 }, { "epoch": 0.7656, "grad_norm": 0.929892897605896, "learning_rate": 6.4775405261810364e-06, "loss": 0.7174, "num_input_tokens_seen": 25640928, "step": 4785 }, { "epoch": 0.7664, "grad_norm": 1.0972721576690674, "learning_rate": 6.435400119669618e-06, "loss": 0.6151, "num_input_tokens_seen": 25667376, "step": 4790 }, { "epoch": 0.7672, "grad_norm": 1.1308437585830688, "learning_rate": 6.3933769772607535e-06, "loss": 0.7291, "num_input_tokens_seen": 25697136, "step": 4795 }, { "epoch": 0.768, "grad_norm": 0.9035334587097168, "learning_rate": 6.3514713643954475e-06, "loss": 0.7215, "num_input_tokens_seen": 25718912, "step": 4800 }, { "epoch": 0.7688, "grad_norm": 0.7563897371292114, "learning_rate": 6.309683545772327e-06, "loss": 0.7092, "num_input_tokens_seen": 25746400, "step": 4805 }, { "epoch": 0.7696, "grad_norm": 0.8460260629653931, "learning_rate": 6.268013785345969e-06, "loss": 0.6675, "num_input_tokens_seen": 25772192, "step": 4810 }, { "epoch": 0.7704, "grad_norm": 0.8550633788108826, "learning_rate": 6.226462346325221e-06, "loss": 0.7428, "num_input_tokens_seen": 25802256, "step": 4815 }, { "epoch": 0.7712, "grad_norm": 1.0745741128921509, "learning_rate": 6.185029491171554e-06, "loss": 0.7039, "num_input_tokens_seen": 25829952, "step": 4820 }, { "epoch": 0.772, "grad_norm": 0.9149506688117981, "learning_rate": 6.143715481597404e-06, "loss": 0.6733, "num_input_tokens_seen": 25854752, "step": 4825 }, { "epoch": 0.7728, "grad_norm": 1.1725239753723145, "learning_rate": 6.102520578564508e-06, "loss": 0.7872, "num_input_tokens_seen": 25881264, "step": 4830 }, { "epoch": 0.7736, "grad_norm": 1.1597212553024292, "learning_rate": 6.061445042282271e-06, "loss": 0.7586, "num_input_tokens_seen": 25906064, "step": 4835 }, { "epoch": 0.7744, "grad_norm": 0.9395809173583984, "learning_rate": 6.020489132206089e-06, "loss": 0.7269, "num_input_tokens_seen": 25931280, "step": 4840 }, { "epoch": 0.7752, "grad_norm": 0.8174002170562744, "learning_rate": 5.979653107035757e-06, "loss": 0.6304, "num_input_tokens_seen": 25958880, "step": 4845 }, { "epoch": 0.776, "grad_norm": 0.9226968884468079, "learning_rate": 5.9389372247138e-06, "loss": 0.6855, "num_input_tokens_seen": 25984528, "step": 4850 }, { "epoch": 0.7768, "grad_norm": 1.0765284299850464, "learning_rate": 5.898341742423865e-06, "loss": 0.7141, "num_input_tokens_seen": 26014272, "step": 4855 }, { "epoch": 0.7776, "grad_norm": 0.7940208315849304, "learning_rate": 5.857866916589089e-06, "loss": 0.668, "num_input_tokens_seen": 26045888, "step": 4860 }, { "epoch": 0.7784, "grad_norm": 0.9069024920463562, "learning_rate": 5.81751300287045e-06, "loss": 0.7404, "num_input_tokens_seen": 26069232, "step": 4865 }, { "epoch": 0.7792, "grad_norm": 1.2687326669692993, "learning_rate": 5.777280256165218e-06, "loss": 0.633, "num_input_tokens_seen": 26095936, "step": 4870 }, { "epoch": 0.78, "grad_norm": 1.0579140186309814, "learning_rate": 5.737168930605272e-06, "loss": 0.6365, "num_input_tokens_seen": 26121184, "step": 4875 }, { "epoch": 0.7808, "grad_norm": 0.8767179846763611, "learning_rate": 5.6971792795555505e-06, "loss": 0.6427, "num_input_tokens_seen": 26147504, "step": 4880 }, { "epoch": 0.7816, "grad_norm": 0.9713358283042908, "learning_rate": 5.6573115556124325e-06, "loss": 0.6509, "num_input_tokens_seen": 26174208, "step": 4885 }, { "epoch": 0.7824, "grad_norm": 0.7532449960708618, "learning_rate": 5.617566010602113e-06, "loss": 0.7382, "num_input_tokens_seen": 26200112, "step": 4890 }, { "epoch": 0.7832, "grad_norm": 1.191658854484558, "learning_rate": 5.577942895579064e-06, "loss": 0.7537, "num_input_tokens_seen": 26227952, "step": 4895 }, { "epoch": 0.784, "grad_norm": 0.9605312943458557, "learning_rate": 5.538442460824417e-06, "loss": 0.673, "num_input_tokens_seen": 26259392, "step": 4900 }, { "epoch": 0.7848, "grad_norm": 1.0074589252471924, "learning_rate": 5.499064955844382e-06, "loss": 0.6684, "num_input_tokens_seen": 26285456, "step": 4905 }, { "epoch": 0.7856, "grad_norm": 0.8559053540229797, "learning_rate": 5.4598106293686916e-06, "loss": 0.7051, "num_input_tokens_seen": 26316544, "step": 4910 }, { "epoch": 0.7864, "grad_norm": 1.0885223150253296, "learning_rate": 5.420679729348993e-06, "loss": 0.6481, "num_input_tokens_seen": 26341840, "step": 4915 }, { "epoch": 0.7872, "grad_norm": 1.0094472169876099, "learning_rate": 5.381672502957324e-06, "loss": 0.7953, "num_input_tokens_seen": 26371008, "step": 4920 }, { "epoch": 0.788, "grad_norm": 1.1841135025024414, "learning_rate": 5.342789196584527e-06, "loss": 0.6966, "num_input_tokens_seen": 26400048, "step": 4925 }, { "epoch": 0.7888, "grad_norm": 0.9742115139961243, "learning_rate": 5.304030055838705e-06, "loss": 0.6804, "num_input_tokens_seen": 26425408, "step": 4930 }, { "epoch": 0.7896, "grad_norm": 1.327427625656128, "learning_rate": 5.26539532554364e-06, "loss": 0.6282, "num_input_tokens_seen": 26452352, "step": 4935 }, { "epoch": 0.7904, "grad_norm": 0.9497706890106201, "learning_rate": 5.226885249737293e-06, "loss": 0.588, "num_input_tokens_seen": 26479456, "step": 4940 }, { "epoch": 0.7912, "grad_norm": 1.0179611444473267, "learning_rate": 5.1885000716702355e-06, "loss": 0.738, "num_input_tokens_seen": 26504912, "step": 4945 }, { "epoch": 0.792, "grad_norm": 1.0633511543273926, "learning_rate": 5.150240033804116e-06, "loss": 0.661, "num_input_tokens_seen": 26528320, "step": 4950 }, { "epoch": 0.7928, "grad_norm": 0.887589693069458, "learning_rate": 5.112105377810128e-06, "loss": 0.8033, "num_input_tokens_seen": 26553984, "step": 4955 }, { "epoch": 0.7936, "grad_norm": 1.0450173616409302, "learning_rate": 5.074096344567475e-06, "loss": 0.6161, "num_input_tokens_seen": 26582768, "step": 4960 }, { "epoch": 0.7944, "grad_norm": 1.2726836204528809, "learning_rate": 5.036213174161877e-06, "loss": 0.7286, "num_input_tokens_seen": 26610272, "step": 4965 }, { "epoch": 0.7952, "grad_norm": 1.0096362829208374, "learning_rate": 4.998456105884025e-06, "loss": 0.7065, "num_input_tokens_seen": 26636352, "step": 4970 }, { "epoch": 0.796, "grad_norm": 1.1659733057022095, "learning_rate": 4.960825378228082e-06, "loss": 0.6842, "num_input_tokens_seen": 26667824, "step": 4975 }, { "epoch": 0.7968, "grad_norm": 0.9078534245491028, "learning_rate": 4.9233212288901845e-06, "loss": 0.7069, "num_input_tokens_seen": 26698272, "step": 4980 }, { "epoch": 0.7976, "grad_norm": 0.8440881967544556, "learning_rate": 4.885943894766909e-06, "loss": 0.5942, "num_input_tokens_seen": 26725984, "step": 4985 }, { "epoch": 0.7984, "grad_norm": 1.0457020998001099, "learning_rate": 4.848693611953825e-06, "loss": 0.8419, "num_input_tokens_seen": 26751360, "step": 4990 }, { "epoch": 0.7992, "grad_norm": 0.947726309299469, "learning_rate": 4.811570615743952e-06, "loss": 0.5888, "num_input_tokens_seen": 26782672, "step": 4995 }, { "epoch": 0.8, "grad_norm": 1.1138640642166138, "learning_rate": 4.7745751406263165e-06, "loss": 0.6504, "num_input_tokens_seen": 26809120, "step": 5000 }, { "epoch": 0.8008, "grad_norm": 0.9211150407791138, "learning_rate": 4.737707420284451e-06, "loss": 0.6603, "num_input_tokens_seen": 26839552, "step": 5005 }, { "epoch": 0.8016, "grad_norm": 1.2926892042160034, "learning_rate": 4.700967687594901e-06, "loss": 0.627, "num_input_tokens_seen": 26864416, "step": 5010 }, { "epoch": 0.8024, "grad_norm": 0.9436898827552795, "learning_rate": 4.664356174625795e-06, "loss": 0.6509, "num_input_tokens_seen": 26890368, "step": 5015 }, { "epoch": 0.8032, "grad_norm": 0.8215711712837219, "learning_rate": 4.627873112635345e-06, "loss": 0.6673, "num_input_tokens_seen": 26916064, "step": 5020 }, { "epoch": 0.804, "grad_norm": 0.9311307072639465, "learning_rate": 4.591518732070402e-06, "loss": 0.7972, "num_input_tokens_seen": 26940528, "step": 5025 }, { "epoch": 0.8048, "grad_norm": 1.1058831214904785, "learning_rate": 4.5552932625649944e-06, "loss": 0.6977, "num_input_tokens_seen": 26965296, "step": 5030 }, { "epoch": 0.8056, "grad_norm": 1.2519973516464233, "learning_rate": 4.5191969329388625e-06, "loss": 0.8094, "num_input_tokens_seen": 26988240, "step": 5035 }, { "epoch": 0.8064, "grad_norm": 0.9225587248802185, "learning_rate": 4.483229971196054e-06, "loss": 0.7441, "num_input_tokens_seen": 27015632, "step": 5040 }, { "epoch": 0.8072, "grad_norm": 0.8636001944541931, "learning_rate": 4.44739260452344e-06, "loss": 0.6618, "num_input_tokens_seen": 27040528, "step": 5045 }, { "epoch": 0.808, "grad_norm": 1.094529151916504, "learning_rate": 4.411685059289314e-06, "loss": 0.7527, "num_input_tokens_seen": 27066560, "step": 5050 }, { "epoch": 0.8088, "grad_norm": 0.9794814586639404, "learning_rate": 4.376107561041937e-06, "loss": 0.7844, "num_input_tokens_seen": 27089408, "step": 5055 }, { "epoch": 0.8096, "grad_norm": 0.8608947396278381, "learning_rate": 4.340660334508115e-06, "loss": 0.7598, "num_input_tokens_seen": 27114832, "step": 5060 }, { "epoch": 0.8104, "grad_norm": 0.8481171727180481, "learning_rate": 4.305343603591802e-06, "loss": 0.6645, "num_input_tokens_seen": 27140320, "step": 5065 }, { "epoch": 0.8112, "grad_norm": 0.9115588068962097, "learning_rate": 4.270157591372667e-06, "loss": 0.7065, "num_input_tokens_seen": 27171200, "step": 5070 }, { "epoch": 0.812, "grad_norm": 1.4465726613998413, "learning_rate": 4.235102520104681e-06, "loss": 0.7481, "num_input_tokens_seen": 27195056, "step": 5075 }, { "epoch": 0.8128, "grad_norm": 1.1063685417175293, "learning_rate": 4.200178611214736e-06, "loss": 0.6154, "num_input_tokens_seen": 27220816, "step": 5080 }, { "epoch": 0.8136, "grad_norm": 1.3894023895263672, "learning_rate": 4.165386085301212e-06, "loss": 0.6661, "num_input_tokens_seen": 27246400, "step": 5085 }, { "epoch": 0.8144, "grad_norm": 1.0136696100234985, "learning_rate": 4.130725162132612e-06, "loss": 0.7043, "num_input_tokens_seen": 27269936, "step": 5090 }, { "epoch": 0.8152, "grad_norm": 1.36388099193573, "learning_rate": 4.096196060646168e-06, "loss": 0.8173, "num_input_tokens_seen": 27293488, "step": 5095 }, { "epoch": 0.816, "grad_norm": 0.9864152669906616, "learning_rate": 4.061798998946459e-06, "loss": 0.7154, "num_input_tokens_seen": 27318592, "step": 5100 }, { "epoch": 0.8168, "grad_norm": 0.9855313301086426, "learning_rate": 4.027534194304005e-06, "loss": 0.6336, "num_input_tokens_seen": 27343616, "step": 5105 }, { "epoch": 0.8176, "grad_norm": 0.9487363696098328, "learning_rate": 3.99340186315395e-06, "loss": 0.7355, "num_input_tokens_seen": 27369216, "step": 5110 }, { "epoch": 0.8184, "grad_norm": 0.8174062371253967, "learning_rate": 3.959402221094635e-06, "loss": 0.6034, "num_input_tokens_seen": 27398704, "step": 5115 }, { "epoch": 0.8192, "grad_norm": 1.17109215259552, "learning_rate": 3.925535482886286e-06, "loss": 0.7962, "num_input_tokens_seen": 27424176, "step": 5120 }, { "epoch": 0.82, "grad_norm": 0.9322016835212708, "learning_rate": 3.891801862449629e-06, "loss": 0.7289, "num_input_tokens_seen": 27452656, "step": 5125 }, { "epoch": 0.8208, "grad_norm": 0.8374980688095093, "learning_rate": 3.858201572864537e-06, "loss": 0.6644, "num_input_tokens_seen": 27478656, "step": 5130 }, { "epoch": 0.8216, "grad_norm": 0.8750137686729431, "learning_rate": 3.824734826368703e-06, "loss": 0.7519, "num_input_tokens_seen": 27507184, "step": 5135 }, { "epoch": 0.8224, "grad_norm": 0.935739278793335, "learning_rate": 3.7914018343562895e-06, "loss": 0.7611, "num_input_tokens_seen": 27536112, "step": 5140 }, { "epoch": 0.8232, "grad_norm": 1.0115796327590942, "learning_rate": 3.75820280737659e-06, "loss": 0.6713, "num_input_tokens_seen": 27563728, "step": 5145 }, { "epoch": 0.824, "grad_norm": 1.1365923881530762, "learning_rate": 3.725137955132707e-06, "loss": 0.6522, "num_input_tokens_seen": 27587120, "step": 5150 }, { "epoch": 0.8248, "grad_norm": 1.2187998294830322, "learning_rate": 3.692207486480209e-06, "loss": 0.7707, "num_input_tokens_seen": 27608240, "step": 5155 }, { "epoch": 0.8256, "grad_norm": 1.2192776203155518, "learning_rate": 3.6594116094258337e-06, "loss": 0.6148, "num_input_tokens_seen": 27637840, "step": 5160 }, { "epoch": 0.8264, "grad_norm": 0.8527853488922119, "learning_rate": 3.626750531126169e-06, "loss": 0.6884, "num_input_tokens_seen": 27662144, "step": 5165 }, { "epoch": 0.8272, "grad_norm": 1.293915033340454, "learning_rate": 3.594224457886336e-06, "loss": 0.6954, "num_input_tokens_seen": 27691184, "step": 5170 }, { "epoch": 0.828, "grad_norm": 0.9655611515045166, "learning_rate": 3.561833595158698e-06, "loss": 0.6736, "num_input_tokens_seen": 27719376, "step": 5175 }, { "epoch": 0.8288, "grad_norm": 0.8014153838157654, "learning_rate": 3.529578147541532e-06, "loss": 0.7575, "num_input_tokens_seen": 27749248, "step": 5180 }, { "epoch": 0.8296, "grad_norm": 0.8693012595176697, "learning_rate": 3.4974583187777852e-06, "loss": 0.6607, "num_input_tokens_seen": 27779136, "step": 5185 }, { "epoch": 0.8304, "grad_norm": 1.2928762435913086, "learning_rate": 3.4654743117537524e-06, "loss": 0.844, "num_input_tokens_seen": 27802592, "step": 5190 }, { "epoch": 0.8312, "grad_norm": 0.9536007046699524, "learning_rate": 3.433626328497805e-06, "loss": 0.6596, "num_input_tokens_seen": 27826864, "step": 5195 }, { "epoch": 0.832, "grad_norm": 1.01909601688385, "learning_rate": 3.4019145701791184e-06, "loss": 0.7834, "num_input_tokens_seen": 27851680, "step": 5200 }, { "epoch": 0.8328, "grad_norm": 0.8944137096405029, "learning_rate": 3.3703392371063845e-06, "loss": 0.6874, "num_input_tokens_seen": 27880208, "step": 5205 }, { "epoch": 0.8336, "grad_norm": 1.2721880674362183, "learning_rate": 3.338900528726571e-06, "loss": 0.6468, "num_input_tokens_seen": 27907392, "step": 5210 }, { "epoch": 0.8344, "grad_norm": 1.3406318426132202, "learning_rate": 3.3075986436236493e-06, "loss": 0.6675, "num_input_tokens_seen": 27934560, "step": 5215 }, { "epoch": 0.8352, "grad_norm": 0.9489244818687439, "learning_rate": 3.2764337795173435e-06, "loss": 0.6704, "num_input_tokens_seen": 27963248, "step": 5220 }, { "epoch": 0.836, "grad_norm": 0.9695754647254944, "learning_rate": 3.245406133261858e-06, "loss": 0.6903, "num_input_tokens_seen": 27989872, "step": 5225 }, { "epoch": 0.8368, "grad_norm": 0.7802848219871521, "learning_rate": 3.2145159008446807e-06, "loss": 0.7441, "num_input_tokens_seen": 28012208, "step": 5230 }, { "epoch": 0.8376, "grad_norm": 1.2693451642990112, "learning_rate": 3.1837632773853098e-06, "loss": 0.6636, "num_input_tokens_seen": 28041200, "step": 5235 }, { "epoch": 0.8384, "grad_norm": 1.1437602043151855, "learning_rate": 3.15314845713402e-06, "loss": 0.7246, "num_input_tokens_seen": 28068272, "step": 5240 }, { "epoch": 0.8392, "grad_norm": 0.8950793743133545, "learning_rate": 3.122671633470664e-06, "loss": 0.6583, "num_input_tokens_seen": 28092768, "step": 5245 }, { "epoch": 0.84, "grad_norm": 1.010606288909912, "learning_rate": 3.0923329989034132e-06, "loss": 0.6823, "num_input_tokens_seen": 28122944, "step": 5250 }, { "epoch": 0.8408, "grad_norm": 1.0807864665985107, "learning_rate": 3.062132745067581e-06, "loss": 0.7467, "num_input_tokens_seen": 28151424, "step": 5255 }, { "epoch": 0.8416, "grad_norm": 1.1845099925994873, "learning_rate": 3.0320710627243813e-06, "loss": 0.7541, "num_input_tokens_seen": 28176752, "step": 5260 }, { "epoch": 0.8424, "grad_norm": 0.7790080904960632, "learning_rate": 3.002148141759739e-06, "loss": 0.6829, "num_input_tokens_seen": 28205456, "step": 5265 }, { "epoch": 0.8432, "grad_norm": 1.364443302154541, "learning_rate": 2.97236417118309e-06, "loss": 0.692, "num_input_tokens_seen": 28230304, "step": 5270 }, { "epoch": 0.844, "grad_norm": 1.1289632320404053, "learning_rate": 2.942719339126171e-06, "loss": 0.7628, "num_input_tokens_seen": 28255536, "step": 5275 }, { "epoch": 0.8448, "grad_norm": 1.3984529972076416, "learning_rate": 2.9132138328418573e-06, "loss": 0.6972, "num_input_tokens_seen": 28279600, "step": 5280 }, { "epoch": 0.8456, "grad_norm": 1.0336651802062988, "learning_rate": 2.8838478387029606e-06, "loss": 0.6801, "num_input_tokens_seen": 28304688, "step": 5285 }, { "epoch": 0.8464, "grad_norm": 0.8985416293144226, "learning_rate": 2.8546215422010638e-06, "loss": 0.6697, "num_input_tokens_seen": 28331584, "step": 5290 }, { "epoch": 0.8472, "grad_norm": 1.3255959749221802, "learning_rate": 2.8255351279453446e-06, "loss": 0.6816, "num_input_tokens_seen": 28360256, "step": 5295 }, { "epoch": 0.848, "grad_norm": 0.9332570433616638, "learning_rate": 2.7965887796613884e-06, "loss": 0.7763, "num_input_tokens_seen": 28385168, "step": 5300 }, { "epoch": 0.8488, "grad_norm": 1.041506290435791, "learning_rate": 2.767782680190073e-06, "loss": 0.7837, "num_input_tokens_seen": 28407248, "step": 5305 }, { "epoch": 0.8496, "grad_norm": 1.075434923171997, "learning_rate": 2.739117011486378e-06, "loss": 0.6975, "num_input_tokens_seen": 28434304, "step": 5310 }, { "epoch": 0.8504, "grad_norm": 1.0231233835220337, "learning_rate": 2.710591954618247e-06, "loss": 0.7119, "num_input_tokens_seen": 28465424, "step": 5315 }, { "epoch": 0.8512, "grad_norm": 1.0655337572097778, "learning_rate": 2.6822076897654452e-06, "loss": 0.7644, "num_input_tokens_seen": 28494416, "step": 5320 }, { "epoch": 0.852, "grad_norm": 1.177964448928833, "learning_rate": 2.6539643962184057e-06, "loss": 0.644, "num_input_tokens_seen": 28519552, "step": 5325 }, { "epoch": 0.8528, "grad_norm": 1.0055785179138184, "learning_rate": 2.6258622523771287e-06, "loss": 0.7378, "num_input_tokens_seen": 28545632, "step": 5330 }, { "epoch": 0.8536, "grad_norm": 1.2302112579345703, "learning_rate": 2.5979014357500248e-06, "loss": 0.7267, "num_input_tokens_seen": 28571440, "step": 5335 }, { "epoch": 0.8544, "grad_norm": 1.2346506118774414, "learning_rate": 2.570082122952816e-06, "loss": 0.6015, "num_input_tokens_seen": 28599472, "step": 5340 }, { "epoch": 0.8552, "grad_norm": 0.8938889503479004, "learning_rate": 2.5424044897073895e-06, "loss": 0.6327, "num_input_tokens_seen": 28629136, "step": 5345 }, { "epoch": 0.856, "grad_norm": 0.8958183526992798, "learning_rate": 2.514868710840723e-06, "loss": 0.739, "num_input_tokens_seen": 28648928, "step": 5350 }, { "epoch": 0.8568, "grad_norm": 1.0484833717346191, "learning_rate": 2.4874749602837697e-06, "loss": 0.7279, "num_input_tokens_seen": 28675056, "step": 5355 }, { "epoch": 0.8576, "grad_norm": 0.887750506401062, "learning_rate": 2.4602234110703364e-06, "loss": 0.726, "num_input_tokens_seen": 28698416, "step": 5360 }, { "epoch": 0.8584, "grad_norm": 0.7482147216796875, "learning_rate": 2.43311423533602e-06, "loss": 0.7046, "num_input_tokens_seen": 28729856, "step": 5365 }, { "epoch": 0.8592, "grad_norm": 1.003188967704773, "learning_rate": 2.406147604317119e-06, "loss": 0.6922, "num_input_tokens_seen": 28757360, "step": 5370 }, { "epoch": 0.86, "grad_norm": 1.5129293203353882, "learning_rate": 2.379323688349516e-06, "loss": 0.6664, "num_input_tokens_seen": 28780624, "step": 5375 }, { "epoch": 0.8608, "grad_norm": 0.9118067622184753, "learning_rate": 2.3526426568676483e-06, "loss": 0.6532, "num_input_tokens_seen": 28805616, "step": 5380 }, { "epoch": 0.8616, "grad_norm": 1.007717251777649, "learning_rate": 2.326104678403415e-06, "loss": 0.6678, "num_input_tokens_seen": 28833504, "step": 5385 }, { "epoch": 0.8624, "grad_norm": 0.9576060771942139, "learning_rate": 2.299709920585108e-06, "loss": 0.6152, "num_input_tokens_seen": 28862704, "step": 5390 }, { "epoch": 0.8632, "grad_norm": 1.1482651233673096, "learning_rate": 2.2734585501363673e-06, "loss": 0.7131, "num_input_tokens_seen": 28886224, "step": 5395 }, { "epoch": 0.864, "grad_norm": 1.0614038705825806, "learning_rate": 2.2473507328751086e-06, "loss": 0.735, "num_input_tokens_seen": 28911760, "step": 5400 }, { "epoch": 0.8648, "grad_norm": 1.0017966032028198, "learning_rate": 2.2213866337125022e-06, "loss": 0.6706, "num_input_tokens_seen": 28941360, "step": 5405 }, { "epoch": 0.8656, "grad_norm": 0.963431179523468, "learning_rate": 2.1955664166519036e-06, "loss": 0.7683, "num_input_tokens_seen": 28965568, "step": 5410 }, { "epoch": 0.8664, "grad_norm": 1.1716382503509521, "learning_rate": 2.1698902447878477e-06, "loss": 0.623, "num_input_tokens_seen": 28994432, "step": 5415 }, { "epoch": 0.8672, "grad_norm": 1.2640058994293213, "learning_rate": 2.1443582803049755e-06, "loss": 0.7774, "num_input_tokens_seen": 29016560, "step": 5420 }, { "epoch": 0.868, "grad_norm": 0.9828547239303589, "learning_rate": 2.118970684477062e-06, "loss": 0.6332, "num_input_tokens_seen": 29043920, "step": 5425 }, { "epoch": 0.8688, "grad_norm": 0.9180524349212646, "learning_rate": 2.093727617665955e-06, "loss": 0.6658, "num_input_tokens_seen": 29073840, "step": 5430 }, { "epoch": 0.8696, "grad_norm": 1.1137315034866333, "learning_rate": 2.068629239320588e-06, "loss": 0.7078, "num_input_tokens_seen": 29102752, "step": 5435 }, { "epoch": 0.8704, "grad_norm": 1.1765251159667969, "learning_rate": 2.043675707975959e-06, "loss": 0.7049, "num_input_tokens_seen": 29126576, "step": 5440 }, { "epoch": 0.8712, "grad_norm": 0.9444310665130615, "learning_rate": 2.0188671812521292e-06, "loss": 0.7931, "num_input_tokens_seen": 29153120, "step": 5445 }, { "epoch": 0.872, "grad_norm": 0.914959192276001, "learning_rate": 1.9942038158532407e-06, "loss": 0.8394, "num_input_tokens_seen": 29182192, "step": 5450 }, { "epoch": 0.8728, "grad_norm": 1.21523916721344, "learning_rate": 1.969685767566512e-06, "loss": 0.6915, "num_input_tokens_seen": 29206368, "step": 5455 }, { "epoch": 0.8736, "grad_norm": 0.8198549151420593, "learning_rate": 1.9453131912612694e-06, "loss": 0.6627, "num_input_tokens_seen": 29235984, "step": 5460 }, { "epoch": 0.8744, "grad_norm": 0.9284049868583679, "learning_rate": 1.921086240887937e-06, "loss": 0.6671, "num_input_tokens_seen": 29260672, "step": 5465 }, { "epoch": 0.8752, "grad_norm": 0.9254517555236816, "learning_rate": 1.8970050694771064e-06, "loss": 0.665, "num_input_tokens_seen": 29287792, "step": 5470 }, { "epoch": 0.876, "grad_norm": 0.8524356484413147, "learning_rate": 1.8730698291385518e-06, "loss": 0.663, "num_input_tokens_seen": 29314656, "step": 5475 }, { "epoch": 0.8768, "grad_norm": 0.8038460612297058, "learning_rate": 1.8492806710602496e-06, "loss": 0.7054, "num_input_tokens_seen": 29338976, "step": 5480 }, { "epoch": 0.8776, "grad_norm": 0.8914519548416138, "learning_rate": 1.8256377455074525e-06, "loss": 0.6905, "num_input_tokens_seen": 29364912, "step": 5485 }, { "epoch": 0.8784, "grad_norm": 0.8928874135017395, "learning_rate": 1.802141201821736e-06, "loss": 0.7641, "num_input_tokens_seen": 29392960, "step": 5490 }, { "epoch": 0.8792, "grad_norm": 0.7842042446136475, "learning_rate": 1.7787911884200314e-06, "loss": 0.6918, "num_input_tokens_seen": 29416848, "step": 5495 }, { "epoch": 0.88, "grad_norm": 1.335999846458435, "learning_rate": 1.7555878527937164e-06, "loss": 0.6281, "num_input_tokens_seen": 29445856, "step": 5500 }, { "epoch": 0.8808, "grad_norm": 0.981039822101593, "learning_rate": 1.7325313415076705e-06, "loss": 0.7199, "num_input_tokens_seen": 29474400, "step": 5505 }, { "epoch": 0.8816, "grad_norm": 1.003205418586731, "learning_rate": 1.7096218001993513e-06, "loss": 0.7022, "num_input_tokens_seen": 29501312, "step": 5510 }, { "epoch": 0.8824, "grad_norm": 0.804409921169281, "learning_rate": 1.686859373577876e-06, "loss": 0.635, "num_input_tokens_seen": 29530160, "step": 5515 }, { "epoch": 0.8832, "grad_norm": 0.6613283157348633, "learning_rate": 1.6642442054230934e-06, "loss": 0.6752, "num_input_tokens_seen": 29557168, "step": 5520 }, { "epoch": 0.884, "grad_norm": 1.0624265670776367, "learning_rate": 1.6417764385846996e-06, "loss": 0.7824, "num_input_tokens_seen": 29584832, "step": 5525 }, { "epoch": 0.8848, "grad_norm": 1.0792144536972046, "learning_rate": 1.6194562149813242e-06, "loss": 0.6823, "num_input_tokens_seen": 29609504, "step": 5530 }, { "epoch": 0.8856, "grad_norm": 1.0641052722930908, "learning_rate": 1.5972836755996285e-06, "loss": 0.6777, "num_input_tokens_seen": 29636768, "step": 5535 }, { "epoch": 0.8864, "grad_norm": 1.2235264778137207, "learning_rate": 1.5752589604934255e-06, "loss": 0.7372, "num_input_tokens_seen": 29660496, "step": 5540 }, { "epoch": 0.8872, "grad_norm": 0.9904717206954956, "learning_rate": 1.5533822087827805e-06, "loss": 0.7126, "num_input_tokens_seen": 29686928, "step": 5545 }, { "epoch": 0.888, "grad_norm": 1.2209672927856445, "learning_rate": 1.5316535586531483e-06, "loss": 0.6564, "num_input_tokens_seen": 29714800, "step": 5550 }, { "epoch": 0.8888, "grad_norm": 1.2465245723724365, "learning_rate": 1.5100731473544933e-06, "loss": 0.8006, "num_input_tokens_seen": 29741808, "step": 5555 }, { "epoch": 0.8896, "grad_norm": 1.0243083238601685, "learning_rate": 1.4886411112004255e-06, "loss": 0.7322, "num_input_tokens_seen": 29763088, "step": 5560 }, { "epoch": 0.8904, "grad_norm": 0.9879063367843628, "learning_rate": 1.4673575855673277e-06, "loss": 0.7243, "num_input_tokens_seen": 29791520, "step": 5565 }, { "epoch": 0.8912, "grad_norm": 0.8593109250068665, "learning_rate": 1.4462227048935183e-06, "loss": 0.6955, "num_input_tokens_seen": 29817600, "step": 5570 }, { "epoch": 0.892, "grad_norm": 0.9585959911346436, "learning_rate": 1.425236602678387e-06, "loss": 0.6658, "num_input_tokens_seen": 29843136, "step": 5575 }, { "epoch": 0.8928, "grad_norm": 1.2681745290756226, "learning_rate": 1.4043994114815661e-06, "loss": 0.7943, "num_input_tokens_seen": 29864864, "step": 5580 }, { "epoch": 0.8936, "grad_norm": 0.8207817673683167, "learning_rate": 1.38371126292208e-06, "loss": 0.7734, "num_input_tokens_seen": 29890416, "step": 5585 }, { "epoch": 0.8944, "grad_norm": 1.0749483108520508, "learning_rate": 1.3631722876775138e-06, "loss": 0.7008, "num_input_tokens_seen": 29916400, "step": 5590 }, { "epoch": 0.8952, "grad_norm": 1.1091291904449463, "learning_rate": 1.3427826154832042e-06, "loss": 0.6434, "num_input_tokens_seen": 29944304, "step": 5595 }, { "epoch": 0.896, "grad_norm": 1.3164221048355103, "learning_rate": 1.3225423751313942e-06, "loss": 0.6952, "num_input_tokens_seen": 29967648, "step": 5600 }, { "epoch": 0.8968, "grad_norm": 1.2402032613754272, "learning_rate": 1.3024516944704496e-06, "loss": 0.6331, "num_input_tokens_seen": 29989312, "step": 5605 }, { "epoch": 0.8976, "grad_norm": 1.1296372413635254, "learning_rate": 1.2825107004040272e-06, "loss": 0.7894, "num_input_tokens_seen": 30012384, "step": 5610 }, { "epoch": 0.8984, "grad_norm": 1.109983205795288, "learning_rate": 1.2627195188902791e-06, "loss": 0.6819, "num_input_tokens_seen": 30042656, "step": 5615 }, { "epoch": 0.8992, "grad_norm": 1.0760643482208252, "learning_rate": 1.2430782749410673e-06, "loss": 0.8208, "num_input_tokens_seen": 30068464, "step": 5620 }, { "epoch": 0.9, "grad_norm": 1.1865812540054321, "learning_rate": 1.2235870926211619e-06, "loss": 0.7203, "num_input_tokens_seen": 30093216, "step": 5625 }, { "epoch": 0.9008, "grad_norm": 1.0090526342391968, "learning_rate": 1.2042460950474648e-06, "loss": 0.7368, "num_input_tokens_seen": 30117488, "step": 5630 }, { "epoch": 0.9016, "grad_norm": 0.8406433463096619, "learning_rate": 1.1850554043882328e-06, "loss": 0.7681, "num_input_tokens_seen": 30144016, "step": 5635 }, { "epoch": 0.9024, "grad_norm": 1.057853102684021, "learning_rate": 1.1660151418622922e-06, "loss": 0.6962, "num_input_tokens_seen": 30177184, "step": 5640 }, { "epoch": 0.9032, "grad_norm": 0.9097471237182617, "learning_rate": 1.1471254277382881e-06, "loss": 0.7056, "num_input_tokens_seen": 30206048, "step": 5645 }, { "epoch": 0.904, "grad_norm": 1.0608327388763428, "learning_rate": 1.1283863813339263e-06, "loss": 0.7089, "num_input_tokens_seen": 30229936, "step": 5650 }, { "epoch": 0.9048, "grad_norm": 1.1075376272201538, "learning_rate": 1.1097981210152043e-06, "loss": 0.7794, "num_input_tokens_seen": 30257760, "step": 5655 }, { "epoch": 0.9056, "grad_norm": 0.9509792327880859, "learning_rate": 1.0913607641956841e-06, "loss": 0.77, "num_input_tokens_seen": 30286464, "step": 5660 }, { "epoch": 0.9064, "grad_norm": 0.981621265411377, "learning_rate": 1.0730744273357213e-06, "loss": 0.708, "num_input_tokens_seen": 30317040, "step": 5665 }, { "epoch": 0.9072, "grad_norm": 0.8428750038146973, "learning_rate": 1.0549392259417646e-06, "loss": 0.6423, "num_input_tokens_seen": 30342672, "step": 5670 }, { "epoch": 0.908, "grad_norm": 0.9921499490737915, "learning_rate": 1.0369552745656013e-06, "loss": 0.8176, "num_input_tokens_seen": 30369952, "step": 5675 }, { "epoch": 0.9088, "grad_norm": 0.9219129681587219, "learning_rate": 1.0191226868036418e-06, "loss": 0.6924, "num_input_tokens_seen": 30400800, "step": 5680 }, { "epoch": 0.9096, "grad_norm": 1.451660394668579, "learning_rate": 1.001441575296208e-06, "loss": 0.576, "num_input_tokens_seen": 30436240, "step": 5685 }, { "epoch": 0.9104, "grad_norm": 0.9631555676460266, "learning_rate": 9.839120517267985e-07, "loss": 0.6206, "num_input_tokens_seen": 30465232, "step": 5690 }, { "epoch": 0.9112, "grad_norm": 1.125351071357727, "learning_rate": 9.665342268214166e-07, "loss": 0.7424, "num_input_tokens_seen": 30489776, "step": 5695 }, { "epoch": 0.912, "grad_norm": 1.0404316186904907, "learning_rate": 9.493082103478517e-07, "loss": 0.6868, "num_input_tokens_seen": 30514592, "step": 5700 }, { "epoch": 0.9128, "grad_norm": 0.9020105004310608, "learning_rate": 9.322341111149852e-07, "loss": 0.7017, "num_input_tokens_seen": 30544112, "step": 5705 }, { "epoch": 0.9136, "grad_norm": 1.0047924518585205, "learning_rate": 9.153120369721046e-07, "loss": 0.6429, "num_input_tokens_seen": 30577440, "step": 5710 }, { "epoch": 0.9144, "grad_norm": 1.087547779083252, "learning_rate": 8.985420948082329e-07, "loss": 0.6507, "num_input_tokens_seen": 30602704, "step": 5715 }, { "epoch": 0.9152, "grad_norm": 0.9647343754768372, "learning_rate": 8.819243905514308e-07, "loss": 0.7508, "num_input_tokens_seen": 30627360, "step": 5720 }, { "epoch": 0.916, "grad_norm": 0.9557966589927673, "learning_rate": 8.65459029168153e-07, "loss": 0.6506, "num_input_tokens_seen": 30657216, "step": 5725 }, { "epoch": 0.9168, "grad_norm": 0.991233229637146, "learning_rate": 8.491461146625773e-07, "loss": 0.6426, "num_input_tokens_seen": 30683472, "step": 5730 }, { "epoch": 0.9176, "grad_norm": 0.8405401706695557, "learning_rate": 8.329857500759292e-07, "loss": 0.6227, "num_input_tokens_seen": 30707392, "step": 5735 }, { "epoch": 0.9184, "grad_norm": 1.2772481441497803, "learning_rate": 8.169780374858577e-07, "loss": 0.839, "num_input_tokens_seen": 30732160, "step": 5740 }, { "epoch": 0.9192, "grad_norm": 0.8571876883506775, "learning_rate": 8.011230780057749e-07, "loss": 0.6817, "num_input_tokens_seen": 30760336, "step": 5745 }, { "epoch": 0.92, "grad_norm": 1.000186562538147, "learning_rate": 7.854209717842231e-07, "loss": 0.7016, "num_input_tokens_seen": 30788800, "step": 5750 }, { "epoch": 0.9208, "grad_norm": 1.0842921733856201, "learning_rate": 7.698718180042392e-07, "loss": 0.702, "num_input_tokens_seen": 30813632, "step": 5755 }, { "epoch": 0.9216, "grad_norm": 1.0399980545043945, "learning_rate": 7.544757148827297e-07, "loss": 0.7203, "num_input_tokens_seen": 30840816, "step": 5760 }, { "epoch": 0.9224, "grad_norm": 1.1896955966949463, "learning_rate": 7.392327596698473e-07, "loss": 0.6873, "num_input_tokens_seen": 30861664, "step": 5765 }, { "epoch": 0.9232, "grad_norm": 1.3494455814361572, "learning_rate": 7.241430486483819e-07, "loss": 0.6975, "num_input_tokens_seen": 30886064, "step": 5770 }, { "epoch": 0.924, "grad_norm": 1.1318798065185547, "learning_rate": 7.092066771331507e-07, "loss": 0.6058, "num_input_tokens_seen": 30910608, "step": 5775 }, { "epoch": 0.9248, "grad_norm": 1.230055332183838, "learning_rate": 6.944237394703984e-07, "loss": 0.8128, "num_input_tokens_seen": 30935008, "step": 5780 }, { "epoch": 0.9256, "grad_norm": 1.0150400400161743, "learning_rate": 6.797943290371839e-07, "loss": 0.7329, "num_input_tokens_seen": 30959792, "step": 5785 }, { "epoch": 0.9264, "grad_norm": 0.9498345255851746, "learning_rate": 6.653185382408194e-07, "loss": 0.673, "num_input_tokens_seen": 30985856, "step": 5790 }, { "epoch": 0.9272, "grad_norm": 1.047587275505066, "learning_rate": 6.509964585182687e-07, "loss": 0.7395, "num_input_tokens_seen": 31013888, "step": 5795 }, { "epoch": 0.928, "grad_norm": 0.9112536907196045, "learning_rate": 6.368281803355691e-07, "loss": 0.753, "num_input_tokens_seen": 31038352, "step": 5800 }, { "epoch": 0.9288, "grad_norm": 0.9714504480361938, "learning_rate": 6.228137931872713e-07, "loss": 0.7573, "num_input_tokens_seen": 31066624, "step": 5805 }, { "epoch": 0.9296, "grad_norm": 1.1300855875015259, "learning_rate": 6.089533855958507e-07, "loss": 0.759, "num_input_tokens_seen": 31093184, "step": 5810 }, { "epoch": 0.9304, "grad_norm": 1.0004905462265015, "learning_rate": 5.95247045111183e-07, "loss": 0.7482, "num_input_tokens_seen": 31118352, "step": 5815 }, { "epoch": 0.9312, "grad_norm": 0.8432052731513977, "learning_rate": 5.816948583099613e-07, "loss": 0.6295, "num_input_tokens_seen": 31145616, "step": 5820 }, { "epoch": 0.932, "grad_norm": 0.923195481300354, "learning_rate": 5.68296910795163e-07, "loss": 0.7596, "num_input_tokens_seen": 31167088, "step": 5825 }, { "epoch": 0.9328, "grad_norm": 1.0346819162368774, "learning_rate": 5.550532871955061e-07, "loss": 0.689, "num_input_tokens_seen": 31192672, "step": 5830 }, { "epoch": 0.9336, "grad_norm": 0.8770239949226379, "learning_rate": 5.419640711649188e-07, "loss": 0.6387, "num_input_tokens_seen": 31224016, "step": 5835 }, { "epoch": 0.9344, "grad_norm": 1.2829992771148682, "learning_rate": 5.290293453819955e-07, "loss": 0.7316, "num_input_tokens_seen": 31247008, "step": 5840 }, { "epoch": 0.9352, "grad_norm": 0.9933931231498718, "learning_rate": 5.162491915495005e-07, "loss": 0.7255, "num_input_tokens_seen": 31273232, "step": 5845 }, { "epoch": 0.936, "grad_norm": 0.8756529092788696, "learning_rate": 5.036236903938285e-07, "loss": 0.7188, "num_input_tokens_seen": 31299504, "step": 5850 }, { "epoch": 0.9368, "grad_norm": 0.85035240650177, "learning_rate": 4.911529216645088e-07, "loss": 0.6763, "num_input_tokens_seen": 31325792, "step": 5855 }, { "epoch": 0.9376, "grad_norm": 0.9068401455879211, "learning_rate": 4.788369641336943e-07, "loss": 0.6109, "num_input_tokens_seen": 31351216, "step": 5860 }, { "epoch": 0.9384, "grad_norm": 1.079689860343933, "learning_rate": 4.666758955956613e-07, "loss": 0.7778, "num_input_tokens_seen": 31376464, "step": 5865 }, { "epoch": 0.9392, "grad_norm": 0.964074969291687, "learning_rate": 4.546697928663357e-07, "loss": 0.6315, "num_input_tokens_seen": 31408832, "step": 5870 }, { "epoch": 0.94, "grad_norm": 1.1026054620742798, "learning_rate": 4.4281873178278475e-07, "loss": 0.7918, "num_input_tokens_seen": 31432352, "step": 5875 }, { "epoch": 0.9408, "grad_norm": 0.914069652557373, "learning_rate": 4.311227872027479e-07, "loss": 0.6983, "num_input_tokens_seen": 31457392, "step": 5880 }, { "epoch": 0.9416, "grad_norm": 0.989115834236145, "learning_rate": 4.1958203300417054e-07, "loss": 0.7233, "num_input_tokens_seen": 31482704, "step": 5885 }, { "epoch": 0.9424, "grad_norm": 1.034597396850586, "learning_rate": 4.0819654208472947e-07, "loss": 0.6402, "num_input_tokens_seen": 31512368, "step": 5890 }, { "epoch": 0.9432, "grad_norm": 1.3309321403503418, "learning_rate": 3.9696638636137206e-07, "loss": 0.6942, "num_input_tokens_seen": 31539040, "step": 5895 }, { "epoch": 0.944, "grad_norm": 1.2237857580184937, "learning_rate": 3.8589163676986674e-07, "loss": 0.7119, "num_input_tokens_seen": 31563712, "step": 5900 }, { "epoch": 0.9448, "grad_norm": 1.0268304347991943, "learning_rate": 3.7497236326434757e-07, "loss": 0.6575, "num_input_tokens_seen": 31587760, "step": 5905 }, { "epoch": 0.9456, "grad_norm": 1.0060738325119019, "learning_rate": 3.6420863481688437e-07, "loss": 0.704, "num_input_tokens_seen": 31612976, "step": 5910 }, { "epoch": 0.9464, "grad_norm": 0.9219969511032104, "learning_rate": 3.536005194170328e-07, "loss": 0.7876, "num_input_tokens_seen": 31639472, "step": 5915 }, { "epoch": 0.9472, "grad_norm": 0.8863883018493652, "learning_rate": 3.431480840714152e-07, "loss": 0.7033, "num_input_tokens_seen": 31670768, "step": 5920 }, { "epoch": 0.948, "grad_norm": 0.9394556879997253, "learning_rate": 3.328513948032991e-07, "loss": 0.7095, "num_input_tokens_seen": 31696624, "step": 5925 }, { "epoch": 0.9488, "grad_norm": 0.8008967638015747, "learning_rate": 3.227105166521638e-07, "loss": 0.6629, "num_input_tokens_seen": 31723840, "step": 5930 }, { "epoch": 0.9496, "grad_norm": 0.9910029172897339, "learning_rate": 3.127255136733093e-07, "loss": 0.591, "num_input_tokens_seen": 31752736, "step": 5935 }, { "epoch": 0.9504, "grad_norm": 0.9355325698852539, "learning_rate": 3.0289644893744527e-07, "loss": 0.6641, "num_input_tokens_seen": 31777760, "step": 5940 }, { "epoch": 0.9512, "grad_norm": 0.9911002516746521, "learning_rate": 2.9322338453028066e-07, "loss": 0.6156, "num_input_tokens_seen": 31805264, "step": 5945 }, { "epoch": 0.952, "grad_norm": 1.4127229452133179, "learning_rate": 2.8370638155215123e-07, "loss": 0.7834, "num_input_tokens_seen": 31828656, "step": 5950 }, { "epoch": 0.9528, "grad_norm": 1.0222047567367554, "learning_rate": 2.743455001176176e-07, "loss": 0.6998, "num_input_tokens_seen": 31855424, "step": 5955 }, { "epoch": 0.9536, "grad_norm": 0.9893736839294434, "learning_rate": 2.6514079935509584e-07, "loss": 0.7458, "num_input_tokens_seen": 31879168, "step": 5960 }, { "epoch": 0.9544, "grad_norm": 0.8317204713821411, "learning_rate": 2.560923374064772e-07, "loss": 0.7061, "num_input_tokens_seen": 31903824, "step": 5965 }, { "epoch": 0.9552, "grad_norm": 1.3220785856246948, "learning_rate": 2.472001714267674e-07, "loss": 0.8603, "num_input_tokens_seen": 31927184, "step": 5970 }, { "epoch": 0.956, "grad_norm": 0.8110103607177734, "learning_rate": 2.384643575837203e-07, "loss": 0.6273, "num_input_tokens_seen": 31955104, "step": 5975 }, { "epoch": 0.9568, "grad_norm": 0.6332679390907288, "learning_rate": 2.298849510574824e-07, "loss": 0.714, "num_input_tokens_seen": 31985888, "step": 5980 }, { "epoch": 0.9576, "grad_norm": 0.9290034174919128, "learning_rate": 2.2146200604024613e-07, "loss": 0.6899, "num_input_tokens_seen": 32013520, "step": 5985 }, { "epoch": 0.9584, "grad_norm": 1.0509424209594727, "learning_rate": 2.1319557573591108e-07, "loss": 0.677, "num_input_tokens_seen": 32038880, "step": 5990 }, { "epoch": 0.9592, "grad_norm": 1.0169018507003784, "learning_rate": 2.050857123597455e-07, "loss": 0.7033, "num_input_tokens_seen": 32062160, "step": 5995 }, { "epoch": 0.96, "grad_norm": 1.053408145904541, "learning_rate": 1.9713246713805588e-07, "loss": 0.6431, "num_input_tokens_seen": 32085712, "step": 6000 }, { "epoch": 0.9608, "grad_norm": 1.1077343225479126, "learning_rate": 1.8933589030785682e-07, "loss": 0.683, "num_input_tokens_seen": 32115232, "step": 6005 }, { "epoch": 0.9616, "grad_norm": 1.2601428031921387, "learning_rate": 1.8169603111656552e-07, "loss": 0.751, "num_input_tokens_seen": 32142992, "step": 6010 }, { "epoch": 0.9624, "grad_norm": 0.7372344136238098, "learning_rate": 1.7421293782168835e-07, "loss": 0.5808, "num_input_tokens_seen": 32176176, "step": 6015 }, { "epoch": 0.9632, "grad_norm": 0.856760561466217, "learning_rate": 1.6688665769050703e-07, "loss": 0.6852, "num_input_tokens_seen": 32204992, "step": 6020 }, { "epoch": 0.964, "grad_norm": 1.110574722290039, "learning_rate": 1.5971723699979013e-07, "loss": 0.6778, "num_input_tokens_seen": 32232960, "step": 6025 }, { "epoch": 0.9648, "grad_norm": 1.084190845489502, "learning_rate": 1.5270472103549315e-07, "loss": 0.7036, "num_input_tokens_seen": 32262672, "step": 6030 }, { "epoch": 0.9656, "grad_norm": 0.9454313516616821, "learning_rate": 1.4584915409248112e-07, "loss": 0.655, "num_input_tokens_seen": 32285504, "step": 6035 }, { "epoch": 0.9664, "grad_norm": 0.9206419587135315, "learning_rate": 1.3915057947423705e-07, "loss": 0.7324, "num_input_tokens_seen": 32312288, "step": 6040 }, { "epoch": 0.9672, "grad_norm": 0.9567137956619263, "learning_rate": 1.3260903949260107e-07, "loss": 0.7166, "num_input_tokens_seen": 32339424, "step": 6045 }, { "epoch": 0.968, "grad_norm": 1.0180697441101074, "learning_rate": 1.2622457546749567e-07, "loss": 0.7, "num_input_tokens_seen": 32362848, "step": 6050 }, { "epoch": 0.9688, "grad_norm": 1.2073848247528076, "learning_rate": 1.1999722772666476e-07, "loss": 0.7519, "num_input_tokens_seen": 32393264, "step": 6055 }, { "epoch": 0.9696, "grad_norm": 1.020180583000183, "learning_rate": 1.1392703560542117e-07, "loss": 0.7524, "num_input_tokens_seen": 32418464, "step": 6060 }, { "epoch": 0.9704, "grad_norm": 1.0939137935638428, "learning_rate": 1.080140374463967e-07, "loss": 0.5829, "num_input_tokens_seen": 32449248, "step": 6065 }, { "epoch": 0.9712, "grad_norm": 1.2307384014129639, "learning_rate": 1.0225827059930083e-07, "loss": 0.7017, "num_input_tokens_seen": 32477312, "step": 6070 }, { "epoch": 0.972, "grad_norm": 1.0141756534576416, "learning_rate": 9.665977142068738e-08, "loss": 0.6852, "num_input_tokens_seen": 32505024, "step": 6075 }, { "epoch": 0.9728, "grad_norm": 1.0366077423095703, "learning_rate": 9.121857527372158e-08, "loss": 0.733, "num_input_tokens_seen": 32530080, "step": 6080 }, { "epoch": 0.9736, "grad_norm": 0.9831274151802063, "learning_rate": 8.593471652794949e-08, "loss": 0.6721, "num_input_tokens_seen": 32557488, "step": 6085 }, { "epoch": 0.9744, "grad_norm": 1.0170478820800781, "learning_rate": 8.080822855909831e-08, "loss": 0.6572, "num_input_tokens_seen": 32589072, "step": 6090 }, { "epoch": 0.9752, "grad_norm": 1.0840100049972534, "learning_rate": 7.583914374885426e-08, "loss": 0.7535, "num_input_tokens_seen": 32613296, "step": 6095 }, { "epoch": 0.976, "grad_norm": 1.1899126768112183, "learning_rate": 7.102749348465165e-08, "loss": 0.639, "num_input_tokens_seen": 32642512, "step": 6100 }, { "epoch": 0.9768, "grad_norm": 1.0756986141204834, "learning_rate": 6.637330815949527e-08, "loss": 0.7558, "num_input_tokens_seen": 32666064, "step": 6105 }, { "epoch": 0.9776, "grad_norm": 0.9403240084648132, "learning_rate": 6.187661717174386e-08, "loss": 0.7228, "num_input_tokens_seen": 32690016, "step": 6110 }, { "epoch": 0.9784, "grad_norm": 0.9194949269294739, "learning_rate": 5.753744892494639e-08, "loss": 0.7079, "num_input_tokens_seen": 32716240, "step": 6115 }, { "epoch": 0.9792, "grad_norm": 0.9947624206542969, "learning_rate": 5.335583082764495e-08, "loss": 0.7692, "num_input_tokens_seen": 32741648, "step": 6120 }, { "epoch": 0.98, "grad_norm": 1.2828369140625, "learning_rate": 4.9331789293211026e-08, "loss": 0.6285, "num_input_tokens_seen": 32770224, "step": 6125 }, { "epoch": 0.9808, "grad_norm": 1.0066205263137817, "learning_rate": 4.546534973968175e-08, "loss": 0.7464, "num_input_tokens_seen": 32798864, "step": 6130 }, { "epoch": 0.9816, "grad_norm": 1.3146965503692627, "learning_rate": 4.1756536589585004e-08, "loss": 0.6632, "num_input_tokens_seen": 32829136, "step": 6135 }, { "epoch": 0.9824, "grad_norm": 0.8514100909233093, "learning_rate": 3.820537326980622e-08, "loss": 0.7378, "num_input_tokens_seen": 32858976, "step": 6140 }, { "epoch": 0.9832, "grad_norm": 0.9065835475921631, "learning_rate": 3.481188221142184e-08, "loss": 0.7125, "num_input_tokens_seen": 32886208, "step": 6145 }, { "epoch": 0.984, "grad_norm": 1.2251099348068237, "learning_rate": 3.157608484956332e-08, "loss": 0.7723, "num_input_tokens_seen": 32912960, "step": 6150 }, { "epoch": 0.9848, "grad_norm": 1.0463021993637085, "learning_rate": 2.8498001623286642e-08, "loss": 0.6472, "num_input_tokens_seen": 32941072, "step": 6155 }, { "epoch": 0.9856, "grad_norm": 1.001555323600769, "learning_rate": 2.557765197543638e-08, "loss": 0.7462, "num_input_tokens_seen": 32968304, "step": 6160 }, { "epoch": 0.9864, "grad_norm": 1.0437195301055908, "learning_rate": 2.281505435253184e-08, "loss": 0.7079, "num_input_tokens_seen": 32993008, "step": 6165 }, { "epoch": 0.9872, "grad_norm": 1.1048009395599365, "learning_rate": 2.0210226204639414e-08, "loss": 0.7058, "num_input_tokens_seen": 33018128, "step": 6170 }, { "epoch": 0.988, "grad_norm": 1.1593177318572998, "learning_rate": 1.7763183985269883e-08, "loss": 0.5942, "num_input_tokens_seen": 33047968, "step": 6175 }, { "epoch": 0.9888, "grad_norm": 1.3036433458328247, "learning_rate": 1.5473943151270153e-08, "loss": 0.7738, "num_input_tokens_seen": 33072560, "step": 6180 }, { "epoch": 0.9896, "grad_norm": 1.0337815284729004, "learning_rate": 1.3342518162728912e-08, "loss": 0.781, "num_input_tokens_seen": 33096928, "step": 6185 }, { "epoch": 0.9904, "grad_norm": 1.0598255395889282, "learning_rate": 1.136892248288779e-08, "loss": 0.6607, "num_input_tokens_seen": 33123808, "step": 6190 }, { "epoch": 0.9912, "grad_norm": 1.658722162246704, "learning_rate": 9.553168578049775e-09, "loss": 0.7506, "num_input_tokens_seen": 33148688, "step": 6195 }, { "epoch": 0.992, "grad_norm": 0.9460881352424622, "learning_rate": 7.895267917501504e-09, "loss": 0.6521, "num_input_tokens_seen": 33176624, "step": 6200 }, { "epoch": 0.9928, "grad_norm": 0.9642547369003296, "learning_rate": 6.395230973443856e-09, "loss": 0.709, "num_input_tokens_seen": 33198160, "step": 6205 }, { "epoch": 0.9936, "grad_norm": 1.1588774919509888, "learning_rate": 5.053067220925356e-09, "loss": 0.6685, "num_input_tokens_seen": 33226336, "step": 6210 }, { "epoch": 0.9944, "grad_norm": 0.7818155288696289, "learning_rate": 3.868785137786657e-09, "loss": 0.6672, "num_input_tokens_seen": 33245824, "step": 6215 }, { "epoch": 0.9952, "grad_norm": 0.8517420887947083, "learning_rate": 2.842392204591149e-09, "loss": 0.7053, "num_input_tokens_seen": 33274176, "step": 6220 }, { "epoch": 0.996, "grad_norm": 1.1486226320266724, "learning_rate": 1.973894904597207e-09, "loss": 0.7184, "num_input_tokens_seen": 33302528, "step": 6225 }, { "epoch": 0.9968, "grad_norm": 0.9121309518814087, "learning_rate": 1.2632987237054528e-09, "loss": 0.7092, "num_input_tokens_seen": 33330384, "step": 6230 }, { "epoch": 0.9976, "grad_norm": 1.1165632009506226, "learning_rate": 7.106081504254514e-10, "loss": 0.6142, "num_input_tokens_seen": 33357968, "step": 6235 }, { "epoch": 0.9984, "grad_norm": 1.0336644649505615, "learning_rate": 3.158266758562789e-10, "loss": 0.7147, "num_input_tokens_seen": 33381536, "step": 6240 }, { "epoch": 0.9992, "grad_norm": 1.1494331359863281, "learning_rate": 7.89567936476665e-11, "loss": 0.7005, "num_input_tokens_seen": 33409984, "step": 6245 }, { "epoch": 1.0, "grad_norm": 1.0414918661117554, "learning_rate": 0.0, "loss": 0.7516, "num_input_tokens_seen": 33437856, "step": 6250 }, { "epoch": 1.0, "num_input_tokens_seen": 33437856, "step": 6250, "total_flos": 7.1914395644928e+16, "train_loss": 0.7151971128082275, "train_runtime": 36754.4929, "train_samples_per_second": 2.721, "train_steps_per_second": 0.17 } ], "logging_steps": 5, "max_steps": 6250, "num_input_tokens_seen": 33437856, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 7.1914395644928e+16, "train_batch_size": 2, "trial_name": null, "trial_params": null }