{ "best_metric": null, "best_model_checkpoint": null, "epoch": 4.98793242156074, "eval_steps": 500, "global_step": 775, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.006436041834271922, "grad_norm": 857.0773315429688, "learning_rate": 5.000000000000001e-07, "loss": 12.2736, "num_input_tokens_seen": 6576, "step": 1 }, { "epoch": 0.012872083668543845, "grad_norm": 899.0701293945312, "learning_rate": 1.0000000000000002e-06, "loss": 12.5518, "num_input_tokens_seen": 13312, "step": 2 }, { "epoch": 0.019308125502815767, "grad_norm": 833.7578125, "learning_rate": 1.5e-06, "loss": 11.8551, "num_input_tokens_seen": 19952, "step": 3 }, { "epoch": 0.02574416733708769, "grad_norm": 712.9902954101562, "learning_rate": 2.0000000000000003e-06, "loss": 10.9045, "num_input_tokens_seen": 26640, "step": 4 }, { "epoch": 0.032180209171359615, "grad_norm": 620.4878540039062, "learning_rate": 2.5e-06, "loss": 8.9845, "num_input_tokens_seen": 33360, "step": 5 }, { "epoch": 0.038616251005631534, "grad_norm": 376.7406921386719, "learning_rate": 3e-06, "loss": 6.35, "num_input_tokens_seen": 39984, "step": 6 }, { "epoch": 0.04505229283990346, "grad_norm": 356.7503967285156, "learning_rate": 3.5e-06, "loss": 5.4864, "num_input_tokens_seen": 46496, "step": 7 }, { "epoch": 0.05148833467417538, "grad_norm": 383.4678955078125, "learning_rate": 4.000000000000001e-06, "loss": 2.8013, "num_input_tokens_seen": 53008, "step": 8 }, { "epoch": 0.057924376508447305, "grad_norm": 285.64483642578125, "learning_rate": 4.5e-06, "loss": 0.9851, "num_input_tokens_seen": 59856, "step": 9 }, { "epoch": 0.06436041834271923, "grad_norm": 222.22850036621094, "learning_rate": 5e-06, "loss": 0.5088, "num_input_tokens_seen": 66352, "step": 10 }, { "epoch": 0.07079646017699115, "grad_norm": 338.3565368652344, "learning_rate": 4.99997891923933e-06, "loss": 2.022, "num_input_tokens_seen": 73184, "step": 11 }, { "epoch": 0.07723250201126307, "grad_norm": 97.87013244628906, "learning_rate": 4.999915677312839e-06, "loss": 0.5842, "num_input_tokens_seen": 80064, "step": 12 }, { "epoch": 0.083668543845535, "grad_norm": 137.05706787109375, "learning_rate": 4.999810275287077e-06, "loss": 0.8538, "num_input_tokens_seen": 86688, "step": 13 }, { "epoch": 0.09010458567980692, "grad_norm": 88.00048828125, "learning_rate": 4.9996627149396075e-06, "loss": 0.5843, "num_input_tokens_seen": 93168, "step": 14 }, { "epoch": 0.09654062751407884, "grad_norm": 61.654090881347656, "learning_rate": 4.999472998758979e-06, "loss": 0.4456, "num_input_tokens_seen": 99696, "step": 15 }, { "epoch": 0.10297666934835076, "grad_norm": 37.3619499206543, "learning_rate": 4.99924112994468e-06, "loss": 0.3542, "num_input_tokens_seen": 106016, "step": 16 }, { "epoch": 0.10941271118262269, "grad_norm": 41.219093322753906, "learning_rate": 4.998967112407087e-06, "loss": 0.3416, "num_input_tokens_seen": 112672, "step": 17 }, { "epoch": 0.11584875301689461, "grad_norm": 22.267297744750977, "learning_rate": 4.9986509507673986e-06, "loss": 0.2803, "num_input_tokens_seen": 119312, "step": 18 }, { "epoch": 0.12228479485116653, "grad_norm": 22.40268898010254, "learning_rate": 4.998292650357558e-06, "loss": 0.2603, "num_input_tokens_seen": 126016, "step": 19 }, { "epoch": 0.12872083668543846, "grad_norm": 23.54829978942871, "learning_rate": 4.99789221722016e-06, "loss": 0.2456, "num_input_tokens_seen": 132704, "step": 20 }, { "epoch": 0.13515687851971037, "grad_norm": 24.01500701904297, "learning_rate": 4.997449658108354e-06, "loss": 0.2269, "num_input_tokens_seen": 139472, "step": 21 }, { "epoch": 0.1415929203539823, "grad_norm": 26.877809524536133, "learning_rate": 4.996964980485725e-06, "loss": 0.2896, "num_input_tokens_seen": 145968, "step": 22 }, { "epoch": 0.14802896218825423, "grad_norm": 11.89682388305664, "learning_rate": 4.996438192526173e-06, "loss": 0.1415, "num_input_tokens_seen": 152144, "step": 23 }, { "epoch": 0.15446500402252614, "grad_norm": 12.671065330505371, "learning_rate": 4.995869303113768e-06, "loss": 0.2228, "num_input_tokens_seen": 158432, "step": 24 }, { "epoch": 0.16090104585679807, "grad_norm": 12.23610782623291, "learning_rate": 4.995258321842611e-06, "loss": 0.1537, "num_input_tokens_seen": 164672, "step": 25 }, { "epoch": 0.16733708769107, "grad_norm": 21.917552947998047, "learning_rate": 4.994605259016658e-06, "loss": 0.2146, "num_input_tokens_seen": 170896, "step": 26 }, { "epoch": 0.1737731295253419, "grad_norm": 6.669096946716309, "learning_rate": 4.993910125649561e-06, "loss": 0.1688, "num_input_tokens_seen": 177904, "step": 27 }, { "epoch": 0.18020917135961384, "grad_norm": 10.865259170532227, "learning_rate": 4.99317293346447e-06, "loss": 0.1592, "num_input_tokens_seen": 184640, "step": 28 }, { "epoch": 0.18664521319388577, "grad_norm": 22.876623153686523, "learning_rate": 4.992393694893844e-06, "loss": 0.2333, "num_input_tokens_seen": 191008, "step": 29 }, { "epoch": 0.19308125502815768, "grad_norm": 12.427145004272461, "learning_rate": 4.991572423079236e-06, "loss": 0.1812, "num_input_tokens_seen": 197568, "step": 30 }, { "epoch": 0.1995172968624296, "grad_norm": 15.451845169067383, "learning_rate": 4.990709131871074e-06, "loss": 0.1925, "num_input_tokens_seen": 204272, "step": 31 }, { "epoch": 0.20595333869670152, "grad_norm": 18.907636642456055, "learning_rate": 4.989803835828426e-06, "loss": 0.1864, "num_input_tokens_seen": 210944, "step": 32 }, { "epoch": 0.21238938053097345, "grad_norm": 15.60383129119873, "learning_rate": 4.988856550218755e-06, "loss": 0.1848, "num_input_tokens_seen": 217584, "step": 33 }, { "epoch": 0.21882542236524538, "grad_norm": 7.129302024841309, "learning_rate": 4.987867291017662e-06, "loss": 0.1466, "num_input_tokens_seen": 224064, "step": 34 }, { "epoch": 0.2252614641995173, "grad_norm": 15.756115913391113, "learning_rate": 4.986836074908616e-06, "loss": 0.1481, "num_input_tokens_seen": 230880, "step": 35 }, { "epoch": 0.23169750603378922, "grad_norm": 8.845354080200195, "learning_rate": 4.985762919282674e-06, "loss": 0.149, "num_input_tokens_seen": 237312, "step": 36 }, { "epoch": 0.23813354786806115, "grad_norm": 15.7093505859375, "learning_rate": 4.984647842238185e-06, "loss": 0.1506, "num_input_tokens_seen": 243648, "step": 37 }, { "epoch": 0.24456958970233306, "grad_norm": 11.331380844116211, "learning_rate": 4.983490862580486e-06, "loss": 0.1709, "num_input_tokens_seen": 250096, "step": 38 }, { "epoch": 0.251005631536605, "grad_norm": 6.254825115203857, "learning_rate": 4.982291999821587e-06, "loss": 0.0898, "num_input_tokens_seen": 256432, "step": 39 }, { "epoch": 0.2574416733708769, "grad_norm": 7.792216777801514, "learning_rate": 4.98105127417984e-06, "loss": 0.1856, "num_input_tokens_seen": 263088, "step": 40 }, { "epoch": 0.26387771520514886, "grad_norm": 4.186593055725098, "learning_rate": 4.979768706579595e-06, "loss": 0.0947, "num_input_tokens_seen": 269904, "step": 41 }, { "epoch": 0.27031375703942073, "grad_norm": 7.599153518676758, "learning_rate": 4.978444318650855e-06, "loss": 0.1367, "num_input_tokens_seen": 276592, "step": 42 }, { "epoch": 0.27674979887369267, "grad_norm": 10.765763282775879, "learning_rate": 4.977078132728901e-06, "loss": 0.1645, "num_input_tokens_seen": 283424, "step": 43 }, { "epoch": 0.2831858407079646, "grad_norm": 6.705766677856445, "learning_rate": 4.975670171853926e-06, "loss": 0.1179, "num_input_tokens_seen": 290176, "step": 44 }, { "epoch": 0.28962188254223653, "grad_norm": 6.055794715881348, "learning_rate": 4.9742204597706386e-06, "loss": 0.133, "num_input_tokens_seen": 296752, "step": 45 }, { "epoch": 0.29605792437650846, "grad_norm": 7.4584760665893555, "learning_rate": 4.972729020927866e-06, "loss": 0.083, "num_input_tokens_seen": 303392, "step": 46 }, { "epoch": 0.3024939662107804, "grad_norm": 10.979104995727539, "learning_rate": 4.9711958804781385e-06, "loss": 0.1748, "num_input_tokens_seen": 310304, "step": 47 }, { "epoch": 0.3089300080450523, "grad_norm": 13.912871360778809, "learning_rate": 4.969621064277271e-06, "loss": 0.1854, "num_input_tokens_seen": 317440, "step": 48 }, { "epoch": 0.3153660498793242, "grad_norm": 6.554210186004639, "learning_rate": 4.968004598883923e-06, "loss": 0.1232, "num_input_tokens_seen": 324304, "step": 49 }, { "epoch": 0.32180209171359614, "grad_norm": 6.358190536499023, "learning_rate": 4.966346511559149e-06, "loss": 0.1172, "num_input_tokens_seen": 330832, "step": 50 }, { "epoch": 0.32823813354786807, "grad_norm": 6.195626258850098, "learning_rate": 4.964646830265944e-06, "loss": 0.1404, "num_input_tokens_seen": 337952, "step": 51 }, { "epoch": 0.33467417538214, "grad_norm": 12.585171699523926, "learning_rate": 4.962905583668766e-06, "loss": 0.137, "num_input_tokens_seen": 344384, "step": 52 }, { "epoch": 0.3411102172164119, "grad_norm": 3.7672178745269775, "learning_rate": 4.961122801133059e-06, "loss": 0.1191, "num_input_tokens_seen": 351184, "step": 53 }, { "epoch": 0.3475462590506838, "grad_norm": 17.48076629638672, "learning_rate": 4.9592985127247525e-06, "loss": 0.1624, "num_input_tokens_seen": 357696, "step": 54 }, { "epoch": 0.35398230088495575, "grad_norm": 7.758498668670654, "learning_rate": 4.957432749209755e-06, "loss": 0.1256, "num_input_tokens_seen": 364368, "step": 55 }, { "epoch": 0.3604183427192277, "grad_norm": 10.048332214355469, "learning_rate": 4.955525542053438e-06, "loss": 0.1274, "num_input_tokens_seen": 370896, "step": 56 }, { "epoch": 0.3668543845534996, "grad_norm": 17.495296478271484, "learning_rate": 4.953576923420105e-06, "loss": 0.174, "num_input_tokens_seen": 377168, "step": 57 }, { "epoch": 0.37329042638777155, "grad_norm": 7.546329021453857, "learning_rate": 4.9515869261724444e-06, "loss": 0.0805, "num_input_tokens_seen": 383728, "step": 58 }, { "epoch": 0.3797264682220434, "grad_norm": 5.6687188148498535, "learning_rate": 4.949555583870983e-06, "loss": 0.1181, "num_input_tokens_seen": 390448, "step": 59 }, { "epoch": 0.38616251005631536, "grad_norm": 9.777739524841309, "learning_rate": 4.9474829307735115e-06, "loss": 0.1613, "num_input_tokens_seen": 396960, "step": 60 }, { "epoch": 0.3925985518905873, "grad_norm": 10.064454078674316, "learning_rate": 4.9453690018345144e-06, "loss": 0.0885, "num_input_tokens_seen": 403680, "step": 61 }, { "epoch": 0.3990345937248592, "grad_norm": 6.493910312652588, "learning_rate": 4.943213832704575e-06, "loss": 0.1526, "num_input_tokens_seen": 410096, "step": 62 }, { "epoch": 0.40547063555913115, "grad_norm": 7.949091911315918, "learning_rate": 4.941017459729778e-06, "loss": 0.1114, "num_input_tokens_seen": 416672, "step": 63 }, { "epoch": 0.41190667739340303, "grad_norm": 8.829463958740234, "learning_rate": 4.938779919951092e-06, "loss": 0.1139, "num_input_tokens_seen": 423136, "step": 64 }, { "epoch": 0.41834271922767496, "grad_norm": 3.495246410369873, "learning_rate": 4.936501251103751e-06, "loss": 0.0878, "num_input_tokens_seen": 429888, "step": 65 }, { "epoch": 0.4247787610619469, "grad_norm": 8.937992095947266, "learning_rate": 4.934181491616613e-06, "loss": 0.1047, "num_input_tokens_seen": 436720, "step": 66 }, { "epoch": 0.43121480289621883, "grad_norm": 12.225470542907715, "learning_rate": 4.9318206806115125e-06, "loss": 0.1323, "num_input_tokens_seen": 443648, "step": 67 }, { "epoch": 0.43765084473049076, "grad_norm": 8.192527770996094, "learning_rate": 4.929418857902603e-06, "loss": 0.095, "num_input_tokens_seen": 450464, "step": 68 }, { "epoch": 0.4440868865647627, "grad_norm": 9.30573844909668, "learning_rate": 4.926976063995687e-06, "loss": 0.2024, "num_input_tokens_seen": 457296, "step": 69 }, { "epoch": 0.4505229283990346, "grad_norm": 16.341676712036133, "learning_rate": 4.9244923400875245e-06, "loss": 0.1614, "num_input_tokens_seen": 463984, "step": 70 }, { "epoch": 0.4569589702333065, "grad_norm": 19.184734344482422, "learning_rate": 4.921967728065147e-06, "loss": 0.2073, "num_input_tokens_seen": 470432, "step": 71 }, { "epoch": 0.46339501206757844, "grad_norm": 9.802066802978516, "learning_rate": 4.91940227050515e-06, "loss": 0.108, "num_input_tokens_seen": 476736, "step": 72 }, { "epoch": 0.46983105390185037, "grad_norm": 8.744816780090332, "learning_rate": 4.916796010672969e-06, "loss": 0.1455, "num_input_tokens_seen": 482896, "step": 73 }, { "epoch": 0.4762670957361223, "grad_norm": 11.598526954650879, "learning_rate": 4.914148992522157e-06, "loss": 0.131, "num_input_tokens_seen": 489504, "step": 74 }, { "epoch": 0.4827031375703942, "grad_norm": 11.42316722869873, "learning_rate": 4.911461260693639e-06, "loss": 0.1416, "num_input_tokens_seen": 496160, "step": 75 }, { "epoch": 0.4891391794046661, "grad_norm": 3.1568145751953125, "learning_rate": 4.908732860514958e-06, "loss": 0.1045, "num_input_tokens_seen": 502528, "step": 76 }, { "epoch": 0.49557522123893805, "grad_norm": 9.457361221313477, "learning_rate": 4.905963837999518e-06, "loss": 0.1466, "num_input_tokens_seen": 509104, "step": 77 }, { "epoch": 0.50201126307321, "grad_norm": 9.220935821533203, "learning_rate": 4.903154239845798e-06, "loss": 0.1502, "num_input_tokens_seen": 515760, "step": 78 }, { "epoch": 0.5084473049074819, "grad_norm": 5.706662654876709, "learning_rate": 4.900304113436571e-06, "loss": 0.1235, "num_input_tokens_seen": 522336, "step": 79 }, { "epoch": 0.5148833467417538, "grad_norm": 8.74252700805664, "learning_rate": 4.897413506838103e-06, "loss": 0.1022, "num_input_tokens_seen": 528960, "step": 80 }, { "epoch": 0.5213193885760258, "grad_norm": 4.498232841491699, "learning_rate": 4.894482468799344e-06, "loss": 0.0922, "num_input_tokens_seen": 535920, "step": 81 }, { "epoch": 0.5277554304102977, "grad_norm": 3.6902291774749756, "learning_rate": 4.891511048751102e-06, "loss": 0.0699, "num_input_tokens_seen": 542496, "step": 82 }, { "epoch": 0.5341914722445696, "grad_norm": 5.754522323608398, "learning_rate": 4.888499296805214e-06, "loss": 0.1057, "num_input_tokens_seen": 548752, "step": 83 }, { "epoch": 0.5406275140788415, "grad_norm": 4.513391017913818, "learning_rate": 4.8854472637536966e-06, "loss": 0.0793, "num_input_tokens_seen": 555696, "step": 84 }, { "epoch": 0.5470635559131134, "grad_norm": 4.931502342224121, "learning_rate": 4.882355001067892e-06, "loss": 0.14, "num_input_tokens_seen": 562192, "step": 85 }, { "epoch": 0.5534995977473853, "grad_norm": 6.896547794342041, "learning_rate": 4.8792225608976e-06, "loss": 0.1538, "num_input_tokens_seen": 568672, "step": 86 }, { "epoch": 0.5599356395816573, "grad_norm": 3.4364850521087646, "learning_rate": 4.8760499960702005e-06, "loss": 0.1135, "num_input_tokens_seen": 575440, "step": 87 }, { "epoch": 0.5663716814159292, "grad_norm": 6.179934501647949, "learning_rate": 4.8728373600897535e-06, "loss": 0.1253, "num_input_tokens_seen": 581808, "step": 88 }, { "epoch": 0.5728077232502011, "grad_norm": 14.744488716125488, "learning_rate": 4.869584707136109e-06, "loss": 0.1408, "num_input_tokens_seen": 588576, "step": 89 }, { "epoch": 0.5792437650844731, "grad_norm": 8.414978981018066, "learning_rate": 4.8662920920639866e-06, "loss": 0.0916, "num_input_tokens_seen": 595328, "step": 90 }, { "epoch": 0.585679806918745, "grad_norm": 3.016206979751587, "learning_rate": 4.86295957040205e-06, "loss": 0.1016, "num_input_tokens_seen": 601808, "step": 91 }, { "epoch": 0.5921158487530169, "grad_norm": 3.084475040435791, "learning_rate": 4.8595871983519705e-06, "loss": 0.0936, "num_input_tokens_seen": 608400, "step": 92 }, { "epoch": 0.5985518905872889, "grad_norm": 5.78838586807251, "learning_rate": 4.856175032787485e-06, "loss": 0.1557, "num_input_tokens_seen": 615296, "step": 93 }, { "epoch": 0.6049879324215608, "grad_norm": 4.525265216827393, "learning_rate": 4.852723131253429e-06, "loss": 0.0879, "num_input_tokens_seen": 621888, "step": 94 }, { "epoch": 0.6114239742558326, "grad_norm": 8.129231452941895, "learning_rate": 4.849231551964771e-06, "loss": 0.1399, "num_input_tokens_seen": 628768, "step": 95 }, { "epoch": 0.6178600160901045, "grad_norm": 7.902085781097412, "learning_rate": 4.845700353805629e-06, "loss": 0.1724, "num_input_tokens_seen": 635056, "step": 96 }, { "epoch": 0.6242960579243765, "grad_norm": 7.928036689758301, "learning_rate": 4.842129596328277e-06, "loss": 0.1018, "num_input_tokens_seen": 641872, "step": 97 }, { "epoch": 0.6307320997586484, "grad_norm": 5.5206756591796875, "learning_rate": 4.838519339752143e-06, "loss": 0.0398, "num_input_tokens_seen": 648752, "step": 98 }, { "epoch": 0.6371681415929203, "grad_norm": 5.910008907318115, "learning_rate": 4.834869644962789e-06, "loss": 0.1094, "num_input_tokens_seen": 655424, "step": 99 }, { "epoch": 0.6436041834271923, "grad_norm": 17.12012481689453, "learning_rate": 4.83118057351089e-06, "loss": 0.1915, "num_input_tokens_seen": 662224, "step": 100 }, { "epoch": 0.6500402252614642, "grad_norm": 13.876479148864746, "learning_rate": 4.827452187611192e-06, "loss": 0.1518, "num_input_tokens_seen": 668576, "step": 101 }, { "epoch": 0.6564762670957361, "grad_norm": 9.082406044006348, "learning_rate": 4.823684550141464e-06, "loss": 0.141, "num_input_tokens_seen": 675232, "step": 102 }, { "epoch": 0.6629123089300081, "grad_norm": 7.364652633666992, "learning_rate": 4.819877724641437e-06, "loss": 0.1564, "num_input_tokens_seen": 681856, "step": 103 }, { "epoch": 0.66934835076428, "grad_norm": 16.54301643371582, "learning_rate": 4.8160317753117326e-06, "loss": 0.1267, "num_input_tokens_seen": 688416, "step": 104 }, { "epoch": 0.6757843925985519, "grad_norm": 25.702648162841797, "learning_rate": 4.81214676701278e-06, "loss": 0.228, "num_input_tokens_seen": 695248, "step": 105 }, { "epoch": 0.6822204344328238, "grad_norm": 17.066158294677734, "learning_rate": 4.808222765263724e-06, "loss": 0.1532, "num_input_tokens_seen": 701952, "step": 106 }, { "epoch": 0.6886564762670957, "grad_norm": 11.833669662475586, "learning_rate": 4.8042598362413175e-06, "loss": 0.1482, "num_input_tokens_seen": 708368, "step": 107 }, { "epoch": 0.6950925181013676, "grad_norm": 2.9714369773864746, "learning_rate": 4.800258046778809e-06, "loss": 0.1074, "num_input_tokens_seen": 714768, "step": 108 }, { "epoch": 0.7015285599356396, "grad_norm": 9.384042739868164, "learning_rate": 4.796217464364808e-06, "loss": 0.1163, "num_input_tokens_seen": 721600, "step": 109 }, { "epoch": 0.7079646017699115, "grad_norm": 12.922999382019043, "learning_rate": 4.792138157142158e-06, "loss": 0.1871, "num_input_tokens_seen": 728448, "step": 110 }, { "epoch": 0.7144006436041834, "grad_norm": 5.947402000427246, "learning_rate": 4.788020193906776e-06, "loss": 0.092, "num_input_tokens_seen": 734720, "step": 111 }, { "epoch": 0.7208366854384554, "grad_norm": 6.692570686340332, "learning_rate": 4.783863644106502e-06, "loss": 0.0959, "num_input_tokens_seen": 741216, "step": 112 }, { "epoch": 0.7272727272727273, "grad_norm": 4.1358771324157715, "learning_rate": 4.779668577839921e-06, "loss": 0.0853, "num_input_tokens_seen": 747664, "step": 113 }, { "epoch": 0.7337087691069992, "grad_norm": 4.695752143859863, "learning_rate": 4.775435065855183e-06, "loss": 0.1581, "num_input_tokens_seen": 754480, "step": 114 }, { "epoch": 0.7401448109412712, "grad_norm": 3.8357784748077393, "learning_rate": 4.771163179548809e-06, "loss": 0.0907, "num_input_tokens_seen": 761344, "step": 115 }, { "epoch": 0.7465808527755431, "grad_norm": 4.455271244049072, "learning_rate": 4.766852990964492e-06, "loss": 0.0826, "num_input_tokens_seen": 768160, "step": 116 }, { "epoch": 0.7530168946098149, "grad_norm": 3.682065963745117, "learning_rate": 4.762504572791873e-06, "loss": 0.0979, "num_input_tokens_seen": 774496, "step": 117 }, { "epoch": 0.7594529364440868, "grad_norm": 6.100201606750488, "learning_rate": 4.7581179983653224e-06, "loss": 0.1617, "num_input_tokens_seen": 781232, "step": 118 }, { "epoch": 0.7658889782783588, "grad_norm": 7.6822991371154785, "learning_rate": 4.753693341662702e-06, "loss": 0.1306, "num_input_tokens_seen": 788064, "step": 119 }, { "epoch": 0.7723250201126307, "grad_norm": 4.73075008392334, "learning_rate": 4.749230677304114e-06, "loss": 0.0955, "num_input_tokens_seen": 794656, "step": 120 }, { "epoch": 0.7787610619469026, "grad_norm": 4.435886859893799, "learning_rate": 4.7447300805506455e-06, "loss": 0.0683, "num_input_tokens_seen": 801184, "step": 121 }, { "epoch": 0.7851971037811746, "grad_norm": 3.517606735229492, "learning_rate": 4.7401916273031e-06, "loss": 0.1117, "num_input_tokens_seen": 808000, "step": 122 }, { "epoch": 0.7916331456154465, "grad_norm": 3.445953130722046, "learning_rate": 4.7356153941007145e-06, "loss": 0.1115, "num_input_tokens_seen": 814608, "step": 123 }, { "epoch": 0.7980691874497184, "grad_norm": 2.4660255908966064, "learning_rate": 4.73100145811987e-06, "loss": 0.0945, "num_input_tokens_seen": 821072, "step": 124 }, { "epoch": 0.8045052292839904, "grad_norm": 6.673710346221924, "learning_rate": 4.726349897172791e-06, "loss": 0.1125, "num_input_tokens_seen": 827840, "step": 125 }, { "epoch": 0.8109412711182623, "grad_norm": 7.910948753356934, "learning_rate": 4.721660789706232e-06, "loss": 0.15, "num_input_tokens_seen": 834880, "step": 126 }, { "epoch": 0.8173773129525342, "grad_norm": 4.5563154220581055, "learning_rate": 4.716934214800155e-06, "loss": 0.1015, "num_input_tokens_seen": 841360, "step": 127 }, { "epoch": 0.8238133547868061, "grad_norm": 5.7200422286987305, "learning_rate": 4.712170252166395e-06, "loss": 0.1271, "num_input_tokens_seen": 847888, "step": 128 }, { "epoch": 0.830249396621078, "grad_norm": 4.4525465965271, "learning_rate": 4.707368982147318e-06, "loss": 0.0762, "num_input_tokens_seen": 854896, "step": 129 }, { "epoch": 0.8366854384553499, "grad_norm": 4.427840232849121, "learning_rate": 4.702530485714462e-06, "loss": 0.1196, "num_input_tokens_seen": 861600, "step": 130 }, { "epoch": 0.8431214802896219, "grad_norm": 3.674197196960449, "learning_rate": 4.697654844467175e-06, "loss": 0.0866, "num_input_tokens_seen": 868272, "step": 131 }, { "epoch": 0.8495575221238938, "grad_norm": 7.5055413246154785, "learning_rate": 4.69274214063124e-06, "loss": 0.0718, "num_input_tokens_seen": 875232, "step": 132 }, { "epoch": 0.8559935639581657, "grad_norm": 7.523169040679932, "learning_rate": 4.687792457057482e-06, "loss": 0.0808, "num_input_tokens_seen": 882112, "step": 133 }, { "epoch": 0.8624296057924377, "grad_norm": 10.57685375213623, "learning_rate": 4.682805877220378e-06, "loss": 0.1069, "num_input_tokens_seen": 888848, "step": 134 }, { "epoch": 0.8688656476267096, "grad_norm": 6.235794544219971, "learning_rate": 4.677782485216644e-06, "loss": 0.0804, "num_input_tokens_seen": 895136, "step": 135 }, { "epoch": 0.8753016894609815, "grad_norm": 5.526005268096924, "learning_rate": 4.672722365763821e-06, "loss": 0.068, "num_input_tokens_seen": 901552, "step": 136 }, { "epoch": 0.8817377312952535, "grad_norm": 6.142871856689453, "learning_rate": 4.667625604198842e-06, "loss": 0.1193, "num_input_tokens_seen": 908272, "step": 137 }, { "epoch": 0.8881737731295254, "grad_norm": 17.300273895263672, "learning_rate": 4.662492286476595e-06, "loss": 0.1535, "num_input_tokens_seen": 914928, "step": 138 }, { "epoch": 0.8946098149637972, "grad_norm": 13.767914772033691, "learning_rate": 4.657322499168475e-06, "loss": 0.1303, "num_input_tokens_seen": 921296, "step": 139 }, { "epoch": 0.9010458567980691, "grad_norm": 5.356888294219971, "learning_rate": 4.65211632946092e-06, "loss": 0.0879, "num_input_tokens_seen": 927728, "step": 140 }, { "epoch": 0.9074818986323411, "grad_norm": 10.261467933654785, "learning_rate": 4.646873865153945e-06, "loss": 0.0986, "num_input_tokens_seen": 934240, "step": 141 }, { "epoch": 0.913917940466613, "grad_norm": 14.075957298278809, "learning_rate": 4.641595194659657e-06, "loss": 0.1219, "num_input_tokens_seen": 940832, "step": 142 }, { "epoch": 0.9203539823008849, "grad_norm": 11.964951515197754, "learning_rate": 4.63628040700077e-06, "loss": 0.1303, "num_input_tokens_seen": 947856, "step": 143 }, { "epoch": 0.9267900241351569, "grad_norm": 6.297915935516357, "learning_rate": 4.630929591809095e-06, "loss": 0.081, "num_input_tokens_seen": 954160, "step": 144 }, { "epoch": 0.9332260659694288, "grad_norm": 4.006863594055176, "learning_rate": 4.625542839324036e-06, "loss": 0.0979, "num_input_tokens_seen": 960848, "step": 145 }, { "epoch": 0.9396621078037007, "grad_norm": 9.041242599487305, "learning_rate": 4.620120240391065e-06, "loss": 0.1446, "num_input_tokens_seen": 967440, "step": 146 }, { "epoch": 0.9460981496379727, "grad_norm": 14.858406066894531, "learning_rate": 4.614661886460191e-06, "loss": 0.1267, "num_input_tokens_seen": 973808, "step": 147 }, { "epoch": 0.9525341914722446, "grad_norm": 12.371238708496094, "learning_rate": 4.609167869584416e-06, "loss": 0.1095, "num_input_tokens_seen": 980352, "step": 148 }, { "epoch": 0.9589702333065165, "grad_norm": 6.89439582824707, "learning_rate": 4.6036382824181836e-06, "loss": 0.1252, "num_input_tokens_seen": 987088, "step": 149 }, { "epoch": 0.9654062751407884, "grad_norm": 3.6482529640197754, "learning_rate": 4.598073218215817e-06, "loss": 0.0645, "num_input_tokens_seen": 993648, "step": 150 }, { "epoch": 0.9718423169750603, "grad_norm": 6.078918933868408, "learning_rate": 4.592472770829945e-06, "loss": 0.0974, "num_input_tokens_seen": 1000272, "step": 151 }, { "epoch": 0.9782783588093322, "grad_norm": 10.974119186401367, "learning_rate": 4.586837034709921e-06, "loss": 0.0833, "num_input_tokens_seen": 1006912, "step": 152 }, { "epoch": 0.9847144006436042, "grad_norm": 8.552461624145508, "learning_rate": 4.581166104900228e-06, "loss": 0.0787, "num_input_tokens_seen": 1013328, "step": 153 }, { "epoch": 0.9911504424778761, "grad_norm": 8.927652359008789, "learning_rate": 4.575460077038877e-06, "loss": 0.0814, "num_input_tokens_seen": 1020128, "step": 154 }, { "epoch": 0.997586484312148, "grad_norm": 2.613471269607544, "learning_rate": 4.569719047355795e-06, "loss": 0.0278, "num_input_tokens_seen": 1026848, "step": 155 }, { "epoch": 1.00402252614642, "grad_norm": 4.19236946105957, "learning_rate": 4.5639431126712e-06, "loss": 0.093, "num_input_tokens_seen": 1033728, "step": 156 }, { "epoch": 1.010458567980692, "grad_norm": 6.943019866943359, "learning_rate": 4.5581323703939685e-06, "loss": 0.073, "num_input_tokens_seen": 1040720, "step": 157 }, { "epoch": 1.0168946098149638, "grad_norm": 6.545025825500488, "learning_rate": 4.552286918519996e-06, "loss": 0.0625, "num_input_tokens_seen": 1047168, "step": 158 }, { "epoch": 1.0233306516492358, "grad_norm": 7.890603065490723, "learning_rate": 4.5464068556305375e-06, "loss": 0.0461, "num_input_tokens_seen": 1053760, "step": 159 }, { "epoch": 1.0297666934835077, "grad_norm": 5.44887638092041, "learning_rate": 4.540492280890555e-06, "loss": 0.0318, "num_input_tokens_seen": 1060176, "step": 160 }, { "epoch": 1.0362027353177796, "grad_norm": 1.036007285118103, "learning_rate": 4.534543294047033e-06, "loss": 0.0068, "num_input_tokens_seen": 1066768, "step": 161 }, { "epoch": 1.0426387771520516, "grad_norm": 5.863292694091797, "learning_rate": 4.528559995427309e-06, "loss": 0.0462, "num_input_tokens_seen": 1073376, "step": 162 }, { "epoch": 1.0490748189863235, "grad_norm": 8.744257926940918, "learning_rate": 4.522542485937369e-06, "loss": 0.0487, "num_input_tokens_seen": 1079952, "step": 163 }, { "epoch": 1.0555108608205954, "grad_norm": 6.485115051269531, "learning_rate": 4.516490867060156e-06, "loss": 0.0664, "num_input_tokens_seen": 1086848, "step": 164 }, { "epoch": 1.0619469026548674, "grad_norm": 3.8945565223693848, "learning_rate": 4.5104052408538545e-06, "loss": 0.0347, "num_input_tokens_seen": 1093328, "step": 165 }, { "epoch": 1.068382944489139, "grad_norm": 3.5805532932281494, "learning_rate": 4.504285709950167e-06, "loss": 0.0202, "num_input_tokens_seen": 1099840, "step": 166 }, { "epoch": 1.074818986323411, "grad_norm": 6.033172607421875, "learning_rate": 4.498132377552587e-06, "loss": 0.0573, "num_input_tokens_seen": 1106528, "step": 167 }, { "epoch": 1.081255028157683, "grad_norm": 8.104386329650879, "learning_rate": 4.491945347434656e-06, "loss": 0.0848, "num_input_tokens_seen": 1113424, "step": 168 }, { "epoch": 1.0876910699919549, "grad_norm": 6.130471229553223, "learning_rate": 4.485724723938215e-06, "loss": 0.0464, "num_input_tokens_seen": 1120064, "step": 169 }, { "epoch": 1.0941271118262268, "grad_norm": 8.871036529541016, "learning_rate": 4.479470611971646e-06, "loss": 0.1004, "num_input_tokens_seen": 1126960, "step": 170 }, { "epoch": 1.1005631536604987, "grad_norm": 7.802618026733398, "learning_rate": 4.473183117008096e-06, "loss": 0.0842, "num_input_tokens_seen": 1133664, "step": 171 }, { "epoch": 1.1069991954947707, "grad_norm": 2.848886489868164, "learning_rate": 4.4668623450837085e-06, "loss": 0.0452, "num_input_tokens_seen": 1140048, "step": 172 }, { "epoch": 1.1134352373290426, "grad_norm": 2.9371185302734375, "learning_rate": 4.460508402795827e-06, "loss": 0.0225, "num_input_tokens_seen": 1146448, "step": 173 }, { "epoch": 1.1198712791633145, "grad_norm": 2.1428751945495605, "learning_rate": 4.4541213973012005e-06, "loss": 0.0058, "num_input_tokens_seen": 1152960, "step": 174 }, { "epoch": 1.1263073209975865, "grad_norm": 6.480560302734375, "learning_rate": 4.447701436314176e-06, "loss": 0.0565, "num_input_tokens_seen": 1159632, "step": 175 }, { "epoch": 1.1327433628318584, "grad_norm": 8.678375244140625, "learning_rate": 4.441248628104884e-06, "loss": 0.0591, "num_input_tokens_seen": 1166640, "step": 176 }, { "epoch": 1.1391794046661303, "grad_norm": 8.184906005859375, "learning_rate": 4.434763081497407e-06, "loss": 0.0488, "num_input_tokens_seen": 1173408, "step": 177 }, { "epoch": 1.1456154465004023, "grad_norm": 3.727961540222168, "learning_rate": 4.428244905867952e-06, "loss": 0.0318, "num_input_tokens_seen": 1179776, "step": 178 }, { "epoch": 1.1520514883346742, "grad_norm": 7.119325160980225, "learning_rate": 4.421694211142998e-06, "loss": 0.064, "num_input_tokens_seen": 1186720, "step": 179 }, { "epoch": 1.1584875301689461, "grad_norm": 3.094886541366577, "learning_rate": 4.415111107797445e-06, "loss": 0.0465, "num_input_tokens_seen": 1193408, "step": 180 }, { "epoch": 1.164923572003218, "grad_norm": 5.577038288116455, "learning_rate": 4.408495706852758e-06, "loss": 0.0344, "num_input_tokens_seen": 1200096, "step": 181 }, { "epoch": 1.17135961383749, "grad_norm": 7.607036590576172, "learning_rate": 4.401848119875081e-06, "loss": 0.0747, "num_input_tokens_seen": 1206848, "step": 182 }, { "epoch": 1.177795655671762, "grad_norm": 5.953075885772705, "learning_rate": 4.395168458973368e-06, "loss": 0.073, "num_input_tokens_seen": 1213632, "step": 183 }, { "epoch": 1.1842316975060339, "grad_norm": 7.784894943237305, "learning_rate": 4.388456836797484e-06, "loss": 0.0652, "num_input_tokens_seen": 1220336, "step": 184 }, { "epoch": 1.1906677393403058, "grad_norm": 6.535793781280518, "learning_rate": 4.381713366536312e-06, "loss": 0.0881, "num_input_tokens_seen": 1226736, "step": 185 }, { "epoch": 1.1971037811745777, "grad_norm": 4.9065093994140625, "learning_rate": 4.374938161915835e-06, "loss": 0.0676, "num_input_tokens_seen": 1233536, "step": 186 }, { "epoch": 1.2035398230088497, "grad_norm": 5.7732648849487305, "learning_rate": 4.368131337197228e-06, "loss": 0.0481, "num_input_tokens_seen": 1240032, "step": 187 }, { "epoch": 1.2099758648431216, "grad_norm": 5.656060218811035, "learning_rate": 4.361293007174926e-06, "loss": 0.0477, "num_input_tokens_seen": 1247008, "step": 188 }, { "epoch": 1.2164119066773935, "grad_norm": 3.2787587642669678, "learning_rate": 4.354423287174686e-06, "loss": 0.0456, "num_input_tokens_seen": 1254032, "step": 189 }, { "epoch": 1.2228479485116655, "grad_norm": 9.526626586914062, "learning_rate": 4.3475222930516484e-06, "loss": 0.133, "num_input_tokens_seen": 1261104, "step": 190 }, { "epoch": 1.2292839903459372, "grad_norm": 4.272536277770996, "learning_rate": 4.340590141188377e-06, "loss": 0.0672, "num_input_tokens_seen": 1267680, "step": 191 }, { "epoch": 1.235720032180209, "grad_norm": 4.5576701164245605, "learning_rate": 4.333626948492898e-06, "loss": 0.0352, "num_input_tokens_seen": 1274112, "step": 192 }, { "epoch": 1.242156074014481, "grad_norm": 2.7765443325042725, "learning_rate": 4.326632832396733e-06, "loss": 0.0361, "num_input_tokens_seen": 1280528, "step": 193 }, { "epoch": 1.248592115848753, "grad_norm": 2.681631565093994, "learning_rate": 4.319607910852911e-06, "loss": 0.0432, "num_input_tokens_seen": 1287232, "step": 194 }, { "epoch": 1.255028157683025, "grad_norm": 7.467050075531006, "learning_rate": 4.3125523023339825e-06, "loss": 0.0398, "num_input_tokens_seen": 1293792, "step": 195 }, { "epoch": 1.2614641995172968, "grad_norm": 2.6330530643463135, "learning_rate": 4.305466125830023e-06, "loss": 0.0473, "num_input_tokens_seen": 1300624, "step": 196 }, { "epoch": 1.2679002413515688, "grad_norm": 5.228641510009766, "learning_rate": 4.2983495008466285e-06, "loss": 0.0394, "num_input_tokens_seen": 1307520, "step": 197 }, { "epoch": 1.2743362831858407, "grad_norm": 5.0004191398620605, "learning_rate": 4.29120254740289e-06, "loss": 0.0645, "num_input_tokens_seen": 1313824, "step": 198 }, { "epoch": 1.2807723250201126, "grad_norm": 5.468844890594482, "learning_rate": 4.284025386029381e-06, "loss": 0.0663, "num_input_tokens_seen": 1319952, "step": 199 }, { "epoch": 1.2872083668543846, "grad_norm": 6.145412921905518, "learning_rate": 4.276818137766118e-06, "loss": 0.1067, "num_input_tokens_seen": 1326352, "step": 200 }, { "epoch": 1.2936444086886565, "grad_norm": 5.632473945617676, "learning_rate": 4.269580924160523e-06, "loss": 0.0603, "num_input_tokens_seen": 1332912, "step": 201 }, { "epoch": 1.3000804505229284, "grad_norm": 1.326751947402954, "learning_rate": 4.262313867265369e-06, "loss": 0.0055, "num_input_tokens_seen": 1339872, "step": 202 }, { "epoch": 1.3065164923572004, "grad_norm": 6.162146091461182, "learning_rate": 4.255017089636725e-06, "loss": 0.0547, "num_input_tokens_seen": 1346240, "step": 203 }, { "epoch": 1.3129525341914723, "grad_norm": 2.500483989715576, "learning_rate": 4.24769071433189e-06, "loss": 0.0684, "num_input_tokens_seen": 1353104, "step": 204 }, { "epoch": 1.3193885760257442, "grad_norm": 5.962297439575195, "learning_rate": 4.240334864907317e-06, "loss": 0.0484, "num_input_tokens_seen": 1359664, "step": 205 }, { "epoch": 1.3258246178600162, "grad_norm": 4.183216571807861, "learning_rate": 4.232949665416526e-06, "loss": 0.0225, "num_input_tokens_seen": 1366112, "step": 206 }, { "epoch": 1.332260659694288, "grad_norm": 2.874197244644165, "learning_rate": 4.225535240408014e-06, "loss": 0.0278, "num_input_tokens_seen": 1372912, "step": 207 }, { "epoch": 1.33869670152856, "grad_norm": 5.831293106079102, "learning_rate": 4.218091714923157e-06, "loss": 0.034, "num_input_tokens_seen": 1379200, "step": 208 }, { "epoch": 1.3451327433628317, "grad_norm": 4.147435665130615, "learning_rate": 4.210619214494099e-06, "loss": 0.0453, "num_input_tokens_seen": 1385520, "step": 209 }, { "epoch": 1.3515687851971037, "grad_norm": 6.03895378112793, "learning_rate": 4.203117865141635e-06, "loss": 0.0564, "num_input_tokens_seen": 1391968, "step": 210 }, { "epoch": 1.3580048270313756, "grad_norm": 4.196593284606934, "learning_rate": 4.195587793373085e-06, "loss": 0.0318, "num_input_tokens_seen": 1398576, "step": 211 }, { "epoch": 1.3644408688656475, "grad_norm": 6.364063739776611, "learning_rate": 4.188029126180161e-06, "loss": 0.0575, "num_input_tokens_seen": 1405280, "step": 212 }, { "epoch": 1.3708769106999195, "grad_norm": 5.420915603637695, "learning_rate": 4.180441991036827e-06, "loss": 0.0448, "num_input_tokens_seen": 1411968, "step": 213 }, { "epoch": 1.3773129525341914, "grad_norm": 5.313647747039795, "learning_rate": 4.172826515897146e-06, "loss": 0.0493, "num_input_tokens_seen": 1418576, "step": 214 }, { "epoch": 1.3837489943684633, "grad_norm": 3.2573652267456055, "learning_rate": 4.165182829193126e-06, "loss": 0.0478, "num_input_tokens_seen": 1425360, "step": 215 }, { "epoch": 1.3901850362027353, "grad_norm": 4.227644443511963, "learning_rate": 4.15751105983255e-06, "loss": 0.0887, "num_input_tokens_seen": 1432144, "step": 216 }, { "epoch": 1.3966210780370072, "grad_norm": 6.514432907104492, "learning_rate": 4.149811337196808e-06, "loss": 0.0634, "num_input_tokens_seen": 1439200, "step": 217 }, { "epoch": 1.4030571198712791, "grad_norm": 3.3998050689697266, "learning_rate": 4.142083791138703e-06, "loss": 0.0349, "num_input_tokens_seen": 1445728, "step": 218 }, { "epoch": 1.409493161705551, "grad_norm": 5.725708961486816, "learning_rate": 4.134328551980279e-06, "loss": 0.0459, "num_input_tokens_seen": 1452384, "step": 219 }, { "epoch": 1.415929203539823, "grad_norm": 3.3524420261383057, "learning_rate": 4.126545750510605e-06, "loss": 0.0304, "num_input_tokens_seen": 1459136, "step": 220 }, { "epoch": 1.422365245374095, "grad_norm": 9.169583320617676, "learning_rate": 4.118735517983584e-06, "loss": 0.0658, "num_input_tokens_seen": 1465632, "step": 221 }, { "epoch": 1.4288012872083669, "grad_norm": 5.310299873352051, "learning_rate": 4.110897986115729e-06, "loss": 0.087, "num_input_tokens_seen": 1472592, "step": 222 }, { "epoch": 1.4352373290426388, "grad_norm": 4.850796222686768, "learning_rate": 4.1030332870839466e-06, "loss": 0.0952, "num_input_tokens_seen": 1479168, "step": 223 }, { "epoch": 1.4416733708769107, "grad_norm": 5.20851993560791, "learning_rate": 4.0951415535233065e-06, "loss": 0.0358, "num_input_tokens_seen": 1485664, "step": 224 }, { "epoch": 1.4481094127111827, "grad_norm": 4.104648113250732, "learning_rate": 4.087222918524807e-06, "loss": 0.0527, "num_input_tokens_seen": 1492368, "step": 225 }, { "epoch": 1.4545454545454546, "grad_norm": 2.5263378620147705, "learning_rate": 4.079277515633127e-06, "loss": 0.0452, "num_input_tokens_seen": 1498752, "step": 226 }, { "epoch": 1.4609814963797265, "grad_norm": 2.5317678451538086, "learning_rate": 4.0713054788443776e-06, "loss": 0.0313, "num_input_tokens_seen": 1505296, "step": 227 }, { "epoch": 1.4674175382139985, "grad_norm": 5.61666202545166, "learning_rate": 4.063306942603835e-06, "loss": 0.0544, "num_input_tokens_seen": 1511584, "step": 228 }, { "epoch": 1.4738535800482704, "grad_norm": 3.4415996074676514, "learning_rate": 4.0552820418036855e-06, "loss": 0.0428, "num_input_tokens_seen": 1517776, "step": 229 }, { "epoch": 1.4802896218825423, "grad_norm": 2.44814395904541, "learning_rate": 4.0472309117807365e-06, "loss": 0.0183, "num_input_tokens_seen": 1524416, "step": 230 }, { "epoch": 1.4867256637168142, "grad_norm": 3.827312469482422, "learning_rate": 4.039153688314146e-06, "loss": 0.0662, "num_input_tokens_seen": 1530864, "step": 231 }, { "epoch": 1.4931617055510862, "grad_norm": 2.9351532459259033, "learning_rate": 4.031050507623125e-06, "loss": 0.0258, "num_input_tokens_seen": 1537216, "step": 232 }, { "epoch": 1.4995977473853581, "grad_norm": 6.040038585662842, "learning_rate": 4.022921506364644e-06, "loss": 0.0584, "num_input_tokens_seen": 1543824, "step": 233 }, { "epoch": 1.50603378921963, "grad_norm": 2.7363831996917725, "learning_rate": 4.014766821631128e-06, "loss": 0.0916, "num_input_tokens_seen": 1550432, "step": 234 }, { "epoch": 1.512469831053902, "grad_norm": 4.466485977172852, "learning_rate": 4.006586590948141e-06, "loss": 0.0397, "num_input_tokens_seen": 1556912, "step": 235 }, { "epoch": 1.518905872888174, "grad_norm": 4.36499547958374, "learning_rate": 3.998380952272073e-06, "loss": 0.0405, "num_input_tokens_seen": 1563456, "step": 236 }, { "epoch": 1.5253419147224458, "grad_norm": 3.1068978309631348, "learning_rate": 3.990150043987806e-06, "loss": 0.0645, "num_input_tokens_seen": 1570240, "step": 237 }, { "epoch": 1.5317779565567178, "grad_norm": 4.554339408874512, "learning_rate": 3.981894004906388e-06, "loss": 0.0389, "num_input_tokens_seen": 1576896, "step": 238 }, { "epoch": 1.5382139983909895, "grad_norm": 2.1207427978515625, "learning_rate": 3.973612974262685e-06, "loss": 0.0341, "num_input_tokens_seen": 1583440, "step": 239 }, { "epoch": 1.5446500402252614, "grad_norm": 4.71979284286499, "learning_rate": 3.965307091713037e-06, "loss": 0.0625, "num_input_tokens_seen": 1589968, "step": 240 }, { "epoch": 1.5510860820595334, "grad_norm": 3.9797351360321045, "learning_rate": 3.956976497332903e-06, "loss": 0.0651, "num_input_tokens_seen": 1596416, "step": 241 }, { "epoch": 1.5575221238938053, "grad_norm": 4.844697952270508, "learning_rate": 3.948621331614495e-06, "loss": 0.0391, "num_input_tokens_seen": 1602944, "step": 242 }, { "epoch": 1.5639581657280772, "grad_norm": 4.572307109832764, "learning_rate": 3.9402417354644115e-06, "loss": 0.0486, "num_input_tokens_seen": 1609632, "step": 243 }, { "epoch": 1.5703942075623492, "grad_norm": 7.0537309646606445, "learning_rate": 3.9318378502012636e-06, "loss": 0.1192, "num_input_tokens_seen": 1616096, "step": 244 }, { "epoch": 1.576830249396621, "grad_norm": 4.42478609085083, "learning_rate": 3.923409817553284e-06, "loss": 0.0679, "num_input_tokens_seen": 1622848, "step": 245 }, { "epoch": 1.583266291230893, "grad_norm": 5.157562255859375, "learning_rate": 3.914957779655946e-06, "loss": 0.0493, "num_input_tokens_seen": 1629600, "step": 246 }, { "epoch": 1.589702333065165, "grad_norm": 2.8394153118133545, "learning_rate": 3.906481879049559e-06, "loss": 0.0456, "num_input_tokens_seen": 1636192, "step": 247 }, { "epoch": 1.5961383748994369, "grad_norm": 2.4742684364318848, "learning_rate": 3.897982258676867e-06, "loss": 0.0391, "num_input_tokens_seen": 1642832, "step": 248 }, { "epoch": 1.6025744167337088, "grad_norm": 4.165124893188477, "learning_rate": 3.8894590618806435e-06, "loss": 0.0501, "num_input_tokens_seen": 1649904, "step": 249 }, { "epoch": 1.6090104585679805, "grad_norm": 2.7913286685943604, "learning_rate": 3.880912432401265e-06, "loss": 0.0397, "num_input_tokens_seen": 1656704, "step": 250 }, { "epoch": 1.6154465004022525, "grad_norm": 4.8400397300720215, "learning_rate": 3.872342514374291e-06, "loss": 0.0846, "num_input_tokens_seen": 1663680, "step": 251 }, { "epoch": 1.6218825422365244, "grad_norm": 3.111396074295044, "learning_rate": 3.863749452328035e-06, "loss": 0.0443, "num_input_tokens_seen": 1670160, "step": 252 }, { "epoch": 1.6283185840707963, "grad_norm": 3.1794304847717285, "learning_rate": 3.855133391181124e-06, "loss": 0.045, "num_input_tokens_seen": 1676832, "step": 253 }, { "epoch": 1.6347546259050683, "grad_norm": 1.6655223369598389, "learning_rate": 3.846494476240057e-06, "loss": 0.0172, "num_input_tokens_seen": 1683664, "step": 254 }, { "epoch": 1.6411906677393402, "grad_norm": 4.251989841461182, "learning_rate": 3.837832853196751e-06, "loss": 0.0949, "num_input_tokens_seen": 1690208, "step": 255 }, { "epoch": 1.6476267095736121, "grad_norm": 7.070593357086182, "learning_rate": 3.8291486681260904e-06, "loss": 0.0277, "num_input_tokens_seen": 1697296, "step": 256 }, { "epoch": 1.654062751407884, "grad_norm": 2.8217155933380127, "learning_rate": 3.820442067483455e-06, "loss": 0.0247, "num_input_tokens_seen": 1703504, "step": 257 }, { "epoch": 1.660498793242156, "grad_norm": 5.125271320343018, "learning_rate": 3.811713198102258e-06, "loss": 0.0549, "num_input_tokens_seen": 1710016, "step": 258 }, { "epoch": 1.666934835076428, "grad_norm": 5.227617263793945, "learning_rate": 3.802962207191463e-06, "loss": 0.0342, "num_input_tokens_seen": 1716960, "step": 259 }, { "epoch": 1.6733708769106999, "grad_norm": 3.3697738647460938, "learning_rate": 3.794189242333107e-06, "loss": 0.0617, "num_input_tokens_seen": 1723504, "step": 260 }, { "epoch": 1.6798069187449718, "grad_norm": 2.9104015827178955, "learning_rate": 3.785394451479806e-06, "loss": 0.0675, "num_input_tokens_seen": 1730160, "step": 261 }, { "epoch": 1.6862429605792437, "grad_norm": 4.513949394226074, "learning_rate": 3.7765779829522674e-06, "loss": 0.1055, "num_input_tokens_seen": 1736752, "step": 262 }, { "epoch": 1.6926790024135157, "grad_norm": 3.0852975845336914, "learning_rate": 3.7677399854367815e-06, "loss": 0.0355, "num_input_tokens_seen": 1743328, "step": 263 }, { "epoch": 1.6991150442477876, "grad_norm": 3.222297191619873, "learning_rate": 3.7588806079827147e-06, "loss": 0.0622, "num_input_tokens_seen": 1749776, "step": 264 }, { "epoch": 1.7055510860820595, "grad_norm": 2.017244338989258, "learning_rate": 3.7500000000000005e-06, "loss": 0.0297, "num_input_tokens_seen": 1756512, "step": 265 }, { "epoch": 1.7119871279163315, "grad_norm": 2.465116262435913, "learning_rate": 3.7410983112566166e-06, "loss": 0.0312, "num_input_tokens_seen": 1762928, "step": 266 }, { "epoch": 1.7184231697506034, "grad_norm": 2.8471832275390625, "learning_rate": 3.7321756918760587e-06, "loss": 0.0811, "num_input_tokens_seen": 1769392, "step": 267 }, { "epoch": 1.7248592115848753, "grad_norm": 3.4750540256500244, "learning_rate": 3.7232322923348093e-06, "loss": 0.067, "num_input_tokens_seen": 1776032, "step": 268 }, { "epoch": 1.7312952534191473, "grad_norm": 2.845557928085327, "learning_rate": 3.7142682634598016e-06, "loss": 0.0553, "num_input_tokens_seen": 1782512, "step": 269 }, { "epoch": 1.7377312952534192, "grad_norm": 2.0945403575897217, "learning_rate": 3.7052837564258728e-06, "loss": 0.021, "num_input_tokens_seen": 1789280, "step": 270 }, { "epoch": 1.7441673370876911, "grad_norm": 2.614729642868042, "learning_rate": 3.6962789227532165e-06, "loss": 0.0589, "num_input_tokens_seen": 1795696, "step": 271 }, { "epoch": 1.750603378921963, "grad_norm": 3.331339120864868, "learning_rate": 3.6872539143048287e-06, "loss": 0.0521, "num_input_tokens_seen": 1802448, "step": 272 }, { "epoch": 1.757039420756235, "grad_norm": 2.845620632171631, "learning_rate": 3.6782088832839436e-06, "loss": 0.0402, "num_input_tokens_seen": 1809264, "step": 273 }, { "epoch": 1.763475462590507, "grad_norm": 3.3971211910247803, "learning_rate": 3.6691439822314672e-06, "loss": 0.0363, "num_input_tokens_seen": 1815808, "step": 274 }, { "epoch": 1.7699115044247788, "grad_norm": 5.249027729034424, "learning_rate": 3.660059364023409e-06, "loss": 0.0523, "num_input_tokens_seen": 1822352, "step": 275 }, { "epoch": 1.7763475462590508, "grad_norm": 3.6546497344970703, "learning_rate": 3.650955181868298e-06, "loss": 0.0255, "num_input_tokens_seen": 1829056, "step": 276 }, { "epoch": 1.7827835880933227, "grad_norm": 7.767543792724609, "learning_rate": 3.641831589304602e-06, "loss": 0.1031, "num_input_tokens_seen": 1835696, "step": 277 }, { "epoch": 1.7892196299275946, "grad_norm": 1.5550068616867065, "learning_rate": 3.6326887401981386e-06, "loss": 0.0452, "num_input_tokens_seen": 1842288, "step": 278 }, { "epoch": 1.7956556717618666, "grad_norm": 4.8318986892700195, "learning_rate": 3.6235267887394774e-06, "loss": 0.0537, "num_input_tokens_seen": 1848960, "step": 279 }, { "epoch": 1.8020917135961385, "grad_norm": 4.691814422607422, "learning_rate": 3.6143458894413463e-06, "loss": 0.0572, "num_input_tokens_seen": 1855648, "step": 280 }, { "epoch": 1.8085277554304104, "grad_norm": 2.6937472820281982, "learning_rate": 3.6051461971360146e-06, "loss": 0.0298, "num_input_tokens_seen": 1862160, "step": 281 }, { "epoch": 1.8149637972646824, "grad_norm": 4.052839279174805, "learning_rate": 3.595927866972694e-06, "loss": 0.037, "num_input_tokens_seen": 1868896, "step": 282 }, { "epoch": 1.8213998390989543, "grad_norm": 5.030338287353516, "learning_rate": 3.586691054414913e-06, "loss": 0.0783, "num_input_tokens_seen": 1875248, "step": 283 }, { "epoch": 1.827835880933226, "grad_norm": 1.9826079607009888, "learning_rate": 3.577435915237899e-06, "loss": 0.0436, "num_input_tokens_seen": 1881728, "step": 284 }, { "epoch": 1.834271922767498, "grad_norm": 1.8905837535858154, "learning_rate": 3.5681626055259526e-06, "loss": 0.0258, "num_input_tokens_seen": 1888384, "step": 285 }, { "epoch": 1.8407079646017699, "grad_norm": 1.9678194522857666, "learning_rate": 3.558871281669811e-06, "loss": 0.0235, "num_input_tokens_seen": 1894864, "step": 286 }, { "epoch": 1.8471440064360418, "grad_norm": 4.199605464935303, "learning_rate": 3.549562100364014e-06, "loss": 0.0541, "num_input_tokens_seen": 1901680, "step": 287 }, { "epoch": 1.8535800482703138, "grad_norm": 4.100510120391846, "learning_rate": 3.5402352186042602e-06, "loss": 0.0767, "num_input_tokens_seen": 1908304, "step": 288 }, { "epoch": 1.8600160901045857, "grad_norm": 6.471580982208252, "learning_rate": 3.530890793684759e-06, "loss": 0.0558, "num_input_tokens_seen": 1914736, "step": 289 }, { "epoch": 1.8664521319388576, "grad_norm": 6.2181525230407715, "learning_rate": 3.521528983195579e-06, "loss": 0.0483, "num_input_tokens_seen": 1921088, "step": 290 }, { "epoch": 1.8728881737731295, "grad_norm": 3.5814297199249268, "learning_rate": 3.512149945019989e-06, "loss": 0.0389, "num_input_tokens_seen": 1927408, "step": 291 }, { "epoch": 1.8793242156074015, "grad_norm": 3.193094491958618, "learning_rate": 3.502753837331797e-06, "loss": 0.034, "num_input_tokens_seen": 1934160, "step": 292 }, { "epoch": 1.8857602574416734, "grad_norm": 3.2676048278808594, "learning_rate": 3.4933408185926805e-06, "loss": 0.0921, "num_input_tokens_seen": 1940912, "step": 293 }, { "epoch": 1.8921962992759453, "grad_norm": 4.060972690582275, "learning_rate": 3.4839110475495153e-06, "loss": 0.0661, "num_input_tokens_seen": 1947488, "step": 294 }, { "epoch": 1.898632341110217, "grad_norm": 4.40585470199585, "learning_rate": 3.4744646832316985e-06, "loss": 0.0301, "num_input_tokens_seen": 1954000, "step": 295 }, { "epoch": 1.905068382944489, "grad_norm": 4.472731113433838, "learning_rate": 3.465001884948468e-06, "loss": 0.0878, "num_input_tokens_seen": 1960400, "step": 296 }, { "epoch": 1.911504424778761, "grad_norm": 3.2221555709838867, "learning_rate": 3.45552281228621e-06, "loss": 0.1126, "num_input_tokens_seen": 1967728, "step": 297 }, { "epoch": 1.9179404666130329, "grad_norm": 3.6210269927978516, "learning_rate": 3.446027625105776e-06, "loss": 0.0679, "num_input_tokens_seen": 1974096, "step": 298 }, { "epoch": 1.9243765084473048, "grad_norm": 2.038454055786133, "learning_rate": 3.436516483539781e-06, "loss": 0.031, "num_input_tokens_seen": 1980672, "step": 299 }, { "epoch": 1.9308125502815767, "grad_norm": 2.2427828311920166, "learning_rate": 3.4269895479899023e-06, "loss": 0.0687, "num_input_tokens_seen": 1987104, "step": 300 }, { "epoch": 1.9372485921158487, "grad_norm": 6.37827730178833, "learning_rate": 3.4174469791241805e-06, "loss": 0.0497, "num_input_tokens_seen": 1994064, "step": 301 }, { "epoch": 1.9436846339501206, "grad_norm": 9.542262077331543, "learning_rate": 3.4078889378743036e-06, "loss": 0.0829, "num_input_tokens_seen": 2001056, "step": 302 }, { "epoch": 1.9501206757843925, "grad_norm": 6.237174034118652, "learning_rate": 3.3983155854328942e-06, "loss": 0.0578, "num_input_tokens_seen": 2007712, "step": 303 }, { "epoch": 1.9565567176186645, "grad_norm": 2.3653266429901123, "learning_rate": 3.388727083250795e-06, "loss": 0.0398, "num_input_tokens_seen": 2014368, "step": 304 }, { "epoch": 1.9629927594529364, "grad_norm": 3.9448723793029785, "learning_rate": 3.379123593034342e-06, "loss": 0.0754, "num_input_tokens_seen": 2020592, "step": 305 }, { "epoch": 1.9694288012872083, "grad_norm": 2.1158804893493652, "learning_rate": 3.369505276742638e-06, "loss": 0.0395, "num_input_tokens_seen": 2026864, "step": 306 }, { "epoch": 1.9758648431214803, "grad_norm": 5.131661891937256, "learning_rate": 3.359872296584821e-06, "loss": 0.0575, "num_input_tokens_seen": 2033440, "step": 307 }, { "epoch": 1.9823008849557522, "grad_norm": 5.0783867835998535, "learning_rate": 3.350224815017331e-06, "loss": 0.0472, "num_input_tokens_seen": 2039712, "step": 308 }, { "epoch": 1.9887369267900241, "grad_norm": 6.688424587249756, "learning_rate": 3.3405629947411687e-06, "loss": 0.0498, "num_input_tokens_seen": 2046576, "step": 309 }, { "epoch": 1.995172968624296, "grad_norm": 5.24268102645874, "learning_rate": 3.3308869986991493e-06, "loss": 0.0447, "num_input_tokens_seen": 2053248, "step": 310 }, { "epoch": 2.001609010458568, "grad_norm": 1.7300570011138916, "learning_rate": 3.32119699007316e-06, "loss": 0.0155, "num_input_tokens_seen": 2059840, "step": 311 }, { "epoch": 2.00804505229284, "grad_norm": 2.5391845703125, "learning_rate": 3.311493132281402e-06, "loss": 0.0183, "num_input_tokens_seen": 2066384, "step": 312 }, { "epoch": 2.014481094127112, "grad_norm": 1.9404152631759644, "learning_rate": 3.3017755889756382e-06, "loss": 0.0102, "num_input_tokens_seen": 2073088, "step": 313 }, { "epoch": 2.020917135961384, "grad_norm": 1.1497960090637207, "learning_rate": 3.292044524038433e-06, "loss": 0.0119, "num_input_tokens_seen": 2079600, "step": 314 }, { "epoch": 2.0273531777956557, "grad_norm": 0.6188907027244568, "learning_rate": 3.2823001015803863e-06, "loss": 0.0037, "num_input_tokens_seen": 2086080, "step": 315 }, { "epoch": 2.0337892196299276, "grad_norm": 2.5652434825897217, "learning_rate": 3.272542485937369e-06, "loss": 0.0048, "num_input_tokens_seen": 2092768, "step": 316 }, { "epoch": 2.0402252614641996, "grad_norm": 1.3636257648468018, "learning_rate": 3.2627718416677484e-06, "loss": 0.004, "num_input_tokens_seen": 2099296, "step": 317 }, { "epoch": 2.0466613032984715, "grad_norm": 3.7406702041625977, "learning_rate": 3.2529883335496163e-06, "loss": 0.0472, "num_input_tokens_seen": 2106176, "step": 318 }, { "epoch": 2.0530973451327434, "grad_norm": 0.2876489460468292, "learning_rate": 3.243192126578007e-06, "loss": 0.0008, "num_input_tokens_seen": 2112560, "step": 319 }, { "epoch": 2.0595333869670154, "grad_norm": 3.388899087905884, "learning_rate": 3.2333833859621155e-06, "loss": 0.0332, "num_input_tokens_seen": 2119296, "step": 320 }, { "epoch": 2.0659694288012873, "grad_norm": 2.6212401390075684, "learning_rate": 3.223562277122513e-06, "loss": 0.0434, "num_input_tokens_seen": 2125632, "step": 321 }, { "epoch": 2.0724054706355592, "grad_norm": 3.6854021549224854, "learning_rate": 3.213728965688356e-06, "loss": 0.0105, "num_input_tokens_seen": 2132096, "step": 322 }, { "epoch": 2.078841512469831, "grad_norm": 3.9269893169403076, "learning_rate": 3.2038836174945907e-06, "loss": 0.0188, "num_input_tokens_seen": 2138336, "step": 323 }, { "epoch": 2.085277554304103, "grad_norm": 2.3363194465637207, "learning_rate": 3.194026398579162e-06, "loss": 0.0382, "num_input_tokens_seen": 2144672, "step": 324 }, { "epoch": 2.091713596138375, "grad_norm": 0.16176919639110565, "learning_rate": 3.184157475180208e-06, "loss": 0.0002, "num_input_tokens_seen": 2151216, "step": 325 }, { "epoch": 2.098149637972647, "grad_norm": 7.4007368087768555, "learning_rate": 3.1742770137332567e-06, "loss": 0.0473, "num_input_tokens_seen": 2158000, "step": 326 }, { "epoch": 2.104585679806919, "grad_norm": 0.3990660607814789, "learning_rate": 3.164385180868425e-06, "loss": 0.0008, "num_input_tokens_seen": 2164448, "step": 327 }, { "epoch": 2.111021721641191, "grad_norm": 5.447741508483887, "learning_rate": 3.1544821434076013e-06, "loss": 0.0123, "num_input_tokens_seen": 2171120, "step": 328 }, { "epoch": 2.1174577634754628, "grad_norm": 4.229776382446289, "learning_rate": 3.144568068361634e-06, "loss": 0.03, "num_input_tokens_seen": 2177648, "step": 329 }, { "epoch": 2.1238938053097347, "grad_norm": 5.920961380004883, "learning_rate": 3.1346431229275197e-06, "loss": 0.0207, "num_input_tokens_seen": 2183856, "step": 330 }, { "epoch": 2.1303298471440066, "grad_norm": 11.779773712158203, "learning_rate": 3.124707474485577e-06, "loss": 0.0172, "num_input_tokens_seen": 2190608, "step": 331 }, { "epoch": 2.136765888978278, "grad_norm": 8.82557201385498, "learning_rate": 3.1147612905966286e-06, "loss": 0.0115, "num_input_tokens_seen": 2197232, "step": 332 }, { "epoch": 2.14320193081255, "grad_norm": 1.1176470518112183, "learning_rate": 3.1048047389991693e-06, "loss": 0.0217, "num_input_tokens_seen": 2203456, "step": 333 }, { "epoch": 2.149637972646822, "grad_norm": 0.8806192278862, "learning_rate": 3.094837987606547e-06, "loss": 0.0035, "num_input_tokens_seen": 2209856, "step": 334 }, { "epoch": 2.156074014481094, "grad_norm": 6.793837547302246, "learning_rate": 3.084861204504122e-06, "loss": 0.0426, "num_input_tokens_seen": 2216400, "step": 335 }, { "epoch": 2.162510056315366, "grad_norm": 5.5410075187683105, "learning_rate": 3.0748745579464347e-06, "loss": 0.0382, "num_input_tokens_seen": 2222864, "step": 336 }, { "epoch": 2.168946098149638, "grad_norm": 1.8118884563446045, "learning_rate": 3.0648782163543696e-06, "loss": 0.0082, "num_input_tokens_seen": 2229760, "step": 337 }, { "epoch": 2.1753821399839097, "grad_norm": 2.607206106185913, "learning_rate": 3.0548723483123157e-06, "loss": 0.0338, "num_input_tokens_seen": 2236368, "step": 338 }, { "epoch": 2.1818181818181817, "grad_norm": 1.080344557762146, "learning_rate": 3.0448571225653195e-06, "loss": 0.0141, "num_input_tokens_seen": 2242816, "step": 339 }, { "epoch": 2.1882542236524536, "grad_norm": 2.380739212036133, "learning_rate": 3.0348327080162438e-06, "loss": 0.0287, "num_input_tokens_seen": 2249488, "step": 340 }, { "epoch": 2.1946902654867255, "grad_norm": 1.0098868608474731, "learning_rate": 3.0247992737229147e-06, "loss": 0.0027, "num_input_tokens_seen": 2255968, "step": 341 }, { "epoch": 2.2011263073209975, "grad_norm": 0.7940512895584106, "learning_rate": 3.014756988895275e-06, "loss": 0.0026, "num_input_tokens_seen": 2262544, "step": 342 }, { "epoch": 2.2075623491552694, "grad_norm": 2.9759926795959473, "learning_rate": 3.0047060228925256e-06, "loss": 0.039, "num_input_tokens_seen": 2269312, "step": 343 }, { "epoch": 2.2139983909895413, "grad_norm": 4.84032678604126, "learning_rate": 2.994646545220275e-06, "loss": 0.0154, "num_input_tokens_seen": 2275968, "step": 344 }, { "epoch": 2.2204344328238133, "grad_norm": 2.9671568870544434, "learning_rate": 2.9845787255276753e-06, "loss": 0.0231, "num_input_tokens_seen": 2282976, "step": 345 }, { "epoch": 2.226870474658085, "grad_norm": 5.410647392272949, "learning_rate": 2.9745027336045652e-06, "loss": 0.04, "num_input_tokens_seen": 2289696, "step": 346 }, { "epoch": 2.233306516492357, "grad_norm": 5.828602313995361, "learning_rate": 2.964418739378603e-06, "loss": 0.0282, "num_input_tokens_seen": 2296272, "step": 347 }, { "epoch": 2.239742558326629, "grad_norm": 1.9481452703475952, "learning_rate": 2.954326912912404e-06, "loss": 0.0143, "num_input_tokens_seen": 2303120, "step": 348 }, { "epoch": 2.246178600160901, "grad_norm": 3.2762415409088135, "learning_rate": 2.9442274244006725e-06, "loss": 0.0194, "num_input_tokens_seen": 2309728, "step": 349 }, { "epoch": 2.252614641995173, "grad_norm": 2.3237709999084473, "learning_rate": 2.9341204441673267e-06, "loss": 0.0051, "num_input_tokens_seen": 2316144, "step": 350 }, { "epoch": 2.259050683829445, "grad_norm": 1.7801238298416138, "learning_rate": 2.924006142662632e-06, "loss": 0.0162, "num_input_tokens_seen": 2322768, "step": 351 }, { "epoch": 2.265486725663717, "grad_norm": 4.876129150390625, "learning_rate": 2.913884690460325e-06, "loss": 0.0313, "num_input_tokens_seen": 2329312, "step": 352 }, { "epoch": 2.2719227674979887, "grad_norm": 0.9637519717216492, "learning_rate": 2.903756258254734e-06, "loss": 0.0041, "num_input_tokens_seen": 2335824, "step": 353 }, { "epoch": 2.2783588093322606, "grad_norm": 2.7481493949890137, "learning_rate": 2.8936210168579043e-06, "loss": 0.0321, "num_input_tokens_seen": 2342272, "step": 354 }, { "epoch": 2.2847948511665326, "grad_norm": 1.682763934135437, "learning_rate": 2.883479137196714e-06, "loss": 0.0064, "num_input_tokens_seen": 2349056, "step": 355 }, { "epoch": 2.2912308930008045, "grad_norm": 5.632142066955566, "learning_rate": 2.8733307903099926e-06, "loss": 0.0237, "num_input_tokens_seen": 2355552, "step": 356 }, { "epoch": 2.2976669348350764, "grad_norm": 2.460470199584961, "learning_rate": 2.8631761473456377e-06, "loss": 0.0152, "num_input_tokens_seen": 2361808, "step": 357 }, { "epoch": 2.3041029766693484, "grad_norm": 0.9998040199279785, "learning_rate": 2.853015379557729e-06, "loss": 0.0038, "num_input_tokens_seen": 2368288, "step": 358 }, { "epoch": 2.3105390185036203, "grad_norm": 3.164407968521118, "learning_rate": 2.842848658303637e-06, "loss": 0.0168, "num_input_tokens_seen": 2374960, "step": 359 }, { "epoch": 2.3169750603378922, "grad_norm": 2.3879611492156982, "learning_rate": 2.832676155041135e-06, "loss": 0.0049, "num_input_tokens_seen": 2381776, "step": 360 }, { "epoch": 2.323411102172164, "grad_norm": 1.3164470195770264, "learning_rate": 2.822498041325509e-06, "loss": 0.0114, "num_input_tokens_seen": 2388112, "step": 361 }, { "epoch": 2.329847144006436, "grad_norm": 2.3726656436920166, "learning_rate": 2.8123144888066623e-06, "loss": 0.022, "num_input_tokens_seen": 2394736, "step": 362 }, { "epoch": 2.336283185840708, "grad_norm": 1.7789826393127441, "learning_rate": 2.802125669226222e-06, "loss": 0.0154, "num_input_tokens_seen": 2401248, "step": 363 }, { "epoch": 2.34271922767498, "grad_norm": 3.68959641456604, "learning_rate": 2.7919317544146405e-06, "loss": 0.0204, "num_input_tokens_seen": 2407872, "step": 364 }, { "epoch": 2.349155269509252, "grad_norm": 2.4927353858947754, "learning_rate": 2.7817329162883033e-06, "loss": 0.0334, "num_input_tokens_seen": 2414432, "step": 365 }, { "epoch": 2.355591311343524, "grad_norm": 4.594964504241943, "learning_rate": 2.7715293268466204e-06, "loss": 0.0132, "num_input_tokens_seen": 2420848, "step": 366 }, { "epoch": 2.3620273531777958, "grad_norm": 4.325422286987305, "learning_rate": 2.761321158169134e-06, "loss": 0.0291, "num_input_tokens_seen": 2427728, "step": 367 }, { "epoch": 2.3684633950120677, "grad_norm": 2.46122407913208, "learning_rate": 2.7511085824126133e-06, "loss": 0.0089, "num_input_tokens_seen": 2434880, "step": 368 }, { "epoch": 2.3748994368463396, "grad_norm": 2.729311227798462, "learning_rate": 2.74089177180815e-06, "loss": 0.0306, "num_input_tokens_seen": 2441168, "step": 369 }, { "epoch": 2.3813354786806116, "grad_norm": 5.095163345336914, "learning_rate": 2.730670898658255e-06, "loss": 0.0297, "num_input_tokens_seen": 2447920, "step": 370 }, { "epoch": 2.3877715205148835, "grad_norm": 1.902287483215332, "learning_rate": 2.7204461353339546e-06, "loss": 0.0247, "num_input_tokens_seen": 2454704, "step": 371 }, { "epoch": 2.3942075623491554, "grad_norm": 3.267244577407837, "learning_rate": 2.7102176542718783e-06, "loss": 0.0234, "num_input_tokens_seen": 2461216, "step": 372 }, { "epoch": 2.4006436041834274, "grad_norm": 4.101126670837402, "learning_rate": 2.699985627971354e-06, "loss": 0.0192, "num_input_tokens_seen": 2468032, "step": 373 }, { "epoch": 2.4070796460176993, "grad_norm": 4.104948997497559, "learning_rate": 2.689750228991503e-06, "loss": 0.0324, "num_input_tokens_seen": 2474544, "step": 374 }, { "epoch": 2.4135156878519712, "grad_norm": 2.1446776390075684, "learning_rate": 2.679511629948319e-06, "loss": 0.0332, "num_input_tokens_seen": 2481312, "step": 375 }, { "epoch": 2.419951729686243, "grad_norm": 0.7457873225212097, "learning_rate": 2.669270003511769e-06, "loss": 0.0043, "num_input_tokens_seen": 2487888, "step": 376 }, { "epoch": 2.426387771520515, "grad_norm": 2.1420276165008545, "learning_rate": 2.6590255224028725e-06, "loss": 0.0197, "num_input_tokens_seen": 2494784, "step": 377 }, { "epoch": 2.432823813354787, "grad_norm": 3.0415239334106445, "learning_rate": 2.648778359390794e-06, "loss": 0.0366, "num_input_tokens_seen": 2501712, "step": 378 }, { "epoch": 2.439259855189059, "grad_norm": 3.6502788066864014, "learning_rate": 2.638528687289925e-06, "loss": 0.0173, "num_input_tokens_seen": 2508592, "step": 379 }, { "epoch": 2.445695897023331, "grad_norm": 2.2913506031036377, "learning_rate": 2.6282766789569742e-06, "loss": 0.0102, "num_input_tokens_seen": 2515216, "step": 380 }, { "epoch": 2.4521319388576024, "grad_norm": 3.3507297039031982, "learning_rate": 2.618022507288049e-06, "loss": 0.0361, "num_input_tokens_seen": 2522064, "step": 381 }, { "epoch": 2.4585679806918743, "grad_norm": 2.98098087310791, "learning_rate": 2.6077663452157398e-06, "loss": 0.0292, "num_input_tokens_seen": 2528608, "step": 382 }, { "epoch": 2.4650040225261463, "grad_norm": 1.4962135553359985, "learning_rate": 2.5975083657062043e-06, "loss": 0.0095, "num_input_tokens_seen": 2535328, "step": 383 }, { "epoch": 2.471440064360418, "grad_norm": 2.0819742679595947, "learning_rate": 2.587248741756253e-06, "loss": 0.015, "num_input_tokens_seen": 2542224, "step": 384 }, { "epoch": 2.47787610619469, "grad_norm": 1.8906433582305908, "learning_rate": 2.576987646390426e-06, "loss": 0.0276, "num_input_tokens_seen": 2548976, "step": 385 }, { "epoch": 2.484312148028962, "grad_norm": 2.451510190963745, "learning_rate": 2.566725252658081e-06, "loss": 0.0284, "num_input_tokens_seen": 2555568, "step": 386 }, { "epoch": 2.490748189863234, "grad_norm": 3.7337939739227295, "learning_rate": 2.5564617336304703e-06, "loss": 0.0366, "num_input_tokens_seen": 2562128, "step": 387 }, { "epoch": 2.497184231697506, "grad_norm": 1.6401593685150146, "learning_rate": 2.546197262397825e-06, "loss": 0.0322, "num_input_tokens_seen": 2568640, "step": 388 }, { "epoch": 2.503620273531778, "grad_norm": 0.9136457443237305, "learning_rate": 2.535932012066434e-06, "loss": 0.0057, "num_input_tokens_seen": 2575024, "step": 389 }, { "epoch": 2.51005631536605, "grad_norm": 1.119612455368042, "learning_rate": 2.525666155755725e-06, "loss": 0.0054, "num_input_tokens_seen": 2581520, "step": 390 }, { "epoch": 2.5164923572003217, "grad_norm": 2.4770889282226562, "learning_rate": 2.515399866595347e-06, "loss": 0.0199, "num_input_tokens_seen": 2588528, "step": 391 }, { "epoch": 2.5229283990345936, "grad_norm": 0.35335639119148254, "learning_rate": 2.5051333177222476e-06, "loss": 0.0045, "num_input_tokens_seen": 2594992, "step": 392 }, { "epoch": 2.5293644408688656, "grad_norm": 2.8933093547821045, "learning_rate": 2.4948666822777536e-06, "loss": 0.0283, "num_input_tokens_seen": 2601568, "step": 393 }, { "epoch": 2.5358004827031375, "grad_norm": 1.7032990455627441, "learning_rate": 2.4846001334046537e-06, "loss": 0.0248, "num_input_tokens_seen": 2608160, "step": 394 }, { "epoch": 2.5422365245374094, "grad_norm": 1.9688091278076172, "learning_rate": 2.474333844244276e-06, "loss": 0.0132, "num_input_tokens_seen": 2614656, "step": 395 }, { "epoch": 2.5486725663716814, "grad_norm": 3.135990619659424, "learning_rate": 2.464067987933567e-06, "loss": 0.04, "num_input_tokens_seen": 2621600, "step": 396 }, { "epoch": 2.5551086082059533, "grad_norm": 0.7140212059020996, "learning_rate": 2.453802737602176e-06, "loss": 0.0029, "num_input_tokens_seen": 2627984, "step": 397 }, { "epoch": 2.5615446500402252, "grad_norm": 3.9643640518188477, "learning_rate": 2.4435382663695305e-06, "loss": 0.0254, "num_input_tokens_seen": 2634720, "step": 398 }, { "epoch": 2.567980691874497, "grad_norm": 2.284302234649658, "learning_rate": 2.4332747473419193e-06, "loss": 0.0108, "num_input_tokens_seen": 2641456, "step": 399 }, { "epoch": 2.574416733708769, "grad_norm": 2.6400082111358643, "learning_rate": 2.4230123536095746e-06, "loss": 0.0269, "num_input_tokens_seen": 2647760, "step": 400 }, { "epoch": 2.580852775543041, "grad_norm": 3.1969995498657227, "learning_rate": 2.4127512582437486e-06, "loss": 0.0111, "num_input_tokens_seen": 2654608, "step": 401 }, { "epoch": 2.587288817377313, "grad_norm": 3.651118516921997, "learning_rate": 2.4024916342937966e-06, "loss": 0.0222, "num_input_tokens_seen": 2661072, "step": 402 }, { "epoch": 2.593724859211585, "grad_norm": 2.1281003952026367, "learning_rate": 2.392233654784262e-06, "loss": 0.0101, "num_input_tokens_seen": 2667712, "step": 403 }, { "epoch": 2.600160901045857, "grad_norm": 2.6782784461975098, "learning_rate": 2.3819774927119523e-06, "loss": 0.0138, "num_input_tokens_seen": 2674496, "step": 404 }, { "epoch": 2.6065969428801288, "grad_norm": 2.2902138233184814, "learning_rate": 2.3717233210430258e-06, "loss": 0.0281, "num_input_tokens_seen": 2680816, "step": 405 }, { "epoch": 2.6130329847144007, "grad_norm": 1.9150536060333252, "learning_rate": 2.3614713127100752e-06, "loss": 0.0042, "num_input_tokens_seen": 2687632, "step": 406 }, { "epoch": 2.6194690265486726, "grad_norm": 0.3568836748600006, "learning_rate": 2.3512216406092066e-06, "loss": 0.0015, "num_input_tokens_seen": 2694464, "step": 407 }, { "epoch": 2.6259050683829446, "grad_norm": 2.3506011962890625, "learning_rate": 2.340974477597128e-06, "loss": 0.0279, "num_input_tokens_seen": 2701344, "step": 408 }, { "epoch": 2.6323411102172165, "grad_norm": 2.780200481414795, "learning_rate": 2.3307299964882314e-06, "loss": 0.0399, "num_input_tokens_seen": 2707536, "step": 409 }, { "epoch": 2.6387771520514884, "grad_norm": 1.1793303489685059, "learning_rate": 2.3204883700516813e-06, "loss": 0.0074, "num_input_tokens_seen": 2714544, "step": 410 }, { "epoch": 2.6452131938857604, "grad_norm": 1.7807022333145142, "learning_rate": 2.310249771008498e-06, "loss": 0.0078, "num_input_tokens_seen": 2721056, "step": 411 }, { "epoch": 2.6516492357200323, "grad_norm": 12.764676094055176, "learning_rate": 2.3000143720286463e-06, "loss": 0.0406, "num_input_tokens_seen": 2727664, "step": 412 }, { "epoch": 2.6580852775543042, "grad_norm": 0.44338610768318176, "learning_rate": 2.2897823457281225e-06, "loss": 0.0023, "num_input_tokens_seen": 2733600, "step": 413 }, { "epoch": 2.664521319388576, "grad_norm": 3.5756232738494873, "learning_rate": 2.2795538646660462e-06, "loss": 0.006, "num_input_tokens_seen": 2740400, "step": 414 }, { "epoch": 2.670957361222848, "grad_norm": 1.4301191568374634, "learning_rate": 2.269329101341745e-06, "loss": 0.0236, "num_input_tokens_seen": 2747248, "step": 415 }, { "epoch": 2.67739340305712, "grad_norm": 2.0859804153442383, "learning_rate": 2.2591082281918507e-06, "loss": 0.0136, "num_input_tokens_seen": 2753776, "step": 416 }, { "epoch": 2.6838294448913915, "grad_norm": 2.9704370498657227, "learning_rate": 2.2488914175873876e-06, "loss": 0.015, "num_input_tokens_seen": 2760720, "step": 417 }, { "epoch": 2.6902654867256635, "grad_norm": 3.1178269386291504, "learning_rate": 2.238678841830867e-06, "loss": 0.0483, "num_input_tokens_seen": 2767136, "step": 418 }, { "epoch": 2.6967015285599354, "grad_norm": 0.6049777269363403, "learning_rate": 2.2284706731533805e-06, "loss": 0.0014, "num_input_tokens_seen": 2773680, "step": 419 }, { "epoch": 2.7031375703942073, "grad_norm": 3.5615270137786865, "learning_rate": 2.2182670837116975e-06, "loss": 0.0279, "num_input_tokens_seen": 2780160, "step": 420 }, { "epoch": 2.7095736122284793, "grad_norm": 3.4241111278533936, "learning_rate": 2.20806824558536e-06, "loss": 0.0705, "num_input_tokens_seen": 2786912, "step": 421 }, { "epoch": 2.716009654062751, "grad_norm": 1.0644826889038086, "learning_rate": 2.197874330773779e-06, "loss": 0.005, "num_input_tokens_seen": 2793888, "step": 422 }, { "epoch": 2.722445695897023, "grad_norm": 5.071107387542725, "learning_rate": 2.1876855111933385e-06, "loss": 0.0453, "num_input_tokens_seen": 2800320, "step": 423 }, { "epoch": 2.728881737731295, "grad_norm": 1.9479647874832153, "learning_rate": 2.1775019586744924e-06, "loss": 0.0095, "num_input_tokens_seen": 2807088, "step": 424 }, { "epoch": 2.735317779565567, "grad_norm": 2.730952262878418, "learning_rate": 2.167323844958867e-06, "loss": 0.0095, "num_input_tokens_seen": 2813312, "step": 425 }, { "epoch": 2.741753821399839, "grad_norm": 2.1456387042999268, "learning_rate": 2.1571513416963647e-06, "loss": 0.0138, "num_input_tokens_seen": 2819936, "step": 426 }, { "epoch": 2.748189863234111, "grad_norm": 2.14911150932312, "learning_rate": 2.1469846204422724e-06, "loss": 0.0272, "num_input_tokens_seen": 2826224, "step": 427 }, { "epoch": 2.754625905068383, "grad_norm": 0.5957837700843811, "learning_rate": 2.136823852654363e-06, "loss": 0.0026, "num_input_tokens_seen": 2832960, "step": 428 }, { "epoch": 2.7610619469026547, "grad_norm": 0.3253982961177826, "learning_rate": 2.126669209690008e-06, "loss": 0.0016, "num_input_tokens_seen": 2839888, "step": 429 }, { "epoch": 2.7674979887369267, "grad_norm": 3.472017765045166, "learning_rate": 2.1165208628032863e-06, "loss": 0.0633, "num_input_tokens_seen": 2846688, "step": 430 }, { "epoch": 2.7739340305711986, "grad_norm": 2.029026985168457, "learning_rate": 2.1063789831420957e-06, "loss": 0.0191, "num_input_tokens_seen": 2853184, "step": 431 }, { "epoch": 2.7803700724054705, "grad_norm": 2.316349506378174, "learning_rate": 2.096243741745266e-06, "loss": 0.0075, "num_input_tokens_seen": 2859632, "step": 432 }, { "epoch": 2.7868061142397424, "grad_norm": 3.786245346069336, "learning_rate": 2.086115309539675e-06, "loss": 0.0371, "num_input_tokens_seen": 2865920, "step": 433 }, { "epoch": 2.7932421560740144, "grad_norm": 1.864402413368225, "learning_rate": 2.0759938573373683e-06, "loss": 0.0275, "num_input_tokens_seen": 2872464, "step": 434 }, { "epoch": 2.7996781979082863, "grad_norm": 8.142292022705078, "learning_rate": 2.0658795558326745e-06, "loss": 0.0441, "num_input_tokens_seen": 2879168, "step": 435 }, { "epoch": 2.8061142397425582, "grad_norm": 1.3945283889770508, "learning_rate": 2.0557725755993283e-06, "loss": 0.0074, "num_input_tokens_seen": 2885520, "step": 436 }, { "epoch": 2.81255028157683, "grad_norm": 1.951145887374878, "learning_rate": 2.0456730870875964e-06, "loss": 0.0412, "num_input_tokens_seen": 2892368, "step": 437 }, { "epoch": 2.818986323411102, "grad_norm": 1.666693925857544, "learning_rate": 2.035581260621398e-06, "loss": 0.007, "num_input_tokens_seen": 2898640, "step": 438 }, { "epoch": 2.825422365245374, "grad_norm": 0.8178473114967346, "learning_rate": 2.0254972663954356e-06, "loss": 0.0195, "num_input_tokens_seen": 2905312, "step": 439 }, { "epoch": 2.831858407079646, "grad_norm": 2.1499900817871094, "learning_rate": 2.015421274472325e-06, "loss": 0.0104, "num_input_tokens_seen": 2911872, "step": 440 }, { "epoch": 2.838294448913918, "grad_norm": 3.162245273590088, "learning_rate": 2.005353454779726e-06, "loss": 0.0196, "num_input_tokens_seen": 2918496, "step": 441 }, { "epoch": 2.84473049074819, "grad_norm": 1.1920592784881592, "learning_rate": 1.995293977107475e-06, "loss": 0.0131, "num_input_tokens_seen": 2924944, "step": 442 }, { "epoch": 2.8511665325824618, "grad_norm": 1.091436743736267, "learning_rate": 1.9852430111047254e-06, "loss": 0.0072, "num_input_tokens_seen": 2931440, "step": 443 }, { "epoch": 2.8576025744167337, "grad_norm": 2.0469212532043457, "learning_rate": 1.9752007262770857e-06, "loss": 0.0058, "num_input_tokens_seen": 2938304, "step": 444 }, { "epoch": 2.8640386162510056, "grad_norm": 1.6995850801467896, "learning_rate": 1.965167291983757e-06, "loss": 0.0242, "num_input_tokens_seen": 2945168, "step": 445 }, { "epoch": 2.8704746580852776, "grad_norm": 5.4955735206604, "learning_rate": 1.955142877434681e-06, "loss": 0.0323, "num_input_tokens_seen": 2951952, "step": 446 }, { "epoch": 2.8769106999195495, "grad_norm": 1.5203238725662231, "learning_rate": 1.9451276516876856e-06, "loss": 0.0186, "num_input_tokens_seen": 2958432, "step": 447 }, { "epoch": 2.8833467417538214, "grad_norm": 1.398633599281311, "learning_rate": 1.9351217836456316e-06, "loss": 0.0071, "num_input_tokens_seen": 2965328, "step": 448 }, { "epoch": 2.8897827835880934, "grad_norm": 1.4775344133377075, "learning_rate": 1.9251254420535665e-06, "loss": 0.0177, "num_input_tokens_seen": 2971872, "step": 449 }, { "epoch": 2.8962188254223653, "grad_norm": 3.7046666145324707, "learning_rate": 1.9151387954958792e-06, "loss": 0.044, "num_input_tokens_seen": 2978784, "step": 450 }, { "epoch": 2.9026548672566372, "grad_norm": 1.9969475269317627, "learning_rate": 1.9051620123934538e-06, "loss": 0.0119, "num_input_tokens_seen": 2985760, "step": 451 }, { "epoch": 2.909090909090909, "grad_norm": 1.3861935138702393, "learning_rate": 1.895195261000831e-06, "loss": 0.0121, "num_input_tokens_seen": 2992352, "step": 452 }, { "epoch": 2.915526950925181, "grad_norm": 2.0632236003875732, "learning_rate": 1.885238709403372e-06, "loss": 0.0319, "num_input_tokens_seen": 2998800, "step": 453 }, { "epoch": 2.921962992759453, "grad_norm": 0.31324344873428345, "learning_rate": 1.8752925255144228e-06, "loss": 0.0024, "num_input_tokens_seen": 3005392, "step": 454 }, { "epoch": 2.928399034593725, "grad_norm": 1.0096696615219116, "learning_rate": 1.8653568770724805e-06, "loss": 0.0102, "num_input_tokens_seen": 3012016, "step": 455 }, { "epoch": 2.934835076427997, "grad_norm": 4.725823879241943, "learning_rate": 1.8554319316383657e-06, "loss": 0.0419, "num_input_tokens_seen": 3018768, "step": 456 }, { "epoch": 2.941271118262269, "grad_norm": 1.6467297077178955, "learning_rate": 1.8455178565923993e-06, "loss": 0.0109, "num_input_tokens_seen": 3025328, "step": 457 }, { "epoch": 2.9477071600965408, "grad_norm": 1.3065979480743408, "learning_rate": 1.8356148191315753e-06, "loss": 0.0092, "num_input_tokens_seen": 3032080, "step": 458 }, { "epoch": 2.9541432019308127, "grad_norm": 2.6485443115234375, "learning_rate": 1.8257229862667437e-06, "loss": 0.0449, "num_input_tokens_seen": 3038880, "step": 459 }, { "epoch": 2.9605792437650846, "grad_norm": 0.9736925363540649, "learning_rate": 1.8158425248197931e-06, "loss": 0.014, "num_input_tokens_seen": 3045552, "step": 460 }, { "epoch": 2.9670152855993566, "grad_norm": 0.423833429813385, "learning_rate": 1.8059736014208388e-06, "loss": 0.0035, "num_input_tokens_seen": 3052288, "step": 461 }, { "epoch": 2.9734513274336285, "grad_norm": 3.7729272842407227, "learning_rate": 1.7961163825054101e-06, "loss": 0.016, "num_input_tokens_seen": 3058768, "step": 462 }, { "epoch": 2.9798873692679004, "grad_norm": 2.9312222003936768, "learning_rate": 1.7862710343116451e-06, "loss": 0.0151, "num_input_tokens_seen": 3065584, "step": 463 }, { "epoch": 2.9863234111021724, "grad_norm": 0.6318484544754028, "learning_rate": 1.7764377228774877e-06, "loss": 0.0039, "num_input_tokens_seen": 3072368, "step": 464 }, { "epoch": 2.9927594529364443, "grad_norm": 5.504857063293457, "learning_rate": 1.7666166140378853e-06, "loss": 0.0361, "num_input_tokens_seen": 3078864, "step": 465 }, { "epoch": 2.9991954947707162, "grad_norm": 2.98315167427063, "learning_rate": 1.7568078734219934e-06, "loss": 0.0609, "num_input_tokens_seen": 3085664, "step": 466 }, { "epoch": 3.0056315366049877, "grad_norm": 0.24189546704292297, "learning_rate": 1.747011666450384e-06, "loss": 0.0027, "num_input_tokens_seen": 3091568, "step": 467 }, { "epoch": 3.0120675784392597, "grad_norm": 3.122098922729492, "learning_rate": 1.737228158332252e-06, "loss": 0.0097, "num_input_tokens_seen": 3098544, "step": 468 }, { "epoch": 3.0185036202735316, "grad_norm": 2.117048740386963, "learning_rate": 1.7274575140626318e-06, "loss": 0.0091, "num_input_tokens_seen": 3105120, "step": 469 }, { "epoch": 3.0249396621078035, "grad_norm": 0.3818783760070801, "learning_rate": 1.7176998984196148e-06, "loss": 0.0026, "num_input_tokens_seen": 3111552, "step": 470 }, { "epoch": 3.0313757039420755, "grad_norm": 3.4925177097320557, "learning_rate": 1.7079554759615685e-06, "loss": 0.0311, "num_input_tokens_seen": 3118192, "step": 471 }, { "epoch": 3.0378117457763474, "grad_norm": 0.1732572317123413, "learning_rate": 1.6982244110243626e-06, "loss": 0.0014, "num_input_tokens_seen": 3124640, "step": 472 }, { "epoch": 3.0442477876106193, "grad_norm": 1.305844783782959, "learning_rate": 1.6885068677185989e-06, "loss": 0.0185, "num_input_tokens_seen": 3130992, "step": 473 }, { "epoch": 3.0506838294448912, "grad_norm": 0.9071294665336609, "learning_rate": 1.678803009926841e-06, "loss": 0.0075, "num_input_tokens_seen": 3137696, "step": 474 }, { "epoch": 3.057119871279163, "grad_norm": 0.9389513731002808, "learning_rate": 1.6691130013008514e-06, "loss": 0.0069, "num_input_tokens_seen": 3144560, "step": 475 }, { "epoch": 3.063555913113435, "grad_norm": 0.15343110263347626, "learning_rate": 1.6594370052588328e-06, "loss": 0.0009, "num_input_tokens_seen": 3151072, "step": 476 }, { "epoch": 3.069991954947707, "grad_norm": 0.5078912973403931, "learning_rate": 1.6497751849826692e-06, "loss": 0.0015, "num_input_tokens_seen": 3158016, "step": 477 }, { "epoch": 3.076427996781979, "grad_norm": 0.14821191132068634, "learning_rate": 1.6401277034151798e-06, "loss": 0.0007, "num_input_tokens_seen": 3164560, "step": 478 }, { "epoch": 3.082864038616251, "grad_norm": 0.3397853672504425, "learning_rate": 1.630494723257363e-06, "loss": 0.0012, "num_input_tokens_seen": 3171088, "step": 479 }, { "epoch": 3.089300080450523, "grad_norm": 0.25013279914855957, "learning_rate": 1.620876406965658e-06, "loss": 0.0018, "num_input_tokens_seen": 3177952, "step": 480 }, { "epoch": 3.0957361222847948, "grad_norm": 0.04799158126115799, "learning_rate": 1.611272916749205e-06, "loss": 0.0003, "num_input_tokens_seen": 3184592, "step": 481 }, { "epoch": 3.1021721641190667, "grad_norm": 2.0195066928863525, "learning_rate": 1.6016844145671062e-06, "loss": 0.0044, "num_input_tokens_seen": 3190896, "step": 482 }, { "epoch": 3.1086082059533386, "grad_norm": 0.6244819164276123, "learning_rate": 1.5921110621256972e-06, "loss": 0.0019, "num_input_tokens_seen": 3197376, "step": 483 }, { "epoch": 3.1150442477876106, "grad_norm": 2.540050506591797, "learning_rate": 1.58255302087582e-06, "loss": 0.0059, "num_input_tokens_seen": 3203776, "step": 484 }, { "epoch": 3.1214802896218825, "grad_norm": 0.7487736344337463, "learning_rate": 1.5730104520100984e-06, "loss": 0.0036, "num_input_tokens_seen": 3210464, "step": 485 }, { "epoch": 3.1279163314561544, "grad_norm": 0.052535440772771835, "learning_rate": 1.56348351646022e-06, "loss": 0.0002, "num_input_tokens_seen": 3217056, "step": 486 }, { "epoch": 3.1343523732904264, "grad_norm": 2.5393643379211426, "learning_rate": 1.5539723748942246e-06, "loss": 0.0019, "num_input_tokens_seen": 3223840, "step": 487 }, { "epoch": 3.1407884151246983, "grad_norm": 0.28790536522865295, "learning_rate": 1.544477187713791e-06, "loss": 0.0009, "num_input_tokens_seen": 3230592, "step": 488 }, { "epoch": 3.1472244569589702, "grad_norm": 2.5697410106658936, "learning_rate": 1.534998115051533e-06, "loss": 0.0318, "num_input_tokens_seen": 3237216, "step": 489 }, { "epoch": 3.153660498793242, "grad_norm": 1.5203006267547607, "learning_rate": 1.5255353167683017e-06, "loss": 0.0216, "num_input_tokens_seen": 3243920, "step": 490 }, { "epoch": 3.160096540627514, "grad_norm": 0.1484091877937317, "learning_rate": 1.5160889524504857e-06, "loss": 0.0004, "num_input_tokens_seen": 3250656, "step": 491 }, { "epoch": 3.166532582461786, "grad_norm": 3.3526744842529297, "learning_rate": 1.50665918140732e-06, "loss": 0.0286, "num_input_tokens_seen": 3257312, "step": 492 }, { "epoch": 3.172968624296058, "grad_norm": 1.3879235982894897, "learning_rate": 1.4972461626682033e-06, "loss": 0.0254, "num_input_tokens_seen": 3264112, "step": 493 }, { "epoch": 3.17940466613033, "grad_norm": 1.1939952373504639, "learning_rate": 1.4878500549800115e-06, "loss": 0.0039, "num_input_tokens_seen": 3270528, "step": 494 }, { "epoch": 3.185840707964602, "grad_norm": 0.20248474180698395, "learning_rate": 1.4784710168044215e-06, "loss": 0.0005, "num_input_tokens_seen": 3277008, "step": 495 }, { "epoch": 3.1922767497988738, "grad_norm": 1.903956413269043, "learning_rate": 1.4691092063152417e-06, "loss": 0.0196, "num_input_tokens_seen": 3283376, "step": 496 }, { "epoch": 3.1987127916331457, "grad_norm": 0.3746008276939392, "learning_rate": 1.459764781395741e-06, "loss": 0.0015, "num_input_tokens_seen": 3289664, "step": 497 }, { "epoch": 3.2051488334674176, "grad_norm": 4.635190486907959, "learning_rate": 1.4504378996359867e-06, "loss": 0.0088, "num_input_tokens_seen": 3296576, "step": 498 }, { "epoch": 3.2115848753016896, "grad_norm": 1.4451507329940796, "learning_rate": 1.4411287183301902e-06, "loss": 0.0023, "num_input_tokens_seen": 3303120, "step": 499 }, { "epoch": 3.2180209171359615, "grad_norm": 1.252470850944519, "learning_rate": 1.4318373944740485e-06, "loss": 0.0071, "num_input_tokens_seen": 3310384, "step": 500 }, { "epoch": 3.2244569589702334, "grad_norm": 0.6509237289428711, "learning_rate": 1.4225640847621006e-06, "loss": 0.0006, "num_input_tokens_seen": 3316768, "step": 501 }, { "epoch": 3.2308930008045054, "grad_norm": 0.2248382717370987, "learning_rate": 1.4133089455850878e-06, "loss": 0.0011, "num_input_tokens_seen": 3323488, "step": 502 }, { "epoch": 3.2373290426387773, "grad_norm": 1.0306220054626465, "learning_rate": 1.4040721330273063e-06, "loss": 0.0057, "num_input_tokens_seen": 3330000, "step": 503 }, { "epoch": 3.2437650844730492, "grad_norm": 0.1734343320131302, "learning_rate": 1.3948538028639851e-06, "loss": 0.0006, "num_input_tokens_seen": 3336592, "step": 504 }, { "epoch": 3.250201126307321, "grad_norm": 0.9872696399688721, "learning_rate": 1.3856541105586545e-06, "loss": 0.0066, "num_input_tokens_seen": 3343136, "step": 505 }, { "epoch": 3.256637168141593, "grad_norm": 0.8048367500305176, "learning_rate": 1.3764732112605223e-06, "loss": 0.0079, "num_input_tokens_seen": 3349680, "step": 506 }, { "epoch": 3.263073209975865, "grad_norm": 1.8275296688079834, "learning_rate": 1.367311259801863e-06, "loss": 0.0215, "num_input_tokens_seen": 3356304, "step": 507 }, { "epoch": 3.2695092518101365, "grad_norm": 1.90727698802948, "learning_rate": 1.3581684106953987e-06, "loss": 0.0031, "num_input_tokens_seen": 3363008, "step": 508 }, { "epoch": 3.2759452936444085, "grad_norm": 2.614037275314331, "learning_rate": 1.3490448181317025e-06, "loss": 0.0024, "num_input_tokens_seen": 3369728, "step": 509 }, { "epoch": 3.2823813354786804, "grad_norm": 1.9239071607589722, "learning_rate": 1.3399406359765921e-06, "loss": 0.0094, "num_input_tokens_seen": 3375968, "step": 510 }, { "epoch": 3.2888173773129523, "grad_norm": 1.1601731777191162, "learning_rate": 1.3308560177685334e-06, "loss": 0.0054, "num_input_tokens_seen": 3383024, "step": 511 }, { "epoch": 3.2952534191472242, "grad_norm": 0.31424281001091003, "learning_rate": 1.3217911167160575e-06, "loss": 0.0008, "num_input_tokens_seen": 3389488, "step": 512 }, { "epoch": 3.301689460981496, "grad_norm": 2.633910655975342, "learning_rate": 1.3127460856951724e-06, "loss": 0.0053, "num_input_tokens_seen": 3395712, "step": 513 }, { "epoch": 3.308125502815768, "grad_norm": 0.9618326425552368, "learning_rate": 1.303721077246784e-06, "loss": 0.006, "num_input_tokens_seen": 3402384, "step": 514 }, { "epoch": 3.31456154465004, "grad_norm": 0.22136647999286652, "learning_rate": 1.2947162435741278e-06, "loss": 0.0004, "num_input_tokens_seen": 3409136, "step": 515 }, { "epoch": 3.320997586484312, "grad_norm": 1.880077838897705, "learning_rate": 1.2857317365401997e-06, "loss": 0.0135, "num_input_tokens_seen": 3415776, "step": 516 }, { "epoch": 3.327433628318584, "grad_norm": 2.234178304672241, "learning_rate": 1.2767677076651913e-06, "loss": 0.0083, "num_input_tokens_seen": 3422496, "step": 517 }, { "epoch": 3.333869670152856, "grad_norm": 0.21132518351078033, "learning_rate": 1.2678243081239421e-06, "loss": 0.0004, "num_input_tokens_seen": 3429312, "step": 518 }, { "epoch": 3.340305711987128, "grad_norm": 1.0334022045135498, "learning_rate": 1.2589016887433846e-06, "loss": 0.0038, "num_input_tokens_seen": 3435840, "step": 519 }, { "epoch": 3.3467417538213997, "grad_norm": 2.751037359237671, "learning_rate": 1.2500000000000007e-06, "loss": 0.0269, "num_input_tokens_seen": 3442176, "step": 520 }, { "epoch": 3.3531777956556716, "grad_norm": 0.7970973253250122, "learning_rate": 1.2411193920172866e-06, "loss": 0.0062, "num_input_tokens_seen": 3448784, "step": 521 }, { "epoch": 3.3596138374899436, "grad_norm": 0.09952107071876526, "learning_rate": 1.2322600145632204e-06, "loss": 0.0004, "num_input_tokens_seen": 3455184, "step": 522 }, { "epoch": 3.3660498793242155, "grad_norm": 0.6218022108078003, "learning_rate": 1.2234220170477332e-06, "loss": 0.0011, "num_input_tokens_seen": 3461792, "step": 523 }, { "epoch": 3.3724859211584874, "grad_norm": 1.4417766332626343, "learning_rate": 1.2146055485201943e-06, "loss": 0.0026, "num_input_tokens_seen": 3468624, "step": 524 }, { "epoch": 3.3789219629927594, "grad_norm": 2.819247245788574, "learning_rate": 1.205810757666894e-06, "loss": 0.0183, "num_input_tokens_seen": 3474976, "step": 525 }, { "epoch": 3.3853580048270313, "grad_norm": 1.7066518068313599, "learning_rate": 1.1970377928085372e-06, "loss": 0.0079, "num_input_tokens_seen": 3481360, "step": 526 }, { "epoch": 3.3917940466613032, "grad_norm": 2.671914577484131, "learning_rate": 1.188286801897743e-06, "loss": 0.0123, "num_input_tokens_seen": 3487904, "step": 527 }, { "epoch": 3.398230088495575, "grad_norm": 0.37451621890068054, "learning_rate": 1.1795579325165448e-06, "loss": 0.0018, "num_input_tokens_seen": 3494368, "step": 528 }, { "epoch": 3.404666130329847, "grad_norm": 0.3565497398376465, "learning_rate": 1.1708513318739096e-06, "loss": 0.0014, "num_input_tokens_seen": 3500704, "step": 529 }, { "epoch": 3.411102172164119, "grad_norm": 0.22408631443977356, "learning_rate": 1.1621671468032495e-06, "loss": 0.0009, "num_input_tokens_seen": 3507216, "step": 530 }, { "epoch": 3.417538213998391, "grad_norm": 1.3339484930038452, "learning_rate": 1.153505523759944e-06, "loss": 0.0078, "num_input_tokens_seen": 3513664, "step": 531 }, { "epoch": 3.423974255832663, "grad_norm": 0.6494855880737305, "learning_rate": 1.1448666088188766e-06, "loss": 0.0027, "num_input_tokens_seen": 3520096, "step": 532 }, { "epoch": 3.430410297666935, "grad_norm": 0.10890411585569382, "learning_rate": 1.1362505476719662e-06, "loss": 0.0004, "num_input_tokens_seen": 3526560, "step": 533 }, { "epoch": 3.4368463395012068, "grad_norm": 13.174049377441406, "learning_rate": 1.1276574856257097e-06, "loss": 0.0064, "num_input_tokens_seen": 3533536, "step": 534 }, { "epoch": 3.4432823813354787, "grad_norm": 2.0001068115234375, "learning_rate": 1.1190875675987355e-06, "loss": 0.007, "num_input_tokens_seen": 3540288, "step": 535 }, { "epoch": 3.4497184231697506, "grad_norm": 0.15650025010108948, "learning_rate": 1.1105409381193572e-06, "loss": 0.0005, "num_input_tokens_seen": 3546720, "step": 536 }, { "epoch": 3.4561544650040226, "grad_norm": 0.13460475206375122, "learning_rate": 1.1020177413231334e-06, "loss": 0.0004, "num_input_tokens_seen": 3553280, "step": 537 }, { "epoch": 3.4625905068382945, "grad_norm": 2.165956735610962, "learning_rate": 1.0935181209504422e-06, "loss": 0.0294, "num_input_tokens_seen": 3559776, "step": 538 }, { "epoch": 3.4690265486725664, "grad_norm": 0.6856318712234497, "learning_rate": 1.0850422203440555e-06, "loss": 0.0036, "num_input_tokens_seen": 3566848, "step": 539 }, { "epoch": 3.4754625905068384, "grad_norm": 1.306766152381897, "learning_rate": 1.0765901824467167e-06, "loss": 0.0051, "num_input_tokens_seen": 3573280, "step": 540 }, { "epoch": 3.4818986323411103, "grad_norm": 0.3889179825782776, "learning_rate": 1.068162149798737e-06, "loss": 0.0012, "num_input_tokens_seen": 3579712, "step": 541 }, { "epoch": 3.4883346741753822, "grad_norm": 1.5245965719223022, "learning_rate": 1.0597582645355891e-06, "loss": 0.0231, "num_input_tokens_seen": 3586480, "step": 542 }, { "epoch": 3.494770716009654, "grad_norm": 0.6708037257194519, "learning_rate": 1.0513786683855062e-06, "loss": 0.0041, "num_input_tokens_seen": 3593136, "step": 543 }, { "epoch": 3.501206757843926, "grad_norm": 2.0138630867004395, "learning_rate": 1.0430235026670979e-06, "loss": 0.0124, "num_input_tokens_seen": 3599968, "step": 544 }, { "epoch": 3.507642799678198, "grad_norm": 7.274059295654297, "learning_rate": 1.034692908286964e-06, "loss": 0.0171, "num_input_tokens_seen": 3606592, "step": 545 }, { "epoch": 3.51407884151247, "grad_norm": 5.609940052032471, "learning_rate": 1.0263870257373162e-06, "loss": 0.008, "num_input_tokens_seen": 3613072, "step": 546 }, { "epoch": 3.520514883346742, "grad_norm": 1.4191588163375854, "learning_rate": 1.0181059950936131e-06, "loss": 0.0035, "num_input_tokens_seen": 3619696, "step": 547 }, { "epoch": 3.526950925181014, "grad_norm": 0.1580982804298401, "learning_rate": 1.0098499560121943e-06, "loss": 0.0006, "num_input_tokens_seen": 3626240, "step": 548 }, { "epoch": 3.5333869670152858, "grad_norm": 0.637765109539032, "learning_rate": 1.0016190477279274e-06, "loss": 0.002, "num_input_tokens_seen": 3632704, "step": 549 }, { "epoch": 3.5398230088495577, "grad_norm": 0.07971790432929993, "learning_rate": 9.934134090518593e-07, "loss": 0.0003, "num_input_tokens_seen": 3639360, "step": 550 }, { "epoch": 3.5462590506838296, "grad_norm": 0.15312433242797852, "learning_rate": 9.852331783688722e-07, "loss": 0.0004, "num_input_tokens_seen": 3646112, "step": 551 }, { "epoch": 3.5526950925181016, "grad_norm": 0.5353730916976929, "learning_rate": 9.770784936353555e-07, "loss": 0.0016, "num_input_tokens_seen": 3652704, "step": 552 }, { "epoch": 3.5591311343523735, "grad_norm": 0.3197666108608246, "learning_rate": 9.689494923768756e-07, "loss": 0.0012, "num_input_tokens_seen": 3659696, "step": 553 }, { "epoch": 3.5655671761866454, "grad_norm": 1.4529962539672852, "learning_rate": 9.608463116858544e-07, "loss": 0.0057, "num_input_tokens_seen": 3666288, "step": 554 }, { "epoch": 3.5720032180209174, "grad_norm": 2.7501587867736816, "learning_rate": 9.527690882192636e-07, "loss": 0.0168, "num_input_tokens_seen": 3673104, "step": 555 }, { "epoch": 3.5784392598551893, "grad_norm": 0.21036742627620697, "learning_rate": 9.447179581963156e-07, "loss": 0.0012, "num_input_tokens_seen": 3679872, "step": 556 }, { "epoch": 3.5848753016894612, "grad_norm": 0.03335335850715637, "learning_rate": 9.366930573961649e-07, "loss": 0.0002, "num_input_tokens_seen": 3686288, "step": 557 }, { "epoch": 3.591311343523733, "grad_norm": 1.3189131021499634, "learning_rate": 9.286945211556231e-07, "loss": 0.0025, "num_input_tokens_seen": 3692976, "step": 558 }, { "epoch": 3.597747385358005, "grad_norm": 1.1787400245666504, "learning_rate": 9.207224843668733e-07, "loss": 0.0194, "num_input_tokens_seen": 3699312, "step": 559 }, { "epoch": 3.604183427192277, "grad_norm": 0.9992094039916992, "learning_rate": 9.127770814751933e-07, "loss": 0.0055, "num_input_tokens_seen": 3705888, "step": 560 }, { "epoch": 3.6106194690265485, "grad_norm": 2.264843702316284, "learning_rate": 9.048584464766938e-07, "loss": 0.0215, "num_input_tokens_seen": 3712688, "step": 561 }, { "epoch": 3.6170555108608204, "grad_norm": 0.06527237594127655, "learning_rate": 8.969667129160547e-07, "loss": 0.0003, "num_input_tokens_seen": 3719168, "step": 562 }, { "epoch": 3.6234915526950924, "grad_norm": 3.79392409324646, "learning_rate": 8.891020138842718e-07, "loss": 0.0242, "num_input_tokens_seen": 3726048, "step": 563 }, { "epoch": 3.6299275945293643, "grad_norm": 0.9232211112976074, "learning_rate": 8.81264482016416e-07, "loss": 0.0206, "num_input_tokens_seen": 3732672, "step": 564 }, { "epoch": 3.6363636363636362, "grad_norm": 0.5276843309402466, "learning_rate": 8.734542494893955e-07, "loss": 0.0024, "num_input_tokens_seen": 3739456, "step": 565 }, { "epoch": 3.642799678197908, "grad_norm": 1.1676807403564453, "learning_rate": 8.65671448019722e-07, "loss": 0.0087, "num_input_tokens_seen": 3746160, "step": 566 }, { "epoch": 3.64923572003218, "grad_norm": 1.3703765869140625, "learning_rate": 8.579162088612974e-07, "loss": 0.0089, "num_input_tokens_seen": 3752560, "step": 567 }, { "epoch": 3.655671761866452, "grad_norm": 0.06538532674312592, "learning_rate": 8.501886628031941e-07, "loss": 0.0003, "num_input_tokens_seen": 3759600, "step": 568 }, { "epoch": 3.662107803700724, "grad_norm": 0.0386020764708519, "learning_rate": 8.424889401674505e-07, "loss": 0.0002, "num_input_tokens_seen": 3766096, "step": 569 }, { "epoch": 3.668543845534996, "grad_norm": 0.20554865896701813, "learning_rate": 8.348171708068748e-07, "loss": 0.0009, "num_input_tokens_seen": 3772944, "step": 570 }, { "epoch": 3.674979887369268, "grad_norm": 0.9973205327987671, "learning_rate": 8.271734841028553e-07, "loss": 0.0154, "num_input_tokens_seen": 3779664, "step": 571 }, { "epoch": 3.6814159292035398, "grad_norm": 0.30160781741142273, "learning_rate": 8.195580089631733e-07, "loss": 0.0012, "num_input_tokens_seen": 3786080, "step": 572 }, { "epoch": 3.6878519710378117, "grad_norm": 0.49049124121665955, "learning_rate": 8.119708738198395e-07, "loss": 0.0008, "num_input_tokens_seen": 3792768, "step": 573 }, { "epoch": 3.6942880128720836, "grad_norm": 1.6590077877044678, "learning_rate": 8.04412206626915e-07, "loss": 0.0081, "num_input_tokens_seen": 3799472, "step": 574 }, { "epoch": 3.7007240547063556, "grad_norm": 1.814943552017212, "learning_rate": 7.968821348583644e-07, "loss": 0.008, "num_input_tokens_seen": 3805984, "step": 575 }, { "epoch": 3.7071600965406275, "grad_norm": 1.6639471054077148, "learning_rate": 7.89380785505901e-07, "loss": 0.0073, "num_input_tokens_seen": 3813088, "step": 576 }, { "epoch": 3.7135961383748994, "grad_norm": 0.946050763130188, "learning_rate": 7.819082850768433e-07, "loss": 0.0062, "num_input_tokens_seen": 3820208, "step": 577 }, { "epoch": 3.7200321802091714, "grad_norm": 0.2189425230026245, "learning_rate": 7.744647595919869e-07, "loss": 0.0015, "num_input_tokens_seen": 3826800, "step": 578 }, { "epoch": 3.7264682220434433, "grad_norm": 2.796231985092163, "learning_rate": 7.670503345834757e-07, "loss": 0.0268, "num_input_tokens_seen": 3833344, "step": 579 }, { "epoch": 3.7329042638777152, "grad_norm": 0.13711552321910858, "learning_rate": 7.596651350926837e-07, "loss": 0.0007, "num_input_tokens_seen": 3839920, "step": 580 }, { "epoch": 3.739340305711987, "grad_norm": 0.3616367280483246, "learning_rate": 7.523092856681099e-07, "loss": 0.0016, "num_input_tokens_seen": 3846432, "step": 581 }, { "epoch": 3.745776347546259, "grad_norm": 2.3357245922088623, "learning_rate": 7.44982910363276e-07, "loss": 0.0631, "num_input_tokens_seen": 3853216, "step": 582 }, { "epoch": 3.752212389380531, "grad_norm": 1.7805283069610596, "learning_rate": 7.376861327346325e-07, "loss": 0.0128, "num_input_tokens_seen": 3859664, "step": 583 }, { "epoch": 3.758648431214803, "grad_norm": 0.5933414101600647, "learning_rate": 7.304190758394775e-07, "loss": 0.0034, "num_input_tokens_seen": 3866208, "step": 584 }, { "epoch": 3.765084473049075, "grad_norm": 3.1310431957244873, "learning_rate": 7.231818622338824e-07, "loss": 0.0633, "num_input_tokens_seen": 3872736, "step": 585 }, { "epoch": 3.771520514883347, "grad_norm": 0.2022082656621933, "learning_rate": 7.159746139706194e-07, "loss": 0.0007, "num_input_tokens_seen": 3879264, "step": 586 }, { "epoch": 3.7779565567176188, "grad_norm": 6.451120376586914, "learning_rate": 7.087974525971103e-07, "loss": 0.0211, "num_input_tokens_seen": 3885744, "step": 587 }, { "epoch": 3.7843925985518907, "grad_norm": 0.8931072354316711, "learning_rate": 7.016504991533727e-07, "loss": 0.009, "num_input_tokens_seen": 3892304, "step": 588 }, { "epoch": 3.7908286403861626, "grad_norm": 1.4347479343414307, "learning_rate": 6.94533874169977e-07, "loss": 0.0152, "num_input_tokens_seen": 3898768, "step": 589 }, { "epoch": 3.7972646822204346, "grad_norm": 0.5323463678359985, "learning_rate": 6.874476976660185e-07, "loss": 0.0022, "num_input_tokens_seen": 3904976, "step": 590 }, { "epoch": 3.8037007240547065, "grad_norm": 1.120011806488037, "learning_rate": 6.803920891470905e-07, "loss": 0.014, "num_input_tokens_seen": 3911360, "step": 591 }, { "epoch": 3.8101367658889784, "grad_norm": 0.6292040348052979, "learning_rate": 6.733671676032674e-07, "loss": 0.0085, "num_input_tokens_seen": 3918224, "step": 592 }, { "epoch": 3.8165728077232504, "grad_norm": 3.3647360801696777, "learning_rate": 6.663730515071019e-07, "loss": 0.0161, "num_input_tokens_seen": 3924960, "step": 593 }, { "epoch": 3.823008849557522, "grad_norm": 1.8465656042099, "learning_rate": 6.594098588116243e-07, "loss": 0.0234, "num_input_tokens_seen": 3931712, "step": 594 }, { "epoch": 3.829444891391794, "grad_norm": 1.0739251375198364, "learning_rate": 6.524777069483526e-07, "loss": 0.0186, "num_input_tokens_seen": 3938304, "step": 595 }, { "epoch": 3.8358809332260657, "grad_norm": 3.146777629852295, "learning_rate": 6.455767128253148e-07, "loss": 0.0199, "num_input_tokens_seen": 3945200, "step": 596 }, { "epoch": 3.8423169750603376, "grad_norm": 1.1694271564483643, "learning_rate": 6.38706992825075e-07, "loss": 0.0052, "num_input_tokens_seen": 3951808, "step": 597 }, { "epoch": 3.8487530168946096, "grad_norm": 1.1547743082046509, "learning_rate": 6.318686628027723e-07, "loss": 0.0165, "num_input_tokens_seen": 3958480, "step": 598 }, { "epoch": 3.8551890587288815, "grad_norm": 1.1595410108566284, "learning_rate": 6.250618380841661e-07, "loss": 0.01, "num_input_tokens_seen": 3965072, "step": 599 }, { "epoch": 3.8616251005631534, "grad_norm": 0.8452915549278259, "learning_rate": 6.182866334636889e-07, "loss": 0.0047, "num_input_tokens_seen": 3971808, "step": 600 }, { "epoch": 3.8680611423974254, "grad_norm": 2.201892375946045, "learning_rate": 6.115431632025154e-07, "loss": 0.0039, "num_input_tokens_seen": 3978480, "step": 601 }, { "epoch": 3.8744971842316973, "grad_norm": 0.24013373255729675, "learning_rate": 6.048315410266326e-07, "loss": 0.0022, "num_input_tokens_seen": 3985216, "step": 602 }, { "epoch": 3.8809332260659692, "grad_norm": 0.442757248878479, "learning_rate": 5.981518801249192e-07, "loss": 0.0042, "num_input_tokens_seen": 3991792, "step": 603 }, { "epoch": 3.887369267900241, "grad_norm": 2.5312795639038086, "learning_rate": 5.915042931472426e-07, "loss": 0.0076, "num_input_tokens_seen": 3998224, "step": 604 }, { "epoch": 3.893805309734513, "grad_norm": 0.3599741756916046, "learning_rate": 5.848888922025553e-07, "loss": 0.0019, "num_input_tokens_seen": 4004960, "step": 605 }, { "epoch": 3.900241351568785, "grad_norm": 0.33045250177383423, "learning_rate": 5.783057888570034e-07, "loss": 0.0014, "num_input_tokens_seen": 4011984, "step": 606 }, { "epoch": 3.906677393403057, "grad_norm": 0.540598452091217, "learning_rate": 5.717550941320482e-07, "loss": 0.0022, "num_input_tokens_seen": 4018912, "step": 607 }, { "epoch": 3.913113435237329, "grad_norm": 0.4901201128959656, "learning_rate": 5.65236918502593e-07, "loss": 0.0024, "num_input_tokens_seen": 4025504, "step": 608 }, { "epoch": 3.919549477071601, "grad_norm": 0.23451536893844604, "learning_rate": 5.587513718951165e-07, "loss": 0.0013, "num_input_tokens_seen": 4031776, "step": 609 }, { "epoch": 3.9259855189058728, "grad_norm": 0.9038437604904175, "learning_rate": 5.522985636858238e-07, "loss": 0.0064, "num_input_tokens_seen": 4038208, "step": 610 }, { "epoch": 3.9324215607401447, "grad_norm": 1.4877148866653442, "learning_rate": 5.458786026988005e-07, "loss": 0.0084, "num_input_tokens_seen": 4044928, "step": 611 }, { "epoch": 3.9388576025744166, "grad_norm": 0.12848466634750366, "learning_rate": 5.394915972041739e-07, "loss": 0.0009, "num_input_tokens_seen": 4051552, "step": 612 }, { "epoch": 3.9452936444086886, "grad_norm": 0.22914128005504608, "learning_rate": 5.33137654916292e-07, "loss": 0.001, "num_input_tokens_seen": 4058304, "step": 613 }, { "epoch": 3.9517296862429605, "grad_norm": 0.7593125700950623, "learning_rate": 5.268168829919046e-07, "loss": 0.0064, "num_input_tokens_seen": 4064720, "step": 614 }, { "epoch": 3.9581657280772324, "grad_norm": 0.6085631251335144, "learning_rate": 5.205293880283552e-07, "loss": 0.0033, "num_input_tokens_seen": 4071216, "step": 615 }, { "epoch": 3.9646017699115044, "grad_norm": 0.9351167678833008, "learning_rate": 5.14275276061785e-07, "loss": 0.0065, "num_input_tokens_seen": 4077904, "step": 616 }, { "epoch": 3.9710378117457763, "grad_norm": 2.1718461513519287, "learning_rate": 5.080546525653448e-07, "loss": 0.0272, "num_input_tokens_seen": 4084656, "step": 617 }, { "epoch": 3.9774738535800482, "grad_norm": 0.39000532031059265, "learning_rate": 5.018676224474139e-07, "loss": 0.0015, "num_input_tokens_seen": 4091584, "step": 618 }, { "epoch": 3.98390989541432, "grad_norm": 0.5723803639411926, "learning_rate": 4.957142900498335e-07, "loss": 0.0013, "num_input_tokens_seen": 4098768, "step": 619 }, { "epoch": 3.990345937248592, "grad_norm": 1.5266039371490479, "learning_rate": 4.895947591461456e-07, "loss": 0.0148, "num_input_tokens_seen": 4105312, "step": 620 }, { "epoch": 3.996781979082864, "grad_norm": 0.7928001880645752, "learning_rate": 4.835091329398436e-07, "loss": 0.0063, "num_input_tokens_seen": 4112000, "step": 621 }, { "epoch": 4.003218020917136, "grad_norm": 0.09017051756381989, "learning_rate": 4.774575140626317e-07, "loss": 0.0003, "num_input_tokens_seen": 4118624, "step": 622 }, { "epoch": 4.009654062751408, "grad_norm": 0.2493676394224167, "learning_rate": 4.714400045726919e-07, "loss": 0.001, "num_input_tokens_seen": 4125408, "step": 623 }, { "epoch": 4.01609010458568, "grad_norm": 0.03381378576159477, "learning_rate": 4.6545670595296686e-07, "loss": 0.0001, "num_input_tokens_seen": 4131936, "step": 624 }, { "epoch": 4.022526146419952, "grad_norm": 2.845327854156494, "learning_rate": 4.5950771910944603e-07, "loss": 0.0193, "num_input_tokens_seen": 4138352, "step": 625 }, { "epoch": 4.028962188254224, "grad_norm": 0.6973279714584351, "learning_rate": 4.5359314436946275e-07, "loss": 0.0049, "num_input_tokens_seen": 4144672, "step": 626 }, { "epoch": 4.035398230088496, "grad_norm": 0.3552819788455963, "learning_rate": 4.4771308148000487e-07, "loss": 0.0015, "num_input_tokens_seen": 4151296, "step": 627 }, { "epoch": 4.041834271922768, "grad_norm": 0.2976234555244446, "learning_rate": 4.418676296060323e-07, "loss": 0.0019, "num_input_tokens_seen": 4157696, "step": 628 }, { "epoch": 4.0482703137570395, "grad_norm": 0.6438854932785034, "learning_rate": 4.3605688732880097e-07, "loss": 0.0034, "num_input_tokens_seen": 4164352, "step": 629 }, { "epoch": 4.054706355591311, "grad_norm": 0.055070556700229645, "learning_rate": 4.302809526442053e-07, "loss": 0.0003, "num_input_tokens_seen": 4170992, "step": 630 }, { "epoch": 4.061142397425583, "grad_norm": 0.5393857359886169, "learning_rate": 4.2453992296112384e-07, "loss": 0.0031, "num_input_tokens_seen": 4177888, "step": 631 }, { "epoch": 4.067578439259855, "grad_norm": 0.10041255503892899, "learning_rate": 4.188338950997728e-07, "loss": 0.0004, "num_input_tokens_seen": 4184800, "step": 632 }, { "epoch": 4.074014481094127, "grad_norm": 0.4824787378311157, "learning_rate": 4.1316296529007955e-07, "loss": 0.0027, "num_input_tokens_seen": 4191136, "step": 633 }, { "epoch": 4.080450522928399, "grad_norm": 0.8842573761940002, "learning_rate": 4.075272291700558e-07, "loss": 0.0047, "num_input_tokens_seen": 4197984, "step": 634 }, { "epoch": 4.086886564762671, "grad_norm": 0.0672411248087883, "learning_rate": 4.019267817841835e-07, "loss": 0.0003, "num_input_tokens_seen": 4204688, "step": 635 }, { "epoch": 4.093322606596943, "grad_norm": 1.144921898841858, "learning_rate": 3.9636171758181657e-07, "loss": 0.0204, "num_input_tokens_seen": 4211360, "step": 636 }, { "epoch": 4.099758648431215, "grad_norm": 1.0628600120544434, "learning_rate": 3.908321304155846e-07, "loss": 0.0043, "num_input_tokens_seen": 4218000, "step": 637 }, { "epoch": 4.106194690265487, "grad_norm": 0.03438463807106018, "learning_rate": 3.853381135398093e-07, "loss": 0.0002, "num_input_tokens_seen": 4224544, "step": 638 }, { "epoch": 4.112630732099759, "grad_norm": 0.22854630649089813, "learning_rate": 3.798797596089351e-07, "loss": 0.0009, "num_input_tokens_seen": 4230992, "step": 639 }, { "epoch": 4.119066773934031, "grad_norm": 0.12790539860725403, "learning_rate": 3.7445716067596506e-07, "loss": 0.0004, "num_input_tokens_seen": 4237808, "step": 640 }, { "epoch": 4.125502815768303, "grad_norm": 0.040783047676086426, "learning_rate": 3.6907040819090604e-07, "loss": 0.0003, "num_input_tokens_seen": 4244032, "step": 641 }, { "epoch": 4.131938857602575, "grad_norm": 0.29912275075912476, "learning_rate": 3.63719592999231e-07, "loss": 0.0015, "num_input_tokens_seen": 4250640, "step": 642 }, { "epoch": 4.1383748994368466, "grad_norm": 0.1869562268257141, "learning_rate": 3.5840480534034355e-07, "loss": 0.0012, "num_input_tokens_seen": 4257440, "step": 643 }, { "epoch": 4.1448109412711185, "grad_norm": 0.6986035108566284, "learning_rate": 3.5312613484605546e-07, "loss": 0.0041, "num_input_tokens_seen": 4263936, "step": 644 }, { "epoch": 4.15124698310539, "grad_norm": 1.672957181930542, "learning_rate": 3.4788367053908087e-07, "loss": 0.0087, "num_input_tokens_seen": 4270144, "step": 645 }, { "epoch": 4.157683024939662, "grad_norm": 0.5007069110870361, "learning_rate": 3.4267750083152587e-07, "loss": 0.0025, "num_input_tokens_seen": 4276944, "step": 646 }, { "epoch": 4.164119066773934, "grad_norm": 0.2200661450624466, "learning_rate": 3.375077135234051e-07, "loss": 0.0012, "num_input_tokens_seen": 4283488, "step": 647 }, { "epoch": 4.170555108608206, "grad_norm": 1.018943190574646, "learning_rate": 3.323743958011588e-07, "loss": 0.0111, "num_input_tokens_seen": 4290000, "step": 648 }, { "epoch": 4.176991150442478, "grad_norm": 0.5488151907920837, "learning_rate": 3.2727763423617915e-07, "loss": 0.0026, "num_input_tokens_seen": 4296544, "step": 649 }, { "epoch": 4.18342719227675, "grad_norm": 0.12664268910884857, "learning_rate": 3.222175147833556e-07, "loss": 0.0004, "num_input_tokens_seen": 4303056, "step": 650 }, { "epoch": 4.189863234111022, "grad_norm": 0.04309312626719475, "learning_rate": 3.171941227796227e-07, "loss": 0.0002, "num_input_tokens_seen": 4309664, "step": 651 }, { "epoch": 4.196299275945294, "grad_norm": 2.9486300945281982, "learning_rate": 3.122075429425184e-07, "loss": 0.0118, "num_input_tokens_seen": 4316112, "step": 652 }, { "epoch": 4.202735317779566, "grad_norm": 0.03176088631153107, "learning_rate": 3.072578593687606e-07, "loss": 0.0001, "num_input_tokens_seen": 4322800, "step": 653 }, { "epoch": 4.209171359613838, "grad_norm": 4.464654445648193, "learning_rate": 3.0234515553282523e-07, "loss": 0.0151, "num_input_tokens_seen": 4329408, "step": 654 }, { "epoch": 4.21560740144811, "grad_norm": 0.017552955076098442, "learning_rate": 2.9746951428553884e-07, "loss": 0.0001, "num_input_tokens_seen": 4335648, "step": 655 }, { "epoch": 4.222043443282382, "grad_norm": 0.385110467672348, "learning_rate": 2.9263101785268253e-07, "loss": 0.0019, "num_input_tokens_seen": 4342256, "step": 656 }, { "epoch": 4.228479485116654, "grad_norm": 0.3891147971153259, "learning_rate": 2.8782974783360534e-07, "loss": 0.0009, "num_input_tokens_seen": 4349280, "step": 657 }, { "epoch": 4.2349155269509255, "grad_norm": 0.687170147895813, "learning_rate": 2.8306578519984526e-07, "loss": 0.0051, "num_input_tokens_seen": 4356128, "step": 658 }, { "epoch": 4.2413515687851975, "grad_norm": 0.16641825437545776, "learning_rate": 2.783392102937682e-07, "loss": 0.0008, "num_input_tokens_seen": 4362672, "step": 659 }, { "epoch": 4.247787610619469, "grad_norm": 0.02807171456515789, "learning_rate": 2.7365010282720954e-07, "loss": 0.0001, "num_input_tokens_seen": 4369440, "step": 660 }, { "epoch": 4.254223652453741, "grad_norm": 1.0298210382461548, "learning_rate": 2.6899854188013054e-07, "loss": 0.0043, "num_input_tokens_seen": 4375760, "step": 661 }, { "epoch": 4.260659694288013, "grad_norm": 0.15670017898082733, "learning_rate": 2.643846058992866e-07, "loss": 0.0004, "num_input_tokens_seen": 4382768, "step": 662 }, { "epoch": 4.267095736122285, "grad_norm": 1.2815680503845215, "learning_rate": 2.5980837269690056e-07, "loss": 0.0092, "num_input_tokens_seen": 4389424, "step": 663 }, { "epoch": 4.273531777956556, "grad_norm": 0.23917140066623688, "learning_rate": 2.552699194493549e-07, "loss": 0.0006, "num_input_tokens_seen": 4395904, "step": 664 }, { "epoch": 4.279967819790828, "grad_norm": 0.8005861043930054, "learning_rate": 2.507693226958871e-07, "loss": 0.0049, "num_input_tokens_seen": 4402144, "step": 665 }, { "epoch": 4.2864038616251, "grad_norm": 0.8631348609924316, "learning_rate": 2.463066583372989e-07, "loss": 0.0058, "num_input_tokens_seen": 4408672, "step": 666 }, { "epoch": 4.292839903459372, "grad_norm": 0.017498647794127464, "learning_rate": 2.418820016346779e-07, "loss": 0.0001, "num_input_tokens_seen": 4415040, "step": 667 }, { "epoch": 4.299275945293644, "grad_norm": 0.8143237829208374, "learning_rate": 2.3749542720812757e-07, "loss": 0.0063, "num_input_tokens_seen": 4421696, "step": 668 }, { "epoch": 4.305711987127916, "grad_norm": 0.5140169262886047, "learning_rate": 2.331470090355084e-07, "loss": 0.0116, "num_input_tokens_seen": 4428096, "step": 669 }, { "epoch": 4.312148028962188, "grad_norm": 0.6097451448440552, "learning_rate": 2.2883682045119066e-07, "loss": 0.003, "num_input_tokens_seen": 4434752, "step": 670 }, { "epoch": 4.31858407079646, "grad_norm": 0.030739160254597664, "learning_rate": 2.2456493414481778e-07, "loss": 0.0001, "num_input_tokens_seen": 4441584, "step": 671 }, { "epoch": 4.325020112630732, "grad_norm": 0.017091860994696617, "learning_rate": 2.2033142216007913e-07, "loss": 0.0001, "num_input_tokens_seen": 4448464, "step": 672 }, { "epoch": 4.331456154465004, "grad_norm": 0.18128401041030884, "learning_rate": 2.1613635589349756e-07, "loss": 0.0008, "num_input_tokens_seen": 4454976, "step": 673 }, { "epoch": 4.337892196299276, "grad_norm": 0.0346699096262455, "learning_rate": 2.1197980609322406e-07, "loss": 0.0002, "num_input_tokens_seen": 4461440, "step": 674 }, { "epoch": 4.3443282381335475, "grad_norm": 0.018729638308286667, "learning_rate": 2.07861842857843e-07, "loss": 0.0001, "num_input_tokens_seen": 4468080, "step": 675 }, { "epoch": 4.3507642799678194, "grad_norm": 1.5200186967849731, "learning_rate": 2.0378253563519247e-07, "loss": 0.0105, "num_input_tokens_seen": 4474944, "step": 676 }, { "epoch": 4.357200321802091, "grad_norm": 0.03133641555905342, "learning_rate": 1.997419532211925e-07, "loss": 0.0001, "num_input_tokens_seen": 4481456, "step": 677 }, { "epoch": 4.363636363636363, "grad_norm": 0.012541470117866993, "learning_rate": 1.9574016375868282e-07, "loss": 0.0001, "num_input_tokens_seen": 4487472, "step": 678 }, { "epoch": 4.370072405470635, "grad_norm": 1.6271870136260986, "learning_rate": 1.9177723473627647e-07, "loss": 0.0076, "num_input_tokens_seen": 4494320, "step": 679 }, { "epoch": 4.376508447304907, "grad_norm": 0.03906352072954178, "learning_rate": 1.8785323298722098e-07, "loss": 0.0001, "num_input_tokens_seen": 4501152, "step": 680 }, { "epoch": 4.382944489139179, "grad_norm": 0.024355776607990265, "learning_rate": 1.839682246882682e-07, "loss": 0.0001, "num_input_tokens_seen": 4507376, "step": 681 }, { "epoch": 4.389380530973451, "grad_norm": 0.4699815511703491, "learning_rate": 1.801222753585638e-07, "loss": 0.002, "num_input_tokens_seen": 4513904, "step": 682 }, { "epoch": 4.395816572807723, "grad_norm": 1.0409318208694458, "learning_rate": 1.7631544985853623e-07, "loss": 0.0088, "num_input_tokens_seen": 4520608, "step": 683 }, { "epoch": 4.402252614641995, "grad_norm": 0.42572081089019775, "learning_rate": 1.725478123888083e-07, "loss": 0.0023, "num_input_tokens_seen": 4527184, "step": 684 }, { "epoch": 4.408688656476267, "grad_norm": 0.04756924882531166, "learning_rate": 1.6881942648911077e-07, "loss": 0.0002, "num_input_tokens_seen": 4533696, "step": 685 }, { "epoch": 4.415124698310539, "grad_norm": 0.21432961523532867, "learning_rate": 1.6513035503721213e-07, "loss": 0.0017, "num_input_tokens_seen": 4540624, "step": 686 }, { "epoch": 4.421560740144811, "grad_norm": 0.035157278180122375, "learning_rate": 1.614806602478583e-07, "loss": 0.0002, "num_input_tokens_seen": 4547056, "step": 687 }, { "epoch": 4.427996781979083, "grad_norm": 1.0397815704345703, "learning_rate": 1.5787040367172379e-07, "loss": 0.0073, "num_input_tokens_seen": 4553712, "step": 688 }, { "epoch": 4.434432823813355, "grad_norm": 0.8960546851158142, "learning_rate": 1.542996461943716e-07, "loss": 0.0013, "num_input_tokens_seen": 4560080, "step": 689 }, { "epoch": 4.4408688656476265, "grad_norm": 0.1009814515709877, "learning_rate": 1.507684480352292e-07, "loss": 0.0003, "num_input_tokens_seen": 4566496, "step": 690 }, { "epoch": 4.447304907481898, "grad_norm": 1.6075918674468994, "learning_rate": 1.4727686874657143e-07, "loss": 0.0149, "num_input_tokens_seen": 4573152, "step": 691 }, { "epoch": 4.45374094931617, "grad_norm": 0.5051795840263367, "learning_rate": 1.4382496721251526e-07, "loss": 0.0026, "num_input_tokens_seen": 4580432, "step": 692 }, { "epoch": 4.460176991150442, "grad_norm": 0.01903243362903595, "learning_rate": 1.4041280164802967e-07, "loss": 0.0001, "num_input_tokens_seen": 4587024, "step": 693 }, { "epoch": 4.466613032984714, "grad_norm": 0.04848824068903923, "learning_rate": 1.3704042959795132e-07, "loss": 0.0002, "num_input_tokens_seen": 4593296, "step": 694 }, { "epoch": 4.473049074818986, "grad_norm": 1.155561923980713, "learning_rate": 1.3370790793601373e-07, "loss": 0.006, "num_input_tokens_seen": 4600000, "step": 695 }, { "epoch": 4.479485116653258, "grad_norm": 0.06723422557115555, "learning_rate": 1.3041529286389078e-07, "loss": 0.0004, "num_input_tokens_seen": 4606560, "step": 696 }, { "epoch": 4.48592115848753, "grad_norm": 0.10645350813865662, "learning_rate": 1.2716263991024712e-07, "loss": 0.0004, "num_input_tokens_seen": 4613424, "step": 697 }, { "epoch": 4.492357200321802, "grad_norm": 0.09232950955629349, "learning_rate": 1.2395000392980057e-07, "loss": 0.0007, "num_input_tokens_seen": 4620064, "step": 698 }, { "epoch": 4.498793242156074, "grad_norm": 0.04445146396756172, "learning_rate": 1.2077743910239998e-07, "loss": 0.0002, "num_input_tokens_seen": 4626608, "step": 699 }, { "epoch": 4.505229283990346, "grad_norm": 0.03232429176568985, "learning_rate": 1.1764499893210879e-07, "loss": 0.0001, "num_input_tokens_seen": 4633280, "step": 700 }, { "epoch": 4.511665325824618, "grad_norm": 0.7786117792129517, "learning_rate": 1.145527362463042e-07, "loss": 0.0033, "num_input_tokens_seen": 4639920, "step": 701 }, { "epoch": 4.51810136765889, "grad_norm": 0.15632830560207367, "learning_rate": 1.1150070319478679e-07, "loss": 0.0009, "num_input_tokens_seen": 4646736, "step": 702 }, { "epoch": 4.524537409493162, "grad_norm": 2.964639186859131, "learning_rate": 1.0848895124889819e-07, "loss": 0.0216, "num_input_tokens_seen": 4653328, "step": 703 }, { "epoch": 4.530973451327434, "grad_norm": 0.8460017442703247, "learning_rate": 1.0551753120065621e-07, "loss": 0.0035, "num_input_tokens_seen": 4660112, "step": 704 }, { "epoch": 4.5374094931617055, "grad_norm": 1.1353970766067505, "learning_rate": 1.0258649316189722e-07, "loss": 0.0073, "num_input_tokens_seen": 4666560, "step": 705 }, { "epoch": 4.543845534995977, "grad_norm": 0.08409194648265839, "learning_rate": 9.969588656342982e-08, "loss": 0.0003, "num_input_tokens_seen": 4673152, "step": 706 }, { "epoch": 4.550281576830249, "grad_norm": 0.1840662956237793, "learning_rate": 9.684576015420277e-08, "loss": 0.0007, "num_input_tokens_seen": 4679360, "step": 707 }, { "epoch": 4.556717618664521, "grad_norm": 0.049431972205638885, "learning_rate": 9.403616200048288e-08, "loss": 0.0003, "num_input_tokens_seen": 4685904, "step": 708 }, { "epoch": 4.563153660498793, "grad_norm": 1.309277892112732, "learning_rate": 9.12671394850423e-08, "loss": 0.0168, "num_input_tokens_seen": 4692320, "step": 709 }, { "epoch": 4.569589702333065, "grad_norm": 0.3398638367652893, "learning_rate": 8.85387393063622e-08, "loss": 0.0015, "num_input_tokens_seen": 4698928, "step": 710 }, { "epoch": 4.576025744167337, "grad_norm": 0.016252102330327034, "learning_rate": 8.585100747784376e-08, "loss": 0.0001, "num_input_tokens_seen": 4706000, "step": 711 }, { "epoch": 4.582461786001609, "grad_norm": 0.6447362303733826, "learning_rate": 8.320398932703145e-08, "loss": 0.0036, "num_input_tokens_seen": 4712352, "step": 712 }, { "epoch": 4.588897827835881, "grad_norm": 0.5575013756752014, "learning_rate": 8.059772949485068e-08, "loss": 0.0022, "num_input_tokens_seen": 4718848, "step": 713 }, { "epoch": 4.595333869670153, "grad_norm": 1.0512995719909668, "learning_rate": 7.803227193485336e-08, "loss": 0.0085, "num_input_tokens_seen": 4725728, "step": 714 }, { "epoch": 4.601769911504425, "grad_norm": 0.012490477412939072, "learning_rate": 7.550765991247655e-08, "loss": 0.0001, "num_input_tokens_seen": 4732352, "step": 715 }, { "epoch": 4.608205953338697, "grad_norm": 0.025810543447732925, "learning_rate": 7.30239360043139e-08, "loss": 0.0001, "num_input_tokens_seen": 4738992, "step": 716 }, { "epoch": 4.614641995172969, "grad_norm": 0.5250550508499146, "learning_rate": 7.058114209739675e-08, "loss": 0.007, "num_input_tokens_seen": 4745872, "step": 717 }, { "epoch": 4.621078037007241, "grad_norm": 0.05289880558848381, "learning_rate": 6.817931938848805e-08, "loss": 0.0002, "num_input_tokens_seen": 4752544, "step": 718 }, { "epoch": 4.627514078841513, "grad_norm": 0.0455067902803421, "learning_rate": 6.581850838338816e-08, "loss": 0.0002, "num_input_tokens_seen": 4759360, "step": 719 }, { "epoch": 4.6339501206757845, "grad_norm": 0.1569470316171646, "learning_rate": 6.349874889624963e-08, "loss": 0.0008, "num_input_tokens_seen": 4766016, "step": 720 }, { "epoch": 4.640386162510056, "grad_norm": 0.7371820211410522, "learning_rate": 6.12200800489085e-08, "loss": 0.0043, "num_input_tokens_seen": 4772624, "step": 721 }, { "epoch": 4.646822204344328, "grad_norm": 0.09805099666118622, "learning_rate": 5.898254027022293e-08, "loss": 0.0004, "num_input_tokens_seen": 4778960, "step": 722 }, { "epoch": 4.6532582461786, "grad_norm": 0.6390008926391602, "learning_rate": 5.678616729542535e-08, "loss": 0.0024, "num_input_tokens_seen": 4785600, "step": 723 }, { "epoch": 4.659694288012872, "grad_norm": 0.4965854585170746, "learning_rate": 5.463099816548578e-08, "loss": 0.0024, "num_input_tokens_seen": 4792208, "step": 724 }, { "epoch": 4.666130329847144, "grad_norm": 0.012886490672826767, "learning_rate": 5.2517069226488694e-08, "loss": 0.0001, "num_input_tokens_seen": 4798816, "step": 725 }, { "epoch": 4.672566371681416, "grad_norm": 0.04072566702961922, "learning_rate": 5.044441612901768e-08, "loss": 0.0001, "num_input_tokens_seen": 4805408, "step": 726 }, { "epoch": 4.679002413515688, "grad_norm": 0.07961362600326538, "learning_rate": 4.841307382755567e-08, "loss": 0.0002, "num_input_tokens_seen": 4812480, "step": 727 }, { "epoch": 4.68543845534996, "grad_norm": 0.006482974626123905, "learning_rate": 4.6423076579895646e-08, "loss": 0.0, "num_input_tokens_seen": 4819360, "step": 728 }, { "epoch": 4.691874497184232, "grad_norm": 0.06561946123838425, "learning_rate": 4.4474457946562245e-08, "loss": 0.0003, "num_input_tokens_seen": 4825904, "step": 729 }, { "epoch": 4.698310539018504, "grad_norm": 0.10200546681880951, "learning_rate": 4.256725079024554e-08, "loss": 0.0004, "num_input_tokens_seen": 4832544, "step": 730 }, { "epoch": 4.704746580852776, "grad_norm": 0.2905844449996948, "learning_rate": 4.070148727524814e-08, "loss": 0.001, "num_input_tokens_seen": 4838960, "step": 731 }, { "epoch": 4.711182622687048, "grad_norm": 0.013804874382913113, "learning_rate": 3.887719886694091e-08, "loss": 0.0001, "num_input_tokens_seen": 4845584, "step": 732 }, { "epoch": 4.71761866452132, "grad_norm": 0.8108282685279846, "learning_rate": 3.709441633123367e-08, "loss": 0.0031, "num_input_tokens_seen": 4852416, "step": 733 }, { "epoch": 4.7240547063555915, "grad_norm": 0.11679325252771378, "learning_rate": 3.535316973405672e-08, "loss": 0.0006, "num_input_tokens_seen": 4858864, "step": 734 }, { "epoch": 4.7304907481898635, "grad_norm": 0.013299252837896347, "learning_rate": 3.3653488440851255e-08, "loss": 0.0001, "num_input_tokens_seen": 4865552, "step": 735 }, { "epoch": 4.736926790024135, "grad_norm": 0.8139033913612366, "learning_rate": 3.1995401116077516e-08, "loss": 0.006, "num_input_tokens_seen": 4871984, "step": 736 }, { "epoch": 4.743362831858407, "grad_norm": 0.023770008236169815, "learning_rate": 3.037893572272937e-08, "loss": 0.0001, "num_input_tokens_seen": 4878688, "step": 737 }, { "epoch": 4.749798873692679, "grad_norm": 0.1465252786874771, "learning_rate": 2.8804119521862183e-08, "loss": 0.001, "num_input_tokens_seen": 4885504, "step": 738 }, { "epoch": 4.756234915526951, "grad_norm": 0.7871643304824829, "learning_rate": 2.7270979072135106e-08, "loss": 0.0019, "num_input_tokens_seen": 4892272, "step": 739 }, { "epoch": 4.762670957361223, "grad_norm": 0.6069658994674683, "learning_rate": 2.5779540229361744e-08, "loss": 0.0047, "num_input_tokens_seen": 4898848, "step": 740 }, { "epoch": 4.769106999195495, "grad_norm": 0.07687046378850937, "learning_rate": 2.4329828146074096e-08, "loss": 0.0002, "num_input_tokens_seen": 4905392, "step": 741 }, { "epoch": 4.775543041029767, "grad_norm": 0.0760512426495552, "learning_rate": 2.2921867271099296e-08, "loss": 0.0002, "num_input_tokens_seen": 4912672, "step": 742 }, { "epoch": 4.781979082864039, "grad_norm": 0.14020369946956635, "learning_rate": 2.155568134914604e-08, "loss": 0.0006, "num_input_tokens_seen": 4919472, "step": 743 }, { "epoch": 4.788415124698311, "grad_norm": 0.020505385473370552, "learning_rate": 2.0231293420405194e-08, "loss": 0.0001, "num_input_tokens_seen": 4926448, "step": 744 }, { "epoch": 4.794851166532583, "grad_norm": 0.6544126868247986, "learning_rate": 1.8948725820160663e-08, "loss": 0.0038, "num_input_tokens_seen": 4933152, "step": 745 }, { "epoch": 4.801287208366855, "grad_norm": 1.0113900899887085, "learning_rate": 1.770800017841301e-08, "loss": 0.0088, "num_input_tokens_seen": 4939568, "step": 746 }, { "epoch": 4.807723250201127, "grad_norm": 0.05956251546740532, "learning_rate": 1.650913741951421e-08, "loss": 0.0002, "num_input_tokens_seen": 4946224, "step": 747 }, { "epoch": 4.814159292035399, "grad_norm": 0.21396887302398682, "learning_rate": 1.5352157761815978e-08, "loss": 0.002, "num_input_tokens_seen": 4952880, "step": 748 }, { "epoch": 4.8205953338696705, "grad_norm": 0.3376445472240448, "learning_rate": 1.4237080717326712e-08, "loss": 0.0012, "num_input_tokens_seen": 4960288, "step": 749 }, { "epoch": 4.8270313757039425, "grad_norm": 0.47905248403549194, "learning_rate": 1.3163925091384532e-08, "loss": 0.0038, "num_input_tokens_seen": 4966656, "step": 750 }, { "epoch": 4.833467417538214, "grad_norm": 0.08333203196525574, "learning_rate": 1.2132708982338925e-08, "loss": 0.0003, "num_input_tokens_seen": 4973184, "step": 751 }, { "epoch": 4.839903459372486, "grad_norm": 0.45923787355422974, "learning_rate": 1.1143449781245985e-08, "loss": 0.0023, "num_input_tokens_seen": 4980080, "step": 752 }, { "epoch": 4.846339501206758, "grad_norm": 0.038865748792886734, "learning_rate": 1.0196164171574762e-08, "loss": 0.0001, "num_input_tokens_seen": 4987104, "step": 753 }, { "epoch": 4.85277554304103, "grad_norm": 0.2948670983314514, "learning_rate": 9.290868128926378e-09, "loss": 0.0044, "num_input_tokens_seen": 4993728, "step": 754 }, { "epoch": 4.859211584875302, "grad_norm": 0.5785093903541565, "learning_rate": 8.427576920763957e-09, "loss": 0.004, "num_input_tokens_seen": 5000368, "step": 755 }, { "epoch": 4.865647626709574, "grad_norm": 0.29199209809303284, "learning_rate": 7.606305106155898e-09, "loss": 0.0042, "num_input_tokens_seen": 5006768, "step": 756 }, { "epoch": 4.872083668543846, "grad_norm": 0.25938865542411804, "learning_rate": 6.827066535529947e-09, "loss": 0.0004, "num_input_tokens_seen": 5013264, "step": 757 }, { "epoch": 4.878519710378118, "grad_norm": 0.1507510095834732, "learning_rate": 6.089874350439507e-09, "loss": 0.0005, "num_input_tokens_seen": 5019744, "step": 758 }, { "epoch": 4.88495575221239, "grad_norm": 0.009750754572451115, "learning_rate": 5.394740983341862e-09, "loss": 0.0, "num_input_tokens_seen": 5026080, "step": 759 }, { "epoch": 4.891391794046662, "grad_norm": 0.049791790544986725, "learning_rate": 4.74167815738974e-09, "loss": 0.0001, "num_input_tokens_seen": 5032544, "step": 760 }, { "epoch": 4.897827835880933, "grad_norm": 0.971774160861969, "learning_rate": 4.130696886231744e-09, "loss": 0.0089, "num_input_tokens_seen": 5039248, "step": 761 }, { "epoch": 4.904263877715205, "grad_norm": 0.9287542700767517, "learning_rate": 3.561807473827783e-09, "loss": 0.0118, "num_input_tokens_seen": 5045792, "step": 762 }, { "epoch": 4.910699919549477, "grad_norm": 0.06942977011203766, "learning_rate": 3.035019514275317e-09, "loss": 0.0003, "num_input_tokens_seen": 5052320, "step": 763 }, { "epoch": 4.917135961383749, "grad_norm": 0.28999796509742737, "learning_rate": 2.5503418916464352e-09, "loss": 0.0013, "num_input_tokens_seen": 5059200, "step": 764 }, { "epoch": 4.923572003218021, "grad_norm": 0.8473367691040039, "learning_rate": 2.1077827798404728e-09, "loss": 0.0058, "num_input_tokens_seen": 5065824, "step": 765 }, { "epoch": 4.9300080450522925, "grad_norm": 0.14826533198356628, "learning_rate": 1.707349642442735e-09, "loss": 0.0002, "num_input_tokens_seen": 5072080, "step": 766 }, { "epoch": 4.936444086886564, "grad_norm": 0.9073830246925354, "learning_rate": 1.349049232601818e-09, "loss": 0.0059, "num_input_tokens_seen": 5078336, "step": 767 }, { "epoch": 4.942880128720836, "grad_norm": 1.048936367034912, "learning_rate": 1.0328875929138671e-09, "loss": 0.0068, "num_input_tokens_seen": 5085152, "step": 768 }, { "epoch": 4.949316170555108, "grad_norm": 0.06418836861848831, "learning_rate": 7.588700553209926e-10, "loss": 0.0003, "num_input_tokens_seen": 5092080, "step": 769 }, { "epoch": 4.95575221238938, "grad_norm": 2.129972457885742, "learning_rate": 5.270012410216185e-10, "loss": 0.0251, "num_input_tokens_seen": 5098960, "step": 770 }, { "epoch": 4.962188254223652, "grad_norm": 0.3871181309223175, "learning_rate": 3.3728506039276686e-10, "loss": 0.0054, "num_input_tokens_seen": 5105648, "step": 771 }, { "epoch": 4.968624296057924, "grad_norm": 0.185493603348732, "learning_rate": 1.8972471292344474e-10, "loss": 0.0007, "num_input_tokens_seen": 5112096, "step": 772 }, { "epoch": 4.975060337892196, "grad_norm": 0.05420377105474472, "learning_rate": 8.432268716135338e-11, "loss": 0.0002, "num_input_tokens_seen": 5118960, "step": 773 }, { "epoch": 4.981496379726468, "grad_norm": 0.5486555695533752, "learning_rate": 2.108076067014464e-11, "loss": 0.0032, "num_input_tokens_seen": 5125872, "step": 774 }, { "epoch": 4.98793242156074, "grad_norm": 1.0398619174957275, "learning_rate": 0.0, "loss": 0.0107, "num_input_tokens_seen": 5132288, "step": 775 }, { "epoch": 4.98793242156074, "num_input_tokens_seen": 5132288, "step": 775, "total_flos": 2.3110461174474342e+17, "train_loss": 0.14078616270634925, "train_runtime": 10317.9615, "train_samples_per_second": 9.634, "train_steps_per_second": 0.075 } ], "logging_steps": 1, "max_steps": 775, "num_input_tokens_seen": 5132288, "num_train_epochs": 5, "save_steps": 5000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.3110461174474342e+17, "train_batch_size": 2, "trial_name": null, "trial_params": null }