{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.2012673791733661, "eval_steps": 133, "global_step": 3180, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0003783221413033198, "grad_norm": 56.01426433664828, "learning_rate": 1e-08, "loss": 8.5655, "step": 1 }, { "epoch": 0.0003783221413033198, "eval_loss": 8.416223526000977, "eval_runtime": 26.8642, "eval_samples_per_second": 32.944, "eval_steps_per_second": 1.042, "step": 1 }, { "epoch": 0.0003783221413033198, "eval_bench_accuracy_arc_challenge": 0.12857142857142856, "eval_bench_accuracy_hellaswag": 0.025, "eval_bench_accuracy_mmlu": 0.21739130434782608, "eval_bench_average_accuracy": 0.1236542443064182, "eval_bench_loss": 10.19635223924068, "eval_bench_total_accuracy": 0.1054945054945055, "step": 1 }, { "epoch": 0.0007566442826066396, "grad_norm": 52.75063804651517, "learning_rate": 2e-08, "loss": 8.4236, "step": 2 }, { "epoch": 0.0011349664239099593, "grad_norm": 54.29511008856074, "learning_rate": 3e-08, "loss": 8.5128, "step": 3 }, { "epoch": 0.0015132885652132792, "grad_norm": 50.84717091006242, "learning_rate": 4e-08, "loss": 8.368, "step": 4 }, { "epoch": 0.0018916107065165989, "grad_norm": 58.682276590467374, "learning_rate": 5e-08, "loss": 8.5171, "step": 5 }, { "epoch": 0.0022699328478199185, "grad_norm": 54.19973526319146, "learning_rate": 6e-08, "loss": 8.4329, "step": 6 }, { "epoch": 0.0026482549891232382, "grad_norm": 52.00177926668044, "learning_rate": 7e-08, "loss": 8.4562, "step": 7 }, { "epoch": 0.0030265771304265584, "grad_norm": 55.9652762703784, "learning_rate": 8e-08, "loss": 8.5017, "step": 8 }, { "epoch": 0.003404899271729878, "grad_norm": 54.88105368356734, "learning_rate": 9e-08, "loss": 8.471, "step": 9 }, { "epoch": 0.0037832214130331977, "grad_norm": 50.22661382824928, "learning_rate": 1e-07, "loss": 8.4042, "step": 10 }, { "epoch": 0.004161543554336518, "grad_norm": 51.712774406266966, "learning_rate": 1.0999999999999999e-07, "loss": 8.4819, "step": 11 }, { "epoch": 0.004539865695639837, "grad_norm": 44.20700801792938, "learning_rate": 1.2e-07, "loss": 8.2981, "step": 12 }, { "epoch": 0.004918187836943157, "grad_norm": 46.914384802444836, "learning_rate": 1.3e-07, "loss": 8.4152, "step": 13 }, { "epoch": 0.0052965099782464765, "grad_norm": 46.66045652280597, "learning_rate": 1.4e-07, "loss": 8.4776, "step": 14 }, { "epoch": 0.005674832119549797, "grad_norm": 45.99567071730722, "learning_rate": 1.5e-07, "loss": 8.4602, "step": 15 }, { "epoch": 0.006053154260853117, "grad_norm": 31.7220420827569, "learning_rate": 1.6e-07, "loss": 8.342, "step": 16 }, { "epoch": 0.006431476402156436, "grad_norm": 31.79821930177939, "learning_rate": 1.7000000000000001e-07, "loss": 8.4073, "step": 17 }, { "epoch": 0.006809798543459756, "grad_norm": 34.99852513062481, "learning_rate": 1.8e-07, "loss": 8.4475, "step": 18 }, { "epoch": 0.007188120684763075, "grad_norm": 32.34312521349501, "learning_rate": 1.8999999999999998e-07, "loss": 8.3691, "step": 19 }, { "epoch": 0.0075664428260663955, "grad_norm": 28.491575199383966, "learning_rate": 2e-07, "loss": 8.2467, "step": 20 }, { "epoch": 0.007944764967369716, "grad_norm": 27.788350456113577, "learning_rate": 2.0999999999999997e-07, "loss": 8.2619, "step": 21 }, { "epoch": 0.008323087108673036, "grad_norm": 23.054768686734494, "learning_rate": 2.1999999999999998e-07, "loss": 8.2719, "step": 22 }, { "epoch": 0.008701409249976354, "grad_norm": 20.862948070445295, "learning_rate": 2.3e-07, "loss": 8.1701, "step": 23 }, { "epoch": 0.009079731391279674, "grad_norm": 23.840305973367958, "learning_rate": 2.4e-07, "loss": 8.2447, "step": 24 }, { "epoch": 0.009458053532582994, "grad_norm": 22.407061285607927, "learning_rate": 2.5e-07, "loss": 8.2056, "step": 25 }, { "epoch": 0.009836375673886314, "grad_norm": 21.55132867797403, "learning_rate": 2.6e-07, "loss": 8.1552, "step": 26 }, { "epoch": 0.010214697815189635, "grad_norm": 20.992840710071967, "learning_rate": 2.7e-07, "loss": 8.188, "step": 27 }, { "epoch": 0.010593019956492953, "grad_norm": 22.39828627182125, "learning_rate": 2.8e-07, "loss": 8.1256, "step": 28 }, { "epoch": 0.010971342097796273, "grad_norm": 18.46346034557574, "learning_rate": 2.9e-07, "loss": 8.0045, "step": 29 }, { "epoch": 0.011349664239099593, "grad_norm": 12.704677816631309, "learning_rate": 3e-07, "loss": 8.0417, "step": 30 }, { "epoch": 0.011727986380402913, "grad_norm": 15.722346563574124, "learning_rate": 3.1e-07, "loss": 7.9647, "step": 31 }, { "epoch": 0.012106308521706233, "grad_norm": 14.31712037195988, "learning_rate": 3.2e-07, "loss": 8.0119, "step": 32 }, { "epoch": 0.012484630663009552, "grad_norm": 13.002942588027526, "learning_rate": 3.3e-07, "loss": 8.029, "step": 33 }, { "epoch": 0.012862952804312872, "grad_norm": 15.303670533896709, "learning_rate": 3.4000000000000003e-07, "loss": 7.9847, "step": 34 }, { "epoch": 0.013241274945616192, "grad_norm": 12.964425414274471, "learning_rate": 3.5e-07, "loss": 8.0026, "step": 35 }, { "epoch": 0.013619597086919512, "grad_norm": 19.040688578500415, "learning_rate": 3.6e-07, "loss": 8.0397, "step": 36 }, { "epoch": 0.013997919228222832, "grad_norm": 14.264527574014561, "learning_rate": 3.7e-07, "loss": 7.8472, "step": 37 }, { "epoch": 0.01437624136952615, "grad_norm": 14.259878980724565, "learning_rate": 3.7999999999999996e-07, "loss": 7.9499, "step": 38 }, { "epoch": 0.01475456351082947, "grad_norm": 21.02927607859569, "learning_rate": 3.8999999999999997e-07, "loss": 7.8521, "step": 39 }, { "epoch": 0.015132885652132791, "grad_norm": 16.308228829260607, "learning_rate": 4e-07, "loss": 7.8008, "step": 40 }, { "epoch": 0.015511207793436111, "grad_norm": 21.835730681754328, "learning_rate": 4.0999999999999994e-07, "loss": 7.7515, "step": 41 }, { "epoch": 0.01588952993473943, "grad_norm": 22.548471887636545, "learning_rate": 4.1999999999999995e-07, "loss": 7.7859, "step": 42 }, { "epoch": 0.01626785207604275, "grad_norm": 23.40758724577002, "learning_rate": 4.2999999999999996e-07, "loss": 7.7679, "step": 43 }, { "epoch": 0.01664617421734607, "grad_norm": 22.806229545212982, "learning_rate": 4.3999999999999997e-07, "loss": 7.7211, "step": 44 }, { "epoch": 0.01702449635864939, "grad_norm": 19.930882370057223, "learning_rate": 4.5e-07, "loss": 7.7017, "step": 45 }, { "epoch": 0.017402818499952708, "grad_norm": 17.292062567746196, "learning_rate": 4.6e-07, "loss": 7.7146, "step": 46 }, { "epoch": 0.01778114064125603, "grad_norm": 18.070618266890932, "learning_rate": 4.6999999999999995e-07, "loss": 7.7119, "step": 47 }, { "epoch": 0.01815946278255935, "grad_norm": 16.65539275683302, "learning_rate": 4.8e-07, "loss": 7.6178, "step": 48 }, { "epoch": 0.01853778492386267, "grad_norm": 19.36073786979339, "learning_rate": 4.9e-07, "loss": 7.6387, "step": 49 }, { "epoch": 0.01891610706516599, "grad_norm": 22.520853767642276, "learning_rate": 5e-07, "loss": 7.6346, "step": 50 }, { "epoch": 0.01929442920646931, "grad_norm": 21.674704957397896, "learning_rate": 5.1e-07, "loss": 7.5339, "step": 51 }, { "epoch": 0.01967275134777263, "grad_norm": 26.85039717209422, "learning_rate": 5.2e-07, "loss": 7.3655, "step": 52 }, { "epoch": 0.02005107348907595, "grad_norm": 29.784500661137994, "learning_rate": 5.3e-07, "loss": 7.3935, "step": 53 }, { "epoch": 0.02042939563037927, "grad_norm": 36.73803214173563, "learning_rate": 5.4e-07, "loss": 7.3942, "step": 54 }, { "epoch": 0.02080771777168259, "grad_norm": 55.998259201380826, "learning_rate": 5.5e-07, "loss": 7.3246, "step": 55 }, { "epoch": 0.021186039912985906, "grad_norm": 54.6219968094922, "learning_rate": 5.6e-07, "loss": 7.2241, "step": 56 }, { "epoch": 0.021564362054289226, "grad_norm": 115.48000957700997, "learning_rate": 5.699999999999999e-07, "loss": 7.3169, "step": 57 }, { "epoch": 0.021942684195592546, "grad_norm": 240.40441808566737, "learning_rate": 5.8e-07, "loss": 7.1243, "step": 58 }, { "epoch": 0.022321006336895866, "grad_norm": 102.2272021984647, "learning_rate": 5.9e-07, "loss": 7.0371, "step": 59 }, { "epoch": 0.022699328478199186, "grad_norm": 256.9288700751086, "learning_rate": 6e-07, "loss": 6.8907, "step": 60 }, { "epoch": 0.023077650619502506, "grad_norm": 131.56800170402965, "learning_rate": 6.1e-07, "loss": 6.854, "step": 61 }, { "epoch": 0.023455972760805827, "grad_norm": 358.2045690657579, "learning_rate": 6.2e-07, "loss": 6.7673, "step": 62 }, { "epoch": 0.023834294902109147, "grad_norm": 259.0360488341225, "learning_rate": 6.3e-07, "loss": 6.6898, "step": 63 }, { "epoch": 0.024212617043412467, "grad_norm": 324.46556421575104, "learning_rate": 6.4e-07, "loss": 6.6792, "step": 64 }, { "epoch": 0.024590939184715787, "grad_norm": 218.90309813691587, "learning_rate": 6.5e-07, "loss": 6.5833, "step": 65 }, { "epoch": 0.024969261326019104, "grad_norm": 345.9947605906595, "learning_rate": 6.6e-07, "loss": 6.5841, "step": 66 }, { "epoch": 0.025347583467322424, "grad_norm": 327.5192852015763, "learning_rate": 6.7e-07, "loss": 6.5379, "step": 67 }, { "epoch": 0.025725905608625744, "grad_norm": 272.0304082708135, "learning_rate": 6.800000000000001e-07, "loss": 6.4003, "step": 68 }, { "epoch": 0.026104227749929064, "grad_norm": 224.03062395364572, "learning_rate": 6.9e-07, "loss": 6.3064, "step": 69 }, { "epoch": 0.026482549891232384, "grad_norm": 326.13516923115037, "learning_rate": 7e-07, "loss": 6.2681, "step": 70 }, { "epoch": 0.026860872032535704, "grad_norm": 236.06386821993763, "learning_rate": 7.1e-07, "loss": 6.1658, "step": 71 }, { "epoch": 0.027239194173839024, "grad_norm": 117.09820504079929, "learning_rate": 7.2e-07, "loss": 6.1013, "step": 72 }, { "epoch": 0.027617516315142344, "grad_norm": 130.77996709008073, "learning_rate": 7.3e-07, "loss": 6.0313, "step": 73 }, { "epoch": 0.027995838456445665, "grad_norm": 184.1694406122909, "learning_rate": 7.4e-07, "loss": 5.9761, "step": 74 }, { "epoch": 0.028374160597748985, "grad_norm": 107.41668355609693, "learning_rate": 7.5e-07, "loss": 5.8533, "step": 75 }, { "epoch": 0.0287524827390523, "grad_norm": 167.17458055865583, "learning_rate": 7.599999999999999e-07, "loss": 5.842, "step": 76 }, { "epoch": 0.02913080488035562, "grad_norm": 83.1018765552699, "learning_rate": 7.699999999999999e-07, "loss": 5.8106, "step": 77 }, { "epoch": 0.02950912702165894, "grad_norm": 930.4199949174266, "learning_rate": 7.799999999999999e-07, "loss": 5.9417, "step": 78 }, { "epoch": 0.02988744916296226, "grad_norm": 344.9243101513464, "learning_rate": 7.9e-07, "loss": 5.9401, "step": 79 }, { "epoch": 0.030265771304265582, "grad_norm": 203.82832876269842, "learning_rate": 8e-07, "loss": 5.8335, "step": 80 }, { "epoch": 0.030644093445568902, "grad_norm": 303.4319382071192, "learning_rate": 8.1e-07, "loss": 5.6823, "step": 81 }, { "epoch": 0.031022415586872222, "grad_norm": 248.28331376619403, "learning_rate": 8.199999999999999e-07, "loss": 5.7745, "step": 82 }, { "epoch": 0.03140073772817554, "grad_norm": 462.20565983043144, "learning_rate": 8.299999999999999e-07, "loss": 5.6386, "step": 83 }, { "epoch": 0.03177905986947886, "grad_norm": 194.41981862598635, "learning_rate": 8.399999999999999e-07, "loss": 5.5997, "step": 84 }, { "epoch": 0.03215738201078218, "grad_norm": 293.3275031516269, "learning_rate": 8.499999999999999e-07, "loss": 5.5106, "step": 85 }, { "epoch": 0.0325357041520855, "grad_norm": 140.97321101678344, "learning_rate": 8.599999999999999e-07, "loss": 5.4563, "step": 86 }, { "epoch": 0.03291402629338882, "grad_norm": 180.15140475284437, "learning_rate": 8.699999999999999e-07, "loss": 5.4357, "step": 87 }, { "epoch": 0.03329234843469214, "grad_norm": 333.3719583206301, "learning_rate": 8.799999999999999e-07, "loss": 5.3168, "step": 88 }, { "epoch": 0.03367067057599546, "grad_norm": 121.82713201522955, "learning_rate": 8.9e-07, "loss": 5.3945, "step": 89 }, { "epoch": 0.03404899271729878, "grad_norm": 582.7969295558685, "learning_rate": 9e-07, "loss": 5.3863, "step": 90 }, { "epoch": 0.0344273148586021, "grad_norm": 217.6434706478821, "learning_rate": 9.1e-07, "loss": 5.2662, "step": 91 }, { "epoch": 0.034805636999905416, "grad_norm": 374.4674448505233, "learning_rate": 9.2e-07, "loss": 5.2355, "step": 92 }, { "epoch": 0.03518395914120874, "grad_norm": 218.23465312606612, "learning_rate": 9.3e-07, "loss": 5.1486, "step": 93 }, { "epoch": 0.03556228128251206, "grad_norm": 98.81927420372956, "learning_rate": 9.399999999999999e-07, "loss": 5.0807, "step": 94 }, { "epoch": 0.03594060342381538, "grad_norm": 211.12146153212487, "learning_rate": 9.499999999999999e-07, "loss": 5.0853, "step": 95 }, { "epoch": 0.0363189255651187, "grad_norm": 190.3736868117524, "learning_rate": 9.6e-07, "loss": 5.0756, "step": 96 }, { "epoch": 0.03669724770642202, "grad_norm": 122.03862248450174, "learning_rate": 9.7e-07, "loss": 4.9252, "step": 97 }, { "epoch": 0.03707556984772534, "grad_norm": 410.81026410608786, "learning_rate": 9.8e-07, "loss": 5.0664, "step": 98 }, { "epoch": 0.03745389198902866, "grad_norm": 269.97951212839484, "learning_rate": 9.9e-07, "loss": 4.9091, "step": 99 }, { "epoch": 0.03783221413033198, "grad_norm": 260.7212338620472, "learning_rate": 1e-06, "loss": 4.8821, "step": 100 }, { "epoch": 0.0382105362716353, "grad_norm": 165.92539323350238, "learning_rate": 1.0099999999999999e-06, "loss": 4.7469, "step": 101 }, { "epoch": 0.03858885841293862, "grad_norm": 281.9862388742268, "learning_rate": 1.02e-06, "loss": 4.7974, "step": 102 }, { "epoch": 0.038967180554241934, "grad_norm": 164.28597977866295, "learning_rate": 1.0299999999999999e-06, "loss": 4.6513, "step": 103 }, { "epoch": 0.03934550269554526, "grad_norm": 315.7550450358392, "learning_rate": 1.04e-06, "loss": 4.7021, "step": 104 }, { "epoch": 0.039723824836848574, "grad_norm": 202.93065604656107, "learning_rate": 1.05e-06, "loss": 4.5712, "step": 105 }, { "epoch": 0.0401021469781519, "grad_norm": 210.26805622762828, "learning_rate": 1.06e-06, "loss": 4.6196, "step": 106 }, { "epoch": 0.040480469119455215, "grad_norm": 187.14917857744504, "learning_rate": 1.07e-06, "loss": 4.5484, "step": 107 }, { "epoch": 0.04085879126075854, "grad_norm": 155.43076076847103, "learning_rate": 1.08e-06, "loss": 4.4144, "step": 108 }, { "epoch": 0.041237113402061855, "grad_norm": 154.98829996861681, "learning_rate": 1.09e-06, "loss": 4.3404, "step": 109 }, { "epoch": 0.04161543554336518, "grad_norm": 141.595366217918, "learning_rate": 1.1e-06, "loss": 4.3111, "step": 110 }, { "epoch": 0.041993757684668495, "grad_norm": 134.27240833451944, "learning_rate": 1.11e-06, "loss": 4.1952, "step": 111 }, { "epoch": 0.04237207982597181, "grad_norm": 95.65375597330166, "learning_rate": 1.12e-06, "loss": 4.0809, "step": 112 }, { "epoch": 0.042750401967275135, "grad_norm": 109.07352101322023, "learning_rate": 1.1299999999999998e-06, "loss": 4.0286, "step": 113 }, { "epoch": 0.04312872410857845, "grad_norm": 114.47547920727833, "learning_rate": 1.1399999999999999e-06, "loss": 3.9147, "step": 114 }, { "epoch": 0.043507046249881776, "grad_norm": 105.22542090856187, "learning_rate": 1.1499999999999998e-06, "loss": 3.888, "step": 115 }, { "epoch": 0.04388536839118509, "grad_norm": 170.85609503557524, "learning_rate": 1.16e-06, "loss": 3.7806, "step": 116 }, { "epoch": 0.044263690532488416, "grad_norm": 132.60484964177928, "learning_rate": 1.1699999999999998e-06, "loss": 3.7388, "step": 117 }, { "epoch": 0.04464201267379173, "grad_norm": 817.4981900388101, "learning_rate": 1.18e-06, "loss": 3.8085, "step": 118 }, { "epoch": 0.045020334815095056, "grad_norm": 277.2968095396992, "learning_rate": 1.1899999999999998e-06, "loss": 3.7519, "step": 119 }, { "epoch": 0.04539865695639837, "grad_norm": 242.3036172020571, "learning_rate": 1.2e-06, "loss": 3.6811, "step": 120 }, { "epoch": 0.045776979097701696, "grad_norm": 147.12958250512, "learning_rate": 1.2099999999999998e-06, "loss": 3.5537, "step": 121 }, { "epoch": 0.04615530123900501, "grad_norm": 304.91416915276426, "learning_rate": 1.22e-06, "loss": 3.5308, "step": 122 }, { "epoch": 0.04653362338030833, "grad_norm": 228.8092972324273, "learning_rate": 1.2299999999999999e-06, "loss": 3.4916, "step": 123 }, { "epoch": 0.04691194552161165, "grad_norm": 197.353832945714, "learning_rate": 1.24e-06, "loss": 3.4215, "step": 124 }, { "epoch": 0.04729026766291497, "grad_norm": 228.72368996651358, "learning_rate": 1.2499999999999999e-06, "loss": 3.371, "step": 125 }, { "epoch": 0.04766858980421829, "grad_norm": 164.2731725612326, "learning_rate": 1.26e-06, "loss": 3.3909, "step": 126 }, { "epoch": 0.04804691194552161, "grad_norm": 186.5826183173996, "learning_rate": 1.27e-06, "loss": 3.3104, "step": 127 }, { "epoch": 0.048425234086824934, "grad_norm": 139.94786192019586, "learning_rate": 1.28e-06, "loss": 3.2437, "step": 128 }, { "epoch": 0.04880355622812825, "grad_norm": 170.89837594203516, "learning_rate": 1.29e-06, "loss": 3.2145, "step": 129 }, { "epoch": 0.049181878369431574, "grad_norm": 124.04755267516651, "learning_rate": 1.3e-06, "loss": 3.1275, "step": 130 }, { "epoch": 0.04956020051073489, "grad_norm": 112.7475091581948, "learning_rate": 1.31e-06, "loss": 3.1021, "step": 131 }, { "epoch": 0.04993852265203821, "grad_norm": 483.6676734928997, "learning_rate": 1.32e-06, "loss": 3.0251, "step": 132 }, { "epoch": 0.05031684479334153, "grad_norm": 131.48794283663062, "learning_rate": 1.33e-06, "loss": 3.0474, "step": 133 }, { "epoch": 0.05031684479334153, "eval_loss": 3.0402355194091797, "eval_runtime": 26.8305, "eval_samples_per_second": 32.985, "eval_steps_per_second": 1.044, "step": 133 }, { "epoch": 0.05031684479334153, "eval_bench_accuracy_arc_challenge": 0.2714285714285714, "eval_bench_accuracy_hellaswag": 0.22, "eval_bench_accuracy_mmlu": 0.23478260869565218, "eval_bench_average_accuracy": 0.2420703933747412, "eval_bench_loss": 6.577301560786733, "eval_bench_total_accuracy": 0.23956043956043957, "step": 133 }, { "epoch": 0.05069516693464485, "grad_norm": 664.2692049220283, "learning_rate": 1.34e-06, "loss": 3.0489, "step": 134 }, { "epoch": 0.05107348907594817, "grad_norm": 164.70902413028506, "learning_rate": 1.35e-06, "loss": 3.0729, "step": 135 }, { "epoch": 0.05145181121725149, "grad_norm": 778.4019675411471, "learning_rate": 1.3600000000000001e-06, "loss": 2.9025, "step": 136 }, { "epoch": 0.05183013335855481, "grad_norm": 141.784859477734, "learning_rate": 1.37e-06, "loss": 2.9153, "step": 137 }, { "epoch": 0.05220845549985813, "grad_norm": 815.6337164546584, "learning_rate": 1.38e-06, "loss": 2.9767, "step": 138 }, { "epoch": 0.05258677764116145, "grad_norm": 387.14144869932585, "learning_rate": 1.3899999999999998e-06, "loss": 2.9545, "step": 139 }, { "epoch": 0.05296509978246477, "grad_norm": 1286.7446765387322, "learning_rate": 1.4e-06, "loss": 2.9779, "step": 140 }, { "epoch": 0.05334342192376809, "grad_norm": 170.85639571110613, "learning_rate": 1.4099999999999998e-06, "loss": 2.8642, "step": 141 }, { "epoch": 0.05372174406507141, "grad_norm": 375.24244542748465, "learning_rate": 1.42e-06, "loss": 2.7942, "step": 142 }, { "epoch": 0.054100066206374725, "grad_norm": 154.53620941237315, "learning_rate": 1.4299999999999999e-06, "loss": 2.7527, "step": 143 }, { "epoch": 0.05447838834767805, "grad_norm": 188.97826644064364, "learning_rate": 1.44e-06, "loss": 2.7492, "step": 144 }, { "epoch": 0.054856710488981365, "grad_norm": 103.19619548153565, "learning_rate": 1.4499999999999999e-06, "loss": 2.6708, "step": 145 }, { "epoch": 0.05523503263028469, "grad_norm": 125.47407228350237, "learning_rate": 1.46e-06, "loss": 2.6737, "step": 146 }, { "epoch": 0.055613354771588006, "grad_norm": 71.31808903587059, "learning_rate": 1.47e-06, "loss": 2.6175, "step": 147 }, { "epoch": 0.05599167691289133, "grad_norm": 158.4470726659215, "learning_rate": 1.48e-06, "loss": 2.5772, "step": 148 }, { "epoch": 0.056369999054194646, "grad_norm": 213.54517556280484, "learning_rate": 1.49e-06, "loss": 2.5397, "step": 149 }, { "epoch": 0.05674832119549797, "grad_norm": 94.87447540886092, "learning_rate": 1.5e-06, "loss": 2.5007, "step": 150 }, { "epoch": 0.057126643336801286, "grad_norm": 140.6331701396571, "learning_rate": 1.51e-06, "loss": 2.4911, "step": 151 }, { "epoch": 0.0575049654781046, "grad_norm": 71.42229734282893, "learning_rate": 1.5199999999999998e-06, "loss": 2.3964, "step": 152 }, { "epoch": 0.057883287619407926, "grad_norm": 100.92797990716835, "learning_rate": 1.53e-06, "loss": 2.3796, "step": 153 }, { "epoch": 0.05826160976071124, "grad_norm": 69.12965458867137, "learning_rate": 1.5399999999999999e-06, "loss": 2.4147, "step": 154 }, { "epoch": 0.058639931902014567, "grad_norm": 68.31144568523656, "learning_rate": 1.55e-06, "loss": 2.285, "step": 155 }, { "epoch": 0.05901825404331788, "grad_norm": 63.86407191747168, "learning_rate": 1.5599999999999999e-06, "loss": 2.2905, "step": 156 }, { "epoch": 0.05939657618462121, "grad_norm": 89.9702991999028, "learning_rate": 1.57e-06, "loss": 2.2642, "step": 157 }, { "epoch": 0.05977489832592452, "grad_norm": 38.70583191014119, "learning_rate": 1.58e-06, "loss": 2.1927, "step": 158 }, { "epoch": 0.06015322046722785, "grad_norm": 150.0176513817121, "learning_rate": 1.59e-06, "loss": 2.2046, "step": 159 }, { "epoch": 0.060531542608531164, "grad_norm": 85.38752600608713, "learning_rate": 1.6e-06, "loss": 2.1777, "step": 160 }, { "epoch": 0.06090986474983449, "grad_norm": 108.46382637315519, "learning_rate": 1.61e-06, "loss": 2.0947, "step": 161 }, { "epoch": 0.061288186891137804, "grad_norm": 72.33751976980996, "learning_rate": 1.62e-06, "loss": 2.1455, "step": 162 }, { "epoch": 0.06166650903244112, "grad_norm": 254.7588636023186, "learning_rate": 1.6299999999999999e-06, "loss": 2.0967, "step": 163 }, { "epoch": 0.062044831173744444, "grad_norm": 143.3727693773649, "learning_rate": 1.6399999999999998e-06, "loss": 2.0443, "step": 164 }, { "epoch": 0.06242315331504776, "grad_norm": 672.6219381081797, "learning_rate": 1.6499999999999999e-06, "loss": 2.2139, "step": 165 }, { "epoch": 0.06280147545635108, "grad_norm": 89.69156829747156, "learning_rate": 1.6599999999999998e-06, "loss": 2.0433, "step": 166 }, { "epoch": 0.06317979759765441, "grad_norm": 47.054580203479496, "learning_rate": 1.6699999999999999e-06, "loss": 1.9805, "step": 167 }, { "epoch": 0.06355811973895772, "grad_norm": 53.90193516042071, "learning_rate": 1.6799999999999998e-06, "loss": 1.8572, "step": 168 }, { "epoch": 0.06393644188026104, "grad_norm": 55.351958687059195, "learning_rate": 1.69e-06, "loss": 1.8879, "step": 169 }, { "epoch": 0.06431476402156436, "grad_norm": 30.956994176305464, "learning_rate": 1.6999999999999998e-06, "loss": 1.8335, "step": 170 }, { "epoch": 0.06469308616286769, "grad_norm": 81.23380900946358, "learning_rate": 1.71e-06, "loss": 1.8101, "step": 171 }, { "epoch": 0.065071408304171, "grad_norm": 46.43733520396148, "learning_rate": 1.7199999999999998e-06, "loss": 1.8177, "step": 172 }, { "epoch": 0.06544973044547432, "grad_norm": 46.90830376181402, "learning_rate": 1.73e-06, "loss": 1.7543, "step": 173 }, { "epoch": 0.06582805258677764, "grad_norm": 69.19161149417722, "learning_rate": 1.7399999999999999e-06, "loss": 1.7712, "step": 174 }, { "epoch": 0.06620637472808096, "grad_norm": 46.99692135130498, "learning_rate": 1.75e-06, "loss": 1.7728, "step": 175 }, { "epoch": 0.06658469686938429, "grad_norm": 85.68605330443327, "learning_rate": 1.7599999999999999e-06, "loss": 1.7186, "step": 176 }, { "epoch": 0.0669630190106876, "grad_norm": 48.57963404347663, "learning_rate": 1.77e-06, "loss": 1.6979, "step": 177 }, { "epoch": 0.06734134115199092, "grad_norm": 111.44637207499896, "learning_rate": 1.78e-06, "loss": 1.734, "step": 178 }, { "epoch": 0.06771966329329424, "grad_norm": 83.89157732570692, "learning_rate": 1.79e-06, "loss": 1.6947, "step": 179 }, { "epoch": 0.06809798543459757, "grad_norm": 50.66006983599147, "learning_rate": 1.8e-06, "loss": 1.6385, "step": 180 }, { "epoch": 0.06847630757590088, "grad_norm": 47.32959657636825, "learning_rate": 1.81e-06, "loss": 1.5717, "step": 181 }, { "epoch": 0.0688546297172042, "grad_norm": 71.70671420810187, "learning_rate": 1.82e-06, "loss": 1.5167, "step": 182 }, { "epoch": 0.06923295185850752, "grad_norm": 48.11379424928171, "learning_rate": 1.83e-06, "loss": 1.5992, "step": 183 }, { "epoch": 0.06961127399981083, "grad_norm": 54.01731463177801, "learning_rate": 1.84e-06, "loss": 1.5217, "step": 184 }, { "epoch": 0.06998959614111416, "grad_norm": 39.52299725178149, "learning_rate": 1.85e-06, "loss": 1.5009, "step": 185 }, { "epoch": 0.07036791828241748, "grad_norm": 63.37058186080119, "learning_rate": 1.86e-06, "loss": 1.5853, "step": 186 }, { "epoch": 0.0707462404237208, "grad_norm": 44.5116426583779, "learning_rate": 1.87e-06, "loss": 1.4865, "step": 187 }, { "epoch": 0.07112456256502411, "grad_norm": 40.56409454228496, "learning_rate": 1.8799999999999998e-06, "loss": 1.4732, "step": 188 }, { "epoch": 0.07150288470632744, "grad_norm": 31.923505092753718, "learning_rate": 1.89e-06, "loss": 1.4519, "step": 189 }, { "epoch": 0.07188120684763076, "grad_norm": 34.50709112981039, "learning_rate": 1.8999999999999998e-06, "loss": 1.4205, "step": 190 }, { "epoch": 0.07225952898893408, "grad_norm": 22.09682402936458, "learning_rate": 1.91e-06, "loss": 1.38, "step": 191 }, { "epoch": 0.0726378511302374, "grad_norm": 25.3767669172789, "learning_rate": 1.92e-06, "loss": 1.3879, "step": 192 }, { "epoch": 0.07301617327154071, "grad_norm": 29.51813748066488, "learning_rate": 1.9299999999999997e-06, "loss": 1.3506, "step": 193 }, { "epoch": 0.07339449541284404, "grad_norm": 21.76501410574832, "learning_rate": 1.94e-06, "loss": 1.3237, "step": 194 }, { "epoch": 0.07377281755414736, "grad_norm": 20.74781891582525, "learning_rate": 1.95e-06, "loss": 1.3639, "step": 195 }, { "epoch": 0.07415113969545067, "grad_norm": 27.66733930317673, "learning_rate": 1.96e-06, "loss": 1.3061, "step": 196 }, { "epoch": 0.07452946183675399, "grad_norm": 21.087698250942193, "learning_rate": 1.9699999999999998e-06, "loss": 1.375, "step": 197 }, { "epoch": 0.07490778397805732, "grad_norm": 22.065927379036225, "learning_rate": 1.98e-06, "loss": 1.3219, "step": 198 }, { "epoch": 0.07528610611936064, "grad_norm": 37.132637966902955, "learning_rate": 1.99e-06, "loss": 1.2424, "step": 199 }, { "epoch": 0.07566442826066395, "grad_norm": 20.85100061426098, "learning_rate": 2e-06, "loss": 1.2973, "step": 200 }, { "epoch": 0.07604275040196727, "grad_norm": 19.748272671220768, "learning_rate": 2.01e-06, "loss": 1.2371, "step": 201 }, { "epoch": 0.0764210725432706, "grad_norm": 24.073543088140834, "learning_rate": 2.0199999999999997e-06, "loss": 1.252, "step": 202 }, { "epoch": 0.07679939468457392, "grad_norm": 34.22154387867275, "learning_rate": 2.0299999999999996e-06, "loss": 1.2911, "step": 203 }, { "epoch": 0.07717771682587724, "grad_norm": 16.511181722757403, "learning_rate": 2.04e-06, "loss": 1.2321, "step": 204 }, { "epoch": 0.07755603896718055, "grad_norm": 12.872226386234452, "learning_rate": 2.05e-06, "loss": 1.1767, "step": 205 }, { "epoch": 0.07793436110848387, "grad_norm": 15.436365816346868, "learning_rate": 2.0599999999999998e-06, "loss": 1.1955, "step": 206 }, { "epoch": 0.0783126832497872, "grad_norm": 12.062107586682833, "learning_rate": 2.0699999999999997e-06, "loss": 1.1799, "step": 207 }, { "epoch": 0.07869100539109052, "grad_norm": 49.38765930014822, "learning_rate": 2.08e-06, "loss": 1.1762, "step": 208 }, { "epoch": 0.07906932753239383, "grad_norm": 23.38441549316206, "learning_rate": 2.09e-06, "loss": 1.1831, "step": 209 }, { "epoch": 0.07944764967369715, "grad_norm": 22.28035230836217, "learning_rate": 2.1e-06, "loss": 1.1858, "step": 210 }, { "epoch": 0.07982597181500048, "grad_norm": 43.05138932031075, "learning_rate": 2.1099999999999997e-06, "loss": 1.2106, "step": 211 }, { "epoch": 0.0802042939563038, "grad_norm": 22.919581037837645, "learning_rate": 2.12e-06, "loss": 1.1872, "step": 212 }, { "epoch": 0.08058261609760711, "grad_norm": 106.27528509092721, "learning_rate": 2.13e-06, "loss": 1.1807, "step": 213 }, { "epoch": 0.08096093823891043, "grad_norm": 62.766496496977574, "learning_rate": 2.14e-06, "loss": 1.1932, "step": 214 }, { "epoch": 0.08133926038021375, "grad_norm": 66.54674237816508, "learning_rate": 2.1499999999999997e-06, "loss": 1.1328, "step": 215 }, { "epoch": 0.08171758252151708, "grad_norm": 66.81453157766589, "learning_rate": 2.16e-06, "loss": 1.1613, "step": 216 }, { "epoch": 0.0820959046628204, "grad_norm": 35.57901795776919, "learning_rate": 2.17e-06, "loss": 1.1821, "step": 217 }, { "epoch": 0.08247422680412371, "grad_norm": 10.30900211340774, "learning_rate": 2.18e-06, "loss": 1.1023, "step": 218 }, { "epoch": 0.08285254894542703, "grad_norm": 29.533042017371177, "learning_rate": 2.1899999999999998e-06, "loss": 1.1669, "step": 219 }, { "epoch": 0.08323087108673036, "grad_norm": 22.47096674174166, "learning_rate": 2.2e-06, "loss": 1.1612, "step": 220 }, { "epoch": 0.08360919322803367, "grad_norm": 13.583126551810135, "learning_rate": 2.21e-06, "loss": 1.0867, "step": 221 }, { "epoch": 0.08398751536933699, "grad_norm": 9.91479302526445, "learning_rate": 2.22e-06, "loss": 1.0916, "step": 222 }, { "epoch": 0.0843658375106403, "grad_norm": 11.269431287067826, "learning_rate": 2.23e-06, "loss": 1.1264, "step": 223 }, { "epoch": 0.08474415965194362, "grad_norm": 7.7465735801712805, "learning_rate": 2.24e-06, "loss": 1.136, "step": 224 }, { "epoch": 0.08512248179324695, "grad_norm": 8.687635755465738, "learning_rate": 2.25e-06, "loss": 1.0803, "step": 225 }, { "epoch": 0.08550080393455027, "grad_norm": 11.628437205512707, "learning_rate": 2.2599999999999995e-06, "loss": 1.1646, "step": 226 }, { "epoch": 0.08587912607585359, "grad_norm": 9.268721256498573, "learning_rate": 2.27e-06, "loss": 1.1015, "step": 227 }, { "epoch": 0.0862574482171569, "grad_norm": 6.187500026884083, "learning_rate": 2.2799999999999998e-06, "loss": 1.0662, "step": 228 }, { "epoch": 0.08663577035846023, "grad_norm": 8.62028463677054, "learning_rate": 2.29e-06, "loss": 1.052, "step": 229 }, { "epoch": 0.08701409249976355, "grad_norm": 9.674790887814405, "learning_rate": 2.2999999999999996e-06, "loss": 1.0978, "step": 230 }, { "epoch": 0.08739241464106687, "grad_norm": 8.326705028491853, "learning_rate": 2.31e-06, "loss": 1.0184, "step": 231 }, { "epoch": 0.08777073678237018, "grad_norm": 7.318027642173224, "learning_rate": 2.32e-06, "loss": 1.0509, "step": 232 }, { "epoch": 0.0881490589236735, "grad_norm": 12.85041462496061, "learning_rate": 2.33e-06, "loss": 1.0556, "step": 233 }, { "epoch": 0.08852738106497683, "grad_norm": 9.328207044954535, "learning_rate": 2.3399999999999996e-06, "loss": 1.0816, "step": 234 }, { "epoch": 0.08890570320628015, "grad_norm": 7.022150416570471, "learning_rate": 2.35e-06, "loss": 1.0466, "step": 235 }, { "epoch": 0.08928402534758346, "grad_norm": 8.86057501782776, "learning_rate": 2.36e-06, "loss": 1.04, "step": 236 }, { "epoch": 0.08966234748888678, "grad_norm": 9.072613041437753, "learning_rate": 2.37e-06, "loss": 1.039, "step": 237 }, { "epoch": 0.09004066963019011, "grad_norm": 11.561198612520238, "learning_rate": 2.3799999999999997e-06, "loss": 1.025, "step": 238 }, { "epoch": 0.09041899177149343, "grad_norm": 5.796410505813014, "learning_rate": 2.39e-06, "loss": 1.0007, "step": 239 }, { "epoch": 0.09079731391279675, "grad_norm": 13.451590053171754, "learning_rate": 2.4e-06, "loss": 1.0051, "step": 240 }, { "epoch": 0.09117563605410006, "grad_norm": 8.917436837849364, "learning_rate": 2.4100000000000002e-06, "loss": 1.0866, "step": 241 }, { "epoch": 0.09155395819540339, "grad_norm": 4.792174398814023, "learning_rate": 2.4199999999999997e-06, "loss": 1.0022, "step": 242 }, { "epoch": 0.09193228033670671, "grad_norm": 6.487991210049911, "learning_rate": 2.43e-06, "loss": 0.976, "step": 243 }, { "epoch": 0.09231060247801003, "grad_norm": 9.885175529767102, "learning_rate": 2.44e-06, "loss": 1.0038, "step": 244 }, { "epoch": 0.09268892461931334, "grad_norm": 5.6067215406645134, "learning_rate": 2.4500000000000003e-06, "loss": 1.0559, "step": 245 }, { "epoch": 0.09306724676061666, "grad_norm": 14.632584569195519, "learning_rate": 2.4599999999999997e-06, "loss": 1.0229, "step": 246 }, { "epoch": 0.09344556890191999, "grad_norm": 6.406784955802286, "learning_rate": 2.47e-06, "loss": 1.0252, "step": 247 }, { "epoch": 0.0938238910432233, "grad_norm": 7.547314965665046, "learning_rate": 2.48e-06, "loss": 0.9838, "step": 248 }, { "epoch": 0.09420221318452662, "grad_norm": 6.44920071987235, "learning_rate": 2.4900000000000003e-06, "loss": 0.9664, "step": 249 }, { "epoch": 0.09458053532582994, "grad_norm": 5.4686676744513765, "learning_rate": 2.4999999999999998e-06, "loss": 0.9781, "step": 250 }, { "epoch": 0.09495885746713327, "grad_norm": 5.951563165398436, "learning_rate": 2.5099999999999997e-06, "loss": 0.9953, "step": 251 }, { "epoch": 0.09533717960843659, "grad_norm": 5.7316411610727105, "learning_rate": 2.52e-06, "loss": 1.0431, "step": 252 }, { "epoch": 0.0957155017497399, "grad_norm": 4.90373215304178, "learning_rate": 2.5299999999999995e-06, "loss": 0.9738, "step": 253 }, { "epoch": 0.09609382389104322, "grad_norm": 4.018027173598048, "learning_rate": 2.54e-06, "loss": 1.0113, "step": 254 }, { "epoch": 0.09647214603234654, "grad_norm": 6.869682846334475, "learning_rate": 2.5499999999999997e-06, "loss": 0.9812, "step": 255 }, { "epoch": 0.09685046817364987, "grad_norm": 5.959477622367862, "learning_rate": 2.56e-06, "loss": 1.0031, "step": 256 }, { "epoch": 0.09722879031495318, "grad_norm": 4.231167141984737, "learning_rate": 2.5699999999999995e-06, "loss": 1.0319, "step": 257 }, { "epoch": 0.0976071124562565, "grad_norm": 6.714523011394094, "learning_rate": 2.58e-06, "loss": 0.9851, "step": 258 }, { "epoch": 0.09798543459755982, "grad_norm": 6.020515136070658, "learning_rate": 2.5899999999999998e-06, "loss": 0.9782, "step": 259 }, { "epoch": 0.09836375673886315, "grad_norm": 4.681331319695956, "learning_rate": 2.6e-06, "loss": 1.014, "step": 260 }, { "epoch": 0.09874207888016646, "grad_norm": 7.4305112606450905, "learning_rate": 2.6099999999999996e-06, "loss": 0.9751, "step": 261 }, { "epoch": 0.09912040102146978, "grad_norm": 3.819753600694035, "learning_rate": 2.62e-06, "loss": 0.968, "step": 262 }, { "epoch": 0.0994987231627731, "grad_norm": 5.789415532330102, "learning_rate": 2.63e-06, "loss": 0.9529, "step": 263 }, { "epoch": 0.09987704530407641, "grad_norm": 4.539898474801753, "learning_rate": 2.64e-06, "loss": 0.978, "step": 264 }, { "epoch": 0.10025536744537974, "grad_norm": 3.2389391663703306, "learning_rate": 2.6499999999999996e-06, "loss": 0.9833, "step": 265 }, { "epoch": 0.10063368958668306, "grad_norm": 5.4718084763112556, "learning_rate": 2.66e-06, "loss": 0.9714, "step": 266 }, { "epoch": 0.10063368958668306, "eval_loss": 0.9851981997489929, "eval_runtime": 27.2115, "eval_samples_per_second": 32.523, "eval_steps_per_second": 1.029, "step": 266 }, { "epoch": 0.10063368958668306, "eval_bench_accuracy_arc_challenge": 0.29285714285714287, "eval_bench_accuracy_hellaswag": 0.215, "eval_bench_accuracy_mmlu": 0.3826086956521739, "eval_bench_average_accuracy": 0.29682194616977225, "eval_bench_loss": 6.3663490696957235, "eval_bench_total_accuracy": 0.2813186813186813, "step": 266 }, { "epoch": 0.10101201172798638, "grad_norm": 4.736473735176666, "learning_rate": 2.67e-06, "loss": 1.0245, "step": 267 }, { "epoch": 0.1013903338692897, "grad_norm": 2.927740836124029, "learning_rate": 2.68e-06, "loss": 0.9906, "step": 268 }, { "epoch": 0.10176865601059303, "grad_norm": 4.622383990826824, "learning_rate": 2.6899999999999997e-06, "loss": 0.9679, "step": 269 }, { "epoch": 0.10214697815189634, "grad_norm": 3.8746535383849836, "learning_rate": 2.7e-06, "loss": 0.9211, "step": 270 }, { "epoch": 0.10252530029319966, "grad_norm": 4.361727224982868, "learning_rate": 2.71e-06, "loss": 0.9779, "step": 271 }, { "epoch": 0.10290362243450298, "grad_norm": 3.2847575684010795, "learning_rate": 2.7200000000000002e-06, "loss": 0.969, "step": 272 }, { "epoch": 0.1032819445758063, "grad_norm": 2.946259099361567, "learning_rate": 2.7299999999999997e-06, "loss": 0.9374, "step": 273 }, { "epoch": 0.10366026671710962, "grad_norm": 3.5163454504687364, "learning_rate": 2.74e-06, "loss": 0.9809, "step": 274 }, { "epoch": 0.10403858885841294, "grad_norm": 4.1448737340815045, "learning_rate": 2.75e-06, "loss": 0.9816, "step": 275 }, { "epoch": 0.10441691099971626, "grad_norm": 3.345900089125294, "learning_rate": 2.76e-06, "loss": 0.94, "step": 276 }, { "epoch": 0.10479523314101957, "grad_norm": 4.756231356260067, "learning_rate": 2.7699999999999997e-06, "loss": 0.9948, "step": 277 }, { "epoch": 0.1051735552823229, "grad_norm": 3.395795830645774, "learning_rate": 2.7799999999999996e-06, "loss": 0.9852, "step": 278 }, { "epoch": 0.10555187742362622, "grad_norm": 3.7361359597792085, "learning_rate": 2.79e-06, "loss": 0.9705, "step": 279 }, { "epoch": 0.10593019956492954, "grad_norm": 2.9021780470974536, "learning_rate": 2.8e-06, "loss": 0.9517, "step": 280 }, { "epoch": 0.10630852170623285, "grad_norm": 3.3140561096891408, "learning_rate": 2.8099999999999998e-06, "loss": 0.9518, "step": 281 }, { "epoch": 0.10668684384753618, "grad_norm": 4.955772041684827, "learning_rate": 2.8199999999999997e-06, "loss": 0.949, "step": 282 }, { "epoch": 0.1070651659888395, "grad_norm": 2.7495737336593447, "learning_rate": 2.83e-06, "loss": 0.9637, "step": 283 }, { "epoch": 0.10744348813014282, "grad_norm": 5.5808851538998745, "learning_rate": 2.84e-06, "loss": 0.9149, "step": 284 }, { "epoch": 0.10782181027144613, "grad_norm": 3.2461608503776582, "learning_rate": 2.85e-06, "loss": 0.9562, "step": 285 }, { "epoch": 0.10820013241274945, "grad_norm": 3.016464443847612, "learning_rate": 2.8599999999999997e-06, "loss": 0.9635, "step": 286 }, { "epoch": 0.10857845455405278, "grad_norm": 3.1653672708590936, "learning_rate": 2.87e-06, "loss": 1.0064, "step": 287 }, { "epoch": 0.1089567766953561, "grad_norm": 2.1243065072255907, "learning_rate": 2.88e-06, "loss": 0.9279, "step": 288 }, { "epoch": 0.10933509883665941, "grad_norm": 3.4080159282806712, "learning_rate": 2.89e-06, "loss": 0.9759, "step": 289 }, { "epoch": 0.10971342097796273, "grad_norm": 2.610557409129719, "learning_rate": 2.8999999999999998e-06, "loss": 0.9787, "step": 290 }, { "epoch": 0.11009174311926606, "grad_norm": 2.2107636510154176, "learning_rate": 2.91e-06, "loss": 0.9296, "step": 291 }, { "epoch": 0.11047006526056938, "grad_norm": 4.245908140335627, "learning_rate": 2.92e-06, "loss": 0.9273, "step": 292 }, { "epoch": 0.1108483874018727, "grad_norm": 2.895847446673922, "learning_rate": 2.93e-06, "loss": 0.9383, "step": 293 }, { "epoch": 0.11122670954317601, "grad_norm": 2.704339168426421, "learning_rate": 2.94e-06, "loss": 0.9153, "step": 294 }, { "epoch": 0.11160503168447933, "grad_norm": 2.701813364341608, "learning_rate": 2.95e-06, "loss": 0.9299, "step": 295 }, { "epoch": 0.11198335382578266, "grad_norm": 2.948359459278812, "learning_rate": 2.96e-06, "loss": 0.9702, "step": 296 }, { "epoch": 0.11236167596708597, "grad_norm": 3.377595158199111, "learning_rate": 2.97e-06, "loss": 0.9554, "step": 297 }, { "epoch": 0.11273999810838929, "grad_norm": 2.5213378940105415, "learning_rate": 2.98e-06, "loss": 0.9312, "step": 298 }, { "epoch": 0.11311832024969261, "grad_norm": 4.796315482527464, "learning_rate": 2.99e-06, "loss": 0.9294, "step": 299 }, { "epoch": 0.11349664239099594, "grad_norm": 2.161917946044457, "learning_rate": 3e-06, "loss": 0.9603, "step": 300 }, { "epoch": 0.11387496453229926, "grad_norm": 4.2290402280104145, "learning_rate": 3.0099999999999996e-06, "loss": 0.9079, "step": 301 }, { "epoch": 0.11425328667360257, "grad_norm": 2.7667893528721867, "learning_rate": 3.02e-06, "loss": 0.953, "step": 302 }, { "epoch": 0.11463160881490589, "grad_norm": 9.065359561610483, "learning_rate": 3.03e-06, "loss": 0.9891, "step": 303 }, { "epoch": 0.1150099309562092, "grad_norm": 3.629194869203107, "learning_rate": 3.0399999999999997e-06, "loss": 0.9434, "step": 304 }, { "epoch": 0.11538825309751254, "grad_norm": 3.2434020969746182, "learning_rate": 3.0499999999999996e-06, "loss": 0.9289, "step": 305 }, { "epoch": 0.11576657523881585, "grad_norm": 3.266784032620147, "learning_rate": 3.06e-06, "loss": 0.941, "step": 306 }, { "epoch": 0.11614489738011917, "grad_norm": 2.2252097372145627, "learning_rate": 3.07e-06, "loss": 0.9197, "step": 307 }, { "epoch": 0.11652321952142249, "grad_norm": 2.2906797269719683, "learning_rate": 3.0799999999999997e-06, "loss": 0.9278, "step": 308 }, { "epoch": 0.11690154166272582, "grad_norm": 2.899028879345415, "learning_rate": 3.0899999999999996e-06, "loss": 0.9177, "step": 309 }, { "epoch": 0.11727986380402913, "grad_norm": 1.9374921205584867, "learning_rate": 3.1e-06, "loss": 0.9049, "step": 310 }, { "epoch": 0.11765818594533245, "grad_norm": 1.90674843142603, "learning_rate": 3.11e-06, "loss": 0.9563, "step": 311 }, { "epoch": 0.11803650808663577, "grad_norm": 1.878846884674951, "learning_rate": 3.1199999999999998e-06, "loss": 0.9139, "step": 312 }, { "epoch": 0.1184148302279391, "grad_norm": 1.8411547245015762, "learning_rate": 3.1299999999999997e-06, "loss": 0.947, "step": 313 }, { "epoch": 0.11879315236924241, "grad_norm": 1.6495211524540856, "learning_rate": 3.14e-06, "loss": 0.8994, "step": 314 }, { "epoch": 0.11917147451054573, "grad_norm": 1.979339834494396, "learning_rate": 3.15e-06, "loss": 0.9425, "step": 315 }, { "epoch": 0.11954979665184905, "grad_norm": 1.6881739152797177, "learning_rate": 3.16e-06, "loss": 0.9079, "step": 316 }, { "epoch": 0.11992811879315236, "grad_norm": 1.7476621404963093, "learning_rate": 3.1699999999999997e-06, "loss": 0.9342, "step": 317 }, { "epoch": 0.1203064409344557, "grad_norm": 1.7825714782443438, "learning_rate": 3.18e-06, "loss": 0.9736, "step": 318 }, { "epoch": 0.12068476307575901, "grad_norm": 1.7904157984440023, "learning_rate": 3.19e-06, "loss": 0.8904, "step": 319 }, { "epoch": 0.12106308521706233, "grad_norm": 1.8488826023075036, "learning_rate": 3.2e-06, "loss": 0.9374, "step": 320 }, { "epoch": 0.12144140735836564, "grad_norm": 1.7466001202181465, "learning_rate": 3.2099999999999998e-06, "loss": 0.9506, "step": 321 }, { "epoch": 0.12181972949966897, "grad_norm": 1.9022275763429817, "learning_rate": 3.22e-06, "loss": 0.9452, "step": 322 }, { "epoch": 0.12219805164097229, "grad_norm": 1.62671365850624, "learning_rate": 3.23e-06, "loss": 0.9063, "step": 323 }, { "epoch": 0.12257637378227561, "grad_norm": 1.537323535673334, "learning_rate": 3.24e-06, "loss": 0.892, "step": 324 }, { "epoch": 0.12295469592357892, "grad_norm": 1.6088280546082747, "learning_rate": 3.25e-06, "loss": 0.9055, "step": 325 }, { "epoch": 0.12333301806488224, "grad_norm": 1.754864511511676, "learning_rate": 3.2599999999999997e-06, "loss": 0.9982, "step": 326 }, { "epoch": 0.12371134020618557, "grad_norm": 1.7110520395582398, "learning_rate": 3.27e-06, "loss": 0.8869, "step": 327 }, { "epoch": 0.12408966234748889, "grad_norm": 2.2210658284362976, "learning_rate": 3.2799999999999995e-06, "loss": 0.9, "step": 328 }, { "epoch": 0.1244679844887922, "grad_norm": 2.0718951481844337, "learning_rate": 3.29e-06, "loss": 0.9474, "step": 329 }, { "epoch": 0.12484630663009552, "grad_norm": 1.6483777638825354, "learning_rate": 3.2999999999999997e-06, "loss": 0.9193, "step": 330 }, { "epoch": 0.12522462877139884, "grad_norm": 1.8408500351694481, "learning_rate": 3.31e-06, "loss": 0.9331, "step": 331 }, { "epoch": 0.12560295091270215, "grad_norm": 1.5886399601274244, "learning_rate": 3.3199999999999996e-06, "loss": 0.9181, "step": 332 }, { "epoch": 0.1259812730540055, "grad_norm": 1.5415700759277726, "learning_rate": 3.33e-06, "loss": 0.9078, "step": 333 }, { "epoch": 0.12635959519530882, "grad_norm": 1.5699378541238653, "learning_rate": 3.3399999999999998e-06, "loss": 0.9415, "step": 334 }, { "epoch": 0.12673791733661213, "grad_norm": 1.4355378270145513, "learning_rate": 3.35e-06, "loss": 0.9328, "step": 335 }, { "epoch": 0.12711623947791545, "grad_norm": 1.4472036059899498, "learning_rate": 3.3599999999999996e-06, "loss": 0.9235, "step": 336 }, { "epoch": 0.12749456161921877, "grad_norm": 1.493466705425371, "learning_rate": 3.37e-06, "loss": 0.917, "step": 337 }, { "epoch": 0.12787288376052208, "grad_norm": 1.725222957788955, "learning_rate": 3.38e-06, "loss": 0.9229, "step": 338 }, { "epoch": 0.1282512059018254, "grad_norm": 1.829546156665469, "learning_rate": 3.39e-06, "loss": 0.9199, "step": 339 }, { "epoch": 0.12862952804312872, "grad_norm": 1.562404556848645, "learning_rate": 3.3999999999999996e-06, "loss": 0.9258, "step": 340 }, { "epoch": 0.12900785018443203, "grad_norm": 1.5503184849860385, "learning_rate": 3.41e-06, "loss": 0.9056, "step": 341 }, { "epoch": 0.12938617232573538, "grad_norm": 2.093643266825353, "learning_rate": 3.42e-06, "loss": 0.9151, "step": 342 }, { "epoch": 0.1297644944670387, "grad_norm": 1.5470351610527242, "learning_rate": 3.43e-06, "loss": 0.9295, "step": 343 }, { "epoch": 0.130142816608342, "grad_norm": 1.6415927498606424, "learning_rate": 3.4399999999999997e-06, "loss": 0.9227, "step": 344 }, { "epoch": 0.13052113874964533, "grad_norm": 1.501364967749395, "learning_rate": 3.45e-06, "loss": 0.9196, "step": 345 }, { "epoch": 0.13089946089094864, "grad_norm": 1.4667926955996313, "learning_rate": 3.46e-06, "loss": 0.9875, "step": 346 }, { "epoch": 0.13127778303225196, "grad_norm": 1.4015397895960147, "learning_rate": 3.4700000000000002e-06, "loss": 0.9174, "step": 347 }, { "epoch": 0.13165610517355528, "grad_norm": 1.6317901839112616, "learning_rate": 3.4799999999999997e-06, "loss": 0.9022, "step": 348 }, { "epoch": 0.1320344273148586, "grad_norm": 1.5495030641920218, "learning_rate": 3.49e-06, "loss": 0.9056, "step": 349 }, { "epoch": 0.1324127494561619, "grad_norm": 1.4169162437828007, "learning_rate": 3.5e-06, "loss": 0.9125, "step": 350 }, { "epoch": 0.13279107159746525, "grad_norm": 1.5269510878366184, "learning_rate": 3.5099999999999994e-06, "loss": 0.9325, "step": 351 }, { "epoch": 0.13316939373876857, "grad_norm": 1.4845731562408333, "learning_rate": 3.5199999999999998e-06, "loss": 0.9119, "step": 352 }, { "epoch": 0.1335477158800719, "grad_norm": 1.2998342684154016, "learning_rate": 3.5299999999999997e-06, "loss": 0.8989, "step": 353 }, { "epoch": 0.1339260380213752, "grad_norm": 1.4867481861923495, "learning_rate": 3.54e-06, "loss": 0.9201, "step": 354 }, { "epoch": 0.13430436016267852, "grad_norm": 1.4212824059163913, "learning_rate": 3.5499999999999995e-06, "loss": 0.9288, "step": 355 }, { "epoch": 0.13468268230398184, "grad_norm": 1.3588961307618976, "learning_rate": 3.56e-06, "loss": 0.9117, "step": 356 }, { "epoch": 0.13506100444528515, "grad_norm": 1.4097313807539793, "learning_rate": 3.5699999999999997e-06, "loss": 0.9139, "step": 357 }, { "epoch": 0.13543932658658847, "grad_norm": 1.490782064831479, "learning_rate": 3.58e-06, "loss": 0.938, "step": 358 }, { "epoch": 0.1358176487278918, "grad_norm": 1.2930048652835795, "learning_rate": 3.5899999999999995e-06, "loss": 0.9023, "step": 359 }, { "epoch": 0.13619597086919513, "grad_norm": 1.824182436515982, "learning_rate": 3.6e-06, "loss": 0.9343, "step": 360 }, { "epoch": 0.13657429301049845, "grad_norm": 1.4837219324976698, "learning_rate": 3.6099999999999997e-06, "loss": 0.9418, "step": 361 }, { "epoch": 0.13695261515180177, "grad_norm": 1.3718729917310193, "learning_rate": 3.62e-06, "loss": 0.9231, "step": 362 }, { "epoch": 0.13733093729310508, "grad_norm": 1.3644818822127356, "learning_rate": 3.6299999999999995e-06, "loss": 0.9093, "step": 363 }, { "epoch": 0.1377092594344084, "grad_norm": 1.4274881326706697, "learning_rate": 3.64e-06, "loss": 0.9077, "step": 364 }, { "epoch": 0.13808758157571172, "grad_norm": 1.3169195252885812, "learning_rate": 3.6499999999999998e-06, "loss": 0.8772, "step": 365 }, { "epoch": 0.13846590371701503, "grad_norm": 1.3505673564506786, "learning_rate": 3.66e-06, "loss": 0.8729, "step": 366 }, { "epoch": 0.13884422585831835, "grad_norm": 1.3728815922981648, "learning_rate": 3.6699999999999996e-06, "loss": 0.91, "step": 367 }, { "epoch": 0.13922254799962167, "grad_norm": 1.4225979847364822, "learning_rate": 3.68e-06, "loss": 0.8862, "step": 368 }, { "epoch": 0.139600870140925, "grad_norm": 1.3363118705656714, "learning_rate": 3.69e-06, "loss": 0.9322, "step": 369 }, { "epoch": 0.13997919228222833, "grad_norm": 1.318614371056809, "learning_rate": 3.7e-06, "loss": 0.926, "step": 370 }, { "epoch": 0.14035751442353164, "grad_norm": 1.330484253084181, "learning_rate": 3.7099999999999996e-06, "loss": 0.9456, "step": 371 }, { "epoch": 0.14073583656483496, "grad_norm": 1.3318506320691512, "learning_rate": 3.72e-06, "loss": 0.9017, "step": 372 }, { "epoch": 0.14111415870613828, "grad_norm": 1.3759434761704756, "learning_rate": 3.73e-06, "loss": 0.8881, "step": 373 }, { "epoch": 0.1414924808474416, "grad_norm": 1.3957619030952084, "learning_rate": 3.74e-06, "loss": 0.9121, "step": 374 }, { "epoch": 0.1418708029887449, "grad_norm": 1.3427799016571502, "learning_rate": 3.7499999999999997e-06, "loss": 0.9106, "step": 375 }, { "epoch": 0.14224912513004823, "grad_norm": 44.30080368963616, "learning_rate": 3.7599999999999996e-06, "loss": 0.8911, "step": 376 }, { "epoch": 0.14262744727135154, "grad_norm": 2.2669972347416127, "learning_rate": 3.77e-06, "loss": 0.933, "step": 377 }, { "epoch": 0.1430057694126549, "grad_norm": 1.4829201626961606, "learning_rate": 3.78e-06, "loss": 0.901, "step": 378 }, { "epoch": 0.1433840915539582, "grad_norm": 4.064663928049432, "learning_rate": 3.7899999999999997e-06, "loss": 0.8942, "step": 379 }, { "epoch": 0.14376241369526152, "grad_norm": 1.8169275430345828, "learning_rate": 3.7999999999999996e-06, "loss": 0.88, "step": 380 }, { "epoch": 0.14414073583656484, "grad_norm": 1.903257571166488, "learning_rate": 3.81e-06, "loss": 0.9286, "step": 381 }, { "epoch": 0.14451905797786815, "grad_norm": 1.662557610937424, "learning_rate": 3.82e-06, "loss": 0.8947, "step": 382 }, { "epoch": 0.14489738011917147, "grad_norm": 1.3504615763712993, "learning_rate": 3.83e-06, "loss": 0.9081, "step": 383 }, { "epoch": 0.1452757022604748, "grad_norm": 2.083053759282353, "learning_rate": 3.84e-06, "loss": 0.9229, "step": 384 }, { "epoch": 0.1456540244017781, "grad_norm": 1.5724819369725127, "learning_rate": 3.8499999999999996e-06, "loss": 0.9019, "step": 385 }, { "epoch": 0.14603234654308142, "grad_norm": 1.2833291006046557, "learning_rate": 3.8599999999999995e-06, "loss": 0.8943, "step": 386 }, { "epoch": 0.14641066868438476, "grad_norm": 1.6810072820257926, "learning_rate": 3.87e-06, "loss": 0.9469, "step": 387 }, { "epoch": 0.14678899082568808, "grad_norm": 1.462137670239198, "learning_rate": 3.88e-06, "loss": 0.885, "step": 388 }, { "epoch": 0.1471673129669914, "grad_norm": 1.3544773507596952, "learning_rate": 3.89e-06, "loss": 0.9223, "step": 389 }, { "epoch": 0.14754563510829471, "grad_norm": 1.305788748108731, "learning_rate": 3.9e-06, "loss": 0.9085, "step": 390 }, { "epoch": 0.14792395724959803, "grad_norm": 1.4728433076805145, "learning_rate": 3.91e-06, "loss": 0.9111, "step": 391 }, { "epoch": 0.14830227939090135, "grad_norm": 1.3023289374881166, "learning_rate": 3.92e-06, "loss": 0.9082, "step": 392 }, { "epoch": 0.14868060153220466, "grad_norm": 1.528856941817902, "learning_rate": 3.93e-06, "loss": 0.8583, "step": 393 }, { "epoch": 0.14905892367350798, "grad_norm": 1.2279025499674738, "learning_rate": 3.9399999999999995e-06, "loss": 0.8943, "step": 394 }, { "epoch": 0.1494372458148113, "grad_norm": 1.5480907504889059, "learning_rate": 3.95e-06, "loss": 0.858, "step": 395 }, { "epoch": 0.14981556795611464, "grad_norm": 1.3146063824478018, "learning_rate": 3.96e-06, "loss": 0.8618, "step": 396 }, { "epoch": 0.15019389009741796, "grad_norm": 1.334057857690303, "learning_rate": 3.97e-06, "loss": 0.9243, "step": 397 }, { "epoch": 0.15057221223872128, "grad_norm": 1.3866128005645164, "learning_rate": 3.98e-06, "loss": 0.9274, "step": 398 }, { "epoch": 0.1509505343800246, "grad_norm": 1.2955294219171367, "learning_rate": 3.99e-06, "loss": 0.9173, "step": 399 }, { "epoch": 0.1509505343800246, "eval_loss": 0.9027320742607117, "eval_runtime": 27.0581, "eval_samples_per_second": 32.707, "eval_steps_per_second": 1.035, "step": 399 }, { "epoch": 0.1509505343800246, "eval_bench_accuracy_arc_challenge": 0.24285714285714285, "eval_bench_accuracy_hellaswag": 0.24, "eval_bench_accuracy_mmlu": 0.3739130434782609, "eval_bench_average_accuracy": 0.2855900621118012, "eval_bench_loss": 4.885084721080044, "eval_bench_total_accuracy": 0.27472527472527475, "step": 399 }, { "epoch": 0.1513288565213279, "grad_norm": 1.4867956471987611, "learning_rate": 4e-06, "loss": 0.8442, "step": 400 }, { "epoch": 0.15170717866263123, "grad_norm": 1.4418482940385888, "learning_rate": 4.01e-06, "loss": 0.8851, "step": 401 }, { "epoch": 0.15208550080393454, "grad_norm": 1.2367816437008439, "learning_rate": 4.02e-06, "loss": 0.9016, "step": 402 }, { "epoch": 0.15246382294523786, "grad_norm": 1.3381669970164036, "learning_rate": 4.03e-06, "loss": 0.8967, "step": 403 }, { "epoch": 0.1528421450865412, "grad_norm": 1.178040710244701, "learning_rate": 4.0399999999999994e-06, "loss": 0.9052, "step": 404 }, { "epoch": 0.15322046722784452, "grad_norm": 1.354680203607332, "learning_rate": 4.049999999999999e-06, "loss": 0.916, "step": 405 }, { "epoch": 0.15359878936914784, "grad_norm": 1.2478760852613116, "learning_rate": 4.059999999999999e-06, "loss": 0.8918, "step": 406 }, { "epoch": 0.15397711151045115, "grad_norm": 1.3580886429686791, "learning_rate": 4.07e-06, "loss": 0.8769, "step": 407 }, { "epoch": 0.15435543365175447, "grad_norm": 1.4849252692119392, "learning_rate": 4.08e-06, "loss": 0.8985, "step": 408 }, { "epoch": 0.1547337557930578, "grad_norm": 1.234446053198778, "learning_rate": 4.09e-06, "loss": 0.8681, "step": 409 }, { "epoch": 0.1551120779343611, "grad_norm": 1.4907001456714162, "learning_rate": 4.1e-06, "loss": 0.9035, "step": 410 }, { "epoch": 0.15549040007566442, "grad_norm": 1.1935520171507346, "learning_rate": 4.1100000000000005e-06, "loss": 0.8939, "step": 411 }, { "epoch": 0.15586872221696774, "grad_norm": 1.3431797561411594, "learning_rate": 4.1199999999999995e-06, "loss": 0.8892, "step": 412 }, { "epoch": 0.15624704435827108, "grad_norm": 1.1858701499867044, "learning_rate": 4.129999999999999e-06, "loss": 0.8952, "step": 413 }, { "epoch": 0.1566253664995744, "grad_norm": 1.3160462921208504, "learning_rate": 4.139999999999999e-06, "loss": 0.9104, "step": 414 }, { "epoch": 0.15700368864087771, "grad_norm": 1.205303163621962, "learning_rate": 4.15e-06, "loss": 0.8989, "step": 415 }, { "epoch": 0.15738201078218103, "grad_norm": 1.2116662309617274, "learning_rate": 4.16e-06, "loss": 0.9178, "step": 416 }, { "epoch": 0.15776033292348435, "grad_norm": 1.1758637546414648, "learning_rate": 4.17e-06, "loss": 0.8792, "step": 417 }, { "epoch": 0.15813865506478766, "grad_norm": 1.2552462548629688, "learning_rate": 4.18e-06, "loss": 0.8981, "step": 418 }, { "epoch": 0.15851697720609098, "grad_norm": 1.206264514397755, "learning_rate": 4.1900000000000005e-06, "loss": 0.9058, "step": 419 }, { "epoch": 0.1588952993473943, "grad_norm": 1.2231014501429258, "learning_rate": 4.2e-06, "loss": 0.899, "step": 420 }, { "epoch": 0.15927362148869761, "grad_norm": 1.2120070273790158, "learning_rate": 4.2099999999999995e-06, "loss": 0.8449, "step": 421 }, { "epoch": 0.15965194363000096, "grad_norm": 1.225434870357441, "learning_rate": 4.219999999999999e-06, "loss": 0.8925, "step": 422 }, { "epoch": 0.16003026577130428, "grad_norm": 1.2700536143173544, "learning_rate": 4.23e-06, "loss": 0.8948, "step": 423 }, { "epoch": 0.1604085879126076, "grad_norm": 1.327617668860312, "learning_rate": 4.24e-06, "loss": 0.8808, "step": 424 }, { "epoch": 0.1607869100539109, "grad_norm": 1.2286005573930583, "learning_rate": 4.25e-06, "loss": 0.8885, "step": 425 }, { "epoch": 0.16116523219521423, "grad_norm": 1.265158345195646, "learning_rate": 4.26e-06, "loss": 0.8973, "step": 426 }, { "epoch": 0.16154355433651754, "grad_norm": 1.2113247771231779, "learning_rate": 4.27e-06, "loss": 0.88, "step": 427 }, { "epoch": 0.16192187647782086, "grad_norm": 1.1981923822069018, "learning_rate": 4.28e-06, "loss": 0.8812, "step": 428 }, { "epoch": 0.16230019861912418, "grad_norm": 1.269210905108754, "learning_rate": 4.29e-06, "loss": 0.951, "step": 429 }, { "epoch": 0.1626785207604275, "grad_norm": 1.270040077896289, "learning_rate": 4.2999999999999995e-06, "loss": 0.8502, "step": 430 }, { "epoch": 0.16305684290173084, "grad_norm": 1.2459835235482208, "learning_rate": 4.309999999999999e-06, "loss": 0.9249, "step": 431 }, { "epoch": 0.16343516504303415, "grad_norm": 1.2065849160511677, "learning_rate": 4.32e-06, "loss": 0.8569, "step": 432 }, { "epoch": 0.16381348718433747, "grad_norm": 1.3240957525319628, "learning_rate": 4.33e-06, "loss": 0.8378, "step": 433 }, { "epoch": 0.1641918093256408, "grad_norm": 1.308494624204772, "learning_rate": 4.34e-06, "loss": 0.8853, "step": 434 }, { "epoch": 0.1645701314669441, "grad_norm": 1.2876226830148083, "learning_rate": 4.35e-06, "loss": 0.8999, "step": 435 }, { "epoch": 0.16494845360824742, "grad_norm": 1.3895344761060464, "learning_rate": 4.36e-06, "loss": 0.8995, "step": 436 }, { "epoch": 0.16532677574955074, "grad_norm": 1.2397074052657744, "learning_rate": 4.37e-06, "loss": 0.8787, "step": 437 }, { "epoch": 0.16570509789085405, "grad_norm": 1.2286411029399464, "learning_rate": 4.3799999999999996e-06, "loss": 0.8968, "step": 438 }, { "epoch": 0.16608342003215737, "grad_norm": 1.231038186520652, "learning_rate": 4.3899999999999995e-06, "loss": 0.8781, "step": 439 }, { "epoch": 0.16646174217346071, "grad_norm": 1.2138487844408843, "learning_rate": 4.4e-06, "loss": 0.8698, "step": 440 }, { "epoch": 0.16684006431476403, "grad_norm": 1.3027744892443913, "learning_rate": 4.41e-06, "loss": 0.9253, "step": 441 }, { "epoch": 0.16721838645606735, "grad_norm": 1.2467659827353952, "learning_rate": 4.42e-06, "loss": 0.9121, "step": 442 }, { "epoch": 0.16759670859737066, "grad_norm": 1.1589200132022377, "learning_rate": 4.43e-06, "loss": 0.8803, "step": 443 }, { "epoch": 0.16797503073867398, "grad_norm": 1.2200621136986902, "learning_rate": 4.44e-06, "loss": 0.9079, "step": 444 }, { "epoch": 0.1683533528799773, "grad_norm": 1.1747935123553643, "learning_rate": 4.45e-06, "loss": 0.8766, "step": 445 }, { "epoch": 0.1687316750212806, "grad_norm": 1.1865214460906777, "learning_rate": 4.46e-06, "loss": 0.9068, "step": 446 }, { "epoch": 0.16910999716258393, "grad_norm": 1.2579950961305297, "learning_rate": 4.4699999999999996e-06, "loss": 0.8815, "step": 447 }, { "epoch": 0.16948831930388725, "grad_norm": 1.226665097174107, "learning_rate": 4.48e-06, "loss": 0.9327, "step": 448 }, { "epoch": 0.1698666414451906, "grad_norm": 1.1931395850546989, "learning_rate": 4.49e-06, "loss": 0.8796, "step": 449 }, { "epoch": 0.1702449635864939, "grad_norm": 1.202501530652917, "learning_rate": 4.5e-06, "loss": 0.8931, "step": 450 }, { "epoch": 0.17062328572779722, "grad_norm": 1.1807025967685065, "learning_rate": 4.509999999999999e-06, "loss": 0.8887, "step": 451 }, { "epoch": 0.17100160786910054, "grad_norm": 1.219222521929812, "learning_rate": 4.519999999999999e-06, "loss": 0.8999, "step": 452 }, { "epoch": 0.17137993001040386, "grad_norm": 1.234613051649134, "learning_rate": 4.53e-06, "loss": 0.8439, "step": 453 }, { "epoch": 0.17175825215170717, "grad_norm": 1.2268814413232634, "learning_rate": 4.54e-06, "loss": 0.8679, "step": 454 }, { "epoch": 0.1721365742930105, "grad_norm": 1.2687792576706662, "learning_rate": 4.55e-06, "loss": 0.9137, "step": 455 }, { "epoch": 0.1725148964343138, "grad_norm": 1.259597511238193, "learning_rate": 4.5599999999999995e-06, "loss": 0.8929, "step": 456 }, { "epoch": 0.17289321857561712, "grad_norm": 1.1601209722807053, "learning_rate": 4.57e-06, "loss": 0.8989, "step": 457 }, { "epoch": 0.17327154071692047, "grad_norm": 1.1337571129482695, "learning_rate": 4.58e-06, "loss": 0.8867, "step": 458 }, { "epoch": 0.17364986285822379, "grad_norm": 1.2315099804928107, "learning_rate": 4.589999999999999e-06, "loss": 0.8766, "step": 459 }, { "epoch": 0.1740281849995271, "grad_norm": 1.1590598116825013, "learning_rate": 4.599999999999999e-06, "loss": 0.8996, "step": 460 }, { "epoch": 0.17440650714083042, "grad_norm": 1.2223724961641853, "learning_rate": 4.61e-06, "loss": 0.8885, "step": 461 }, { "epoch": 0.17478482928213374, "grad_norm": 1.2563659855924223, "learning_rate": 4.62e-06, "loss": 0.9316, "step": 462 }, { "epoch": 0.17516315142343705, "grad_norm": 1.2219308373205684, "learning_rate": 4.63e-06, "loss": 0.9402, "step": 463 }, { "epoch": 0.17554147356474037, "grad_norm": 1.2529933281060042, "learning_rate": 4.64e-06, "loss": 0.8425, "step": 464 }, { "epoch": 0.17591979570604369, "grad_norm": 1.1519152308086784, "learning_rate": 4.65e-06, "loss": 0.8335, "step": 465 }, { "epoch": 0.176298117847347, "grad_norm": 1.1993447663063845, "learning_rate": 4.66e-06, "loss": 0.8423, "step": 466 }, { "epoch": 0.17667643998865035, "grad_norm": 1.2393551988442821, "learning_rate": 4.669999999999999e-06, "loss": 0.8766, "step": 467 }, { "epoch": 0.17705476212995366, "grad_norm": 1.1568166146377072, "learning_rate": 4.679999999999999e-06, "loss": 0.913, "step": 468 }, { "epoch": 0.17743308427125698, "grad_norm": 1.2535994832897241, "learning_rate": 4.69e-06, "loss": 0.8611, "step": 469 }, { "epoch": 0.1778114064125603, "grad_norm": 1.2581510292576754, "learning_rate": 4.7e-06, "loss": 0.852, "step": 470 }, { "epoch": 0.1781897285538636, "grad_norm": 1.185843568335289, "learning_rate": 4.71e-06, "loss": 0.8712, "step": 471 }, { "epoch": 0.17856805069516693, "grad_norm": 1.1762961141384334, "learning_rate": 4.72e-06, "loss": 0.8848, "step": 472 }, { "epoch": 0.17894637283647025, "grad_norm": 1.2378038953878985, "learning_rate": 4.7300000000000005e-06, "loss": 0.89, "step": 473 }, { "epoch": 0.17932469497777356, "grad_norm": 1.2303598909876003, "learning_rate": 4.74e-06, "loss": 0.9019, "step": 474 }, { "epoch": 0.1797030171190769, "grad_norm": 1.3055168080029775, "learning_rate": 4.749999999999999e-06, "loss": 0.8886, "step": 475 }, { "epoch": 0.18008133926038022, "grad_norm": 1.263816208541402, "learning_rate": 4.759999999999999e-06, "loss": 0.8934, "step": 476 }, { "epoch": 0.18045966140168354, "grad_norm": 1.2304160263194301, "learning_rate": 4.769999999999999e-06, "loss": 0.8334, "step": 477 }, { "epoch": 0.18083798354298686, "grad_norm": 1.16427739617554, "learning_rate": 4.78e-06, "loss": 0.8933, "step": 478 }, { "epoch": 0.18121630568429017, "grad_norm": 1.2928340654165948, "learning_rate": 4.79e-06, "loss": 0.9091, "step": 479 }, { "epoch": 0.1815946278255935, "grad_norm": 1.2237270548636812, "learning_rate": 4.8e-06, "loss": 0.8894, "step": 480 }, { "epoch": 0.1819729499668968, "grad_norm": 1.2973745239107866, "learning_rate": 4.81e-06, "loss": 0.8827, "step": 481 }, { "epoch": 0.18235127210820012, "grad_norm": 1.2192171355443393, "learning_rate": 4.8200000000000004e-06, "loss": 0.842, "step": 482 }, { "epoch": 0.18272959424950344, "grad_norm": 1.1825464816429376, "learning_rate": 4.8299999999999995e-06, "loss": 0.8974, "step": 483 }, { "epoch": 0.18310791639080679, "grad_norm": 1.2357877717915002, "learning_rate": 4.839999999999999e-06, "loss": 0.8713, "step": 484 }, { "epoch": 0.1834862385321101, "grad_norm": 1.2724832467234655, "learning_rate": 4.849999999999999e-06, "loss": 0.8916, "step": 485 }, { "epoch": 0.18386456067341342, "grad_norm": 1.2402819428437333, "learning_rate": 4.86e-06, "loss": 0.9006, "step": 486 }, { "epoch": 0.18424288281471674, "grad_norm": 1.253080289206958, "learning_rate": 4.87e-06, "loss": 0.8552, "step": 487 }, { "epoch": 0.18462120495602005, "grad_norm": 1.20114987062819, "learning_rate": 4.88e-06, "loss": 0.8646, "step": 488 }, { "epoch": 0.18499952709732337, "grad_norm": 1.2698388666443412, "learning_rate": 4.89e-06, "loss": 0.9058, "step": 489 }, { "epoch": 0.18537784923862669, "grad_norm": 1.255138008138629, "learning_rate": 4.9000000000000005e-06, "loss": 0.9045, "step": 490 }, { "epoch": 0.18575617137993, "grad_norm": 1.173366935458501, "learning_rate": 4.91e-06, "loss": 0.8653, "step": 491 }, { "epoch": 0.18613449352123332, "grad_norm": 1.2544859383454867, "learning_rate": 4.9199999999999995e-06, "loss": 0.8577, "step": 492 }, { "epoch": 0.18651281566253666, "grad_norm": 1.1732808685881084, "learning_rate": 4.929999999999999e-06, "loss": 0.8551, "step": 493 }, { "epoch": 0.18689113780383998, "grad_norm": 1.2265764031917046, "learning_rate": 4.94e-06, "loss": 0.8726, "step": 494 }, { "epoch": 0.1872694599451433, "grad_norm": 1.2234524388802157, "learning_rate": 4.95e-06, "loss": 0.8833, "step": 495 }, { "epoch": 0.1876477820864466, "grad_norm": 1.2488343163013593, "learning_rate": 4.96e-06, "loss": 0.8704, "step": 496 }, { "epoch": 0.18802610422774993, "grad_norm": 1.1667370629188312, "learning_rate": 4.97e-06, "loss": 0.8637, "step": 497 }, { "epoch": 0.18840442636905325, "grad_norm": 1.1300202443780525, "learning_rate": 4.980000000000001e-06, "loss": 0.8222, "step": 498 }, { "epoch": 0.18878274851035656, "grad_norm": 1.2105094043051028, "learning_rate": 4.99e-06, "loss": 0.8172, "step": 499 }, { "epoch": 0.18916107065165988, "grad_norm": 1.147109513607525, "learning_rate": 4.9999999999999996e-06, "loss": 0.8718, "step": 500 }, { "epoch": 0.1895393927929632, "grad_norm": 1.186254501579871, "learning_rate": 5.0099999999999995e-06, "loss": 0.8672, "step": 501 }, { "epoch": 0.18991771493426654, "grad_norm": 1.1921470006777564, "learning_rate": 5.019999999999999e-06, "loss": 0.8984, "step": 502 }, { "epoch": 0.19029603707556986, "grad_norm": 1.204441588496536, "learning_rate": 5.03e-06, "loss": 0.8933, "step": 503 }, { "epoch": 0.19067435921687317, "grad_norm": 1.176488402672726, "learning_rate": 5.04e-06, "loss": 0.8179, "step": 504 }, { "epoch": 0.1910526813581765, "grad_norm": 1.1591890939118275, "learning_rate": 5.05e-06, "loss": 0.8994, "step": 505 }, { "epoch": 0.1914310034994798, "grad_norm": 1.1844780849489716, "learning_rate": 5.059999999999999e-06, "loss": 0.9002, "step": 506 }, { "epoch": 0.19180932564078312, "grad_norm": 1.1340897482563235, "learning_rate": 5.07e-06, "loss": 0.8629, "step": 507 }, { "epoch": 0.19218764778208644, "grad_norm": 1.242695087632576, "learning_rate": 5.08e-06, "loss": 0.893, "step": 508 }, { "epoch": 0.19256596992338976, "grad_norm": 1.21618537349293, "learning_rate": 5.0899999999999995e-06, "loss": 0.8874, "step": 509 }, { "epoch": 0.19294429206469307, "grad_norm": 1.2081469798752933, "learning_rate": 5.0999999999999995e-06, "loss": 0.8672, "step": 510 }, { "epoch": 0.19332261420599642, "grad_norm": 1.1486757711757551, "learning_rate": 5.11e-06, "loss": 0.8445, "step": 511 }, { "epoch": 0.19370093634729973, "grad_norm": 1.160176382154706, "learning_rate": 5.12e-06, "loss": 0.8689, "step": 512 }, { "epoch": 0.19407925848860305, "grad_norm": 1.1842115955863446, "learning_rate": 5.13e-06, "loss": 0.887, "step": 513 }, { "epoch": 0.19445758062990637, "grad_norm": 1.1622953235550992, "learning_rate": 5.139999999999999e-06, "loss": 0.8891, "step": 514 }, { "epoch": 0.19483590277120968, "grad_norm": 1.2278834007146076, "learning_rate": 5.15e-06, "loss": 0.9542, "step": 515 }, { "epoch": 0.195214224912513, "grad_norm": 1.1688897803585725, "learning_rate": 5.16e-06, "loss": 0.842, "step": 516 }, { "epoch": 0.19559254705381632, "grad_norm": 1.169443235508946, "learning_rate": 5.17e-06, "loss": 0.926, "step": 517 }, { "epoch": 0.19597086919511963, "grad_norm": 1.190101722103473, "learning_rate": 5.1799999999999995e-06, "loss": 0.9012, "step": 518 }, { "epoch": 0.19634919133642295, "grad_norm": 1.1139938105404836, "learning_rate": 5.19e-06, "loss": 0.8355, "step": 519 }, { "epoch": 0.1967275134777263, "grad_norm": 1.1644272208548614, "learning_rate": 5.2e-06, "loss": 0.8508, "step": 520 }, { "epoch": 0.1971058356190296, "grad_norm": 1.188005585447595, "learning_rate": 5.21e-06, "loss": 0.8884, "step": 521 }, { "epoch": 0.19748415776033293, "grad_norm": 1.162381129570287, "learning_rate": 5.219999999999999e-06, "loss": 0.8494, "step": 522 }, { "epoch": 0.19786247990163625, "grad_norm": 1.1379792376540319, "learning_rate": 5.23e-06, "loss": 0.8427, "step": 523 }, { "epoch": 0.19824080204293956, "grad_norm": 1.163441860737916, "learning_rate": 5.24e-06, "loss": 0.8831, "step": 524 }, { "epoch": 0.19861912418424288, "grad_norm": 1.1604063632172568, "learning_rate": 5.25e-06, "loss": 0.8898, "step": 525 }, { "epoch": 0.1989974463255462, "grad_norm": 1.1325670759545932, "learning_rate": 5.26e-06, "loss": 0.8735, "step": 526 }, { "epoch": 0.1993757684668495, "grad_norm": 1.1790821072251718, "learning_rate": 5.2699999999999995e-06, "loss": 0.8343, "step": 527 }, { "epoch": 0.19975409060815283, "grad_norm": 1.1453742135606537, "learning_rate": 5.28e-06, "loss": 0.8566, "step": 528 }, { "epoch": 0.20013241274945617, "grad_norm": 1.13296207138768, "learning_rate": 5.29e-06, "loss": 0.8659, "step": 529 }, { "epoch": 0.2005107348907595, "grad_norm": 1.1666609028219261, "learning_rate": 5.299999999999999e-06, "loss": 0.8853, "step": 530 }, { "epoch": 0.2008890570320628, "grad_norm": 1.1656374685369397, "learning_rate": 5.309999999999999e-06, "loss": 0.9086, "step": 531 }, { "epoch": 0.20126737917336612, "grad_norm": 1.1343885551812507, "learning_rate": 5.32e-06, "loss": 0.8379, "step": 532 }, { "epoch": 0.20126737917336612, "eval_loss": 0.8767463564872742, "eval_runtime": 26.8872, "eval_samples_per_second": 32.915, "eval_steps_per_second": 1.041, "step": 532 }, { "epoch": 0.20126737917336612, "eval_bench_accuracy_arc_challenge": 0.24285714285714285, "eval_bench_accuracy_hellaswag": 0.275, "eval_bench_accuracy_mmlu": 0.3391304347826087, "eval_bench_average_accuracy": 0.2856625258799172, "eval_bench_loss": 5.605643824527138, "eval_bench_total_accuracy": 0.2813186813186813, "step": 532 }, { "epoch": 0.20164570131466944, "grad_norm": 1.1898287763707267, "learning_rate": 5.33e-06, "loss": 0.8633, "step": 533 }, { "epoch": 0.20202402345597276, "grad_norm": 1.2061752853772802, "learning_rate": 5.34e-06, "loss": 0.8537, "step": 534 }, { "epoch": 0.20240234559727607, "grad_norm": 1.1524730070815266, "learning_rate": 5.35e-06, "loss": 0.8658, "step": 535 }, { "epoch": 0.2027806677385794, "grad_norm": 1.2112053959243978, "learning_rate": 5.36e-06, "loss": 0.8658, "step": 536 }, { "epoch": 0.2031589898798827, "grad_norm": 1.1062007713391508, "learning_rate": 5.37e-06, "loss": 0.8695, "step": 537 }, { "epoch": 0.20353731202118605, "grad_norm": 1.1454209056836882, "learning_rate": 5.379999999999999e-06, "loss": 0.8411, "step": 538 }, { "epoch": 0.20391563416248937, "grad_norm": 1.1969213700372077, "learning_rate": 5.389999999999999e-06, "loss": 0.8262, "step": 539 }, { "epoch": 0.20429395630379268, "grad_norm": 1.1817755878296146, "learning_rate": 5.4e-06, "loss": 0.8928, "step": 540 }, { "epoch": 0.204672278445096, "grad_norm": 1.2881214697120862, "learning_rate": 5.41e-06, "loss": 0.8755, "step": 541 }, { "epoch": 0.20505060058639932, "grad_norm": 1.1803409039809667, "learning_rate": 5.42e-06, "loss": 0.8728, "step": 542 }, { "epoch": 0.20542892272770263, "grad_norm": 1.2147547833072705, "learning_rate": 5.43e-06, "loss": 0.8673, "step": 543 }, { "epoch": 0.20580724486900595, "grad_norm": 1.111022507543289, "learning_rate": 5.4400000000000004e-06, "loss": 0.8572, "step": 544 }, { "epoch": 0.20618556701030927, "grad_norm": 1.229625708529713, "learning_rate": 5.45e-06, "loss": 0.9064, "step": 545 }, { "epoch": 0.2065638891516126, "grad_norm": 1.1293738392645483, "learning_rate": 5.459999999999999e-06, "loss": 0.8504, "step": 546 }, { "epoch": 0.20694221129291593, "grad_norm": 1.1526707564326522, "learning_rate": 5.469999999999999e-06, "loss": 0.8722, "step": 547 }, { "epoch": 0.20732053343421925, "grad_norm": 1.1056906302195102, "learning_rate": 5.48e-06, "loss": 0.8253, "step": 548 }, { "epoch": 0.20769885557552256, "grad_norm": 1.1541954114677542, "learning_rate": 5.49e-06, "loss": 0.8475, "step": 549 }, { "epoch": 0.20807717771682588, "grad_norm": 1.151670600398325, "learning_rate": 5.5e-06, "loss": 0.8372, "step": 550 }, { "epoch": 0.2084554998581292, "grad_norm": 1.157820909806914, "learning_rate": 5.51e-06, "loss": 0.8595, "step": 551 }, { "epoch": 0.2088338219994325, "grad_norm": 1.1605316476134264, "learning_rate": 5.52e-06, "loss": 0.8595, "step": 552 }, { "epoch": 0.20921214414073583, "grad_norm": 1.1898854269979218, "learning_rate": 5.53e-06, "loss": 0.8499, "step": 553 }, { "epoch": 0.20959046628203915, "grad_norm": 1.1432985309555297, "learning_rate": 5.5399999999999995e-06, "loss": 0.9105, "step": 554 }, { "epoch": 0.2099687884233425, "grad_norm": 1.1991072095190312, "learning_rate": 5.549999999999999e-06, "loss": 0.9184, "step": 555 }, { "epoch": 0.2103471105646458, "grad_norm": 1.140264913482887, "learning_rate": 5.559999999999999e-06, "loss": 0.8663, "step": 556 }, { "epoch": 0.21072543270594912, "grad_norm": 1.1185725137493638, "learning_rate": 5.57e-06, "loss": 0.9098, "step": 557 }, { "epoch": 0.21110375484725244, "grad_norm": 1.156695278835195, "learning_rate": 5.58e-06, "loss": 0.8781, "step": 558 }, { "epoch": 0.21148207698855576, "grad_norm": 1.145333592771482, "learning_rate": 5.59e-06, "loss": 0.882, "step": 559 }, { "epoch": 0.21186039912985907, "grad_norm": 1.1762140502072864, "learning_rate": 5.6e-06, "loss": 0.8269, "step": 560 }, { "epoch": 0.2122387212711624, "grad_norm": 1.1607104680787836, "learning_rate": 5.61e-06, "loss": 0.8718, "step": 561 }, { "epoch": 0.2126170434124657, "grad_norm": 1.1469573147450298, "learning_rate": 5.6199999999999996e-06, "loss": 0.9056, "step": 562 }, { "epoch": 0.21299536555376902, "grad_norm": 1.1193447632576843, "learning_rate": 5.6299999999999995e-06, "loss": 0.8501, "step": 563 }, { "epoch": 0.21337368769507237, "grad_norm": 1.136879874832253, "learning_rate": 5.639999999999999e-06, "loss": 0.8124, "step": 564 }, { "epoch": 0.21375200983637568, "grad_norm": 1.1284818158744658, "learning_rate": 5.65e-06, "loss": 0.8676, "step": 565 }, { "epoch": 0.214130331977679, "grad_norm": 1.2698716712465286, "learning_rate": 5.66e-06, "loss": 0.8661, "step": 566 }, { "epoch": 0.21450865411898232, "grad_norm": 1.153073394080358, "learning_rate": 5.67e-06, "loss": 0.8164, "step": 567 }, { "epoch": 0.21488697626028563, "grad_norm": 1.187929464303015, "learning_rate": 5.68e-06, "loss": 0.8803, "step": 568 }, { "epoch": 0.21526529840158895, "grad_norm": 1.1011027732459755, "learning_rate": 5.69e-06, "loss": 0.8709, "step": 569 }, { "epoch": 0.21564362054289227, "grad_norm": 1.104661943825339, "learning_rate": 5.7e-06, "loss": 0.8408, "step": 570 }, { "epoch": 0.21602194268419558, "grad_norm": 1.1237999429331513, "learning_rate": 5.7099999999999995e-06, "loss": 0.8316, "step": 571 }, { "epoch": 0.2164002648254989, "grad_norm": 1.188002832097036, "learning_rate": 5.7199999999999994e-06, "loss": 0.8431, "step": 572 }, { "epoch": 0.21677858696680224, "grad_norm": 1.1510459825305048, "learning_rate": 5.73e-06, "loss": 0.8847, "step": 573 }, { "epoch": 0.21715690910810556, "grad_norm": 1.0954180332540966, "learning_rate": 5.74e-06, "loss": 0.8544, "step": 574 }, { "epoch": 0.21753523124940888, "grad_norm": 1.1472545717374318, "learning_rate": 5.75e-06, "loss": 0.8249, "step": 575 }, { "epoch": 0.2179135533907122, "grad_norm": 1.175641095732617, "learning_rate": 5.76e-06, "loss": 0.8614, "step": 576 }, { "epoch": 0.2182918755320155, "grad_norm": 1.116355053736543, "learning_rate": 5.769999999999999e-06, "loss": 0.8405, "step": 577 }, { "epoch": 0.21867019767331883, "grad_norm": 1.1157321259442492, "learning_rate": 5.78e-06, "loss": 0.8786, "step": 578 }, { "epoch": 0.21904851981462214, "grad_norm": 1.1931582815103652, "learning_rate": 5.79e-06, "loss": 0.8904, "step": 579 }, { "epoch": 0.21942684195592546, "grad_norm": 1.184066717780273, "learning_rate": 5.7999999999999995e-06, "loss": 0.8508, "step": 580 }, { "epoch": 0.21980516409722878, "grad_norm": 1.161154664599336, "learning_rate": 5.8099999999999994e-06, "loss": 0.9202, "step": 581 }, { "epoch": 0.22018348623853212, "grad_norm": 1.2235874832602252, "learning_rate": 5.82e-06, "loss": 0.8361, "step": 582 }, { "epoch": 0.22056180837983544, "grad_norm": 1.1262137082837416, "learning_rate": 5.83e-06, "loss": 0.8566, "step": 583 }, { "epoch": 0.22094013052113876, "grad_norm": 1.2072112047436216, "learning_rate": 5.84e-06, "loss": 0.8632, "step": 584 }, { "epoch": 0.22131845266244207, "grad_norm": 1.1490940800541938, "learning_rate": 5.849999999999999e-06, "loss": 0.8593, "step": 585 }, { "epoch": 0.2216967748037454, "grad_norm": 1.207791799143847, "learning_rate": 5.86e-06, "loss": 0.8556, "step": 586 }, { "epoch": 0.2220750969450487, "grad_norm": 1.1526196801211563, "learning_rate": 5.87e-06, "loss": 0.8606, "step": 587 }, { "epoch": 0.22245341908635202, "grad_norm": 1.1397609148470536, "learning_rate": 5.88e-06, "loss": 0.8469, "step": 588 }, { "epoch": 0.22283174122765534, "grad_norm": 1.1785117139043815, "learning_rate": 5.8899999999999995e-06, "loss": 0.9147, "step": 589 }, { "epoch": 0.22321006336895866, "grad_norm": 1.1858125002539965, "learning_rate": 5.9e-06, "loss": 0.8849, "step": 590 }, { "epoch": 0.223588385510262, "grad_norm": 1.1941323389502188, "learning_rate": 5.91e-06, "loss": 0.869, "step": 591 }, { "epoch": 0.22396670765156532, "grad_norm": 1.1418623190210022, "learning_rate": 5.92e-06, "loss": 0.8308, "step": 592 }, { "epoch": 0.22434502979286863, "grad_norm": 1.0743417979986591, "learning_rate": 5.929999999999999e-06, "loss": 0.843, "step": 593 }, { "epoch": 0.22472335193417195, "grad_norm": 1.1529208818856194, "learning_rate": 5.94e-06, "loss": 0.8235, "step": 594 }, { "epoch": 0.22510167407547527, "grad_norm": 1.0767273225154363, "learning_rate": 5.95e-06, "loss": 0.8247, "step": 595 }, { "epoch": 0.22547999621677858, "grad_norm": 1.1070019054712885, "learning_rate": 5.96e-06, "loss": 0.8426, "step": 596 }, { "epoch": 0.2258583183580819, "grad_norm": 1.166373551635366, "learning_rate": 5.97e-06, "loss": 0.8732, "step": 597 }, { "epoch": 0.22623664049938522, "grad_norm": 1.123857925375413, "learning_rate": 5.98e-06, "loss": 0.8464, "step": 598 }, { "epoch": 0.22661496264068853, "grad_norm": 1.08557960856811, "learning_rate": 5.99e-06, "loss": 0.821, "step": 599 }, { "epoch": 0.22699328478199188, "grad_norm": 1.1164890662505647, "learning_rate": 6e-06, "loss": 0.8846, "step": 600 }, { "epoch": 0.2273716069232952, "grad_norm": 1.1514037573784872, "learning_rate": 6.009999999999999e-06, "loss": 0.8552, "step": 601 }, { "epoch": 0.2277499290645985, "grad_norm": 1.1511174146769416, "learning_rate": 6.019999999999999e-06, "loss": 0.9014, "step": 602 }, { "epoch": 0.22812825120590183, "grad_norm": 1.1696423261594386, "learning_rate": 6.03e-06, "loss": 0.8605, "step": 603 }, { "epoch": 0.22850657334720514, "grad_norm": 1.1207706559785515, "learning_rate": 6.04e-06, "loss": 0.8382, "step": 604 }, { "epoch": 0.22888489548850846, "grad_norm": 1.1767521633404514, "learning_rate": 6.05e-06, "loss": 0.9206, "step": 605 }, { "epoch": 0.22926321762981178, "grad_norm": 1.1758374604143937, "learning_rate": 6.06e-06, "loss": 0.8883, "step": 606 }, { "epoch": 0.2296415397711151, "grad_norm": 1.148791521470335, "learning_rate": 6.07e-06, "loss": 0.9091, "step": 607 }, { "epoch": 0.2300198619124184, "grad_norm": 1.1533752302256568, "learning_rate": 6.079999999999999e-06, "loss": 0.915, "step": 608 }, { "epoch": 0.23039818405372176, "grad_norm": 1.1082862913426186, "learning_rate": 6.089999999999999e-06, "loss": 0.8259, "step": 609 }, { "epoch": 0.23077650619502507, "grad_norm": 1.1400168808816862, "learning_rate": 6.099999999999999e-06, "loss": 0.8417, "step": 610 }, { "epoch": 0.2311548283363284, "grad_norm": 1.149922499835282, "learning_rate": 6.11e-06, "loss": 0.8736, "step": 611 }, { "epoch": 0.2315331504776317, "grad_norm": 1.1611344187938348, "learning_rate": 6.12e-06, "loss": 0.8376, "step": 612 }, { "epoch": 0.23191147261893502, "grad_norm": 1.1787603376828737, "learning_rate": 6.13e-06, "loss": 0.8558, "step": 613 }, { "epoch": 0.23228979476023834, "grad_norm": 1.155525289243939, "learning_rate": 6.14e-06, "loss": 0.8463, "step": 614 }, { "epoch": 0.23266811690154166, "grad_norm": 1.1589832886045384, "learning_rate": 6.15e-06, "loss": 0.8182, "step": 615 }, { "epoch": 0.23304643904284497, "grad_norm": 1.1033596458549921, "learning_rate": 6.1599999999999995e-06, "loss": 0.8324, "step": 616 }, { "epoch": 0.23342476118414832, "grad_norm": 1.2358470403500466, "learning_rate": 6.169999999999999e-06, "loss": 0.8682, "step": 617 }, { "epoch": 0.23380308332545163, "grad_norm": 1.0984535537652391, "learning_rate": 6.179999999999999e-06, "loss": 0.8332, "step": 618 }, { "epoch": 0.23418140546675495, "grad_norm": 1.2128396124349823, "learning_rate": 6.19e-06, "loss": 0.8747, "step": 619 }, { "epoch": 0.23455972760805827, "grad_norm": 1.2275794235621071, "learning_rate": 6.2e-06, "loss": 0.8953, "step": 620 }, { "epoch": 0.23493804974936158, "grad_norm": 1.2542101409168016, "learning_rate": 6.21e-06, "loss": 0.8892, "step": 621 }, { "epoch": 0.2353163718906649, "grad_norm": 1.204474995156125, "learning_rate": 6.22e-06, "loss": 0.8491, "step": 622 }, { "epoch": 0.23569469403196822, "grad_norm": 1.1548886283677673, "learning_rate": 6.2300000000000005e-06, "loss": 0.8581, "step": 623 }, { "epoch": 0.23607301617327153, "grad_norm": 1.251297532099902, "learning_rate": 6.2399999999999995e-06, "loss": 0.851, "step": 624 }, { "epoch": 0.23645133831457485, "grad_norm": 1.218716341983368, "learning_rate": 6.2499999999999995e-06, "loss": 0.917, "step": 625 }, { "epoch": 0.2368296604558782, "grad_norm": 1.1845662251647084, "learning_rate": 6.259999999999999e-06, "loss": 0.9132, "step": 626 }, { "epoch": 0.2372079825971815, "grad_norm": 1.1620810200029381, "learning_rate": 6.269999999999999e-06, "loss": 0.8652, "step": 627 }, { "epoch": 0.23758630473848483, "grad_norm": 1.1563059559969693, "learning_rate": 6.28e-06, "loss": 0.8474, "step": 628 }, { "epoch": 0.23796462687978814, "grad_norm": 1.1388389502769878, "learning_rate": 6.29e-06, "loss": 0.8314, "step": 629 }, { "epoch": 0.23834294902109146, "grad_norm": 1.1551456623854715, "learning_rate": 6.3e-06, "loss": 0.8902, "step": 630 }, { "epoch": 0.23872127116239478, "grad_norm": 1.1459750574525491, "learning_rate": 6.31e-06, "loss": 0.8505, "step": 631 }, { "epoch": 0.2390995933036981, "grad_norm": 1.0925608036319805, "learning_rate": 6.32e-06, "loss": 0.8651, "step": 632 }, { "epoch": 0.2394779154450014, "grad_norm": 1.1607966985031983, "learning_rate": 6.3299999999999995e-06, "loss": 0.8156, "step": 633 }, { "epoch": 0.23985623758630473, "grad_norm": 1.112649862871437, "learning_rate": 6.3399999999999994e-06, "loss": 0.823, "step": 634 }, { "epoch": 0.24023455972760807, "grad_norm": 1.1213541389814015, "learning_rate": 6.349999999999999e-06, "loss": 0.8397, "step": 635 }, { "epoch": 0.2406128818689114, "grad_norm": 1.134629038613528, "learning_rate": 6.36e-06, "loss": 0.8503, "step": 636 }, { "epoch": 0.2409912040102147, "grad_norm": 1.1342734785655144, "learning_rate": 6.37e-06, "loss": 0.8497, "step": 637 }, { "epoch": 0.24136952615151802, "grad_norm": 1.1277526276470056, "learning_rate": 6.38e-06, "loss": 0.8348, "step": 638 }, { "epoch": 0.24174784829282134, "grad_norm": 1.1313262215365258, "learning_rate": 6.39e-06, "loss": 0.8746, "step": 639 }, { "epoch": 0.24212617043412465, "grad_norm": 1.0984126709233168, "learning_rate": 6.4e-06, "loss": 0.8296, "step": 640 }, { "epoch": 0.24250449257542797, "grad_norm": 1.0888784783993595, "learning_rate": 6.41e-06, "loss": 0.8129, "step": 641 }, { "epoch": 0.2428828147167313, "grad_norm": 1.1461818324642985, "learning_rate": 6.4199999999999995e-06, "loss": 0.8834, "step": 642 }, { "epoch": 0.2432611368580346, "grad_norm": 1.1427506153934843, "learning_rate": 6.429999999999999e-06, "loss": 0.8706, "step": 643 }, { "epoch": 0.24363945899933795, "grad_norm": 1.144102199065487, "learning_rate": 6.44e-06, "loss": 0.8877, "step": 644 }, { "epoch": 0.24401778114064127, "grad_norm": 1.1231424595451174, "learning_rate": 6.45e-06, "loss": 0.8939, "step": 645 }, { "epoch": 0.24439610328194458, "grad_norm": 1.1218026132749124, "learning_rate": 6.46e-06, "loss": 0.8366, "step": 646 }, { "epoch": 0.2447744254232479, "grad_norm": 1.2086540508049943, "learning_rate": 6.469999999999999e-06, "loss": 0.892, "step": 647 }, { "epoch": 0.24515274756455122, "grad_norm": 1.0868363589750187, "learning_rate": 6.48e-06, "loss": 0.8581, "step": 648 }, { "epoch": 0.24553106970585453, "grad_norm": 1.1504181380058272, "learning_rate": 6.49e-06, "loss": 0.8942, "step": 649 }, { "epoch": 0.24590939184715785, "grad_norm": 1.1874832509790985, "learning_rate": 6.5e-06, "loss": 0.8379, "step": 650 }, { "epoch": 0.24628771398846117, "grad_norm": 1.1066886977698138, "learning_rate": 6.5099999999999995e-06, "loss": 0.8645, "step": 651 }, { "epoch": 0.24666603612976448, "grad_norm": 1.1091171121306154, "learning_rate": 6.519999999999999e-06, "loss": 0.8866, "step": 652 }, { "epoch": 0.24704435827106783, "grad_norm": 1.1168392333785764, "learning_rate": 6.53e-06, "loss": 0.8377, "step": 653 }, { "epoch": 0.24742268041237114, "grad_norm": 1.1333024723334617, "learning_rate": 6.54e-06, "loss": 0.8404, "step": 654 }, { "epoch": 0.24780100255367446, "grad_norm": 1.1624311607412376, "learning_rate": 6.549999999999999e-06, "loss": 0.8578, "step": 655 }, { "epoch": 0.24817932469497778, "grad_norm": 1.140510520926876, "learning_rate": 6.559999999999999e-06, "loss": 0.7948, "step": 656 }, { "epoch": 0.2485576468362811, "grad_norm": 1.1241297695775005, "learning_rate": 6.57e-06, "loss": 0.8455, "step": 657 }, { "epoch": 0.2489359689775844, "grad_norm": 1.1171688585786779, "learning_rate": 6.58e-06, "loss": 0.8347, "step": 658 }, { "epoch": 0.24931429111888773, "grad_norm": 1.131716974118065, "learning_rate": 6.59e-06, "loss": 0.8624, "step": 659 }, { "epoch": 0.24969261326019104, "grad_norm": 1.1586113355227856, "learning_rate": 6.5999999999999995e-06, "loss": 0.8937, "step": 660 }, { "epoch": 0.2500709354014944, "grad_norm": 1.186938370866149, "learning_rate": 6.61e-06, "loss": 0.8523, "step": 661 }, { "epoch": 0.2504492575427977, "grad_norm": 1.1500652838613878, "learning_rate": 6.62e-06, "loss": 0.8537, "step": 662 }, { "epoch": 0.250827579684101, "grad_norm": 1.2121811392488833, "learning_rate": 6.629999999999999e-06, "loss": 0.8477, "step": 663 }, { "epoch": 0.2512059018254043, "grad_norm": 1.1348675624901883, "learning_rate": 6.639999999999999e-06, "loss": 0.8502, "step": 664 }, { "epoch": 0.25158422396670765, "grad_norm": 1.102535269461347, "learning_rate": 6.65e-06, "loss": 0.8745, "step": 665 }, { "epoch": 0.25158422396670765, "eval_loss": 0.8625780940055847, "eval_runtime": 27.0021, "eval_samples_per_second": 32.775, "eval_steps_per_second": 1.037, "step": 665 }, { "epoch": 0.25158422396670765, "eval_bench_accuracy_arc_challenge": 0.24285714285714285, "eval_bench_accuracy_hellaswag": 0.225, "eval_bench_accuracy_mmlu": 0.2782608695652174, "eval_bench_average_accuracy": 0.24870600414078672, "eval_bench_loss": 5.327823571991503, "eval_bench_total_accuracy": 0.24395604395604395, "step": 665 }, { "epoch": 0.251962546108011, "grad_norm": 1.149499114356956, "learning_rate": 6.66e-06, "loss": 0.8693, "step": 666 }, { "epoch": 0.2523408682493143, "grad_norm": 1.161075438749712, "learning_rate": 6.67e-06, "loss": 0.9075, "step": 667 }, { "epoch": 0.25271919039061763, "grad_norm": 1.141541764628487, "learning_rate": 6.6799999999999996e-06, "loss": 0.8643, "step": 668 }, { "epoch": 0.2530975125319209, "grad_norm": 1.1390764097501647, "learning_rate": 6.69e-06, "loss": 0.8752, "step": 669 }, { "epoch": 0.25347583467322427, "grad_norm": 1.1198865085900025, "learning_rate": 6.7e-06, "loss": 0.8403, "step": 670 }, { "epoch": 0.25385415681452755, "grad_norm": 1.143235453200182, "learning_rate": 6.709999999999999e-06, "loss": 0.8347, "step": 671 }, { "epoch": 0.2542324789558309, "grad_norm": 1.105054342960603, "learning_rate": 6.719999999999999e-06, "loss": 0.877, "step": 672 }, { "epoch": 0.2546108010971342, "grad_norm": 1.1899413861555337, "learning_rate": 6.73e-06, "loss": 0.8239, "step": 673 }, { "epoch": 0.25498912323843753, "grad_norm": 1.1305008415556128, "learning_rate": 6.74e-06, "loss": 0.8598, "step": 674 }, { "epoch": 0.2553674453797409, "grad_norm": 1.168034799536073, "learning_rate": 6.75e-06, "loss": 0.8294, "step": 675 }, { "epoch": 0.25574576752104416, "grad_norm": 1.1472097884900647, "learning_rate": 6.76e-06, "loss": 0.9007, "step": 676 }, { "epoch": 0.2561240896623475, "grad_norm": 1.0931411919432397, "learning_rate": 6.7699999999999996e-06, "loss": 0.8326, "step": 677 }, { "epoch": 0.2565024118036508, "grad_norm": 1.1510688024969498, "learning_rate": 6.78e-06, "loss": 0.8828, "step": 678 }, { "epoch": 0.25688073394495414, "grad_norm": 1.1191461068866526, "learning_rate": 6.789999999999999e-06, "loss": 0.8461, "step": 679 }, { "epoch": 0.25725905608625743, "grad_norm": 1.1041404496614182, "learning_rate": 6.799999999999999e-06, "loss": 0.8285, "step": 680 }, { "epoch": 0.2576373782275608, "grad_norm": 1.1012877673575499, "learning_rate": 6.809999999999999e-06, "loss": 0.8548, "step": 681 }, { "epoch": 0.25801570036886406, "grad_norm": 1.1057501522176822, "learning_rate": 6.82e-06, "loss": 0.8591, "step": 682 }, { "epoch": 0.2583940225101674, "grad_norm": 1.1498742481849225, "learning_rate": 6.83e-06, "loss": 0.8661, "step": 683 }, { "epoch": 0.25877234465147075, "grad_norm": 1.1378178315852814, "learning_rate": 6.84e-06, "loss": 0.8759, "step": 684 }, { "epoch": 0.25915066679277404, "grad_norm": 1.1011069671017035, "learning_rate": 6.85e-06, "loss": 0.823, "step": 685 }, { "epoch": 0.2595289889340774, "grad_norm": 1.160807734407358, "learning_rate": 6.86e-06, "loss": 0.8732, "step": 686 }, { "epoch": 0.2599073110753807, "grad_norm": 1.0867868118261128, "learning_rate": 6.8699999999999994e-06, "loss": 0.8367, "step": 687 }, { "epoch": 0.260285633216684, "grad_norm": 1.0969221739263768, "learning_rate": 6.879999999999999e-06, "loss": 0.8647, "step": 688 }, { "epoch": 0.2606639553579873, "grad_norm": 1.0995292401504533, "learning_rate": 6.889999999999999e-06, "loss": 0.8524, "step": 689 }, { "epoch": 0.26104227749929065, "grad_norm": 1.1692507904848903, "learning_rate": 6.9e-06, "loss": 0.8519, "step": 690 }, { "epoch": 0.26142059964059394, "grad_norm": 1.0998400071794445, "learning_rate": 6.91e-06, "loss": 0.8287, "step": 691 }, { "epoch": 0.2617989217818973, "grad_norm": 1.1968950530047644, "learning_rate": 6.92e-06, "loss": 0.8138, "step": 692 }, { "epoch": 0.26217724392320063, "grad_norm": 1.095854905073934, "learning_rate": 6.93e-06, "loss": 0.8568, "step": 693 }, { "epoch": 0.2625555660645039, "grad_norm": 1.1079273378796317, "learning_rate": 6.9400000000000005e-06, "loss": 0.8353, "step": 694 }, { "epoch": 0.26293388820580726, "grad_norm": 1.1606191819435765, "learning_rate": 6.9499999999999995e-06, "loss": 0.8561, "step": 695 }, { "epoch": 0.26331221034711055, "grad_norm": 1.0902425837878627, "learning_rate": 6.9599999999999994e-06, "loss": 0.8391, "step": 696 }, { "epoch": 0.2636905324884139, "grad_norm": 1.1206727493642596, "learning_rate": 6.969999999999999e-06, "loss": 0.8233, "step": 697 }, { "epoch": 0.2640688546297172, "grad_norm": 1.0982647837307586, "learning_rate": 6.98e-06, "loss": 0.8602, "step": 698 }, { "epoch": 0.26444717677102053, "grad_norm": 1.0871328583668558, "learning_rate": 6.99e-06, "loss": 0.8299, "step": 699 }, { "epoch": 0.2648254989123238, "grad_norm": 1.1008815238203256, "learning_rate": 7e-06, "loss": 0.8341, "step": 700 }, { "epoch": 0.26520382105362716, "grad_norm": 1.1750095526723472, "learning_rate": 7.01e-06, "loss": 0.8682, "step": 701 }, { "epoch": 0.2655821431949305, "grad_norm": 1.1415931541767914, "learning_rate": 7.019999999999999e-06, "loss": 0.8932, "step": 702 }, { "epoch": 0.2659604653362338, "grad_norm": 1.0981715817655127, "learning_rate": 7.03e-06, "loss": 0.838, "step": 703 }, { "epoch": 0.26633878747753714, "grad_norm": 1.0986067356062597, "learning_rate": 7.0399999999999995e-06, "loss": 0.8503, "step": 704 }, { "epoch": 0.26671710961884043, "grad_norm": 1.1084347528867848, "learning_rate": 7.049999999999999e-06, "loss": 0.8958, "step": 705 }, { "epoch": 0.2670954317601438, "grad_norm": 1.1475294765378516, "learning_rate": 7.059999999999999e-06, "loss": 0.8496, "step": 706 }, { "epoch": 0.26747375390144706, "grad_norm": 1.117143691203432, "learning_rate": 7.07e-06, "loss": 0.875, "step": 707 }, { "epoch": 0.2678520760427504, "grad_norm": 1.1331250955748378, "learning_rate": 7.08e-06, "loss": 0.854, "step": 708 }, { "epoch": 0.2682303981840537, "grad_norm": 1.0837995640069416, "learning_rate": 7.09e-06, "loss": 0.8461, "step": 709 }, { "epoch": 0.26860872032535704, "grad_norm": 1.0933867992273585, "learning_rate": 7.099999999999999e-06, "loss": 0.8383, "step": 710 }, { "epoch": 0.2689870424666604, "grad_norm": 1.0862191237112888, "learning_rate": 7.11e-06, "loss": 0.7976, "step": 711 }, { "epoch": 0.2693653646079637, "grad_norm": 1.1151836826262986, "learning_rate": 7.12e-06, "loss": 0.8224, "step": 712 }, { "epoch": 0.269743686749267, "grad_norm": 1.189062828656012, "learning_rate": 7.1299999999999995e-06, "loss": 0.8917, "step": 713 }, { "epoch": 0.2701220088905703, "grad_norm": 1.1119181389921133, "learning_rate": 7.139999999999999e-06, "loss": 0.8291, "step": 714 }, { "epoch": 0.27050033103187365, "grad_norm": 1.114538144475484, "learning_rate": 7.15e-06, "loss": 0.8996, "step": 715 }, { "epoch": 0.27087865317317694, "grad_norm": 1.1005437857491667, "learning_rate": 7.16e-06, "loss": 0.7888, "step": 716 }, { "epoch": 0.2712569753144803, "grad_norm": 1.1146994809955666, "learning_rate": 7.17e-06, "loss": 0.8878, "step": 717 }, { "epoch": 0.2716352974557836, "grad_norm": 1.0936279250904897, "learning_rate": 7.179999999999999e-06, "loss": 0.8672, "step": 718 }, { "epoch": 0.2720136195970869, "grad_norm": 1.1366251894998205, "learning_rate": 7.19e-06, "loss": 0.8858, "step": 719 }, { "epoch": 0.27239194173839026, "grad_norm": 1.1195931324613553, "learning_rate": 7.2e-06, "loss": 0.8507, "step": 720 }, { "epoch": 0.27277026387969355, "grad_norm": 1.0935327911384591, "learning_rate": 7.21e-06, "loss": 0.8424, "step": 721 }, { "epoch": 0.2731485860209969, "grad_norm": 1.0953372322434138, "learning_rate": 7.2199999999999995e-06, "loss": 0.8831, "step": 722 }, { "epoch": 0.2735269081623002, "grad_norm": 1.0904032768722667, "learning_rate": 7.23e-06, "loss": 0.8334, "step": 723 }, { "epoch": 0.27390523030360353, "grad_norm": 1.1346874176897102, "learning_rate": 7.24e-06, "loss": 0.8506, "step": 724 }, { "epoch": 0.2742835524449068, "grad_norm": 1.154262444900059, "learning_rate": 7.25e-06, "loss": 0.8393, "step": 725 }, { "epoch": 0.27466187458621016, "grad_norm": 1.1336981217637951, "learning_rate": 7.259999999999999e-06, "loss": 0.8371, "step": 726 }, { "epoch": 0.27504019672751345, "grad_norm": 1.1530922109530841, "learning_rate": 7.269999999999999e-06, "loss": 0.9141, "step": 727 }, { "epoch": 0.2754185188688168, "grad_norm": 1.1414400257725132, "learning_rate": 7.28e-06, "loss": 0.8615, "step": 728 }, { "epoch": 0.27579684101012014, "grad_norm": 1.0747602134856014, "learning_rate": 7.29e-06, "loss": 0.8507, "step": 729 }, { "epoch": 0.27617516315142343, "grad_norm": 1.1341332656767107, "learning_rate": 7.2999999999999996e-06, "loss": 0.8771, "step": 730 }, { "epoch": 0.2765534852927268, "grad_norm": 1.127774756748704, "learning_rate": 7.3099999999999995e-06, "loss": 0.8559, "step": 731 }, { "epoch": 0.27693180743403006, "grad_norm": 1.106246473020497, "learning_rate": 7.32e-06, "loss": 0.8333, "step": 732 }, { "epoch": 0.2773101295753334, "grad_norm": 1.072619886572064, "learning_rate": 7.33e-06, "loss": 0.8138, "step": 733 }, { "epoch": 0.2776884517166367, "grad_norm": 1.1053237591292755, "learning_rate": 7.339999999999999e-06, "loss": 0.8929, "step": 734 }, { "epoch": 0.27806677385794004, "grad_norm": 1.0590657569440343, "learning_rate": 7.349999999999999e-06, "loss": 0.8657, "step": 735 }, { "epoch": 0.27844509599924333, "grad_norm": 1.0990511323540157, "learning_rate": 7.36e-06, "loss": 0.831, "step": 736 }, { "epoch": 0.2788234181405467, "grad_norm": 1.0960494967933392, "learning_rate": 7.37e-06, "loss": 0.8672, "step": 737 }, { "epoch": 0.27920174028185, "grad_norm": 1.0923972930315522, "learning_rate": 7.38e-06, "loss": 0.8359, "step": 738 }, { "epoch": 0.2795800624231533, "grad_norm": 1.117398170352597, "learning_rate": 7.3899999999999995e-06, "loss": 0.8678, "step": 739 }, { "epoch": 0.27995838456445665, "grad_norm": 1.0964334876514574, "learning_rate": 7.4e-06, "loss": 0.8175, "step": 740 }, { "epoch": 0.28033670670575994, "grad_norm": 1.137429209179925, "learning_rate": 7.41e-06, "loss": 0.8469, "step": 741 }, { "epoch": 0.2807150288470633, "grad_norm": 1.1550309848051612, "learning_rate": 7.419999999999999e-06, "loss": 0.8326, "step": 742 }, { "epoch": 0.2810933509883666, "grad_norm": 1.1935237789558146, "learning_rate": 7.429999999999999e-06, "loss": 0.8568, "step": 743 }, { "epoch": 0.2814716731296699, "grad_norm": 1.1694982973025607, "learning_rate": 7.44e-06, "loss": 0.8869, "step": 744 }, { "epoch": 0.2818499952709732, "grad_norm": 1.1920139094347593, "learning_rate": 7.45e-06, "loss": 0.8487, "step": 745 }, { "epoch": 0.28222831741227655, "grad_norm": 1.1367845567285337, "learning_rate": 7.46e-06, "loss": 0.8554, "step": 746 }, { "epoch": 0.2826066395535799, "grad_norm": 1.1505063717374056, "learning_rate": 7.47e-06, "loss": 0.8371, "step": 747 }, { "epoch": 0.2829849616948832, "grad_norm": 1.1339987287473563, "learning_rate": 7.48e-06, "loss": 0.8256, "step": 748 }, { "epoch": 0.28336328383618653, "grad_norm": 1.158977003616627, "learning_rate": 7.49e-06, "loss": 0.8913, "step": 749 }, { "epoch": 0.2837416059774898, "grad_norm": 1.1022707433616572, "learning_rate": 7.499999999999999e-06, "loss": 0.8117, "step": 750 }, { "epoch": 0.28411992811879316, "grad_norm": 1.1550634309139105, "learning_rate": 7.509999999999999e-06, "loss": 0.8906, "step": 751 }, { "epoch": 0.28449825026009645, "grad_norm": 1.090317910646282, "learning_rate": 7.519999999999999e-06, "loss": 0.8799, "step": 752 }, { "epoch": 0.2848765724013998, "grad_norm": 1.0677643984555838, "learning_rate": 7.53e-06, "loss": 0.8653, "step": 753 }, { "epoch": 0.2852548945427031, "grad_norm": 1.1663544994037678, "learning_rate": 7.54e-06, "loss": 0.8737, "step": 754 }, { "epoch": 0.28563321668400643, "grad_norm": 1.0973153975053445, "learning_rate": 7.55e-06, "loss": 0.8485, "step": 755 }, { "epoch": 0.2860115388253098, "grad_norm": 1.0761549351444184, "learning_rate": 7.56e-06, "loss": 0.8284, "step": 756 }, { "epoch": 0.28638986096661306, "grad_norm": 1.1355050591654032, "learning_rate": 7.5699999999999995e-06, "loss": 0.8348, "step": 757 }, { "epoch": 0.2867681831079164, "grad_norm": 1.116699730612722, "learning_rate": 7.5799999999999994e-06, "loss": 0.8405, "step": 758 }, { "epoch": 0.2871465052492197, "grad_norm": 1.1037588379626753, "learning_rate": 7.589999999999999e-06, "loss": 0.8652, "step": 759 }, { "epoch": 0.28752482739052304, "grad_norm": 1.092569661781677, "learning_rate": 7.599999999999999e-06, "loss": 0.8786, "step": 760 }, { "epoch": 0.28790314953182633, "grad_norm": 1.1079207038423997, "learning_rate": 7.61e-06, "loss": 0.8731, "step": 761 }, { "epoch": 0.2882814716731297, "grad_norm": 1.0840455559100046, "learning_rate": 7.62e-06, "loss": 0.8533, "step": 762 }, { "epoch": 0.28865979381443296, "grad_norm": 1.1088308729059055, "learning_rate": 7.63e-06, "loss": 0.8407, "step": 763 }, { "epoch": 0.2890381159557363, "grad_norm": 1.070788168887275, "learning_rate": 7.64e-06, "loss": 0.8919, "step": 764 }, { "epoch": 0.28941643809703965, "grad_norm": 1.060969292922543, "learning_rate": 7.65e-06, "loss": 0.812, "step": 765 }, { "epoch": 0.28979476023834294, "grad_norm": 1.1301219505514637, "learning_rate": 7.66e-06, "loss": 0.8336, "step": 766 }, { "epoch": 0.2901730823796463, "grad_norm": 1.0534794694384884, "learning_rate": 7.67e-06, "loss": 0.8329, "step": 767 }, { "epoch": 0.2905514045209496, "grad_norm": 1.1347313685498166, "learning_rate": 7.68e-06, "loss": 0.8793, "step": 768 }, { "epoch": 0.2909297266622529, "grad_norm": 1.1475444842715925, "learning_rate": 7.69e-06, "loss": 0.8508, "step": 769 }, { "epoch": 0.2913080488035562, "grad_norm": 1.131952349011137, "learning_rate": 7.699999999999999e-06, "loss": 0.845, "step": 770 }, { "epoch": 0.29168637094485955, "grad_norm": 1.1447781586459667, "learning_rate": 7.709999999999999e-06, "loss": 0.8726, "step": 771 }, { "epoch": 0.29206469308616284, "grad_norm": 1.1327583004535982, "learning_rate": 7.719999999999999e-06, "loss": 0.8104, "step": 772 }, { "epoch": 0.2924430152274662, "grad_norm": 1.128617220703407, "learning_rate": 7.73e-06, "loss": 0.8176, "step": 773 }, { "epoch": 0.29282133736876953, "grad_norm": 1.1023174787003673, "learning_rate": 7.74e-06, "loss": 0.8428, "step": 774 }, { "epoch": 0.2931996595100728, "grad_norm": 1.1676360521088707, "learning_rate": 7.75e-06, "loss": 0.8811, "step": 775 }, { "epoch": 0.29357798165137616, "grad_norm": 1.1926785192763554, "learning_rate": 7.76e-06, "loss": 0.8814, "step": 776 }, { "epoch": 0.29395630379267945, "grad_norm": 1.0926242154672956, "learning_rate": 7.769999999999998e-06, "loss": 0.8697, "step": 777 }, { "epoch": 0.2943346259339828, "grad_norm": 1.1477061183634145, "learning_rate": 7.78e-06, "loss": 0.883, "step": 778 }, { "epoch": 0.2947129480752861, "grad_norm": 1.0524242129666213, "learning_rate": 7.79e-06, "loss": 0.8285, "step": 779 }, { "epoch": 0.29509127021658943, "grad_norm": 1.1003220338231798, "learning_rate": 7.8e-06, "loss": 0.873, "step": 780 }, { "epoch": 0.2954695923578927, "grad_norm": 1.0924766297335016, "learning_rate": 7.81e-06, "loss": 0.8388, "step": 781 }, { "epoch": 0.29584791449919606, "grad_norm": 1.0905974324189436, "learning_rate": 7.82e-06, "loss": 0.8456, "step": 782 }, { "epoch": 0.2962262366404994, "grad_norm": 1.0784036223330382, "learning_rate": 7.83e-06, "loss": 0.8732, "step": 783 }, { "epoch": 0.2966045587818027, "grad_norm": 1.0471596415042548, "learning_rate": 7.84e-06, "loss": 0.8396, "step": 784 }, { "epoch": 0.29698288092310604, "grad_norm": 1.080443491875735, "learning_rate": 7.85e-06, "loss": 0.8458, "step": 785 }, { "epoch": 0.29736120306440933, "grad_norm": 1.0828576066417819, "learning_rate": 7.86e-06, "loss": 0.813, "step": 786 }, { "epoch": 0.2977395252057127, "grad_norm": 1.0752539748255008, "learning_rate": 7.87e-06, "loss": 0.8564, "step": 787 }, { "epoch": 0.29811784734701596, "grad_norm": 1.0994217833391198, "learning_rate": 7.879999999999999e-06, "loss": 0.8263, "step": 788 }, { "epoch": 0.2984961694883193, "grad_norm": 1.086381772786406, "learning_rate": 7.889999999999999e-06, "loss": 0.8563, "step": 789 }, { "epoch": 0.2988744916296226, "grad_norm": 1.1088374241291266, "learning_rate": 7.9e-06, "loss": 0.864, "step": 790 }, { "epoch": 0.29925281377092594, "grad_norm": 1.1571412075379082, "learning_rate": 7.91e-06, "loss": 0.8171, "step": 791 }, { "epoch": 0.2996311359122293, "grad_norm": 1.1203389931533279, "learning_rate": 7.92e-06, "loss": 0.8441, "step": 792 }, { "epoch": 0.3000094580535326, "grad_norm": 1.0955306189611171, "learning_rate": 7.929999999999999e-06, "loss": 0.8367, "step": 793 }, { "epoch": 0.3003877801948359, "grad_norm": 1.0518036198212661, "learning_rate": 7.94e-06, "loss": 0.8115, "step": 794 }, { "epoch": 0.3007661023361392, "grad_norm": 1.1024545203471212, "learning_rate": 7.95e-06, "loss": 0.8981, "step": 795 }, { "epoch": 0.30114442447744255, "grad_norm": 1.1408707488859684, "learning_rate": 7.96e-06, "loss": 0.8574, "step": 796 }, { "epoch": 0.30152274661874584, "grad_norm": 1.0664606162956756, "learning_rate": 7.97e-06, "loss": 0.851, "step": 797 }, { "epoch": 0.3019010687600492, "grad_norm": 1.1045392245613144, "learning_rate": 7.98e-06, "loss": 0.8472, "step": 798 }, { "epoch": 0.3019010687600492, "eval_loss": 0.850925862789154, "eval_runtime": 26.6744, "eval_samples_per_second": 33.178, "eval_steps_per_second": 1.05, "step": 798 }, { "epoch": 0.3019010687600492, "eval_bench_accuracy_arc_challenge": 0.21428571428571427, "eval_bench_accuracy_hellaswag": 0.235, "eval_bench_accuracy_mmlu": 0.28695652173913044, "eval_bench_average_accuracy": 0.24541407867494824, "eval_bench_loss": 4.9830322265625, "eval_bench_total_accuracy": 0.24175824175824176, "step": 798 }, { "epoch": 0.30227939090135253, "grad_norm": 1.1188259399602403, "learning_rate": 7.99e-06, "loss": 0.8468, "step": 799 }, { "epoch": 0.3026577130426558, "grad_norm": 1.1431484110606045, "learning_rate": 8e-06, "loss": 0.8401, "step": 800 }, { "epoch": 0.30303603518395916, "grad_norm": 1.083646592987573, "learning_rate": 7.999999611606006e-06, "loss": 0.8062, "step": 801 }, { "epoch": 0.30341435732526245, "grad_norm": 1.1319556143394125, "learning_rate": 7.999998446424103e-06, "loss": 0.8818, "step": 802 }, { "epoch": 0.3037926794665658, "grad_norm": 1.0994025822887656, "learning_rate": 7.999996504454512e-06, "loss": 0.8509, "step": 803 }, { "epoch": 0.3041710016078691, "grad_norm": 1.0755886346693961, "learning_rate": 7.999993785697617e-06, "loss": 0.8004, "step": 804 }, { "epoch": 0.30454932374917243, "grad_norm": 1.1441919264010905, "learning_rate": 7.99999029015394e-06, "loss": 0.808, "step": 805 }, { "epoch": 0.3049276458904757, "grad_norm": 1.1065610412104439, "learning_rate": 7.999986017824165e-06, "loss": 0.8549, "step": 806 }, { "epoch": 0.30530596803177906, "grad_norm": 1.0882701082696518, "learning_rate": 7.999980968709117e-06, "loss": 0.8468, "step": 807 }, { "epoch": 0.3056842901730824, "grad_norm": 1.1088124295992208, "learning_rate": 7.999975142809778e-06, "loss": 0.8736, "step": 808 }, { "epoch": 0.3060626123143857, "grad_norm": 1.1033663016693673, "learning_rate": 7.99996854012728e-06, "loss": 0.8476, "step": 809 }, { "epoch": 0.30644093445568904, "grad_norm": 1.13603689058083, "learning_rate": 7.999961160662905e-06, "loss": 0.8445, "step": 810 }, { "epoch": 0.30681925659699233, "grad_norm": 1.160741078547518, "learning_rate": 7.999953004418086e-06, "loss": 0.8858, "step": 811 }, { "epoch": 0.3071975787382957, "grad_norm": 1.1137885301105297, "learning_rate": 7.999944071394408e-06, "loss": 0.8468, "step": 812 }, { "epoch": 0.30757590087959896, "grad_norm": 1.0950922126362728, "learning_rate": 7.999934361593606e-06, "loss": 0.8277, "step": 813 }, { "epoch": 0.3079542230209023, "grad_norm": 1.0705498486629084, "learning_rate": 7.999923875017561e-06, "loss": 0.8542, "step": 814 }, { "epoch": 0.3083325451622056, "grad_norm": 1.0320443969916053, "learning_rate": 7.999912611668314e-06, "loss": 0.8311, "step": 815 }, { "epoch": 0.30871086730350894, "grad_norm": 1.1098560201406311, "learning_rate": 7.999900571548054e-06, "loss": 0.8285, "step": 816 }, { "epoch": 0.3090891894448123, "grad_norm": 1.117956788545042, "learning_rate": 7.999887754659112e-06, "loss": 0.8062, "step": 817 }, { "epoch": 0.3094675115861156, "grad_norm": 1.0815055115388574, "learning_rate": 7.999874161003984e-06, "loss": 0.825, "step": 818 }, { "epoch": 0.3098458337274189, "grad_norm": 1.1258610055051623, "learning_rate": 7.999859790585307e-06, "loss": 0.8544, "step": 819 }, { "epoch": 0.3102241558687222, "grad_norm": 1.0792203366803435, "learning_rate": 7.99984464340587e-06, "loss": 0.8371, "step": 820 }, { "epoch": 0.31060247801002555, "grad_norm": 1.0857066217255478, "learning_rate": 7.999828719468619e-06, "loss": 0.8025, "step": 821 }, { "epoch": 0.31098080015132884, "grad_norm": 1.0345681012946357, "learning_rate": 7.999812018776642e-06, "loss": 0.7961, "step": 822 }, { "epoch": 0.3113591222926322, "grad_norm": 1.0880871394519303, "learning_rate": 7.999794541333184e-06, "loss": 0.867, "step": 823 }, { "epoch": 0.3117374444339355, "grad_norm": 1.0734362647252, "learning_rate": 7.99977628714164e-06, "loss": 0.8504, "step": 824 }, { "epoch": 0.3121157665752388, "grad_norm": 1.0651195855212972, "learning_rate": 7.999757256205554e-06, "loss": 0.836, "step": 825 }, { "epoch": 0.31249408871654216, "grad_norm": 1.0952088927990486, "learning_rate": 7.99973744852862e-06, "loss": 0.8685, "step": 826 }, { "epoch": 0.31287241085784545, "grad_norm": 1.1189908995835645, "learning_rate": 7.999716864114687e-06, "loss": 0.8612, "step": 827 }, { "epoch": 0.3132507329991488, "grad_norm": 1.1107627441762915, "learning_rate": 7.999695502967753e-06, "loss": 0.887, "step": 828 }, { "epoch": 0.3136290551404521, "grad_norm": 1.0910830318775155, "learning_rate": 7.999673365091965e-06, "loss": 0.8149, "step": 829 }, { "epoch": 0.31400737728175543, "grad_norm": 1.0878738960197105, "learning_rate": 7.99965045049162e-06, "loss": 0.8543, "step": 830 }, { "epoch": 0.3143856994230587, "grad_norm": 1.1304840925957875, "learning_rate": 7.999626759171173e-06, "loss": 0.8607, "step": 831 }, { "epoch": 0.31476402156436206, "grad_norm": 1.0977832972523356, "learning_rate": 7.99960229113522e-06, "loss": 0.8238, "step": 832 }, { "epoch": 0.31514234370566535, "grad_norm": 1.1056029713906521, "learning_rate": 7.999577046388514e-06, "loss": 0.8449, "step": 833 }, { "epoch": 0.3155206658469687, "grad_norm": 1.1263279045653014, "learning_rate": 7.999551024935959e-06, "loss": 0.8996, "step": 834 }, { "epoch": 0.31589898798827204, "grad_norm": 1.1023495304424114, "learning_rate": 7.999524226782608e-06, "loss": 0.8059, "step": 835 }, { "epoch": 0.31627731012957533, "grad_norm": 1.0710753056086557, "learning_rate": 7.999496651933662e-06, "loss": 0.8364, "step": 836 }, { "epoch": 0.3166556322708787, "grad_norm": 1.1628408036471776, "learning_rate": 7.999468300394481e-06, "loss": 0.8491, "step": 837 }, { "epoch": 0.31703395441218196, "grad_norm": 1.1011205956685801, "learning_rate": 7.999439172170566e-06, "loss": 0.8371, "step": 838 }, { "epoch": 0.3174122765534853, "grad_norm": 1.067716374321139, "learning_rate": 7.999409267267577e-06, "loss": 0.8257, "step": 839 }, { "epoch": 0.3177905986947886, "grad_norm": 1.1358374860128349, "learning_rate": 7.99937858569132e-06, "loss": 0.8317, "step": 840 }, { "epoch": 0.31816892083609194, "grad_norm": 1.0779959631518108, "learning_rate": 7.999347127447752e-06, "loss": 0.7981, "step": 841 }, { "epoch": 0.31854724297739523, "grad_norm": 1.1254796876535107, "learning_rate": 7.999314892542985e-06, "loss": 0.8971, "step": 842 }, { "epoch": 0.3189255651186986, "grad_norm": 1.0901729922813403, "learning_rate": 7.999281880983277e-06, "loss": 0.8506, "step": 843 }, { "epoch": 0.3193038872600019, "grad_norm": 1.0709160400913234, "learning_rate": 7.999248092775039e-06, "loss": 0.8468, "step": 844 }, { "epoch": 0.3196822094013052, "grad_norm": 1.1223182444160262, "learning_rate": 7.999213527924831e-06, "loss": 0.8217, "step": 845 }, { "epoch": 0.32006053154260855, "grad_norm": 1.1033066311400137, "learning_rate": 7.99917818643937e-06, "loss": 0.8646, "step": 846 }, { "epoch": 0.32043885368391184, "grad_norm": 1.1122943393613496, "learning_rate": 7.999142068325514e-06, "loss": 0.8343, "step": 847 }, { "epoch": 0.3208171758252152, "grad_norm": 1.1197740571480894, "learning_rate": 7.999105173590281e-06, "loss": 0.8408, "step": 848 }, { "epoch": 0.3211954979665185, "grad_norm": 1.0680302459683109, "learning_rate": 7.999067502240835e-06, "loss": 0.8527, "step": 849 }, { "epoch": 0.3215738201078218, "grad_norm": 1.0872491602723373, "learning_rate": 7.99902905428449e-06, "loss": 0.8417, "step": 850 }, { "epoch": 0.3219521422491251, "grad_norm": 1.106663351318103, "learning_rate": 7.998989829728712e-06, "loss": 0.8055, "step": 851 }, { "epoch": 0.32233046439042845, "grad_norm": 1.0809694317490106, "learning_rate": 7.998949828581122e-06, "loss": 0.8614, "step": 852 }, { "epoch": 0.3227087865317318, "grad_norm": 1.102190346138006, "learning_rate": 7.998909050849484e-06, "loss": 0.8716, "step": 853 }, { "epoch": 0.3230871086730351, "grad_norm": 1.0436133036323463, "learning_rate": 7.998867496541719e-06, "loss": 0.8575, "step": 854 }, { "epoch": 0.32346543081433843, "grad_norm": 1.0545933388006492, "learning_rate": 7.998825165665894e-06, "loss": 0.8208, "step": 855 }, { "epoch": 0.3238437529556417, "grad_norm": 1.066597036199654, "learning_rate": 7.998782058230237e-06, "loss": 0.7723, "step": 856 }, { "epoch": 0.32422207509694506, "grad_norm": 1.053365311188067, "learning_rate": 7.998738174243111e-06, "loss": 0.8102, "step": 857 }, { "epoch": 0.32460039723824835, "grad_norm": 1.0581107038361595, "learning_rate": 7.99869351371304e-06, "loss": 0.7999, "step": 858 }, { "epoch": 0.3249787193795517, "grad_norm": 1.1008953546338276, "learning_rate": 7.998648076648702e-06, "loss": 0.8568, "step": 859 }, { "epoch": 0.325357041520855, "grad_norm": 1.1417115474045594, "learning_rate": 7.998601863058915e-06, "loss": 0.8183, "step": 860 }, { "epoch": 0.32573536366215833, "grad_norm": 1.0221082409435902, "learning_rate": 7.998554872952656e-06, "loss": 0.8236, "step": 861 }, { "epoch": 0.3261136858034617, "grad_norm": 1.0319653291858766, "learning_rate": 7.99850710633905e-06, "loss": 0.8268, "step": 862 }, { "epoch": 0.32649200794476496, "grad_norm": 1.0741619232930077, "learning_rate": 7.998458563227374e-06, "loss": 0.8635, "step": 863 }, { "epoch": 0.3268703300860683, "grad_norm": 1.084988318258729, "learning_rate": 7.998409243627051e-06, "loss": 0.807, "step": 864 }, { "epoch": 0.3272486522273716, "grad_norm": 1.0687498037098355, "learning_rate": 7.998359147547665e-06, "loss": 0.852, "step": 865 }, { "epoch": 0.32762697436867494, "grad_norm": 1.125647258256957, "learning_rate": 7.99830827499894e-06, "loss": 0.8153, "step": 866 }, { "epoch": 0.32800529650997823, "grad_norm": 1.1182770611625017, "learning_rate": 7.998256625990756e-06, "loss": 0.8103, "step": 867 }, { "epoch": 0.3283836186512816, "grad_norm": 1.0564435912408205, "learning_rate": 7.998204200533144e-06, "loss": 0.8119, "step": 868 }, { "epoch": 0.32876194079258486, "grad_norm": 1.1460131223742922, "learning_rate": 7.998150998636284e-06, "loss": 0.8289, "step": 869 }, { "epoch": 0.3291402629338882, "grad_norm": 1.0575674306051868, "learning_rate": 7.998097020310509e-06, "loss": 0.8428, "step": 870 }, { "epoch": 0.32951858507519155, "grad_norm": 1.1137833102998567, "learning_rate": 7.9980422655663e-06, "loss": 0.8218, "step": 871 }, { "epoch": 0.32989690721649484, "grad_norm": 1.1107427833797017, "learning_rate": 7.997986734414291e-06, "loss": 0.851, "step": 872 }, { "epoch": 0.3302752293577982, "grad_norm": 1.1272405856822123, "learning_rate": 7.997930426865266e-06, "loss": 0.8604, "step": 873 }, { "epoch": 0.3306535514991015, "grad_norm": 1.0539626107226423, "learning_rate": 7.997873342930158e-06, "loss": 0.8531, "step": 874 }, { "epoch": 0.3310318736404048, "grad_norm": 1.0696538969484604, "learning_rate": 7.997815482620057e-06, "loss": 0.838, "step": 875 }, { "epoch": 0.3314101957817081, "grad_norm": 1.1460143163401961, "learning_rate": 7.997756845946193e-06, "loss": 0.7944, "step": 876 }, { "epoch": 0.33178851792301145, "grad_norm": 1.1082280014219137, "learning_rate": 7.997697432919957e-06, "loss": 0.9019, "step": 877 }, { "epoch": 0.33216684006431474, "grad_norm": 1.0841358926479614, "learning_rate": 7.997637243552888e-06, "loss": 0.7975, "step": 878 }, { "epoch": 0.3325451622056181, "grad_norm": 1.056009898365743, "learning_rate": 7.997576277856674e-06, "loss": 0.8574, "step": 879 }, { "epoch": 0.33292348434692143, "grad_norm": 1.0802951235255627, "learning_rate": 7.99751453584315e-06, "loss": 0.8155, "step": 880 }, { "epoch": 0.3333018064882247, "grad_norm": 1.077889148763545, "learning_rate": 7.99745201752431e-06, "loss": 0.7963, "step": 881 }, { "epoch": 0.33368012862952806, "grad_norm": 1.1621065299950686, "learning_rate": 7.997388722912295e-06, "loss": 0.8548, "step": 882 }, { "epoch": 0.33405845077083135, "grad_norm": 1.1322105218350456, "learning_rate": 7.997324652019394e-06, "loss": 0.8795, "step": 883 }, { "epoch": 0.3344367729121347, "grad_norm": 1.136478913491314, "learning_rate": 7.997259804858054e-06, "loss": 0.8053, "step": 884 }, { "epoch": 0.334815095053438, "grad_norm": 1.132941842896281, "learning_rate": 7.997194181440863e-06, "loss": 0.8753, "step": 885 }, { "epoch": 0.3351934171947413, "grad_norm": 1.072088751980564, "learning_rate": 7.997127781780567e-06, "loss": 0.8471, "step": 886 }, { "epoch": 0.3355717393360446, "grad_norm": 1.136959198020949, "learning_rate": 7.997060605890062e-06, "loss": 0.8805, "step": 887 }, { "epoch": 0.33595006147734796, "grad_norm": 1.1411444801682626, "learning_rate": 7.996992653782392e-06, "loss": 0.8241, "step": 888 }, { "epoch": 0.3363283836186513, "grad_norm": 1.0911333474121823, "learning_rate": 7.996923925470752e-06, "loss": 0.8134, "step": 889 }, { "epoch": 0.3367067057599546, "grad_norm": 1.0929540349841498, "learning_rate": 7.996854420968492e-06, "loss": 0.8362, "step": 890 }, { "epoch": 0.33708502790125794, "grad_norm": 1.1142134518728692, "learning_rate": 7.996784140289106e-06, "loss": 0.8583, "step": 891 }, { "epoch": 0.3374633500425612, "grad_norm": 1.0776120467255657, "learning_rate": 7.996713083446245e-06, "loss": 0.8405, "step": 892 }, { "epoch": 0.33784167218386457, "grad_norm": 1.0315550349351374, "learning_rate": 7.996641250453707e-06, "loss": 0.8233, "step": 893 }, { "epoch": 0.33821999432516786, "grad_norm": 1.1320956870150307, "learning_rate": 7.996568641325441e-06, "loss": 0.8497, "step": 894 }, { "epoch": 0.3385983164664712, "grad_norm": 1.0891148355471727, "learning_rate": 7.996495256075548e-06, "loss": 0.8338, "step": 895 }, { "epoch": 0.3389766386077745, "grad_norm": 1.1104610577848222, "learning_rate": 7.99642109471828e-06, "loss": 0.8166, "step": 896 }, { "epoch": 0.33935496074907784, "grad_norm": 1.0961276245110951, "learning_rate": 7.996346157268037e-06, "loss": 0.8213, "step": 897 }, { "epoch": 0.3397332828903812, "grad_norm": 1.053397674073016, "learning_rate": 7.996270443739375e-06, "loss": 0.8269, "step": 898 }, { "epoch": 0.34011160503168447, "grad_norm": 1.05985869383675, "learning_rate": 7.996193954146995e-06, "loss": 0.8632, "step": 899 }, { "epoch": 0.3404899271729878, "grad_norm": 1.0747332831609127, "learning_rate": 7.996116688505749e-06, "loss": 0.8308, "step": 900 }, { "epoch": 0.3408682493142911, "grad_norm": 1.0617958908539586, "learning_rate": 7.996038646830645e-06, "loss": 0.8003, "step": 901 }, { "epoch": 0.34124657145559445, "grad_norm": 1.0595674189471762, "learning_rate": 7.995959829136837e-06, "loss": 0.7948, "step": 902 }, { "epoch": 0.34162489359689774, "grad_norm": 1.0753382871745762, "learning_rate": 7.995880235439632e-06, "loss": 0.8399, "step": 903 }, { "epoch": 0.3420032157382011, "grad_norm": 1.1183441140434693, "learning_rate": 7.995799865754487e-06, "loss": 0.8221, "step": 904 }, { "epoch": 0.34238153787950437, "grad_norm": 1.0929766123596374, "learning_rate": 7.995718720097011e-06, "loss": 0.8309, "step": 905 }, { "epoch": 0.3427598600208077, "grad_norm": 1.0179073548109145, "learning_rate": 7.995636798482959e-06, "loss": 0.8355, "step": 906 }, { "epoch": 0.34313818216211106, "grad_norm": 1.1183732645745317, "learning_rate": 7.99555410092824e-06, "loss": 0.8376, "step": 907 }, { "epoch": 0.34351650430341435, "grad_norm": 1.165733705514543, "learning_rate": 7.995470627448915e-06, "loss": 0.86, "step": 908 }, { "epoch": 0.3438948264447177, "grad_norm": 1.0552618018743587, "learning_rate": 7.995386378061196e-06, "loss": 0.8468, "step": 909 }, { "epoch": 0.344273148586021, "grad_norm": 1.131651010498469, "learning_rate": 7.995301352781439e-06, "loss": 0.8489, "step": 910 }, { "epoch": 0.3446514707273243, "grad_norm": 1.1028826199732988, "learning_rate": 7.995215551626162e-06, "loss": 0.8721, "step": 911 }, { "epoch": 0.3450297928686276, "grad_norm": 1.1380255943103783, "learning_rate": 7.995128974612022e-06, "loss": 0.8484, "step": 912 }, { "epoch": 0.34540811500993096, "grad_norm": 1.0659393620350812, "learning_rate": 7.995041621755835e-06, "loss": 0.8198, "step": 913 }, { "epoch": 0.34578643715123425, "grad_norm": 1.059819166817385, "learning_rate": 7.994953493074562e-06, "loss": 0.8601, "step": 914 }, { "epoch": 0.3461647592925376, "grad_norm": 1.1168724106612267, "learning_rate": 7.994864588585323e-06, "loss": 0.8314, "step": 915 }, { "epoch": 0.34654308143384094, "grad_norm": 1.0696755810222651, "learning_rate": 7.994774908305377e-06, "loss": 0.8488, "step": 916 }, { "epoch": 0.3469214035751442, "grad_norm": 1.1571812110459856, "learning_rate": 7.99468445225214e-06, "loss": 0.8157, "step": 917 }, { "epoch": 0.34729972571644757, "grad_norm": 1.114611745775756, "learning_rate": 7.994593220443181e-06, "loss": 0.8368, "step": 918 }, { "epoch": 0.34767804785775086, "grad_norm": 1.152864146273239, "learning_rate": 7.994501212896218e-06, "loss": 0.861, "step": 919 }, { "epoch": 0.3480563699990542, "grad_norm": 1.1345158690879138, "learning_rate": 7.994408429629113e-06, "loss": 0.8163, "step": 920 }, { "epoch": 0.3484346921403575, "grad_norm": 1.0577940861565938, "learning_rate": 7.994314870659892e-06, "loss": 0.7803, "step": 921 }, { "epoch": 0.34881301428166084, "grad_norm": 1.04106331488491, "learning_rate": 7.994220536006717e-06, "loss": 0.8291, "step": 922 }, { "epoch": 0.3491913364229641, "grad_norm": 1.0394935151014175, "learning_rate": 7.99412542568791e-06, "loss": 0.7819, "step": 923 }, { "epoch": 0.34956965856426747, "grad_norm": 1.1306507694533081, "learning_rate": 7.994029539721941e-06, "loss": 0.8594, "step": 924 }, { "epoch": 0.3499479807055708, "grad_norm": 1.0984697906601044, "learning_rate": 7.993932878127433e-06, "loss": 0.872, "step": 925 }, { "epoch": 0.3503263028468741, "grad_norm": 1.0848529154386723, "learning_rate": 7.993835440923154e-06, "loss": 0.8668, "step": 926 }, { "epoch": 0.35070462498817745, "grad_norm": 1.074249076888769, "learning_rate": 7.993737228128028e-06, "loss": 0.88, "step": 927 }, { "epoch": 0.35108294712948074, "grad_norm": 1.0595559434730502, "learning_rate": 7.993638239761127e-06, "loss": 0.8448, "step": 928 }, { "epoch": 0.3514612692707841, "grad_norm": 1.0586225742216135, "learning_rate": 7.993538475841674e-06, "loss": 0.806, "step": 929 }, { "epoch": 0.35183959141208737, "grad_norm": 1.0965639423993851, "learning_rate": 7.993437936389045e-06, "loss": 0.8532, "step": 930 }, { "epoch": 0.3522179135533907, "grad_norm": 1.0635648509605742, "learning_rate": 7.99333662142276e-06, "loss": 0.8659, "step": 931 }, { "epoch": 0.3522179135533907, "eval_loss": 0.8405433893203735, "eval_runtime": 26.7827, "eval_samples_per_second": 33.044, "eval_steps_per_second": 1.045, "step": 931 }, { "epoch": 0.3522179135533907, "eval_bench_accuracy_arc_challenge": 0.2, "eval_bench_accuracy_hellaswag": 0.265, "eval_bench_accuracy_mmlu": 0.20869565217391303, "eval_bench_average_accuracy": 0.22456521739130433, "eval_bench_loss": 4.116911503306606, "eval_bench_total_accuracy": 0.23076923076923078, "step": 931 }, { "epoch": 0.352596235694694, "grad_norm": 1.071445968085627, "learning_rate": 7.993234530962498e-06, "loss": 0.8349, "step": 932 }, { "epoch": 0.35297455783599735, "grad_norm": 1.1138872222419933, "learning_rate": 7.993131665028082e-06, "loss": 0.8369, "step": 933 }, { "epoch": 0.3533528799773007, "grad_norm": 1.034081458809988, "learning_rate": 7.993028023639493e-06, "loss": 0.8302, "step": 934 }, { "epoch": 0.353731202118604, "grad_norm": 1.0615568247982479, "learning_rate": 7.992923606816852e-06, "loss": 0.7956, "step": 935 }, { "epoch": 0.3541095242599073, "grad_norm": 1.0966324306911683, "learning_rate": 7.992818414580439e-06, "loss": 0.8157, "step": 936 }, { "epoch": 0.3544878464012106, "grad_norm": 1.0499428116789347, "learning_rate": 7.992712446950682e-06, "loss": 0.8448, "step": 937 }, { "epoch": 0.35486616854251396, "grad_norm": 1.0929166781794446, "learning_rate": 7.99260570394816e-06, "loss": 0.838, "step": 938 }, { "epoch": 0.35524449068381725, "grad_norm": 1.0784478665113866, "learning_rate": 7.9924981855936e-06, "loss": 0.8477, "step": 939 }, { "epoch": 0.3556228128251206, "grad_norm": 1.112873701673093, "learning_rate": 7.992389891907885e-06, "loss": 0.837, "step": 940 }, { "epoch": 0.3560011349664239, "grad_norm": 1.0396578216523251, "learning_rate": 7.992280822912044e-06, "loss": 0.7867, "step": 941 }, { "epoch": 0.3563794571077272, "grad_norm": 1.1025438788531285, "learning_rate": 7.992170978627258e-06, "loss": 0.8588, "step": 942 }, { "epoch": 0.35675777924903057, "grad_norm": 1.0567533995232752, "learning_rate": 7.992060359074857e-06, "loss": 0.8415, "step": 943 }, { "epoch": 0.35713610139033386, "grad_norm": 1.0876544163342308, "learning_rate": 7.991948964276324e-06, "loss": 0.8139, "step": 944 }, { "epoch": 0.3575144235316372, "grad_norm": 1.1119965568409491, "learning_rate": 7.991836794253291e-06, "loss": 0.8236, "step": 945 }, { "epoch": 0.3578927456729405, "grad_norm": 1.050449035576396, "learning_rate": 7.991723849027543e-06, "loss": 0.8683, "step": 946 }, { "epoch": 0.35827106781424384, "grad_norm": 1.0727809938491701, "learning_rate": 7.991610128621012e-06, "loss": 0.8637, "step": 947 }, { "epoch": 0.3586493899555471, "grad_norm": 1.1142250081446294, "learning_rate": 7.991495633055782e-06, "loss": 0.8173, "step": 948 }, { "epoch": 0.35902771209685047, "grad_norm": 1.0422992081938323, "learning_rate": 7.99138036235409e-06, "loss": 0.8247, "step": 949 }, { "epoch": 0.3594060342381538, "grad_norm": 1.0683985452632145, "learning_rate": 7.991264316538315e-06, "loss": 0.7835, "step": 950 }, { "epoch": 0.3597843563794571, "grad_norm": 1.1389275468673155, "learning_rate": 7.991147495631001e-06, "loss": 0.8263, "step": 951 }, { "epoch": 0.36016267852076045, "grad_norm": 1.0300732494637694, "learning_rate": 7.99102989965483e-06, "loss": 0.8382, "step": 952 }, { "epoch": 0.36054100066206374, "grad_norm": 1.1134877059951171, "learning_rate": 7.990911528632637e-06, "loss": 0.8301, "step": 953 }, { "epoch": 0.3609193228033671, "grad_norm": 1.1556214956120872, "learning_rate": 7.990792382587413e-06, "loss": 0.8339, "step": 954 }, { "epoch": 0.36129764494467037, "grad_norm": 1.0496596260111375, "learning_rate": 7.990672461542295e-06, "loss": 0.855, "step": 955 }, { "epoch": 0.3616759670859737, "grad_norm": 1.0631933354628074, "learning_rate": 7.99055176552057e-06, "loss": 0.8028, "step": 956 }, { "epoch": 0.362054289227277, "grad_norm": 1.112630845203049, "learning_rate": 7.990430294545676e-06, "loss": 0.8324, "step": 957 }, { "epoch": 0.36243261136858035, "grad_norm": 1.047199242259213, "learning_rate": 7.990308048641205e-06, "loss": 0.8113, "step": 958 }, { "epoch": 0.3628109335098837, "grad_norm": 1.027441822648717, "learning_rate": 7.990185027830895e-06, "loss": 0.818, "step": 959 }, { "epoch": 0.363189255651187, "grad_norm": 1.1215384265121908, "learning_rate": 7.990061232138636e-06, "loss": 0.8105, "step": 960 }, { "epoch": 0.3635675777924903, "grad_norm": 1.068442952320319, "learning_rate": 7.989936661588471e-06, "loss": 0.7921, "step": 961 }, { "epoch": 0.3639458999337936, "grad_norm": 1.1092839541563482, "learning_rate": 7.989811316204588e-06, "loss": 0.8604, "step": 962 }, { "epoch": 0.36432422207509696, "grad_norm": 1.071801311807864, "learning_rate": 7.989685196011332e-06, "loss": 0.8309, "step": 963 }, { "epoch": 0.36470254421640025, "grad_norm": 1.0755045364863953, "learning_rate": 7.989558301033193e-06, "loss": 0.8281, "step": 964 }, { "epoch": 0.3650808663577036, "grad_norm": 1.0267320983799983, "learning_rate": 7.989430631294813e-06, "loss": 0.8354, "step": 965 }, { "epoch": 0.3654591884990069, "grad_norm": 1.137253491825624, "learning_rate": 7.98930218682099e-06, "loss": 0.879, "step": 966 }, { "epoch": 0.3658375106403102, "grad_norm": 1.078336142946193, "learning_rate": 7.989172967636661e-06, "loss": 0.7937, "step": 967 }, { "epoch": 0.36621583278161357, "grad_norm": 1.249220122221408, "learning_rate": 7.98904297376692e-06, "loss": 0.8719, "step": 968 }, { "epoch": 0.36659415492291686, "grad_norm": 1.0553052489470098, "learning_rate": 7.988912205237018e-06, "loss": 0.8343, "step": 969 }, { "epoch": 0.3669724770642202, "grad_norm": 1.0825650361601242, "learning_rate": 7.988780662072345e-06, "loss": 0.8708, "step": 970 }, { "epoch": 0.3673507992055235, "grad_norm": 1.0492113257783737, "learning_rate": 7.988648344298449e-06, "loss": 0.8158, "step": 971 }, { "epoch": 0.36772912134682684, "grad_norm": 1.1098170719484017, "learning_rate": 7.988515251941022e-06, "loss": 0.8072, "step": 972 }, { "epoch": 0.3681074434881301, "grad_norm": 1.0470408388006793, "learning_rate": 7.988381385025913e-06, "loss": 0.8254, "step": 973 }, { "epoch": 0.36848576562943347, "grad_norm": 1.1223023650314936, "learning_rate": 7.988246743579118e-06, "loss": 0.8422, "step": 974 }, { "epoch": 0.36886408777073676, "grad_norm": 1.0378189816707217, "learning_rate": 7.988111327626781e-06, "loss": 0.7986, "step": 975 }, { "epoch": 0.3692424099120401, "grad_norm": 1.0879026599404655, "learning_rate": 7.987975137195206e-06, "loss": 0.8239, "step": 976 }, { "epoch": 0.36962073205334345, "grad_norm": 1.0445944467404071, "learning_rate": 7.987838172310836e-06, "loss": 0.7856, "step": 977 }, { "epoch": 0.36999905419464674, "grad_norm": 1.0952504464513027, "learning_rate": 7.987700433000268e-06, "loss": 0.8474, "step": 978 }, { "epoch": 0.3703773763359501, "grad_norm": 1.0976482765823483, "learning_rate": 7.987561919290254e-06, "loss": 0.8067, "step": 979 }, { "epoch": 0.37075569847725337, "grad_norm": 1.0673215016151512, "learning_rate": 7.987422631207691e-06, "loss": 0.7747, "step": 980 }, { "epoch": 0.3711340206185567, "grad_norm": 1.1205110055136513, "learning_rate": 7.98728256877963e-06, "loss": 0.7892, "step": 981 }, { "epoch": 0.37151234275986, "grad_norm": 1.092436787430483, "learning_rate": 7.987141732033268e-06, "loss": 0.8332, "step": 982 }, { "epoch": 0.37189066490116335, "grad_norm": 1.091564370951629, "learning_rate": 7.987000120995958e-06, "loss": 0.8318, "step": 983 }, { "epoch": 0.37226898704246664, "grad_norm": 1.0840271784135682, "learning_rate": 7.986857735695197e-06, "loss": 0.8343, "step": 984 }, { "epoch": 0.37264730918377, "grad_norm": 1.1224128911012572, "learning_rate": 7.98671457615864e-06, "loss": 0.8084, "step": 985 }, { "epoch": 0.3730256313250733, "grad_norm": 1.0744788507306402, "learning_rate": 7.986570642414086e-06, "loss": 0.8468, "step": 986 }, { "epoch": 0.3734039534663766, "grad_norm": 1.0627524449061605, "learning_rate": 7.986425934489486e-06, "loss": 0.794, "step": 987 }, { "epoch": 0.37378227560767996, "grad_norm": 1.1606049685680029, "learning_rate": 7.986280452412942e-06, "loss": 0.8599, "step": 988 }, { "epoch": 0.37416059774898325, "grad_norm": 1.1453346028219251, "learning_rate": 7.986134196212707e-06, "loss": 0.839, "step": 989 }, { "epoch": 0.3745389198902866, "grad_norm": 1.047560845313498, "learning_rate": 7.985987165917182e-06, "loss": 0.838, "step": 990 }, { "epoch": 0.3749172420315899, "grad_norm": 1.0691648190671164, "learning_rate": 7.985839361554922e-06, "loss": 0.8349, "step": 991 }, { "epoch": 0.3752955641728932, "grad_norm": 1.0728147519090105, "learning_rate": 7.985690783154628e-06, "loss": 0.8082, "step": 992 }, { "epoch": 0.3756738863141965, "grad_norm": 1.0710609346244502, "learning_rate": 7.985541430745155e-06, "loss": 0.8367, "step": 993 }, { "epoch": 0.37605220845549986, "grad_norm": 1.0345097180466358, "learning_rate": 7.985391304355508e-06, "loss": 0.8235, "step": 994 }, { "epoch": 0.3764305305968032, "grad_norm": 1.0627329252549442, "learning_rate": 7.985240404014836e-06, "loss": 0.8361, "step": 995 }, { "epoch": 0.3768088527381065, "grad_norm": 1.055170154515539, "learning_rate": 7.98508872975245e-06, "loss": 0.7913, "step": 996 }, { "epoch": 0.37718717487940984, "grad_norm": 1.0799095201174227, "learning_rate": 7.9849362815978e-06, "loss": 0.8143, "step": 997 }, { "epoch": 0.3775654970207131, "grad_norm": 1.1004168575034028, "learning_rate": 7.984783059580493e-06, "loss": 0.8325, "step": 998 }, { "epoch": 0.37794381916201647, "grad_norm": 1.064297565177233, "learning_rate": 7.984629063730284e-06, "loss": 0.7825, "step": 999 }, { "epoch": 0.37832214130331976, "grad_norm": 1.0635329039354893, "learning_rate": 7.984474294077078e-06, "loss": 0.843, "step": 1000 }, { "epoch": 0.3787004634446231, "grad_norm": 1.0134149947950788, "learning_rate": 7.98431875065093e-06, "loss": 0.8407, "step": 1001 }, { "epoch": 0.3790787855859264, "grad_norm": 1.1003240739229772, "learning_rate": 7.984162433482048e-06, "loss": 0.8757, "step": 1002 }, { "epoch": 0.37945710772722974, "grad_norm": 1.0704123729576063, "learning_rate": 7.984005342600789e-06, "loss": 0.8385, "step": 1003 }, { "epoch": 0.3798354298685331, "grad_norm": 1.082489049237877, "learning_rate": 7.983847478037655e-06, "loss": 0.8494, "step": 1004 }, { "epoch": 0.38021375200983637, "grad_norm": 1.080752264367249, "learning_rate": 7.983688839823308e-06, "loss": 0.8609, "step": 1005 }, { "epoch": 0.3805920741511397, "grad_norm": 1.1968418204384677, "learning_rate": 7.983529427988552e-06, "loss": 0.8564, "step": 1006 }, { "epoch": 0.380970396292443, "grad_norm": 1.061469890379153, "learning_rate": 7.983369242564346e-06, "loss": 0.7891, "step": 1007 }, { "epoch": 0.38134871843374635, "grad_norm": 1.0621745023983624, "learning_rate": 7.983208283581796e-06, "loss": 0.864, "step": 1008 }, { "epoch": 0.38172704057504964, "grad_norm": 1.1002758271639341, "learning_rate": 7.98304655107216e-06, "loss": 0.8511, "step": 1009 }, { "epoch": 0.382105362716353, "grad_norm": 1.2982365803931801, "learning_rate": 7.982884045066848e-06, "loss": 0.8707, "step": 1010 }, { "epoch": 0.38248368485765627, "grad_norm": 1.0481998500890215, "learning_rate": 7.982720765597416e-06, "loss": 0.808, "step": 1011 }, { "epoch": 0.3828620069989596, "grad_norm": 1.0843657280284922, "learning_rate": 7.982556712695573e-06, "loss": 0.8033, "step": 1012 }, { "epoch": 0.38324032914026296, "grad_norm": 1.056797859890995, "learning_rate": 7.982391886393176e-06, "loss": 0.8109, "step": 1013 }, { "epoch": 0.38361865128156625, "grad_norm": 1.060307047043872, "learning_rate": 7.982226286722239e-06, "loss": 0.8485, "step": 1014 }, { "epoch": 0.3839969734228696, "grad_norm": 1.0880414860647125, "learning_rate": 7.982059913714915e-06, "loss": 0.829, "step": 1015 }, { "epoch": 0.3843752955641729, "grad_norm": 1.0647653565219015, "learning_rate": 7.981892767403516e-06, "loss": 0.831, "step": 1016 }, { "epoch": 0.3847536177054762, "grad_norm": 1.1245340497823308, "learning_rate": 7.9817248478205e-06, "loss": 0.8633, "step": 1017 }, { "epoch": 0.3851319398467795, "grad_norm": 1.083643967559738, "learning_rate": 7.981556154998477e-06, "loss": 0.8694, "step": 1018 }, { "epoch": 0.38551026198808286, "grad_norm": 1.0892685401414424, "learning_rate": 7.981386688970209e-06, "loss": 0.8455, "step": 1019 }, { "epoch": 0.38588858412938615, "grad_norm": 1.080573813534876, "learning_rate": 7.981216449768603e-06, "loss": 0.8028, "step": 1020 }, { "epoch": 0.3862669062706895, "grad_norm": 1.0697257333484091, "learning_rate": 7.981045437426718e-06, "loss": 0.8254, "step": 1021 }, { "epoch": 0.38664522841199284, "grad_norm": 1.1482898982014345, "learning_rate": 7.980873651977768e-06, "loss": 0.8434, "step": 1022 }, { "epoch": 0.3870235505532961, "grad_norm": 1.066295131291774, "learning_rate": 7.98070109345511e-06, "loss": 0.7966, "step": 1023 }, { "epoch": 0.38740187269459947, "grad_norm": 1.0329631074824188, "learning_rate": 7.980527761892255e-06, "loss": 0.7914, "step": 1024 }, { "epoch": 0.38778019483590276, "grad_norm": 1.0857069666875103, "learning_rate": 7.980353657322863e-06, "loss": 0.8622, "step": 1025 }, { "epoch": 0.3881585169772061, "grad_norm": 1.060211010001084, "learning_rate": 7.980178779780747e-06, "loss": 0.8381, "step": 1026 }, { "epoch": 0.3885368391185094, "grad_norm": 1.0543634996329088, "learning_rate": 7.980003129299865e-06, "loss": 0.8378, "step": 1027 }, { "epoch": 0.38891516125981274, "grad_norm": 1.1081388338013471, "learning_rate": 7.979826705914328e-06, "loss": 0.8338, "step": 1028 }, { "epoch": 0.389293483401116, "grad_norm": 1.104557100267363, "learning_rate": 7.9796495096584e-06, "loss": 0.795, "step": 1029 }, { "epoch": 0.38967180554241937, "grad_norm": 1.0655072241835162, "learning_rate": 7.979471540566489e-06, "loss": 0.8237, "step": 1030 }, { "epoch": 0.3900501276837227, "grad_norm": 1.0796326933387017, "learning_rate": 7.979292798673156e-06, "loss": 0.8556, "step": 1031 }, { "epoch": 0.390428449825026, "grad_norm": 1.0380712383913533, "learning_rate": 7.979113284013114e-06, "loss": 0.839, "step": 1032 }, { "epoch": 0.39080677196632935, "grad_norm": 1.085425876568373, "learning_rate": 7.97893299662122e-06, "loss": 0.8516, "step": 1033 }, { "epoch": 0.39118509410763264, "grad_norm": 1.2207322749435598, "learning_rate": 7.978751936532491e-06, "loss": 0.8549, "step": 1034 }, { "epoch": 0.391563416248936, "grad_norm": 1.088319428223248, "learning_rate": 7.978570103782086e-06, "loss": 0.8573, "step": 1035 }, { "epoch": 0.39194173839023927, "grad_norm": 1.0545678177926456, "learning_rate": 7.978387498405317e-06, "loss": 0.8325, "step": 1036 }, { "epoch": 0.3923200605315426, "grad_norm": 1.0921146086499482, "learning_rate": 7.978204120437641e-06, "loss": 0.7912, "step": 1037 }, { "epoch": 0.3926983826728459, "grad_norm": 1.1156394836322963, "learning_rate": 7.978019969914676e-06, "loss": 0.8344, "step": 1038 }, { "epoch": 0.39307670481414925, "grad_norm": 1.1163141481746923, "learning_rate": 7.97783504687218e-06, "loss": 0.8039, "step": 1039 }, { "epoch": 0.3934550269554526, "grad_norm": 1.1055832393565042, "learning_rate": 7.977649351346065e-06, "loss": 0.8098, "step": 1040 }, { "epoch": 0.3938333490967559, "grad_norm": 1.0475102246909884, "learning_rate": 7.97746288337239e-06, "loss": 0.7868, "step": 1041 }, { "epoch": 0.3942116712380592, "grad_norm": 1.0630199431469338, "learning_rate": 7.977275642987371e-06, "loss": 0.7965, "step": 1042 }, { "epoch": 0.3945899933793625, "grad_norm": 1.1096476912788604, "learning_rate": 7.977087630227368e-06, "loss": 0.8052, "step": 1043 }, { "epoch": 0.39496831552066586, "grad_norm": 1.0863091134871783, "learning_rate": 7.976898845128891e-06, "loss": 0.8435, "step": 1044 }, { "epoch": 0.39534663766196915, "grad_norm": 1.0492836175021802, "learning_rate": 7.976709287728602e-06, "loss": 0.8083, "step": 1045 }, { "epoch": 0.3957249598032725, "grad_norm": 1.0529300466346392, "learning_rate": 7.976518958063315e-06, "loss": 0.8274, "step": 1046 }, { "epoch": 0.3961032819445758, "grad_norm": 1.070473727548606, "learning_rate": 7.976327856169989e-06, "loss": 0.7971, "step": 1047 }, { "epoch": 0.3964816040858791, "grad_norm": 1.0617092300636013, "learning_rate": 7.976135982085734e-06, "loss": 0.8536, "step": 1048 }, { "epoch": 0.39685992622718247, "grad_norm": 1.0606504595804507, "learning_rate": 7.975943335847815e-06, "loss": 0.777, "step": 1049 }, { "epoch": 0.39723824836848576, "grad_norm": 1.1335961432026964, "learning_rate": 7.97574991749364e-06, "loss": 0.8707, "step": 1050 }, { "epoch": 0.3976165705097891, "grad_norm": 1.0932495202458485, "learning_rate": 7.975555727060773e-06, "loss": 0.8476, "step": 1051 }, { "epoch": 0.3979948926510924, "grad_norm": 1.0904729718461323, "learning_rate": 7.975360764586923e-06, "loss": 0.8325, "step": 1052 }, { "epoch": 0.39837321479239574, "grad_norm": 1.060481887356713, "learning_rate": 7.975165030109953e-06, "loss": 0.8293, "step": 1053 }, { "epoch": 0.398751536933699, "grad_norm": 1.0594136483291037, "learning_rate": 7.974968523667874e-06, "loss": 0.8333, "step": 1054 }, { "epoch": 0.39912985907500237, "grad_norm": 1.072066755016977, "learning_rate": 7.974771245298845e-06, "loss": 0.8588, "step": 1055 }, { "epoch": 0.39950818121630566, "grad_norm": 1.0407488984374065, "learning_rate": 7.974573195041179e-06, "loss": 0.8119, "step": 1056 }, { "epoch": 0.399886503357609, "grad_norm": 1.0897696384583164, "learning_rate": 7.974374372933333e-06, "loss": 0.8729, "step": 1057 }, { "epoch": 0.40026482549891235, "grad_norm": 1.0395716067441272, "learning_rate": 7.974174779013923e-06, "loss": 0.844, "step": 1058 }, { "epoch": 0.40064314764021564, "grad_norm": 1.0440432063315428, "learning_rate": 7.973974413321706e-06, "loss": 0.8311, "step": 1059 }, { "epoch": 0.401021469781519, "grad_norm": 1.085811930524537, "learning_rate": 7.973773275895593e-06, "loss": 0.8506, "step": 1060 }, { "epoch": 0.40139979192282227, "grad_norm": 1.017123583458792, "learning_rate": 7.973571366774646e-06, "loss": 0.7491, "step": 1061 }, { "epoch": 0.4017781140641256, "grad_norm": 1.041022717188848, "learning_rate": 7.973368685998074e-06, "loss": 0.8189, "step": 1062 }, { "epoch": 0.4021564362054289, "grad_norm": 1.0150607929017172, "learning_rate": 7.973165233605234e-06, "loss": 0.814, "step": 1063 }, { "epoch": 0.40253475834673225, "grad_norm": 1.0458554860554623, "learning_rate": 7.972961009635642e-06, "loss": 0.8123, "step": 1064 }, { "epoch": 0.40253475834673225, "eval_loss": 0.8304316997528076, "eval_runtime": 26.6669, "eval_samples_per_second": 33.187, "eval_steps_per_second": 1.05, "step": 1064 }, { "epoch": 0.40253475834673225, "eval_bench_accuracy_arc_challenge": 0.25, "eval_bench_accuracy_hellaswag": 0.285, "eval_bench_accuracy_mmlu": 0.2782608695652174, "eval_bench_average_accuracy": 0.2710869565217391, "eval_bench_loss": 4.517480147512336, "eval_bench_total_accuracy": 0.2725274725274725, "step": 1064 }, { "epoch": 0.40291308048803554, "grad_norm": 1.037409138160307, "learning_rate": 7.972756014128952e-06, "loss": 0.8159, "step": 1065 }, { "epoch": 0.4032914026293389, "grad_norm": 1.0836167448402902, "learning_rate": 7.972550247124976e-06, "loss": 0.8131, "step": 1066 }, { "epoch": 0.4036697247706422, "grad_norm": 1.0933137283571555, "learning_rate": 7.972343708663674e-06, "loss": 0.8183, "step": 1067 }, { "epoch": 0.4040480469119455, "grad_norm": 1.03216484709328, "learning_rate": 7.972136398785154e-06, "loss": 0.8569, "step": 1068 }, { "epoch": 0.40442636905324886, "grad_norm": 1.0656155608965763, "learning_rate": 7.971928317529676e-06, "loss": 0.8453, "step": 1069 }, { "epoch": 0.40480469119455215, "grad_norm": 1.0708238570639999, "learning_rate": 7.971719464937647e-06, "loss": 0.8367, "step": 1070 }, { "epoch": 0.4051830133358555, "grad_norm": 1.0621498480602682, "learning_rate": 7.971509841049628e-06, "loss": 0.8589, "step": 1071 }, { "epoch": 0.4055613354771588, "grad_norm": 1.0072315129856741, "learning_rate": 7.971299445906324e-06, "loss": 0.8379, "step": 1072 }, { "epoch": 0.4059396576184621, "grad_norm": 1.033456153626471, "learning_rate": 7.971088279548597e-06, "loss": 0.8079, "step": 1073 }, { "epoch": 0.4063179797597654, "grad_norm": 1.0079272901425842, "learning_rate": 7.970876342017452e-06, "loss": 0.7868, "step": 1074 }, { "epoch": 0.40669630190106876, "grad_norm": 1.0073805003714849, "learning_rate": 7.970663633354047e-06, "loss": 0.7988, "step": 1075 }, { "epoch": 0.4070746240423721, "grad_norm": 1.0708487426838318, "learning_rate": 7.97045015359969e-06, "loss": 0.8026, "step": 1076 }, { "epoch": 0.4074529461836754, "grad_norm": 1.069671541329999, "learning_rate": 7.970235902795838e-06, "loss": 0.8462, "step": 1077 }, { "epoch": 0.40783126832497874, "grad_norm": 1.0250427566221285, "learning_rate": 7.9700208809841e-06, "loss": 0.819, "step": 1078 }, { "epoch": 0.408209590466282, "grad_norm": 1.035811754086645, "learning_rate": 7.969805088206226e-06, "loss": 0.8192, "step": 1079 }, { "epoch": 0.40858791260758537, "grad_norm": 1.0919846226041652, "learning_rate": 7.96958852450413e-06, "loss": 0.8463, "step": 1080 }, { "epoch": 0.40896623474888866, "grad_norm": 1.0922304905923719, "learning_rate": 7.969371189919865e-06, "loss": 0.8505, "step": 1081 }, { "epoch": 0.409344556890192, "grad_norm": 1.0327335666733615, "learning_rate": 7.969153084495636e-06, "loss": 0.8054, "step": 1082 }, { "epoch": 0.4097228790314953, "grad_norm": 1.069756821894608, "learning_rate": 7.968934208273798e-06, "loss": 0.8348, "step": 1083 }, { "epoch": 0.41010120117279864, "grad_norm": 1.0472686446394408, "learning_rate": 7.968714561296859e-06, "loss": 0.8302, "step": 1084 }, { "epoch": 0.410479523314102, "grad_norm": 1.0462638623089058, "learning_rate": 7.96849414360747e-06, "loss": 0.8249, "step": 1085 }, { "epoch": 0.41085784545540527, "grad_norm": 1.0056327093077677, "learning_rate": 7.96827295524844e-06, "loss": 0.7795, "step": 1086 }, { "epoch": 0.4112361675967086, "grad_norm": 1.0244037556207601, "learning_rate": 7.968050996262716e-06, "loss": 0.7905, "step": 1087 }, { "epoch": 0.4116144897380119, "grad_norm": 1.0346973741005767, "learning_rate": 7.967828266693409e-06, "loss": 0.8371, "step": 1088 }, { "epoch": 0.41199281187931525, "grad_norm": 1.0958021967982934, "learning_rate": 7.96760476658377e-06, "loss": 0.8479, "step": 1089 }, { "epoch": 0.41237113402061853, "grad_norm": 1.0136255102022522, "learning_rate": 7.967380495977201e-06, "loss": 0.8055, "step": 1090 }, { "epoch": 0.4127494561619219, "grad_norm": 1.0687414316917077, "learning_rate": 7.967155454917255e-06, "loss": 0.8481, "step": 1091 }, { "epoch": 0.4131277783032252, "grad_norm": 1.0765456661292323, "learning_rate": 7.966929643447634e-06, "loss": 0.8115, "step": 1092 }, { "epoch": 0.4135061004445285, "grad_norm": 1.078258124622418, "learning_rate": 7.966703061612192e-06, "loss": 0.8319, "step": 1093 }, { "epoch": 0.41388442258583186, "grad_norm": 1.0491237525414794, "learning_rate": 7.966475709454928e-06, "loss": 0.8592, "step": 1094 }, { "epoch": 0.41426274472713515, "grad_norm": 1.0719668981104609, "learning_rate": 7.966247587019994e-06, "loss": 0.821, "step": 1095 }, { "epoch": 0.4146410668684385, "grad_norm": 1.026254989024167, "learning_rate": 7.966018694351691e-06, "loss": 0.8168, "step": 1096 }, { "epoch": 0.4150193890097418, "grad_norm": 1.0321711854785867, "learning_rate": 7.96578903149447e-06, "loss": 0.8255, "step": 1097 }, { "epoch": 0.4153977111510451, "grad_norm": 1.0513898483857722, "learning_rate": 7.965558598492929e-06, "loss": 0.7748, "step": 1098 }, { "epoch": 0.4157760332923484, "grad_norm": 1.0364175851458883, "learning_rate": 7.965327395391819e-06, "loss": 0.7978, "step": 1099 }, { "epoch": 0.41615435543365176, "grad_norm": 0.985307760157813, "learning_rate": 7.965095422236038e-06, "loss": 0.801, "step": 1100 }, { "epoch": 0.4165326775749551, "grad_norm": 1.0813628193591218, "learning_rate": 7.964862679070634e-06, "loss": 0.845, "step": 1101 }, { "epoch": 0.4169109997162584, "grad_norm": 1.0734207809402587, "learning_rate": 7.964629165940808e-06, "loss": 0.8817, "step": 1102 }, { "epoch": 0.41728932185756173, "grad_norm": 1.0599230797124688, "learning_rate": 7.964394882891904e-06, "loss": 0.8085, "step": 1103 }, { "epoch": 0.417667643998865, "grad_norm": 1.078793670107089, "learning_rate": 7.96415982996942e-06, "loss": 0.7938, "step": 1104 }, { "epoch": 0.41804596614016837, "grad_norm": 1.0350357122236093, "learning_rate": 7.963924007219002e-06, "loss": 0.8207, "step": 1105 }, { "epoch": 0.41842428828147166, "grad_norm": 1.041240999715739, "learning_rate": 7.963687414686449e-06, "loss": 0.7737, "step": 1106 }, { "epoch": 0.418802610422775, "grad_norm": 1.1066667842190356, "learning_rate": 7.963450052417703e-06, "loss": 0.8191, "step": 1107 }, { "epoch": 0.4191809325640783, "grad_norm": 1.0866062695241046, "learning_rate": 7.963211920458863e-06, "loss": 0.8098, "step": 1108 }, { "epoch": 0.41955925470538163, "grad_norm": 1.0628974307927237, "learning_rate": 7.962973018856169e-06, "loss": 0.836, "step": 1109 }, { "epoch": 0.419937576846685, "grad_norm": 1.0490148472801595, "learning_rate": 7.962733347656018e-06, "loss": 0.8074, "step": 1110 }, { "epoch": 0.42031589898798827, "grad_norm": 1.056521276681419, "learning_rate": 7.962492906904953e-06, "loss": 0.7798, "step": 1111 }, { "epoch": 0.4206942211292916, "grad_norm": 1.0568484786859005, "learning_rate": 7.962251696649665e-06, "loss": 0.832, "step": 1112 }, { "epoch": 0.4210725432705949, "grad_norm": 1.022548771593414, "learning_rate": 7.962009716937e-06, "loss": 0.8576, "step": 1113 }, { "epoch": 0.42145086541189825, "grad_norm": 1.0376517279626776, "learning_rate": 7.961766967813946e-06, "loss": 0.7709, "step": 1114 }, { "epoch": 0.42182918755320153, "grad_norm": 1.057176802372392, "learning_rate": 7.961523449327646e-06, "loss": 0.8684, "step": 1115 }, { "epoch": 0.4222075096945049, "grad_norm": 1.0278310719203412, "learning_rate": 7.961279161525389e-06, "loss": 0.7934, "step": 1116 }, { "epoch": 0.42258583183580817, "grad_norm": 1.0116937469277474, "learning_rate": 7.961034104454618e-06, "loss": 0.8288, "step": 1117 }, { "epoch": 0.4229641539771115, "grad_norm": 1.0791508367529585, "learning_rate": 7.960788278162918e-06, "loss": 0.8295, "step": 1118 }, { "epoch": 0.42334247611841486, "grad_norm": 1.0482664569638203, "learning_rate": 7.960541682698034e-06, "loss": 0.8044, "step": 1119 }, { "epoch": 0.42372079825971815, "grad_norm": 1.026033507367731, "learning_rate": 7.960294318107847e-06, "loss": 0.8086, "step": 1120 }, { "epoch": 0.4240991204010215, "grad_norm": 1.0713832704640005, "learning_rate": 7.960046184440399e-06, "loss": 0.8421, "step": 1121 }, { "epoch": 0.4244774425423248, "grad_norm": 1.0635267452769637, "learning_rate": 7.959797281743876e-06, "loss": 0.8452, "step": 1122 }, { "epoch": 0.4248557646836281, "grad_norm": 1.046318335512741, "learning_rate": 7.959547610066613e-06, "loss": 0.7944, "step": 1123 }, { "epoch": 0.4252340868249314, "grad_norm": 1.0788089412291229, "learning_rate": 7.959297169457097e-06, "loss": 0.8338, "step": 1124 }, { "epoch": 0.42561240896623476, "grad_norm": 1.0582140885008549, "learning_rate": 7.959045959963962e-06, "loss": 0.7914, "step": 1125 }, { "epoch": 0.42599073110753805, "grad_norm": 1.0773203264262958, "learning_rate": 7.958793981635991e-06, "loss": 0.8549, "step": 1126 }, { "epoch": 0.4263690532488414, "grad_norm": 1.0738918058139102, "learning_rate": 7.958541234522119e-06, "loss": 0.7836, "step": 1127 }, { "epoch": 0.42674737539014473, "grad_norm": 1.0307363548970123, "learning_rate": 7.958287718671429e-06, "loss": 0.829, "step": 1128 }, { "epoch": 0.427125697531448, "grad_norm": 1.0223432647328048, "learning_rate": 7.958033434133152e-06, "loss": 0.8421, "step": 1129 }, { "epoch": 0.42750401967275137, "grad_norm": 1.0402584891579054, "learning_rate": 7.95777838095667e-06, "loss": 0.7836, "step": 1130 }, { "epoch": 0.42788234181405466, "grad_norm": 1.0761841482760737, "learning_rate": 7.957522559191514e-06, "loss": 0.7933, "step": 1131 }, { "epoch": 0.428260663955358, "grad_norm": 1.0391476619745978, "learning_rate": 7.957265968887361e-06, "loss": 0.811, "step": 1132 }, { "epoch": 0.4286389860966613, "grad_norm": 1.026814188051067, "learning_rate": 7.957008610094043e-06, "loss": 0.8078, "step": 1133 }, { "epoch": 0.42901730823796463, "grad_norm": 1.0406330571564124, "learning_rate": 7.956750482861538e-06, "loss": 0.8359, "step": 1134 }, { "epoch": 0.4293956303792679, "grad_norm": 1.0642979501183267, "learning_rate": 7.956491587239971e-06, "loss": 0.8045, "step": 1135 }, { "epoch": 0.42977395252057127, "grad_norm": 1.0393212545559525, "learning_rate": 7.956231923279624e-06, "loss": 0.8348, "step": 1136 }, { "epoch": 0.4301522746618746, "grad_norm": 1.0470124602821342, "learning_rate": 7.955971491030917e-06, "loss": 0.8148, "step": 1137 }, { "epoch": 0.4305305968031779, "grad_norm": 1.0676455383028118, "learning_rate": 7.955710290544428e-06, "loss": 0.8336, "step": 1138 }, { "epoch": 0.43090891894448125, "grad_norm": 1.0721667527067038, "learning_rate": 7.955448321870882e-06, "loss": 0.831, "step": 1139 }, { "epoch": 0.43128724108578453, "grad_norm": 1.064318000094558, "learning_rate": 7.955185585061151e-06, "loss": 0.8335, "step": 1140 }, { "epoch": 0.4316655632270879, "grad_norm": 1.0302584817777816, "learning_rate": 7.95492208016626e-06, "loss": 0.791, "step": 1141 }, { "epoch": 0.43204388536839117, "grad_norm": 1.0256366632375336, "learning_rate": 7.954657807237379e-06, "loss": 0.8253, "step": 1142 }, { "epoch": 0.4324222075096945, "grad_norm": 1.0251051777197329, "learning_rate": 7.954392766325828e-06, "loss": 0.8223, "step": 1143 }, { "epoch": 0.4328005296509978, "grad_norm": 1.045445405795435, "learning_rate": 7.954126957483077e-06, "loss": 0.7606, "step": 1144 }, { "epoch": 0.43317885179230115, "grad_norm": 1.0425200750958303, "learning_rate": 7.95386038076075e-06, "loss": 0.8537, "step": 1145 }, { "epoch": 0.4335571739336045, "grad_norm": 1.0419269404142824, "learning_rate": 7.953593036210611e-06, "loss": 0.8277, "step": 1146 }, { "epoch": 0.4339354960749078, "grad_norm": 1.084574429840746, "learning_rate": 7.953324923884578e-06, "loss": 0.803, "step": 1147 }, { "epoch": 0.4343138182162111, "grad_norm": 1.0419638253671073, "learning_rate": 7.953056043834717e-06, "loss": 0.8334, "step": 1148 }, { "epoch": 0.4346921403575144, "grad_norm": 1.0168098031537844, "learning_rate": 7.952786396113248e-06, "loss": 0.7849, "step": 1149 }, { "epoch": 0.43507046249881776, "grad_norm": 1.0391261866313206, "learning_rate": 7.95251598077253e-06, "loss": 0.792, "step": 1150 }, { "epoch": 0.43544878464012104, "grad_norm": 1.0145928185391837, "learning_rate": 7.95224479786508e-06, "loss": 0.8069, "step": 1151 }, { "epoch": 0.4358271067814244, "grad_norm": 1.0145834983924735, "learning_rate": 7.951972847443561e-06, "loss": 0.8045, "step": 1152 }, { "epoch": 0.4362054289227277, "grad_norm": 1.0385429868897398, "learning_rate": 7.951700129560786e-06, "loss": 0.8091, "step": 1153 }, { "epoch": 0.436583751064031, "grad_norm": 1.0484204110539974, "learning_rate": 7.951426644269712e-06, "loss": 0.8118, "step": 1154 }, { "epoch": 0.43696207320533437, "grad_norm": 1.059201104727976, "learning_rate": 7.951152391623452e-06, "loss": 0.8335, "step": 1155 }, { "epoch": 0.43734039534663766, "grad_norm": 1.0061721443896443, "learning_rate": 7.950877371675265e-06, "loss": 0.7489, "step": 1156 }, { "epoch": 0.437718717487941, "grad_norm": 1.0920232553881484, "learning_rate": 7.950601584478557e-06, "loss": 0.8012, "step": 1157 }, { "epoch": 0.4380970396292443, "grad_norm": 1.0519115174631195, "learning_rate": 7.950325030086889e-06, "loss": 0.7923, "step": 1158 }, { "epoch": 0.43847536177054763, "grad_norm": 1.0813679052789027, "learning_rate": 7.950047708553962e-06, "loss": 0.8313, "step": 1159 }, { "epoch": 0.4388536839118509, "grad_norm": 1.0854599046397435, "learning_rate": 7.949769619933634e-06, "loss": 0.8616, "step": 1160 }, { "epoch": 0.43923200605315427, "grad_norm": 1.1104488658598137, "learning_rate": 7.94949076427991e-06, "loss": 0.7878, "step": 1161 }, { "epoch": 0.43961032819445756, "grad_norm": 1.1346641422155257, "learning_rate": 7.949211141646941e-06, "loss": 0.8287, "step": 1162 }, { "epoch": 0.4399886503357609, "grad_norm": 1.0632008460543734, "learning_rate": 7.948930752089029e-06, "loss": 0.8278, "step": 1163 }, { "epoch": 0.44036697247706424, "grad_norm": 1.0770714736885665, "learning_rate": 7.948649595660626e-06, "loss": 0.794, "step": 1164 }, { "epoch": 0.44074529461836753, "grad_norm": 1.0320296674718166, "learning_rate": 7.948367672416329e-06, "loss": 0.7973, "step": 1165 }, { "epoch": 0.4411236167596709, "grad_norm": 1.037195297637391, "learning_rate": 7.94808498241089e-06, "loss": 0.8124, "step": 1166 }, { "epoch": 0.44150193890097417, "grad_norm": 1.07174382564237, "learning_rate": 7.947801525699204e-06, "loss": 0.8501, "step": 1167 }, { "epoch": 0.4418802610422775, "grad_norm": 1.0423383360705205, "learning_rate": 7.947517302336321e-06, "loss": 0.8023, "step": 1168 }, { "epoch": 0.4422585831835808, "grad_norm": 1.0225149206809994, "learning_rate": 7.947232312377431e-06, "loss": 0.8082, "step": 1169 }, { "epoch": 0.44263690532488414, "grad_norm": 1.0490213514112987, "learning_rate": 7.946946555877883e-06, "loss": 0.8553, "step": 1170 }, { "epoch": 0.44301522746618743, "grad_norm": 1.0565295484573578, "learning_rate": 7.946660032893168e-06, "loss": 0.8334, "step": 1171 }, { "epoch": 0.4433935496074908, "grad_norm": 1.096379949923879, "learning_rate": 7.946372743478928e-06, "loss": 0.7885, "step": 1172 }, { "epoch": 0.4437718717487941, "grad_norm": 1.0635010257740696, "learning_rate": 7.946084687690952e-06, "loss": 0.867, "step": 1173 }, { "epoch": 0.4441501938900974, "grad_norm": 1.046045957242929, "learning_rate": 7.945795865585184e-06, "loss": 0.7794, "step": 1174 }, { "epoch": 0.44452851603140076, "grad_norm": 1.1358219370976814, "learning_rate": 7.945506277217707e-06, "loss": 0.8048, "step": 1175 }, { "epoch": 0.44490683817270404, "grad_norm": 1.0850391747638126, "learning_rate": 7.945215922644764e-06, "loss": 0.8056, "step": 1176 }, { "epoch": 0.4452851603140074, "grad_norm": 1.1532691295951847, "learning_rate": 7.944924801922734e-06, "loss": 0.8176, "step": 1177 }, { "epoch": 0.4456634824553107, "grad_norm": 1.0915907522482993, "learning_rate": 7.944632915108158e-06, "loss": 0.7994, "step": 1178 }, { "epoch": 0.446041804596614, "grad_norm": 1.0282978902411528, "learning_rate": 7.944340262257718e-06, "loss": 0.8263, "step": 1179 }, { "epoch": 0.4464201267379173, "grad_norm": 1.1021567277496518, "learning_rate": 7.944046843428244e-06, "loss": 0.829, "step": 1180 }, { "epoch": 0.44679844887922066, "grad_norm": 1.0694612963890957, "learning_rate": 7.94375265867672e-06, "loss": 0.8565, "step": 1181 }, { "epoch": 0.447176771020524, "grad_norm": 1.0750903881599976, "learning_rate": 7.943457708060272e-06, "loss": 0.8396, "step": 1182 }, { "epoch": 0.4475550931618273, "grad_norm": 1.0453024844416716, "learning_rate": 7.943161991636183e-06, "loss": 0.8096, "step": 1183 }, { "epoch": 0.44793341530313063, "grad_norm": 1.0657511458371332, "learning_rate": 7.942865509461879e-06, "loss": 0.7964, "step": 1184 }, { "epoch": 0.4483117374444339, "grad_norm": 1.0565556737130861, "learning_rate": 7.942568261594931e-06, "loss": 0.8254, "step": 1185 }, { "epoch": 0.44869005958573727, "grad_norm": 1.0811193147116154, "learning_rate": 7.942270248093072e-06, "loss": 0.8741, "step": 1186 }, { "epoch": 0.44906838172704056, "grad_norm": 1.0468093016525521, "learning_rate": 7.941971469014168e-06, "loss": 0.8379, "step": 1187 }, { "epoch": 0.4494467038683439, "grad_norm": 1.06315933336805, "learning_rate": 7.941671924416245e-06, "loss": 0.8294, "step": 1188 }, { "epoch": 0.4498250260096472, "grad_norm": 1.044215685157516, "learning_rate": 7.941371614357473e-06, "loss": 0.8093, "step": 1189 }, { "epoch": 0.45020334815095053, "grad_norm": 1.0172723595558777, "learning_rate": 7.941070538896172e-06, "loss": 0.777, "step": 1190 }, { "epoch": 0.4505816702922539, "grad_norm": 1.0750120304696666, "learning_rate": 7.940768698090809e-06, "loss": 0.8105, "step": 1191 }, { "epoch": 0.45095999243355717, "grad_norm": 1.0440692979176232, "learning_rate": 7.940466091999999e-06, "loss": 0.8537, "step": 1192 }, { "epoch": 0.4513383145748605, "grad_norm": 1.031643540251273, "learning_rate": 7.940162720682508e-06, "loss": 0.8362, "step": 1193 }, { "epoch": 0.4517166367161638, "grad_norm": 1.0019678147671374, "learning_rate": 7.939858584197252e-06, "loss": 0.8142, "step": 1194 }, { "epoch": 0.45209495885746714, "grad_norm": 1.060840824446392, "learning_rate": 7.939553682603292e-06, "loss": 0.7826, "step": 1195 }, { "epoch": 0.45247328099877043, "grad_norm": 1.0604407355830034, "learning_rate": 7.939248015959839e-06, "loss": 0.8276, "step": 1196 }, { "epoch": 0.4528516031400738, "grad_norm": 1.0445689437408072, "learning_rate": 7.938941584326251e-06, "loss": 0.7994, "step": 1197 }, { "epoch": 0.4528516031400738, "eval_loss": 0.8220446705818176, "eval_runtime": 26.7666, "eval_samples_per_second": 33.064, "eval_steps_per_second": 1.046, "step": 1197 }, { "epoch": 0.4528516031400738, "eval_bench_accuracy_arc_challenge": 0.2571428571428571, "eval_bench_accuracy_hellaswag": 0.225, "eval_bench_accuracy_mmlu": 0.23478260869565218, "eval_bench_average_accuracy": 0.23897515527950308, "eval_bench_loss": 5.286834716796875, "eval_bench_total_accuracy": 0.23736263736263735, "step": 1197 }, { "epoch": 0.45322992528137707, "grad_norm": 1.0158388274699295, "learning_rate": 7.938634387762039e-06, "loss": 0.8241, "step": 1198 }, { "epoch": 0.4536082474226804, "grad_norm": 1.165515743538843, "learning_rate": 7.938326426326857e-06, "loss": 0.8526, "step": 1199 }, { "epoch": 0.45398656956398376, "grad_norm": 1.0460295029244764, "learning_rate": 7.938017700080514e-06, "loss": 0.7998, "step": 1200 }, { "epoch": 0.45436489170528704, "grad_norm": 1.0837173342344641, "learning_rate": 7.93770820908296e-06, "loss": 0.7997, "step": 1201 }, { "epoch": 0.4547432138465904, "grad_norm": 1.0243169477083875, "learning_rate": 7.937397953394296e-06, "loss": 0.7991, "step": 1202 }, { "epoch": 0.4551215359878937, "grad_norm": 1.0695328376321132, "learning_rate": 7.937086933074777e-06, "loss": 0.7884, "step": 1203 }, { "epoch": 0.455499858129197, "grad_norm": 1.0594971537497897, "learning_rate": 7.9367751481848e-06, "loss": 0.793, "step": 1204 }, { "epoch": 0.4558781802705003, "grad_norm": 1.0554812656920887, "learning_rate": 7.936462598784913e-06, "loss": 0.8283, "step": 1205 }, { "epoch": 0.45625650241180365, "grad_norm": 1.0592140535117982, "learning_rate": 7.936149284935811e-06, "loss": 0.8323, "step": 1206 }, { "epoch": 0.45663482455310694, "grad_norm": 1.026196033728254, "learning_rate": 7.935835206698342e-06, "loss": 0.8024, "step": 1207 }, { "epoch": 0.4570131466944103, "grad_norm": 1.0292414805578125, "learning_rate": 7.935520364133494e-06, "loss": 0.7895, "step": 1208 }, { "epoch": 0.45739146883571363, "grad_norm": 1.0251629830106175, "learning_rate": 7.935204757302413e-06, "loss": 0.8086, "step": 1209 }, { "epoch": 0.4577697909770169, "grad_norm": 1.0757191280770386, "learning_rate": 7.934888386266387e-06, "loss": 0.8562, "step": 1210 }, { "epoch": 0.45814811311832027, "grad_norm": 1.0698429731328996, "learning_rate": 7.934571251086853e-06, "loss": 0.8518, "step": 1211 }, { "epoch": 0.45852643525962355, "grad_norm": 1.074189860162607, "learning_rate": 7.934253351825402e-06, "loss": 0.7941, "step": 1212 }, { "epoch": 0.4589047574009269, "grad_norm": 1.0538357299975836, "learning_rate": 7.933934688543764e-06, "loss": 0.8394, "step": 1213 }, { "epoch": 0.4592830795422302, "grad_norm": 1.0421117329655678, "learning_rate": 7.933615261303826e-06, "loss": 0.7609, "step": 1214 }, { "epoch": 0.45966140168353353, "grad_norm": 1.0391554404129049, "learning_rate": 7.933295070167617e-06, "loss": 0.8257, "step": 1215 }, { "epoch": 0.4600397238248368, "grad_norm": 1.0446148939643307, "learning_rate": 7.93297411519732e-06, "loss": 0.8104, "step": 1216 }, { "epoch": 0.46041804596614017, "grad_norm": 1.0344384305012022, "learning_rate": 7.932652396455262e-06, "loss": 0.8044, "step": 1217 }, { "epoch": 0.4607963681074435, "grad_norm": 1.0733053009164926, "learning_rate": 7.932329914003919e-06, "loss": 0.8174, "step": 1218 }, { "epoch": 0.4611746902487468, "grad_norm": 1.0714389655461505, "learning_rate": 7.932006667905917e-06, "loss": 0.8255, "step": 1219 }, { "epoch": 0.46155301239005014, "grad_norm": 1.028255926596019, "learning_rate": 7.93168265822403e-06, "loss": 0.8132, "step": 1220 }, { "epoch": 0.46193133453135343, "grad_norm": 1.0523184669233379, "learning_rate": 7.93135788502118e-06, "loss": 0.8428, "step": 1221 }, { "epoch": 0.4623096566726568, "grad_norm": 1.0557227987751663, "learning_rate": 7.931032348360435e-06, "loss": 0.8332, "step": 1222 }, { "epoch": 0.46268797881396007, "grad_norm": 1.0609398608821474, "learning_rate": 7.930706048305015e-06, "loss": 0.8254, "step": 1223 }, { "epoch": 0.4630663009552634, "grad_norm": 1.0113270947271225, "learning_rate": 7.930378984918286e-06, "loss": 0.8335, "step": 1224 }, { "epoch": 0.4634446230965667, "grad_norm": 1.0131305243085915, "learning_rate": 7.93005115826376e-06, "loss": 0.7971, "step": 1225 }, { "epoch": 0.46382294523787004, "grad_norm": 1.0569179946125011, "learning_rate": 7.929722568405108e-06, "loss": 0.8166, "step": 1226 }, { "epoch": 0.4642012673791734, "grad_norm": 1.042578338856108, "learning_rate": 7.929393215406131e-06, "loss": 0.8204, "step": 1227 }, { "epoch": 0.4645795895204767, "grad_norm": 1.0748606201799873, "learning_rate": 7.929063099330795e-06, "loss": 0.8152, "step": 1228 }, { "epoch": 0.46495791166178, "grad_norm": 1.0587959397105573, "learning_rate": 7.928732220243206e-06, "loss": 0.8452, "step": 1229 }, { "epoch": 0.4653362338030833, "grad_norm": 1.0914151462165957, "learning_rate": 7.928400578207617e-06, "loss": 0.8131, "step": 1230 }, { "epoch": 0.46571455594438665, "grad_norm": 1.0396349529813116, "learning_rate": 7.928068173288438e-06, "loss": 0.8113, "step": 1231 }, { "epoch": 0.46609287808568994, "grad_norm": 1.0607390438435043, "learning_rate": 7.927735005550215e-06, "loss": 0.8368, "step": 1232 }, { "epoch": 0.4664712002269933, "grad_norm": 1.0290648955783543, "learning_rate": 7.927401075057652e-06, "loss": 0.808, "step": 1233 }, { "epoch": 0.46684952236829663, "grad_norm": 1.0438273949617254, "learning_rate": 7.927066381875595e-06, "loss": 0.8109, "step": 1234 }, { "epoch": 0.4672278445095999, "grad_norm": 1.0492773898494756, "learning_rate": 7.926730926069041e-06, "loss": 0.8263, "step": 1235 }, { "epoch": 0.46760616665090327, "grad_norm": 1.0898615275461312, "learning_rate": 7.926394707703133e-06, "loss": 0.8417, "step": 1236 }, { "epoch": 0.46798448879220655, "grad_norm": 1.0371312864392424, "learning_rate": 7.926057726843167e-06, "loss": 0.7853, "step": 1237 }, { "epoch": 0.4683628109335099, "grad_norm": 1.0311331135840094, "learning_rate": 7.925719983554582e-06, "loss": 0.8433, "step": 1238 }, { "epoch": 0.4687411330748132, "grad_norm": 1.0104501833340858, "learning_rate": 7.925381477902967e-06, "loss": 0.8246, "step": 1239 }, { "epoch": 0.46911945521611653, "grad_norm": 1.033351900846643, "learning_rate": 7.92504220995406e-06, "loss": 0.801, "step": 1240 }, { "epoch": 0.4694977773574198, "grad_norm": 1.0678576004897766, "learning_rate": 7.92470217977374e-06, "loss": 0.7953, "step": 1241 }, { "epoch": 0.46987609949872317, "grad_norm": 1.049154054889686, "learning_rate": 7.924361387428047e-06, "loss": 0.8034, "step": 1242 }, { "epoch": 0.4702544216400265, "grad_norm": 1.0501910151623293, "learning_rate": 7.924019832983159e-06, "loss": 0.8421, "step": 1243 }, { "epoch": 0.4706327437813298, "grad_norm": 1.0265699705882914, "learning_rate": 7.923677516505404e-06, "loss": 0.7909, "step": 1244 }, { "epoch": 0.47101106592263314, "grad_norm": 1.0395280931797561, "learning_rate": 7.92333443806126e-06, "loss": 0.8283, "step": 1245 }, { "epoch": 0.47138938806393643, "grad_norm": 1.006365421675378, "learning_rate": 7.922990597717352e-06, "loss": 0.8065, "step": 1246 }, { "epoch": 0.4717677102052398, "grad_norm": 1.0276097967827926, "learning_rate": 7.922645995540453e-06, "loss": 0.808, "step": 1247 }, { "epoch": 0.47214603234654307, "grad_norm": 0.990132630477362, "learning_rate": 7.922300631597482e-06, "loss": 0.8006, "step": 1248 }, { "epoch": 0.4725243544878464, "grad_norm": 1.047163368722463, "learning_rate": 7.921954505955508e-06, "loss": 0.7698, "step": 1249 }, { "epoch": 0.4729026766291497, "grad_norm": 1.0735335320173403, "learning_rate": 7.921607618681748e-06, "loss": 0.807, "step": 1250 }, { "epoch": 0.47328099877045304, "grad_norm": 1.0461927309518722, "learning_rate": 7.921259969843568e-06, "loss": 0.8158, "step": 1251 }, { "epoch": 0.4736593209117564, "grad_norm": 1.0478396570827158, "learning_rate": 7.920911559508476e-06, "loss": 0.8386, "step": 1252 }, { "epoch": 0.4740376430530597, "grad_norm": 1.0449949458790635, "learning_rate": 7.920562387744139e-06, "loss": 0.769, "step": 1253 }, { "epoch": 0.474415965194363, "grad_norm": 1.0333564168358704, "learning_rate": 7.92021245461836e-06, "loss": 0.7821, "step": 1254 }, { "epoch": 0.4747942873356663, "grad_norm": 1.0160573616445434, "learning_rate": 7.919861760199095e-06, "loss": 0.8134, "step": 1255 }, { "epoch": 0.47517260947696965, "grad_norm": 1.113593494987971, "learning_rate": 7.91951030455445e-06, "loss": 0.8009, "step": 1256 }, { "epoch": 0.47555093161827294, "grad_norm": 1.0583016464392816, "learning_rate": 7.919158087752675e-06, "loss": 0.8338, "step": 1257 }, { "epoch": 0.4759292537595763, "grad_norm": 1.0274177510689335, "learning_rate": 7.918805109862172e-06, "loss": 0.7701, "step": 1258 }, { "epoch": 0.4763075759008796, "grad_norm": 0.9716066799511451, "learning_rate": 7.918451370951486e-06, "loss": 0.7624, "step": 1259 }, { "epoch": 0.4766858980421829, "grad_norm": 1.0417278811736634, "learning_rate": 7.91809687108931e-06, "loss": 0.8515, "step": 1260 }, { "epoch": 0.47706422018348627, "grad_norm": 1.0815755118948713, "learning_rate": 7.917741610344492e-06, "loss": 0.826, "step": 1261 }, { "epoch": 0.47744254232478955, "grad_norm": 0.994132013241377, "learning_rate": 7.917385588786019e-06, "loss": 0.8112, "step": 1262 }, { "epoch": 0.4778208644660929, "grad_norm": 1.0835320028786077, "learning_rate": 7.91702880648303e-06, "loss": 0.8283, "step": 1263 }, { "epoch": 0.4781991866073962, "grad_norm": 1.0656905256693705, "learning_rate": 7.916671263504812e-06, "loss": 0.8112, "step": 1264 }, { "epoch": 0.47857750874869953, "grad_norm": 1.0642356494274112, "learning_rate": 7.916312959920796e-06, "loss": 0.8187, "step": 1265 }, { "epoch": 0.4789558308900028, "grad_norm": 1.1132626507153238, "learning_rate": 7.915953895800568e-06, "loss": 0.8333, "step": 1266 }, { "epoch": 0.47933415303130616, "grad_norm": 1.0964935829984281, "learning_rate": 7.915594071213852e-06, "loss": 0.8555, "step": 1267 }, { "epoch": 0.47971247517260945, "grad_norm": 1.0333616049038883, "learning_rate": 7.915233486230529e-06, "loss": 0.8002, "step": 1268 }, { "epoch": 0.4800907973139128, "grad_norm": 1.0938509373019147, "learning_rate": 7.914872140920622e-06, "loss": 0.8222, "step": 1269 }, { "epoch": 0.48046911945521614, "grad_norm": 1.0500659271586612, "learning_rate": 7.914510035354302e-06, "loss": 0.7984, "step": 1270 }, { "epoch": 0.48084744159651943, "grad_norm": 1.0412102283401292, "learning_rate": 7.914147169601891e-06, "loss": 0.8178, "step": 1271 }, { "epoch": 0.4812257637378228, "grad_norm": 0.9740307673809164, "learning_rate": 7.913783543733856e-06, "loss": 0.7733, "step": 1272 }, { "epoch": 0.48160408587912606, "grad_norm": 1.069013806380367, "learning_rate": 7.91341915782081e-06, "loss": 0.8355, "step": 1273 }, { "epoch": 0.4819824080204294, "grad_norm": 1.020794082270209, "learning_rate": 7.913054011933518e-06, "loss": 0.8066, "step": 1274 }, { "epoch": 0.4823607301617327, "grad_norm": 1.0710477291242142, "learning_rate": 7.91268810614289e-06, "loss": 0.822, "step": 1275 }, { "epoch": 0.48273905230303604, "grad_norm": 1.021706668635038, "learning_rate": 7.912321440519982e-06, "loss": 0.8393, "step": 1276 }, { "epoch": 0.48311737444433933, "grad_norm": 1.0381317605620335, "learning_rate": 7.911954015136e-06, "loss": 0.8001, "step": 1277 }, { "epoch": 0.4834956965856427, "grad_norm": 1.0491889355455017, "learning_rate": 7.9115858300623e-06, "loss": 0.8424, "step": 1278 }, { "epoch": 0.483874018726946, "grad_norm": 1.027527176211447, "learning_rate": 7.911216885370377e-06, "loss": 0.7934, "step": 1279 }, { "epoch": 0.4842523408682493, "grad_norm": 1.0241159829134092, "learning_rate": 7.910847181131883e-06, "loss": 0.8632, "step": 1280 }, { "epoch": 0.48463066300955265, "grad_norm": 1.050840821158761, "learning_rate": 7.910476717418613e-06, "loss": 0.8341, "step": 1281 }, { "epoch": 0.48500898515085594, "grad_norm": 1.0312020050809032, "learning_rate": 7.910105494302508e-06, "loss": 0.8124, "step": 1282 }, { "epoch": 0.4853873072921593, "grad_norm": 1.058895959078315, "learning_rate": 7.90973351185566e-06, "loss": 0.8179, "step": 1283 }, { "epoch": 0.4857656294334626, "grad_norm": 1.0442278097312725, "learning_rate": 7.909360770150308e-06, "loss": 0.8251, "step": 1284 }, { "epoch": 0.4861439515747659, "grad_norm": 1.0685857966408454, "learning_rate": 7.908987269258834e-06, "loss": 0.8506, "step": 1285 }, { "epoch": 0.4865222737160692, "grad_norm": 1.1080322429830538, "learning_rate": 7.908613009253774e-06, "loss": 0.825, "step": 1286 }, { "epoch": 0.48690059585737255, "grad_norm": 1.0340810208381146, "learning_rate": 7.908237990207805e-06, "loss": 0.7916, "step": 1287 }, { "epoch": 0.4872789179986759, "grad_norm": 1.0420175323828418, "learning_rate": 7.907862212193758e-06, "loss": 0.822, "step": 1288 }, { "epoch": 0.4876572401399792, "grad_norm": 1.0199603577395158, "learning_rate": 7.907485675284604e-06, "loss": 0.8082, "step": 1289 }, { "epoch": 0.48803556228128253, "grad_norm": 1.0282638290755661, "learning_rate": 7.907108379553467e-06, "loss": 0.8308, "step": 1290 }, { "epoch": 0.4884138844225858, "grad_norm": 1.0699234725043125, "learning_rate": 7.90673032507362e-06, "loss": 0.809, "step": 1291 }, { "epoch": 0.48879220656388916, "grad_norm": 1.0537759557907738, "learning_rate": 7.906351511918477e-06, "loss": 0.8244, "step": 1292 }, { "epoch": 0.48917052870519245, "grad_norm": 1.0220073412783424, "learning_rate": 7.905971940161603e-06, "loss": 0.8313, "step": 1293 }, { "epoch": 0.4895488508464958, "grad_norm": 1.0751723455689177, "learning_rate": 7.905591609876708e-06, "loss": 0.8373, "step": 1294 }, { "epoch": 0.4899271729877991, "grad_norm": 1.0162597179792359, "learning_rate": 7.905210521137654e-06, "loss": 0.8142, "step": 1295 }, { "epoch": 0.49030549512910243, "grad_norm": 1.0733965520897772, "learning_rate": 7.904828674018446e-06, "loss": 0.8325, "step": 1296 }, { "epoch": 0.4906838172704058, "grad_norm": 1.0275444217813758, "learning_rate": 7.904446068593236e-06, "loss": 0.812, "step": 1297 }, { "epoch": 0.49106213941170906, "grad_norm": 1.0074767810899912, "learning_rate": 7.904062704936325e-06, "loss": 0.8072, "step": 1298 }, { "epoch": 0.4914404615530124, "grad_norm": 1.0390065488319102, "learning_rate": 7.903678583122165e-06, "loss": 0.8008, "step": 1299 }, { "epoch": 0.4918187836943157, "grad_norm": 0.9868065507715447, "learning_rate": 7.903293703225345e-06, "loss": 0.816, "step": 1300 }, { "epoch": 0.49219710583561904, "grad_norm": 1.0553901493428994, "learning_rate": 7.902908065320615e-06, "loss": 0.835, "step": 1301 }, { "epoch": 0.49257542797692233, "grad_norm": 1.0153758567731757, "learning_rate": 7.902521669482858e-06, "loss": 0.7622, "step": 1302 }, { "epoch": 0.4929537501182257, "grad_norm": 1.039524643535567, "learning_rate": 7.902134515787115e-06, "loss": 0.8219, "step": 1303 }, { "epoch": 0.49333207225952896, "grad_norm": 1.0193352620631986, "learning_rate": 7.901746604308567e-06, "loss": 0.7745, "step": 1304 }, { "epoch": 0.4937103944008323, "grad_norm": 1.0237247993056149, "learning_rate": 7.901357935122549e-06, "loss": 0.7918, "step": 1305 }, { "epoch": 0.49408871654213565, "grad_norm": 1.018379832975063, "learning_rate": 7.900968508304535e-06, "loss": 0.8111, "step": 1306 }, { "epoch": 0.49446703868343894, "grad_norm": 1.116472085720671, "learning_rate": 7.900578323930154e-06, "loss": 0.7942, "step": 1307 }, { "epoch": 0.4948453608247423, "grad_norm": 1.0587349903275387, "learning_rate": 7.900187382075179e-06, "loss": 0.7992, "step": 1308 }, { "epoch": 0.4952236829660456, "grad_norm": 1.0058048161089288, "learning_rate": 7.899795682815525e-06, "loss": 0.7812, "step": 1309 }, { "epoch": 0.4956020051073489, "grad_norm": 1.0466221891639538, "learning_rate": 7.899403226227265e-06, "loss": 0.8172, "step": 1310 }, { "epoch": 0.4959803272486522, "grad_norm": 1.021072365800396, "learning_rate": 7.899010012386609e-06, "loss": 0.7917, "step": 1311 }, { "epoch": 0.49635864938995555, "grad_norm": 1.0276680529834, "learning_rate": 7.898616041369919e-06, "loss": 0.806, "step": 1312 }, { "epoch": 0.49673697153125884, "grad_norm": 1.0080935461504426, "learning_rate": 7.898221313253703e-06, "loss": 0.7839, "step": 1313 }, { "epoch": 0.4971152936725622, "grad_norm": 1.045973831410194, "learning_rate": 7.897825828114615e-06, "loss": 0.8396, "step": 1314 }, { "epoch": 0.49749361581386553, "grad_norm": 1.0314643332651545, "learning_rate": 7.897429586029458e-06, "loss": 0.845, "step": 1315 }, { "epoch": 0.4978719379551688, "grad_norm": 1.0214806015923183, "learning_rate": 7.897032587075181e-06, "loss": 0.8178, "step": 1316 }, { "epoch": 0.49825026009647216, "grad_norm": 1.0739578792818636, "learning_rate": 7.896634831328881e-06, "loss": 0.803, "step": 1317 }, { "epoch": 0.49862858223777545, "grad_norm": 1.1075886688146952, "learning_rate": 7.8962363188678e-06, "loss": 0.7869, "step": 1318 }, { "epoch": 0.4990069043790788, "grad_norm": 1.0212558702854573, "learning_rate": 7.895837049769326e-06, "loss": 0.8181, "step": 1319 }, { "epoch": 0.4993852265203821, "grad_norm": 1.0781905029615857, "learning_rate": 7.895437024111e-06, "loss": 0.8469, "step": 1320 }, { "epoch": 0.49976354866168543, "grad_norm": 1.0970231389243905, "learning_rate": 7.895036241970501e-06, "loss": 0.8268, "step": 1321 }, { "epoch": 0.5001418708029888, "grad_norm": 0.9979190002347814, "learning_rate": 7.894634703425664e-06, "loss": 0.82, "step": 1322 }, { "epoch": 0.5005201929442921, "grad_norm": 1.011211832148979, "learning_rate": 7.894232408554466e-06, "loss": 0.7793, "step": 1323 }, { "epoch": 0.5008985150855954, "grad_norm": 1.058479892971991, "learning_rate": 7.893829357435027e-06, "loss": 0.8557, "step": 1324 }, { "epoch": 0.5012768372268988, "grad_norm": 1.067675718676119, "learning_rate": 7.893425550145624e-06, "loss": 0.8075, "step": 1325 }, { "epoch": 0.501655159368202, "grad_norm": 1.0748158502027498, "learning_rate": 7.893020986764671e-06, "loss": 0.8217, "step": 1326 }, { "epoch": 0.5020334815095053, "grad_norm": 1.0371866926324267, "learning_rate": 7.892615667370736e-06, "loss": 0.786, "step": 1327 }, { "epoch": 0.5024118036508086, "grad_norm": 1.0227845872267822, "learning_rate": 7.892209592042528e-06, "loss": 0.851, "step": 1328 }, { "epoch": 0.502790125792112, "grad_norm": 1.053385595871815, "learning_rate": 7.891802760858909e-06, "loss": 0.8131, "step": 1329 }, { "epoch": 0.5031684479334153, "grad_norm": 1.0858668827753901, "learning_rate": 7.89139517389888e-06, "loss": 0.8178, "step": 1330 }, { "epoch": 0.5031684479334153, "eval_loss": 0.8155249357223511, "eval_runtime": 26.9154, "eval_samples_per_second": 32.881, "eval_steps_per_second": 1.04, "step": 1330 }, { "epoch": 0.5031684479334153, "eval_bench_accuracy_arc_challenge": 0.22857142857142856, "eval_bench_accuracy_hellaswag": 0.255, "eval_bench_accuracy_mmlu": 0.2782608695652174, "eval_bench_average_accuracy": 0.253944099378882, "eval_bench_loss": 5.252888461999726, "eval_bench_total_accuracy": 0.25274725274725274, "step": 1330 }, { "epoch": 0.5035467700747186, "grad_norm": 1.0418553186067219, "learning_rate": 7.890986831241598e-06, "loss": 0.7842, "step": 1331 }, { "epoch": 0.503925092216022, "grad_norm": 1.027783298562076, "learning_rate": 7.890577732966358e-06, "loss": 0.7925, "step": 1332 }, { "epoch": 0.5043034143573253, "grad_norm": 1.0399175596382164, "learning_rate": 7.890167879152609e-06, "loss": 0.8595, "step": 1333 }, { "epoch": 0.5046817364986286, "grad_norm": 1.0324556300456535, "learning_rate": 7.88975726987994e-06, "loss": 0.8402, "step": 1334 }, { "epoch": 0.5050600586399319, "grad_norm": 1.0669911175427689, "learning_rate": 7.889345905228092e-06, "loss": 0.8132, "step": 1335 }, { "epoch": 0.5054383807812353, "grad_norm": 1.07761249948945, "learning_rate": 7.888933785276951e-06, "loss": 0.8122, "step": 1336 }, { "epoch": 0.5058167029225386, "grad_norm": 1.0315582279231172, "learning_rate": 7.888520910106548e-06, "loss": 0.8063, "step": 1337 }, { "epoch": 0.5061950250638418, "grad_norm": 1.028383480686869, "learning_rate": 7.888107279797064e-06, "loss": 0.8115, "step": 1338 }, { "epoch": 0.5065733472051451, "grad_norm": 1.1084019164549017, "learning_rate": 7.887692894428822e-06, "loss": 0.8586, "step": 1339 }, { "epoch": 0.5069516693464485, "grad_norm": 1.0246273881178, "learning_rate": 7.887277754082298e-06, "loss": 0.7968, "step": 1340 }, { "epoch": 0.5073299914877518, "grad_norm": 1.0537510788483588, "learning_rate": 7.886861858838109e-06, "loss": 0.7794, "step": 1341 }, { "epoch": 0.5077083136290551, "grad_norm": 1.025698434441957, "learning_rate": 7.88644520877702e-06, "loss": 0.7983, "step": 1342 }, { "epoch": 0.5080866357703585, "grad_norm": 1.0480085776508747, "learning_rate": 7.886027803979946e-06, "loss": 0.8016, "step": 1343 }, { "epoch": 0.5084649579116618, "grad_norm": 1.0461816558010573, "learning_rate": 7.885609644527943e-06, "loss": 0.8189, "step": 1344 }, { "epoch": 0.5088432800529651, "grad_norm": 0.993326821555258, "learning_rate": 7.885190730502215e-06, "loss": 0.7957, "step": 1345 }, { "epoch": 0.5092216021942684, "grad_norm": 1.0745480385635238, "learning_rate": 7.884771061984118e-06, "loss": 0.8019, "step": 1346 }, { "epoch": 0.5095999243355718, "grad_norm": 1.0384805298302937, "learning_rate": 7.884350639055147e-06, "loss": 0.8395, "step": 1347 }, { "epoch": 0.5099782464768751, "grad_norm": 1.020760024227472, "learning_rate": 7.883929461796949e-06, "loss": 0.7919, "step": 1348 }, { "epoch": 0.5103565686181784, "grad_norm": 1.0426222802625165, "learning_rate": 7.883507530291315e-06, "loss": 0.8133, "step": 1349 }, { "epoch": 0.5107348907594818, "grad_norm": 1.0236106718012763, "learning_rate": 7.883084844620181e-06, "loss": 0.7525, "step": 1350 }, { "epoch": 0.511113212900785, "grad_norm": 1.0752909757757687, "learning_rate": 7.882661404865635e-06, "loss": 0.8363, "step": 1351 }, { "epoch": 0.5114915350420883, "grad_norm": 1.0496011841679878, "learning_rate": 7.882237211109903e-06, "loss": 0.825, "step": 1352 }, { "epoch": 0.5118698571833916, "grad_norm": 1.052905405929199, "learning_rate": 7.881812263435365e-06, "loss": 0.7808, "step": 1353 }, { "epoch": 0.512248179324695, "grad_norm": 1.0383149467870931, "learning_rate": 7.881386561924544e-06, "loss": 0.8258, "step": 1354 }, { "epoch": 0.5126265014659983, "grad_norm": 1.0142846574710827, "learning_rate": 7.880960106660112e-06, "loss": 0.832, "step": 1355 }, { "epoch": 0.5130048236073016, "grad_norm": 1.0162105056610324, "learning_rate": 7.880532897724882e-06, "loss": 0.8271, "step": 1356 }, { "epoch": 0.5133831457486049, "grad_norm": 1.0111397828819904, "learning_rate": 7.880104935201817e-06, "loss": 0.7716, "step": 1357 }, { "epoch": 0.5137614678899083, "grad_norm": 1.0387312593547113, "learning_rate": 7.879676219174028e-06, "loss": 0.7856, "step": 1358 }, { "epoch": 0.5141397900312116, "grad_norm": 1.0976300200992746, "learning_rate": 7.879246749724769e-06, "loss": 0.8214, "step": 1359 }, { "epoch": 0.5145181121725149, "grad_norm": 1.0225148649560976, "learning_rate": 7.878816526937443e-06, "loss": 0.8154, "step": 1360 }, { "epoch": 0.5148964343138183, "grad_norm": 1.0564511900500775, "learning_rate": 7.878385550895597e-06, "loss": 0.7706, "step": 1361 }, { "epoch": 0.5152747564551216, "grad_norm": 1.065194818654382, "learning_rate": 7.877953821682924e-06, "loss": 0.7806, "step": 1362 }, { "epoch": 0.5156530785964248, "grad_norm": 1.0318627975030588, "learning_rate": 7.877521339383267e-06, "loss": 0.8317, "step": 1363 }, { "epoch": 0.5160314007377281, "grad_norm": 1.0660496042471788, "learning_rate": 7.877088104080612e-06, "loss": 0.8116, "step": 1364 }, { "epoch": 0.5164097228790315, "grad_norm": 1.0084811396262128, "learning_rate": 7.87665411585909e-06, "loss": 0.8233, "step": 1365 }, { "epoch": 0.5167880450203348, "grad_norm": 1.0061856631615549, "learning_rate": 7.876219374802983e-06, "loss": 0.8226, "step": 1366 }, { "epoch": 0.5171663671616381, "grad_norm": 0.9962092519447693, "learning_rate": 7.875783880996717e-06, "loss": 0.7949, "step": 1367 }, { "epoch": 0.5175446893029415, "grad_norm": 1.0320181154699064, "learning_rate": 7.87534763452486e-06, "loss": 0.8078, "step": 1368 }, { "epoch": 0.5179230114442448, "grad_norm": 1.0366220904643662, "learning_rate": 7.87491063547213e-06, "loss": 0.7915, "step": 1369 }, { "epoch": 0.5183013335855481, "grad_norm": 0.9990483570523689, "learning_rate": 7.874472883923396e-06, "loss": 0.7962, "step": 1370 }, { "epoch": 0.5186796557268514, "grad_norm": 1.072712099895109, "learning_rate": 7.874034379963663e-06, "loss": 0.8201, "step": 1371 }, { "epoch": 0.5190579778681548, "grad_norm": 1.0469398611990606, "learning_rate": 7.873595123678088e-06, "loss": 0.8295, "step": 1372 }, { "epoch": 0.5194363000094581, "grad_norm": 1.0258466230718022, "learning_rate": 7.873155115151976e-06, "loss": 0.7962, "step": 1373 }, { "epoch": 0.5198146221507614, "grad_norm": 1.0150744464405486, "learning_rate": 7.872714354470771e-06, "loss": 0.8091, "step": 1374 }, { "epoch": 0.5201929442920646, "grad_norm": 1.0877815460579687, "learning_rate": 7.87227284172007e-06, "loss": 0.8449, "step": 1375 }, { "epoch": 0.520571266433368, "grad_norm": 0.9989012315656198, "learning_rate": 7.871830576985613e-06, "loss": 0.7904, "step": 1376 }, { "epoch": 0.5209495885746713, "grad_norm": 1.0281663493359343, "learning_rate": 7.871387560353288e-06, "loss": 0.8235, "step": 1377 }, { "epoch": 0.5213279107159746, "grad_norm": 1.013255314723829, "learning_rate": 7.870943791909124e-06, "loss": 0.8137, "step": 1378 }, { "epoch": 0.521706232857278, "grad_norm": 1.0404202767535178, "learning_rate": 7.870499271739304e-06, "loss": 0.8331, "step": 1379 }, { "epoch": 0.5220845549985813, "grad_norm": 1.0008843854289766, "learning_rate": 7.870053999930149e-06, "loss": 0.7985, "step": 1380 }, { "epoch": 0.5224628771398846, "grad_norm": 1.115907702208107, "learning_rate": 7.869607976568131e-06, "loss": 0.8444, "step": 1381 }, { "epoch": 0.5228411992811879, "grad_norm": 1.0499698053880258, "learning_rate": 7.869161201739866e-06, "loss": 0.7875, "step": 1382 }, { "epoch": 0.5232195214224913, "grad_norm": 1.0086891227734494, "learning_rate": 7.868713675532115e-06, "loss": 0.7981, "step": 1383 }, { "epoch": 0.5235978435637946, "grad_norm": 1.0416968121742411, "learning_rate": 7.868265398031788e-06, "loss": 0.8082, "step": 1384 }, { "epoch": 0.5239761657050979, "grad_norm": 0.9956171233693443, "learning_rate": 7.86781636932594e-06, "loss": 0.8497, "step": 1385 }, { "epoch": 0.5243544878464013, "grad_norm": 1.0366372693126888, "learning_rate": 7.867366589501767e-06, "loss": 0.7878, "step": 1386 }, { "epoch": 0.5247328099877046, "grad_norm": 1.0252929211171813, "learning_rate": 7.86691605864662e-06, "loss": 0.8254, "step": 1387 }, { "epoch": 0.5251111321290078, "grad_norm": 1.0349722097719734, "learning_rate": 7.866464776847987e-06, "loss": 0.8092, "step": 1388 }, { "epoch": 0.5254894542703111, "grad_norm": 1.0775801625166288, "learning_rate": 7.866012744193508e-06, "loss": 0.8032, "step": 1389 }, { "epoch": 0.5258677764116145, "grad_norm": 1.025158242287074, "learning_rate": 7.865559960770964e-06, "loss": 0.7777, "step": 1390 }, { "epoch": 0.5262460985529178, "grad_norm": 1.0261907345479138, "learning_rate": 7.865106426668287e-06, "loss": 0.7656, "step": 1391 }, { "epoch": 0.5266244206942211, "grad_norm": 1.0119949142526334, "learning_rate": 7.864652141973549e-06, "loss": 0.817, "step": 1392 }, { "epoch": 0.5270027428355244, "grad_norm": 0.9887922738590984, "learning_rate": 7.864197106774973e-06, "loss": 0.7871, "step": 1393 }, { "epoch": 0.5273810649768278, "grad_norm": 1.0473369889166892, "learning_rate": 7.863741321160924e-06, "loss": 0.7885, "step": 1394 }, { "epoch": 0.5277593871181311, "grad_norm": 1.021975230127612, "learning_rate": 7.863284785219916e-06, "loss": 0.7862, "step": 1395 }, { "epoch": 0.5281377092594344, "grad_norm": 1.0624890686836679, "learning_rate": 7.862827499040604e-06, "loss": 0.8445, "step": 1396 }, { "epoch": 0.5285160314007378, "grad_norm": 1.0159701351719927, "learning_rate": 7.862369462711795e-06, "loss": 0.8084, "step": 1397 }, { "epoch": 0.5288943535420411, "grad_norm": 1.0307854419947649, "learning_rate": 7.861910676322434e-06, "loss": 0.7957, "step": 1398 }, { "epoch": 0.5292726756833444, "grad_norm": 1.088274510577477, "learning_rate": 7.861451139961622e-06, "loss": 0.8134, "step": 1399 }, { "epoch": 0.5296509978246476, "grad_norm": 1.1610468987478788, "learning_rate": 7.860990853718593e-06, "loss": 0.7706, "step": 1400 }, { "epoch": 0.530029319965951, "grad_norm": 1.0709949089292212, "learning_rate": 7.860529817682737e-06, "loss": 0.839, "step": 1401 }, { "epoch": 0.5304076421072543, "grad_norm": 1.0641189768424455, "learning_rate": 7.860068031943586e-06, "loss": 0.7794, "step": 1402 }, { "epoch": 0.5307859642485576, "grad_norm": 1.0425801957230985, "learning_rate": 7.859605496590816e-06, "loss": 0.7982, "step": 1403 }, { "epoch": 0.531164286389861, "grad_norm": 1.0561738214600724, "learning_rate": 7.859142211714251e-06, "loss": 0.8298, "step": 1404 }, { "epoch": 0.5315426085311643, "grad_norm": 1.0034598628819673, "learning_rate": 7.858678177403859e-06, "loss": 0.842, "step": 1405 }, { "epoch": 0.5319209306724676, "grad_norm": 1.0174154185360578, "learning_rate": 7.858213393749755e-06, "loss": 0.8024, "step": 1406 }, { "epoch": 0.5322992528137709, "grad_norm": 1.002603647328177, "learning_rate": 7.857747860842196e-06, "loss": 0.8186, "step": 1407 }, { "epoch": 0.5326775749550743, "grad_norm": 1.0285530234043798, "learning_rate": 7.857281578771589e-06, "loss": 0.8156, "step": 1408 }, { "epoch": 0.5330558970963776, "grad_norm": 1.02768116084931, "learning_rate": 7.856814547628485e-06, "loss": 0.8165, "step": 1409 }, { "epoch": 0.5334342192376809, "grad_norm": 1.1031829681313992, "learning_rate": 7.85634676750358e-06, "loss": 0.8579, "step": 1410 }, { "epoch": 0.5338125413789842, "grad_norm": 1.027426941839886, "learning_rate": 7.855878238487714e-06, "loss": 0.7945, "step": 1411 }, { "epoch": 0.5341908635202876, "grad_norm": 1.0561714395136612, "learning_rate": 7.855408960671875e-06, "loss": 0.7641, "step": 1412 }, { "epoch": 0.5345691856615908, "grad_norm": 1.090238437190781, "learning_rate": 7.854938934147195e-06, "loss": 0.8063, "step": 1413 }, { "epoch": 0.5349475078028941, "grad_norm": 1.2074317498906901, "learning_rate": 7.854468159004952e-06, "loss": 0.7921, "step": 1414 }, { "epoch": 0.5353258299441975, "grad_norm": 1.0749934432108652, "learning_rate": 7.85399663533657e-06, "loss": 0.8165, "step": 1415 }, { "epoch": 0.5357041520855008, "grad_norm": 1.0472554586470812, "learning_rate": 7.853524363233614e-06, "loss": 0.8232, "step": 1416 }, { "epoch": 0.5360824742268041, "grad_norm": 1.0321608082815132, "learning_rate": 7.853051342787802e-06, "loss": 0.8207, "step": 1417 }, { "epoch": 0.5364607963681074, "grad_norm": 1.010186032847584, "learning_rate": 7.852577574090992e-06, "loss": 0.7875, "step": 1418 }, { "epoch": 0.5368391185094108, "grad_norm": 1.0585550633979846, "learning_rate": 7.852103057235187e-06, "loss": 0.7872, "step": 1419 }, { "epoch": 0.5372174406507141, "grad_norm": 1.0424950696245099, "learning_rate": 7.851627792312539e-06, "loss": 0.7871, "step": 1420 }, { "epoch": 0.5375957627920174, "grad_norm": 1.0123853847303819, "learning_rate": 7.85115177941534e-06, "loss": 0.7915, "step": 1421 }, { "epoch": 0.5379740849333208, "grad_norm": 1.0357173714573609, "learning_rate": 7.850675018636034e-06, "loss": 0.7829, "step": 1422 }, { "epoch": 0.5383524070746241, "grad_norm": 1.4395615442604752, "learning_rate": 7.850197510067203e-06, "loss": 0.8255, "step": 1423 }, { "epoch": 0.5387307292159274, "grad_norm": 1.0121918462650672, "learning_rate": 7.849719253801578e-06, "loss": 0.8553, "step": 1424 }, { "epoch": 0.5391090513572306, "grad_norm": 0.9837030660961567, "learning_rate": 7.849240249932039e-06, "loss": 0.7586, "step": 1425 }, { "epoch": 0.539487373498534, "grad_norm": 1.018520798880126, "learning_rate": 7.848760498551603e-06, "loss": 0.8266, "step": 1426 }, { "epoch": 0.5398656956398373, "grad_norm": 1.0215594842474691, "learning_rate": 7.848279999753438e-06, "loss": 0.8115, "step": 1427 }, { "epoch": 0.5402440177811406, "grad_norm": 1.0166660418304827, "learning_rate": 7.847798753630854e-06, "loss": 0.7822, "step": 1428 }, { "epoch": 0.5406223399224439, "grad_norm": 1.0027140748494623, "learning_rate": 7.84731676027731e-06, "loss": 0.8033, "step": 1429 }, { "epoch": 0.5410006620637473, "grad_norm": 1.0627188785846766, "learning_rate": 7.846834019786404e-06, "loss": 0.8265, "step": 1430 }, { "epoch": 0.5413789842050506, "grad_norm": 1.0264202021796238, "learning_rate": 7.846350532251887e-06, "loss": 0.8109, "step": 1431 }, { "epoch": 0.5417573063463539, "grad_norm": 1.0850130197305035, "learning_rate": 7.845866297767647e-06, "loss": 0.8166, "step": 1432 }, { "epoch": 0.5421356284876573, "grad_norm": 1.0443803197744415, "learning_rate": 7.845381316427724e-06, "loss": 0.8134, "step": 1433 }, { "epoch": 0.5425139506289606, "grad_norm": 1.0216121613789444, "learning_rate": 7.844895588326298e-06, "loss": 0.8248, "step": 1434 }, { "epoch": 0.5428922727702639, "grad_norm": 1.0528680390786613, "learning_rate": 7.844409113557698e-06, "loss": 0.8306, "step": 1435 }, { "epoch": 0.5432705949115672, "grad_norm": 1.056376944389717, "learning_rate": 7.843921892216392e-06, "loss": 0.7733, "step": 1436 }, { "epoch": 0.5436489170528706, "grad_norm": 1.0054617166141346, "learning_rate": 7.843433924397002e-06, "loss": 0.7937, "step": 1437 }, { "epoch": 0.5440272391941738, "grad_norm": 1.0047703505362153, "learning_rate": 7.842945210194286e-06, "loss": 0.7923, "step": 1438 }, { "epoch": 0.5444055613354771, "grad_norm": 1.0096110719940172, "learning_rate": 7.842455749703151e-06, "loss": 0.7994, "step": 1439 }, { "epoch": 0.5447838834767805, "grad_norm": 1.0605981769829262, "learning_rate": 7.841965543018651e-06, "loss": 0.8085, "step": 1440 }, { "epoch": 0.5451622056180838, "grad_norm": 1.0471718815415907, "learning_rate": 7.841474590235981e-06, "loss": 0.8463, "step": 1441 }, { "epoch": 0.5455405277593871, "grad_norm": 1.0505867574083267, "learning_rate": 7.840982891450483e-06, "loss": 0.8242, "step": 1442 }, { "epoch": 0.5459188499006904, "grad_norm": 1.0445952963424892, "learning_rate": 7.840490446757645e-06, "loss": 0.7749, "step": 1443 }, { "epoch": 0.5462971720419938, "grad_norm": 1.0068778649332644, "learning_rate": 7.839997256253096e-06, "loss": 0.8116, "step": 1444 }, { "epoch": 0.5466754941832971, "grad_norm": 1.00961692913919, "learning_rate": 7.839503320032612e-06, "loss": 0.7901, "step": 1445 }, { "epoch": 0.5470538163246004, "grad_norm": 0.9780075250092127, "learning_rate": 7.839008638192115e-06, "loss": 0.7885, "step": 1446 }, { "epoch": 0.5474321384659037, "grad_norm": 1.100812581357096, "learning_rate": 7.838513210827671e-06, "loss": 0.8001, "step": 1447 }, { "epoch": 0.5478104606072071, "grad_norm": 1.0494389505966184, "learning_rate": 7.83801703803549e-06, "loss": 0.7977, "step": 1448 }, { "epoch": 0.5481887827485104, "grad_norm": 1.034386181938751, "learning_rate": 7.837520119911927e-06, "loss": 0.8244, "step": 1449 }, { "epoch": 0.5485671048898136, "grad_norm": 1.0112131883045796, "learning_rate": 7.837022456553482e-06, "loss": 0.7537, "step": 1450 }, { "epoch": 0.548945427031117, "grad_norm": 1.0542214842469684, "learning_rate": 7.836524048056801e-06, "loss": 0.8436, "step": 1451 }, { "epoch": 0.5493237491724203, "grad_norm": 1.0139124551358574, "learning_rate": 7.836024894518673e-06, "loss": 0.7765, "step": 1452 }, { "epoch": 0.5497020713137236, "grad_norm": 1.0370438053735662, "learning_rate": 7.835524996036031e-06, "loss": 0.7957, "step": 1453 }, { "epoch": 0.5500803934550269, "grad_norm": 1.0403261101993466, "learning_rate": 7.835024352705953e-06, "loss": 0.8082, "step": 1454 }, { "epoch": 0.5504587155963303, "grad_norm": 1.0223772000926137, "learning_rate": 7.834522964625665e-06, "loss": 0.8091, "step": 1455 }, { "epoch": 0.5508370377376336, "grad_norm": 0.9867288417868126, "learning_rate": 7.834020831892534e-06, "loss": 0.7971, "step": 1456 }, { "epoch": 0.5512153598789369, "grad_norm": 1.038419907192562, "learning_rate": 7.833517954604074e-06, "loss": 0.7774, "step": 1457 }, { "epoch": 0.5515936820202403, "grad_norm": 1.0143771814537008, "learning_rate": 7.833014332857939e-06, "loss": 0.7763, "step": 1458 }, { "epoch": 0.5519720041615436, "grad_norm": 1.0001756819325087, "learning_rate": 7.832509966751933e-06, "loss": 0.7889, "step": 1459 }, { "epoch": 0.5523503263028469, "grad_norm": 1.036257856076326, "learning_rate": 7.832004856384001e-06, "loss": 0.7901, "step": 1460 }, { "epoch": 0.5527286484441502, "grad_norm": 1.0355156315068814, "learning_rate": 7.831499001852236e-06, "loss": 0.7742, "step": 1461 }, { "epoch": 0.5531069705854536, "grad_norm": 1.1407334044483102, "learning_rate": 7.830992403254873e-06, "loss": 0.8265, "step": 1462 }, { "epoch": 0.5534852927267568, "grad_norm": 1.0063557289156941, "learning_rate": 7.83048506069029e-06, "loss": 0.7994, "step": 1463 }, { "epoch": 0.5534852927267568, "eval_loss": 0.8094308972358704, "eval_runtime": 26.9598, "eval_samples_per_second": 32.827, "eval_steps_per_second": 1.039, "step": 1463 }, { "epoch": 0.5534852927267568, "eval_bench_accuracy_arc_challenge": 0.25, "eval_bench_accuracy_hellaswag": 0.215, "eval_bench_accuracy_mmlu": 0.2608695652173913, "eval_bench_average_accuracy": 0.24195652173913043, "eval_bench_loss": 6.063661274157073, "eval_bench_total_accuracy": 0.23736263736263735, "step": 1463 }, { "epoch": 0.5538636148680601, "grad_norm": 1.0744841523132298, "learning_rate": 7.829976974257012e-06, "loss": 0.8504, "step": 1464 }, { "epoch": 0.5542419370093635, "grad_norm": 1.0186917057516884, "learning_rate": 7.829468144053712e-06, "loss": 0.8052, "step": 1465 }, { "epoch": 0.5546202591506668, "grad_norm": 1.0107687681368964, "learning_rate": 7.828958570179196e-06, "loss": 0.8094, "step": 1466 }, { "epoch": 0.5549985812919701, "grad_norm": 1.0349853318053726, "learning_rate": 7.828448252732428e-06, "loss": 0.8303, "step": 1467 }, { "epoch": 0.5553769034332734, "grad_norm": 1.0450694598466956, "learning_rate": 7.827937191812508e-06, "loss": 0.7924, "step": 1468 }, { "epoch": 0.5557552255745768, "grad_norm": 1.0278598268440422, "learning_rate": 7.82742538751868e-06, "loss": 0.7701, "step": 1469 }, { "epoch": 0.5561335477158801, "grad_norm": 1.0315097348678433, "learning_rate": 7.826912839950338e-06, "loss": 0.7643, "step": 1470 }, { "epoch": 0.5565118698571834, "grad_norm": 1.0630245419936848, "learning_rate": 7.826399549207016e-06, "loss": 0.8334, "step": 1471 }, { "epoch": 0.5568901919984867, "grad_norm": 1.057495631028003, "learning_rate": 7.825885515388394e-06, "loss": 0.8098, "step": 1472 }, { "epoch": 0.5572685141397901, "grad_norm": 1.0485936898987425, "learning_rate": 7.825370738594296e-06, "loss": 0.8524, "step": 1473 }, { "epoch": 0.5576468362810933, "grad_norm": 1.089800751911175, "learning_rate": 7.82485521892469e-06, "loss": 0.7807, "step": 1474 }, { "epoch": 0.5580251584223966, "grad_norm": 1.008238694676228, "learning_rate": 7.824338956479687e-06, "loss": 0.7641, "step": 1475 }, { "epoch": 0.5584034805637, "grad_norm": 0.9866356509513795, "learning_rate": 7.823821951359546e-06, "loss": 0.8072, "step": 1476 }, { "epoch": 0.5587818027050033, "grad_norm": 1.0159932518028019, "learning_rate": 7.823304203664665e-06, "loss": 0.7563, "step": 1477 }, { "epoch": 0.5591601248463066, "grad_norm": 1.0691391299613169, "learning_rate": 7.82278571349559e-06, "loss": 0.7666, "step": 1478 }, { "epoch": 0.5595384469876099, "grad_norm": 1.069708560088697, "learning_rate": 7.822266480953014e-06, "loss": 0.8094, "step": 1479 }, { "epoch": 0.5599167691289133, "grad_norm": 1.0399404229309808, "learning_rate": 7.821746506137766e-06, "loss": 0.8041, "step": 1480 }, { "epoch": 0.5602950912702166, "grad_norm": 1.0528966086217326, "learning_rate": 7.821225789150823e-06, "loss": 0.8186, "step": 1481 }, { "epoch": 0.5606734134115199, "grad_norm": 1.078154168587184, "learning_rate": 7.820704330093309e-06, "loss": 0.7697, "step": 1482 }, { "epoch": 0.5610517355528233, "grad_norm": 0.9974199242655317, "learning_rate": 7.82018212906649e-06, "loss": 0.7627, "step": 1483 }, { "epoch": 0.5614300576941266, "grad_norm": 1.0441157570327169, "learning_rate": 7.819659186171774e-06, "loss": 0.7637, "step": 1484 }, { "epoch": 0.5618083798354299, "grad_norm": 1.0350192453023053, "learning_rate": 7.819135501510717e-06, "loss": 0.7863, "step": 1485 }, { "epoch": 0.5621867019767331, "grad_norm": 1.0314197771080482, "learning_rate": 7.818611075185016e-06, "loss": 0.7761, "step": 1486 }, { "epoch": 0.5625650241180365, "grad_norm": 1.1142918188982494, "learning_rate": 7.818085907296514e-06, "loss": 0.8451, "step": 1487 }, { "epoch": 0.5629433462593398, "grad_norm": 1.0635918190610065, "learning_rate": 7.817559997947194e-06, "loss": 0.7987, "step": 1488 }, { "epoch": 0.5633216684006431, "grad_norm": 1.0137296000615337, "learning_rate": 7.817033347239188e-06, "loss": 0.7849, "step": 1489 }, { "epoch": 0.5636999905419464, "grad_norm": 1.0465836630867722, "learning_rate": 7.816505955274772e-06, "loss": 0.7609, "step": 1490 }, { "epoch": 0.5640783126832498, "grad_norm": 1.0227869394316658, "learning_rate": 7.81597782215636e-06, "loss": 0.7658, "step": 1491 }, { "epoch": 0.5644566348245531, "grad_norm": 1.025273340871076, "learning_rate": 7.815448947986518e-06, "loss": 0.7943, "step": 1492 }, { "epoch": 0.5648349569658564, "grad_norm": 1.0788965118297305, "learning_rate": 7.814919332867948e-06, "loss": 0.7825, "step": 1493 }, { "epoch": 0.5652132791071598, "grad_norm": 1.0290788502294095, "learning_rate": 7.814388976903501e-06, "loss": 0.7686, "step": 1494 }, { "epoch": 0.5655916012484631, "grad_norm": 1.0043872677988737, "learning_rate": 7.813857880196172e-06, "loss": 0.765, "step": 1495 }, { "epoch": 0.5659699233897664, "grad_norm": 1.0416556353562665, "learning_rate": 7.813326042849096e-06, "loss": 0.7905, "step": 1496 }, { "epoch": 0.5663482455310697, "grad_norm": 1.0403767458597168, "learning_rate": 7.812793464965557e-06, "loss": 0.8392, "step": 1497 }, { "epoch": 0.5667265676723731, "grad_norm": 1.0804135578705913, "learning_rate": 7.812260146648978e-06, "loss": 0.8042, "step": 1498 }, { "epoch": 0.5671048898136763, "grad_norm": 1.0525290992619953, "learning_rate": 7.811726088002928e-06, "loss": 0.8125, "step": 1499 }, { "epoch": 0.5674832119549796, "grad_norm": 1.0443809449733452, "learning_rate": 7.81119128913112e-06, "loss": 0.8449, "step": 1500 }, { "epoch": 0.567861534096283, "grad_norm": 1.0484442830821317, "learning_rate": 7.810655750137408e-06, "loss": 0.791, "step": 1501 }, { "epoch": 0.5682398562375863, "grad_norm": 1.0322889324418691, "learning_rate": 7.810119471125797e-06, "loss": 0.7638, "step": 1502 }, { "epoch": 0.5686181783788896, "grad_norm": 1.0251619422017846, "learning_rate": 7.809582452200428e-06, "loss": 0.7971, "step": 1503 }, { "epoch": 0.5689965005201929, "grad_norm": 1.0150926516902954, "learning_rate": 7.809044693465587e-06, "loss": 0.7734, "step": 1504 }, { "epoch": 0.5693748226614963, "grad_norm": 1.0663474541629985, "learning_rate": 7.808506195025707e-06, "loss": 0.8411, "step": 1505 }, { "epoch": 0.5697531448027996, "grad_norm": 1.0708265848333849, "learning_rate": 7.807966956985363e-06, "loss": 0.8428, "step": 1506 }, { "epoch": 0.5701314669441029, "grad_norm": 1.0294311898641297, "learning_rate": 7.807426979449273e-06, "loss": 0.8016, "step": 1507 }, { "epoch": 0.5705097890854062, "grad_norm": 1.072155935601359, "learning_rate": 7.806886262522298e-06, "loss": 0.7896, "step": 1508 }, { "epoch": 0.5708881112267096, "grad_norm": 1.0602457428763656, "learning_rate": 7.806344806309445e-06, "loss": 0.8306, "step": 1509 }, { "epoch": 0.5712664333680129, "grad_norm": 1.0410264668234372, "learning_rate": 7.805802610915862e-06, "loss": 0.7708, "step": 1510 }, { "epoch": 0.5716447555093161, "grad_norm": 1.0323609766839155, "learning_rate": 7.805259676446843e-06, "loss": 0.7731, "step": 1511 }, { "epoch": 0.5720230776506195, "grad_norm": 1.0629777585594808, "learning_rate": 7.804716003007825e-06, "loss": 0.8667, "step": 1512 }, { "epoch": 0.5724013997919228, "grad_norm": 0.9991092397744588, "learning_rate": 7.804171590704384e-06, "loss": 0.8158, "step": 1513 }, { "epoch": 0.5727797219332261, "grad_norm": 1.0691406196971251, "learning_rate": 7.803626439642245e-06, "loss": 0.8439, "step": 1514 }, { "epoch": 0.5731580440745294, "grad_norm": 1.003105717691004, "learning_rate": 7.803080549927276e-06, "loss": 0.8294, "step": 1515 }, { "epoch": 0.5735363662158328, "grad_norm": 1.03908547211568, "learning_rate": 7.802533921665487e-06, "loss": 0.7924, "step": 1516 }, { "epoch": 0.5739146883571361, "grad_norm": 1.0879350896154778, "learning_rate": 7.801986554963032e-06, "loss": 0.8214, "step": 1517 }, { "epoch": 0.5742930104984394, "grad_norm": 1.0215923317383557, "learning_rate": 7.801438449926204e-06, "loss": 0.7672, "step": 1518 }, { "epoch": 0.5746713326397428, "grad_norm": 1.0667625852082359, "learning_rate": 7.800889606661448e-06, "loss": 0.779, "step": 1519 }, { "epoch": 0.5750496547810461, "grad_norm": 1.0265205651578218, "learning_rate": 7.800340025275346e-06, "loss": 0.8048, "step": 1520 }, { "epoch": 0.5754279769223494, "grad_norm": 1.07228233508983, "learning_rate": 7.799789705874626e-06, "loss": 0.7798, "step": 1521 }, { "epoch": 0.5758062990636527, "grad_norm": 1.0864037890509946, "learning_rate": 7.799238648566155e-06, "loss": 0.8061, "step": 1522 }, { "epoch": 0.5761846212049561, "grad_norm": 1.024552729289987, "learning_rate": 7.79868685345695e-06, "loss": 0.7923, "step": 1523 }, { "epoch": 0.5765629433462593, "grad_norm": 1.050893206442173, "learning_rate": 7.798134320654169e-06, "loss": 0.7922, "step": 1524 }, { "epoch": 0.5769412654875626, "grad_norm": 1.0361508996059923, "learning_rate": 7.797581050265108e-06, "loss": 0.7934, "step": 1525 }, { "epoch": 0.5773195876288659, "grad_norm": 1.0710969406799804, "learning_rate": 7.797027042397215e-06, "loss": 0.8126, "step": 1526 }, { "epoch": 0.5776979097701693, "grad_norm": 1.0658020905692465, "learning_rate": 7.796472297158071e-06, "loss": 0.825, "step": 1527 }, { "epoch": 0.5780762319114726, "grad_norm": 1.0530236797299208, "learning_rate": 7.79591681465541e-06, "loss": 0.8297, "step": 1528 }, { "epoch": 0.5784545540527759, "grad_norm": 1.0375398854054054, "learning_rate": 7.795360594997107e-06, "loss": 0.8184, "step": 1529 }, { "epoch": 0.5788328761940793, "grad_norm": 1.0223176641231346, "learning_rate": 7.794803638291175e-06, "loss": 0.8081, "step": 1530 }, { "epoch": 0.5792111983353826, "grad_norm": 1.0392507145784662, "learning_rate": 7.794245944645772e-06, "loss": 0.8473, "step": 1531 }, { "epoch": 0.5795895204766859, "grad_norm": 1.022490501012432, "learning_rate": 7.793687514169201e-06, "loss": 0.7883, "step": 1532 }, { "epoch": 0.5799678426179892, "grad_norm": 1.0564202458689138, "learning_rate": 7.793128346969911e-06, "loss": 0.7797, "step": 1533 }, { "epoch": 0.5803461647592926, "grad_norm": 1.0741330485557585, "learning_rate": 7.792568443156489e-06, "loss": 0.808, "step": 1534 }, { "epoch": 0.5807244869005959, "grad_norm": 0.9936986392860392, "learning_rate": 7.792007802837665e-06, "loss": 0.7748, "step": 1535 }, { "epoch": 0.5811028090418991, "grad_norm": 1.04388957808874, "learning_rate": 7.791446426122313e-06, "loss": 0.8282, "step": 1536 }, { "epoch": 0.5814811311832025, "grad_norm": 1.0718346958784504, "learning_rate": 7.790884313119454e-06, "loss": 0.7922, "step": 1537 }, { "epoch": 0.5818594533245058, "grad_norm": 1.0477864953037763, "learning_rate": 7.790321463938246e-06, "loss": 0.8141, "step": 1538 }, { "epoch": 0.5822377754658091, "grad_norm": 1.026774949013717, "learning_rate": 7.789757878687995e-06, "loss": 0.7598, "step": 1539 }, { "epoch": 0.5826160976071124, "grad_norm": 1.015538072369435, "learning_rate": 7.789193557478143e-06, "loss": 0.7877, "step": 1540 }, { "epoch": 0.5829944197484158, "grad_norm": 1.0348274415641654, "learning_rate": 7.788628500418287e-06, "loss": 0.8258, "step": 1541 }, { "epoch": 0.5833727418897191, "grad_norm": 1.02268572106111, "learning_rate": 7.788062707618151e-06, "loss": 0.8323, "step": 1542 }, { "epoch": 0.5837510640310224, "grad_norm": 1.0046192564851208, "learning_rate": 7.787496179187618e-06, "loss": 0.7522, "step": 1543 }, { "epoch": 0.5841293861723257, "grad_norm": 1.0526322563558683, "learning_rate": 7.7869289152367e-06, "loss": 0.8168, "step": 1544 }, { "epoch": 0.5845077083136291, "grad_norm": 0.9819648563646498, "learning_rate": 7.78636091587556e-06, "loss": 0.7441, "step": 1545 }, { "epoch": 0.5848860304549324, "grad_norm": 1.0131957579824842, "learning_rate": 7.785792181214504e-06, "loss": 0.7716, "step": 1546 }, { "epoch": 0.5852643525962357, "grad_norm": 1.0442706083972597, "learning_rate": 7.785222711363975e-06, "loss": 0.783, "step": 1547 }, { "epoch": 0.5856426747375391, "grad_norm": 1.024417321524946, "learning_rate": 7.784652506434564e-06, "loss": 0.808, "step": 1548 }, { "epoch": 0.5860209968788423, "grad_norm": 1.0597851794054838, "learning_rate": 7.784081566537004e-06, "loss": 0.8209, "step": 1549 }, { "epoch": 0.5863993190201456, "grad_norm": 1.0122874466478462, "learning_rate": 7.783509891782168e-06, "loss": 0.7717, "step": 1550 }, { "epoch": 0.5867776411614489, "grad_norm": 1.0075483569470989, "learning_rate": 7.782937482281076e-06, "loss": 0.7653, "step": 1551 }, { "epoch": 0.5871559633027523, "grad_norm": 1.021446573700645, "learning_rate": 7.782364338144885e-06, "loss": 0.7696, "step": 1552 }, { "epoch": 0.5875342854440556, "grad_norm": 1.0432444836660548, "learning_rate": 7.781790459484901e-06, "loss": 0.7933, "step": 1553 }, { "epoch": 0.5879126075853589, "grad_norm": 1.0051174216679133, "learning_rate": 7.781215846412565e-06, "loss": 0.7867, "step": 1554 }, { "epoch": 0.5882909297266623, "grad_norm": 1.0867512164890576, "learning_rate": 7.78064049903947e-06, "loss": 0.7725, "step": 1555 }, { "epoch": 0.5886692518679656, "grad_norm": 1.04980942321374, "learning_rate": 7.780064417477346e-06, "loss": 0.8114, "step": 1556 }, { "epoch": 0.5890475740092689, "grad_norm": 1.0617568995349125, "learning_rate": 7.779487601838065e-06, "loss": 0.7859, "step": 1557 }, { "epoch": 0.5894258961505722, "grad_norm": 1.0628832051157708, "learning_rate": 7.778910052233642e-06, "loss": 0.8021, "step": 1558 }, { "epoch": 0.5898042182918756, "grad_norm": 1.0898131031337233, "learning_rate": 7.778331768776237e-06, "loss": 0.802, "step": 1559 }, { "epoch": 0.5901825404331789, "grad_norm": 1.0649413521341573, "learning_rate": 7.77775275157815e-06, "loss": 0.8217, "step": 1560 }, { "epoch": 0.5905608625744821, "grad_norm": 1.0368511400497493, "learning_rate": 7.777173000751825e-06, "loss": 0.7819, "step": 1561 }, { "epoch": 0.5909391847157854, "grad_norm": 1.020241580639323, "learning_rate": 7.776592516409848e-06, "loss": 0.8435, "step": 1562 }, { "epoch": 0.5913175068570888, "grad_norm": 1.039218236167864, "learning_rate": 7.776011298664945e-06, "loss": 0.822, "step": 1563 }, { "epoch": 0.5916958289983921, "grad_norm": 1.0277738056724017, "learning_rate": 7.775429347629992e-06, "loss": 0.7755, "step": 1564 }, { "epoch": 0.5920741511396954, "grad_norm": 0.9767055405759969, "learning_rate": 7.774846663417996e-06, "loss": 0.8259, "step": 1565 }, { "epoch": 0.5924524732809988, "grad_norm": 1.0409555633420142, "learning_rate": 7.774263246142116e-06, "loss": 0.7829, "step": 1566 }, { "epoch": 0.5928307954223021, "grad_norm": 1.0275312312209073, "learning_rate": 7.77367909591565e-06, "loss": 0.7724, "step": 1567 }, { "epoch": 0.5932091175636054, "grad_norm": 1.0128232786560865, "learning_rate": 7.773094212852036e-06, "loss": 0.778, "step": 1568 }, { "epoch": 0.5935874397049087, "grad_norm": 1.010220293379828, "learning_rate": 7.77250859706486e-06, "loss": 0.8122, "step": 1569 }, { "epoch": 0.5939657618462121, "grad_norm": 1.0377569519031766, "learning_rate": 7.771922248667843e-06, "loss": 0.7944, "step": 1570 }, { "epoch": 0.5943440839875154, "grad_norm": 1.0056143743542545, "learning_rate": 7.771335167774855e-06, "loss": 0.8184, "step": 1571 }, { "epoch": 0.5947224061288187, "grad_norm": 1.0823167997700618, "learning_rate": 7.770747354499902e-06, "loss": 0.793, "step": 1572 }, { "epoch": 0.5951007282701221, "grad_norm": 1.005554310069684, "learning_rate": 7.770158808957142e-06, "loss": 0.8294, "step": 1573 }, { "epoch": 0.5954790504114253, "grad_norm": 1.016774447299906, "learning_rate": 7.769569531260861e-06, "loss": 0.7916, "step": 1574 }, { "epoch": 0.5958573725527286, "grad_norm": 0.9815704963237092, "learning_rate": 7.7689795215255e-06, "loss": 0.7873, "step": 1575 }, { "epoch": 0.5962356946940319, "grad_norm": 1.054358096080715, "learning_rate": 7.768388779865636e-06, "loss": 0.8164, "step": 1576 }, { "epoch": 0.5966140168353353, "grad_norm": 0.9774109882411877, "learning_rate": 7.767797306395988e-06, "loss": 0.791, "step": 1577 }, { "epoch": 0.5969923389766386, "grad_norm": 1.0358457305091455, "learning_rate": 7.76720510123142e-06, "loss": 0.7707, "step": 1578 }, { "epoch": 0.5973706611179419, "grad_norm": 1.0624591531096403, "learning_rate": 7.766612164486936e-06, "loss": 0.8472, "step": 1579 }, { "epoch": 0.5977489832592452, "grad_norm": 0.9928836589328845, "learning_rate": 7.766018496277682e-06, "loss": 0.7902, "step": 1580 }, { "epoch": 0.5981273054005486, "grad_norm": 1.0280490587815976, "learning_rate": 7.765424096718946e-06, "loss": 0.7841, "step": 1581 }, { "epoch": 0.5985056275418519, "grad_norm": 0.9873621543820231, "learning_rate": 7.76482896592616e-06, "loss": 0.8006, "step": 1582 }, { "epoch": 0.5988839496831552, "grad_norm": 1.0709729821860812, "learning_rate": 7.764233104014897e-06, "loss": 0.8682, "step": 1583 }, { "epoch": 0.5992622718244586, "grad_norm": 0.9867939695157474, "learning_rate": 7.76363651110087e-06, "loss": 0.7879, "step": 1584 }, { "epoch": 0.5996405939657619, "grad_norm": 1.0795152732921542, "learning_rate": 7.763039187299937e-06, "loss": 0.815, "step": 1585 }, { "epoch": 0.6000189161070651, "grad_norm": 0.9899000945502743, "learning_rate": 7.762441132728095e-06, "loss": 0.7855, "step": 1586 }, { "epoch": 0.6003972382483684, "grad_norm": 1.0252908086535142, "learning_rate": 7.761842347501485e-06, "loss": 0.8165, "step": 1587 }, { "epoch": 0.6007755603896718, "grad_norm": 1.0423466115896767, "learning_rate": 7.76124283173639e-06, "loss": 0.8567, "step": 1588 }, { "epoch": 0.6011538825309751, "grad_norm": 0.9948472361654808, "learning_rate": 7.760642585549233e-06, "loss": 0.7931, "step": 1589 }, { "epoch": 0.6015322046722784, "grad_norm": 0.9998595808495474, "learning_rate": 7.760041609056582e-06, "loss": 0.7922, "step": 1590 }, { "epoch": 0.6019105268135818, "grad_norm": 1.0113044627393564, "learning_rate": 7.759439902375141e-06, "loss": 0.7983, "step": 1591 }, { "epoch": 0.6022888489548851, "grad_norm": 1.052771258939431, "learning_rate": 7.758837465621764e-06, "loss": 0.8088, "step": 1592 }, { "epoch": 0.6026671710961884, "grad_norm": 1.0123858085251436, "learning_rate": 7.758234298913439e-06, "loss": 0.784, "step": 1593 }, { "epoch": 0.6030454932374917, "grad_norm": 1.0337794095975905, "learning_rate": 7.757630402367303e-06, "loss": 0.7997, "step": 1594 }, { "epoch": 0.6034238153787951, "grad_norm": 0.9846999031423823, "learning_rate": 7.757025776100625e-06, "loss": 0.7447, "step": 1595 }, { "epoch": 0.6038021375200984, "grad_norm": 1.0462409901802558, "learning_rate": 7.756420420230828e-06, "loss": 0.7686, "step": 1596 }, { "epoch": 0.6038021375200984, "eval_loss": 0.8007391691207886, "eval_runtime": 27.0514, "eval_samples_per_second": 32.715, "eval_steps_per_second": 1.035, "step": 1596 }, { "epoch": 0.6038021375200984, "eval_bench_accuracy_arc_challenge": 0.25, "eval_bench_accuracy_hellaswag": 0.21, "eval_bench_accuracy_mmlu": 0.25217391304347825, "eval_bench_average_accuracy": 0.23739130434782607, "eval_bench_loss": 6.375945509525767, "eval_bench_total_accuracy": 0.23296703296703297, "step": 1596 }, { "epoch": 0.6041804596614017, "grad_norm": 1.0790625835061922, "learning_rate": 7.755814334875466e-06, "loss": 0.8091, "step": 1597 }, { "epoch": 0.6045587818027051, "grad_norm": 0.9802043723000299, "learning_rate": 7.75520752015224e-06, "loss": 0.7256, "step": 1598 }, { "epoch": 0.6049371039440083, "grad_norm": 0.9923431981852016, "learning_rate": 7.754599976178994e-06, "loss": 0.8054, "step": 1599 }, { "epoch": 0.6053154260853116, "grad_norm": 1.0242822958979938, "learning_rate": 7.753991703073709e-06, "loss": 0.7947, "step": 1600 }, { "epoch": 0.6056937482266149, "grad_norm": 1.0693420250669043, "learning_rate": 7.75338270095451e-06, "loss": 0.7714, "step": 1601 }, { "epoch": 0.6060720703679183, "grad_norm": 1.0393417772805222, "learning_rate": 7.752772969939662e-06, "loss": 0.7984, "step": 1602 }, { "epoch": 0.6064503925092216, "grad_norm": 1.0193556335184584, "learning_rate": 7.752162510147576e-06, "loss": 0.7845, "step": 1603 }, { "epoch": 0.6068287146505249, "grad_norm": 1.0439223450090194, "learning_rate": 7.751551321696798e-06, "loss": 0.7902, "step": 1604 }, { "epoch": 0.6072070367918282, "grad_norm": 1.0458764132750307, "learning_rate": 7.75093940470602e-06, "loss": 0.8277, "step": 1605 }, { "epoch": 0.6075853589331316, "grad_norm": 1.0304823323522874, "learning_rate": 7.750326759294077e-06, "loss": 0.7936, "step": 1606 }, { "epoch": 0.6079636810744349, "grad_norm": 1.037572458907066, "learning_rate": 7.749713385579942e-06, "loss": 0.779, "step": 1607 }, { "epoch": 0.6083420032157382, "grad_norm": 1.0233220079303753, "learning_rate": 7.749099283682727e-06, "loss": 0.7924, "step": 1608 }, { "epoch": 0.6087203253570416, "grad_norm": 1.0490780083116327, "learning_rate": 7.748484453721694e-06, "loss": 0.8337, "step": 1609 }, { "epoch": 0.6090986474983449, "grad_norm": 1.0173257743419322, "learning_rate": 7.747868895816236e-06, "loss": 0.7673, "step": 1610 }, { "epoch": 0.6094769696396481, "grad_norm": 1.0573789547993953, "learning_rate": 7.747252610085895e-06, "loss": 0.8377, "step": 1611 }, { "epoch": 0.6098552917809514, "grad_norm": 1.0257255841383113, "learning_rate": 7.746635596650352e-06, "loss": 0.7728, "step": 1612 }, { "epoch": 0.6102336139222548, "grad_norm": 1.0160660389387, "learning_rate": 7.746017855629429e-06, "loss": 0.8025, "step": 1613 }, { "epoch": 0.6106119360635581, "grad_norm": 1.0602513504043805, "learning_rate": 7.74539938714309e-06, "loss": 0.7925, "step": 1614 }, { "epoch": 0.6109902582048614, "grad_norm": 1.0377020898351703, "learning_rate": 7.744780191311437e-06, "loss": 0.804, "step": 1615 }, { "epoch": 0.6113685803461648, "grad_norm": 0.9962327806446186, "learning_rate": 7.744160268254718e-06, "loss": 0.7463, "step": 1616 }, { "epoch": 0.6117469024874681, "grad_norm": 1.03576395621217, "learning_rate": 7.743539618093323e-06, "loss": 0.8125, "step": 1617 }, { "epoch": 0.6121252246287714, "grad_norm": 1.0791330433766595, "learning_rate": 7.742918240947774e-06, "loss": 0.7497, "step": 1618 }, { "epoch": 0.6125035467700747, "grad_norm": 1.0186732713870292, "learning_rate": 7.742296136938745e-06, "loss": 0.7715, "step": 1619 }, { "epoch": 0.6128818689113781, "grad_norm": 1.0549459798818361, "learning_rate": 7.741673306187047e-06, "loss": 0.7663, "step": 1620 }, { "epoch": 0.6132601910526814, "grad_norm": 0.9830530108058492, "learning_rate": 7.74104974881363e-06, "loss": 0.8146, "step": 1621 }, { "epoch": 0.6136385131939847, "grad_norm": 1.0384186325465743, "learning_rate": 7.74042546493959e-06, "loss": 0.7864, "step": 1622 }, { "epoch": 0.614016835335288, "grad_norm": 1.050915873907994, "learning_rate": 7.739800454686156e-06, "loss": 0.7966, "step": 1623 }, { "epoch": 0.6143951574765913, "grad_norm": 1.0241953725880033, "learning_rate": 7.739174718174705e-06, "loss": 0.7659, "step": 1624 }, { "epoch": 0.6147734796178946, "grad_norm": 1.0278047735993348, "learning_rate": 7.738548255526757e-06, "loss": 0.7753, "step": 1625 }, { "epoch": 0.6151518017591979, "grad_norm": 1.0028879958633992, "learning_rate": 7.737921066863963e-06, "loss": 0.798, "step": 1626 }, { "epoch": 0.6155301239005013, "grad_norm": 1.046709030024919, "learning_rate": 7.737293152308125e-06, "loss": 0.8318, "step": 1627 }, { "epoch": 0.6159084460418046, "grad_norm": 1.053664353449831, "learning_rate": 7.736664511981184e-06, "loss": 0.8518, "step": 1628 }, { "epoch": 0.6162867681831079, "grad_norm": 0.9978105688058767, "learning_rate": 7.736035146005216e-06, "loss": 0.7807, "step": 1629 }, { "epoch": 0.6166650903244112, "grad_norm": 1.0998599207938173, "learning_rate": 7.735405054502443e-06, "loss": 0.8517, "step": 1630 }, { "epoch": 0.6170434124657146, "grad_norm": 1.0347549984516864, "learning_rate": 7.734774237595227e-06, "loss": 0.7861, "step": 1631 }, { "epoch": 0.6174217346070179, "grad_norm": 1.0604030894353325, "learning_rate": 7.734142695406072e-06, "loss": 0.8444, "step": 1632 }, { "epoch": 0.6178000567483212, "grad_norm": 0.9995358654268639, "learning_rate": 7.73351042805762e-06, "loss": 0.7982, "step": 1633 }, { "epoch": 0.6181783788896246, "grad_norm": 1.012063791302332, "learning_rate": 7.732877435672656e-06, "loss": 0.7891, "step": 1634 }, { "epoch": 0.6185567010309279, "grad_norm": 1.062079535667684, "learning_rate": 7.732243718374105e-06, "loss": 0.7953, "step": 1635 }, { "epoch": 0.6189350231722311, "grad_norm": 1.0049506132948145, "learning_rate": 7.731609276285034e-06, "loss": 0.8185, "step": 1636 }, { "epoch": 0.6193133453135344, "grad_norm": 0.9787699976228371, "learning_rate": 7.730974109528651e-06, "loss": 0.8099, "step": 1637 }, { "epoch": 0.6196916674548378, "grad_norm": 0.9716390457115083, "learning_rate": 7.730338218228298e-06, "loss": 0.7695, "step": 1638 }, { "epoch": 0.6200699895961411, "grad_norm": 0.9806455110749785, "learning_rate": 7.729701602507469e-06, "loss": 0.7199, "step": 1639 }, { "epoch": 0.6204483117374444, "grad_norm": 1.0303904399928674, "learning_rate": 7.729064262489791e-06, "loss": 0.8018, "step": 1640 }, { "epoch": 0.6208266338787477, "grad_norm": 1.0184745198287024, "learning_rate": 7.72842619829903e-06, "loss": 0.8168, "step": 1641 }, { "epoch": 0.6212049560200511, "grad_norm": 1.0350761019221557, "learning_rate": 7.727787410059102e-06, "loss": 0.8063, "step": 1642 }, { "epoch": 0.6215832781613544, "grad_norm": 0.9997598615132083, "learning_rate": 7.727147897894055e-06, "loss": 0.7692, "step": 1643 }, { "epoch": 0.6219616003026577, "grad_norm": 1.0317018080080016, "learning_rate": 7.72650766192808e-06, "loss": 0.7963, "step": 1644 }, { "epoch": 0.6223399224439611, "grad_norm": 1.058330305743686, "learning_rate": 7.725866702285508e-06, "loss": 0.7778, "step": 1645 }, { "epoch": 0.6227182445852644, "grad_norm": 1.050475543436919, "learning_rate": 7.725225019090813e-06, "loss": 0.8052, "step": 1646 }, { "epoch": 0.6230965667265677, "grad_norm": 1.0381951307937078, "learning_rate": 7.724582612468609e-06, "loss": 0.7643, "step": 1647 }, { "epoch": 0.623474888867871, "grad_norm": 0.9960696467209328, "learning_rate": 7.723939482543647e-06, "loss": 0.781, "step": 1648 }, { "epoch": 0.6238532110091743, "grad_norm": 1.0235710160288658, "learning_rate": 7.723295629440823e-06, "loss": 0.7818, "step": 1649 }, { "epoch": 0.6242315331504776, "grad_norm": 0.9987662526618373, "learning_rate": 7.722651053285168e-06, "loss": 0.7532, "step": 1650 }, { "epoch": 0.6246098552917809, "grad_norm": 1.038603322649077, "learning_rate": 7.722005754201863e-06, "loss": 0.7995, "step": 1651 }, { "epoch": 0.6249881774330843, "grad_norm": 1.0372844825153233, "learning_rate": 7.721359732316216e-06, "loss": 0.7982, "step": 1652 }, { "epoch": 0.6253664995743876, "grad_norm": 1.0075983510701718, "learning_rate": 7.720712987753687e-06, "loss": 0.771, "step": 1653 }, { "epoch": 0.6257448217156909, "grad_norm": 1.060885095951037, "learning_rate": 7.72006552063987e-06, "loss": 0.8095, "step": 1654 }, { "epoch": 0.6261231438569942, "grad_norm": 1.024942261074342, "learning_rate": 7.719417331100501e-06, "loss": 0.8175, "step": 1655 }, { "epoch": 0.6265014659982976, "grad_norm": 1.0259969128854978, "learning_rate": 7.718768419261458e-06, "loss": 0.7614, "step": 1656 }, { "epoch": 0.6268797881396009, "grad_norm": 1.0032297451874017, "learning_rate": 7.718118785248759e-06, "loss": 0.7612, "step": 1657 }, { "epoch": 0.6272581102809042, "grad_norm": 1.0210932763381098, "learning_rate": 7.717468429188556e-06, "loss": 0.7755, "step": 1658 }, { "epoch": 0.6276364324222075, "grad_norm": 1.046603168853803, "learning_rate": 7.71681735120715e-06, "loss": 0.7888, "step": 1659 }, { "epoch": 0.6280147545635109, "grad_norm": 1.0302944601931032, "learning_rate": 7.716165551430978e-06, "loss": 0.8215, "step": 1660 }, { "epoch": 0.6283930767048141, "grad_norm": 1.0538426037667707, "learning_rate": 7.715513029986616e-06, "loss": 0.8277, "step": 1661 }, { "epoch": 0.6287713988461174, "grad_norm": 1.0079131456868133, "learning_rate": 7.714859787000784e-06, "loss": 0.7898, "step": 1662 }, { "epoch": 0.6291497209874208, "grad_norm": 1.0091132558305784, "learning_rate": 7.714205822600338e-06, "loss": 0.7628, "step": 1663 }, { "epoch": 0.6295280431287241, "grad_norm": 1.0370707510362853, "learning_rate": 7.713551136912277e-06, "loss": 0.7847, "step": 1664 }, { "epoch": 0.6299063652700274, "grad_norm": 1.0254976981220805, "learning_rate": 7.712895730063737e-06, "loss": 0.8251, "step": 1665 }, { "epoch": 0.6302846874113307, "grad_norm": 1.0129086665617333, "learning_rate": 7.712239602181998e-06, "loss": 0.813, "step": 1666 }, { "epoch": 0.6306630095526341, "grad_norm": 1.0211770501504658, "learning_rate": 7.711582753394478e-06, "loss": 0.7909, "step": 1667 }, { "epoch": 0.6310413316939374, "grad_norm": 1.2302756712980163, "learning_rate": 7.710925183828736e-06, "loss": 0.782, "step": 1668 }, { "epoch": 0.6314196538352407, "grad_norm": 1.0606820966683679, "learning_rate": 7.710266893612468e-06, "loss": 0.8001, "step": 1669 }, { "epoch": 0.6317979759765441, "grad_norm": 1.0257958327969605, "learning_rate": 7.70960788287351e-06, "loss": 0.7715, "step": 1670 }, { "epoch": 0.6321762981178474, "grad_norm": 1.033181617178253, "learning_rate": 7.708948151739847e-06, "loss": 0.7884, "step": 1671 }, { "epoch": 0.6325546202591507, "grad_norm": 1.0142271201151716, "learning_rate": 7.708287700339588e-06, "loss": 0.7846, "step": 1672 }, { "epoch": 0.632932942400454, "grad_norm": 1.0581952369577206, "learning_rate": 7.707626528800999e-06, "loss": 0.835, "step": 1673 }, { "epoch": 0.6333112645417573, "grad_norm": 1.031831226064096, "learning_rate": 7.706964637252472e-06, "loss": 0.7808, "step": 1674 }, { "epoch": 0.6336895866830606, "grad_norm": 1.034926042820135, "learning_rate": 7.706302025822546e-06, "loss": 0.8133, "step": 1675 }, { "epoch": 0.6340679088243639, "grad_norm": 0.9974796232689039, "learning_rate": 7.705638694639897e-06, "loss": 0.8022, "step": 1676 }, { "epoch": 0.6344462309656672, "grad_norm": 0.9991746871631939, "learning_rate": 7.704974643833345e-06, "loss": 0.7768, "step": 1677 }, { "epoch": 0.6348245531069706, "grad_norm": 1.0647934668234986, "learning_rate": 7.704309873531842e-06, "loss": 0.7784, "step": 1678 }, { "epoch": 0.6352028752482739, "grad_norm": 1.0706641503151557, "learning_rate": 7.70364438386449e-06, "loss": 0.7549, "step": 1679 }, { "epoch": 0.6355811973895772, "grad_norm": 1.5575289700539314, "learning_rate": 7.70297817496052e-06, "loss": 0.7869, "step": 1680 }, { "epoch": 0.6359595195308806, "grad_norm": 1.0441884975223152, "learning_rate": 7.702311246949312e-06, "loss": 0.8212, "step": 1681 }, { "epoch": 0.6363378416721839, "grad_norm": 1.0184875000693254, "learning_rate": 7.701643599960377e-06, "loss": 0.7783, "step": 1682 }, { "epoch": 0.6367161638134872, "grad_norm": 1.056484375092538, "learning_rate": 7.700975234123374e-06, "loss": 0.7997, "step": 1683 }, { "epoch": 0.6370944859547905, "grad_norm": 1.0158431220473627, "learning_rate": 7.700306149568096e-06, "loss": 0.7887, "step": 1684 }, { "epoch": 0.6374728080960939, "grad_norm": 1.005886147632736, "learning_rate": 7.699636346424476e-06, "loss": 0.8146, "step": 1685 }, { "epoch": 0.6378511302373971, "grad_norm": 0.9516674282028371, "learning_rate": 7.698965824822591e-06, "loss": 0.7617, "step": 1686 }, { "epoch": 0.6382294523787004, "grad_norm": 1.0354398239486777, "learning_rate": 7.698294584892653e-06, "loss": 0.7698, "step": 1687 }, { "epoch": 0.6386077745200038, "grad_norm": 1.0412153778199809, "learning_rate": 7.69762262676501e-06, "loss": 0.7741, "step": 1688 }, { "epoch": 0.6389860966613071, "grad_norm": 1.0038063833719368, "learning_rate": 7.696949950570162e-06, "loss": 0.7726, "step": 1689 }, { "epoch": 0.6393644188026104, "grad_norm": 1.0041297661402129, "learning_rate": 7.696276556438736e-06, "loss": 0.8076, "step": 1690 }, { "epoch": 0.6397427409439137, "grad_norm": 1.052469874333398, "learning_rate": 7.695602444501503e-06, "loss": 0.7906, "step": 1691 }, { "epoch": 0.6401210630852171, "grad_norm": 0.9490194460452617, "learning_rate": 7.694927614889376e-06, "loss": 0.7188, "step": 1692 }, { "epoch": 0.6404993852265204, "grad_norm": 0.974323163548883, "learning_rate": 7.694252067733404e-06, "loss": 0.753, "step": 1693 }, { "epoch": 0.6408777073678237, "grad_norm": 1.0319007840691403, "learning_rate": 7.693575803164774e-06, "loss": 0.7962, "step": 1694 }, { "epoch": 0.641256029509127, "grad_norm": 1.0299952133041577, "learning_rate": 7.692898821314816e-06, "loss": 0.7723, "step": 1695 }, { "epoch": 0.6416343516504304, "grad_norm": 1.0632785008902024, "learning_rate": 7.692221122315e-06, "loss": 0.7536, "step": 1696 }, { "epoch": 0.6420126737917337, "grad_norm": 1.0478356927175443, "learning_rate": 7.69154270629693e-06, "loss": 0.7759, "step": 1697 }, { "epoch": 0.642390995933037, "grad_norm": 1.0207221782050084, "learning_rate": 7.690863573392355e-06, "loss": 0.8025, "step": 1698 }, { "epoch": 0.6427693180743403, "grad_norm": 1.0307450911725362, "learning_rate": 7.690183723733158e-06, "loss": 0.8126, "step": 1699 }, { "epoch": 0.6431476402156436, "grad_norm": 0.9558201805744811, "learning_rate": 7.689503157451366e-06, "loss": 0.7926, "step": 1700 }, { "epoch": 0.6435259623569469, "grad_norm": 0.9839314509833194, "learning_rate": 7.68882187467914e-06, "loss": 0.7982, "step": 1701 }, { "epoch": 0.6439042844982502, "grad_norm": 1.0446036605229558, "learning_rate": 7.688139875548786e-06, "loss": 0.7424, "step": 1702 }, { "epoch": 0.6442826066395536, "grad_norm": 0.9747599328413645, "learning_rate": 7.687457160192746e-06, "loss": 0.7769, "step": 1703 }, { "epoch": 0.6446609287808569, "grad_norm": 1.0017104708165576, "learning_rate": 7.6867737287436e-06, "loss": 0.7779, "step": 1704 }, { "epoch": 0.6450392509221602, "grad_norm": 1.0396981093860427, "learning_rate": 7.686089581334069e-06, "loss": 0.7966, "step": 1705 }, { "epoch": 0.6454175730634636, "grad_norm": 1.0077578946931687, "learning_rate": 7.685404718097011e-06, "loss": 0.7658, "step": 1706 }, { "epoch": 0.6457958952047669, "grad_norm": 1.0045936301109948, "learning_rate": 7.684719139165426e-06, "loss": 0.8215, "step": 1707 }, { "epoch": 0.6461742173460702, "grad_norm": 1.0059220607870412, "learning_rate": 7.684032844672452e-06, "loss": 0.784, "step": 1708 }, { "epoch": 0.6465525394873735, "grad_norm": 1.002030780249217, "learning_rate": 7.683345834751362e-06, "loss": 0.754, "step": 1709 }, { "epoch": 0.6469308616286769, "grad_norm": 1.0524082695853973, "learning_rate": 7.682658109535575e-06, "loss": 0.8141, "step": 1710 }, { "epoch": 0.6473091837699801, "grad_norm": 1.023391717099541, "learning_rate": 7.681969669158643e-06, "loss": 0.8029, "step": 1711 }, { "epoch": 0.6476875059112834, "grad_norm": 1.0537878870256816, "learning_rate": 7.68128051375426e-06, "loss": 0.8026, "step": 1712 }, { "epoch": 0.6480658280525867, "grad_norm": 0.9946301646936768, "learning_rate": 7.680590643456258e-06, "loss": 0.8154, "step": 1713 }, { "epoch": 0.6484441501938901, "grad_norm": 1.0129808485922718, "learning_rate": 7.679900058398606e-06, "loss": 0.7482, "step": 1714 }, { "epoch": 0.6488224723351934, "grad_norm": 1.1366026781982712, "learning_rate": 7.679208758715417e-06, "loss": 0.7844, "step": 1715 }, { "epoch": 0.6492007944764967, "grad_norm": 1.0252138838659255, "learning_rate": 7.678516744540936e-06, "loss": 0.7827, "step": 1716 }, { "epoch": 0.6495791166178001, "grad_norm": 1.0483329033578623, "learning_rate": 7.67782401600955e-06, "loss": 0.7995, "step": 1717 }, { "epoch": 0.6499574387591034, "grad_norm": 0.9954302178962173, "learning_rate": 7.677130573255787e-06, "loss": 0.7528, "step": 1718 }, { "epoch": 0.6503357609004067, "grad_norm": 1.0342284002896778, "learning_rate": 7.67643641641431e-06, "loss": 0.7967, "step": 1719 }, { "epoch": 0.65071408304171, "grad_norm": 1.0744541931554912, "learning_rate": 7.675741545619926e-06, "loss": 0.7959, "step": 1720 }, { "epoch": 0.6510924051830134, "grad_norm": 0.9960576642926111, "learning_rate": 7.675045961007571e-06, "loss": 0.7644, "step": 1721 }, { "epoch": 0.6514707273243167, "grad_norm": 1.0388432797415568, "learning_rate": 7.674349662712328e-06, "loss": 0.8452, "step": 1722 }, { "epoch": 0.65184904946562, "grad_norm": 1.0809172859395315, "learning_rate": 7.673652650869415e-06, "loss": 0.8068, "step": 1723 }, { "epoch": 0.6522273716069233, "grad_norm": 1.0066539502318497, "learning_rate": 7.672954925614193e-06, "loss": 0.7709, "step": 1724 }, { "epoch": 0.6526056937482266, "grad_norm": 1.0418268199259764, "learning_rate": 7.672256487082155e-06, "loss": 0.7932, "step": 1725 }, { "epoch": 0.6529840158895299, "grad_norm": 1.0245053090908052, "learning_rate": 7.671557335408935e-06, "loss": 0.798, "step": 1726 }, { "epoch": 0.6533623380308332, "grad_norm": 1.0356795152001224, "learning_rate": 7.670857470730309e-06, "loss": 0.7573, "step": 1727 }, { "epoch": 0.6537406601721366, "grad_norm": 1.0311220411463944, "learning_rate": 7.670156893182188e-06, "loss": 0.8159, "step": 1728 }, { "epoch": 0.6541189823134399, "grad_norm": 0.9968740214468425, "learning_rate": 7.66945560290062e-06, "loss": 0.8174, "step": 1729 }, { "epoch": 0.6541189823134399, "eval_loss": 0.7927515506744385, "eval_runtime": 26.7774, "eval_samples_per_second": 33.05, "eval_steps_per_second": 1.046, "step": 1729 }, { "epoch": 0.6541189823134399, "eval_bench_accuracy_arc_challenge": 0.0, "eval_bench_accuracy_hellaswag": 0.21, "eval_bench_accuracy_mmlu": 0.23478260869565218, "eval_bench_average_accuracy": 0.1482608695652174, "eval_bench_loss": 7.814903928522478, "eval_bench_total_accuracy": 0.15164835164835164, "step": 1729 }, { "epoch": 0.6544973044547432, "grad_norm": 1.0536869570872927, "learning_rate": 7.668753600021795e-06, "loss": 0.7894, "step": 1730 }, { "epoch": 0.6548756265960465, "grad_norm": 1.0802849973303468, "learning_rate": 7.66805088468204e-06, "loss": 0.8128, "step": 1731 }, { "epoch": 0.6552539487373499, "grad_norm": 1.0195535501035122, "learning_rate": 7.66734745701782e-06, "loss": 0.7698, "step": 1732 }, { "epoch": 0.6556322708786532, "grad_norm": 0.9866819845303567, "learning_rate": 7.666643317165737e-06, "loss": 0.7632, "step": 1733 }, { "epoch": 0.6560105930199565, "grad_norm": 1.0362620307566515, "learning_rate": 7.665938465262536e-06, "loss": 0.8242, "step": 1734 }, { "epoch": 0.6563889151612599, "grad_norm": 1.005122320879091, "learning_rate": 7.665232901445093e-06, "loss": 0.8128, "step": 1735 }, { "epoch": 0.6567672373025631, "grad_norm": 0.9968147052835493, "learning_rate": 7.66452662585043e-06, "loss": 0.7765, "step": 1736 }, { "epoch": 0.6571455594438664, "grad_norm": 1.0160098359583503, "learning_rate": 7.663819638615705e-06, "loss": 0.769, "step": 1737 }, { "epoch": 0.6575238815851697, "grad_norm": 0.9957799905329473, "learning_rate": 7.663111939878207e-06, "loss": 0.75, "step": 1738 }, { "epoch": 0.6579022037264731, "grad_norm": 0.9817964252654222, "learning_rate": 7.662403529775372e-06, "loss": 0.7814, "step": 1739 }, { "epoch": 0.6582805258677764, "grad_norm": 0.9928916742992132, "learning_rate": 7.661694408444773e-06, "loss": 0.7904, "step": 1740 }, { "epoch": 0.6586588480090797, "grad_norm": 1.0410892155118083, "learning_rate": 7.660984576024117e-06, "loss": 0.8191, "step": 1741 }, { "epoch": 0.6590371701503831, "grad_norm": 1.0021028586166405, "learning_rate": 7.660274032651249e-06, "loss": 0.7712, "step": 1742 }, { "epoch": 0.6594154922916864, "grad_norm": 0.9990600675172764, "learning_rate": 7.65956277846416e-06, "loss": 0.7857, "step": 1743 }, { "epoch": 0.6597938144329897, "grad_norm": 1.0992751750590166, "learning_rate": 7.658850813600969e-06, "loss": 0.7878, "step": 1744 }, { "epoch": 0.660172136574293, "grad_norm": 1.0189976892843522, "learning_rate": 7.65813813819994e-06, "loss": 0.77, "step": 1745 }, { "epoch": 0.6605504587155964, "grad_norm": 1.0468429508760897, "learning_rate": 7.657424752399471e-06, "loss": 0.7768, "step": 1746 }, { "epoch": 0.6609287808568997, "grad_norm": 1.0374665153019, "learning_rate": 7.6567106563381e-06, "loss": 0.8103, "step": 1747 }, { "epoch": 0.661307102998203, "grad_norm": 1.0713460469365848, "learning_rate": 7.655995850154501e-06, "loss": 0.7646, "step": 1748 }, { "epoch": 0.6616854251395063, "grad_norm": 1.048711304359486, "learning_rate": 7.655280333987491e-06, "loss": 0.7852, "step": 1749 }, { "epoch": 0.6620637472808096, "grad_norm": 1.0319143016049546, "learning_rate": 7.654564107976017e-06, "loss": 0.7979, "step": 1750 }, { "epoch": 0.6624420694221129, "grad_norm": 1.0575930996275595, "learning_rate": 7.653847172259169e-06, "loss": 0.7768, "step": 1751 }, { "epoch": 0.6628203915634162, "grad_norm": 0.9638702778680636, "learning_rate": 7.653129526976173e-06, "loss": 0.7979, "step": 1752 }, { "epoch": 0.6631987137047196, "grad_norm": 0.9690337454201767, "learning_rate": 7.652411172266398e-06, "loss": 0.7894, "step": 1753 }, { "epoch": 0.6635770358460229, "grad_norm": 1.0072303768845905, "learning_rate": 7.65169210826934e-06, "loss": 0.7302, "step": 1754 }, { "epoch": 0.6639553579873262, "grad_norm": 1.0168462219112109, "learning_rate": 7.650972335124644e-06, "loss": 0.7918, "step": 1755 }, { "epoch": 0.6643336801286295, "grad_norm": 0.9845272479814176, "learning_rate": 7.650251852972084e-06, "loss": 0.7798, "step": 1756 }, { "epoch": 0.6647120022699329, "grad_norm": 1.0559359255774574, "learning_rate": 7.649530661951578e-06, "loss": 0.7835, "step": 1757 }, { "epoch": 0.6650903244112362, "grad_norm": 1.0127474528668845, "learning_rate": 7.64880876220318e-06, "loss": 0.7566, "step": 1758 }, { "epoch": 0.6654686465525395, "grad_norm": 1.067173774382862, "learning_rate": 7.648086153867078e-06, "loss": 0.7738, "step": 1759 }, { "epoch": 0.6658469686938429, "grad_norm": 1.0262747793123224, "learning_rate": 7.6473628370836e-06, "loss": 0.7833, "step": 1760 }, { "epoch": 0.6662252908351461, "grad_norm": 1.0515582564211456, "learning_rate": 7.646638811993216e-06, "loss": 0.7538, "step": 1761 }, { "epoch": 0.6666036129764494, "grad_norm": 1.0329994771612065, "learning_rate": 7.645914078736526e-06, "loss": 0.8164, "step": 1762 }, { "epoch": 0.6669819351177527, "grad_norm": 1.0311907540077614, "learning_rate": 7.645188637454272e-06, "loss": 0.7706, "step": 1763 }, { "epoch": 0.6673602572590561, "grad_norm": 1.0409947640223565, "learning_rate": 7.644462488287334e-06, "loss": 0.7885, "step": 1764 }, { "epoch": 0.6677385794003594, "grad_norm": 0.988219756000234, "learning_rate": 7.643735631376724e-06, "loss": 0.7408, "step": 1765 }, { "epoch": 0.6681169015416627, "grad_norm": 1.027004288225805, "learning_rate": 7.643008066863598e-06, "loss": 0.8121, "step": 1766 }, { "epoch": 0.6684952236829661, "grad_norm": 1.0184065601333092, "learning_rate": 7.642279794889249e-06, "loss": 0.7576, "step": 1767 }, { "epoch": 0.6688735458242694, "grad_norm": 1.043603934502605, "learning_rate": 7.641550815595102e-06, "loss": 0.771, "step": 1768 }, { "epoch": 0.6692518679655727, "grad_norm": 1.060392114018632, "learning_rate": 7.640821129122723e-06, "loss": 0.8247, "step": 1769 }, { "epoch": 0.669630190106876, "grad_norm": 1.0126323816870029, "learning_rate": 7.640090735613818e-06, "loss": 0.8022, "step": 1770 }, { "epoch": 0.6700085122481794, "grad_norm": 1.1648366101787067, "learning_rate": 7.639359635210222e-06, "loss": 0.7826, "step": 1771 }, { "epoch": 0.6703868343894827, "grad_norm": 1.0724674686904885, "learning_rate": 7.638627828053918e-06, "loss": 0.7897, "step": 1772 }, { "epoch": 0.6707651565307859, "grad_norm": 1.0540972019117152, "learning_rate": 7.637895314287016e-06, "loss": 0.7645, "step": 1773 }, { "epoch": 0.6711434786720892, "grad_norm": 1.0057331810331451, "learning_rate": 7.63716209405177e-06, "loss": 0.816, "step": 1774 }, { "epoch": 0.6715218008133926, "grad_norm": 0.9970921236923102, "learning_rate": 7.63642816749057e-06, "loss": 0.7671, "step": 1775 }, { "epoch": 0.6719001229546959, "grad_norm": 1.002453880727358, "learning_rate": 7.635693534745941e-06, "loss": 0.7885, "step": 1776 }, { "epoch": 0.6722784450959992, "grad_norm": 1.0312771975163908, "learning_rate": 7.634958195960548e-06, "loss": 0.7951, "step": 1777 }, { "epoch": 0.6726567672373026, "grad_norm": 1.0177245342291783, "learning_rate": 7.634222151277188e-06, "loss": 0.773, "step": 1778 }, { "epoch": 0.6730350893786059, "grad_norm": 1.060998481737934, "learning_rate": 7.633485400838804e-06, "loss": 0.7924, "step": 1779 }, { "epoch": 0.6734134115199092, "grad_norm": 1.0340561242421995, "learning_rate": 7.632747944788468e-06, "loss": 0.8451, "step": 1780 }, { "epoch": 0.6737917336612125, "grad_norm": 1.0461873170538059, "learning_rate": 7.63200978326939e-06, "loss": 0.7896, "step": 1781 }, { "epoch": 0.6741700558025159, "grad_norm": 1.0320131696114871, "learning_rate": 7.631270916424923e-06, "loss": 0.7914, "step": 1782 }, { "epoch": 0.6745483779438192, "grad_norm": 1.0291951526102714, "learning_rate": 7.630531344398549e-06, "loss": 0.7273, "step": 1783 }, { "epoch": 0.6749267000851225, "grad_norm": 1.0352838518441736, "learning_rate": 7.62979106733389e-06, "loss": 0.8042, "step": 1784 }, { "epoch": 0.6753050222264259, "grad_norm": 0.999179215624018, "learning_rate": 7.629050085374709e-06, "loss": 0.8106, "step": 1785 }, { "epoch": 0.6756833443677291, "grad_norm": 1.002781374078623, "learning_rate": 7.6283083986649e-06, "loss": 0.7478, "step": 1786 }, { "epoch": 0.6760616665090324, "grad_norm": 1.0578987973117508, "learning_rate": 7.627566007348498e-06, "loss": 0.767, "step": 1787 }, { "epoch": 0.6764399886503357, "grad_norm": 1.018623825083434, "learning_rate": 7.626822911569673e-06, "loss": 0.7603, "step": 1788 }, { "epoch": 0.6768183107916391, "grad_norm": 1.0691359310227244, "learning_rate": 7.62607911147273e-06, "loss": 0.8033, "step": 1789 }, { "epoch": 0.6771966329329424, "grad_norm": 1.0473330500599638, "learning_rate": 7.625334607202115e-06, "loss": 0.799, "step": 1790 }, { "epoch": 0.6775749550742457, "grad_norm": 1.0276960283606948, "learning_rate": 7.624589398902408e-06, "loss": 0.7882, "step": 1791 }, { "epoch": 0.677953277215549, "grad_norm": 1.0216841452284737, "learning_rate": 7.623843486718325e-06, "loss": 0.7753, "step": 1792 }, { "epoch": 0.6783315993568524, "grad_norm": 1.017840190852707, "learning_rate": 7.623096870794722e-06, "loss": 0.7944, "step": 1793 }, { "epoch": 0.6787099214981557, "grad_norm": 1.0234534365543315, "learning_rate": 7.6223495512765865e-06, "loss": 0.7607, "step": 1794 }, { "epoch": 0.679088243639459, "grad_norm": 1.0142595858519063, "learning_rate": 7.621601528309049e-06, "loss": 0.7665, "step": 1795 }, { "epoch": 0.6794665657807624, "grad_norm": 1.0071219703193526, "learning_rate": 7.620852802037371e-06, "loss": 0.791, "step": 1796 }, { "epoch": 0.6798448879220657, "grad_norm": 1.0031377757032336, "learning_rate": 7.620103372606954e-06, "loss": 0.7502, "step": 1797 }, { "epoch": 0.6802232100633689, "grad_norm": 1.014284865797237, "learning_rate": 7.619353240163334e-06, "loss": 0.8012, "step": 1798 }, { "epoch": 0.6806015322046722, "grad_norm": 1.0281456730858456, "learning_rate": 7.618602404852186e-06, "loss": 0.8308, "step": 1799 }, { "epoch": 0.6809798543459756, "grad_norm": 1.0358974761664392, "learning_rate": 7.617850866819319e-06, "loss": 0.8116, "step": 1800 }, { "epoch": 0.6813581764872789, "grad_norm": 1.0233639481564207, "learning_rate": 7.61709862621068e-06, "loss": 0.8062, "step": 1801 }, { "epoch": 0.6817364986285822, "grad_norm": 0.9776086740367372, "learning_rate": 7.61634568317235e-06, "loss": 0.7926, "step": 1802 }, { "epoch": 0.6821148207698856, "grad_norm": 0.9900234559536079, "learning_rate": 7.61559203785055e-06, "loss": 0.8129, "step": 1803 }, { "epoch": 0.6824931429111889, "grad_norm": 1.040154643226836, "learning_rate": 7.614837690391636e-06, "loss": 0.8262, "step": 1804 }, { "epoch": 0.6828714650524922, "grad_norm": 1.0403643598930472, "learning_rate": 7.6140826409421e-06, "loss": 0.7831, "step": 1805 }, { "epoch": 0.6832497871937955, "grad_norm": 1.0496211590491318, "learning_rate": 7.613326889648568e-06, "loss": 0.7929, "step": 1806 }, { "epoch": 0.6836281093350989, "grad_norm": 1.0256924582611977, "learning_rate": 7.612570436657808e-06, "loss": 0.7801, "step": 1807 }, { "epoch": 0.6840064314764022, "grad_norm": 1.0346392707882297, "learning_rate": 7.611813282116718e-06, "loss": 0.7603, "step": 1808 }, { "epoch": 0.6843847536177055, "grad_norm": 1.058185361602599, "learning_rate": 7.611055426172336e-06, "loss": 0.8167, "step": 1809 }, { "epoch": 0.6847630757590087, "grad_norm": 1.0289714760483837, "learning_rate": 7.610296868971836e-06, "loss": 0.7822, "step": 1810 }, { "epoch": 0.6851413979003121, "grad_norm": 0.9890953669530531, "learning_rate": 7.609537610662528e-06, "loss": 0.7714, "step": 1811 }, { "epoch": 0.6855197200416154, "grad_norm": 1.01307277761867, "learning_rate": 7.608777651391857e-06, "loss": 0.7925, "step": 1812 }, { "epoch": 0.6858980421829187, "grad_norm": 1.0261501203430148, "learning_rate": 7.608016991307404e-06, "loss": 0.8062, "step": 1813 }, { "epoch": 0.6862763643242221, "grad_norm": 0.9935911309387935, "learning_rate": 7.607255630556888e-06, "loss": 0.774, "step": 1814 }, { "epoch": 0.6866546864655254, "grad_norm": 0.9904144924679894, "learning_rate": 7.606493569288161e-06, "loss": 0.7868, "step": 1815 }, { "epoch": 0.6870330086068287, "grad_norm": 1.0371418151427427, "learning_rate": 7.605730807649218e-06, "loss": 0.8109, "step": 1816 }, { "epoch": 0.687411330748132, "grad_norm": 1.044727797122997, "learning_rate": 7.604967345788178e-06, "loss": 0.8072, "step": 1817 }, { "epoch": 0.6877896528894354, "grad_norm": 0.9838280587194601, "learning_rate": 7.604203183853309e-06, "loss": 0.7536, "step": 1818 }, { "epoch": 0.6881679750307387, "grad_norm": 0.9950885054311281, "learning_rate": 7.603438321993005e-06, "loss": 0.7839, "step": 1819 }, { "epoch": 0.688546297172042, "grad_norm": 1.0428299690199871, "learning_rate": 7.602672760355803e-06, "loss": 0.7959, "step": 1820 }, { "epoch": 0.6889246193133454, "grad_norm": 1.0581206043121922, "learning_rate": 7.60190649909037e-06, "loss": 0.8037, "step": 1821 }, { "epoch": 0.6893029414546487, "grad_norm": 1.0096469456861266, "learning_rate": 7.601139538345513e-06, "loss": 0.8094, "step": 1822 }, { "epoch": 0.6896812635959519, "grad_norm": 1.0304785414820892, "learning_rate": 7.600371878270174e-06, "loss": 0.7653, "step": 1823 }, { "epoch": 0.6900595857372552, "grad_norm": 1.00229358608765, "learning_rate": 7.5996035190134295e-06, "loss": 0.7677, "step": 1824 }, { "epoch": 0.6904379078785586, "grad_norm": 0.9916315428613167, "learning_rate": 7.598834460724492e-06, "loss": 0.7942, "step": 1825 }, { "epoch": 0.6908162300198619, "grad_norm": 1.0389539705802777, "learning_rate": 7.5980647035527116e-06, "loss": 0.8177, "step": 1826 }, { "epoch": 0.6911945521611652, "grad_norm": 1.0662869985139787, "learning_rate": 7.597294247647571e-06, "loss": 0.768, "step": 1827 }, { "epoch": 0.6915728743024685, "grad_norm": 0.9714515658714595, "learning_rate": 7.596523093158693e-06, "loss": 0.7437, "step": 1828 }, { "epoch": 0.6919511964437719, "grad_norm": 1.0706404496033293, "learning_rate": 7.595751240235832e-06, "loss": 0.754, "step": 1829 }, { "epoch": 0.6923295185850752, "grad_norm": 0.9706657443652832, "learning_rate": 7.594978689028879e-06, "loss": 0.7612, "step": 1830 }, { "epoch": 0.6927078407263785, "grad_norm": 0.976615002955624, "learning_rate": 7.594205439687862e-06, "loss": 0.7912, "step": 1831 }, { "epoch": 0.6930861628676819, "grad_norm": 1.03426076744833, "learning_rate": 7.593431492362944e-06, "loss": 0.7917, "step": 1832 }, { "epoch": 0.6934644850089852, "grad_norm": 1.013505870254394, "learning_rate": 7.592656847204422e-06, "loss": 0.8032, "step": 1833 }, { "epoch": 0.6938428071502885, "grad_norm": 1.051455387215275, "learning_rate": 7.591881504362731e-06, "loss": 0.7912, "step": 1834 }, { "epoch": 0.6942211292915917, "grad_norm": 1.0263087231043657, "learning_rate": 7.591105463988439e-06, "loss": 0.8123, "step": 1835 }, { "epoch": 0.6945994514328951, "grad_norm": 1.0098333468819833, "learning_rate": 7.590328726232252e-06, "loss": 0.7629, "step": 1836 }, { "epoch": 0.6949777735741984, "grad_norm": 0.9721827671980219, "learning_rate": 7.589551291245009e-06, "loss": 0.7849, "step": 1837 }, { "epoch": 0.6953560957155017, "grad_norm": 1.0098176187649164, "learning_rate": 7.588773159177687e-06, "loss": 0.8025, "step": 1838 }, { "epoch": 0.6957344178568051, "grad_norm": 1.0299979129769126, "learning_rate": 7.587994330181395e-06, "loss": 0.7608, "step": 1839 }, { "epoch": 0.6961127399981084, "grad_norm": 1.0151953505875915, "learning_rate": 7.58721480440738e-06, "loss": 0.7337, "step": 1840 }, { "epoch": 0.6964910621394117, "grad_norm": 0.9904303117829623, "learning_rate": 7.5864345820070255e-06, "loss": 0.7747, "step": 1841 }, { "epoch": 0.696869384280715, "grad_norm": 1.0279175000834502, "learning_rate": 7.585653663131847e-06, "loss": 0.7918, "step": 1842 }, { "epoch": 0.6972477064220184, "grad_norm": 1.0295871862766321, "learning_rate": 7.584872047933494e-06, "loss": 0.7894, "step": 1843 }, { "epoch": 0.6976260285633217, "grad_norm": 1.0094453762458517, "learning_rate": 7.584089736563758e-06, "loss": 0.7206, "step": 1844 }, { "epoch": 0.698004350704625, "grad_norm": 1.0328426196699965, "learning_rate": 7.583306729174558e-06, "loss": 0.8252, "step": 1845 }, { "epoch": 0.6983826728459283, "grad_norm": 0.9905606234505574, "learning_rate": 7.582523025917954e-06, "loss": 0.7955, "step": 1846 }, { "epoch": 0.6987609949872317, "grad_norm": 1.0204271704285062, "learning_rate": 7.581738626946139e-06, "loss": 0.7896, "step": 1847 }, { "epoch": 0.6991393171285349, "grad_norm": 1.004700198708063, "learning_rate": 7.580953532411438e-06, "loss": 0.7748, "step": 1848 }, { "epoch": 0.6995176392698382, "grad_norm": 0.994279251316371, "learning_rate": 7.580167742466319e-06, "loss": 0.7584, "step": 1849 }, { "epoch": 0.6998959614111416, "grad_norm": 0.9865439763922379, "learning_rate": 7.579381257263375e-06, "loss": 0.7909, "step": 1850 }, { "epoch": 0.7002742835524449, "grad_norm": 0.9905670255963472, "learning_rate": 7.578594076955341e-06, "loss": 0.7919, "step": 1851 }, { "epoch": 0.7006526056937482, "grad_norm": 0.9976129555294091, "learning_rate": 7.577806201695086e-06, "loss": 0.7908, "step": 1852 }, { "epoch": 0.7010309278350515, "grad_norm": 0.9839141705023251, "learning_rate": 7.577017631635612e-06, "loss": 0.7895, "step": 1853 }, { "epoch": 0.7014092499763549, "grad_norm": 1.0306759419946148, "learning_rate": 7.576228366930057e-06, "loss": 0.8157, "step": 1854 }, { "epoch": 0.7017875721176582, "grad_norm": 0.9880194656345703, "learning_rate": 7.575438407731695e-06, "loss": 0.7366, "step": 1855 }, { "epoch": 0.7021658942589615, "grad_norm": 1.0081324236040694, "learning_rate": 7.57464775419393e-06, "loss": 0.784, "step": 1856 }, { "epoch": 0.7025442164002649, "grad_norm": 1.0280116129523802, "learning_rate": 7.573856406470311e-06, "loss": 0.7952, "step": 1857 }, { "epoch": 0.7029225385415682, "grad_norm": 1.0591816551488713, "learning_rate": 7.573064364714509e-06, "loss": 0.8319, "step": 1858 }, { "epoch": 0.7033008606828715, "grad_norm": 1.0373058408510851, "learning_rate": 7.57227162908034e-06, "loss": 0.8253, "step": 1859 }, { "epoch": 0.7036791828241747, "grad_norm": 0.9955132075131762, "learning_rate": 7.571478199721749e-06, "loss": 0.7884, "step": 1860 }, { "epoch": 0.7040575049654781, "grad_norm": 1.0411974533080333, "learning_rate": 7.570684076792817e-06, "loss": 0.7535, "step": 1861 }, { "epoch": 0.7044358271067814, "grad_norm": 1.0177208315160653, "learning_rate": 7.569889260447763e-06, "loss": 0.7789, "step": 1862 }, { "epoch": 0.7044358271067814, "eval_loss": 0.7890699505805969, "eval_runtime": 27.0832, "eval_samples_per_second": 32.677, "eval_steps_per_second": 1.034, "step": 1862 }, { "epoch": 0.7044358271067814, "eval_bench_accuracy_arc_challenge": 0.21428571428571427, "eval_bench_accuracy_hellaswag": 0.205, "eval_bench_accuracy_mmlu": 0.2608695652173913, "eval_bench_average_accuracy": 0.22671842650103516, "eval_bench_loss": 7.512841475637336, "eval_bench_total_accuracy": 0.22197802197802197, "step": 1862 }, { "epoch": 0.7048141492480847, "grad_norm": 1.028606903970918, "learning_rate": 7.5690937508409365e-06, "loss": 0.7457, "step": 1863 }, { "epoch": 0.705192471389388, "grad_norm": 1.0180150186079828, "learning_rate": 7.568297548126823e-06, "loss": 0.7938, "step": 1864 }, { "epoch": 0.7055707935306914, "grad_norm": 1.0876809046443543, "learning_rate": 7.567500652460042e-06, "loss": 0.8056, "step": 1865 }, { "epoch": 0.7059491156719947, "grad_norm": 1.029358737326028, "learning_rate": 7.56670306399535e-06, "loss": 0.7877, "step": 1866 }, { "epoch": 0.706327437813298, "grad_norm": 0.9844925839209022, "learning_rate": 7.565904782887634e-06, "loss": 0.7778, "step": 1867 }, { "epoch": 0.7067057599546014, "grad_norm": 0.9869706615525013, "learning_rate": 7.56510580929192e-06, "loss": 0.7339, "step": 1868 }, { "epoch": 0.7070840820959047, "grad_norm": 0.9992927633790497, "learning_rate": 7.564306143363364e-06, "loss": 0.7653, "step": 1869 }, { "epoch": 0.707462404237208, "grad_norm": 1.068436019657602, "learning_rate": 7.563505785257261e-06, "loss": 0.8271, "step": 1870 }, { "epoch": 0.7078407263785113, "grad_norm": 1.0179752248859633, "learning_rate": 7.5627047351290365e-06, "loss": 0.7727, "step": 1871 }, { "epoch": 0.7082190485198147, "grad_norm": 0.9956202748104152, "learning_rate": 7.561902993134254e-06, "loss": 0.7958, "step": 1872 }, { "epoch": 0.7085973706611179, "grad_norm": 1.0177361566307879, "learning_rate": 7.561100559428607e-06, "loss": 0.7779, "step": 1873 }, { "epoch": 0.7089756928024212, "grad_norm": 1.0189767447866593, "learning_rate": 7.560297434167926e-06, "loss": 0.7347, "step": 1874 }, { "epoch": 0.7093540149437246, "grad_norm": 0.9831880323651756, "learning_rate": 7.559493617508178e-06, "loss": 0.7652, "step": 1875 }, { "epoch": 0.7097323370850279, "grad_norm": 1.0200045931632868, "learning_rate": 7.5586891096054595e-06, "loss": 0.7824, "step": 1876 }, { "epoch": 0.7101106592263312, "grad_norm": 1.0558728606715007, "learning_rate": 7.557883910616004e-06, "loss": 0.7815, "step": 1877 }, { "epoch": 0.7104889813676345, "grad_norm": 1.0299315505060327, "learning_rate": 7.557078020696178e-06, "loss": 0.7576, "step": 1878 }, { "epoch": 0.7108673035089379, "grad_norm": 1.0001848779593938, "learning_rate": 7.556271440002485e-06, "loss": 0.7543, "step": 1879 }, { "epoch": 0.7112456256502412, "grad_norm": 1.043812227462618, "learning_rate": 7.555464168691559e-06, "loss": 0.7788, "step": 1880 }, { "epoch": 0.7116239477915445, "grad_norm": 1.0014451581659014, "learning_rate": 7.554656206920169e-06, "loss": 0.7978, "step": 1881 }, { "epoch": 0.7120022699328478, "grad_norm": 0.9973212648551616, "learning_rate": 7.55384755484522e-06, "loss": 0.7993, "step": 1882 }, { "epoch": 0.7123805920741512, "grad_norm": 1.061339436306225, "learning_rate": 7.5530382126237505e-06, "loss": 0.7972, "step": 1883 }, { "epoch": 0.7127589142154545, "grad_norm": 1.0699331120694706, "learning_rate": 7.55222818041293e-06, "loss": 0.8062, "step": 1884 }, { "epoch": 0.7131372363567577, "grad_norm": 0.9914442135257014, "learning_rate": 7.551417458370067e-06, "loss": 0.7791, "step": 1885 }, { "epoch": 0.7135155584980611, "grad_norm": 1.0130251115118372, "learning_rate": 7.5506060466525985e-06, "loss": 0.7875, "step": 1886 }, { "epoch": 0.7138938806393644, "grad_norm": 0.9695805264090013, "learning_rate": 7.5497939454181e-06, "loss": 0.7535, "step": 1887 }, { "epoch": 0.7142722027806677, "grad_norm": 1.0835656228512733, "learning_rate": 7.54898115482428e-06, "loss": 0.7466, "step": 1888 }, { "epoch": 0.714650524921971, "grad_norm": 1.0017379598251102, "learning_rate": 7.548167675028978e-06, "loss": 0.7588, "step": 1889 }, { "epoch": 0.7150288470632744, "grad_norm": 0.9751664976679292, "learning_rate": 7.5473535061901695e-06, "loss": 0.7555, "step": 1890 }, { "epoch": 0.7154071692045777, "grad_norm": 1.0011106387960145, "learning_rate": 7.546538648465965e-06, "loss": 0.8022, "step": 1891 }, { "epoch": 0.715785491345881, "grad_norm": 1.0166681934683515, "learning_rate": 7.545723102014606e-06, "loss": 0.7816, "step": 1892 }, { "epoch": 0.7161638134871844, "grad_norm": 0.9951953671889361, "learning_rate": 7.54490686699447e-06, "loss": 0.7837, "step": 1893 }, { "epoch": 0.7165421356284877, "grad_norm": 0.9976797977790014, "learning_rate": 7.544089943564067e-06, "loss": 0.7493, "step": 1894 }, { "epoch": 0.716920457769791, "grad_norm": 0.9855810780058022, "learning_rate": 7.543272331882042e-06, "loss": 0.7652, "step": 1895 }, { "epoch": 0.7172987799110943, "grad_norm": 1.020136564118587, "learning_rate": 7.542454032107171e-06, "loss": 0.7616, "step": 1896 }, { "epoch": 0.7176771020523977, "grad_norm": 1.0096765840235757, "learning_rate": 7.541635044398367e-06, "loss": 0.7981, "step": 1897 }, { "epoch": 0.7180554241937009, "grad_norm": 1.0416863254077597, "learning_rate": 7.540815368914675e-06, "loss": 0.7925, "step": 1898 }, { "epoch": 0.7184337463350042, "grad_norm": 1.055665744511811, "learning_rate": 7.539995005815272e-06, "loss": 0.7543, "step": 1899 }, { "epoch": 0.7188120684763076, "grad_norm": 1.0231952699867581, "learning_rate": 7.539173955259471e-06, "loss": 0.732, "step": 1900 }, { "epoch": 0.7191903906176109, "grad_norm": 1.0146131355524912, "learning_rate": 7.538352217406718e-06, "loss": 0.7587, "step": 1901 }, { "epoch": 0.7195687127589142, "grad_norm": 0.9865661154079507, "learning_rate": 7.53752979241659e-06, "loss": 0.7232, "step": 1902 }, { "epoch": 0.7199470349002175, "grad_norm": 1.0419811772033911, "learning_rate": 7.536706680448801e-06, "loss": 0.8073, "step": 1903 }, { "epoch": 0.7203253570415209, "grad_norm": 1.0083594152031234, "learning_rate": 7.535882881663199e-06, "loss": 0.7571, "step": 1904 }, { "epoch": 0.7207036791828242, "grad_norm": 1.0318684584286235, "learning_rate": 7.53505839621976e-06, "loss": 0.7831, "step": 1905 }, { "epoch": 0.7210820013241275, "grad_norm": 1.0401925339971567, "learning_rate": 7.534233224278598e-06, "loss": 0.7943, "step": 1906 }, { "epoch": 0.7214603234654308, "grad_norm": 1.0589062419109112, "learning_rate": 7.533407365999957e-06, "loss": 0.8116, "step": 1907 }, { "epoch": 0.7218386456067342, "grad_norm": 1.0189400220308538, "learning_rate": 7.532580821544218e-06, "loss": 0.7823, "step": 1908 }, { "epoch": 0.7222169677480375, "grad_norm": 1.0272792020263282, "learning_rate": 7.531753591071895e-06, "loss": 0.8002, "step": 1909 }, { "epoch": 0.7225952898893407, "grad_norm": 1.0050187200621885, "learning_rate": 7.530925674743631e-06, "loss": 0.7846, "step": 1910 }, { "epoch": 0.7229736120306441, "grad_norm": 1.0465639853681472, "learning_rate": 7.530097072720206e-06, "loss": 0.7764, "step": 1911 }, { "epoch": 0.7233519341719474, "grad_norm": 1.0069767691188105, "learning_rate": 7.529267785162531e-06, "loss": 0.7697, "step": 1912 }, { "epoch": 0.7237302563132507, "grad_norm": 1.025828867260226, "learning_rate": 7.528437812231653e-06, "loss": 0.762, "step": 1913 }, { "epoch": 0.724108578454554, "grad_norm": 1.0945775759380407, "learning_rate": 7.527607154088748e-06, "loss": 0.7684, "step": 1914 }, { "epoch": 0.7244869005958574, "grad_norm": 1.0614450729883849, "learning_rate": 7.526775810895129e-06, "loss": 0.7812, "step": 1915 }, { "epoch": 0.7248652227371607, "grad_norm": 1.01850920934482, "learning_rate": 7.525943782812239e-06, "loss": 0.7859, "step": 1916 }, { "epoch": 0.725243544878464, "grad_norm": 1.065129913789358, "learning_rate": 7.525111070001658e-06, "loss": 0.7982, "step": 1917 }, { "epoch": 0.7256218670197674, "grad_norm": 1.0194785224572207, "learning_rate": 7.524277672625093e-06, "loss": 0.7671, "step": 1918 }, { "epoch": 0.7260001891610707, "grad_norm": 1.030258817312861, "learning_rate": 7.52344359084439e-06, "loss": 0.8275, "step": 1919 }, { "epoch": 0.726378511302374, "grad_norm": 1.059266761958295, "learning_rate": 7.5226088248215224e-06, "loss": 0.7816, "step": 1920 }, { "epoch": 0.7267568334436773, "grad_norm": 1.0105478640349743, "learning_rate": 7.521773374718602e-06, "loss": 0.8033, "step": 1921 }, { "epoch": 0.7271351555849807, "grad_norm": 1.0463475130298905, "learning_rate": 7.52093724069787e-06, "loss": 0.7677, "step": 1922 }, { "epoch": 0.7275134777262839, "grad_norm": 1.0521557210010872, "learning_rate": 7.5201004229217e-06, "loss": 0.7524, "step": 1923 }, { "epoch": 0.7278917998675872, "grad_norm": 0.9739239170042839, "learning_rate": 7.519262921552601e-06, "loss": 0.773, "step": 1924 }, { "epoch": 0.7282701220088905, "grad_norm": 1.0603019289781683, "learning_rate": 7.51842473675321e-06, "loss": 0.7932, "step": 1925 }, { "epoch": 0.7286484441501939, "grad_norm": 1.028483949941829, "learning_rate": 7.517585868686305e-06, "loss": 0.7672, "step": 1926 }, { "epoch": 0.7290267662914972, "grad_norm": 0.9969415525100354, "learning_rate": 7.516746317514788e-06, "loss": 0.7703, "step": 1927 }, { "epoch": 0.7294050884328005, "grad_norm": 1.035817511317999, "learning_rate": 7.515906083401698e-06, "loss": 0.7737, "step": 1928 }, { "epoch": 0.7297834105741039, "grad_norm": 1.0160982960240883, "learning_rate": 7.515065166510206e-06, "loss": 0.7742, "step": 1929 }, { "epoch": 0.7301617327154072, "grad_norm": 1.0008758277167877, "learning_rate": 7.5142235670036164e-06, "loss": 0.8051, "step": 1930 }, { "epoch": 0.7305400548567105, "grad_norm": 1.009681305204692, "learning_rate": 7.513381285045365e-06, "loss": 0.751, "step": 1931 }, { "epoch": 0.7309183769980138, "grad_norm": 0.99959999867296, "learning_rate": 7.51253832079902e-06, "loss": 0.7618, "step": 1932 }, { "epoch": 0.7312966991393172, "grad_norm": 1.0184821171543585, "learning_rate": 7.511694674428282e-06, "loss": 0.7186, "step": 1933 }, { "epoch": 0.7316750212806205, "grad_norm": 1.023960557325352, "learning_rate": 7.510850346096987e-06, "loss": 0.7694, "step": 1934 }, { "epoch": 0.7320533434219237, "grad_norm": 1.0237615508066427, "learning_rate": 7.510005335969097e-06, "loss": 0.8001, "step": 1935 }, { "epoch": 0.7324316655632271, "grad_norm": 0.9818280466649454, "learning_rate": 7.509159644208714e-06, "loss": 0.7652, "step": 1936 }, { "epoch": 0.7328099877045304, "grad_norm": 1.0261934879214791, "learning_rate": 7.508313270980068e-06, "loss": 0.7822, "step": 1937 }, { "epoch": 0.7331883098458337, "grad_norm": 0.9617847765522927, "learning_rate": 7.50746621644752e-06, "loss": 0.752, "step": 1938 }, { "epoch": 0.733566631987137, "grad_norm": 1.003520826414359, "learning_rate": 7.506618480775568e-06, "loss": 0.7748, "step": 1939 }, { "epoch": 0.7339449541284404, "grad_norm": 1.0063475669200121, "learning_rate": 7.505770064128839e-06, "loss": 0.7951, "step": 1940 }, { "epoch": 0.7343232762697437, "grad_norm": 0.9752758302807805, "learning_rate": 7.5049209666720935e-06, "loss": 0.7644, "step": 1941 }, { "epoch": 0.734701598411047, "grad_norm": 0.9800795339764257, "learning_rate": 7.504071188570222e-06, "loss": 0.7628, "step": 1942 }, { "epoch": 0.7350799205523503, "grad_norm": 1.0479867078497374, "learning_rate": 7.50322072998825e-06, "loss": 0.8126, "step": 1943 }, { "epoch": 0.7354582426936537, "grad_norm": 1.0176104213424269, "learning_rate": 7.502369591091334e-06, "loss": 0.7787, "step": 1944 }, { "epoch": 0.735836564834957, "grad_norm": 0.987423177279228, "learning_rate": 7.501517772044762e-06, "loss": 0.781, "step": 1945 }, { "epoch": 0.7362148869762603, "grad_norm": 0.9770247411952715, "learning_rate": 7.500665273013954e-06, "loss": 0.7582, "step": 1946 }, { "epoch": 0.7365932091175637, "grad_norm": 1.0135337755615874, "learning_rate": 7.499812094164466e-06, "loss": 0.7715, "step": 1947 }, { "epoch": 0.7369715312588669, "grad_norm": 0.9649514740163105, "learning_rate": 7.498958235661979e-06, "loss": 0.7925, "step": 1948 }, { "epoch": 0.7373498534001702, "grad_norm": 0.98552867703349, "learning_rate": 7.4981036976723125e-06, "loss": 0.7722, "step": 1949 }, { "epoch": 0.7377281755414735, "grad_norm": 1.0267547692233978, "learning_rate": 7.497248480361414e-06, "loss": 0.8065, "step": 1950 }, { "epoch": 0.7381064976827769, "grad_norm": 1.0067029147076334, "learning_rate": 7.496392583895364e-06, "loss": 0.7494, "step": 1951 }, { "epoch": 0.7384848198240802, "grad_norm": 0.9372187525958833, "learning_rate": 7.495536008440376e-06, "loss": 0.7642, "step": 1952 }, { "epoch": 0.7388631419653835, "grad_norm": 1.0002942186824528, "learning_rate": 7.494678754162792e-06, "loss": 0.7457, "step": 1953 }, { "epoch": 0.7392414641066869, "grad_norm": 1.0052504063399132, "learning_rate": 7.493820821229091e-06, "loss": 0.7617, "step": 1954 }, { "epoch": 0.7396197862479902, "grad_norm": 1.0024323794208965, "learning_rate": 7.492962209805881e-06, "loss": 0.7465, "step": 1955 }, { "epoch": 0.7399981083892935, "grad_norm": 1.0067827748562472, "learning_rate": 7.492102920059899e-06, "loss": 0.7508, "step": 1956 }, { "epoch": 0.7403764305305968, "grad_norm": 0.9864844708039016, "learning_rate": 7.4912429521580186e-06, "loss": 0.7579, "step": 1957 }, { "epoch": 0.7407547526719002, "grad_norm": 1.040735707256207, "learning_rate": 7.490382306267243e-06, "loss": 0.8319, "step": 1958 }, { "epoch": 0.7411330748132035, "grad_norm": 0.9812310587310225, "learning_rate": 7.489520982554705e-06, "loss": 0.7983, "step": 1959 }, { "epoch": 0.7415113969545067, "grad_norm": 0.9981142486146017, "learning_rate": 7.488658981187674e-06, "loss": 0.7803, "step": 1960 }, { "epoch": 0.74188971909581, "grad_norm": 1.0369678771246449, "learning_rate": 7.4877963023335455e-06, "loss": 0.8067, "step": 1961 }, { "epoch": 0.7422680412371134, "grad_norm": 1.0070311838543111, "learning_rate": 7.486932946159852e-06, "loss": 0.7509, "step": 1962 }, { "epoch": 0.7426463633784167, "grad_norm": 1.0251496220350131, "learning_rate": 7.486068912834252e-06, "loss": 0.7541, "step": 1963 }, { "epoch": 0.74302468551972, "grad_norm": 1.009329085229992, "learning_rate": 7.485204202524539e-06, "loss": 0.7551, "step": 1964 }, { "epoch": 0.7434030076610234, "grad_norm": 1.0308386620704468, "learning_rate": 7.484338815398638e-06, "loss": 0.7837, "step": 1965 }, { "epoch": 0.7437813298023267, "grad_norm": 0.9828567433732203, "learning_rate": 7.483472751624603e-06, "loss": 0.7694, "step": 1966 }, { "epoch": 0.74415965194363, "grad_norm": 1.0075865335280576, "learning_rate": 7.4826060113706235e-06, "loss": 0.7624, "step": 1967 }, { "epoch": 0.7445379740849333, "grad_norm": 1.0271007536264374, "learning_rate": 7.481738594805015e-06, "loss": 0.7909, "step": 1968 }, { "epoch": 0.7449162962262367, "grad_norm": 0.958220580447762, "learning_rate": 7.480870502096229e-06, "loss": 0.7026, "step": 1969 }, { "epoch": 0.74529461836754, "grad_norm": 1.0319911779382374, "learning_rate": 7.480001733412845e-06, "loss": 0.7413, "step": 1970 }, { "epoch": 0.7456729405088433, "grad_norm": 0.9847198749453457, "learning_rate": 7.479132288923578e-06, "loss": 0.7656, "step": 1971 }, { "epoch": 0.7460512626501467, "grad_norm": 1.049323663151444, "learning_rate": 7.478262168797268e-06, "loss": 0.7937, "step": 1972 }, { "epoch": 0.7464295847914499, "grad_norm": 1.0325483940129563, "learning_rate": 7.477391373202892e-06, "loss": 0.7644, "step": 1973 }, { "epoch": 0.7468079069327532, "grad_norm": 1.0254999992955873, "learning_rate": 7.476519902309556e-06, "loss": 0.7744, "step": 1974 }, { "epoch": 0.7471862290740565, "grad_norm": 1.033661504020455, "learning_rate": 7.4756477562864946e-06, "loss": 0.7207, "step": 1975 }, { "epoch": 0.7475645512153599, "grad_norm": 0.9636241621995311, "learning_rate": 7.474774935303079e-06, "loss": 0.7721, "step": 1976 }, { "epoch": 0.7479428733566632, "grad_norm": 1.0189537631498706, "learning_rate": 7.473901439528807e-06, "loss": 0.808, "step": 1977 }, { "epoch": 0.7483211954979665, "grad_norm": 1.042367999759497, "learning_rate": 7.473027269133309e-06, "loss": 0.7827, "step": 1978 }, { "epoch": 0.7486995176392698, "grad_norm": 0.9999270083990087, "learning_rate": 7.472152424286347e-06, "loss": 0.7556, "step": 1979 }, { "epoch": 0.7490778397805732, "grad_norm": 0.9965813344103838, "learning_rate": 7.471276905157811e-06, "loss": 0.7642, "step": 1980 }, { "epoch": 0.7494561619218765, "grad_norm": 1.0578606892499327, "learning_rate": 7.470400711917726e-06, "loss": 0.7978, "step": 1981 }, { "epoch": 0.7498344840631798, "grad_norm": 1.0328463598563002, "learning_rate": 7.469523844736247e-06, "loss": 0.8026, "step": 1982 }, { "epoch": 0.7502128062044832, "grad_norm": 1.0198496774898518, "learning_rate": 7.468646303783656e-06, "loss": 0.8032, "step": 1983 }, { "epoch": 0.7505911283457865, "grad_norm": 1.0490904168720407, "learning_rate": 7.4677680892303714e-06, "loss": 0.7968, "step": 1984 }, { "epoch": 0.7509694504870897, "grad_norm": 1.0171290291066204, "learning_rate": 7.466889201246939e-06, "loss": 0.7675, "step": 1985 }, { "epoch": 0.751347772628393, "grad_norm": 1.0240538399264303, "learning_rate": 7.4660096400040365e-06, "loss": 0.8093, "step": 1986 }, { "epoch": 0.7517260947696964, "grad_norm": 1.0074355195502016, "learning_rate": 7.46512940567247e-06, "loss": 0.7546, "step": 1987 }, { "epoch": 0.7521044169109997, "grad_norm": 1.0381601440358865, "learning_rate": 7.464248498423183e-06, "loss": 0.792, "step": 1988 }, { "epoch": 0.752482739052303, "grad_norm": 0.9970266187532506, "learning_rate": 7.46336691842724e-06, "loss": 0.8471, "step": 1989 }, { "epoch": 0.7528610611936064, "grad_norm": 1.013601500628038, "learning_rate": 7.462484665855844e-06, "loss": 0.7304, "step": 1990 }, { "epoch": 0.7532393833349097, "grad_norm": 0.9998815765776821, "learning_rate": 7.4616017408803245e-06, "loss": 0.7493, "step": 1991 }, { "epoch": 0.753617705476213, "grad_norm": 0.9989366138847888, "learning_rate": 7.460718143672144e-06, "loss": 0.7753, "step": 1992 }, { "epoch": 0.7539960276175163, "grad_norm": 0.9848996402329897, "learning_rate": 7.459833874402895e-06, "loss": 0.7577, "step": 1993 }, { "epoch": 0.7543743497588197, "grad_norm": 1.0444974227287713, "learning_rate": 7.458948933244297e-06, "loss": 0.7452, "step": 1994 }, { "epoch": 0.754752671900123, "grad_norm": 0.9593608419190396, "learning_rate": 7.458063320368206e-06, "loss": 0.7433, "step": 1995 }, { "epoch": 0.754752671900123, "eval_loss": 0.7822305560112, "eval_runtime": 25.6112, "eval_samples_per_second": 34.555, "eval_steps_per_second": 1.093, "step": 1995 }, { "epoch": 0.754752671900123, "eval_bench_accuracy_arc_challenge": 0.2, "eval_bench_accuracy_hellaswag": 0.22, "eval_bench_accuracy_mmlu": 0.2608695652173913, "eval_bench_average_accuracy": 0.22695652173913042, "eval_bench_loss": 7.169555128666392, "eval_bench_total_accuracy": 0.22417582417582418, "step": 1995 }, { "epoch": 0.7551309940414263, "grad_norm": 1.0022855306714904, "learning_rate": 7.4571770359466035e-06, "loss": 0.8049, "step": 1996 }, { "epoch": 0.7555093161827295, "grad_norm": 1.0173023456897172, "learning_rate": 7.456290080151603e-06, "loss": 0.7863, "step": 1997 }, { "epoch": 0.7558876383240329, "grad_norm": 1.0126638410642659, "learning_rate": 7.455402453155452e-06, "loss": 0.8349, "step": 1998 }, { "epoch": 0.7562659604653362, "grad_norm": 1.0248458791368407, "learning_rate": 7.454514155130521e-06, "loss": 0.7771, "step": 1999 }, { "epoch": 0.7566442826066395, "grad_norm": 0.999304174568552, "learning_rate": 7.453625186249316e-06, "loss": 0.7706, "step": 2000 }, { "epoch": 0.7570226047479429, "grad_norm": 1.019954314390147, "learning_rate": 7.4527355466844736e-06, "loss": 0.7668, "step": 2001 }, { "epoch": 0.7574009268892462, "grad_norm": 0.9999396906025881, "learning_rate": 7.451845236608757e-06, "loss": 0.7533, "step": 2002 }, { "epoch": 0.7577792490305495, "grad_norm": 1.0169439537665028, "learning_rate": 7.450954256195064e-06, "loss": 0.7502, "step": 2003 }, { "epoch": 0.7581575711718528, "grad_norm": 0.9899425510634611, "learning_rate": 7.450062605616418e-06, "loss": 0.7925, "step": 2004 }, { "epoch": 0.7585358933131562, "grad_norm": 1.0122957620065443, "learning_rate": 7.4491702850459755e-06, "loss": 0.7688, "step": 2005 }, { "epoch": 0.7589142154544595, "grad_norm": 1.0068372233691978, "learning_rate": 7.4482772946570235e-06, "loss": 0.779, "step": 2006 }, { "epoch": 0.7592925375957628, "grad_norm": 1.07032907647764, "learning_rate": 7.447383634622978e-06, "loss": 0.7842, "step": 2007 }, { "epoch": 0.7596708597370662, "grad_norm": 1.0136781921965063, "learning_rate": 7.446489305117383e-06, "loss": 0.7471, "step": 2008 }, { "epoch": 0.7600491818783695, "grad_norm": 0.9990290235709554, "learning_rate": 7.445594306313918e-06, "loss": 0.7971, "step": 2009 }, { "epoch": 0.7604275040196727, "grad_norm": 0.9974204264253046, "learning_rate": 7.4446986383863855e-06, "loss": 0.7887, "step": 2010 }, { "epoch": 0.760805826160976, "grad_norm": 1.0297268325086462, "learning_rate": 7.443802301508725e-06, "loss": 0.7889, "step": 2011 }, { "epoch": 0.7611841483022794, "grad_norm": 1.0466207299533286, "learning_rate": 7.442905295854999e-06, "loss": 0.8049, "step": 2012 }, { "epoch": 0.7615624704435827, "grad_norm": 0.9835574884450916, "learning_rate": 7.442007621599407e-06, "loss": 0.7712, "step": 2013 }, { "epoch": 0.761940792584886, "grad_norm": 1.0395832668769522, "learning_rate": 7.44110927891627e-06, "loss": 0.7918, "step": 2014 }, { "epoch": 0.7623191147261893, "grad_norm": 1.0327351525083417, "learning_rate": 7.440210267980048e-06, "loss": 0.8035, "step": 2015 }, { "epoch": 0.7626974368674927, "grad_norm": 1.0098765426066463, "learning_rate": 7.4393105889653244e-06, "loss": 0.775, "step": 2016 }, { "epoch": 0.763075759008796, "grad_norm": 0.9906376521733136, "learning_rate": 7.438410242046813e-06, "loss": 0.7692, "step": 2017 }, { "epoch": 0.7634540811500993, "grad_norm": 1.0188390810770034, "learning_rate": 7.43750922739936e-06, "loss": 0.7768, "step": 2018 }, { "epoch": 0.7638324032914027, "grad_norm": 1.004186081889002, "learning_rate": 7.436607545197939e-06, "loss": 0.7865, "step": 2019 }, { "epoch": 0.764210725432706, "grad_norm": 1.0358970241783556, "learning_rate": 7.435705195617655e-06, "loss": 0.7778, "step": 2020 }, { "epoch": 0.7645890475740093, "grad_norm": 1.0401669544728585, "learning_rate": 7.43480217883374e-06, "loss": 0.759, "step": 2021 }, { "epoch": 0.7649673697153125, "grad_norm": 0.9824157827172492, "learning_rate": 7.433898495021558e-06, "loss": 0.7786, "step": 2022 }, { "epoch": 0.7653456918566159, "grad_norm": 0.9943895923909359, "learning_rate": 7.4329941443566015e-06, "loss": 0.7656, "step": 2023 }, { "epoch": 0.7657240139979192, "grad_norm": 1.0016606175405596, "learning_rate": 7.432089127014494e-06, "loss": 0.7614, "step": 2024 }, { "epoch": 0.7661023361392225, "grad_norm": 0.9865397528957869, "learning_rate": 7.431183443170985e-06, "loss": 0.7622, "step": 2025 }, { "epoch": 0.7664806582805259, "grad_norm": 0.989160673615647, "learning_rate": 7.430277093001956e-06, "loss": 0.7731, "step": 2026 }, { "epoch": 0.7668589804218292, "grad_norm": 0.9915228634736527, "learning_rate": 7.429370076683419e-06, "loss": 0.7431, "step": 2027 }, { "epoch": 0.7672373025631325, "grad_norm": 1.006928288896759, "learning_rate": 7.428462394391513e-06, "loss": 0.774, "step": 2028 }, { "epoch": 0.7676156247044358, "grad_norm": 1.0198721058181943, "learning_rate": 7.427554046302507e-06, "loss": 0.7973, "step": 2029 }, { "epoch": 0.7679939468457392, "grad_norm": 1.0271414827896679, "learning_rate": 7.426645032592798e-06, "loss": 0.7784, "step": 2030 }, { "epoch": 0.7683722689870425, "grad_norm": 0.9817303579553227, "learning_rate": 7.425735353438917e-06, "loss": 0.7843, "step": 2031 }, { "epoch": 0.7687505911283458, "grad_norm": 1.0056122125969114, "learning_rate": 7.424825009017519e-06, "loss": 0.7438, "step": 2032 }, { "epoch": 0.7691289132696492, "grad_norm": 1.008163384788162, "learning_rate": 7.42391399950539e-06, "loss": 0.7757, "step": 2033 }, { "epoch": 0.7695072354109525, "grad_norm": 1.0329565289060225, "learning_rate": 7.423002325079446e-06, "loss": 0.7577, "step": 2034 }, { "epoch": 0.7698855575522557, "grad_norm": 1.032231211492487, "learning_rate": 7.422089985916731e-06, "loss": 0.8083, "step": 2035 }, { "epoch": 0.770263879693559, "grad_norm": 0.9900284260015346, "learning_rate": 7.4211769821944185e-06, "loss": 0.7658, "step": 2036 }, { "epoch": 0.7706422018348624, "grad_norm": 1.0170156545119755, "learning_rate": 7.420263314089811e-06, "loss": 0.8177, "step": 2037 }, { "epoch": 0.7710205239761657, "grad_norm": 0.9781641011100699, "learning_rate": 7.419348981780341e-06, "loss": 0.7339, "step": 2038 }, { "epoch": 0.771398846117469, "grad_norm": 1.0105682612914932, "learning_rate": 7.418433985443567e-06, "loss": 0.7891, "step": 2039 }, { "epoch": 0.7717771682587723, "grad_norm": 1.008948813484373, "learning_rate": 7.417518325257182e-06, "loss": 0.8386, "step": 2040 }, { "epoch": 0.7721554904000757, "grad_norm": 1.0060473452659786, "learning_rate": 7.416602001399001e-06, "loss": 0.7597, "step": 2041 }, { "epoch": 0.772533812541379, "grad_norm": 1.044835723953059, "learning_rate": 7.415685014046973e-06, "loss": 0.81, "step": 2042 }, { "epoch": 0.7729121346826823, "grad_norm": 1.0214227811554681, "learning_rate": 7.4147673633791735e-06, "loss": 0.7697, "step": 2043 }, { "epoch": 0.7732904568239857, "grad_norm": 1.0023520288888492, "learning_rate": 7.4138490495738085e-06, "loss": 0.7406, "step": 2044 }, { "epoch": 0.773668778965289, "grad_norm": 0.9916370657279895, "learning_rate": 7.412930072809212e-06, "loss": 0.7852, "step": 2045 }, { "epoch": 0.7740471011065923, "grad_norm": 0.9951977955788776, "learning_rate": 7.412010433263844e-06, "loss": 0.7775, "step": 2046 }, { "epoch": 0.7744254232478955, "grad_norm": 0.9994091529470467, "learning_rate": 7.411090131116299e-06, "loss": 0.7954, "step": 2047 }, { "epoch": 0.7748037453891989, "grad_norm": 0.9793893415326917, "learning_rate": 7.410169166545295e-06, "loss": 0.7927, "step": 2048 }, { "epoch": 0.7751820675305022, "grad_norm": 1.0284758106957166, "learning_rate": 7.4092475397296815e-06, "loss": 0.8008, "step": 2049 }, { "epoch": 0.7755603896718055, "grad_norm": 1.0141147851186638, "learning_rate": 7.4083252508484346e-06, "loss": 0.772, "step": 2050 }, { "epoch": 0.7759387118131089, "grad_norm": 1.0301282378593513, "learning_rate": 7.4074023000806594e-06, "loss": 0.7919, "step": 2051 }, { "epoch": 0.7763170339544122, "grad_norm": 0.99410650693796, "learning_rate": 7.4064786876055934e-06, "loss": 0.7731, "step": 2052 }, { "epoch": 0.7766953560957155, "grad_norm": 0.9789727563437306, "learning_rate": 7.405554413602596e-06, "loss": 0.7796, "step": 2053 }, { "epoch": 0.7770736782370188, "grad_norm": 0.9735007692717235, "learning_rate": 7.404629478251161e-06, "loss": 0.7818, "step": 2054 }, { "epoch": 0.7774520003783222, "grad_norm": 1.0192472684096552, "learning_rate": 7.403703881730905e-06, "loss": 0.7495, "step": 2055 }, { "epoch": 0.7778303225196255, "grad_norm": 1.0193180834460873, "learning_rate": 7.402777624221579e-06, "loss": 0.7555, "step": 2056 }, { "epoch": 0.7782086446609288, "grad_norm": 1.013296156079769, "learning_rate": 7.401850705903058e-06, "loss": 0.7637, "step": 2057 }, { "epoch": 0.778586966802232, "grad_norm": 1.0204115231026534, "learning_rate": 7.400923126955347e-06, "loss": 0.7541, "step": 2058 }, { "epoch": 0.7789652889435354, "grad_norm": 0.9996050323401531, "learning_rate": 7.3999948875585785e-06, "loss": 0.787, "step": 2059 }, { "epoch": 0.7793436110848387, "grad_norm": 1.0291924784446762, "learning_rate": 7.399065987893015e-06, "loss": 0.7387, "step": 2060 }, { "epoch": 0.779721933226142, "grad_norm": 0.9494949051999276, "learning_rate": 7.398136428139046e-06, "loss": 0.7774, "step": 2061 }, { "epoch": 0.7801002553674454, "grad_norm": 0.9916891479050064, "learning_rate": 7.397206208477188e-06, "loss": 0.7781, "step": 2062 }, { "epoch": 0.7804785775087487, "grad_norm": 1.0104703821927092, "learning_rate": 7.396275329088088e-06, "loss": 0.7484, "step": 2063 }, { "epoch": 0.780856899650052, "grad_norm": 1.0241907972818591, "learning_rate": 7.395343790152518e-06, "loss": 0.753, "step": 2064 }, { "epoch": 0.7812352217913553, "grad_norm": 0.9817159276301853, "learning_rate": 7.394411591851383e-06, "loss": 0.7595, "step": 2065 }, { "epoch": 0.7816135439326587, "grad_norm": 0.9742244579787934, "learning_rate": 7.393478734365711e-06, "loss": 0.7678, "step": 2066 }, { "epoch": 0.781991866073962, "grad_norm": 1.0017399083453065, "learning_rate": 7.392545217876661e-06, "loss": 0.7632, "step": 2067 }, { "epoch": 0.7823701882152653, "grad_norm": 0.971974697646123, "learning_rate": 7.3916110425655196e-06, "loss": 0.7721, "step": 2068 }, { "epoch": 0.7827485103565687, "grad_norm": 0.9945216337737849, "learning_rate": 7.390676208613699e-06, "loss": 0.7591, "step": 2069 }, { "epoch": 0.783126832497872, "grad_norm": 1.0317158232325052, "learning_rate": 7.389740716202743e-06, "loss": 0.7744, "step": 2070 }, { "epoch": 0.7835051546391752, "grad_norm": 0.9861408269181425, "learning_rate": 7.38880456551432e-06, "loss": 0.7749, "step": 2071 }, { "epoch": 0.7838834767804785, "grad_norm": 1.002529687480502, "learning_rate": 7.387867756730228e-06, "loss": 0.8016, "step": 2072 }, { "epoch": 0.7842617989217819, "grad_norm": 0.9900168216741768, "learning_rate": 7.386930290032394e-06, "loss": 0.8002, "step": 2073 }, { "epoch": 0.7846401210630852, "grad_norm": 1.001127685901358, "learning_rate": 7.385992165602869e-06, "loss": 0.8009, "step": 2074 }, { "epoch": 0.7850184432043885, "grad_norm": 0.9617676340929628, "learning_rate": 7.385053383623835e-06, "loss": 0.7617, "step": 2075 }, { "epoch": 0.7853967653456918, "grad_norm": 1.0269257302626273, "learning_rate": 7.3841139442776006e-06, "loss": 0.7546, "step": 2076 }, { "epoch": 0.7857750874869952, "grad_norm": 1.044935488503257, "learning_rate": 7.383173847746602e-06, "loss": 0.8083, "step": 2077 }, { "epoch": 0.7861534096282985, "grad_norm": 1.038961339513026, "learning_rate": 7.382233094213404e-06, "loss": 0.7768, "step": 2078 }, { "epoch": 0.7865317317696018, "grad_norm": 0.9793045756612131, "learning_rate": 7.381291683860697e-06, "loss": 0.7517, "step": 2079 }, { "epoch": 0.7869100539109052, "grad_norm": 1.0764860961773461, "learning_rate": 7.3803496168713e-06, "loss": 0.8043, "step": 2080 }, { "epoch": 0.7872883760522085, "grad_norm": 1.0597785400343005, "learning_rate": 7.379406893428161e-06, "loss": 0.7798, "step": 2081 }, { "epoch": 0.7876666981935118, "grad_norm": 1.0419863282832178, "learning_rate": 7.378463513714352e-06, "loss": 0.7553, "step": 2082 }, { "epoch": 0.788045020334815, "grad_norm": 1.027454957481235, "learning_rate": 7.377519477913076e-06, "loss": 0.758, "step": 2083 }, { "epoch": 0.7884233424761184, "grad_norm": 1.011078690570445, "learning_rate": 7.37657478620766e-06, "loss": 0.7769, "step": 2084 }, { "epoch": 0.7888016646174217, "grad_norm": 0.9885065657321271, "learning_rate": 7.375629438781564e-06, "loss": 0.766, "step": 2085 }, { "epoch": 0.789179986758725, "grad_norm": 0.9782589251822759, "learning_rate": 7.374683435818367e-06, "loss": 0.7623, "step": 2086 }, { "epoch": 0.7895583089000284, "grad_norm": 0.9983247516328617, "learning_rate": 7.373736777501784e-06, "loss": 0.7632, "step": 2087 }, { "epoch": 0.7899366310413317, "grad_norm": 1.02973823604015, "learning_rate": 7.372789464015651e-06, "loss": 0.7813, "step": 2088 }, { "epoch": 0.790314953182635, "grad_norm": 1.0373760885627215, "learning_rate": 7.371841495543935e-06, "loss": 0.7672, "step": 2089 }, { "epoch": 0.7906932753239383, "grad_norm": 1.0077909409610364, "learning_rate": 7.370892872270726e-06, "loss": 0.7886, "step": 2090 }, { "epoch": 0.7910715974652417, "grad_norm": 0.9614715248979373, "learning_rate": 7.369943594380245e-06, "loss": 0.7615, "step": 2091 }, { "epoch": 0.791449919606545, "grad_norm": 1.0543567606170345, "learning_rate": 7.36899366205684e-06, "loss": 0.774, "step": 2092 }, { "epoch": 0.7918282417478483, "grad_norm": 1.0545293176252877, "learning_rate": 7.368043075484985e-06, "loss": 0.814, "step": 2093 }, { "epoch": 0.7922065638891516, "grad_norm": 1.0410560127064061, "learning_rate": 7.367091834849279e-06, "loss": 0.7662, "step": 2094 }, { "epoch": 0.792584886030455, "grad_norm": 1.0369567485364843, "learning_rate": 7.366139940334452e-06, "loss": 0.7745, "step": 2095 }, { "epoch": 0.7929632081717582, "grad_norm": 1.001965000397155, "learning_rate": 7.36518739212536e-06, "loss": 0.7521, "step": 2096 }, { "epoch": 0.7933415303130615, "grad_norm": 1.0716119168512355, "learning_rate": 7.364234190406982e-06, "loss": 0.7998, "step": 2097 }, { "epoch": 0.7937198524543649, "grad_norm": 0.9981024417773467, "learning_rate": 7.363280335364428e-06, "loss": 0.8172, "step": 2098 }, { "epoch": 0.7940981745956682, "grad_norm": 0.9780788155909068, "learning_rate": 7.362325827182934e-06, "loss": 0.7746, "step": 2099 }, { "epoch": 0.7944764967369715, "grad_norm": 0.9956472621886054, "learning_rate": 7.361370666047864e-06, "loss": 0.766, "step": 2100 }, { "epoch": 0.7948548188782748, "grad_norm": 1.0398013021998502, "learning_rate": 7.360414852144705e-06, "loss": 0.7726, "step": 2101 }, { "epoch": 0.7952331410195782, "grad_norm": 1.004170385720903, "learning_rate": 7.359458385659076e-06, "loss": 0.7704, "step": 2102 }, { "epoch": 0.7956114631608815, "grad_norm": 1.0441183212710015, "learning_rate": 7.358501266776717e-06, "loss": 0.7899, "step": 2103 }, { "epoch": 0.7959897853021848, "grad_norm": 1.00292777553299, "learning_rate": 7.357543495683499e-06, "loss": 0.7531, "step": 2104 }, { "epoch": 0.7963681074434882, "grad_norm": 0.9854050349889603, "learning_rate": 7.356585072565418e-06, "loss": 0.708, "step": 2105 }, { "epoch": 0.7967464295847915, "grad_norm": 0.9941467257050194, "learning_rate": 7.355625997608598e-06, "loss": 0.7413, "step": 2106 }, { "epoch": 0.7971247517260948, "grad_norm": 1.0428097298246077, "learning_rate": 7.354666270999287e-06, "loss": 0.8238, "step": 2107 }, { "epoch": 0.797503073867398, "grad_norm": 1.0286793733390376, "learning_rate": 7.3537058929238616e-06, "loss": 0.8368, "step": 2108 }, { "epoch": 0.7978813960087014, "grad_norm": 0.9891004290656986, "learning_rate": 7.352744863568825e-06, "loss": 0.7488, "step": 2109 }, { "epoch": 0.7982597181500047, "grad_norm": 1.0692597075643575, "learning_rate": 7.351783183120805e-06, "loss": 0.7634, "step": 2110 }, { "epoch": 0.798638040291308, "grad_norm": 1.0569774308781557, "learning_rate": 7.350820851766556e-06, "loss": 0.8056, "step": 2111 }, { "epoch": 0.7990163624326113, "grad_norm": 1.0044318866752053, "learning_rate": 7.349857869692964e-06, "loss": 0.7794, "step": 2112 }, { "epoch": 0.7993946845739147, "grad_norm": 1.0259526053127963, "learning_rate": 7.348894237087033e-06, "loss": 0.7229, "step": 2113 }, { "epoch": 0.799773006715218, "grad_norm": 1.0013521979151376, "learning_rate": 7.347929954135899e-06, "loss": 0.7403, "step": 2114 }, { "epoch": 0.8001513288565213, "grad_norm": 1.0479908669312097, "learning_rate": 7.346965021026824e-06, "loss": 0.8059, "step": 2115 }, { "epoch": 0.8005296509978247, "grad_norm": 0.988460255277449, "learning_rate": 7.345999437947195e-06, "loss": 0.7691, "step": 2116 }, { "epoch": 0.800907973139128, "grad_norm": 1.023874402122828, "learning_rate": 7.345033205084523e-06, "loss": 0.7792, "step": 2117 }, { "epoch": 0.8012862952804313, "grad_norm": 0.9784616959666119, "learning_rate": 7.34406632262645e-06, "loss": 0.7597, "step": 2118 }, { "epoch": 0.8016646174217346, "grad_norm": 0.9773736918754649, "learning_rate": 7.343098790760741e-06, "loss": 0.7679, "step": 2119 }, { "epoch": 0.802042939563038, "grad_norm": 1.0741761264395726, "learning_rate": 7.342130609675286e-06, "loss": 0.8127, "step": 2120 }, { "epoch": 0.8024212617043412, "grad_norm": 1.0374383515762358, "learning_rate": 7.341161779558105e-06, "loss": 0.8014, "step": 2121 }, { "epoch": 0.8027995838456445, "grad_norm": 0.987349661347248, "learning_rate": 7.340192300597342e-06, "loss": 0.7502, "step": 2122 }, { "epoch": 0.8031779059869479, "grad_norm": 1.0037779958682087, "learning_rate": 7.339222172981266e-06, "loss": 0.8052, "step": 2123 }, { "epoch": 0.8035562281282512, "grad_norm": 1.0113297595316113, "learning_rate": 7.338251396898272e-06, "loss": 0.7407, "step": 2124 }, { "epoch": 0.8039345502695545, "grad_norm": 1.0000848733537029, "learning_rate": 7.337279972536883e-06, "loss": 0.7888, "step": 2125 }, { "epoch": 0.8043128724108578, "grad_norm": 1.0033318854689852, "learning_rate": 7.3363079000857475e-06, "loss": 0.7625, "step": 2126 }, { "epoch": 0.8046911945521612, "grad_norm": 0.988981929427202, "learning_rate": 7.335335179733638e-06, "loss": 0.7535, "step": 2127 }, { "epoch": 0.8050695166934645, "grad_norm": 1.0222744589441084, "learning_rate": 7.334361811669454e-06, "loss": 0.8012, "step": 2128 }, { "epoch": 0.8050695166934645, "eval_loss": 0.7778414487838745, "eval_runtime": 25.4758, "eval_samples_per_second": 34.739, "eval_steps_per_second": 1.099, "step": 2128 }, { "epoch": 0.8050695166934645, "eval_bench_accuracy_arc_challenge": 0.0, "eval_bench_accuracy_hellaswag": 0.21, "eval_bench_accuracy_mmlu": 0.2782608695652174, "eval_bench_average_accuracy": 0.1627536231884058, "eval_bench_loss": 7.308511298999452, "eval_bench_total_accuracy": 0.16263736263736264, "step": 2128 }, { "epoch": 0.8054478388347678, "grad_norm": 1.0250776720341548, "learning_rate": 7.33338779608222e-06, "loss": 0.7883, "step": 2129 }, { "epoch": 0.8058261609760711, "grad_norm": 1.0203503130278309, "learning_rate": 7.332413133161088e-06, "loss": 0.7503, "step": 2130 }, { "epoch": 0.8062044831173745, "grad_norm": 1.0054642073739408, "learning_rate": 7.331437823095333e-06, "loss": 0.7955, "step": 2131 }, { "epoch": 0.8065828052586778, "grad_norm": 0.9959030950007364, "learning_rate": 7.33046186607436e-06, "loss": 0.7767, "step": 2132 }, { "epoch": 0.806961127399981, "grad_norm": 1.025147857248923, "learning_rate": 7.329485262287695e-06, "loss": 0.7966, "step": 2133 }, { "epoch": 0.8073394495412844, "grad_norm": 1.0187310556975737, "learning_rate": 7.328508011924991e-06, "loss": 0.7702, "step": 2134 }, { "epoch": 0.8077177716825877, "grad_norm": 0.9632571683519583, "learning_rate": 7.327530115176028e-06, "loss": 0.7435, "step": 2135 }, { "epoch": 0.808096093823891, "grad_norm": 1.01283490823059, "learning_rate": 7.326551572230711e-06, "loss": 0.7586, "step": 2136 }, { "epoch": 0.8084744159651943, "grad_norm": 0.9885051792437358, "learning_rate": 7.3255723832790695e-06, "loss": 0.7906, "step": 2137 }, { "epoch": 0.8088527381064977, "grad_norm": 1.0020516467578051, "learning_rate": 7.32459254851126e-06, "loss": 0.8075, "step": 2138 }, { "epoch": 0.809231060247801, "grad_norm": 1.0310575396515522, "learning_rate": 7.323612068117562e-06, "loss": 0.7599, "step": 2139 }, { "epoch": 0.8096093823891043, "grad_norm": 1.0294050765759222, "learning_rate": 7.322630942288382e-06, "loss": 0.8039, "step": 2140 }, { "epoch": 0.8099877045304077, "grad_norm": 1.0283423256981907, "learning_rate": 7.321649171214252e-06, "loss": 0.7693, "step": 2141 }, { "epoch": 0.810366026671711, "grad_norm": 1.008896097498659, "learning_rate": 7.320666755085831e-06, "loss": 0.7625, "step": 2142 }, { "epoch": 0.8107443488130143, "grad_norm": 0.9603118568373574, "learning_rate": 7.319683694093898e-06, "loss": 0.7718, "step": 2143 }, { "epoch": 0.8111226709543176, "grad_norm": 0.970477910624227, "learning_rate": 7.318699988429361e-06, "loss": 0.7044, "step": 2144 }, { "epoch": 0.811500993095621, "grad_norm": 0.9806308104560464, "learning_rate": 7.317715638283256e-06, "loss": 0.7908, "step": 2145 }, { "epoch": 0.8118793152369242, "grad_norm": 0.9867131475291727, "learning_rate": 7.316730643846737e-06, "loss": 0.7641, "step": 2146 }, { "epoch": 0.8122576373782275, "grad_norm": 1.0063847944117592, "learning_rate": 7.315745005311089e-06, "loss": 0.7885, "step": 2147 }, { "epoch": 0.8126359595195308, "grad_norm": 0.9925485105760502, "learning_rate": 7.314758722867718e-06, "loss": 0.7856, "step": 2148 }, { "epoch": 0.8130142816608342, "grad_norm": 1.029601921057074, "learning_rate": 7.313771796708161e-06, "loss": 0.7466, "step": 2149 }, { "epoch": 0.8133926038021375, "grad_norm": 0.98773252043624, "learning_rate": 7.312784227024073e-06, "loss": 0.7761, "step": 2150 }, { "epoch": 0.8137709259434408, "grad_norm": 1.0289427441863017, "learning_rate": 7.311796014007237e-06, "loss": 0.7589, "step": 2151 }, { "epoch": 0.8141492480847442, "grad_norm": 1.0115981501772968, "learning_rate": 7.310807157849562e-06, "loss": 0.7252, "step": 2152 }, { "epoch": 0.8145275702260475, "grad_norm": 1.0128188745545885, "learning_rate": 7.30981765874308e-06, "loss": 0.7512, "step": 2153 }, { "epoch": 0.8149058923673508, "grad_norm": 1.0162383280113574, "learning_rate": 7.308827516879951e-06, "loss": 0.7966, "step": 2154 }, { "epoch": 0.8152842145086541, "grad_norm": 0.9925052349677926, "learning_rate": 7.307836732452454e-06, "loss": 0.7457, "step": 2155 }, { "epoch": 0.8156625366499575, "grad_norm": 1.0482788371593144, "learning_rate": 7.306845305652999e-06, "loss": 0.7974, "step": 2156 }, { "epoch": 0.8160408587912608, "grad_norm": 1.0283411397442508, "learning_rate": 7.305853236674118e-06, "loss": 0.8108, "step": 2157 }, { "epoch": 0.816419180932564, "grad_norm": 1.0882228349020795, "learning_rate": 7.304860525708467e-06, "loss": 0.7715, "step": 2158 }, { "epoch": 0.8167975030738674, "grad_norm": 1.0579286859912629, "learning_rate": 7.303867172948828e-06, "loss": 0.71, "step": 2159 }, { "epoch": 0.8171758252151707, "grad_norm": 0.9985249082658622, "learning_rate": 7.302873178588106e-06, "loss": 0.7745, "step": 2160 }, { "epoch": 0.817554147356474, "grad_norm": 0.993440700317102, "learning_rate": 7.301878542819333e-06, "loss": 0.7691, "step": 2161 }, { "epoch": 0.8179324694977773, "grad_norm": 0.9913179768021124, "learning_rate": 7.300883265835665e-06, "loss": 0.7705, "step": 2162 }, { "epoch": 0.8183107916390807, "grad_norm": 0.9770635892112639, "learning_rate": 7.2998873478303796e-06, "loss": 0.767, "step": 2163 }, { "epoch": 0.818689113780384, "grad_norm": 1.0059497338070138, "learning_rate": 7.298890788996882e-06, "loss": 0.7965, "step": 2164 }, { "epoch": 0.8190674359216873, "grad_norm": 0.9884431887016643, "learning_rate": 7.297893589528701e-06, "loss": 0.7757, "step": 2165 }, { "epoch": 0.8194457580629906, "grad_norm": 1.0030750546593847, "learning_rate": 7.29689574961949e-06, "loss": 0.7767, "step": 2166 }, { "epoch": 0.819824080204294, "grad_norm": 1.0068282866037144, "learning_rate": 7.2958972694630255e-06, "loss": 0.7163, "step": 2167 }, { "epoch": 0.8202024023455973, "grad_norm": 1.270800973142674, "learning_rate": 7.294898149253212e-06, "loss": 0.81, "step": 2168 }, { "epoch": 0.8205807244869006, "grad_norm": 0.9920148831344642, "learning_rate": 7.2938983891840735e-06, "loss": 0.7473, "step": 2169 }, { "epoch": 0.820959046628204, "grad_norm": 1.0046876632532424, "learning_rate": 7.292897989449759e-06, "loss": 0.7939, "step": 2170 }, { "epoch": 0.8213373687695072, "grad_norm": 0.9825002163170027, "learning_rate": 7.2918969502445475e-06, "loss": 0.7218, "step": 2171 }, { "epoch": 0.8217156909108105, "grad_norm": 1.006900614246493, "learning_rate": 7.290895271762833e-06, "loss": 0.757, "step": 2172 }, { "epoch": 0.8220940130521138, "grad_norm": 1.0110578914640533, "learning_rate": 7.289892954199141e-06, "loss": 0.7536, "step": 2173 }, { "epoch": 0.8224723351934172, "grad_norm": 0.9996959192531956, "learning_rate": 7.288889997748118e-06, "loss": 0.7667, "step": 2174 }, { "epoch": 0.8228506573347205, "grad_norm": 1.037358589589491, "learning_rate": 7.2878864026045365e-06, "loss": 0.7883, "step": 2175 }, { "epoch": 0.8232289794760238, "grad_norm": 1.0492600455435164, "learning_rate": 7.286882168963289e-06, "loss": 0.7286, "step": 2176 }, { "epoch": 0.8236073016173272, "grad_norm": 1.0154884370457191, "learning_rate": 7.285877297019396e-06, "loss": 0.7718, "step": 2177 }, { "epoch": 0.8239856237586305, "grad_norm": 0.9822455344502207, "learning_rate": 7.284871786968002e-06, "loss": 0.7319, "step": 2178 }, { "epoch": 0.8243639458999338, "grad_norm": 0.9947974308410656, "learning_rate": 7.2838656390043735e-06, "loss": 0.7902, "step": 2179 }, { "epoch": 0.8247422680412371, "grad_norm": 0.9989701876963061, "learning_rate": 7.2828588533238985e-06, "loss": 0.7746, "step": 2180 }, { "epoch": 0.8251205901825405, "grad_norm": 1.0012200273146032, "learning_rate": 7.281851430122095e-06, "loss": 0.761, "step": 2181 }, { "epoch": 0.8254989123238438, "grad_norm": 0.9971724282492034, "learning_rate": 7.280843369594601e-06, "loss": 0.7464, "step": 2182 }, { "epoch": 0.825877234465147, "grad_norm": 0.9992511338369285, "learning_rate": 7.279834671937177e-06, "loss": 0.7713, "step": 2183 }, { "epoch": 0.8262555566064504, "grad_norm": 0.9940143408783403, "learning_rate": 7.278825337345711e-06, "loss": 0.7486, "step": 2184 }, { "epoch": 0.8266338787477537, "grad_norm": 0.9975957100214181, "learning_rate": 7.277815366016212e-06, "loss": 0.7622, "step": 2185 }, { "epoch": 0.827012200889057, "grad_norm": 1.0053659263013601, "learning_rate": 7.276804758144815e-06, "loss": 0.8013, "step": 2186 }, { "epoch": 0.8273905230303603, "grad_norm": 1.001627529272579, "learning_rate": 7.275793513927773e-06, "loss": 0.759, "step": 2187 }, { "epoch": 0.8277688451716637, "grad_norm": 0.9996111611603178, "learning_rate": 7.27478163356147e-06, "loss": 0.8009, "step": 2188 }, { "epoch": 0.828147167312967, "grad_norm": 1.0391320056927147, "learning_rate": 7.27376911724241e-06, "loss": 0.7749, "step": 2189 }, { "epoch": 0.8285254894542703, "grad_norm": 1.0007479964257233, "learning_rate": 7.272755965167219e-06, "loss": 0.7745, "step": 2190 }, { "epoch": 0.8289038115955736, "grad_norm": 0.9917344049682559, "learning_rate": 7.271742177532649e-06, "loss": 0.7984, "step": 2191 }, { "epoch": 0.829282133736877, "grad_norm": 1.0388689859693216, "learning_rate": 7.270727754535576e-06, "loss": 0.7835, "step": 2192 }, { "epoch": 0.8296604558781803, "grad_norm": 0.9627429518391599, "learning_rate": 7.269712696372995e-06, "loss": 0.7437, "step": 2193 }, { "epoch": 0.8300387780194836, "grad_norm": 1.005153806625211, "learning_rate": 7.2686970032420284e-06, "loss": 0.726, "step": 2194 }, { "epoch": 0.830417100160787, "grad_norm": 1.0338770873839198, "learning_rate": 7.267680675339922e-06, "loss": 0.7939, "step": 2195 }, { "epoch": 0.8307954223020902, "grad_norm": 1.0017247758066885, "learning_rate": 7.266663712864041e-06, "loss": 0.7308, "step": 2196 }, { "epoch": 0.8311737444433935, "grad_norm": 1.0154861075122885, "learning_rate": 7.265646116011879e-06, "loss": 0.7554, "step": 2197 }, { "epoch": 0.8315520665846968, "grad_norm": 1.0234178291342528, "learning_rate": 7.264627884981051e-06, "loss": 0.8224, "step": 2198 }, { "epoch": 0.8319303887260002, "grad_norm": 1.0125985915447324, "learning_rate": 7.26360901996929e-06, "loss": 0.8038, "step": 2199 }, { "epoch": 0.8323087108673035, "grad_norm": 0.9825269778842675, "learning_rate": 7.262589521174462e-06, "loss": 0.789, "step": 2200 }, { "epoch": 0.8326870330086068, "grad_norm": 0.9764307725469166, "learning_rate": 7.261569388794545e-06, "loss": 0.7588, "step": 2201 }, { "epoch": 0.8330653551499102, "grad_norm": 1.016402207074279, "learning_rate": 7.260548623027651e-06, "loss": 0.748, "step": 2202 }, { "epoch": 0.8334436772912135, "grad_norm": 0.994810826945665, "learning_rate": 7.259527224072006e-06, "loss": 0.7359, "step": 2203 }, { "epoch": 0.8338219994325168, "grad_norm": 1.0083747782827215, "learning_rate": 7.2585051921259625e-06, "loss": 0.7521, "step": 2204 }, { "epoch": 0.8342003215738201, "grad_norm": 0.9946978450754933, "learning_rate": 7.257482527387999e-06, "loss": 0.7485, "step": 2205 }, { "epoch": 0.8345786437151235, "grad_norm": 1.0451273527111948, "learning_rate": 7.256459230056712e-06, "loss": 0.7978, "step": 2206 }, { "epoch": 0.8349569658564268, "grad_norm": 1.0296210997400246, "learning_rate": 7.255435300330822e-06, "loss": 0.7345, "step": 2207 }, { "epoch": 0.83533528799773, "grad_norm": 1.0535232921353834, "learning_rate": 7.2544107384091755e-06, "loss": 0.7763, "step": 2208 }, { "epoch": 0.8357136101390333, "grad_norm": 0.9768577408534046, "learning_rate": 7.253385544490736e-06, "loss": 0.751, "step": 2209 }, { "epoch": 0.8360919322803367, "grad_norm": 1.0580469696656276, "learning_rate": 7.252359718774597e-06, "loss": 0.758, "step": 2210 }, { "epoch": 0.83647025442164, "grad_norm": 0.9929715749347985, "learning_rate": 7.2513332614599675e-06, "loss": 0.7896, "step": 2211 }, { "epoch": 0.8368485765629433, "grad_norm": 1.1099082719051916, "learning_rate": 7.250306172746184e-06, "loss": 0.7772, "step": 2212 }, { "epoch": 0.8372268987042467, "grad_norm": 0.9940485867460078, "learning_rate": 7.249278452832705e-06, "loss": 0.7947, "step": 2213 }, { "epoch": 0.83760522084555, "grad_norm": 1.0246061053371784, "learning_rate": 7.248250101919108e-06, "loss": 0.7893, "step": 2214 }, { "epoch": 0.8379835429868533, "grad_norm": 0.9893371029349722, "learning_rate": 7.2472211202050965e-06, "loss": 0.7719, "step": 2215 }, { "epoch": 0.8383618651281566, "grad_norm": 0.9718025440808898, "learning_rate": 7.246191507890497e-06, "loss": 0.7851, "step": 2216 }, { "epoch": 0.83874018726946, "grad_norm": 1.0577875282663591, "learning_rate": 7.245161265175256e-06, "loss": 0.7631, "step": 2217 }, { "epoch": 0.8391185094107633, "grad_norm": 0.9666158593068822, "learning_rate": 7.2441303922594444e-06, "loss": 0.7525, "step": 2218 }, { "epoch": 0.8394968315520666, "grad_norm": 1.0269689394601913, "learning_rate": 7.243098889343253e-06, "loss": 0.7803, "step": 2219 }, { "epoch": 0.83987515369337, "grad_norm": 1.0433224090730777, "learning_rate": 7.242066756626998e-06, "loss": 0.7779, "step": 2220 }, { "epoch": 0.8402534758346732, "grad_norm": 1.00399384602572, "learning_rate": 7.241033994311116e-06, "loss": 0.8182, "step": 2221 }, { "epoch": 0.8406317979759765, "grad_norm": 1.014003363462672, "learning_rate": 7.240000602596168e-06, "loss": 0.7524, "step": 2222 }, { "epoch": 0.8410101201172798, "grad_norm": 0.9860232096412463, "learning_rate": 7.2389665816828325e-06, "loss": 0.7476, "step": 2223 }, { "epoch": 0.8413884422585832, "grad_norm": 0.9783448367100456, "learning_rate": 7.237931931771915e-06, "loss": 0.6886, "step": 2224 }, { "epoch": 0.8417667643998865, "grad_norm": 0.992553581438228, "learning_rate": 7.23689665306434e-06, "loss": 0.8045, "step": 2225 }, { "epoch": 0.8421450865411898, "grad_norm": 1.0208432653178292, "learning_rate": 7.235860745761159e-06, "loss": 0.7921, "step": 2226 }, { "epoch": 0.8425234086824931, "grad_norm": 1.0158819065635032, "learning_rate": 7.234824210063539e-06, "loss": 0.7985, "step": 2227 }, { "epoch": 0.8429017308237965, "grad_norm": 0.9819725891779805, "learning_rate": 7.233787046172772e-06, "loss": 0.7393, "step": 2228 }, { "epoch": 0.8432800529650998, "grad_norm": 1.051892253060586, "learning_rate": 7.232749254290274e-06, "loss": 0.7596, "step": 2229 }, { "epoch": 0.8436583751064031, "grad_norm": 0.9845966887135575, "learning_rate": 7.231710834617579e-06, "loss": 0.7815, "step": 2230 }, { "epoch": 0.8440366972477065, "grad_norm": 1.052693697055086, "learning_rate": 7.230671787356346e-06, "loss": 0.8102, "step": 2231 }, { "epoch": 0.8444150193890098, "grad_norm": 1.0814629634335617, "learning_rate": 7.2296321127083565e-06, "loss": 0.786, "step": 2232 }, { "epoch": 0.844793341530313, "grad_norm": 1.0487315558993753, "learning_rate": 7.228591810875509e-06, "loss": 0.7718, "step": 2233 }, { "epoch": 0.8451716636716163, "grad_norm": 0.998389444486355, "learning_rate": 7.227550882059829e-06, "loss": 0.7136, "step": 2234 }, { "epoch": 0.8455499858129197, "grad_norm": 1.0160592607513808, "learning_rate": 7.226509326463462e-06, "loss": 0.7694, "step": 2235 }, { "epoch": 0.845928307954223, "grad_norm": 1.0296557439399707, "learning_rate": 7.225467144288673e-06, "loss": 0.7836, "step": 2236 }, { "epoch": 0.8463066300955263, "grad_norm": 0.9899949990998219, "learning_rate": 7.224424335737854e-06, "loss": 0.7606, "step": 2237 }, { "epoch": 0.8466849522368297, "grad_norm": 1.032233571909769, "learning_rate": 7.223380901013511e-06, "loss": 0.8011, "step": 2238 }, { "epoch": 0.847063274378133, "grad_norm": 1.032906429409592, "learning_rate": 7.2223368403182795e-06, "loss": 0.7845, "step": 2239 }, { "epoch": 0.8474415965194363, "grad_norm": 1.001821699955479, "learning_rate": 7.221292153854911e-06, "loss": 0.713, "step": 2240 }, { "epoch": 0.8478199186607396, "grad_norm": 1.0229036832276601, "learning_rate": 7.220246841826281e-06, "loss": 0.7572, "step": 2241 }, { "epoch": 0.848198240802043, "grad_norm": 1.0387317062679455, "learning_rate": 7.219200904435388e-06, "loss": 0.8093, "step": 2242 }, { "epoch": 0.8485765629433463, "grad_norm": 1.0130632257254073, "learning_rate": 7.218154341885345e-06, "loss": 0.7425, "step": 2243 }, { "epoch": 0.8489548850846496, "grad_norm": 1.0046700347238025, "learning_rate": 7.217107154379396e-06, "loss": 0.8065, "step": 2244 }, { "epoch": 0.8493332072259528, "grad_norm": 1.0444439529419776, "learning_rate": 7.216059342120899e-06, "loss": 0.7937, "step": 2245 }, { "epoch": 0.8497115293672562, "grad_norm": 1.017761286392246, "learning_rate": 7.215010905313337e-06, "loss": 0.7786, "step": 2246 }, { "epoch": 0.8500898515085595, "grad_norm": 0.9842033756574665, "learning_rate": 7.213961844160314e-06, "loss": 0.738, "step": 2247 }, { "epoch": 0.8504681736498628, "grad_norm": 1.0440733410115375, "learning_rate": 7.212912158865553e-06, "loss": 0.795, "step": 2248 }, { "epoch": 0.8508464957911662, "grad_norm": 1.0534323824570648, "learning_rate": 7.2118618496329e-06, "loss": 0.789, "step": 2249 }, { "epoch": 0.8512248179324695, "grad_norm": 1.0128288698320076, "learning_rate": 7.210810916666323e-06, "loss": 0.7601, "step": 2250 }, { "epoch": 0.8516031400737728, "grad_norm": 0.9982166918026206, "learning_rate": 7.20975936016991e-06, "loss": 0.7584, "step": 2251 }, { "epoch": 0.8519814622150761, "grad_norm": 1.0104750805001261, "learning_rate": 7.2087071803478674e-06, "loss": 0.7648, "step": 2252 }, { "epoch": 0.8523597843563795, "grad_norm": 1.024270507762229, "learning_rate": 7.207654377404528e-06, "loss": 0.7253, "step": 2253 }, { "epoch": 0.8527381064976828, "grad_norm": 1.0542205370524342, "learning_rate": 7.2066009515443435e-06, "loss": 0.8327, "step": 2254 }, { "epoch": 0.8531164286389861, "grad_norm": 0.9912066811718258, "learning_rate": 7.205546902971885e-06, "loss": 0.7751, "step": 2255 }, { "epoch": 0.8534947507802895, "grad_norm": 0.999231349972615, "learning_rate": 7.204492231891844e-06, "loss": 0.7673, "step": 2256 }, { "epoch": 0.8538730729215928, "grad_norm": 1.03750701214942, "learning_rate": 7.2034369385090375e-06, "loss": 0.7655, "step": 2257 }, { "epoch": 0.854251395062896, "grad_norm": 0.9797275715016183, "learning_rate": 7.202381023028399e-06, "loss": 0.7531, "step": 2258 }, { "epoch": 0.8546297172041993, "grad_norm": 1.0301926734882334, "learning_rate": 7.201324485654982e-06, "loss": 0.7387, "step": 2259 }, { "epoch": 0.8550080393455027, "grad_norm": 1.003615638108583, "learning_rate": 7.200267326593966e-06, "loss": 0.7803, "step": 2260 }, { "epoch": 0.855386361486806, "grad_norm": 1.0162154996402277, "learning_rate": 7.199209546050646e-06, "loss": 0.7674, "step": 2261 }, { "epoch": 0.855386361486806, "eval_loss": 0.7731848955154419, "eval_runtime": 25.6495, "eval_samples_per_second": 34.504, "eval_steps_per_second": 1.092, "step": 2261 }, { "epoch": 0.855386361486806, "eval_bench_accuracy_arc_challenge": 0.0, "eval_bench_accuracy_hellaswag": 0.21, "eval_bench_accuracy_mmlu": 0.20869565217391303, "eval_bench_average_accuracy": 0.13956521739130434, "eval_bench_loss": 7.6725517406798245, "eval_bench_total_accuracy": 0.14505494505494507, "step": 2261 }, { "epoch": 0.8557646836281093, "grad_norm": 1.0412009573747818, "learning_rate": 7.198151144230442e-06, "loss": 0.7652, "step": 2262 }, { "epoch": 0.8561430057694126, "grad_norm": 0.9912857024371587, "learning_rate": 7.19709212133889e-06, "loss": 0.7311, "step": 2263 }, { "epoch": 0.856521327910716, "grad_norm": 1.0104795488446559, "learning_rate": 7.196032477581651e-06, "loss": 0.7253, "step": 2264 }, { "epoch": 0.8568996500520193, "grad_norm": 1.0126990695901916, "learning_rate": 7.194972213164503e-06, "loss": 0.8314, "step": 2265 }, { "epoch": 0.8572779721933226, "grad_norm": 1.049341981781742, "learning_rate": 7.193911328293347e-06, "loss": 0.7698, "step": 2266 }, { "epoch": 0.857656294334626, "grad_norm": 1.0214946966423453, "learning_rate": 7.192849823174205e-06, "loss": 0.7427, "step": 2267 }, { "epoch": 0.8580346164759293, "grad_norm": 1.0124359559388285, "learning_rate": 7.191787698013215e-06, "loss": 0.7708, "step": 2268 }, { "epoch": 0.8584129386172326, "grad_norm": 1.0288994749349432, "learning_rate": 7.190724953016641e-06, "loss": 0.7601, "step": 2269 }, { "epoch": 0.8587912607585358, "grad_norm": 0.9911988587136489, "learning_rate": 7.189661588390864e-06, "loss": 0.7539, "step": 2270 }, { "epoch": 0.8591695828998392, "grad_norm": 0.9737289335443161, "learning_rate": 7.188597604342387e-06, "loss": 0.8045, "step": 2271 }, { "epoch": 0.8595479050411425, "grad_norm": 1.0299900683875711, "learning_rate": 7.187533001077831e-06, "loss": 0.7469, "step": 2272 }, { "epoch": 0.8599262271824458, "grad_norm": 0.9692550794741338, "learning_rate": 7.1864677788039405e-06, "loss": 0.7832, "step": 2273 }, { "epoch": 0.8603045493237492, "grad_norm": 0.9913944045773587, "learning_rate": 7.185401937727577e-06, "loss": 0.7538, "step": 2274 }, { "epoch": 0.8606828714650525, "grad_norm": 0.9327497790355345, "learning_rate": 7.184335478055725e-06, "loss": 0.7619, "step": 2275 }, { "epoch": 0.8610611936063558, "grad_norm": 1.0349847589552532, "learning_rate": 7.183268399995485e-06, "loss": 0.7661, "step": 2276 }, { "epoch": 0.8614395157476591, "grad_norm": 0.9838455296647253, "learning_rate": 7.182200703754084e-06, "loss": 0.7494, "step": 2277 }, { "epoch": 0.8618178378889625, "grad_norm": 1.036668994652531, "learning_rate": 7.181132389538864e-06, "loss": 0.7378, "step": 2278 }, { "epoch": 0.8621961600302658, "grad_norm": 0.9966143490121748, "learning_rate": 7.180063457557288e-06, "loss": 0.7201, "step": 2279 }, { "epoch": 0.8625744821715691, "grad_norm": 1.0216855086157217, "learning_rate": 7.178993908016939e-06, "loss": 0.7221, "step": 2280 }, { "epoch": 0.8629528043128724, "grad_norm": 0.9504238884891278, "learning_rate": 7.177923741125521e-06, "loss": 0.7101, "step": 2281 }, { "epoch": 0.8633311264541758, "grad_norm": 0.9898970091046093, "learning_rate": 7.176852957090857e-06, "loss": 0.7593, "step": 2282 }, { "epoch": 0.863709448595479, "grad_norm": 1.0112937497346717, "learning_rate": 7.17578155612089e-06, "loss": 0.7447, "step": 2283 }, { "epoch": 0.8640877707367823, "grad_norm": 1.0185639349131448, "learning_rate": 7.174709538423684e-06, "loss": 0.752, "step": 2284 }, { "epoch": 0.8644660928780857, "grad_norm": 0.9549086201000673, "learning_rate": 7.17363690420742e-06, "loss": 0.7571, "step": 2285 }, { "epoch": 0.864844415019389, "grad_norm": 1.03834118938008, "learning_rate": 7.1725636536804e-06, "loss": 0.732, "step": 2286 }, { "epoch": 0.8652227371606923, "grad_norm": 0.994855471346937, "learning_rate": 7.1714897870510475e-06, "loss": 0.7754, "step": 2287 }, { "epoch": 0.8656010593019956, "grad_norm": 1.0144762721512632, "learning_rate": 7.170415304527904e-06, "loss": 0.779, "step": 2288 }, { "epoch": 0.865979381443299, "grad_norm": 1.0173462667637359, "learning_rate": 7.169340206319629e-06, "loss": 0.775, "step": 2289 }, { "epoch": 0.8663577035846023, "grad_norm": 1.0393665262607155, "learning_rate": 7.168264492635007e-06, "loss": 0.7625, "step": 2290 }, { "epoch": 0.8667360257259056, "grad_norm": 1.0146930229552926, "learning_rate": 7.167188163682935e-06, "loss": 0.7842, "step": 2291 }, { "epoch": 0.867114347867209, "grad_norm": 1.0431135201455501, "learning_rate": 7.166111219672433e-06, "loss": 0.8075, "step": 2292 }, { "epoch": 0.8674926700085123, "grad_norm": 1.0113058473321697, "learning_rate": 7.165033660812643e-06, "loss": 0.7573, "step": 2293 }, { "epoch": 0.8678709921498156, "grad_norm": 1.00907022185597, "learning_rate": 7.1639554873128215e-06, "loss": 0.7648, "step": 2294 }, { "epoch": 0.8682493142911188, "grad_norm": 1.0060435969797255, "learning_rate": 7.162876699382346e-06, "loss": 0.7145, "step": 2295 }, { "epoch": 0.8686276364324222, "grad_norm": 0.9807023329510167, "learning_rate": 7.161797297230716e-06, "loss": 0.7476, "step": 2296 }, { "epoch": 0.8690059585737255, "grad_norm": 1.0242347975147827, "learning_rate": 7.160717281067547e-06, "loss": 0.8138, "step": 2297 }, { "epoch": 0.8693842807150288, "grad_norm": 1.0298773248234774, "learning_rate": 7.159636651102574e-06, "loss": 0.8271, "step": 2298 }, { "epoch": 0.8697626028563321, "grad_norm": 0.9934220738457435, "learning_rate": 7.158555407545654e-06, "loss": 0.7879, "step": 2299 }, { "epoch": 0.8701409249976355, "grad_norm": 0.9999472676122242, "learning_rate": 7.157473550606759e-06, "loss": 0.7665, "step": 2300 }, { "epoch": 0.8705192471389388, "grad_norm": 1.0177734120658402, "learning_rate": 7.156391080495984e-06, "loss": 0.7531, "step": 2301 }, { "epoch": 0.8708975692802421, "grad_norm": 1.0086388990884834, "learning_rate": 7.155307997423541e-06, "loss": 0.7457, "step": 2302 }, { "epoch": 0.8712758914215455, "grad_norm": 0.9796780747447114, "learning_rate": 7.154224301599763e-06, "loss": 0.7418, "step": 2303 }, { "epoch": 0.8716542135628488, "grad_norm": 0.9961091459135317, "learning_rate": 7.153139993235098e-06, "loss": 0.7291, "step": 2304 }, { "epoch": 0.8720325357041521, "grad_norm": 1.0116589450888551, "learning_rate": 7.152055072540117e-06, "loss": 0.7725, "step": 2305 }, { "epoch": 0.8724108578454554, "grad_norm": 1.0194100669046098, "learning_rate": 7.150969539725506e-06, "loss": 0.8202, "step": 2306 }, { "epoch": 0.8727891799867588, "grad_norm": 0.9919069346041747, "learning_rate": 7.149883395002076e-06, "loss": 0.7728, "step": 2307 }, { "epoch": 0.873167502128062, "grad_norm": 0.9765364177481878, "learning_rate": 7.1487966385807494e-06, "loss": 0.7212, "step": 2308 }, { "epoch": 0.8735458242693653, "grad_norm": 0.9872934189660968, "learning_rate": 7.147709270672575e-06, "loss": 0.8015, "step": 2309 }, { "epoch": 0.8739241464106687, "grad_norm": 1.0148200655162085, "learning_rate": 7.1466212914887115e-06, "loss": 0.7715, "step": 2310 }, { "epoch": 0.874302468551972, "grad_norm": 1.0652420727645235, "learning_rate": 7.145532701240446e-06, "loss": 0.7819, "step": 2311 }, { "epoch": 0.8746807906932753, "grad_norm": 0.9826811637144443, "learning_rate": 7.1444435001391755e-06, "loss": 0.7231, "step": 2312 }, { "epoch": 0.8750591128345786, "grad_norm": 1.038520876786242, "learning_rate": 7.143353688396421e-06, "loss": 0.7896, "step": 2313 }, { "epoch": 0.875437434975882, "grad_norm": 1.0215557969029825, "learning_rate": 7.142263266223823e-06, "loss": 0.7963, "step": 2314 }, { "epoch": 0.8758157571171853, "grad_norm": 1.0484026660723595, "learning_rate": 7.141172233833135e-06, "loss": 0.7561, "step": 2315 }, { "epoch": 0.8761940792584886, "grad_norm": 1.015997044144397, "learning_rate": 7.140080591436234e-06, "loss": 0.7695, "step": 2316 }, { "epoch": 0.8765724013997919, "grad_norm": 1.04281586679857, "learning_rate": 7.138988339245113e-06, "loss": 0.7769, "step": 2317 }, { "epoch": 0.8769507235410953, "grad_norm": 0.9806032529211375, "learning_rate": 7.137895477471883e-06, "loss": 0.7673, "step": 2318 }, { "epoch": 0.8773290456823986, "grad_norm": 0.9765881424303234, "learning_rate": 7.1368020063287766e-06, "loss": 0.7803, "step": 2319 }, { "epoch": 0.8777073678237018, "grad_norm": 1.0121519077458367, "learning_rate": 7.135707926028141e-06, "loss": 0.7763, "step": 2320 }, { "epoch": 0.8780856899650052, "grad_norm": 1.0195213279030102, "learning_rate": 7.134613236782445e-06, "loss": 0.7601, "step": 2321 }, { "epoch": 0.8784640121063085, "grad_norm": 0.9926848801780171, "learning_rate": 7.133517938804272e-06, "loss": 0.773, "step": 2322 }, { "epoch": 0.8788423342476118, "grad_norm": 1.013335711603836, "learning_rate": 7.132422032306327e-06, "loss": 0.7492, "step": 2323 }, { "epoch": 0.8792206563889151, "grad_norm": 1.0269526120318977, "learning_rate": 7.131325517501431e-06, "loss": 0.7475, "step": 2324 }, { "epoch": 0.8795989785302185, "grad_norm": 0.9532233993510265, "learning_rate": 7.130228394602525e-06, "loss": 0.7637, "step": 2325 }, { "epoch": 0.8799773006715218, "grad_norm": 1.0383963544087091, "learning_rate": 7.129130663822665e-06, "loss": 0.7718, "step": 2326 }, { "epoch": 0.8803556228128251, "grad_norm": 0.9999763719084931, "learning_rate": 7.128032325375029e-06, "loss": 0.7945, "step": 2327 }, { "epoch": 0.8807339449541285, "grad_norm": 1.0056490057535235, "learning_rate": 7.126933379472909e-06, "loss": 0.8182, "step": 2328 }, { "epoch": 0.8811122670954318, "grad_norm": 1.0023181500357063, "learning_rate": 7.125833826329719e-06, "loss": 0.7814, "step": 2329 }, { "epoch": 0.8814905892367351, "grad_norm": 1.003862101973702, "learning_rate": 7.124733666158988e-06, "loss": 0.7705, "step": 2330 }, { "epoch": 0.8818689113780384, "grad_norm": 1.018661074932826, "learning_rate": 7.123632899174363e-06, "loss": 0.7518, "step": 2331 }, { "epoch": 0.8822472335193418, "grad_norm": 0.9603199805069197, "learning_rate": 7.122531525589611e-06, "loss": 0.7422, "step": 2332 }, { "epoch": 0.882625555660645, "grad_norm": 1.0327434608146935, "learning_rate": 7.121429545618616e-06, "loss": 0.7267, "step": 2333 }, { "epoch": 0.8830038778019483, "grad_norm": 1.0461189415748307, "learning_rate": 7.120326959475377e-06, "loss": 0.7738, "step": 2334 }, { "epoch": 0.8833821999432517, "grad_norm": 1.035602613015718, "learning_rate": 7.119223767374015e-06, "loss": 0.7892, "step": 2335 }, { "epoch": 0.883760522084555, "grad_norm": 0.968027276282092, "learning_rate": 7.118119969528765e-06, "loss": 0.7885, "step": 2336 }, { "epoch": 0.8841388442258583, "grad_norm": 1.0112556770782621, "learning_rate": 7.117015566153981e-06, "loss": 0.7884, "step": 2337 }, { "epoch": 0.8845171663671616, "grad_norm": 1.0047739746600866, "learning_rate": 7.115910557464136e-06, "loss": 0.7522, "step": 2338 }, { "epoch": 0.884895488508465, "grad_norm": 1.0141841752207075, "learning_rate": 7.1148049436738205e-06, "loss": 0.791, "step": 2339 }, { "epoch": 0.8852738106497683, "grad_norm": 1.0208993885255921, "learning_rate": 7.113698724997739e-06, "loss": 0.7761, "step": 2340 }, { "epoch": 0.8856521327910716, "grad_norm": 1.033124967213344, "learning_rate": 7.112591901650717e-06, "loss": 0.7807, "step": 2341 }, { "epoch": 0.8860304549323749, "grad_norm": 1.0067075556573555, "learning_rate": 7.111484473847696e-06, "loss": 0.7383, "step": 2342 }, { "epoch": 0.8864087770736783, "grad_norm": 1.0177677912510785, "learning_rate": 7.110376441803735e-06, "loss": 0.7602, "step": 2343 }, { "epoch": 0.8867870992149816, "grad_norm": 1.0465141477050506, "learning_rate": 7.109267805734011e-06, "loss": 0.7735, "step": 2344 }, { "epoch": 0.8871654213562848, "grad_norm": 0.9963213834693853, "learning_rate": 7.108158565853817e-06, "loss": 0.7375, "step": 2345 }, { "epoch": 0.8875437434975882, "grad_norm": 1.0163808135186847, "learning_rate": 7.107048722378565e-06, "loss": 0.7742, "step": 2346 }, { "epoch": 0.8879220656388915, "grad_norm": 0.990324738375918, "learning_rate": 7.105938275523783e-06, "loss": 0.7716, "step": 2347 }, { "epoch": 0.8883003877801948, "grad_norm": 1.0075180979561382, "learning_rate": 7.1048272255051155e-06, "loss": 0.7749, "step": 2348 }, { "epoch": 0.8886787099214981, "grad_norm": 1.0025806289786972, "learning_rate": 7.103715572538327e-06, "loss": 0.7633, "step": 2349 }, { "epoch": 0.8890570320628015, "grad_norm": 0.9434765318565549, "learning_rate": 7.1026033168392955e-06, "loss": 0.7206, "step": 2350 }, { "epoch": 0.8894353542041048, "grad_norm": 1.031863789278887, "learning_rate": 7.101490458624019e-06, "loss": 0.772, "step": 2351 }, { "epoch": 0.8898136763454081, "grad_norm": 0.9998099190042471, "learning_rate": 7.100376998108609e-06, "loss": 0.7844, "step": 2352 }, { "epoch": 0.8901919984867115, "grad_norm": 1.051581162222378, "learning_rate": 7.099262935509298e-06, "loss": 0.7347, "step": 2353 }, { "epoch": 0.8905703206280148, "grad_norm": 1.0353178470338773, "learning_rate": 7.098148271042434e-06, "loss": 0.7863, "step": 2354 }, { "epoch": 0.8909486427693181, "grad_norm": 1.0234255328326827, "learning_rate": 7.0970330049244796e-06, "loss": 0.748, "step": 2355 }, { "epoch": 0.8913269649106214, "grad_norm": 1.0295240998477164, "learning_rate": 7.0959171373720185e-06, "loss": 0.7542, "step": 2356 }, { "epoch": 0.8917052870519248, "grad_norm": 1.0198352782855347, "learning_rate": 7.094800668601747e-06, "loss": 0.7536, "step": 2357 }, { "epoch": 0.892083609193228, "grad_norm": 1.0224916784724847, "learning_rate": 7.093683598830481e-06, "loss": 0.7686, "step": 2358 }, { "epoch": 0.8924619313345313, "grad_norm": 0.9932697810286107, "learning_rate": 7.092565928275151e-06, "loss": 0.7587, "step": 2359 }, { "epoch": 0.8928402534758346, "grad_norm": 0.9996944130579072, "learning_rate": 7.091447657152806e-06, "loss": 0.7424, "step": 2360 }, { "epoch": 0.893218575617138, "grad_norm": 1.0357994416158893, "learning_rate": 7.0903287856806115e-06, "loss": 0.7565, "step": 2361 }, { "epoch": 0.8935968977584413, "grad_norm": 1.0362986915824348, "learning_rate": 7.089209314075848e-06, "loss": 0.8214, "step": 2362 }, { "epoch": 0.8939752198997446, "grad_norm": 1.0413906937629382, "learning_rate": 7.0880892425559125e-06, "loss": 0.8131, "step": 2363 }, { "epoch": 0.894353542041048, "grad_norm": 1.049270279884044, "learning_rate": 7.0869685713383224e-06, "loss": 0.7611, "step": 2364 }, { "epoch": 0.8947318641823513, "grad_norm": 1.006126195838057, "learning_rate": 7.085847300640708e-06, "loss": 0.779, "step": 2365 }, { "epoch": 0.8951101863236546, "grad_norm": 1.0247424321465883, "learning_rate": 7.084725430680814e-06, "loss": 0.7754, "step": 2366 }, { "epoch": 0.8954885084649579, "grad_norm": 0.9999318763539188, "learning_rate": 7.083602961676508e-06, "loss": 0.7531, "step": 2367 }, { "epoch": 0.8958668306062613, "grad_norm": 1.0433712571818266, "learning_rate": 7.082479893845766e-06, "loss": 0.7944, "step": 2368 }, { "epoch": 0.8962451527475646, "grad_norm": 1.041108943613929, "learning_rate": 7.081356227406688e-06, "loss": 0.7678, "step": 2369 }, { "epoch": 0.8966234748888678, "grad_norm": 0.98532548481682, "learning_rate": 7.080231962577484e-06, "loss": 0.7438, "step": 2370 }, { "epoch": 0.8970017970301712, "grad_norm": 1.0156341127442319, "learning_rate": 7.079107099576486e-06, "loss": 0.7598, "step": 2371 }, { "epoch": 0.8973801191714745, "grad_norm": 0.9953543675844062, "learning_rate": 7.077981638622136e-06, "loss": 0.7864, "step": 2372 }, { "epoch": 0.8977584413127778, "grad_norm": 1.0022146793260511, "learning_rate": 7.076855579932996e-06, "loss": 0.7825, "step": 2373 }, { "epoch": 0.8981367634540811, "grad_norm": 1.0427697492763814, "learning_rate": 7.075728923727743e-06, "loss": 0.7888, "step": 2374 }, { "epoch": 0.8985150855953845, "grad_norm": 0.9958817090752511, "learning_rate": 7.0746016702251705e-06, "loss": 0.7963, "step": 2375 }, { "epoch": 0.8988934077366878, "grad_norm": 0.9950789001489996, "learning_rate": 7.073473819644188e-06, "loss": 0.748, "step": 2376 }, { "epoch": 0.8992717298779911, "grad_norm": 1.043391376102214, "learning_rate": 7.072345372203821e-06, "loss": 0.7611, "step": 2377 }, { "epoch": 0.8996500520192944, "grad_norm": 0.9894685409366812, "learning_rate": 7.071216328123211e-06, "loss": 0.7709, "step": 2378 }, { "epoch": 0.9000283741605978, "grad_norm": 0.9558223257112808, "learning_rate": 7.070086687621614e-06, "loss": 0.7475, "step": 2379 }, { "epoch": 0.9004066963019011, "grad_norm": 1.0622467722809246, "learning_rate": 7.068956450918402e-06, "loss": 0.7831, "step": 2380 }, { "epoch": 0.9007850184432044, "grad_norm": 1.0492486815404989, "learning_rate": 7.067825618233064e-06, "loss": 0.7823, "step": 2381 }, { "epoch": 0.9011633405845078, "grad_norm": 1.0547751788072182, "learning_rate": 7.066694189785207e-06, "loss": 0.7386, "step": 2382 }, { "epoch": 0.901541662725811, "grad_norm": 0.9973094858870168, "learning_rate": 7.065562165794548e-06, "loss": 0.7391, "step": 2383 }, { "epoch": 0.9019199848671143, "grad_norm": 0.98446276280215, "learning_rate": 7.064429546480923e-06, "loss": 0.7646, "step": 2384 }, { "epoch": 0.9022983070084176, "grad_norm": 0.9987151345524554, "learning_rate": 7.0632963320642835e-06, "loss": 0.7447, "step": 2385 }, { "epoch": 0.902676629149721, "grad_norm": 1.0090087604609146, "learning_rate": 7.062162522764697e-06, "loss": 0.7624, "step": 2386 }, { "epoch": 0.9030549512910243, "grad_norm": 0.986158312317026, "learning_rate": 7.0610281188023456e-06, "loss": 0.7468, "step": 2387 }, { "epoch": 0.9034332734323276, "grad_norm": 0.9908417969281411, "learning_rate": 7.0598931203975265e-06, "loss": 0.7445, "step": 2388 }, { "epoch": 0.903811595573631, "grad_norm": 1.00470221573058, "learning_rate": 7.058757527770654e-06, "loss": 0.7553, "step": 2389 }, { "epoch": 0.9041899177149343, "grad_norm": 1.0142525028955387, "learning_rate": 7.057621341142257e-06, "loss": 0.7672, "step": 2390 }, { "epoch": 0.9045682398562376, "grad_norm": 1.0089711471165104, "learning_rate": 7.056484560732978e-06, "loss": 0.741, "step": 2391 }, { "epoch": 0.9049465619975409, "grad_norm": 1.021059183337408, "learning_rate": 7.055347186763578e-06, "loss": 0.7847, "step": 2392 }, { "epoch": 0.9053248841388443, "grad_norm": 0.9798134218874166, "learning_rate": 7.0542092194549285e-06, "loss": 0.7293, "step": 2393 }, { "epoch": 0.9057032062801476, "grad_norm": 0.9706880657392031, "learning_rate": 7.053070659028024e-06, "loss": 0.7794, "step": 2394 }, { "epoch": 0.9057032062801476, "eval_loss": 0.7680747509002686, "eval_runtime": 25.5451, "eval_samples_per_second": 34.645, "eval_steps_per_second": 1.096, "step": 2394 }, { "epoch": 0.9057032062801476, "eval_bench_accuracy_arc_challenge": 0.0, "eval_bench_accuracy_hellaswag": 0.205, "eval_bench_accuracy_mmlu": 0.20869565217391303, "eval_bench_average_accuracy": 0.13789855072463766, "eval_bench_loss": 7.451362476014254, "eval_bench_total_accuracy": 0.14285714285714285, "step": 2394 }, { "epoch": 0.9060815284214508, "grad_norm": 0.999929570943935, "learning_rate": 7.051931505703967e-06, "loss": 0.7769, "step": 2395 }, { "epoch": 0.9064598505627541, "grad_norm": 0.9790157544755365, "learning_rate": 7.0507917597039765e-06, "loss": 0.7608, "step": 2396 }, { "epoch": 0.9068381727040575, "grad_norm": 0.9955393136714028, "learning_rate": 7.04965142124939e-06, "loss": 0.7653, "step": 2397 }, { "epoch": 0.9072164948453608, "grad_norm": 0.9875148869531033, "learning_rate": 7.048510490561655e-06, "loss": 0.7028, "step": 2398 }, { "epoch": 0.9075948169866641, "grad_norm": 1.0421138342467342, "learning_rate": 7.047368967862342e-06, "loss": 0.7424, "step": 2399 }, { "epoch": 0.9079731391279675, "grad_norm": 0.9937866860623155, "learning_rate": 7.046226853373125e-06, "loss": 0.7624, "step": 2400 }, { "epoch": 0.9083514612692708, "grad_norm": 0.9795947439211311, "learning_rate": 7.0450841473158026e-06, "loss": 0.791, "step": 2401 }, { "epoch": 0.9087297834105741, "grad_norm": 1.0271100426041748, "learning_rate": 7.043940849912285e-06, "loss": 0.7433, "step": 2402 }, { "epoch": 0.9091081055518774, "grad_norm": 0.9630746502196654, "learning_rate": 7.042796961384596e-06, "loss": 0.8024, "step": 2403 }, { "epoch": 0.9094864276931808, "grad_norm": 1.0087369290802906, "learning_rate": 7.041652481954877e-06, "loss": 0.739, "step": 2404 }, { "epoch": 0.9098647498344841, "grad_norm": 1.0258708665585585, "learning_rate": 7.04050741184538e-06, "loss": 0.7313, "step": 2405 }, { "epoch": 0.9102430719757874, "grad_norm": 0.9987870895868104, "learning_rate": 7.039361751278477e-06, "loss": 0.7503, "step": 2406 }, { "epoch": 0.9106213941170908, "grad_norm": 0.9916725633064274, "learning_rate": 7.038215500476649e-06, "loss": 0.765, "step": 2407 }, { "epoch": 0.910999716258394, "grad_norm": 0.995717773961888, "learning_rate": 7.0370686596624965e-06, "loss": 0.7825, "step": 2408 }, { "epoch": 0.9113780383996973, "grad_norm": 1.0289434503270574, "learning_rate": 7.035921229058731e-06, "loss": 0.7977, "step": 2409 }, { "epoch": 0.9117563605410006, "grad_norm": 0.9958712530297943, "learning_rate": 7.034773208888181e-06, "loss": 0.7509, "step": 2410 }, { "epoch": 0.912134682682304, "grad_norm": 1.0550382211959204, "learning_rate": 7.033624599373789e-06, "loss": 0.8412, "step": 2411 }, { "epoch": 0.9125130048236073, "grad_norm": 0.9790726542421223, "learning_rate": 7.032475400738612e-06, "loss": 0.7599, "step": 2412 }, { "epoch": 0.9128913269649106, "grad_norm": 0.9981719741601273, "learning_rate": 7.031325613205818e-06, "loss": 0.7909, "step": 2413 }, { "epoch": 0.9132696491062139, "grad_norm": 1.0149813854916758, "learning_rate": 7.030175236998695e-06, "loss": 0.7659, "step": 2414 }, { "epoch": 0.9136479712475173, "grad_norm": 0.9765548569394097, "learning_rate": 7.029024272340642e-06, "loss": 0.7303, "step": 2415 }, { "epoch": 0.9140262933888206, "grad_norm": 1.0216848187832497, "learning_rate": 7.0278727194551725e-06, "loss": 0.7618, "step": 2416 }, { "epoch": 0.9144046155301239, "grad_norm": 0.9725811064405785, "learning_rate": 7.0267205785659145e-06, "loss": 0.7233, "step": 2417 }, { "epoch": 0.9147829376714273, "grad_norm": 0.9964538932707074, "learning_rate": 7.025567849896611e-06, "loss": 0.7403, "step": 2418 }, { "epoch": 0.9151612598127306, "grad_norm": 1.0054694776115864, "learning_rate": 7.024414533671118e-06, "loss": 0.781, "step": 2419 }, { "epoch": 0.9155395819540338, "grad_norm": 1.0154337801886115, "learning_rate": 7.023260630113405e-06, "loss": 0.7367, "step": 2420 }, { "epoch": 0.9159179040953371, "grad_norm": 1.0203341167769693, "learning_rate": 7.02210613944756e-06, "loss": 0.8087, "step": 2421 }, { "epoch": 0.9162962262366405, "grad_norm": 0.947650552338592, "learning_rate": 7.0209510618977774e-06, "loss": 0.7155, "step": 2422 }, { "epoch": 0.9166745483779438, "grad_norm": 1.0090150775616802, "learning_rate": 7.019795397688373e-06, "loss": 0.7794, "step": 2423 }, { "epoch": 0.9170528705192471, "grad_norm": 0.9571882335592847, "learning_rate": 7.018639147043771e-06, "loss": 0.7076, "step": 2424 }, { "epoch": 0.9174311926605505, "grad_norm": 1.0063061885389344, "learning_rate": 7.017482310188513e-06, "loss": 0.7758, "step": 2425 }, { "epoch": 0.9178095148018538, "grad_norm": 0.9629301761239673, "learning_rate": 7.016324887347253e-06, "loss": 0.7154, "step": 2426 }, { "epoch": 0.9181878369431571, "grad_norm": 0.9573104082993997, "learning_rate": 7.01516687874476e-06, "loss": 0.7459, "step": 2427 }, { "epoch": 0.9185661590844604, "grad_norm": 0.9695229912058232, "learning_rate": 7.014008284605914e-06, "loss": 0.7633, "step": 2428 }, { "epoch": 0.9189444812257638, "grad_norm": 0.9916405550897835, "learning_rate": 7.012849105155712e-06, "loss": 0.7573, "step": 2429 }, { "epoch": 0.9193228033670671, "grad_norm": 0.9986524062176373, "learning_rate": 7.011689340619262e-06, "loss": 0.761, "step": 2430 }, { "epoch": 0.9197011255083704, "grad_norm": 1.0236441362619242, "learning_rate": 7.010528991221788e-06, "loss": 0.7761, "step": 2431 }, { "epoch": 0.9200794476496736, "grad_norm": 1.0067393073595168, "learning_rate": 7.009368057188626e-06, "loss": 0.7842, "step": 2432 }, { "epoch": 0.920457769790977, "grad_norm": 0.996899721738463, "learning_rate": 7.0082065387452245e-06, "loss": 0.7538, "step": 2433 }, { "epoch": 0.9208360919322803, "grad_norm": 0.99997058243727, "learning_rate": 7.007044436117148e-06, "loss": 0.7578, "step": 2434 }, { "epoch": 0.9212144140735836, "grad_norm": 0.9674444295516136, "learning_rate": 7.005881749530074e-06, "loss": 0.7395, "step": 2435 }, { "epoch": 0.921592736214887, "grad_norm": 1.0149587199300343, "learning_rate": 7.004718479209792e-06, "loss": 0.74, "step": 2436 }, { "epoch": 0.9219710583561903, "grad_norm": 0.982017702615209, "learning_rate": 7.003554625382206e-06, "loss": 0.7598, "step": 2437 }, { "epoch": 0.9223493804974936, "grad_norm": 1.044378601494924, "learning_rate": 7.002390188273333e-06, "loss": 0.7415, "step": 2438 }, { "epoch": 0.9227277026387969, "grad_norm": 1.03750251778964, "learning_rate": 7.001225168109302e-06, "loss": 0.7551, "step": 2439 }, { "epoch": 0.9231060247801003, "grad_norm": 1.026894279674952, "learning_rate": 7.000059565116357e-06, "loss": 0.8046, "step": 2440 }, { "epoch": 0.9234843469214036, "grad_norm": 0.9996548023341705, "learning_rate": 6.998893379520856e-06, "loss": 0.7929, "step": 2441 }, { "epoch": 0.9238626690627069, "grad_norm": 0.9654800832683814, "learning_rate": 6.997726611549266e-06, "loss": 0.7578, "step": 2442 }, { "epoch": 0.9242409912040103, "grad_norm": 0.9969850361991277, "learning_rate": 6.996559261428172e-06, "loss": 0.7517, "step": 2443 }, { "epoch": 0.9246193133453136, "grad_norm": 1.0137763399761017, "learning_rate": 6.99539132938427e-06, "loss": 0.7805, "step": 2444 }, { "epoch": 0.9249976354866168, "grad_norm": 1.0004491877637214, "learning_rate": 6.994222815644369e-06, "loss": 0.7702, "step": 2445 }, { "epoch": 0.9253759576279201, "grad_norm": 0.9850295964295233, "learning_rate": 6.993053720435388e-06, "loss": 0.7421, "step": 2446 }, { "epoch": 0.9257542797692235, "grad_norm": 1.0072809709395627, "learning_rate": 6.991884043984365e-06, "loss": 0.7427, "step": 2447 }, { "epoch": 0.9261326019105268, "grad_norm": 0.9603120675902171, "learning_rate": 6.990713786518446e-06, "loss": 0.7741, "step": 2448 }, { "epoch": 0.9265109240518301, "grad_norm": 1.0661697966070056, "learning_rate": 6.989542948264892e-06, "loss": 0.808, "step": 2449 }, { "epoch": 0.9268892461931334, "grad_norm": 1.0156326518533005, "learning_rate": 6.9883715294510775e-06, "loss": 0.7244, "step": 2450 }, { "epoch": 0.9272675683344368, "grad_norm": 1.0150184151775952, "learning_rate": 6.987199530304485e-06, "loss": 0.7383, "step": 2451 }, { "epoch": 0.9276458904757401, "grad_norm": 0.9638266394360308, "learning_rate": 6.986026951052717e-06, "loss": 0.7351, "step": 2452 }, { "epoch": 0.9280242126170434, "grad_norm": 0.9671862930562964, "learning_rate": 6.984853791923483e-06, "loss": 0.7519, "step": 2453 }, { "epoch": 0.9284025347583468, "grad_norm": 1.0291068762144968, "learning_rate": 6.983680053144607e-06, "loss": 0.7215, "step": 2454 }, { "epoch": 0.9287808568996501, "grad_norm": 0.994052763478094, "learning_rate": 6.982505734944027e-06, "loss": 0.7515, "step": 2455 }, { "epoch": 0.9291591790409534, "grad_norm": 1.011400101160467, "learning_rate": 6.981330837549789e-06, "loss": 0.7376, "step": 2456 }, { "epoch": 0.9295375011822566, "grad_norm": 0.9845034325436185, "learning_rate": 6.980155361190058e-06, "loss": 0.7343, "step": 2457 }, { "epoch": 0.92991582332356, "grad_norm": 1.0348865837496204, "learning_rate": 6.978979306093106e-06, "loss": 0.7893, "step": 2458 }, { "epoch": 0.9302941454648633, "grad_norm": 1.0327825604593301, "learning_rate": 6.97780267248732e-06, "loss": 0.7673, "step": 2459 }, { "epoch": 0.9306724676061666, "grad_norm": 0.9996717380132116, "learning_rate": 6.9766254606011984e-06, "loss": 0.7377, "step": 2460 }, { "epoch": 0.93105078974747, "grad_norm": 1.061972465076445, "learning_rate": 6.975447670663353e-06, "loss": 0.758, "step": 2461 }, { "epoch": 0.9314291118887733, "grad_norm": 1.030373026544204, "learning_rate": 6.974269302902506e-06, "loss": 0.8104, "step": 2462 }, { "epoch": 0.9318074340300766, "grad_norm": 0.9948219612555772, "learning_rate": 6.973090357547492e-06, "loss": 0.7363, "step": 2463 }, { "epoch": 0.9321857561713799, "grad_norm": 0.992557973151347, "learning_rate": 6.971910834827262e-06, "loss": 0.7694, "step": 2464 }, { "epoch": 0.9325640783126833, "grad_norm": 1.0094587028107398, "learning_rate": 6.9707307349708725e-06, "loss": 0.7427, "step": 2465 }, { "epoch": 0.9329424004539866, "grad_norm": 1.010298174001919, "learning_rate": 6.969550058207497e-06, "loss": 0.7599, "step": 2466 }, { "epoch": 0.9333207225952899, "grad_norm": 1.0032088715401652, "learning_rate": 6.968368804766418e-06, "loss": 0.7477, "step": 2467 }, { "epoch": 0.9336990447365933, "grad_norm": 1.0071202379716264, "learning_rate": 6.9671869748770335e-06, "loss": 0.7919, "step": 2468 }, { "epoch": 0.9340773668778966, "grad_norm": 1.0344244342248399, "learning_rate": 6.9660045687688505e-06, "loss": 0.7267, "step": 2469 }, { "epoch": 0.9344556890191998, "grad_norm": 1.035268996263067, "learning_rate": 6.964821586671487e-06, "loss": 0.7749, "step": 2470 }, { "epoch": 0.9348340111605031, "grad_norm": 0.9844668317387452, "learning_rate": 6.963638028814676e-06, "loss": 0.7384, "step": 2471 }, { "epoch": 0.9352123333018065, "grad_norm": 1.0353908950550423, "learning_rate": 6.9624538954282615e-06, "loss": 0.7707, "step": 2472 }, { "epoch": 0.9355906554431098, "grad_norm": 0.9744535950334421, "learning_rate": 6.961269186742198e-06, "loss": 0.7677, "step": 2473 }, { "epoch": 0.9359689775844131, "grad_norm": 0.9678869026043476, "learning_rate": 6.960083902986552e-06, "loss": 0.7356, "step": 2474 }, { "epoch": 0.9363472997257164, "grad_norm": 1.0015537336569855, "learning_rate": 6.958898044391503e-06, "loss": 0.7943, "step": 2475 }, { "epoch": 0.9367256218670198, "grad_norm": 1.0081891030967614, "learning_rate": 6.95771161118734e-06, "loss": 0.776, "step": 2476 }, { "epoch": 0.9371039440083231, "grad_norm": 1.0185677288406634, "learning_rate": 6.956524603604465e-06, "loss": 0.7033, "step": 2477 }, { "epoch": 0.9374822661496264, "grad_norm": 0.9879459481418753, "learning_rate": 6.955337021873391e-06, "loss": 0.7211, "step": 2478 }, { "epoch": 0.9378605882909298, "grad_norm": 1.0433490833526935, "learning_rate": 6.954148866224745e-06, "loss": 0.7636, "step": 2479 }, { "epoch": 0.9382389104322331, "grad_norm": 0.9972883410935348, "learning_rate": 6.952960136889261e-06, "loss": 0.768, "step": 2480 }, { "epoch": 0.9386172325735364, "grad_norm": 1.025294442567176, "learning_rate": 6.951770834097787e-06, "loss": 0.7445, "step": 2481 }, { "epoch": 0.9389955547148396, "grad_norm": 1.007913600666321, "learning_rate": 6.9505809580812836e-06, "loss": 0.749, "step": 2482 }, { "epoch": 0.939373876856143, "grad_norm": 1.0138278380784167, "learning_rate": 6.949390509070819e-06, "loss": 0.7565, "step": 2483 }, { "epoch": 0.9397521989974463, "grad_norm": 0.9656456748427612, "learning_rate": 6.948199487297575e-06, "loss": 0.7317, "step": 2484 }, { "epoch": 0.9401305211387496, "grad_norm": 1.0198475672197798, "learning_rate": 6.947007892992846e-06, "loss": 0.7532, "step": 2485 }, { "epoch": 0.940508843280053, "grad_norm": 0.9833135283489965, "learning_rate": 6.945815726388036e-06, "loss": 0.7374, "step": 2486 }, { "epoch": 0.9408871654213563, "grad_norm": 1.0277735201431621, "learning_rate": 6.944622987714659e-06, "loss": 0.7749, "step": 2487 }, { "epoch": 0.9412654875626596, "grad_norm": 0.9787162380579348, "learning_rate": 6.94342967720434e-06, "loss": 0.7542, "step": 2488 }, { "epoch": 0.9416438097039629, "grad_norm": 1.0147682759693886, "learning_rate": 6.94223579508882e-06, "loss": 0.7906, "step": 2489 }, { "epoch": 0.9420221318452663, "grad_norm": 1.0439307019996933, "learning_rate": 6.941041341599945e-06, "loss": 0.7673, "step": 2490 }, { "epoch": 0.9424004539865696, "grad_norm": 0.9810887443680518, "learning_rate": 6.939846316969675e-06, "loss": 0.7533, "step": 2491 }, { "epoch": 0.9427787761278729, "grad_norm": 0.9942535918520738, "learning_rate": 6.938650721430078e-06, "loss": 0.7394, "step": 2492 }, { "epoch": 0.9431570982691762, "grad_norm": 1.010817474142347, "learning_rate": 6.937454555213338e-06, "loss": 0.7677, "step": 2493 }, { "epoch": 0.9435354204104796, "grad_norm": 1.0015437245581833, "learning_rate": 6.9362578185517455e-06, "loss": 0.7292, "step": 2494 }, { "epoch": 0.9439137425517828, "grad_norm": 0.9682063645921414, "learning_rate": 6.935060511677704e-06, "loss": 0.727, "step": 2495 }, { "epoch": 0.9442920646930861, "grad_norm": 1.0122923799282402, "learning_rate": 6.9338626348237256e-06, "loss": 0.7578, "step": 2496 }, { "epoch": 0.9446703868343895, "grad_norm": 0.9983061913086282, "learning_rate": 6.932664188222435e-06, "loss": 0.7346, "step": 2497 }, { "epoch": 0.9450487089756928, "grad_norm": 1.005172892504285, "learning_rate": 6.931465172106567e-06, "loss": 0.7541, "step": 2498 }, { "epoch": 0.9454270311169961, "grad_norm": 1.0223316589568823, "learning_rate": 6.930265586708967e-06, "loss": 0.7639, "step": 2499 }, { "epoch": 0.9458053532582994, "grad_norm": 0.9855359648722788, "learning_rate": 6.9290654322625915e-06, "loss": 0.7134, "step": 2500 }, { "epoch": 0.9461836753996028, "grad_norm": 1.0483624433781076, "learning_rate": 6.927864709000506e-06, "loss": 0.7872, "step": 2501 }, { "epoch": 0.9465619975409061, "grad_norm": 0.9942036517451098, "learning_rate": 6.926663417155887e-06, "loss": 0.7326, "step": 2502 }, { "epoch": 0.9469403196822094, "grad_norm": 1.0319359016167458, "learning_rate": 6.9254615569620235e-06, "loss": 0.7813, "step": 2503 }, { "epoch": 0.9473186418235128, "grad_norm": 0.9619477887439621, "learning_rate": 6.92425912865231e-06, "loss": 0.7322, "step": 2504 }, { "epoch": 0.9476969639648161, "grad_norm": 1.0190096318322415, "learning_rate": 6.923056132460258e-06, "loss": 0.7769, "step": 2505 }, { "epoch": 0.9480752861061194, "grad_norm": 0.9756697967629381, "learning_rate": 6.921852568619483e-06, "loss": 0.7438, "step": 2506 }, { "epoch": 0.9484536082474226, "grad_norm": 1.003280277014329, "learning_rate": 6.9206484373637165e-06, "loss": 0.7453, "step": 2507 }, { "epoch": 0.948831930388726, "grad_norm": 1.0112490844068194, "learning_rate": 6.919443738926794e-06, "loss": 0.7803, "step": 2508 }, { "epoch": 0.9492102525300293, "grad_norm": 1.0306685166804053, "learning_rate": 6.9182384735426654e-06, "loss": 0.7802, "step": 2509 }, { "epoch": 0.9495885746713326, "grad_norm": 0.969869383344729, "learning_rate": 6.917032641445391e-06, "loss": 0.79, "step": 2510 }, { "epoch": 0.9499668968126359, "grad_norm": 0.9908621526334449, "learning_rate": 6.915826242869138e-06, "loss": 0.7506, "step": 2511 }, { "epoch": 0.9503452189539393, "grad_norm": 0.9709032809743106, "learning_rate": 6.914619278048185e-06, "loss": 0.7539, "step": 2512 }, { "epoch": 0.9507235410952426, "grad_norm": 0.9958274771264295, "learning_rate": 6.913411747216924e-06, "loss": 0.7462, "step": 2513 }, { "epoch": 0.9511018632365459, "grad_norm": 1.0015069209840703, "learning_rate": 6.912203650609851e-06, "loss": 0.7439, "step": 2514 }, { "epoch": 0.9514801853778493, "grad_norm": 0.9806871283872736, "learning_rate": 6.910994988461576e-06, "loss": 0.7331, "step": 2515 }, { "epoch": 0.9518585075191526, "grad_norm": 1.036302839942013, "learning_rate": 6.909785761006816e-06, "loss": 0.748, "step": 2516 }, { "epoch": 0.9522368296604559, "grad_norm": 0.9709785882663116, "learning_rate": 6.908575968480401e-06, "loss": 0.7554, "step": 2517 }, { "epoch": 0.9526151518017592, "grad_norm": 0.9541323356782881, "learning_rate": 6.907365611117269e-06, "loss": 0.7561, "step": 2518 }, { "epoch": 0.9529934739430626, "grad_norm": 0.9962944133799874, "learning_rate": 6.906154689152467e-06, "loss": 0.7159, "step": 2519 }, { "epoch": 0.9533717960843658, "grad_norm": 0.9788295892046283, "learning_rate": 6.904943202821153e-06, "loss": 0.746, "step": 2520 }, { "epoch": 0.9537501182256691, "grad_norm": 1.0377198409334003, "learning_rate": 6.903731152358593e-06, "loss": 0.766, "step": 2521 }, { "epoch": 0.9541284403669725, "grad_norm": 1.024423082132459, "learning_rate": 6.902518538000165e-06, "loss": 0.7647, "step": 2522 }, { "epoch": 0.9545067625082758, "grad_norm": 0.988327252141755, "learning_rate": 6.901305359981354e-06, "loss": 0.753, "step": 2523 }, { "epoch": 0.9548850846495791, "grad_norm": 1.0005349737717881, "learning_rate": 6.900091618537756e-06, "loss": 0.775, "step": 2524 }, { "epoch": 0.9552634067908824, "grad_norm": 1.0426641056319086, "learning_rate": 6.8988773139050745e-06, "loss": 0.7613, "step": 2525 }, { "epoch": 0.9556417289321858, "grad_norm": 0.9979157392015524, "learning_rate": 6.897662446319128e-06, "loss": 0.7299, "step": 2526 }, { "epoch": 0.9560200510734891, "grad_norm": 1.0522253646500737, "learning_rate": 6.8964470160158345e-06, "loss": 0.7861, "step": 2527 }, { "epoch": 0.9560200510734891, "eval_loss": 0.7629713416099548, "eval_runtime": 25.5791, "eval_samples_per_second": 34.599, "eval_steps_per_second": 1.095, "step": 2527 }, { "epoch": 0.9560200510734891, "eval_bench_accuracy_arc_challenge": 0.0, "eval_bench_accuracy_hellaswag": 0.21, "eval_bench_accuracy_mmlu": 0.19130434782608696, "eval_bench_average_accuracy": 0.13376811594202898, "eval_bench_loss": 8.016543739720396, "eval_bench_total_accuracy": 0.14065934065934066, "step": 2527 }, { "epoch": 0.9563983732147924, "grad_norm": 0.9719408998085209, "learning_rate": 6.895231023231231e-06, "loss": 0.7946, "step": 2528 }, { "epoch": 0.9567766953560957, "grad_norm": 0.9813191938097949, "learning_rate": 6.894014468201458e-06, "loss": 0.764, "step": 2529 }, { "epoch": 0.9571550174973991, "grad_norm": 1.010212028765319, "learning_rate": 6.892797351162768e-06, "loss": 0.7418, "step": 2530 }, { "epoch": 0.9575333396387024, "grad_norm": 1.0420773855711358, "learning_rate": 6.89157967235152e-06, "loss": 0.7921, "step": 2531 }, { "epoch": 0.9579116617800056, "grad_norm": 1.0456777816900407, "learning_rate": 6.8903614320041835e-06, "loss": 0.7838, "step": 2532 }, { "epoch": 0.958289983921309, "grad_norm": 1.0367052099970122, "learning_rate": 6.889142630357339e-06, "loss": 0.7521, "step": 2533 }, { "epoch": 0.9586683060626123, "grad_norm": 0.983548886380504, "learning_rate": 6.887923267647671e-06, "loss": 0.7953, "step": 2534 }, { "epoch": 0.9590466282039156, "grad_norm": 1.0131073035888476, "learning_rate": 6.88670334411198e-06, "loss": 0.7092, "step": 2535 }, { "epoch": 0.9594249503452189, "grad_norm": 0.9749156032738976, "learning_rate": 6.885482859987169e-06, "loss": 0.7731, "step": 2536 }, { "epoch": 0.9598032724865223, "grad_norm": 0.9831602779891812, "learning_rate": 6.884261815510253e-06, "loss": 0.7585, "step": 2537 }, { "epoch": 0.9601815946278256, "grad_norm": 0.976286974741624, "learning_rate": 6.883040210918356e-06, "loss": 0.7527, "step": 2538 }, { "epoch": 0.9605599167691289, "grad_norm": 1.0051856815206066, "learning_rate": 6.881818046448707e-06, "loss": 0.7875, "step": 2539 }, { "epoch": 0.9609382389104323, "grad_norm": 1.0268754278037553, "learning_rate": 6.880595322338649e-06, "loss": 0.6872, "step": 2540 }, { "epoch": 0.9613165610517356, "grad_norm": 1.0196169965505448, "learning_rate": 6.879372038825632e-06, "loss": 0.7848, "step": 2541 }, { "epoch": 0.9616948831930389, "grad_norm": 1.0277688828480922, "learning_rate": 6.878148196147211e-06, "loss": 0.7249, "step": 2542 }, { "epoch": 0.9620732053343422, "grad_norm": 1.014679951502152, "learning_rate": 6.876923794541057e-06, "loss": 0.7959, "step": 2543 }, { "epoch": 0.9624515274756456, "grad_norm": 1.0094052253264938, "learning_rate": 6.8756988342449415e-06, "loss": 0.7948, "step": 2544 }, { "epoch": 0.9628298496169488, "grad_norm": 1.0222030259917847, "learning_rate": 6.87447331549675e-06, "loss": 0.7467, "step": 2545 }, { "epoch": 0.9632081717582521, "grad_norm": 0.9490633472327878, "learning_rate": 6.873247238534473e-06, "loss": 0.7156, "step": 2546 }, { "epoch": 0.9635864938995554, "grad_norm": 0.9659043484413028, "learning_rate": 6.872020603596212e-06, "loss": 0.7341, "step": 2547 }, { "epoch": 0.9639648160408588, "grad_norm": 0.9858824739287257, "learning_rate": 6.870793410920175e-06, "loss": 0.7685, "step": 2548 }, { "epoch": 0.9643431381821621, "grad_norm": 0.9885841700355127, "learning_rate": 6.869565660744681e-06, "loss": 0.7629, "step": 2549 }, { "epoch": 0.9647214603234654, "grad_norm": 1.0306977685598762, "learning_rate": 6.868337353308153e-06, "loss": 0.7833, "step": 2550 }, { "epoch": 0.9650997824647688, "grad_norm": 1.0148687078689322, "learning_rate": 6.867108488849126e-06, "loss": 0.7642, "step": 2551 }, { "epoch": 0.9654781046060721, "grad_norm": 1.0112932767179983, "learning_rate": 6.865879067606243e-06, "loss": 0.7478, "step": 2552 }, { "epoch": 0.9658564267473754, "grad_norm": 0.9974100963832561, "learning_rate": 6.86464908981825e-06, "loss": 0.7628, "step": 2553 }, { "epoch": 0.9662347488886787, "grad_norm": 1.0415420686599979, "learning_rate": 6.86341855572401e-06, "loss": 0.7545, "step": 2554 }, { "epoch": 0.9666130710299821, "grad_norm": 0.9851459657811695, "learning_rate": 6.862187465562485e-06, "loss": 0.7359, "step": 2555 }, { "epoch": 0.9669913931712854, "grad_norm": 0.9873551970494169, "learning_rate": 6.86095581957275e-06, "loss": 0.7258, "step": 2556 }, { "epoch": 0.9673697153125886, "grad_norm": 1.0161870548595722, "learning_rate": 6.859723617993989e-06, "loss": 0.7714, "step": 2557 }, { "epoch": 0.967748037453892, "grad_norm": 0.9854492125883005, "learning_rate": 6.858490861065489e-06, "loss": 0.7764, "step": 2558 }, { "epoch": 0.9681263595951953, "grad_norm": 0.9909688995229221, "learning_rate": 6.857257549026649e-06, "loss": 0.7338, "step": 2559 }, { "epoch": 0.9685046817364986, "grad_norm": 1.0015511350207071, "learning_rate": 6.856023682116975e-06, "loss": 0.7351, "step": 2560 }, { "epoch": 0.9688830038778019, "grad_norm": 1.0306667770006128, "learning_rate": 6.854789260576079e-06, "loss": 0.7469, "step": 2561 }, { "epoch": 0.9692613260191053, "grad_norm": 1.014683694807874, "learning_rate": 6.853554284643684e-06, "loss": 0.767, "step": 2562 }, { "epoch": 0.9696396481604086, "grad_norm": 0.9988436672004211, "learning_rate": 6.8523187545596165e-06, "loss": 0.7309, "step": 2563 }, { "epoch": 0.9700179703017119, "grad_norm": 0.9893662673634439, "learning_rate": 6.8510826705638135e-06, "loss": 0.7239, "step": 2564 }, { "epoch": 0.9703962924430152, "grad_norm": 1.004789379171909, "learning_rate": 6.849846032896319e-06, "loss": 0.7639, "step": 2565 }, { "epoch": 0.9707746145843186, "grad_norm": 1.008734951569213, "learning_rate": 6.848608841797284e-06, "loss": 0.7376, "step": 2566 }, { "epoch": 0.9711529367256219, "grad_norm": 1.0067190655281641, "learning_rate": 6.847371097506967e-06, "loss": 0.7836, "step": 2567 }, { "epoch": 0.9715312588669252, "grad_norm": 1.0112440088877177, "learning_rate": 6.846132800265736e-06, "loss": 0.7604, "step": 2568 }, { "epoch": 0.9719095810082286, "grad_norm": 0.9702346714144766, "learning_rate": 6.844893950314063e-06, "loss": 0.7703, "step": 2569 }, { "epoch": 0.9722879031495318, "grad_norm": 0.963816870257285, "learning_rate": 6.8436545478925286e-06, "loss": 0.7616, "step": 2570 }, { "epoch": 0.9726662252908351, "grad_norm": 0.9740187939076729, "learning_rate": 6.842414593241821e-06, "loss": 0.732, "step": 2571 }, { "epoch": 0.9730445474321384, "grad_norm": 1.0185019066247203, "learning_rate": 6.841174086602737e-06, "loss": 0.707, "step": 2572 }, { "epoch": 0.9734228695734418, "grad_norm": 0.9769015104859489, "learning_rate": 6.8399330282161775e-06, "loss": 0.7524, "step": 2573 }, { "epoch": 0.9738011917147451, "grad_norm": 0.9935222297751732, "learning_rate": 6.838691418323155e-06, "loss": 0.7235, "step": 2574 }, { "epoch": 0.9741795138560484, "grad_norm": 0.9755675348566133, "learning_rate": 6.8374492571647846e-06, "loss": 0.7371, "step": 2575 }, { "epoch": 0.9745578359973518, "grad_norm": 1.0174788484875912, "learning_rate": 6.83620654498229e-06, "loss": 0.773, "step": 2576 }, { "epoch": 0.9749361581386551, "grad_norm": 0.9952747883195646, "learning_rate": 6.834963282017003e-06, "loss": 0.7432, "step": 2577 }, { "epoch": 0.9753144802799584, "grad_norm": 0.9818219715864143, "learning_rate": 6.8337194685103604e-06, "loss": 0.7096, "step": 2578 }, { "epoch": 0.9756928024212617, "grad_norm": 1.0982997866275739, "learning_rate": 6.832475104703908e-06, "loss": 0.7845, "step": 2579 }, { "epoch": 0.9760711245625651, "grad_norm": 0.9923536018305118, "learning_rate": 6.831230190839297e-06, "loss": 0.7294, "step": 2580 }, { "epoch": 0.9764494467038684, "grad_norm": 1.0161981092231336, "learning_rate": 6.829984727158288e-06, "loss": 0.7757, "step": 2581 }, { "epoch": 0.9768277688451716, "grad_norm": 0.9898052353491911, "learning_rate": 6.828738713902744e-06, "loss": 0.7376, "step": 2582 }, { "epoch": 0.9772060909864749, "grad_norm": 1.0339212935816466, "learning_rate": 6.827492151314637e-06, "loss": 0.7894, "step": 2583 }, { "epoch": 0.9775844131277783, "grad_norm": 0.99586876292627, "learning_rate": 6.826245039636045e-06, "loss": 0.7453, "step": 2584 }, { "epoch": 0.9779627352690816, "grad_norm": 0.9977838972337282, "learning_rate": 6.824997379109157e-06, "loss": 0.7476, "step": 2585 }, { "epoch": 0.9783410574103849, "grad_norm": 0.9807588599647941, "learning_rate": 6.823749169976262e-06, "loss": 0.7326, "step": 2586 }, { "epoch": 0.9787193795516883, "grad_norm": 0.9871054260982461, "learning_rate": 6.822500412479758e-06, "loss": 0.7834, "step": 2587 }, { "epoch": 0.9790977016929916, "grad_norm": 1.005353898928565, "learning_rate": 6.821251106862151e-06, "loss": 0.8029, "step": 2588 }, { "epoch": 0.9794760238342949, "grad_norm": 1.0007901496373823, "learning_rate": 6.820001253366054e-06, "loss": 0.7657, "step": 2589 }, { "epoch": 0.9798543459755982, "grad_norm": 0.961146391440765, "learning_rate": 6.8187508522341824e-06, "loss": 0.771, "step": 2590 }, { "epoch": 0.9802326681169016, "grad_norm": 1.0191182202751332, "learning_rate": 6.81749990370936e-06, "loss": 0.7499, "step": 2591 }, { "epoch": 0.9806109902582049, "grad_norm": 1.0492831958955406, "learning_rate": 6.8162484080345195e-06, "loss": 0.8166, "step": 2592 }, { "epoch": 0.9809893123995082, "grad_norm": 1.003747249368723, "learning_rate": 6.814996365452697e-06, "loss": 0.7708, "step": 2593 }, { "epoch": 0.9813676345408116, "grad_norm": 0.9995620996279818, "learning_rate": 6.813743776207033e-06, "loss": 0.7733, "step": 2594 }, { "epoch": 0.9817459566821148, "grad_norm": 1.0439456765568043, "learning_rate": 6.812490640540781e-06, "loss": 0.7573, "step": 2595 }, { "epoch": 0.9821242788234181, "grad_norm": 0.9625012629288878, "learning_rate": 6.811236958697292e-06, "loss": 0.7629, "step": 2596 }, { "epoch": 0.9825026009647214, "grad_norm": 0.9821346338663993, "learning_rate": 6.80998273092003e-06, "loss": 0.7633, "step": 2597 }, { "epoch": 0.9828809231060248, "grad_norm": 0.983402196881471, "learning_rate": 6.808727957452559e-06, "loss": 0.7614, "step": 2598 }, { "epoch": 0.9832592452473281, "grad_norm": 0.9582697389421767, "learning_rate": 6.807472638538557e-06, "loss": 0.7712, "step": 2599 }, { "epoch": 0.9836375673886314, "grad_norm": 0.9806751562256456, "learning_rate": 6.806216774421798e-06, "loss": 0.7719, "step": 2600 }, { "epoch": 0.9840158895299347, "grad_norm": 1.0014627507544995, "learning_rate": 6.804960365346172e-06, "loss": 0.7139, "step": 2601 }, { "epoch": 0.9843942116712381, "grad_norm": 1.0259899027499157, "learning_rate": 6.803703411555666e-06, "loss": 0.7326, "step": 2602 }, { "epoch": 0.9847725338125414, "grad_norm": 1.0170597423047276, "learning_rate": 6.802445913294379e-06, "loss": 0.7561, "step": 2603 }, { "epoch": 0.9851508559538447, "grad_norm": 1.0122881137848365, "learning_rate": 6.801187870806511e-06, "loss": 0.7709, "step": 2604 }, { "epoch": 0.9855291780951481, "grad_norm": 0.9411910936975904, "learning_rate": 6.7999292843363735e-06, "loss": 0.6868, "step": 2605 }, { "epoch": 0.9859075002364514, "grad_norm": 1.0040103823130304, "learning_rate": 6.798670154128378e-06, "loss": 0.7538, "step": 2606 }, { "epoch": 0.9862858223777546, "grad_norm": 1.081648156784344, "learning_rate": 6.797410480427043e-06, "loss": 0.7273, "step": 2607 }, { "epoch": 0.9866641445190579, "grad_norm": 0.9921108406273706, "learning_rate": 6.7961502634769955e-06, "loss": 0.7676, "step": 2608 }, { "epoch": 0.9870424666603613, "grad_norm": 0.9965895359346758, "learning_rate": 6.794889503522964e-06, "loss": 0.7194, "step": 2609 }, { "epoch": 0.9874207888016646, "grad_norm": 1.0090109120441377, "learning_rate": 6.793628200809785e-06, "loss": 0.7372, "step": 2610 }, { "epoch": 0.9877991109429679, "grad_norm": 0.9451434042232735, "learning_rate": 6.792366355582401e-06, "loss": 0.7743, "step": 2611 }, { "epoch": 0.9881774330842713, "grad_norm": 0.9880689330958867, "learning_rate": 6.791103968085856e-06, "loss": 0.7204, "step": 2612 }, { "epoch": 0.9885557552255746, "grad_norm": 0.9736956319515304, "learning_rate": 6.789841038565304e-06, "loss": 0.7636, "step": 2613 }, { "epoch": 0.9889340773668779, "grad_norm": 0.9848109619143756, "learning_rate": 6.788577567266001e-06, "loss": 0.7474, "step": 2614 }, { "epoch": 0.9893123995081812, "grad_norm": 1.001022251188608, "learning_rate": 6.78731355443331e-06, "loss": 0.7731, "step": 2615 }, { "epoch": 0.9896907216494846, "grad_norm": 0.9743148882263098, "learning_rate": 6.786049000312697e-06, "loss": 0.7397, "step": 2616 }, { "epoch": 0.9900690437907879, "grad_norm": 0.9841148128573903, "learning_rate": 6.784783905149737e-06, "loss": 0.7657, "step": 2617 }, { "epoch": 0.9904473659320912, "grad_norm": 1.0204564463659629, "learning_rate": 6.783518269190107e-06, "loss": 0.7563, "step": 2618 }, { "epoch": 0.9908256880733946, "grad_norm": 1.0167205459149449, "learning_rate": 6.782252092679588e-06, "loss": 0.7788, "step": 2619 }, { "epoch": 0.9912040102146978, "grad_norm": 1.0217942415012797, "learning_rate": 6.7809853758640684e-06, "loss": 0.7276, "step": 2620 }, { "epoch": 0.9915823323560011, "grad_norm": 1.014021497078255, "learning_rate": 6.779718118989542e-06, "loss": 0.7652, "step": 2621 }, { "epoch": 0.9919606544973044, "grad_norm": 0.9971205510444979, "learning_rate": 6.778450322302105e-06, "loss": 0.7496, "step": 2622 }, { "epoch": 0.9923389766386078, "grad_norm": 1.0410834620990588, "learning_rate": 6.7771819860479605e-06, "loss": 0.8007, "step": 2623 }, { "epoch": 0.9927172987799111, "grad_norm": 1.011348733055697, "learning_rate": 6.775913110473416e-06, "loss": 0.7896, "step": 2624 }, { "epoch": 0.9930956209212144, "grad_norm": 1.0305703452734365, "learning_rate": 6.774643695824883e-06, "loss": 0.7479, "step": 2625 }, { "epoch": 0.9934739430625177, "grad_norm": 1.0159355655493194, "learning_rate": 6.773373742348876e-06, "loss": 0.7385, "step": 2626 }, { "epoch": 0.9938522652038211, "grad_norm": 1.0233312299910446, "learning_rate": 6.7721032502920185e-06, "loss": 0.7784, "step": 2627 }, { "epoch": 0.9942305873451244, "grad_norm": 1.0243399462973277, "learning_rate": 6.770832219901036e-06, "loss": 0.7792, "step": 2628 }, { "epoch": 0.9946089094864277, "grad_norm": 1.0039264708461368, "learning_rate": 6.7695606514227576e-06, "loss": 0.7846, "step": 2629 }, { "epoch": 0.9949872316277311, "grad_norm": 0.9765424555517976, "learning_rate": 6.7682885451041185e-06, "loss": 0.7627, "step": 2630 }, { "epoch": 0.9953655537690344, "grad_norm": 0.9941382116530366, "learning_rate": 6.767015901192159e-06, "loss": 0.8234, "step": 2631 }, { "epoch": 0.9957438759103376, "grad_norm": 0.9983004803904528, "learning_rate": 6.7657427199340215e-06, "loss": 0.7517, "step": 2632 }, { "epoch": 0.9961221980516409, "grad_norm": 0.9958601494252852, "learning_rate": 6.764469001576955e-06, "loss": 0.7655, "step": 2633 }, { "epoch": 0.9965005201929443, "grad_norm": 1.049616042000357, "learning_rate": 6.763194746368311e-06, "loss": 0.8428, "step": 2634 }, { "epoch": 0.9968788423342476, "grad_norm": 1.0153634037350727, "learning_rate": 6.761919954555546e-06, "loss": 0.7547, "step": 2635 }, { "epoch": 0.9972571644755509, "grad_norm": 0.9678026635849788, "learning_rate": 6.76064462638622e-06, "loss": 0.7332, "step": 2636 }, { "epoch": 0.9976354866168543, "grad_norm": 0.9563515295105731, "learning_rate": 6.759368762108001e-06, "loss": 0.7502, "step": 2637 }, { "epoch": 0.9980138087581576, "grad_norm": 1.0439875845792919, "learning_rate": 6.758092361968655e-06, "loss": 0.7788, "step": 2638 }, { "epoch": 0.9983921308994609, "grad_norm": 0.9943166463669236, "learning_rate": 6.756815426216055e-06, "loss": 0.7305, "step": 2639 }, { "epoch": 0.9987704530407642, "grad_norm": 1.000562755460714, "learning_rate": 6.75553795509818e-06, "loss": 0.7474, "step": 2640 }, { "epoch": 0.9991487751820676, "grad_norm": 0.9789926629525856, "learning_rate": 6.7542599488631095e-06, "loss": 0.7191, "step": 2641 }, { "epoch": 0.9995270973233709, "grad_norm": 0.9997150620047266, "learning_rate": 6.75298140775903e-06, "loss": 0.7475, "step": 2642 }, { "epoch": 0.9999054194646742, "grad_norm": 0.9978723880114577, "learning_rate": 6.751702332034229e-06, "loss": 0.7285, "step": 2643 }, { "epoch": 1.0002837416059775, "grad_norm": 1.019988669909004, "learning_rate": 6.750422721937099e-06, "loss": 0.7441, "step": 2644 }, { "epoch": 1.0006620637472807, "grad_norm": 0.9993844123830328, "learning_rate": 6.7491425777161385e-06, "loss": 0.7277, "step": 2645 }, { "epoch": 1.0010403858885841, "grad_norm": 1.037289949707294, "learning_rate": 6.7478618996199444e-06, "loss": 0.744, "step": 2646 }, { "epoch": 1.0014187080298875, "grad_norm": 0.9642837824860211, "learning_rate": 6.746580687897223e-06, "loss": 0.7671, "step": 2647 }, { "epoch": 1.0017970301711907, "grad_norm": 1.0378638874808468, "learning_rate": 6.745298942796783e-06, "loss": 0.7711, "step": 2648 }, { "epoch": 1.0003783221413034, "grad_norm": 1.2960225263466278, "learning_rate": 6.744016664567532e-06, "loss": 1.0291, "step": 2649 }, { "epoch": 1.0007566442826066, "grad_norm": 1.6045538345200387, "learning_rate": 6.742733853458485e-06, "loss": 0.6262, "step": 2650 }, { "epoch": 1.00113496642391, "grad_norm": 1.434253850313174, "learning_rate": 6.741450509718761e-06, "loss": 0.62, "step": 2651 }, { "epoch": 1.0015132885652134, "grad_norm": 1.1008479018702804, "learning_rate": 6.740166633597583e-06, "loss": 0.5954, "step": 2652 }, { "epoch": 1.0018916107065166, "grad_norm": 1.3680593625954045, "learning_rate": 6.738882225344276e-06, "loss": 0.6455, "step": 2653 }, { "epoch": 1.00226993284782, "grad_norm": 1.6880769312041324, "learning_rate": 6.737597285208265e-06, "loss": 0.6093, "step": 2654 }, { "epoch": 1.0026482549891231, "grad_norm": 1.638055896899775, "learning_rate": 6.736311813439084e-06, "loss": 0.5981, "step": 2655 }, { "epoch": 1.0030265771304265, "grad_norm": 1.311686250422591, "learning_rate": 6.735025810286366e-06, "loss": 0.597, "step": 2656 }, { "epoch": 1.00340489927173, "grad_norm": 1.2466705587608804, "learning_rate": 6.73373927599985e-06, "loss": 0.5987, "step": 2657 }, { "epoch": 1.003783221413033, "grad_norm": 1.1771167828974005, "learning_rate": 6.732452210829378e-06, "loss": 0.61, "step": 2658 }, { "epoch": 1.0041615435543365, "grad_norm": 1.1134868919849727, "learning_rate": 6.731164615024893e-06, "loss": 0.5931, "step": 2659 }, { "epoch": 1.00453986569564, "grad_norm": 1.1298337066049222, "learning_rate": 6.729876488836443e-06, "loss": 0.5689, "step": 2660 }, { "epoch": 1.00453986569564, "eval_loss": 0.7680820226669312, "eval_runtime": 22.7008, "eval_samples_per_second": 38.985, "eval_steps_per_second": 1.233, "step": 2660 }, { "epoch": 1.00453986569564, "eval_bench_accuracy_arc_challenge": 0.0, "eval_bench_accuracy_hellaswag": 0.215, "eval_bench_accuracy_mmlu": 0.24347826086956523, "eval_bench_average_accuracy": 0.15282608695652175, "eval_bench_loss": 8.457613225568805, "eval_bench_total_accuracy": 0.15604395604395604, "step": 2660 }, { "epoch": 1.004918187836943, "grad_norm": 1.1143554969459137, "learning_rate": 6.728587832514177e-06, "loss": 0.5962, "step": 2661 }, { "epoch": 1.0052965099782465, "grad_norm": 1.1243429820420148, "learning_rate": 6.72729864630835e-06, "loss": 0.603, "step": 2662 }, { "epoch": 1.0056748321195499, "grad_norm": 1.1486817603312562, "learning_rate": 6.726008930469316e-06, "loss": 0.6311, "step": 2663 }, { "epoch": 1.006053154260853, "grad_norm": 1.1595678038663924, "learning_rate": 6.724718685247536e-06, "loss": 0.6305, "step": 2664 }, { "epoch": 1.0064314764021565, "grad_norm": 1.1574946842049172, "learning_rate": 6.7234279108935695e-06, "loss": 0.5891, "step": 2665 }, { "epoch": 1.0068097985434599, "grad_norm": 1.1580865412334664, "learning_rate": 6.7221366076580835e-06, "loss": 0.6021, "step": 2666 }, { "epoch": 1.007188120684763, "grad_norm": 1.1239458896873717, "learning_rate": 6.720844775791843e-06, "loss": 0.5624, "step": 2667 }, { "epoch": 1.0075664428260664, "grad_norm": 1.109348251252506, "learning_rate": 6.719552415545719e-06, "loss": 0.5937, "step": 2668 }, { "epoch": 1.0079447649673696, "grad_norm": 1.0743163318050015, "learning_rate": 6.718259527170685e-06, "loss": 0.6224, "step": 2669 }, { "epoch": 1.008323087108673, "grad_norm": 1.0654956200008854, "learning_rate": 6.716966110917814e-06, "loss": 0.6153, "step": 2670 }, { "epoch": 1.0087014092499764, "grad_norm": 1.1034062304953003, "learning_rate": 6.7156721670382845e-06, "loss": 0.6339, "step": 2671 }, { "epoch": 1.0090797313912796, "grad_norm": 1.094323227986904, "learning_rate": 6.714377695783376e-06, "loss": 0.5888, "step": 2672 }, { "epoch": 1.009458053532583, "grad_norm": 1.060387202863176, "learning_rate": 6.713082697404471e-06, "loss": 0.5821, "step": 2673 }, { "epoch": 1.0098363756738864, "grad_norm": 1.0528419193778633, "learning_rate": 6.711787172153055e-06, "loss": 0.5799, "step": 2674 }, { "epoch": 1.0102146978151896, "grad_norm": 1.0796156745606953, "learning_rate": 6.710491120280715e-06, "loss": 0.595, "step": 2675 }, { "epoch": 1.010593019956493, "grad_norm": 1.0710249844294826, "learning_rate": 6.7091945420391405e-06, "loss": 0.6002, "step": 2676 }, { "epoch": 1.0109713420977964, "grad_norm": 1.064042009582335, "learning_rate": 6.707897437680122e-06, "loss": 0.5725, "step": 2677 }, { "epoch": 1.0113496642390996, "grad_norm": 1.070124967363786, "learning_rate": 6.706599807455556e-06, "loss": 0.6183, "step": 2678 }, { "epoch": 1.011727986380403, "grad_norm": 1.0612174775790282, "learning_rate": 6.705301651617434e-06, "loss": 0.6024, "step": 2679 }, { "epoch": 1.0121063085217061, "grad_norm": 1.1401140132534746, "learning_rate": 6.704002970417857e-06, "loss": 0.6188, "step": 2680 }, { "epoch": 1.0124846306630095, "grad_norm": 1.0463107536511211, "learning_rate": 6.702703764109024e-06, "loss": 0.5775, "step": 2681 }, { "epoch": 1.012862952804313, "grad_norm": 1.0990093291612235, "learning_rate": 6.701404032943237e-06, "loss": 0.6326, "step": 2682 }, { "epoch": 1.013241274945616, "grad_norm": 1.0884271445142926, "learning_rate": 6.700103777172902e-06, "loss": 0.5965, "step": 2683 }, { "epoch": 1.0136195970869195, "grad_norm": 1.100271092818893, "learning_rate": 6.698802997050522e-06, "loss": 0.5893, "step": 2684 }, { "epoch": 1.013997919228223, "grad_norm": 1.08645469114748, "learning_rate": 6.697501692828705e-06, "loss": 0.657, "step": 2685 }, { "epoch": 1.014376241369526, "grad_norm": 1.1084364337638348, "learning_rate": 6.696199864760162e-06, "loss": 0.5858, "step": 2686 }, { "epoch": 1.0147545635108295, "grad_norm": 1.0643950719372635, "learning_rate": 6.694897513097702e-06, "loss": 0.5932, "step": 2687 }, { "epoch": 1.0151328856521329, "grad_norm": 1.0772292275045006, "learning_rate": 6.693594638094239e-06, "loss": 0.5954, "step": 2688 }, { "epoch": 1.015511207793436, "grad_norm": 1.1460007039310949, "learning_rate": 6.692291240002788e-06, "loss": 0.635, "step": 2689 }, { "epoch": 1.0158895299347395, "grad_norm": 1.0833748168296655, "learning_rate": 6.6909873190764644e-06, "loss": 0.6011, "step": 2690 }, { "epoch": 1.0162678520760426, "grad_norm": 1.0151560868995457, "learning_rate": 6.689682875568485e-06, "loss": 0.6106, "step": 2691 }, { "epoch": 1.016646174217346, "grad_norm": 1.0633233343271067, "learning_rate": 6.688377909732169e-06, "loss": 0.6153, "step": 2692 }, { "epoch": 1.0170244963586494, "grad_norm": 1.0729102331474902, "learning_rate": 6.687072421820937e-06, "loss": 0.591, "step": 2693 }, { "epoch": 1.0174028184999526, "grad_norm": 1.027376177863638, "learning_rate": 6.685766412088312e-06, "loss": 0.5804, "step": 2694 }, { "epoch": 1.017781140641256, "grad_norm": 1.0881918331010754, "learning_rate": 6.684459880787915e-06, "loss": 0.5929, "step": 2695 }, { "epoch": 1.0181594627825594, "grad_norm": 1.0773718471003868, "learning_rate": 6.6831528281734726e-06, "loss": 0.5751, "step": 2696 }, { "epoch": 1.0185377849238626, "grad_norm": 1.0731412200657173, "learning_rate": 6.681845254498809e-06, "loss": 0.6228, "step": 2697 }, { "epoch": 1.018916107065166, "grad_norm": 1.085723469308112, "learning_rate": 6.680537160017852e-06, "loss": 0.5825, "step": 2698 }, { "epoch": 1.0192944292064694, "grad_norm": 1.069544360919215, "learning_rate": 6.67922854498463e-06, "loss": 0.5978, "step": 2699 }, { "epoch": 1.0196727513477726, "grad_norm": 1.074673299315061, "learning_rate": 6.67791940965327e-06, "loss": 0.5945, "step": 2700 }, { "epoch": 1.020051073489076, "grad_norm": 1.1365428490420595, "learning_rate": 6.676609754278004e-06, "loss": 0.603, "step": 2701 }, { "epoch": 1.0204293956303794, "grad_norm": 1.1038996629682, "learning_rate": 6.675299579113163e-06, "loss": 0.5841, "step": 2702 }, { "epoch": 1.0208077177716826, "grad_norm": 1.048251282008455, "learning_rate": 6.673988884413178e-06, "loss": 0.5865, "step": 2703 }, { "epoch": 1.021186039912986, "grad_norm": 1.0992286608098654, "learning_rate": 6.672677670432584e-06, "loss": 0.6298, "step": 2704 }, { "epoch": 1.0215643620542891, "grad_norm": 1.0627840105571968, "learning_rate": 6.671365937426013e-06, "loss": 0.5763, "step": 2705 }, { "epoch": 1.0219426841955925, "grad_norm": 1.0819101685440222, "learning_rate": 6.670053685648201e-06, "loss": 0.6267, "step": 2706 }, { "epoch": 1.022321006336896, "grad_norm": 1.0991732623026937, "learning_rate": 6.668740915353981e-06, "loss": 0.616, "step": 2707 }, { "epoch": 1.022699328478199, "grad_norm": 1.0662924149150412, "learning_rate": 6.66742762679829e-06, "loss": 0.6111, "step": 2708 }, { "epoch": 1.0230776506195025, "grad_norm": 1.1578701050290159, "learning_rate": 6.6661138202361665e-06, "loss": 0.5838, "step": 2709 }, { "epoch": 1.023455972760806, "grad_norm": 1.0557307528642985, "learning_rate": 6.664799495922746e-06, "loss": 0.6024, "step": 2710 }, { "epoch": 1.023834294902109, "grad_norm": 1.0697654686744051, "learning_rate": 6.663484654113266e-06, "loss": 0.6372, "step": 2711 }, { "epoch": 1.0242126170434125, "grad_norm": 1.0894837082943014, "learning_rate": 6.662169295063068e-06, "loss": 0.5576, "step": 2712 }, { "epoch": 1.0245909391847159, "grad_norm": 1.1746471691072748, "learning_rate": 6.660853419027588e-06, "loss": 0.6634, "step": 2713 }, { "epoch": 1.024969261326019, "grad_norm": 1.1333166533866583, "learning_rate": 6.659537026262364e-06, "loss": 0.6248, "step": 2714 }, { "epoch": 1.0253475834673225, "grad_norm": 1.1167798031199665, "learning_rate": 6.658220117023038e-06, "loss": 0.6241, "step": 2715 }, { "epoch": 1.0257259056086256, "grad_norm": 1.1422831671384859, "learning_rate": 6.656902691565349e-06, "loss": 0.5891, "step": 2716 }, { "epoch": 1.026104227749929, "grad_norm": 1.1198346720452148, "learning_rate": 6.655584750145137e-06, "loss": 0.6088, "step": 2717 }, { "epoch": 1.0264825498912324, "grad_norm": 1.0477422285686226, "learning_rate": 6.654266293018342e-06, "loss": 0.5915, "step": 2718 }, { "epoch": 1.0268608720325356, "grad_norm": 1.093902006549265, "learning_rate": 6.652947320441006e-06, "loss": 0.6109, "step": 2719 }, { "epoch": 1.027239194173839, "grad_norm": 1.093919628889555, "learning_rate": 6.651627832669267e-06, "loss": 0.593, "step": 2720 }, { "epoch": 1.0276175163151424, "grad_norm": 1.1162752285980133, "learning_rate": 6.6503078299593665e-06, "loss": 0.584, "step": 2721 }, { "epoch": 1.0279958384564456, "grad_norm": 1.06965206133156, "learning_rate": 6.648987312567646e-06, "loss": 0.6592, "step": 2722 }, { "epoch": 1.028374160597749, "grad_norm": 1.0883837174849054, "learning_rate": 6.647666280750545e-06, "loss": 0.6032, "step": 2723 }, { "epoch": 1.0287524827390524, "grad_norm": 1.1074197481547494, "learning_rate": 6.646344734764606e-06, "loss": 0.6242, "step": 2724 }, { "epoch": 1.0291308048803556, "grad_norm": 1.0999734415758928, "learning_rate": 6.645022674866465e-06, "loss": 0.6111, "step": 2725 }, { "epoch": 1.029509127021659, "grad_norm": 1.0943109626585719, "learning_rate": 6.643700101312866e-06, "loss": 0.5804, "step": 2726 }, { "epoch": 1.0298874491629622, "grad_norm": 1.0724232041327344, "learning_rate": 6.642377014360647e-06, "loss": 0.6169, "step": 2727 }, { "epoch": 1.0302657713042656, "grad_norm": 1.11016667382968, "learning_rate": 6.641053414266748e-06, "loss": 0.5934, "step": 2728 }, { "epoch": 1.030644093445569, "grad_norm": 1.105192777962378, "learning_rate": 6.639729301288209e-06, "loss": 0.6439, "step": 2729 }, { "epoch": 1.0310224155868721, "grad_norm": 1.0673633356752081, "learning_rate": 6.638404675682167e-06, "loss": 0.5892, "step": 2730 }, { "epoch": 1.0314007377281755, "grad_norm": 1.0680562475087432, "learning_rate": 6.6370795377058615e-06, "loss": 0.5952, "step": 2731 }, { "epoch": 1.031779059869479, "grad_norm": 1.0738981447230416, "learning_rate": 6.635753887616629e-06, "loss": 0.584, "step": 2732 }, { "epoch": 1.032157382010782, "grad_norm": 1.0785882851048183, "learning_rate": 6.634427725671909e-06, "loss": 0.6374, "step": 2733 }, { "epoch": 1.0325357041520855, "grad_norm": 1.076443272017146, "learning_rate": 6.633101052129236e-06, "loss": 0.5839, "step": 2734 }, { "epoch": 1.032914026293389, "grad_norm": 1.0603789845495335, "learning_rate": 6.631773867246247e-06, "loss": 0.588, "step": 2735 }, { "epoch": 1.033292348434692, "grad_norm": 1.08184770989303, "learning_rate": 6.630446171280678e-06, "loss": 0.6008, "step": 2736 }, { "epoch": 1.0336706705759955, "grad_norm": 1.0967987837787552, "learning_rate": 6.629117964490363e-06, "loss": 0.6084, "step": 2737 }, { "epoch": 1.0340489927172989, "grad_norm": 1.0929847738762193, "learning_rate": 6.627789247133236e-06, "loss": 0.6204, "step": 2738 }, { "epoch": 1.034427314858602, "grad_norm": 1.1124337620049105, "learning_rate": 6.626460019467327e-06, "loss": 0.5917, "step": 2739 }, { "epoch": 1.0348056369999055, "grad_norm": 1.0608012348242233, "learning_rate": 6.625130281750772e-06, "loss": 0.6204, "step": 2740 }, { "epoch": 1.0351839591412086, "grad_norm": 1.0827756336779857, "learning_rate": 6.6238000342418016e-06, "loss": 0.6136, "step": 2741 }, { "epoch": 1.035562281282512, "grad_norm": 1.1403236667837964, "learning_rate": 6.6224692771987435e-06, "loss": 0.5852, "step": 2742 }, { "epoch": 1.0359406034238154, "grad_norm": 1.1342145147069507, "learning_rate": 6.621138010880029e-06, "loss": 0.6074, "step": 2743 }, { "epoch": 1.0363189255651186, "grad_norm": 1.1532502566585765, "learning_rate": 6.619806235544184e-06, "loss": 0.622, "step": 2744 }, { "epoch": 1.036697247706422, "grad_norm": 1.075042839911617, "learning_rate": 6.6184739514498375e-06, "loss": 0.5793, "step": 2745 }, { "epoch": 1.0370755698477254, "grad_norm": 1.0972773225112782, "learning_rate": 6.6171411588557135e-06, "loss": 0.6571, "step": 2746 }, { "epoch": 1.0374538919890286, "grad_norm": 1.0648521041187424, "learning_rate": 6.615807858020637e-06, "loss": 0.5813, "step": 2747 }, { "epoch": 1.037832214130332, "grad_norm": 1.1281029569879641, "learning_rate": 6.614474049203531e-06, "loss": 0.6232, "step": 2748 }, { "epoch": 1.0382105362716354, "grad_norm": 1.1139023350580175, "learning_rate": 6.6131397326634165e-06, "loss": 0.6231, "step": 2749 }, { "epoch": 1.0385888584129386, "grad_norm": 1.087661969678301, "learning_rate": 6.611804908659414e-06, "loss": 0.6145, "step": 2750 }, { "epoch": 1.038967180554242, "grad_norm": 1.050267017698137, "learning_rate": 6.610469577450743e-06, "loss": 0.6187, "step": 2751 }, { "epoch": 1.0393455026955452, "grad_norm": 1.1133488798260815, "learning_rate": 6.6091337392967195e-06, "loss": 0.5907, "step": 2752 }, { "epoch": 1.0397238248368486, "grad_norm": 1.141462547663806, "learning_rate": 6.607797394456761e-06, "loss": 0.6146, "step": 2753 }, { "epoch": 1.040102146978152, "grad_norm": 1.1350155018431922, "learning_rate": 6.606460543190381e-06, "loss": 0.6055, "step": 2754 }, { "epoch": 1.0404804691194551, "grad_norm": 1.133690939941688, "learning_rate": 6.605123185757192e-06, "loss": 0.5792, "step": 2755 }, { "epoch": 1.0408587912607585, "grad_norm": 1.1504040033841838, "learning_rate": 6.603785322416902e-06, "loss": 0.6257, "step": 2756 }, { "epoch": 1.041237113402062, "grad_norm": 1.1018282110182764, "learning_rate": 6.602446953429325e-06, "loss": 0.6226, "step": 2757 }, { "epoch": 1.041615435543365, "grad_norm": 1.0973286890712581, "learning_rate": 6.601108079054366e-06, "loss": 0.5848, "step": 2758 }, { "epoch": 1.0419937576846685, "grad_norm": 1.1239040489828953, "learning_rate": 6.599768699552029e-06, "loss": 0.5958, "step": 2759 }, { "epoch": 1.042372079825972, "grad_norm": 1.0917807839762754, "learning_rate": 6.598428815182419e-06, "loss": 0.6014, "step": 2760 }, { "epoch": 1.042750401967275, "grad_norm": 1.0712800925576116, "learning_rate": 6.5970884262057384e-06, "loss": 0.5823, "step": 2761 }, { "epoch": 1.0431287241085785, "grad_norm": 1.079612565596154, "learning_rate": 6.595747532882284e-06, "loss": 0.5936, "step": 2762 }, { "epoch": 1.0435070462498817, "grad_norm": 1.1368744893133789, "learning_rate": 6.594406135472455e-06, "loss": 0.6005, "step": 2763 }, { "epoch": 1.043885368391185, "grad_norm": 1.1284525435968846, "learning_rate": 6.593064234236747e-06, "loss": 0.6194, "step": 2764 }, { "epoch": 1.0442636905324885, "grad_norm": 1.0918438286119356, "learning_rate": 6.591721829435753e-06, "loss": 0.6074, "step": 2765 }, { "epoch": 1.0446420126737916, "grad_norm": 1.1728746338404048, "learning_rate": 6.590378921330163e-06, "loss": 0.635, "step": 2766 }, { "epoch": 1.045020334815095, "grad_norm": 1.1387975427757375, "learning_rate": 6.589035510180766e-06, "loss": 0.6032, "step": 2767 }, { "epoch": 1.0453986569563984, "grad_norm": 1.0912484299886718, "learning_rate": 6.587691596248451e-06, "loss": 0.5956, "step": 2768 }, { "epoch": 1.0457769790977016, "grad_norm": 1.0934051938921485, "learning_rate": 6.586347179794198e-06, "loss": 0.567, "step": 2769 }, { "epoch": 1.046155301239005, "grad_norm": 1.0838741876834017, "learning_rate": 6.585002261079091e-06, "loss": 0.5922, "step": 2770 }, { "epoch": 1.0465336233803084, "grad_norm": 1.1169061017052488, "learning_rate": 6.583656840364309e-06, "loss": 0.6005, "step": 2771 }, { "epoch": 1.0469119455216116, "grad_norm": 1.0546019024427187, "learning_rate": 6.582310917911128e-06, "loss": 0.6003, "step": 2772 }, { "epoch": 1.047290267662915, "grad_norm": 1.1140364513664793, "learning_rate": 6.580964493980923e-06, "loss": 0.5871, "step": 2773 }, { "epoch": 1.0476685898042184, "grad_norm": 1.5651374874541824, "learning_rate": 6.579617568835163e-06, "loss": 0.5884, "step": 2774 }, { "epoch": 1.0480469119455216, "grad_norm": 1.1080717837472673, "learning_rate": 6.578270142735422e-06, "loss": 0.5962, "step": 2775 }, { "epoch": 1.048425234086825, "grad_norm": 1.0863735483605719, "learning_rate": 6.57692221594336e-06, "loss": 0.6323, "step": 2776 }, { "epoch": 1.0488035562281282, "grad_norm": 1.1674382392240956, "learning_rate": 6.575573788720744e-06, "loss": 0.6026, "step": 2777 }, { "epoch": 1.0491818783694316, "grad_norm": 1.1153179736567085, "learning_rate": 6.574224861329434e-06, "loss": 0.636, "step": 2778 }, { "epoch": 1.049560200510735, "grad_norm": 1.0912285783470717, "learning_rate": 6.572875434031388e-06, "loss": 0.6069, "step": 2779 }, { "epoch": 1.0499385226520381, "grad_norm": 1.1343269303316939, "learning_rate": 6.57152550708866e-06, "loss": 0.6166, "step": 2780 }, { "epoch": 1.0503168447933415, "grad_norm": 1.0940804068025969, "learning_rate": 6.5701750807634e-06, "loss": 0.6085, "step": 2781 }, { "epoch": 1.050695166934645, "grad_norm": 1.1170676036050409, "learning_rate": 6.56882415531786e-06, "loss": 0.6083, "step": 2782 }, { "epoch": 1.051073489075948, "grad_norm": 1.072773895384744, "learning_rate": 6.567472731014385e-06, "loss": 0.5814, "step": 2783 }, { "epoch": 1.0514518112172515, "grad_norm": 1.0859363709331855, "learning_rate": 6.566120808115416e-06, "loss": 0.589, "step": 2784 }, { "epoch": 1.051830133358555, "grad_norm": 1.0636139805625147, "learning_rate": 6.564768386883493e-06, "loss": 0.6261, "step": 2785 }, { "epoch": 1.052208455499858, "grad_norm": 1.0592048162256658, "learning_rate": 6.563415467581253e-06, "loss": 0.5695, "step": 2786 }, { "epoch": 1.0525867776411615, "grad_norm": 1.1045618236850399, "learning_rate": 6.562062050471427e-06, "loss": 0.5858, "step": 2787 }, { "epoch": 1.0529650997824647, "grad_norm": 1.0798015164090737, "learning_rate": 6.560708135816846e-06, "loss": 0.6144, "step": 2788 }, { "epoch": 1.053343421923768, "grad_norm": 1.1212612042998356, "learning_rate": 6.559353723880436e-06, "loss": 0.6137, "step": 2789 }, { "epoch": 1.0537217440650715, "grad_norm": 1.085472033827373, "learning_rate": 6.55799881492522e-06, "loss": 0.5787, "step": 2790 }, { "epoch": 1.0541000662063746, "grad_norm": 1.0771593566721889, "learning_rate": 6.5566434092143166e-06, "loss": 0.5978, "step": 2791 }, { "epoch": 1.054478388347678, "grad_norm": 1.0918802997077126, "learning_rate": 6.555287507010941e-06, "loss": 0.6472, "step": 2792 }, { "epoch": 1.0548567104889814, "grad_norm": 1.123015558169342, "learning_rate": 6.5539311085784064e-06, "loss": 0.6069, "step": 2793 }, { "epoch": 1.0548567104889814, "eval_loss": 0.7674793601036072, "eval_runtime": 22.9197, "eval_samples_per_second": 38.613, "eval_steps_per_second": 1.222, "step": 2793 }, { "epoch": 1.0548567104889814, "eval_bench_accuracy_arc_challenge": 0.10714285714285714, "eval_bench_accuracy_hellaswag": 0.225, "eval_bench_accuracy_mmlu": 0.25217391304347825, "eval_bench_average_accuracy": 0.1947722567287785, "eval_bench_loss": 8.895098769873904, "eval_bench_total_accuracy": 0.1956043956043956, "step": 2793 }, { "epoch": 1.0552350326302846, "grad_norm": 1.063121911475721, "learning_rate": 6.552574214180122e-06, "loss": 0.6165, "step": 2794 }, { "epoch": 1.055613354771588, "grad_norm": 1.136690231965498, "learning_rate": 6.551216824079591e-06, "loss": 0.5886, "step": 2795 }, { "epoch": 1.0559916769128914, "grad_norm": 1.0926208179674248, "learning_rate": 6.549858938540415e-06, "loss": 0.6059, "step": 2796 }, { "epoch": 1.0563699990541946, "grad_norm": 1.159280762617995, "learning_rate": 6.548500557826292e-06, "loss": 0.6049, "step": 2797 }, { "epoch": 1.056748321195498, "grad_norm": 1.0987424681046232, "learning_rate": 6.547141682201013e-06, "loss": 0.6176, "step": 2798 }, { "epoch": 1.0571266433368014, "grad_norm": 1.1195326409944681, "learning_rate": 6.545782311928471e-06, "loss": 0.6069, "step": 2799 }, { "epoch": 1.0575049654781046, "grad_norm": 1.1122477760961818, "learning_rate": 6.544422447272651e-06, "loss": 0.6123, "step": 2800 }, { "epoch": 1.057883287619408, "grad_norm": 1.0805239012593884, "learning_rate": 6.543062088497632e-06, "loss": 0.5603, "step": 2801 }, { "epoch": 1.0582616097607112, "grad_norm": 1.0742528596313945, "learning_rate": 6.541701235867594e-06, "loss": 0.5939, "step": 2802 }, { "epoch": 1.0586399319020146, "grad_norm": 1.0701044494027945, "learning_rate": 6.540339889646809e-06, "loss": 0.5881, "step": 2803 }, { "epoch": 1.059018254043318, "grad_norm": 1.1192253410768613, "learning_rate": 6.538978050099648e-06, "loss": 0.6029, "step": 2804 }, { "epoch": 1.0593965761846211, "grad_norm": 1.1399119586355322, "learning_rate": 6.5376157174905736e-06, "loss": 0.5836, "step": 2805 }, { "epoch": 1.0597748983259245, "grad_norm": 1.053554887782157, "learning_rate": 6.5362528920841495e-06, "loss": 0.6202, "step": 2806 }, { "epoch": 1.060153220467228, "grad_norm": 1.1006415862921535, "learning_rate": 6.534889574145031e-06, "loss": 0.5928, "step": 2807 }, { "epoch": 1.060531542608531, "grad_norm": 1.093418359680058, "learning_rate": 6.533525763937971e-06, "loss": 0.613, "step": 2808 }, { "epoch": 1.0609098647498345, "grad_norm": 1.2168317509722493, "learning_rate": 6.532161461727817e-06, "loss": 0.6087, "step": 2809 }, { "epoch": 1.061288186891138, "grad_norm": 1.1325852875949003, "learning_rate": 6.530796667779512e-06, "loss": 0.589, "step": 2810 }, { "epoch": 1.061666509032441, "grad_norm": 1.0794237445540322, "learning_rate": 6.529431382358095e-06, "loss": 0.6167, "step": 2811 }, { "epoch": 1.0620448311737445, "grad_norm": 1.1740194043276255, "learning_rate": 6.5280656057287e-06, "loss": 0.5906, "step": 2812 }, { "epoch": 1.0624231533150477, "grad_norm": 1.1066949269479105, "learning_rate": 6.5266993381565576e-06, "loss": 0.6096, "step": 2813 }, { "epoch": 1.062801475456351, "grad_norm": 1.0874460595681357, "learning_rate": 6.5253325799069924e-06, "loss": 0.6225, "step": 2814 }, { "epoch": 1.0631797975976545, "grad_norm": 1.1418971085661331, "learning_rate": 6.523965331245424e-06, "loss": 0.6276, "step": 2815 }, { "epoch": 1.0635581197389576, "grad_norm": 1.0564960638972678, "learning_rate": 6.5225975924373695e-06, "loss": 0.5886, "step": 2816 }, { "epoch": 1.063936441880261, "grad_norm": 1.1430774469204301, "learning_rate": 6.521229363748439e-06, "loss": 0.6437, "step": 2817 }, { "epoch": 1.0643147640215644, "grad_norm": 1.1253217138489495, "learning_rate": 6.519860645444339e-06, "loss": 0.6118, "step": 2818 }, { "epoch": 1.0646930861628676, "grad_norm": 1.1031052720391512, "learning_rate": 6.518491437790869e-06, "loss": 0.5948, "step": 2819 }, { "epoch": 1.065071408304171, "grad_norm": 1.1152719824533943, "learning_rate": 6.517121741053925e-06, "loss": 0.6083, "step": 2820 }, { "epoch": 1.0654497304454744, "grad_norm": 1.064206367043127, "learning_rate": 6.5157515554995005e-06, "loss": 0.6205, "step": 2821 }, { "epoch": 1.0658280525867776, "grad_norm": 1.1233058992847154, "learning_rate": 6.514380881393678e-06, "loss": 0.6173, "step": 2822 }, { "epoch": 1.066206374728081, "grad_norm": 1.1002282020103138, "learning_rate": 6.5130097190026406e-06, "loss": 0.6006, "step": 2823 }, { "epoch": 1.0665846968693842, "grad_norm": 1.1433147577521796, "learning_rate": 6.511638068592664e-06, "loss": 0.6041, "step": 2824 }, { "epoch": 1.0669630190106876, "grad_norm": 1.1059521945155957, "learning_rate": 6.510265930430118e-06, "loss": 0.6323, "step": 2825 }, { "epoch": 1.067341341151991, "grad_norm": 1.097061975167435, "learning_rate": 6.508893304781467e-06, "loss": 0.6024, "step": 2826 }, { "epoch": 1.0677196632932942, "grad_norm": 1.1229207491603603, "learning_rate": 6.507520191913271e-06, "loss": 0.6051, "step": 2827 }, { "epoch": 1.0680979854345976, "grad_norm": 1.1170341905034227, "learning_rate": 6.506146592092186e-06, "loss": 0.6114, "step": 2828 }, { "epoch": 1.068476307575901, "grad_norm": 1.0868572793587328, "learning_rate": 6.50477250558496e-06, "loss": 0.6127, "step": 2829 }, { "epoch": 1.0688546297172041, "grad_norm": 1.0989916723272581, "learning_rate": 6.503397932658434e-06, "loss": 0.6019, "step": 2830 }, { "epoch": 1.0692329518585075, "grad_norm": 1.0787911069795304, "learning_rate": 6.50202287357955e-06, "loss": 0.6155, "step": 2831 }, { "epoch": 1.069611273999811, "grad_norm": 1.1069922383175101, "learning_rate": 6.500647328615339e-06, "loss": 0.5872, "step": 2832 }, { "epoch": 1.069989596141114, "grad_norm": 1.1314723796920492, "learning_rate": 6.499271298032926e-06, "loss": 0.624, "step": 2833 }, { "epoch": 1.0703679182824175, "grad_norm": 1.0679301946389819, "learning_rate": 6.497894782099534e-06, "loss": 0.5662, "step": 2834 }, { "epoch": 1.0707462404237207, "grad_norm": 1.0576329002755303, "learning_rate": 6.496517781082478e-06, "loss": 0.5931, "step": 2835 }, { "epoch": 1.071124562565024, "grad_norm": 1.1234936057539915, "learning_rate": 6.495140295249165e-06, "loss": 0.6089, "step": 2836 }, { "epoch": 1.0715028847063275, "grad_norm": 1.0994258336299476, "learning_rate": 6.493762324867102e-06, "loss": 0.6253, "step": 2837 }, { "epoch": 1.0718812068476307, "grad_norm": 1.075300022809444, "learning_rate": 6.492383870203885e-06, "loss": 0.6181, "step": 2838 }, { "epoch": 1.072259528988934, "grad_norm": 1.104060875724737, "learning_rate": 6.4910049315272056e-06, "loss": 0.6222, "step": 2839 }, { "epoch": 1.0726378511302375, "grad_norm": 1.0772894357800407, "learning_rate": 6.489625509104851e-06, "loss": 0.6232, "step": 2840 }, { "epoch": 1.0730161732715406, "grad_norm": 1.0739648027353192, "learning_rate": 6.488245603204699e-06, "loss": 0.6247, "step": 2841 }, { "epoch": 1.073394495412844, "grad_norm": 1.1304543775628406, "learning_rate": 6.486865214094724e-06, "loss": 0.6069, "step": 2842 }, { "epoch": 1.0737728175541474, "grad_norm": 1.1088974959408873, "learning_rate": 6.485484342042994e-06, "loss": 0.6085, "step": 2843 }, { "epoch": 1.0741511396954506, "grad_norm": 1.1163280282207435, "learning_rate": 6.484102987317669e-06, "loss": 0.6281, "step": 2844 }, { "epoch": 1.074529461836754, "grad_norm": 1.1049862073566041, "learning_rate": 6.482721150187005e-06, "loss": 0.5752, "step": 2845 }, { "epoch": 1.0749077839780574, "grad_norm": 1.0841279650014217, "learning_rate": 6.4813388309193515e-06, "loss": 0.5929, "step": 2846 }, { "epoch": 1.0752861061193606, "grad_norm": 1.1191141791395105, "learning_rate": 6.4799560297831475e-06, "loss": 0.6159, "step": 2847 }, { "epoch": 1.075664428260664, "grad_norm": 1.1483205014925426, "learning_rate": 6.478572747046932e-06, "loss": 0.6162, "step": 2848 }, { "epoch": 1.0760427504019672, "grad_norm": 1.1172546945739006, "learning_rate": 6.477188982979333e-06, "loss": 0.5945, "step": 2849 }, { "epoch": 1.0764210725432706, "grad_norm": 1.094078560906285, "learning_rate": 6.475804737849074e-06, "loss": 0.5861, "step": 2850 }, { "epoch": 1.076799394684574, "grad_norm": 1.113717456313015, "learning_rate": 6.474420011924969e-06, "loss": 0.6061, "step": 2851 }, { "epoch": 1.0771777168258772, "grad_norm": 1.0850333571769746, "learning_rate": 6.473034805475931e-06, "loss": 0.629, "step": 2852 }, { "epoch": 1.0775560389671806, "grad_norm": 1.0615304053220176, "learning_rate": 6.471649118770961e-06, "loss": 0.6097, "step": 2853 }, { "epoch": 1.077934361108484, "grad_norm": 1.1085699883011384, "learning_rate": 6.470262952079155e-06, "loss": 0.6356, "step": 2854 }, { "epoch": 1.0783126832497871, "grad_norm": 1.0974176053166769, "learning_rate": 6.468876305669703e-06, "loss": 0.5754, "step": 2855 }, { "epoch": 1.0786910053910905, "grad_norm": 1.094707593335946, "learning_rate": 6.467489179811888e-06, "loss": 0.5876, "step": 2856 }, { "epoch": 1.079069327532394, "grad_norm": 1.0793722208975378, "learning_rate": 6.466101574775085e-06, "loss": 0.5997, "step": 2857 }, { "epoch": 1.079447649673697, "grad_norm": 1.0968219817022344, "learning_rate": 6.464713490828762e-06, "loss": 0.5762, "step": 2858 }, { "epoch": 1.0798259718150005, "grad_norm": 1.1193087668072932, "learning_rate": 6.463324928242483e-06, "loss": 0.6406, "step": 2859 }, { "epoch": 1.080204293956304, "grad_norm": 1.158553138092932, "learning_rate": 6.4619358872859e-06, "loss": 0.601, "step": 2860 }, { "epoch": 1.080582616097607, "grad_norm": 1.0560307084665936, "learning_rate": 6.460546368228763e-06, "loss": 0.5876, "step": 2861 }, { "epoch": 1.0809609382389105, "grad_norm": 1.1280386212424045, "learning_rate": 6.45915637134091e-06, "loss": 0.608, "step": 2862 }, { "epoch": 1.0813392603802137, "grad_norm": 1.0574302798873645, "learning_rate": 6.4577658968922766e-06, "loss": 0.5791, "step": 2863 }, { "epoch": 1.081717582521517, "grad_norm": 1.1150273327727904, "learning_rate": 6.4563749451528875e-06, "loss": 0.6502, "step": 2864 }, { "epoch": 1.0820959046628205, "grad_norm": 1.1367661212026843, "learning_rate": 6.454983516392861e-06, "loss": 0.6063, "step": 2865 }, { "epoch": 1.0824742268041236, "grad_norm": 1.0456913210857686, "learning_rate": 6.4535916108824095e-06, "loss": 0.6346, "step": 2866 }, { "epoch": 1.082852548945427, "grad_norm": 1.2392345666078972, "learning_rate": 6.452199228891837e-06, "loss": 0.5899, "step": 2867 }, { "epoch": 1.0832308710867304, "grad_norm": 1.1039834206366064, "learning_rate": 6.450806370691537e-06, "loss": 0.6013, "step": 2868 }, { "epoch": 1.0836091932280336, "grad_norm": 1.1134513667317554, "learning_rate": 6.449413036552002e-06, "loss": 0.5994, "step": 2869 }, { "epoch": 1.083987515369337, "grad_norm": 1.1650004504353773, "learning_rate": 6.448019226743813e-06, "loss": 0.619, "step": 2870 }, { "epoch": 1.0843658375106404, "grad_norm": 1.0985700628831452, "learning_rate": 6.446624941537641e-06, "loss": 0.5808, "step": 2871 }, { "epoch": 1.0847441596519436, "grad_norm": 1.0514771993012972, "learning_rate": 6.445230181204253e-06, "loss": 0.5855, "step": 2872 }, { "epoch": 1.085122481793247, "grad_norm": 1.1550714439771754, "learning_rate": 6.443834946014509e-06, "loss": 0.6238, "step": 2873 }, { "epoch": 1.0855008039345502, "grad_norm": 1.1110870400800863, "learning_rate": 6.442439236239358e-06, "loss": 0.5789, "step": 2874 }, { "epoch": 1.0858791260758536, "grad_norm": 1.1097000978334268, "learning_rate": 6.441043052149843e-06, "loss": 0.6015, "step": 2875 }, { "epoch": 1.086257448217157, "grad_norm": 1.090750930465743, "learning_rate": 6.439646394017098e-06, "loss": 0.5965, "step": 2876 }, { "epoch": 1.0866357703584602, "grad_norm": 1.095919604598907, "learning_rate": 6.438249262112352e-06, "loss": 0.5681, "step": 2877 }, { "epoch": 1.0870140924997636, "grad_norm": 1.1536923323428168, "learning_rate": 6.43685165670692e-06, "loss": 0.5974, "step": 2878 }, { "epoch": 1.087392414641067, "grad_norm": 1.341838622128703, "learning_rate": 6.435453578072218e-06, "loss": 0.5753, "step": 2879 }, { "epoch": 1.0877707367823701, "grad_norm": 1.1893056531865611, "learning_rate": 6.4340550264797434e-06, "loss": 0.6217, "step": 2880 }, { "epoch": 1.0881490589236735, "grad_norm": 1.1625714412381882, "learning_rate": 6.432656002201094e-06, "loss": 0.623, "step": 2881 }, { "epoch": 1.088527381064977, "grad_norm": 1.135829315043605, "learning_rate": 6.431256505507956e-06, "loss": 0.5915, "step": 2882 }, { "epoch": 1.08890570320628, "grad_norm": 1.1067881135088327, "learning_rate": 6.4298565366721045e-06, "loss": 0.5984, "step": 2883 }, { "epoch": 1.0892840253475835, "grad_norm": 1.1198136681481627, "learning_rate": 6.4284560959654135e-06, "loss": 0.5755, "step": 2884 }, { "epoch": 1.0896623474888867, "grad_norm": 1.094534253436391, "learning_rate": 6.427055183659842e-06, "loss": 0.6325, "step": 2885 }, { "epoch": 1.09004066963019, "grad_norm": 1.6094939269599173, "learning_rate": 6.4256538000274425e-06, "loss": 0.6042, "step": 2886 }, { "epoch": 1.0904189917714935, "grad_norm": 1.1731472259427178, "learning_rate": 6.424251945340361e-06, "loss": 0.6198, "step": 2887 }, { "epoch": 1.0907973139127967, "grad_norm": 1.1461591379945404, "learning_rate": 6.422849619870833e-06, "loss": 0.6106, "step": 2888 }, { "epoch": 1.0911756360541, "grad_norm": 1.1040806331194917, "learning_rate": 6.421446823891185e-06, "loss": 0.5985, "step": 2889 }, { "epoch": 1.0915539581954035, "grad_norm": 1.1119449829415362, "learning_rate": 6.420043557673836e-06, "loss": 0.6461, "step": 2890 }, { "epoch": 1.0919322803367066, "grad_norm": 1.111485207605467, "learning_rate": 6.418639821491297e-06, "loss": 0.5846, "step": 2891 }, { "epoch": 1.09231060247801, "grad_norm": 1.0891225578274688, "learning_rate": 6.417235615616169e-06, "loss": 0.6111, "step": 2892 }, { "epoch": 1.0926889246193134, "grad_norm": 1.1181638114751344, "learning_rate": 6.415830940321143e-06, "loss": 0.5924, "step": 2893 }, { "epoch": 1.0930672467606166, "grad_norm": 1.2299453659498227, "learning_rate": 6.4144257958790055e-06, "loss": 0.5982, "step": 2894 }, { "epoch": 1.09344556890192, "grad_norm": 1.4493877022892196, "learning_rate": 6.413020182562629e-06, "loss": 0.5741, "step": 2895 }, { "epoch": 1.0938238910432232, "grad_norm": 1.1403500438029623, "learning_rate": 6.411614100644982e-06, "loss": 0.6221, "step": 2896 }, { "epoch": 1.0942022131845266, "grad_norm": 1.0998618983813124, "learning_rate": 6.410207550399117e-06, "loss": 0.5934, "step": 2897 }, { "epoch": 1.09458053532583, "grad_norm": 1.1243476132816683, "learning_rate": 6.4088005320981865e-06, "loss": 0.6299, "step": 2898 }, { "epoch": 1.0949588574671332, "grad_norm": 1.1183757598217452, "learning_rate": 6.407393046015428e-06, "loss": 0.6024, "step": 2899 }, { "epoch": 1.0953371796084366, "grad_norm": 1.0925616555798907, "learning_rate": 6.4059850924241686e-06, "loss": 0.5805, "step": 2900 }, { "epoch": 1.09571550174974, "grad_norm": 1.087553481219626, "learning_rate": 6.404576671597832e-06, "loss": 0.5972, "step": 2901 }, { "epoch": 1.0960938238910432, "grad_norm": 1.0406285601726264, "learning_rate": 6.403167783809927e-06, "loss": 0.5849, "step": 2902 }, { "epoch": 1.0964721460323466, "grad_norm": 1.8490666179905808, "learning_rate": 6.4017584293340555e-06, "loss": 0.6362, "step": 2903 }, { "epoch": 1.09685046817365, "grad_norm": 1.0948130651820664, "learning_rate": 6.400348608443909e-06, "loss": 0.594, "step": 2904 }, { "epoch": 1.0972287903149531, "grad_norm": 1.1059993754335438, "learning_rate": 6.398938321413274e-06, "loss": 0.589, "step": 2905 }, { "epoch": 1.0976071124562565, "grad_norm": 1.0776585328385655, "learning_rate": 6.397527568516023e-06, "loss": 0.5661, "step": 2906 }, { "epoch": 1.0979854345975597, "grad_norm": 1.1869940580514498, "learning_rate": 6.396116350026117e-06, "loss": 0.6045, "step": 2907 }, { "epoch": 1.098363756738863, "grad_norm": 1.1370907337901461, "learning_rate": 6.3947046662176135e-06, "loss": 0.6115, "step": 2908 }, { "epoch": 1.0987420788801665, "grad_norm": 1.1294426725345343, "learning_rate": 6.393292517364655e-06, "loss": 0.6074, "step": 2909 }, { "epoch": 1.0991204010214697, "grad_norm": 1.0604438875382727, "learning_rate": 6.3918799037414785e-06, "loss": 0.5706, "step": 2910 }, { "epoch": 1.099498723162773, "grad_norm": 1.0672456659536043, "learning_rate": 6.390466825622408e-06, "loss": 0.6002, "step": 2911 }, { "epoch": 1.0998770453040765, "grad_norm": 1.0422813585571513, "learning_rate": 6.389053283281858e-06, "loss": 0.6114, "step": 2912 }, { "epoch": 1.1002553674453797, "grad_norm": 1.1536181098968183, "learning_rate": 6.387639276994338e-06, "loss": 0.5908, "step": 2913 }, { "epoch": 1.100633689586683, "grad_norm": 1.0392572222921632, "learning_rate": 6.386224807034441e-06, "loss": 0.609, "step": 2914 }, { "epoch": 1.1010120117279865, "grad_norm": 1.1007188679082354, "learning_rate": 6.384809873676853e-06, "loss": 0.5723, "step": 2915 }, { "epoch": 1.1013903338692896, "grad_norm": 1.0886888704359527, "learning_rate": 6.38339447719635e-06, "loss": 0.5999, "step": 2916 }, { "epoch": 1.101768656010593, "grad_norm": 1.0538117291652995, "learning_rate": 6.381978617867798e-06, "loss": 0.5827, "step": 2917 }, { "epoch": 1.1021469781518964, "grad_norm": 1.1022840703948513, "learning_rate": 6.380562295966152e-06, "loss": 0.5885, "step": 2918 }, { "epoch": 1.1025253002931996, "grad_norm": 1.088713478811435, "learning_rate": 6.379145511766457e-06, "loss": 0.6272, "step": 2919 }, { "epoch": 1.102903622434503, "grad_norm": 1.0906087022105642, "learning_rate": 6.377728265543852e-06, "loss": 0.6089, "step": 2920 }, { "epoch": 1.1032819445758064, "grad_norm": 1.1163729261688122, "learning_rate": 6.376310557573557e-06, "loss": 0.5917, "step": 2921 }, { "epoch": 1.1036602667171096, "grad_norm": 1.0921235163203418, "learning_rate": 6.37489238813089e-06, "loss": 0.6058, "step": 2922 }, { "epoch": 1.104038588858413, "grad_norm": 1.1154149500807253, "learning_rate": 6.3734737574912525e-06, "loss": 0.6252, "step": 2923 }, { "epoch": 1.1044169109997162, "grad_norm": 1.121759678056962, "learning_rate": 6.372054665930141e-06, "loss": 0.5907, "step": 2924 }, { "epoch": 1.1047952331410196, "grad_norm": 1.0804882702031746, "learning_rate": 6.370635113723137e-06, "loss": 0.5639, "step": 2925 }, { "epoch": 1.105173555282323, "grad_norm": 1.1249030144235395, "learning_rate": 6.369215101145913e-06, "loss": 0.5926, "step": 2926 }, { "epoch": 1.105173555282323, "eval_loss": 0.7699943780899048, "eval_runtime": 22.7329, "eval_samples_per_second": 38.93, "eval_steps_per_second": 1.232, "step": 2926 }, { "epoch": 1.105173555282323, "eval_bench_accuracy_arc_challenge": 0.0, "eval_bench_accuracy_hellaswag": 0.225, "eval_bench_accuracy_mmlu": 0.23478260869565218, "eval_bench_average_accuracy": 0.1532608695652174, "eval_bench_loss": 8.313174063699288, "eval_bench_total_accuracy": 0.15824175824175823, "step": 2926 }, { "epoch": 1.1055518774236262, "grad_norm": 1.1694671343010261, "learning_rate": 6.367794628474234e-06, "loss": 0.5917, "step": 2927 }, { "epoch": 1.1059301995649296, "grad_norm": 1.1033049688490255, "learning_rate": 6.366373695983949e-06, "loss": 0.5905, "step": 2928 }, { "epoch": 1.106308521706233, "grad_norm": 1.095040817737354, "learning_rate": 6.364952303950998e-06, "loss": 0.603, "step": 2929 }, { "epoch": 1.1066868438475361, "grad_norm": 1.1009885796031318, "learning_rate": 6.363530452651414e-06, "loss": 0.5664, "step": 2930 }, { "epoch": 1.1070651659888395, "grad_norm": 1.0885986691913176, "learning_rate": 6.362108142361314e-06, "loss": 0.5978, "step": 2931 }, { "epoch": 1.107443488130143, "grad_norm": 1.112935862107649, "learning_rate": 6.360685373356908e-06, "loss": 0.589, "step": 2932 }, { "epoch": 1.107821810271446, "grad_norm": 1.0638084218303834, "learning_rate": 6.359262145914492e-06, "loss": 0.6006, "step": 2933 }, { "epoch": 1.1082001324127495, "grad_norm": 1.076804094844068, "learning_rate": 6.357838460310453e-06, "loss": 0.6028, "step": 2934 }, { "epoch": 1.1085784545540527, "grad_norm": 1.0557582456043746, "learning_rate": 6.356414316821267e-06, "loss": 0.5776, "step": 2935 }, { "epoch": 1.108956776695356, "grad_norm": 1.1100016393285486, "learning_rate": 6.354989715723497e-06, "loss": 0.5593, "step": 2936 }, { "epoch": 1.1093350988366595, "grad_norm": 1.0742404921148858, "learning_rate": 6.353564657293798e-06, "loss": 0.6232, "step": 2937 }, { "epoch": 1.1097134209779627, "grad_norm": 1.088260462687826, "learning_rate": 6.352139141808911e-06, "loss": 0.5909, "step": 2938 }, { "epoch": 1.110091743119266, "grad_norm": 1.0733185812324364, "learning_rate": 6.350713169545667e-06, "loss": 0.6047, "step": 2939 }, { "epoch": 1.1104700652605695, "grad_norm": 1.0633085341561832, "learning_rate": 6.349286740780986e-06, "loss": 0.5677, "step": 2940 }, { "epoch": 1.1108483874018726, "grad_norm": 1.0818850000245366, "learning_rate": 6.3478598557918746e-06, "loss": 0.5787, "step": 2941 }, { "epoch": 1.111226709543176, "grad_norm": 1.0561523005435232, "learning_rate": 6.346432514855433e-06, "loss": 0.582, "step": 2942 }, { "epoch": 1.1116050316844794, "grad_norm": 1.1060615731805643, "learning_rate": 6.345004718248842e-06, "loss": 0.6343, "step": 2943 }, { "epoch": 1.1119833538257826, "grad_norm": 1.1585516423096434, "learning_rate": 6.343576466249379e-06, "loss": 0.6502, "step": 2944 }, { "epoch": 1.112361675967086, "grad_norm": 1.1568429114309136, "learning_rate": 6.342147759134404e-06, "loss": 0.6068, "step": 2945 }, { "epoch": 1.1127399981083892, "grad_norm": 1.1317020260149575, "learning_rate": 6.340718597181369e-06, "loss": 0.6181, "step": 2946 }, { "epoch": 1.1131183202496926, "grad_norm": 1.1399537664077894, "learning_rate": 6.339288980667813e-06, "loss": 0.6045, "step": 2947 }, { "epoch": 1.113496642390996, "grad_norm": 1.0766148804522018, "learning_rate": 6.337858909871363e-06, "loss": 0.6124, "step": 2948 }, { "epoch": 1.1138749645322992, "grad_norm": 1.0707807388024961, "learning_rate": 6.336428385069733e-06, "loss": 0.5915, "step": 2949 }, { "epoch": 1.1142532866736026, "grad_norm": 1.1450188453384489, "learning_rate": 6.3349974065407285e-06, "loss": 0.5857, "step": 2950 }, { "epoch": 1.114631608814906, "grad_norm": 1.136001690059522, "learning_rate": 6.33356597456224e-06, "loss": 0.6367, "step": 2951 }, { "epoch": 1.1150099309562091, "grad_norm": 1.0967112071824558, "learning_rate": 6.3321340894122495e-06, "loss": 0.5943, "step": 2952 }, { "epoch": 1.1153882530975125, "grad_norm": 1.1063302432913777, "learning_rate": 6.330701751368822e-06, "loss": 0.6328, "step": 2953 }, { "epoch": 1.115766575238816, "grad_norm": 1.0832237975219734, "learning_rate": 6.329268960710115e-06, "loss": 0.5975, "step": 2954 }, { "epoch": 1.1161448973801191, "grad_norm": 1.1158970628739584, "learning_rate": 6.32783571771437e-06, "loss": 0.5802, "step": 2955 }, { "epoch": 1.1165232195214225, "grad_norm": 1.0941982609931546, "learning_rate": 6.3264020226599226e-06, "loss": 0.6089, "step": 2956 }, { "epoch": 1.1169015416627257, "grad_norm": 1.1193990404740914, "learning_rate": 6.324967875825187e-06, "loss": 0.6057, "step": 2957 }, { "epoch": 1.117279863804029, "grad_norm": 1.1095722096911433, "learning_rate": 6.3235332774886745e-06, "loss": 0.627, "step": 2958 }, { "epoch": 1.1176581859453325, "grad_norm": 1.0981923812352097, "learning_rate": 6.322098227928977e-06, "loss": 0.5965, "step": 2959 }, { "epoch": 1.1180365080866357, "grad_norm": 1.1184126406881407, "learning_rate": 6.320662727424778e-06, "loss": 0.5781, "step": 2960 }, { "epoch": 1.118414830227939, "grad_norm": 1.1151186971248008, "learning_rate": 6.319226776254847e-06, "loss": 0.6248, "step": 2961 }, { "epoch": 1.1187931523692425, "grad_norm": 1.0857075532409588, "learning_rate": 6.317790374698043e-06, "loss": 0.6092, "step": 2962 }, { "epoch": 1.1191714745105457, "grad_norm": 1.0845470846309295, "learning_rate": 6.316353523033309e-06, "loss": 0.5982, "step": 2963 }, { "epoch": 1.119549796651849, "grad_norm": 1.0945263927259892, "learning_rate": 6.3149162215396775e-06, "loss": 0.6168, "step": 2964 }, { "epoch": 1.1199281187931525, "grad_norm": 1.119534561570255, "learning_rate": 6.313478470496267e-06, "loss": 0.5973, "step": 2965 }, { "epoch": 1.1203064409344556, "grad_norm": 1.1136486829327252, "learning_rate": 6.312040270182289e-06, "loss": 0.6142, "step": 2966 }, { "epoch": 1.120684763075759, "grad_norm": 1.1571952059634305, "learning_rate": 6.310601620877031e-06, "loss": 0.6239, "step": 2967 }, { "epoch": 1.1210630852170622, "grad_norm": 1.0835967440445073, "learning_rate": 6.30916252285988e-06, "loss": 0.5821, "step": 2968 }, { "epoch": 1.1214414073583656, "grad_norm": 1.047175519560597, "learning_rate": 6.307722976410302e-06, "loss": 0.6168, "step": 2969 }, { "epoch": 1.121819729499669, "grad_norm": 1.0577983534766697, "learning_rate": 6.306282981807853e-06, "loss": 0.6358, "step": 2970 }, { "epoch": 1.1221980516409722, "grad_norm": 1.168613093646754, "learning_rate": 6.3048425393321746e-06, "loss": 0.6365, "step": 2971 }, { "epoch": 1.1225763737822756, "grad_norm": 1.0786034102911601, "learning_rate": 6.3034016492629995e-06, "loss": 0.6354, "step": 2972 }, { "epoch": 1.122954695923579, "grad_norm": 1.111989141081951, "learning_rate": 6.301960311880141e-06, "loss": 0.5689, "step": 2973 }, { "epoch": 1.1233330180648822, "grad_norm": 1.1074939643939372, "learning_rate": 6.300518527463502e-06, "loss": 0.6143, "step": 2974 }, { "epoch": 1.1237113402061856, "grad_norm": 1.1439421554238876, "learning_rate": 6.299076296293078e-06, "loss": 0.6369, "step": 2975 }, { "epoch": 1.124089662347489, "grad_norm": 1.173720413674196, "learning_rate": 6.297633618648939e-06, "loss": 0.5669, "step": 2976 }, { "epoch": 1.1244679844887921, "grad_norm": 1.0734992198920015, "learning_rate": 6.296190494811254e-06, "loss": 0.6059, "step": 2977 }, { "epoch": 1.1248463066300955, "grad_norm": 1.083993195598927, "learning_rate": 6.29474692506027e-06, "loss": 0.5804, "step": 2978 }, { "epoch": 1.1252246287713987, "grad_norm": 1.687682770941875, "learning_rate": 6.293302909676326e-06, "loss": 0.5925, "step": 2979 }, { "epoch": 1.1256029509127021, "grad_norm": 1.1466832381838077, "learning_rate": 6.291858448939845e-06, "loss": 0.6149, "step": 2980 }, { "epoch": 1.1259812730540055, "grad_norm": 1.1561377459749906, "learning_rate": 6.2904135431313355e-06, "loss": 0.6137, "step": 2981 }, { "epoch": 1.126359595195309, "grad_norm": 1.118710750426949, "learning_rate": 6.2889681925313955e-06, "loss": 0.6363, "step": 2982 }, { "epoch": 1.126737917336612, "grad_norm": 1.1112565724316716, "learning_rate": 6.287522397420707e-06, "loss": 0.5811, "step": 2983 }, { "epoch": 1.1271162394779155, "grad_norm": 1.0776033340786777, "learning_rate": 6.2860761580800395e-06, "loss": 0.5664, "step": 2984 }, { "epoch": 1.1274945616192187, "grad_norm": 1.112244696617646, "learning_rate": 6.284629474790249e-06, "loss": 0.6194, "step": 2985 }, { "epoch": 1.127872883760522, "grad_norm": 1.1163608284842275, "learning_rate": 6.283182347832275e-06, "loss": 0.6127, "step": 2986 }, { "epoch": 1.1282512059018255, "grad_norm": 1.1038870052883005, "learning_rate": 6.281734777487146e-06, "loss": 0.5945, "step": 2987 }, { "epoch": 1.1286295280431287, "grad_norm": 1.1576044510196872, "learning_rate": 6.2802867640359765e-06, "loss": 0.5826, "step": 2988 }, { "epoch": 1.129007850184432, "grad_norm": 1.0759586162897274, "learning_rate": 6.2788383077599665e-06, "loss": 0.5647, "step": 2989 }, { "epoch": 1.1293861723257355, "grad_norm": 1.10006956440662, "learning_rate": 6.277389408940401e-06, "loss": 0.6034, "step": 2990 }, { "epoch": 1.1297644944670386, "grad_norm": 1.1153385637875122, "learning_rate": 6.275940067858652e-06, "loss": 0.5733, "step": 2991 }, { "epoch": 1.130142816608342, "grad_norm": 1.1637742336834638, "learning_rate": 6.2744902847961785e-06, "loss": 0.6388, "step": 2992 }, { "epoch": 1.1305211387496454, "grad_norm": 1.071387857503032, "learning_rate": 6.2730400600345225e-06, "loss": 0.5961, "step": 2993 }, { "epoch": 1.1308994608909486, "grad_norm": 1.093134676797188, "learning_rate": 6.271589393855313e-06, "loss": 0.5631, "step": 2994 }, { "epoch": 1.131277783032252, "grad_norm": 1.1378266739181366, "learning_rate": 6.270138286540266e-06, "loss": 0.6105, "step": 2995 }, { "epoch": 1.1316561051735552, "grad_norm": 1.095271581280312, "learning_rate": 6.2686867383711815e-06, "loss": 0.5777, "step": 2996 }, { "epoch": 1.1320344273148586, "grad_norm": 1.1401342065141504, "learning_rate": 6.267234749629947e-06, "loss": 0.5903, "step": 2997 }, { "epoch": 1.132412749456162, "grad_norm": 1.074323035052679, "learning_rate": 6.265782320598534e-06, "loss": 0.5889, "step": 2998 }, { "epoch": 1.1327910715974652, "grad_norm": 1.1195059679487418, "learning_rate": 6.264329451558998e-06, "loss": 0.6224, "step": 2999 }, { "epoch": 1.1331693937387686, "grad_norm": 1.139005811272008, "learning_rate": 6.262876142793483e-06, "loss": 0.653, "step": 3000 }, { "epoch": 1.133547715880072, "grad_norm": 1.1473595969566768, "learning_rate": 6.2614223945842185e-06, "loss": 0.6227, "step": 3001 }, { "epoch": 1.1339260380213751, "grad_norm": 1.148699042472204, "learning_rate": 6.259968207213518e-06, "loss": 0.6193, "step": 3002 }, { "epoch": 1.1343043601626785, "grad_norm": 1.1238519792084694, "learning_rate": 6.258513580963777e-06, "loss": 0.5851, "step": 3003 }, { "epoch": 1.134682682303982, "grad_norm": 1.1191221759321432, "learning_rate": 6.257058516117483e-06, "loss": 0.5518, "step": 3004 }, { "epoch": 1.1350610044452851, "grad_norm": 1.1389113202769825, "learning_rate": 6.255603012957203e-06, "loss": 0.5996, "step": 3005 }, { "epoch": 1.1354393265865885, "grad_norm": 1.1468366015314928, "learning_rate": 6.254147071765593e-06, "loss": 0.6248, "step": 3006 }, { "epoch": 1.1358176487278917, "grad_norm": 1.120264018594437, "learning_rate": 6.252690692825393e-06, "loss": 0.5909, "step": 3007 }, { "epoch": 1.136195970869195, "grad_norm": 1.1418383417192708, "learning_rate": 6.2512338764194245e-06, "loss": 0.6044, "step": 3008 }, { "epoch": 1.1365742930104985, "grad_norm": 1.1157032732072756, "learning_rate": 6.2497766228306e-06, "loss": 0.6154, "step": 3009 }, { "epoch": 1.1369526151518017, "grad_norm": 1.1025553551561673, "learning_rate": 6.24831893234191e-06, "loss": 0.5968, "step": 3010 }, { "epoch": 1.137330937293105, "grad_norm": 1.0703331822956041, "learning_rate": 6.246860805236438e-06, "loss": 0.614, "step": 3011 }, { "epoch": 1.1377092594344085, "grad_norm": 1.0941387590020248, "learning_rate": 6.245402241797345e-06, "loss": 0.5877, "step": 3012 }, { "epoch": 1.1380875815757117, "grad_norm": 1.1036186168758577, "learning_rate": 6.24394324230788e-06, "loss": 0.6276, "step": 3013 }, { "epoch": 1.138465903717015, "grad_norm": 1.1233817079550372, "learning_rate": 6.242483807051379e-06, "loss": 0.6417, "step": 3014 }, { "epoch": 1.1388442258583185, "grad_norm": 1.0955895613598414, "learning_rate": 6.241023936311256e-06, "loss": 0.5863, "step": 3015 }, { "epoch": 1.1392225479996216, "grad_norm": 1.0858293029589012, "learning_rate": 6.239563630371016e-06, "loss": 0.5758, "step": 3016 }, { "epoch": 1.139600870140925, "grad_norm": 1.116798091733882, "learning_rate": 6.238102889514244e-06, "loss": 0.593, "step": 3017 }, { "epoch": 1.1399791922822282, "grad_norm": 1.1089444839993028, "learning_rate": 6.236641714024614e-06, "loss": 0.6112, "step": 3018 }, { "epoch": 1.1403575144235316, "grad_norm": 1.1986200839787384, "learning_rate": 6.23518010418588e-06, "loss": 0.6141, "step": 3019 }, { "epoch": 1.140735836564835, "grad_norm": 1.0950059129640932, "learning_rate": 6.233718060281883e-06, "loss": 0.6003, "step": 3020 }, { "epoch": 1.1411141587061382, "grad_norm": 1.1293585342809132, "learning_rate": 6.232255582596547e-06, "loss": 0.635, "step": 3021 }, { "epoch": 1.1414924808474416, "grad_norm": 1.0997542937583942, "learning_rate": 6.230792671413882e-06, "loss": 0.5964, "step": 3022 }, { "epoch": 1.141870802988745, "grad_norm": 1.1446241667917194, "learning_rate": 6.22932932701798e-06, "loss": 0.6462, "step": 3023 }, { "epoch": 1.1422491251300482, "grad_norm": 1.152395993262499, "learning_rate": 6.227865549693019e-06, "loss": 0.6318, "step": 3024 }, { "epoch": 1.1426274472713516, "grad_norm": 1.1238060326584278, "learning_rate": 6.226401339723258e-06, "loss": 0.6557, "step": 3025 }, { "epoch": 1.143005769412655, "grad_norm": 1.0616125607827556, "learning_rate": 6.224936697393045e-06, "loss": 0.6035, "step": 3026 }, { "epoch": 1.1433840915539581, "grad_norm": 1.1629946176571666, "learning_rate": 6.2234716229868065e-06, "loss": 0.59, "step": 3027 }, { "epoch": 1.1437624136952615, "grad_norm": 1.1231668221914821, "learning_rate": 6.222006116789058e-06, "loss": 0.6174, "step": 3028 }, { "epoch": 1.1441407358365647, "grad_norm": 1.103161146917699, "learning_rate": 6.220540179084395e-06, "loss": 0.6252, "step": 3029 }, { "epoch": 1.1445190579778681, "grad_norm": 1.1086815203504103, "learning_rate": 6.219073810157498e-06, "loss": 0.6112, "step": 3030 }, { "epoch": 1.1448973801191715, "grad_norm": 1.147629035931138, "learning_rate": 6.21760701029313e-06, "loss": 0.6469, "step": 3031 }, { "epoch": 1.1452757022604747, "grad_norm": 1.1037589470324547, "learning_rate": 6.216139779776144e-06, "loss": 0.6028, "step": 3032 }, { "epoch": 1.145654024401778, "grad_norm": 1.1166980196423228, "learning_rate": 6.214672118891467e-06, "loss": 0.6142, "step": 3033 }, { "epoch": 1.1460323465430815, "grad_norm": 1.1530131434815551, "learning_rate": 6.213204027924117e-06, "loss": 0.5712, "step": 3034 }, { "epoch": 1.1464106686843847, "grad_norm": 1.13750762308757, "learning_rate": 6.211735507159192e-06, "loss": 0.582, "step": 3035 }, { "epoch": 1.146788990825688, "grad_norm": 1.0729278852374269, "learning_rate": 6.210266556881874e-06, "loss": 0.5828, "step": 3036 }, { "epoch": 1.1471673129669915, "grad_norm": 1.1185939936755518, "learning_rate": 6.2087971773774286e-06, "loss": 0.6101, "step": 3037 }, { "epoch": 1.1475456351082947, "grad_norm": 1.165932125246575, "learning_rate": 6.207327368931204e-06, "loss": 0.6175, "step": 3038 }, { "epoch": 1.147923957249598, "grad_norm": 1.1391811371745488, "learning_rate": 6.205857131828636e-06, "loss": 0.6093, "step": 3039 }, { "epoch": 1.1483022793909012, "grad_norm": 1.1299174262619414, "learning_rate": 6.204386466355237e-06, "loss": 0.6192, "step": 3040 }, { "epoch": 1.1486806015322046, "grad_norm": 1.246389380402845, "learning_rate": 6.202915372796606e-06, "loss": 0.595, "step": 3041 }, { "epoch": 1.149058923673508, "grad_norm": 1.1049465243645995, "learning_rate": 6.201443851438428e-06, "loss": 0.5843, "step": 3042 }, { "epoch": 1.1494372458148112, "grad_norm": 1.1156462314879443, "learning_rate": 6.199971902566465e-06, "loss": 0.5973, "step": 3043 }, { "epoch": 1.1498155679561146, "grad_norm": 1.1417777395895234, "learning_rate": 6.198499526466566e-06, "loss": 0.5823, "step": 3044 }, { "epoch": 1.150193890097418, "grad_norm": 1.1161675598044707, "learning_rate": 6.1970267234246614e-06, "loss": 0.6077, "step": 3045 }, { "epoch": 1.1505722122387212, "grad_norm": 1.107178211545901, "learning_rate": 6.195553493726766e-06, "loss": 0.569, "step": 3046 }, { "epoch": 1.1509505343800246, "grad_norm": 1.0801710868852208, "learning_rate": 6.1940798376589765e-06, "loss": 0.581, "step": 3047 }, { "epoch": 1.151328856521328, "grad_norm": 1.1076189504091067, "learning_rate": 6.1926057555074714e-06, "loss": 0.6116, "step": 3048 }, { "epoch": 1.1517071786626312, "grad_norm": 1.1427962338694941, "learning_rate": 6.191131247558515e-06, "loss": 0.5818, "step": 3049 }, { "epoch": 1.1520855008039346, "grad_norm": 1.1114501628255513, "learning_rate": 6.189656314098451e-06, "loss": 0.6247, "step": 3050 }, { "epoch": 1.1524638229452377, "grad_norm": 1.0682006350023248, "learning_rate": 6.188180955413707e-06, "loss": 0.65, "step": 3051 }, { "epoch": 1.1528421450865411, "grad_norm": 1.1150247448570672, "learning_rate": 6.186705171790793e-06, "loss": 0.6079, "step": 3052 }, { "epoch": 1.1532204672278445, "grad_norm": 1.1589938155434434, "learning_rate": 6.185228963516303e-06, "loss": 0.5607, "step": 3053 }, { "epoch": 1.153598789369148, "grad_norm": 1.0858975144409513, "learning_rate": 6.183752330876911e-06, "loss": 0.5727, "step": 3054 }, { "epoch": 1.1539771115104511, "grad_norm": 1.1619601222313594, "learning_rate": 6.182275274159374e-06, "loss": 0.6134, "step": 3055 }, { "epoch": 1.1543554336517545, "grad_norm": 1.1671042432509766, "learning_rate": 6.180797793650534e-06, "loss": 0.5968, "step": 3056 }, { "epoch": 1.1547337557930577, "grad_norm": 1.1356445374431496, "learning_rate": 6.1793198896373126e-06, "loss": 0.6019, "step": 3057 }, { "epoch": 1.155112077934361, "grad_norm": 1.0909299031939843, "learning_rate": 6.177841562406714e-06, "loss": 0.5887, "step": 3058 }, { "epoch": 1.1554904000756645, "grad_norm": 1.1846517046294107, "learning_rate": 6.176362812245823e-06, "loss": 0.6484, "step": 3059 }, { "epoch": 1.1554904000756645, "eval_loss": 0.7661470174789429, "eval_runtime": 22.7544, "eval_samples_per_second": 38.894, "eval_steps_per_second": 1.231, "step": 3059 }, { "epoch": 1.1554904000756645, "eval_bench_accuracy_arc_challenge": 0.0, "eval_bench_accuracy_hellaswag": 0.245, "eval_bench_accuracy_mmlu": 0.24347826086956523, "eval_bench_average_accuracy": 0.16282608695652176, "eval_bench_loss": 8.959730181777687, "eval_bench_total_accuracy": 0.16923076923076924, "step": 3059 }, { "epoch": 1.1558687222169677, "grad_norm": 1.1014654966135637, "learning_rate": 6.174883639441813e-06, "loss": 0.625, "step": 3060 }, { "epoch": 1.156247044358271, "grad_norm": 1.144441845078975, "learning_rate": 6.1734040442819314e-06, "loss": 0.5787, "step": 3061 }, { "epoch": 1.1566253664995745, "grad_norm": 1.0976478073487557, "learning_rate": 6.1719240270535115e-06, "loss": 0.6015, "step": 3062 }, { "epoch": 1.1570036886408777, "grad_norm": 1.0815352868590902, "learning_rate": 6.170443588043969e-06, "loss": 0.6148, "step": 3063 }, { "epoch": 1.157382010782181, "grad_norm": 1.1223114267592562, "learning_rate": 6.1689627275408015e-06, "loss": 0.5857, "step": 3064 }, { "epoch": 1.1577603329234845, "grad_norm": 1.1617228326207247, "learning_rate": 6.167481445831584e-06, "loss": 0.62, "step": 3065 }, { "epoch": 1.1581386550647876, "grad_norm": 1.148043302777634, "learning_rate": 6.165999743203981e-06, "loss": 0.5818, "step": 3066 }, { "epoch": 1.158516977206091, "grad_norm": 1.1190927031819493, "learning_rate": 6.164517619945734e-06, "loss": 0.5709, "step": 3067 }, { "epoch": 1.1588952993473942, "grad_norm": 1.10150683907697, "learning_rate": 6.163035076344664e-06, "loss": 0.5875, "step": 3068 }, { "epoch": 1.1592736214886976, "grad_norm": 1.1470686822843648, "learning_rate": 6.1615521126886805e-06, "loss": 0.6181, "step": 3069 }, { "epoch": 1.159651943630001, "grad_norm": 1.1669518566001118, "learning_rate": 6.1600687292657685e-06, "loss": 0.6471, "step": 3070 }, { "epoch": 1.1600302657713042, "grad_norm": 1.1257835102040736, "learning_rate": 6.158584926363997e-06, "loss": 0.6182, "step": 3071 }, { "epoch": 1.1604085879126076, "grad_norm": 1.0978786636902147, "learning_rate": 6.1571007042715155e-06, "loss": 0.6069, "step": 3072 }, { "epoch": 1.160786910053911, "grad_norm": 1.1449694321634942, "learning_rate": 6.155616063276556e-06, "loss": 0.5924, "step": 3073 }, { "epoch": 1.1611652321952142, "grad_norm": 1.1434155453631858, "learning_rate": 6.15413100366743e-06, "loss": 0.6177, "step": 3074 }, { "epoch": 1.1615435543365176, "grad_norm": 1.1459946058923065, "learning_rate": 6.152645525732535e-06, "loss": 0.6204, "step": 3075 }, { "epoch": 1.161921876477821, "grad_norm": 1.1227094309904373, "learning_rate": 6.151159629760342e-06, "loss": 0.6139, "step": 3076 }, { "epoch": 1.1623001986191241, "grad_norm": 1.1804223490177808, "learning_rate": 6.1496733160394115e-06, "loss": 0.574, "step": 3077 }, { "epoch": 1.1626785207604275, "grad_norm": 1.0582193383269725, "learning_rate": 6.148186584858378e-06, "loss": 0.5983, "step": 3078 }, { "epoch": 1.1630568429017307, "grad_norm": 1.1154872954870327, "learning_rate": 6.146699436505963e-06, "loss": 0.5956, "step": 3079 }, { "epoch": 1.1634351650430341, "grad_norm": 1.143241761123763, "learning_rate": 6.145211871270963e-06, "loss": 0.5889, "step": 3080 }, { "epoch": 1.1638134871843375, "grad_norm": 1.0975871616117818, "learning_rate": 6.143723889442262e-06, "loss": 0.5749, "step": 3081 }, { "epoch": 1.1641918093256407, "grad_norm": 1.0768501813224447, "learning_rate": 6.14223549130882e-06, "loss": 0.5908, "step": 3082 }, { "epoch": 1.164570131466944, "grad_norm": 1.1000086934856, "learning_rate": 6.140746677159679e-06, "loss": 0.6056, "step": 3083 }, { "epoch": 1.1649484536082475, "grad_norm": 1.0862626426516695, "learning_rate": 6.139257447283963e-06, "loss": 0.6144, "step": 3084 }, { "epoch": 1.1653267757495507, "grad_norm": 1.0626159122200098, "learning_rate": 6.137767801970876e-06, "loss": 0.572, "step": 3085 }, { "epoch": 1.165705097890854, "grad_norm": 1.1652572888144201, "learning_rate": 6.1362777415097026e-06, "loss": 0.6185, "step": 3086 }, { "epoch": 1.1660834200321575, "grad_norm": 1.125328152204312, "learning_rate": 6.134787266189807e-06, "loss": 0.5675, "step": 3087 }, { "epoch": 1.1664617421734607, "grad_norm": 1.101464020267817, "learning_rate": 6.133296376300636e-06, "loss": 0.602, "step": 3088 }, { "epoch": 1.166840064314764, "grad_norm": 1.0793577092769573, "learning_rate": 6.131805072131717e-06, "loss": 0.6254, "step": 3089 }, { "epoch": 1.1672183864560672, "grad_norm": 1.10768102540295, "learning_rate": 6.130313353972656e-06, "loss": 0.6187, "step": 3090 }, { "epoch": 1.1675967085973706, "grad_norm": 1.1047177544035363, "learning_rate": 6.128821222113139e-06, "loss": 0.5984, "step": 3091 }, { "epoch": 1.167975030738674, "grad_norm": 1.124376538904714, "learning_rate": 6.127328676842933e-06, "loss": 0.6087, "step": 3092 }, { "epoch": 1.1683533528799772, "grad_norm": 1.1459850760127377, "learning_rate": 6.125835718451888e-06, "loss": 0.6288, "step": 3093 }, { "epoch": 1.1687316750212806, "grad_norm": 1.1027973853013475, "learning_rate": 6.124342347229932e-06, "loss": 0.5924, "step": 3094 }, { "epoch": 1.169109997162584, "grad_norm": 1.0864042237830664, "learning_rate": 6.122848563467071e-06, "loss": 0.6354, "step": 3095 }, { "epoch": 1.1694883193038872, "grad_norm": 1.1155124321690744, "learning_rate": 6.121354367453398e-06, "loss": 0.5762, "step": 3096 }, { "epoch": 1.1698666414451906, "grad_norm": 1.0905032596292725, "learning_rate": 6.119859759479075e-06, "loss": 0.627, "step": 3097 }, { "epoch": 1.170244963586494, "grad_norm": 1.1022953197555945, "learning_rate": 6.118364739834354e-06, "loss": 0.5953, "step": 3098 }, { "epoch": 1.1706232857277972, "grad_norm": 1.1358007173182298, "learning_rate": 6.1168693088095635e-06, "loss": 0.617, "step": 3099 }, { "epoch": 1.1710016078691006, "grad_norm": 1.1876865568157937, "learning_rate": 6.115373466695111e-06, "loss": 0.6061, "step": 3100 }, { "epoch": 1.1713799300104037, "grad_norm": 1.1178878130878194, "learning_rate": 6.113877213781483e-06, "loss": 0.6008, "step": 3101 }, { "epoch": 1.1717582521517071, "grad_norm": 1.1233219656555455, "learning_rate": 6.112380550359251e-06, "loss": 0.6188, "step": 3102 }, { "epoch": 1.1721365742930105, "grad_norm": 1.1145918683259326, "learning_rate": 6.11088347671906e-06, "loss": 0.6169, "step": 3103 }, { "epoch": 1.1725148964343137, "grad_norm": 1.1146999376932287, "learning_rate": 6.109385993151638e-06, "loss": 0.601, "step": 3104 }, { "epoch": 1.1728932185756171, "grad_norm": 1.1205607984942194, "learning_rate": 6.107888099947791e-06, "loss": 0.585, "step": 3105 }, { "epoch": 1.1732715407169205, "grad_norm": 1.0973590433692995, "learning_rate": 6.106389797398405e-06, "loss": 0.5858, "step": 3106 }, { "epoch": 1.1736498628582237, "grad_norm": 1.0876705012991348, "learning_rate": 6.104891085794447e-06, "loss": 0.62, "step": 3107 }, { "epoch": 1.174028184999527, "grad_norm": 1.1614327701889045, "learning_rate": 6.103391965426963e-06, "loss": 0.6232, "step": 3108 }, { "epoch": 1.1744065071408305, "grad_norm": 1.131877847532318, "learning_rate": 6.101892436587076e-06, "loss": 0.6237, "step": 3109 }, { "epoch": 1.1747848292821337, "grad_norm": 1.1319323473611833, "learning_rate": 6.100392499565991e-06, "loss": 0.6022, "step": 3110 }, { "epoch": 1.175163151423437, "grad_norm": 1.0576522311211, "learning_rate": 6.09889215465499e-06, "loss": 0.586, "step": 3111 }, { "epoch": 1.1755414735647403, "grad_norm": 1.1252321397801972, "learning_rate": 6.097391402145437e-06, "loss": 0.6199, "step": 3112 }, { "epoch": 1.1759197957060437, "grad_norm": 1.1287515518957223, "learning_rate": 6.095890242328773e-06, "loss": 0.6113, "step": 3113 }, { "epoch": 1.176298117847347, "grad_norm": 1.126107097869058, "learning_rate": 6.094388675496519e-06, "loss": 0.6223, "step": 3114 }, { "epoch": 1.1766764399886505, "grad_norm": 1.1526993007983606, "learning_rate": 6.092886701940274e-06, "loss": 0.6167, "step": 3115 }, { "epoch": 1.1770547621299536, "grad_norm": 1.1350268180478196, "learning_rate": 6.091384321951718e-06, "loss": 0.5991, "step": 3116 }, { "epoch": 1.177433084271257, "grad_norm": 1.1192535187618968, "learning_rate": 6.089881535822607e-06, "loss": 0.6108, "step": 3117 }, { "epoch": 1.1778114064125602, "grad_norm": 1.0835937335147057, "learning_rate": 6.088378343844779e-06, "loss": 0.5882, "step": 3118 }, { "epoch": 1.1781897285538636, "grad_norm": 1.1181941826509343, "learning_rate": 6.086874746310148e-06, "loss": 0.5997, "step": 3119 }, { "epoch": 1.178568050695167, "grad_norm": 1.0941801054936213, "learning_rate": 6.0853707435107105e-06, "loss": 0.6178, "step": 3120 }, { "epoch": 1.1789463728364702, "grad_norm": 1.1404538263291477, "learning_rate": 6.083866335738536e-06, "loss": 0.6074, "step": 3121 }, { "epoch": 1.1793246949777736, "grad_norm": 1.1233703000761228, "learning_rate": 6.0823615232857795e-06, "loss": 0.618, "step": 3122 }, { "epoch": 1.179703017119077, "grad_norm": 1.061722362668965, "learning_rate": 6.080856306444669e-06, "loss": 0.5729, "step": 3123 }, { "epoch": 1.1800813392603802, "grad_norm": 1.2285987844787836, "learning_rate": 6.079350685507513e-06, "loss": 0.6225, "step": 3124 }, { "epoch": 1.1804596614016836, "grad_norm": 1.1391413881110837, "learning_rate": 6.0778446607667e-06, "loss": 0.6211, "step": 3125 }, { "epoch": 1.180837983542987, "grad_norm": 1.1347880214836306, "learning_rate": 6.076338232514693e-06, "loss": 0.5871, "step": 3126 }, { "epoch": 1.1812163056842901, "grad_norm": 1.117904811244799, "learning_rate": 6.074831401044039e-06, "loss": 0.5793, "step": 3127 }, { "epoch": 1.1815946278255935, "grad_norm": 1.084133372191697, "learning_rate": 6.0733241666473565e-06, "loss": 0.5838, "step": 3128 }, { "epoch": 1.1819729499668967, "grad_norm": 1.1153196913003163, "learning_rate": 6.071816529617348e-06, "loss": 0.608, "step": 3129 }, { "epoch": 1.1823512721082001, "grad_norm": 1.0887386114515758, "learning_rate": 6.070308490246793e-06, "loss": 0.6145, "step": 3130 }, { "epoch": 1.1827295942495035, "grad_norm": 1.1005700243285668, "learning_rate": 6.068800048828548e-06, "loss": 0.6299, "step": 3131 }, { "epoch": 1.1831079163908067, "grad_norm": 1.1292296497121013, "learning_rate": 6.067291205655545e-06, "loss": 0.6279, "step": 3132 }, { "epoch": 1.18348623853211, "grad_norm": 1.1046638350785665, "learning_rate": 6.065781961020799e-06, "loss": 0.607, "step": 3133 }, { "epoch": 1.1838645606734135, "grad_norm": 1.1100596635759847, "learning_rate": 6.064272315217401e-06, "loss": 0.6239, "step": 3134 }, { "epoch": 1.1842428828147167, "grad_norm": 1.1312199660678584, "learning_rate": 6.06276226853852e-06, "loss": 0.651, "step": 3135 }, { "epoch": 1.18462120495602, "grad_norm": 1.1160073692560897, "learning_rate": 6.0612518212774e-06, "loss": 0.6049, "step": 3136 }, { "epoch": 1.1849995270973235, "grad_norm": 1.1218444399043954, "learning_rate": 6.059740973727369e-06, "loss": 0.6163, "step": 3137 }, { "epoch": 1.1853778492386267, "grad_norm": 1.0863749776288811, "learning_rate": 6.058229726181826e-06, "loss": 0.6095, "step": 3138 }, { "epoch": 1.18575617137993, "grad_norm": 1.127680892696607, "learning_rate": 6.0567180789342525e-06, "loss": 0.6037, "step": 3139 }, { "epoch": 1.1861344935212332, "grad_norm": 1.1014547201971496, "learning_rate": 6.0552060322782045e-06, "loss": 0.6162, "step": 3140 }, { "epoch": 1.1865128156625366, "grad_norm": 1.1405927062243015, "learning_rate": 6.053693586507319e-06, "loss": 0.62, "step": 3141 }, { "epoch": 1.18689113780384, "grad_norm": 1.1780565477235185, "learning_rate": 6.052180741915306e-06, "loss": 0.5812, "step": 3142 }, { "epoch": 1.1872694599451432, "grad_norm": 1.1253911441085132, "learning_rate": 6.050667498795956e-06, "loss": 0.6137, "step": 3143 }, { "epoch": 1.1876477820864466, "grad_norm": 1.0775669546796678, "learning_rate": 6.049153857443137e-06, "loss": 0.5969, "step": 3144 }, { "epoch": 1.18802610422775, "grad_norm": 1.0861884773066142, "learning_rate": 6.047639818150795e-06, "loss": 0.6117, "step": 3145 }, { "epoch": 1.1884044263690532, "grad_norm": 1.1310277466445289, "learning_rate": 6.046125381212949e-06, "loss": 0.6118, "step": 3146 }, { "epoch": 1.1887827485103566, "grad_norm": 1.1631052653954754, "learning_rate": 6.044610546923698e-06, "loss": 0.5609, "step": 3147 }, { "epoch": 1.18916107065166, "grad_norm": 1.1024865118483864, "learning_rate": 6.0430953155772215e-06, "loss": 0.5653, "step": 3148 }, { "epoch": 1.1895393927929632, "grad_norm": 1.1668630314886717, "learning_rate": 6.04157968746777e-06, "loss": 0.6325, "step": 3149 }, { "epoch": 1.1899177149342666, "grad_norm": 1.0901216562065503, "learning_rate": 6.040063662889675e-06, "loss": 0.6244, "step": 3150 }, { "epoch": 1.1902960370755697, "grad_norm": 1.1617014317650085, "learning_rate": 6.038547242137344e-06, "loss": 0.6044, "step": 3151 }, { "epoch": 1.1906743592168731, "grad_norm": 1.1296776021297936, "learning_rate": 6.037030425505261e-06, "loss": 0.638, "step": 3152 }, { "epoch": 1.1910526813581765, "grad_norm": 1.079714496664035, "learning_rate": 6.035513213287987e-06, "loss": 0.5866, "step": 3153 }, { "epoch": 1.1914310034994797, "grad_norm": 1.0912986398853903, "learning_rate": 6.033995605780161e-06, "loss": 0.5908, "step": 3154 }, { "epoch": 1.1918093256407831, "grad_norm": 1.1169197719182176, "learning_rate": 6.032477603276497e-06, "loss": 0.6022, "step": 3155 }, { "epoch": 1.1921876477820865, "grad_norm": 1.11128834897286, "learning_rate": 6.030959206071786e-06, "loss": 0.6032, "step": 3156 }, { "epoch": 1.1925659699233897, "grad_norm": 1.1183302435906126, "learning_rate": 6.029440414460898e-06, "loss": 0.5627, "step": 3157 }, { "epoch": 1.192944292064693, "grad_norm": 1.119355782558496, "learning_rate": 6.027921228738777e-06, "loss": 0.6068, "step": 3158 }, { "epoch": 1.1933226142059965, "grad_norm": 1.1655451135308021, "learning_rate": 6.026401649200444e-06, "loss": 0.6177, "step": 3159 }, { "epoch": 1.1937009363472997, "grad_norm": 1.1600929433990705, "learning_rate": 6.024881676140996e-06, "loss": 0.6186, "step": 3160 }, { "epoch": 1.194079258488603, "grad_norm": 1.144791435213969, "learning_rate": 6.023361309855609e-06, "loss": 0.6238, "step": 3161 }, { "epoch": 1.1944575806299063, "grad_norm": 1.1340255387092586, "learning_rate": 6.0218405506395315e-06, "loss": 0.5745, "step": 3162 }, { "epoch": 1.1948359027712097, "grad_norm": 1.0814059408267755, "learning_rate": 6.020319398788093e-06, "loss": 0.6003, "step": 3163 }, { "epoch": 1.195214224912513, "grad_norm": 1.0639110568237347, "learning_rate": 6.018797854596694e-06, "loss": 0.5751, "step": 3164 }, { "epoch": 1.1955925470538162, "grad_norm": 1.0873969004474242, "learning_rate": 6.017275918360814e-06, "loss": 0.6398, "step": 3165 }, { "epoch": 1.1959708691951196, "grad_norm": 1.135078901391362, "learning_rate": 6.015753590376011e-06, "loss": 0.601, "step": 3166 }, { "epoch": 1.196349191336423, "grad_norm": 1.1019279505333324, "learning_rate": 6.014230870937914e-06, "loss": 0.583, "step": 3167 }, { "epoch": 1.1967275134777262, "grad_norm": 1.1240106767893345, "learning_rate": 6.012707760342231e-06, "loss": 0.6299, "step": 3168 }, { "epoch": 1.1971058356190296, "grad_norm": 1.1398317734698087, "learning_rate": 6.011184258884747e-06, "loss": 0.621, "step": 3169 }, { "epoch": 1.197484157760333, "grad_norm": 1.1880903489231742, "learning_rate": 6.00966036686132e-06, "loss": 0.6076, "step": 3170 }, { "epoch": 1.1978624799016362, "grad_norm": 1.1365124579399424, "learning_rate": 6.008136084567885e-06, "loss": 0.5864, "step": 3171 }, { "epoch": 1.1982408020429396, "grad_norm": 1.0794429115781958, "learning_rate": 6.006611412300454e-06, "loss": 0.6252, "step": 3172 }, { "epoch": 1.1986191241842428, "grad_norm": 1.0942349106787228, "learning_rate": 6.005086350355114e-06, "loss": 0.5927, "step": 3173 }, { "epoch": 1.1989974463255462, "grad_norm": 1.1161042073899385, "learning_rate": 6.003560899028027e-06, "loss": 0.6057, "step": 3174 }, { "epoch": 1.1993757684668496, "grad_norm": 1.1148606152894032, "learning_rate": 6.002035058615429e-06, "loss": 0.5962, "step": 3175 }, { "epoch": 1.1997540906081527, "grad_norm": 1.1778025476796163, "learning_rate": 6.000508829413638e-06, "loss": 0.5944, "step": 3176 }, { "epoch": 1.2001324127494561, "grad_norm": 1.1289169545828777, "learning_rate": 5.998982211719038e-06, "loss": 0.6115, "step": 3177 }, { "epoch": 1.2005107348907595, "grad_norm": 1.116100986314231, "learning_rate": 5.997455205828099e-06, "loss": 0.6029, "step": 3178 }, { "epoch": 1.2008890570320627, "grad_norm": 1.1467918953457878, "learning_rate": 5.995927812037356e-06, "loss": 0.6169, "step": 3179 }, { "epoch": 1.2012673791733661, "grad_norm": 1.0730437473704995, "learning_rate": 5.9944000306434275e-06, "loss": 0.5959, "step": 3180 } ], "logging_steps": 1, "max_steps": 7929, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 53, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 5.237643116311216e+19, "train_batch_size": 4, "trial_name": null, "trial_params": null }