{ "best_metric": 0.7140286564826965, "best_model_checkpoint": "data/Llama-31-8B_task-2_180-samples_config-3/checkpoint-323", "epoch": 26.0, "eval_steps": 500, "global_step": 442, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.058823529411764705, "grad_norm": 0.7734732031822205, "learning_rate": 3.9215686274509804e-08, "loss": 1.1412, "step": 1 }, { "epoch": 0.11764705882352941, "grad_norm": 0.6782335638999939, "learning_rate": 7.843137254901961e-08, "loss": 1.1733, "step": 2 }, { "epoch": 0.23529411764705882, "grad_norm": 0.8408952951431274, "learning_rate": 1.5686274509803921e-07, "loss": 1.1666, "step": 4 }, { "epoch": 0.35294117647058826, "grad_norm": 0.7287867069244385, "learning_rate": 2.3529411764705883e-07, "loss": 1.1801, "step": 6 }, { "epoch": 0.47058823529411764, "grad_norm": 0.7565770745277405, "learning_rate": 3.1372549019607843e-07, "loss": 1.1116, "step": 8 }, { "epoch": 0.5882352941176471, "grad_norm": 0.6836811304092407, "learning_rate": 3.921568627450981e-07, "loss": 1.1475, "step": 10 }, { "epoch": 0.7058823529411765, "grad_norm": 0.6788910627365112, "learning_rate": 4.7058823529411767e-07, "loss": 1.0752, "step": 12 }, { "epoch": 0.8235294117647058, "grad_norm": 0.6463932991027832, "learning_rate": 5.490196078431373e-07, "loss": 1.1428, "step": 14 }, { "epoch": 0.9411764705882353, "grad_norm": 0.7503110766410828, "learning_rate": 6.274509803921569e-07, "loss": 1.0365, "step": 16 }, { "epoch": 1.0, "eval_loss": 1.131649374961853, "eval_runtime": 37.0173, "eval_samples_per_second": 0.973, "eval_steps_per_second": 0.973, "step": 17 }, { "epoch": 1.0588235294117647, "grad_norm": 0.7130016088485718, "learning_rate": 7.058823529411766e-07, "loss": 1.1439, "step": 18 }, { "epoch": 1.1764705882352942, "grad_norm": 0.69338458776474, "learning_rate": 7.843137254901962e-07, "loss": 1.1374, "step": 20 }, { "epoch": 1.2941176470588236, "grad_norm": 0.6517343521118164, "learning_rate": 8.627450980392157e-07, "loss": 1.1213, "step": 22 }, { "epoch": 1.4117647058823528, "grad_norm": 0.685614287853241, "learning_rate": 9.411764705882353e-07, "loss": 1.1182, "step": 24 }, { "epoch": 1.5294117647058822, "grad_norm": 0.6998293995857239, "learning_rate": 1.019607843137255e-06, "loss": 1.0508, "step": 26 }, { "epoch": 1.6470588235294117, "grad_norm": 0.6757376194000244, "learning_rate": 1.0980392156862745e-06, "loss": 1.0773, "step": 28 }, { "epoch": 1.7647058823529411, "grad_norm": 0.7003068327903748, "learning_rate": 1.1764705882352942e-06, "loss": 1.1765, "step": 30 }, { "epoch": 1.8823529411764706, "grad_norm": 0.75652676820755, "learning_rate": 1.2549019607843137e-06, "loss": 1.0897, "step": 32 }, { "epoch": 2.0, "grad_norm": 0.8910195231437683, "learning_rate": 1.3333333333333334e-06, "loss": 1.1746, "step": 34 }, { "epoch": 2.0, "eval_loss": 1.119631290435791, "eval_runtime": 37.0058, "eval_samples_per_second": 0.973, "eval_steps_per_second": 0.973, "step": 34 }, { "epoch": 2.1176470588235294, "grad_norm": 0.6963354349136353, "learning_rate": 1.4117647058823531e-06, "loss": 1.0914, "step": 36 }, { "epoch": 2.235294117647059, "grad_norm": 0.6639174818992615, "learning_rate": 1.4901960784313726e-06, "loss": 1.1038, "step": 38 }, { "epoch": 2.3529411764705883, "grad_norm": 0.7396222949028015, "learning_rate": 1.5686274509803923e-06, "loss": 1.0596, "step": 40 }, { "epoch": 2.4705882352941178, "grad_norm": 0.7213342189788818, "learning_rate": 1.6470588235294118e-06, "loss": 1.0962, "step": 42 }, { "epoch": 2.588235294117647, "grad_norm": 0.7507779002189636, "learning_rate": 1.7254901960784315e-06, "loss": 1.0945, "step": 44 }, { "epoch": 2.7058823529411766, "grad_norm": 0.803626298904419, "learning_rate": 1.8039215686274512e-06, "loss": 1.1417, "step": 46 }, { "epoch": 2.8235294117647056, "grad_norm": 0.8166558742523193, "learning_rate": 1.8823529411764707e-06, "loss": 1.1391, "step": 48 }, { "epoch": 2.9411764705882355, "grad_norm": 0.7504469752311707, "learning_rate": 1.96078431372549e-06, "loss": 1.0933, "step": 50 }, { "epoch": 3.0, "eval_loss": 1.0957257747650146, "eval_runtime": 37.0081, "eval_samples_per_second": 0.973, "eval_steps_per_second": 0.973, "step": 51 }, { "epoch": 3.0588235294117645, "grad_norm": 0.7341092228889465, "learning_rate": 2.03921568627451e-06, "loss": 1.0645, "step": 52 }, { "epoch": 3.176470588235294, "grad_norm": 0.7262235283851624, "learning_rate": 2.1176470588235296e-06, "loss": 1.1047, "step": 54 }, { "epoch": 3.2941176470588234, "grad_norm": 0.6933767795562744, "learning_rate": 2.196078431372549e-06, "loss": 1.0305, "step": 56 }, { "epoch": 3.411764705882353, "grad_norm": 0.8481059670448303, "learning_rate": 2.274509803921569e-06, "loss": 1.1403, "step": 58 }, { "epoch": 3.5294117647058822, "grad_norm": 0.7597863078117371, "learning_rate": 2.3529411764705885e-06, "loss": 1.0537, "step": 60 }, { "epoch": 3.6470588235294117, "grad_norm": 0.7542691826820374, "learning_rate": 2.431372549019608e-06, "loss": 1.0815, "step": 62 }, { "epoch": 3.764705882352941, "grad_norm": 0.7457835674285889, "learning_rate": 2.5098039215686274e-06, "loss": 1.0747, "step": 64 }, { "epoch": 3.8823529411764706, "grad_norm": 0.6987655758857727, "learning_rate": 2.5882352941176473e-06, "loss": 1.1015, "step": 66 }, { "epoch": 4.0, "grad_norm": 0.7653221487998962, "learning_rate": 2.666666666666667e-06, "loss": 0.985, "step": 68 }, { "epoch": 4.0, "eval_loss": 1.0540263652801514, "eval_runtime": 37.0036, "eval_samples_per_second": 0.973, "eval_steps_per_second": 0.973, "step": 68 }, { "epoch": 4.117647058823529, "grad_norm": 0.6138588190078735, "learning_rate": 2.7450980392156867e-06, "loss": 1.0818, "step": 70 }, { "epoch": 4.235294117647059, "grad_norm": 0.7161767482757568, "learning_rate": 2.8235294117647062e-06, "loss": 1.0233, "step": 72 }, { "epoch": 4.352941176470588, "grad_norm": 0.5691388845443726, "learning_rate": 2.901960784313726e-06, "loss": 1.0152, "step": 74 }, { "epoch": 4.470588235294118, "grad_norm": 0.6562293171882629, "learning_rate": 2.980392156862745e-06, "loss": 1.0299, "step": 76 }, { "epoch": 4.588235294117647, "grad_norm": 0.6522495150566101, "learning_rate": 3.058823529411765e-06, "loss": 1.0341, "step": 78 }, { "epoch": 4.705882352941177, "grad_norm": 0.6249567270278931, "learning_rate": 3.1372549019607846e-06, "loss": 0.9573, "step": 80 }, { "epoch": 4.823529411764706, "grad_norm": 0.6513485312461853, "learning_rate": 3.2156862745098045e-06, "loss": 1.0373, "step": 82 }, { "epoch": 4.9411764705882355, "grad_norm": 0.6425153613090515, "learning_rate": 3.2941176470588236e-06, "loss": 0.9741, "step": 84 }, { "epoch": 5.0, "eval_loss": 0.99503093957901, "eval_runtime": 37.0007, "eval_samples_per_second": 0.973, "eval_steps_per_second": 0.973, "step": 85 }, { "epoch": 5.0588235294117645, "grad_norm": 0.645844042301178, "learning_rate": 3.3725490196078435e-06, "loss": 0.9862, "step": 86 }, { "epoch": 5.176470588235294, "grad_norm": 0.6497894525527954, "learning_rate": 3.450980392156863e-06, "loss": 0.9873, "step": 88 }, { "epoch": 5.294117647058823, "grad_norm": 0.5283099412918091, "learning_rate": 3.529411764705883e-06, "loss": 0.9488, "step": 90 }, { "epoch": 5.411764705882353, "grad_norm": 0.4905494451522827, "learning_rate": 3.6078431372549024e-06, "loss": 0.9606, "step": 92 }, { "epoch": 5.529411764705882, "grad_norm": 0.5407238602638245, "learning_rate": 3.6862745098039223e-06, "loss": 0.888, "step": 94 }, { "epoch": 5.647058823529412, "grad_norm": 0.5735270977020264, "learning_rate": 3.7647058823529414e-06, "loss": 0.916, "step": 96 }, { "epoch": 5.764705882352941, "grad_norm": 0.5201605558395386, "learning_rate": 3.843137254901962e-06, "loss": 0.982, "step": 98 }, { "epoch": 5.882352941176471, "grad_norm": 0.4470115303993225, "learning_rate": 3.92156862745098e-06, "loss": 0.9389, "step": 100 }, { "epoch": 6.0, "grad_norm": 0.3844221532344818, "learning_rate": 4.000000000000001e-06, "loss": 1.0008, "step": 102 }, { "epoch": 6.0, "eval_loss": 0.937701940536499, "eval_runtime": 37.0036, "eval_samples_per_second": 0.973, "eval_steps_per_second": 0.973, "step": 102 }, { "epoch": 6.117647058823529, "grad_norm": 0.4652542173862457, "learning_rate": 4.07843137254902e-06, "loss": 0.915, "step": 104 }, { "epoch": 6.235294117647059, "grad_norm": 0.4144337773323059, "learning_rate": 4.15686274509804e-06, "loss": 0.8833, "step": 106 }, { "epoch": 6.352941176470588, "grad_norm": 0.4218851327896118, "learning_rate": 4.235294117647059e-06, "loss": 0.8672, "step": 108 }, { "epoch": 6.470588235294118, "grad_norm": 0.4070567786693573, "learning_rate": 4.313725490196079e-06, "loss": 0.9036, "step": 110 }, { "epoch": 6.588235294117647, "grad_norm": 0.36765947937965393, "learning_rate": 4.392156862745098e-06, "loss": 0.9384, "step": 112 }, { "epoch": 6.705882352941177, "grad_norm": 0.3613109886646271, "learning_rate": 4.4705882352941184e-06, "loss": 0.904, "step": 114 }, { "epoch": 6.823529411764706, "grad_norm": 0.36222243309020996, "learning_rate": 4.549019607843138e-06, "loss": 0.8641, "step": 116 }, { "epoch": 6.9411764705882355, "grad_norm": 0.42111915349960327, "learning_rate": 4.627450980392157e-06, "loss": 0.8935, "step": 118 }, { "epoch": 7.0, "eval_loss": 0.8938563466072083, "eval_runtime": 37.0541, "eval_samples_per_second": 0.972, "eval_steps_per_second": 0.972, "step": 119 }, { "epoch": 7.0588235294117645, "grad_norm": 0.35050034523010254, "learning_rate": 4.705882352941177e-06, "loss": 0.9626, "step": 120 }, { "epoch": 7.176470588235294, "grad_norm": 0.37680429220199585, "learning_rate": 4.784313725490196e-06, "loss": 0.8818, "step": 122 }, { "epoch": 7.294117647058823, "grad_norm": 0.3310965895652771, "learning_rate": 4.862745098039216e-06, "loss": 0.8631, "step": 124 }, { "epoch": 7.411764705882353, "grad_norm": 0.3843957781791687, "learning_rate": 4.941176470588236e-06, "loss": 0.8506, "step": 126 }, { "epoch": 7.529411764705882, "grad_norm": 0.3577839732170105, "learning_rate": 5.019607843137255e-06, "loss": 0.841, "step": 128 }, { "epoch": 7.647058823529412, "grad_norm": 0.3117690682411194, "learning_rate": 5.098039215686274e-06, "loss": 0.8576, "step": 130 }, { "epoch": 7.764705882352941, "grad_norm": 0.37441667914390564, "learning_rate": 5.176470588235295e-06, "loss": 0.8127, "step": 132 }, { "epoch": 7.882352941176471, "grad_norm": 0.3178256154060364, "learning_rate": 5.254901960784314e-06, "loss": 0.8554, "step": 134 }, { "epoch": 8.0, "grad_norm": 0.3216207027435303, "learning_rate": 5.333333333333334e-06, "loss": 0.8862, "step": 136 }, { "epoch": 8.0, "eval_loss": 0.8579282760620117, "eval_runtime": 37.0044, "eval_samples_per_second": 0.973, "eval_steps_per_second": 0.973, "step": 136 }, { "epoch": 8.117647058823529, "grad_norm": 0.341949462890625, "learning_rate": 5.411764705882353e-06, "loss": 0.8519, "step": 138 }, { "epoch": 8.235294117647058, "grad_norm": 0.3150944113731384, "learning_rate": 5.4901960784313735e-06, "loss": 0.8338, "step": 140 }, { "epoch": 8.352941176470589, "grad_norm": 0.3646303713321686, "learning_rate": 5.568627450980393e-06, "loss": 0.8381, "step": 142 }, { "epoch": 8.470588235294118, "grad_norm": 0.334943950176239, "learning_rate": 5.6470588235294125e-06, "loss": 0.879, "step": 144 }, { "epoch": 8.588235294117647, "grad_norm": 0.3029642403125763, "learning_rate": 5.725490196078431e-06, "loss": 0.7807, "step": 146 }, { "epoch": 8.705882352941176, "grad_norm": 0.30906420946121216, "learning_rate": 5.803921568627452e-06, "loss": 0.7946, "step": 148 }, { "epoch": 8.823529411764707, "grad_norm": 0.34527918696403503, "learning_rate": 5.882352941176471e-06, "loss": 0.8397, "step": 150 }, { "epoch": 8.941176470588236, "grad_norm": 0.3913155794143677, "learning_rate": 5.96078431372549e-06, "loss": 0.8266, "step": 152 }, { "epoch": 9.0, "eval_loss": 0.8293797969818115, "eval_runtime": 37.0023, "eval_samples_per_second": 0.973, "eval_steps_per_second": 0.973, "step": 153 }, { "epoch": 9.058823529411764, "grad_norm": 0.3411412835121155, "learning_rate": 6.03921568627451e-06, "loss": 0.8249, "step": 154 }, { "epoch": 9.176470588235293, "grad_norm": 0.28964897990226746, "learning_rate": 6.11764705882353e-06, "loss": 0.7887, "step": 156 }, { "epoch": 9.294117647058824, "grad_norm": 0.36540400981903076, "learning_rate": 6.19607843137255e-06, "loss": 0.7434, "step": 158 }, { "epoch": 9.411764705882353, "grad_norm": 0.3152979016304016, "learning_rate": 6.274509803921569e-06, "loss": 0.7864, "step": 160 }, { "epoch": 9.529411764705882, "grad_norm": 0.28152012825012207, "learning_rate": 6.352941176470589e-06, "loss": 0.777, "step": 162 }, { "epoch": 9.647058823529411, "grad_norm": 0.37814223766326904, "learning_rate": 6.431372549019609e-06, "loss": 0.8319, "step": 164 }, { "epoch": 9.764705882352942, "grad_norm": 0.3085719645023346, "learning_rate": 6.5098039215686285e-06, "loss": 0.7986, "step": 166 }, { "epoch": 9.882352941176471, "grad_norm": 0.29185226559638977, "learning_rate": 6.588235294117647e-06, "loss": 0.8292, "step": 168 }, { "epoch": 10.0, "grad_norm": 0.30540931224823, "learning_rate": 6.666666666666667e-06, "loss": 0.7797, "step": 170 }, { "epoch": 10.0, "eval_loss": 0.807489275932312, "eval_runtime": 37.0049, "eval_samples_per_second": 0.973, "eval_steps_per_second": 0.973, "step": 170 }, { "epoch": 10.117647058823529, "grad_norm": 0.3091186285018921, "learning_rate": 6.745098039215687e-06, "loss": 0.7765, "step": 172 }, { "epoch": 10.235294117647058, "grad_norm": 0.30727335810661316, "learning_rate": 6.8235294117647065e-06, "loss": 0.7505, "step": 174 }, { "epoch": 10.352941176470589, "grad_norm": 0.2999902665615082, "learning_rate": 6.901960784313726e-06, "loss": 0.7233, "step": 176 }, { "epoch": 10.470588235294118, "grad_norm": 0.2925303876399994, "learning_rate": 6.9803921568627454e-06, "loss": 0.7613, "step": 178 }, { "epoch": 10.588235294117647, "grad_norm": 0.28137609362602234, "learning_rate": 7.058823529411766e-06, "loss": 0.7336, "step": 180 }, { "epoch": 10.705882352941176, "grad_norm": 0.3438097834587097, "learning_rate": 7.137254901960785e-06, "loss": 0.822, "step": 182 }, { "epoch": 10.823529411764707, "grad_norm": 0.3309740722179413, "learning_rate": 7.215686274509805e-06, "loss": 0.8488, "step": 184 }, { "epoch": 10.941176470588236, "grad_norm": 0.33822691440582275, "learning_rate": 7.294117647058823e-06, "loss": 0.8158, "step": 186 }, { "epoch": 11.0, "eval_loss": 0.790329098701477, "eval_runtime": 37.0086, "eval_samples_per_second": 0.973, "eval_steps_per_second": 0.973, "step": 187 }, { "epoch": 11.058823529411764, "grad_norm": 0.3196800649166107, "learning_rate": 7.372549019607845e-06, "loss": 0.7996, "step": 188 }, { "epoch": 11.176470588235293, "grad_norm": 0.2967463731765747, "learning_rate": 7.450980392156863e-06, "loss": 0.7102, "step": 190 }, { "epoch": 11.294117647058824, "grad_norm": 0.32051464915275574, "learning_rate": 7.529411764705883e-06, "loss": 0.8008, "step": 192 }, { "epoch": 11.411764705882353, "grad_norm": 0.399475634098053, "learning_rate": 7.607843137254902e-06, "loss": 0.8271, "step": 194 }, { "epoch": 11.529411764705882, "grad_norm": 0.3100043535232544, "learning_rate": 7.686274509803923e-06, "loss": 0.7954, "step": 196 }, { "epoch": 11.647058823529411, "grad_norm": 0.31522953510284424, "learning_rate": 7.764705882352941e-06, "loss": 0.712, "step": 198 }, { "epoch": 11.764705882352942, "grad_norm": 0.3450244665145874, "learning_rate": 7.84313725490196e-06, "loss": 0.7329, "step": 200 }, { "epoch": 11.882352941176471, "grad_norm": 0.3267311453819275, "learning_rate": 7.92156862745098e-06, "loss": 0.7275, "step": 202 }, { "epoch": 12.0, "grad_norm": 0.42111936211586, "learning_rate": 8.000000000000001e-06, "loss": 0.6845, "step": 204 }, { "epoch": 12.0, "eval_loss": 0.7741987705230713, "eval_runtime": 37.0028, "eval_samples_per_second": 0.973, "eval_steps_per_second": 0.973, "step": 204 }, { "epoch": 12.117647058823529, "grad_norm": 0.33904674649238586, "learning_rate": 8.07843137254902e-06, "loss": 0.7635, "step": 206 }, { "epoch": 12.235294117647058, "grad_norm": 0.33812686800956726, "learning_rate": 8.15686274509804e-06, "loss": 0.8239, "step": 208 }, { "epoch": 12.352941176470589, "grad_norm": 0.34535664319992065, "learning_rate": 8.23529411764706e-06, "loss": 0.7084, "step": 210 }, { "epoch": 12.470588235294118, "grad_norm": 0.330837219953537, "learning_rate": 8.31372549019608e-06, "loss": 0.6769, "step": 212 }, { "epoch": 12.588235294117647, "grad_norm": 0.31368908286094666, "learning_rate": 8.392156862745099e-06, "loss": 0.7651, "step": 214 }, { "epoch": 12.705882352941176, "grad_norm": 0.35061123967170715, "learning_rate": 8.470588235294118e-06, "loss": 0.7452, "step": 216 }, { "epoch": 12.823529411764707, "grad_norm": 0.32111451029777527, "learning_rate": 8.549019607843138e-06, "loss": 0.718, "step": 218 }, { "epoch": 12.941176470588236, "grad_norm": 0.3150181174278259, "learning_rate": 8.627450980392157e-06, "loss": 0.6819, "step": 220 }, { "epoch": 13.0, "eval_loss": 0.7598351240158081, "eval_runtime": 37.0068, "eval_samples_per_second": 0.973, "eval_steps_per_second": 0.973, "step": 221 }, { "epoch": 13.058823529411764, "grad_norm": 0.3078894019126892, "learning_rate": 8.705882352941177e-06, "loss": 0.7151, "step": 222 }, { "epoch": 13.176470588235293, "grad_norm": 0.3273066282272339, "learning_rate": 8.784313725490196e-06, "loss": 0.6766, "step": 224 }, { "epoch": 13.294117647058824, "grad_norm": 0.30240604281425476, "learning_rate": 8.862745098039216e-06, "loss": 0.665, "step": 226 }, { "epoch": 13.411764705882353, "grad_norm": 0.3365361988544464, "learning_rate": 8.941176470588237e-06, "loss": 0.73, "step": 228 }, { "epoch": 13.529411764705882, "grad_norm": 0.3776240646839142, "learning_rate": 9.019607843137256e-06, "loss": 0.7319, "step": 230 }, { "epoch": 13.647058823529411, "grad_norm": 0.32826074957847595, "learning_rate": 9.098039215686276e-06, "loss": 0.7021, "step": 232 }, { "epoch": 13.764705882352942, "grad_norm": 0.3192596435546875, "learning_rate": 9.176470588235294e-06, "loss": 0.7333, "step": 234 }, { "epoch": 13.882352941176471, "grad_norm": 0.36111271381378174, "learning_rate": 9.254901960784315e-06, "loss": 0.7633, "step": 236 }, { "epoch": 14.0, "grad_norm": 0.3694405257701874, "learning_rate": 9.333333333333334e-06, "loss": 0.7241, "step": 238 }, { "epoch": 14.0, "eval_loss": 0.7471711039543152, "eval_runtime": 37.0047, "eval_samples_per_second": 0.973, "eval_steps_per_second": 0.973, "step": 238 }, { "epoch": 14.117647058823529, "grad_norm": 0.3688935339450836, "learning_rate": 9.411764705882354e-06, "loss": 0.742, "step": 240 }, { "epoch": 14.235294117647058, "grad_norm": 0.41638877987861633, "learning_rate": 9.490196078431373e-06, "loss": 0.7049, "step": 242 }, { "epoch": 14.352941176470589, "grad_norm": 0.3489099442958832, "learning_rate": 9.568627450980393e-06, "loss": 0.6517, "step": 244 }, { "epoch": 14.470588235294118, "grad_norm": 0.3671415448188782, "learning_rate": 9.647058823529412e-06, "loss": 0.7233, "step": 246 }, { "epoch": 14.588235294117647, "grad_norm": 0.37552204728126526, "learning_rate": 9.725490196078432e-06, "loss": 0.701, "step": 248 }, { "epoch": 14.705882352941176, "grad_norm": 0.43261757493019104, "learning_rate": 9.803921568627451e-06, "loss": 0.6979, "step": 250 }, { "epoch": 14.823529411764707, "grad_norm": 0.37599849700927734, "learning_rate": 9.882352941176472e-06, "loss": 0.6462, "step": 252 }, { "epoch": 14.941176470588236, "grad_norm": 0.3583243191242218, "learning_rate": 9.960784313725492e-06, "loss": 0.695, "step": 254 }, { "epoch": 15.0, "eval_loss": 0.7365108132362366, "eval_runtime": 37.003, "eval_samples_per_second": 0.973, "eval_steps_per_second": 0.973, "step": 255 }, { "epoch": 15.058823529411764, "grad_norm": 0.4206070005893707, "learning_rate": 9.999995315380667e-06, "loss": 0.6862, "step": 256 }, { "epoch": 15.176470588235293, "grad_norm": 0.47259286046028137, "learning_rate": 9.99995783847866e-06, "loss": 0.6933, "step": 258 }, { "epoch": 15.294117647058824, "grad_norm": 0.3962372839450836, "learning_rate": 9.999882884955554e-06, "loss": 0.6643, "step": 260 }, { "epoch": 15.411764705882353, "grad_norm": 0.37228551506996155, "learning_rate": 9.99977045537315e-06, "loss": 0.6676, "step": 262 }, { "epoch": 15.529411764705882, "grad_norm": 0.4293304979801178, "learning_rate": 9.999620550574155e-06, "loss": 0.6497, "step": 264 }, { "epoch": 15.647058823529411, "grad_norm": 0.4316819906234741, "learning_rate": 9.999433171682158e-06, "loss": 0.6687, "step": 266 }, { "epoch": 15.764705882352942, "grad_norm": 0.4474015533924103, "learning_rate": 9.999208320101643e-06, "loss": 0.6935, "step": 268 }, { "epoch": 15.882352941176471, "grad_norm": 0.4514879584312439, "learning_rate": 9.998945997517957e-06, "loss": 0.6542, "step": 270 }, { "epoch": 16.0, "grad_norm": 0.5015867948532104, "learning_rate": 9.99864620589731e-06, "loss": 0.6982, "step": 272 }, { "epoch": 16.0, "eval_loss": 0.7271888256072998, "eval_runtime": 37.0158, "eval_samples_per_second": 0.973, "eval_steps_per_second": 0.973, "step": 272 }, { "epoch": 16.11764705882353, "grad_norm": 0.5128858685493469, "learning_rate": 9.998308947486753e-06, "loss": 0.6414, "step": 274 }, { "epoch": 16.235294117647058, "grad_norm": 0.40119990706443787, "learning_rate": 9.997934224814173e-06, "loss": 0.628, "step": 276 }, { "epoch": 16.352941176470587, "grad_norm": 0.4133683145046234, "learning_rate": 9.997522040688258e-06, "loss": 0.7072, "step": 278 }, { "epoch": 16.470588235294116, "grad_norm": 0.41614237427711487, "learning_rate": 9.997072398198492e-06, "loss": 0.6526, "step": 280 }, { "epoch": 16.58823529411765, "grad_norm": 0.4795832931995392, "learning_rate": 9.996585300715117e-06, "loss": 0.6583, "step": 282 }, { "epoch": 16.705882352941178, "grad_norm": 0.4348108470439911, "learning_rate": 9.996060751889114e-06, "loss": 0.6728, "step": 284 }, { "epoch": 16.823529411764707, "grad_norm": 0.4476607143878937, "learning_rate": 9.995498755652186e-06, "loss": 0.626, "step": 286 }, { "epoch": 16.941176470588236, "grad_norm": 0.4528913199901581, "learning_rate": 9.994899316216709e-06, "loss": 0.622, "step": 288 }, { "epoch": 17.0, "eval_loss": 0.7215262055397034, "eval_runtime": 36.9993, "eval_samples_per_second": 0.973, "eval_steps_per_second": 0.973, "step": 289 }, { "epoch": 17.058823529411764, "grad_norm": 0.5009466409683228, "learning_rate": 9.994262438075713e-06, "loss": 0.7081, "step": 290 }, { "epoch": 17.176470588235293, "grad_norm": 0.475374698638916, "learning_rate": 9.993588126002848e-06, "loss": 0.6229, "step": 292 }, { "epoch": 17.294117647058822, "grad_norm": 0.4642236828804016, "learning_rate": 9.992876385052346e-06, "loss": 0.5844, "step": 294 }, { "epoch": 17.41176470588235, "grad_norm": 0.4536580741405487, "learning_rate": 9.992127220558976e-06, "loss": 0.6135, "step": 296 }, { "epoch": 17.529411764705884, "grad_norm": 0.45987674593925476, "learning_rate": 9.991340638138022e-06, "loss": 0.6511, "step": 298 }, { "epoch": 17.647058823529413, "grad_norm": 0.4978199899196625, "learning_rate": 9.990516643685222e-06, "loss": 0.6424, "step": 300 }, { "epoch": 17.764705882352942, "grad_norm": 0.49036839604377747, "learning_rate": 9.98965524337673e-06, "loss": 0.6773, "step": 302 }, { "epoch": 17.88235294117647, "grad_norm": 0.5300158858299255, "learning_rate": 9.988756443669081e-06, "loss": 0.6364, "step": 304 }, { "epoch": 18.0, "grad_norm": 0.4950679540634155, "learning_rate": 9.987820251299121e-06, "loss": 0.5905, "step": 306 }, { "epoch": 18.0, "eval_loss": 0.7155699133872986, "eval_runtime": 37.0042, "eval_samples_per_second": 0.973, "eval_steps_per_second": 0.973, "step": 306 }, { "epoch": 18.11764705882353, "grad_norm": 0.5069499611854553, "learning_rate": 9.98684667328398e-06, "loss": 0.6123, "step": 308 }, { "epoch": 18.235294117647058, "grad_norm": 0.46550482511520386, "learning_rate": 9.985835716921e-06, "loss": 0.6102, "step": 310 }, { "epoch": 18.352941176470587, "grad_norm": 0.7860422730445862, "learning_rate": 9.984787389787689e-06, "loss": 0.6273, "step": 312 }, { "epoch": 18.470588235294116, "grad_norm": 0.49825742840766907, "learning_rate": 9.983701699741668e-06, "loss": 0.6297, "step": 314 }, { "epoch": 18.58823529411765, "grad_norm": 0.5290573239326477, "learning_rate": 9.982578654920601e-06, "loss": 0.6238, "step": 316 }, { "epoch": 18.705882352941178, "grad_norm": 0.5185476541519165, "learning_rate": 9.981418263742148e-06, "loss": 0.5702, "step": 318 }, { "epoch": 18.823529411764707, "grad_norm": 0.5360794067382812, "learning_rate": 9.980220534903889e-06, "loss": 0.6082, "step": 320 }, { "epoch": 18.941176470588236, "grad_norm": 0.5598728656768799, "learning_rate": 9.978985477383264e-06, "loss": 0.6121, "step": 322 }, { "epoch": 19.0, "eval_loss": 0.7140286564826965, "eval_runtime": 36.9943, "eval_samples_per_second": 0.973, "eval_steps_per_second": 0.973, "step": 323 }, { "epoch": 19.058823529411764, "grad_norm": 0.5005812048912048, "learning_rate": 9.97771310043751e-06, "loss": 0.5964, "step": 324 }, { "epoch": 19.176470588235293, "grad_norm": 0.5243679881095886, "learning_rate": 9.97640341360358e-06, "loss": 0.643, "step": 326 }, { "epoch": 19.294117647058822, "grad_norm": 0.5746588706970215, "learning_rate": 9.975056426698094e-06, "loss": 0.587, "step": 328 }, { "epoch": 19.41176470588235, "grad_norm": 0.5213414430618286, "learning_rate": 9.973672149817232e-06, "loss": 0.5835, "step": 330 }, { "epoch": 19.529411764705884, "grad_norm": 0.6261019706726074, "learning_rate": 9.972250593336689e-06, "loss": 0.5884, "step": 332 }, { "epoch": 19.647058823529413, "grad_norm": 0.5905741453170776, "learning_rate": 9.970791767911581e-06, "loss": 0.588, "step": 334 }, { "epoch": 19.764705882352942, "grad_norm": 0.5681843757629395, "learning_rate": 9.96929568447637e-06, "loss": 0.554, "step": 336 }, { "epoch": 19.88235294117647, "grad_norm": 0.6130782961845398, "learning_rate": 9.967762354244778e-06, "loss": 0.6003, "step": 338 }, { "epoch": 20.0, "grad_norm": 0.6039074659347534, "learning_rate": 9.966191788709716e-06, "loss": 0.567, "step": 340 }, { "epoch": 20.0, "eval_loss": 0.7166150808334351, "eval_runtime": 36.9981, "eval_samples_per_second": 0.973, "eval_steps_per_second": 0.973, "step": 340 }, { "epoch": 20.11764705882353, "grad_norm": 0.661340594291687, "learning_rate": 9.964583999643174e-06, "loss": 0.5771, "step": 342 }, { "epoch": 20.235294117647058, "grad_norm": 0.6773268580436707, "learning_rate": 9.962938999096159e-06, "loss": 0.5611, "step": 344 }, { "epoch": 20.352941176470587, "grad_norm": 0.6907221674919128, "learning_rate": 9.961256799398584e-06, "loss": 0.5961, "step": 346 }, { "epoch": 20.470588235294116, "grad_norm": 0.6799242496490479, "learning_rate": 9.95953741315919e-06, "loss": 0.605, "step": 348 }, { "epoch": 20.58823529411765, "grad_norm": 0.6507272720336914, "learning_rate": 9.957780853265441e-06, "loss": 0.5564, "step": 350 }, { "epoch": 20.705882352941178, "grad_norm": 0.7043225765228271, "learning_rate": 9.955987132883435e-06, "loss": 0.5069, "step": 352 }, { "epoch": 20.823529411764707, "grad_norm": 0.7170704007148743, "learning_rate": 9.954156265457801e-06, "loss": 0.5671, "step": 354 }, { "epoch": 20.941176470588236, "grad_norm": 0.6098182797431946, "learning_rate": 9.952288264711601e-06, "loss": 0.5471, "step": 356 }, { "epoch": 21.0, "eval_loss": 0.7171775698661804, "eval_runtime": 36.9995, "eval_samples_per_second": 0.973, "eval_steps_per_second": 0.973, "step": 357 }, { "epoch": 21.058823529411764, "grad_norm": 0.7199786305427551, "learning_rate": 9.950383144646221e-06, "loss": 0.5856, "step": 358 }, { "epoch": 21.176470588235293, "grad_norm": 0.748616635799408, "learning_rate": 9.948440919541277e-06, "loss": 0.5347, "step": 360 }, { "epoch": 21.294117647058822, "grad_norm": 0.7035663723945618, "learning_rate": 9.946461603954499e-06, "loss": 0.5433, "step": 362 }, { "epoch": 21.41176470588235, "grad_norm": 0.7293503880500793, "learning_rate": 9.944445212721619e-06, "loss": 0.5473, "step": 364 }, { "epoch": 21.529411764705884, "grad_norm": 0.7407766580581665, "learning_rate": 9.942391760956277e-06, "loss": 0.527, "step": 366 }, { "epoch": 21.647058823529413, "grad_norm": 0.7856358289718628, "learning_rate": 9.940301264049885e-06, "loss": 0.5372, "step": 368 }, { "epoch": 21.764705882352942, "grad_norm": 0.7008533477783203, "learning_rate": 9.938173737671531e-06, "loss": 0.5375, "step": 370 }, { "epoch": 21.88235294117647, "grad_norm": 0.7586143612861633, "learning_rate": 9.936009197767847e-06, "loss": 0.5467, "step": 372 }, { "epoch": 22.0, "grad_norm": 0.8423267602920532, "learning_rate": 9.933807660562898e-06, "loss": 0.4761, "step": 374 }, { "epoch": 22.0, "eval_loss": 0.7233606576919556, "eval_runtime": 37.0054, "eval_samples_per_second": 0.973, "eval_steps_per_second": 0.973, "step": 374 }, { "epoch": 22.11764705882353, "grad_norm": 0.882168710231781, "learning_rate": 9.931569142558057e-06, "loss": 0.562, "step": 376 }, { "epoch": 22.235294117647058, "grad_norm": 0.8616783618927002, "learning_rate": 9.929293660531889e-06, "loss": 0.5039, "step": 378 }, { "epoch": 22.352941176470587, "grad_norm": 0.9095832705497742, "learning_rate": 9.926981231540007e-06, "loss": 0.4744, "step": 380 }, { "epoch": 22.470588235294116, "grad_norm": 0.8473853468894958, "learning_rate": 9.924631872914967e-06, "loss": 0.4976, "step": 382 }, { "epoch": 22.58823529411765, "grad_norm": 0.844835102558136, "learning_rate": 9.922245602266119e-06, "loss": 0.4721, "step": 384 }, { "epoch": 22.705882352941178, "grad_norm": 0.820750892162323, "learning_rate": 9.919822437479488e-06, "loss": 0.4995, "step": 386 }, { "epoch": 22.823529411764707, "grad_norm": 0.823648989200592, "learning_rate": 9.91736239671763e-06, "loss": 0.5061, "step": 388 }, { "epoch": 22.941176470588236, "grad_norm": 0.8103899359703064, "learning_rate": 9.91486549841951e-06, "loss": 0.4967, "step": 390 }, { "epoch": 23.0, "eval_loss": 0.7357540726661682, "eval_runtime": 37.0101, "eval_samples_per_second": 0.973, "eval_steps_per_second": 0.973, "step": 391 }, { "epoch": 23.058823529411764, "grad_norm": 0.7696477174758911, "learning_rate": 9.912331761300341e-06, "loss": 0.5004, "step": 392 }, { "epoch": 23.176470588235293, "grad_norm": 0.7803938984870911, "learning_rate": 9.909761204351469e-06, "loss": 0.4903, "step": 394 }, { "epoch": 23.294117647058822, "grad_norm": 0.9065766930580139, "learning_rate": 9.90715384684021e-06, "loss": 0.505, "step": 396 }, { "epoch": 23.41176470588235, "grad_norm": 1.1152830123901367, "learning_rate": 9.904509708309723e-06, "loss": 0.4312, "step": 398 }, { "epoch": 23.529411764705884, "grad_norm": 1.0776093006134033, "learning_rate": 9.901828808578846e-06, "loss": 0.4503, "step": 400 }, { "epoch": 23.647058823529413, "grad_norm": 1.0322400331497192, "learning_rate": 9.899111167741966e-06, "loss": 0.4685, "step": 402 }, { "epoch": 23.764705882352942, "grad_norm": 0.9414217472076416, "learning_rate": 9.896356806168851e-06, "loss": 0.4757, "step": 404 }, { "epoch": 23.88235294117647, "grad_norm": 1.1830779314041138, "learning_rate": 9.89356574450451e-06, "loss": 0.4408, "step": 406 }, { "epoch": 24.0, "grad_norm": 0.9795882105827332, "learning_rate": 9.890738003669029e-06, "loss": 0.4833, "step": 408 }, { "epoch": 24.0, "eval_loss": 0.7643568515777588, "eval_runtime": 37.0004, "eval_samples_per_second": 0.973, "eval_steps_per_second": 0.973, "step": 408 }, { "epoch": 24.11764705882353, "grad_norm": 1.2952356338500977, "learning_rate": 9.887873604857424e-06, "loss": 0.4299, "step": 410 }, { "epoch": 24.235294117647058, "grad_norm": 1.2539509534835815, "learning_rate": 9.884972569539471e-06, "loss": 0.4488, "step": 412 }, { "epoch": 24.352941176470587, "grad_norm": 1.0663094520568848, "learning_rate": 9.882034919459556e-06, "loss": 0.4152, "step": 414 }, { "epoch": 24.470588235294116, "grad_norm": 1.03883957862854, "learning_rate": 9.879060676636502e-06, "loss": 0.4435, "step": 416 }, { "epoch": 24.58823529411765, "grad_norm": 0.9475598931312561, "learning_rate": 9.876049863363415e-06, "loss": 0.4408, "step": 418 }, { "epoch": 24.705882352941178, "grad_norm": 0.9946249723434448, "learning_rate": 9.873002502207502e-06, "loss": 0.4292, "step": 420 }, { "epoch": 24.823529411764707, "grad_norm": 1.135332465171814, "learning_rate": 9.86991861600992e-06, "loss": 0.4057, "step": 422 }, { "epoch": 24.941176470588236, "grad_norm": 1.7003220319747925, "learning_rate": 9.866798227885588e-06, "loss": 0.4071, "step": 424 }, { "epoch": 25.0, "eval_loss": 0.8012258410453796, "eval_runtime": 37.0059, "eval_samples_per_second": 0.973, "eval_steps_per_second": 0.973, "step": 425 }, { "epoch": 25.058823529411764, "grad_norm": 1.124939203262329, "learning_rate": 9.863641361223025e-06, "loss": 0.3795, "step": 426 }, { "epoch": 25.176470588235293, "grad_norm": 1.2946455478668213, "learning_rate": 9.860448039684169e-06, "loss": 0.3631, "step": 428 }, { "epoch": 25.294117647058822, "grad_norm": 1.2063255310058594, "learning_rate": 9.857218287204204e-06, "loss": 0.3786, "step": 430 }, { "epoch": 25.41176470588235, "grad_norm": 1.1833475828170776, "learning_rate": 9.853952127991374e-06, "loss": 0.403, "step": 432 }, { "epoch": 25.529411764705884, "grad_norm": 1.2458750009536743, "learning_rate": 9.850649586526808e-06, "loss": 0.3626, "step": 434 }, { "epoch": 25.647058823529413, "grad_norm": 1.2201582193374634, "learning_rate": 9.847310687564335e-06, "loss": 0.3774, "step": 436 }, { "epoch": 25.764705882352942, "grad_norm": 1.2703592777252197, "learning_rate": 9.843935456130295e-06, "loss": 0.4078, "step": 438 }, { "epoch": 25.88235294117647, "grad_norm": 1.8023614883422852, "learning_rate": 9.840523917523354e-06, "loss": 0.406, "step": 440 }, { "epoch": 26.0, "grad_norm": 1.3355423212051392, "learning_rate": 9.83707609731432e-06, "loss": 0.3567, "step": 442 }, { "epoch": 26.0, "eval_loss": 0.8288628458976746, "eval_runtime": 37.0058, "eval_samples_per_second": 0.973, "eval_steps_per_second": 0.973, "step": 442 }, { "epoch": 26.0, "step": 442, "total_flos": 3.38458179962667e+17, "train_loss": 0.7497456199173475, "train_runtime": 12081.7026, "train_samples_per_second": 1.689, "train_steps_per_second": 0.211 } ], "logging_steps": 2, "max_steps": 2550, "num_input_tokens_seen": 0, "num_train_epochs": 150, "save_steps": 25, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 7, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 0 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3.38458179962667e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }