{ "best_metric": 0.10599859803915024, "best_model_checkpoint": "riskmanagementv1/checkpoint-155", "epoch": 5.0, "eval_steps": 500, "global_step": 155, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.03225806451612903, "grad_norm": 1.648207187652588, "learning_rate": 6.25e-07, "loss": 1.0951, "step": 1 }, { "epoch": 0.06451612903225806, "grad_norm": 1.5398106575012207, "learning_rate": 1.25e-06, "loss": 1.116, "step": 2 }, { "epoch": 0.0967741935483871, "grad_norm": 4.653311729431152, "learning_rate": 1.8750000000000003e-06, "loss": 1.0443, "step": 3 }, { "epoch": 0.12903225806451613, "grad_norm": 3.8882646560668945, "learning_rate": 2.5e-06, "loss": 1.113, "step": 4 }, { "epoch": 0.16129032258064516, "grad_norm": 3.754126787185669, "learning_rate": 3.125e-06, "loss": 1.1012, "step": 5 }, { "epoch": 0.1935483870967742, "grad_norm": 1.5717527866363525, "learning_rate": 3.7500000000000005e-06, "loss": 1.0946, "step": 6 }, { "epoch": 0.22580645161290322, "grad_norm": 2.106679677963257, "learning_rate": 4.3750000000000005e-06, "loss": 1.0793, "step": 7 }, { "epoch": 0.25806451612903225, "grad_norm": 2.2854700088500977, "learning_rate": 5e-06, "loss": 1.1227, "step": 8 }, { "epoch": 0.2903225806451613, "grad_norm": 2.0785741806030273, "learning_rate": 5.625e-06, "loss": 1.0784, "step": 9 }, { "epoch": 0.3225806451612903, "grad_norm": 3.856192111968994, "learning_rate": 6.25e-06, "loss": 1.1229, "step": 10 }, { "epoch": 0.3548387096774194, "grad_norm": 2.9804818630218506, "learning_rate": 6.875e-06, "loss": 1.0941, "step": 11 }, { "epoch": 0.3870967741935484, "grad_norm": 2.701162099838257, "learning_rate": 7.500000000000001e-06, "loss": 1.0792, "step": 12 }, { "epoch": 0.41935483870967744, "grad_norm": 5.755159854888916, "learning_rate": 8.125000000000001e-06, "loss": 1.08, "step": 13 }, { "epoch": 0.45161290322580644, "grad_norm": 4.275572299957275, "learning_rate": 8.750000000000001e-06, "loss": 1.0991, "step": 14 }, { "epoch": 0.4838709677419355, "grad_norm": 1.5670201778411865, "learning_rate": 9.375000000000001e-06, "loss": 1.0728, "step": 15 }, { "epoch": 0.5161290322580645, "grad_norm": 2.7937171459198, "learning_rate": 1e-05, "loss": 1.1006, "step": 16 }, { "epoch": 0.5483870967741935, "grad_norm": 2.048818349838257, "learning_rate": 9.928057553956835e-06, "loss": 1.0779, "step": 17 }, { "epoch": 0.5806451612903226, "grad_norm": 4.063783168792725, "learning_rate": 9.85611510791367e-06, "loss": 1.0858, "step": 18 }, { "epoch": 0.6129032258064516, "grad_norm": 2.6299703121185303, "learning_rate": 9.784172661870505e-06, "loss": 1.0715, "step": 19 }, { "epoch": 0.6451612903225806, "grad_norm": 2.324522018432617, "learning_rate": 9.712230215827338e-06, "loss": 1.0964, "step": 20 }, { "epoch": 0.6774193548387096, "grad_norm": 2.1976382732391357, "learning_rate": 9.640287769784174e-06, "loss": 1.032, "step": 21 }, { "epoch": 0.7096774193548387, "grad_norm": 2.2792370319366455, "learning_rate": 9.568345323741008e-06, "loss": 1.0902, "step": 22 }, { "epoch": 0.7419354838709677, "grad_norm": 1.996429204940796, "learning_rate": 9.496402877697842e-06, "loss": 1.0665, "step": 23 }, { "epoch": 0.7741935483870968, "grad_norm": 4.1794891357421875, "learning_rate": 9.424460431654678e-06, "loss": 1.0737, "step": 24 }, { "epoch": 0.8064516129032258, "grad_norm": 4.505804061889648, "learning_rate": 9.35251798561151e-06, "loss": 1.0948, "step": 25 }, { "epoch": 0.8387096774193549, "grad_norm": 3.4325923919677734, "learning_rate": 9.280575539568346e-06, "loss": 1.0494, "step": 26 }, { "epoch": 0.8709677419354839, "grad_norm": 2.4758219718933105, "learning_rate": 9.20863309352518e-06, "loss": 1.052, "step": 27 }, { "epoch": 0.9032258064516129, "grad_norm": 3.858365297317505, "learning_rate": 9.136690647482015e-06, "loss": 1.051, "step": 28 }, { "epoch": 0.9354838709677419, "grad_norm": 2.598480701446533, "learning_rate": 9.064748201438849e-06, "loss": 1.0568, "step": 29 }, { "epoch": 0.967741935483871, "grad_norm": 1.8019946813583374, "learning_rate": 8.992805755395683e-06, "loss": 1.0231, "step": 30 }, { "epoch": 1.0, "grad_norm": 3.940290927886963, "learning_rate": 8.92086330935252e-06, "loss": 1.0415, "step": 31 }, { "epoch": 1.0, "eval_accuracy": 0.9032258064516129, "eval_f1_macro": 0.9035612535612536, "eval_f1_micro": 0.9032258064516129, "eval_f1_weighted": 0.9040736145574856, "eval_loss": 1.0341954231262207, "eval_precision_macro": 0.9093567251461989, "eval_precision_micro": 0.9032258064516129, "eval_precision_weighted": 0.9095925297113752, "eval_recall_macro": 0.9023809523809523, "eval_recall_micro": 0.9032258064516129, "eval_recall_weighted": 0.9032258064516129, "eval_runtime": 0.3613, "eval_samples_per_second": 171.597, "eval_steps_per_second": 11.071, "step": 31 }, { "epoch": 1.032258064516129, "grad_norm": 4.133128643035889, "learning_rate": 8.848920863309353e-06, "loss": 1.0548, "step": 32 }, { "epoch": 1.064516129032258, "grad_norm": 3.782399892807007, "learning_rate": 8.776978417266188e-06, "loss": 1.0703, "step": 33 }, { "epoch": 1.096774193548387, "grad_norm": 4.876072883605957, "learning_rate": 8.705035971223022e-06, "loss": 1.0334, "step": 34 }, { "epoch": 1.129032258064516, "grad_norm": 4.667634963989258, "learning_rate": 8.633093525179856e-06, "loss": 0.9948, "step": 35 }, { "epoch": 1.1612903225806452, "grad_norm": 2.1760268211364746, "learning_rate": 8.561151079136692e-06, "loss": 0.972, "step": 36 }, { "epoch": 1.1935483870967742, "grad_norm": 4.767606735229492, "learning_rate": 8.489208633093526e-06, "loss": 0.9336, "step": 37 }, { "epoch": 1.2258064516129032, "grad_norm": 7.815669536590576, "learning_rate": 8.41726618705036e-06, "loss": 0.9686, "step": 38 }, { "epoch": 1.2580645161290323, "grad_norm": 4.481734752655029, "learning_rate": 8.345323741007195e-06, "loss": 0.9709, "step": 39 }, { "epoch": 1.2903225806451613, "grad_norm": 2.848688840866089, "learning_rate": 8.273381294964029e-06, "loss": 0.962, "step": 40 }, { "epoch": 1.3225806451612903, "grad_norm": 4.002137184143066, "learning_rate": 8.201438848920865e-06, "loss": 0.937, "step": 41 }, { "epoch": 1.3548387096774195, "grad_norm": 4.138007640838623, "learning_rate": 8.129496402877699e-06, "loss": 0.9601, "step": 42 }, { "epoch": 1.3870967741935485, "grad_norm": 4.144357681274414, "learning_rate": 8.057553956834533e-06, "loss": 0.9457, "step": 43 }, { "epoch": 1.4193548387096775, "grad_norm": 5.553256034851074, "learning_rate": 7.985611510791367e-06, "loss": 0.9491, "step": 44 }, { "epoch": 1.4516129032258065, "grad_norm": 3.699176549911499, "learning_rate": 7.913669064748202e-06, "loss": 0.9378, "step": 45 }, { "epoch": 1.4838709677419355, "grad_norm": 4.687703609466553, "learning_rate": 7.841726618705036e-06, "loss": 0.8746, "step": 46 }, { "epoch": 1.5161290322580645, "grad_norm": 4.312544822692871, "learning_rate": 7.769784172661872e-06, "loss": 0.9033, "step": 47 }, { "epoch": 1.5483870967741935, "grad_norm": 7.0601067543029785, "learning_rate": 7.697841726618706e-06, "loss": 0.7922, "step": 48 }, { "epoch": 1.5806451612903225, "grad_norm": 5.145197868347168, "learning_rate": 7.62589928057554e-06, "loss": 0.7994, "step": 49 }, { "epoch": 1.6129032258064515, "grad_norm": 7.315432071685791, "learning_rate": 7.5539568345323745e-06, "loss": 0.9052, "step": 50 }, { "epoch": 1.6451612903225805, "grad_norm": 5.903252601623535, "learning_rate": 7.48201438848921e-06, "loss": 0.7659, "step": 51 }, { "epoch": 1.6774193548387095, "grad_norm": 6.903425216674805, "learning_rate": 7.410071942446043e-06, "loss": 0.6835, "step": 52 }, { "epoch": 1.7096774193548387, "grad_norm": 5.649851322174072, "learning_rate": 7.338129496402878e-06, "loss": 0.7886, "step": 53 }, { "epoch": 1.7419354838709677, "grad_norm": 7.038477420806885, "learning_rate": 7.266187050359713e-06, "loss": 0.8199, "step": 54 }, { "epoch": 1.7741935483870968, "grad_norm": 5.405828475952148, "learning_rate": 7.194244604316547e-06, "loss": 0.7793, "step": 55 }, { "epoch": 1.8064516129032258, "grad_norm": 4.914242744445801, "learning_rate": 7.122302158273382e-06, "loss": 0.6027, "step": 56 }, { "epoch": 1.838709677419355, "grad_norm": 3.8369204998016357, "learning_rate": 7.050359712230216e-06, "loss": 0.6166, "step": 57 }, { "epoch": 1.870967741935484, "grad_norm": 5.300614356994629, "learning_rate": 6.978417266187051e-06, "loss": 0.5933, "step": 58 }, { "epoch": 1.903225806451613, "grad_norm": 6.020833969116211, "learning_rate": 6.906474820143886e-06, "loss": 0.6644, "step": 59 }, { "epoch": 1.935483870967742, "grad_norm": 7.810927867889404, "learning_rate": 6.834532374100719e-06, "loss": 0.6088, "step": 60 }, { "epoch": 1.967741935483871, "grad_norm": 4.888729572296143, "learning_rate": 6.762589928057554e-06, "loss": 0.601, "step": 61 }, { "epoch": 2.0, "grad_norm": 9.534357070922852, "learning_rate": 6.6906474820143886e-06, "loss": 0.5048, "step": 62 }, { "epoch": 2.0, "eval_accuracy": 0.9032258064516129, "eval_f1_macro": 0.9009661835748792, "eval_f1_micro": 0.9032258064516129, "eval_f1_weighted": 0.90205703599813, "eval_loss": 0.5168083310127258, "eval_precision_macro": 0.9125, "eval_precision_micro": 0.9032258064516129, "eval_precision_weighted": 0.9120967741935483, "eval_recall_macro": 0.9007936507936508, "eval_recall_micro": 0.9032258064516129, "eval_recall_weighted": 0.9032258064516129, "eval_runtime": 0.362, "eval_samples_per_second": 171.293, "eval_steps_per_second": 11.051, "step": 62 }, { "epoch": 2.032258064516129, "grad_norm": 5.264218807220459, "learning_rate": 6.618705035971224e-06, "loss": 0.5298, "step": 63 }, { "epoch": 2.064516129032258, "grad_norm": 4.860868453979492, "learning_rate": 6.546762589928059e-06, "loss": 0.4793, "step": 64 }, { "epoch": 2.096774193548387, "grad_norm": 4.750568389892578, "learning_rate": 6.474820143884892e-06, "loss": 0.6173, "step": 65 }, { "epoch": 2.129032258064516, "grad_norm": 6.425321102142334, "learning_rate": 6.402877697841727e-06, "loss": 0.6019, "step": 66 }, { "epoch": 2.161290322580645, "grad_norm": 9.132750511169434, "learning_rate": 6.330935251798561e-06, "loss": 0.5513, "step": 67 }, { "epoch": 2.193548387096774, "grad_norm": 3.2714285850524902, "learning_rate": 6.2589928057553964e-06, "loss": 0.5036, "step": 68 }, { "epoch": 2.225806451612903, "grad_norm": 6.814634323120117, "learning_rate": 6.1870503597122315e-06, "loss": 0.537, "step": 69 }, { "epoch": 2.258064516129032, "grad_norm": 3.728926420211792, "learning_rate": 6.115107913669065e-06, "loss": 0.4946, "step": 70 }, { "epoch": 2.2903225806451615, "grad_norm": 6.5363545417785645, "learning_rate": 6.0431654676259e-06, "loss": 0.4285, "step": 71 }, { "epoch": 2.3225806451612905, "grad_norm": 3.544241428375244, "learning_rate": 5.971223021582734e-06, "loss": 0.5499, "step": 72 }, { "epoch": 2.3548387096774195, "grad_norm": 4.439380168914795, "learning_rate": 5.899280575539568e-06, "loss": 0.4158, "step": 73 }, { "epoch": 2.3870967741935485, "grad_norm": 3.382199287414551, "learning_rate": 5.8273381294964035e-06, "loss": 0.5042, "step": 74 }, { "epoch": 2.4193548387096775, "grad_norm": 3.844313144683838, "learning_rate": 5.755395683453238e-06, "loss": 0.4456, "step": 75 }, { "epoch": 2.4516129032258065, "grad_norm": 6.994678497314453, "learning_rate": 5.683453237410073e-06, "loss": 0.3322, "step": 76 }, { "epoch": 2.4838709677419355, "grad_norm": 3.745391845703125, "learning_rate": 5.611510791366906e-06, "loss": 0.4974, "step": 77 }, { "epoch": 2.5161290322580645, "grad_norm": 3.9794974327087402, "learning_rate": 5.539568345323741e-06, "loss": 0.6066, "step": 78 }, { "epoch": 2.5483870967741935, "grad_norm": 4.3371901512146, "learning_rate": 5.467625899280576e-06, "loss": 0.3384, "step": 79 }, { "epoch": 2.5806451612903225, "grad_norm": 3.778343915939331, "learning_rate": 5.3956834532374105e-06, "loss": 0.3747, "step": 80 }, { "epoch": 2.6129032258064515, "grad_norm": 5.110599994659424, "learning_rate": 5.3237410071942456e-06, "loss": 0.3761, "step": 81 }, { "epoch": 2.6451612903225805, "grad_norm": 4.0779128074646, "learning_rate": 5.251798561151079e-06, "loss": 0.3718, "step": 82 }, { "epoch": 2.6774193548387095, "grad_norm": 3.576948881149292, "learning_rate": 5.179856115107914e-06, "loss": 0.3232, "step": 83 }, { "epoch": 2.709677419354839, "grad_norm": 4.824243068695068, "learning_rate": 5.107913669064749e-06, "loss": 0.3021, "step": 84 }, { "epoch": 2.741935483870968, "grad_norm": 7.285948276519775, "learning_rate": 5.035971223021583e-06, "loss": 0.3227, "step": 85 }, { "epoch": 2.774193548387097, "grad_norm": 4.5859246253967285, "learning_rate": 4.9640287769784175e-06, "loss": 0.2029, "step": 86 }, { "epoch": 2.806451612903226, "grad_norm": 3.3746910095214844, "learning_rate": 4.892086330935253e-06, "loss": 0.3087, "step": 87 }, { "epoch": 2.838709677419355, "grad_norm": 5.0139875411987305, "learning_rate": 4.820143884892087e-06, "loss": 0.3919, "step": 88 }, { "epoch": 2.870967741935484, "grad_norm": 3.168724775314331, "learning_rate": 4.748201438848921e-06, "loss": 0.2469, "step": 89 }, { "epoch": 2.903225806451613, "grad_norm": 6.2994914054870605, "learning_rate": 4.676258992805755e-06, "loss": 0.2812, "step": 90 }, { "epoch": 2.935483870967742, "grad_norm": 1.8329119682312012, "learning_rate": 4.60431654676259e-06, "loss": 0.148, "step": 91 }, { "epoch": 2.967741935483871, "grad_norm": 3.138279914855957, "learning_rate": 4.5323741007194245e-06, "loss": 0.1359, "step": 92 }, { "epoch": 3.0, "grad_norm": 4.396747589111328, "learning_rate": 4.46043165467626e-06, "loss": 0.1879, "step": 93 }, { "epoch": 3.0, "eval_accuracy": 0.967741935483871, "eval_f1_macro": 0.9674603174603175, "eval_f1_micro": 0.967741935483871, "eval_f1_weighted": 0.967741935483871, "eval_loss": 0.22156503796577454, "eval_precision_macro": 0.9674603174603175, "eval_precision_micro": 0.967741935483871, "eval_precision_weighted": 0.967741935483871, "eval_recall_macro": 0.9674603174603175, "eval_recall_micro": 0.967741935483871, "eval_recall_weighted": 0.967741935483871, "eval_runtime": 0.3612, "eval_samples_per_second": 171.656, "eval_steps_per_second": 11.075, "step": 93 }, { "epoch": 3.032258064516129, "grad_norm": 2.482665538787842, "learning_rate": 4.388489208633094e-06, "loss": 0.203, "step": 94 }, { "epoch": 3.064516129032258, "grad_norm": 3.7393405437469482, "learning_rate": 4.316546762589928e-06, "loss": 0.4162, "step": 95 }, { "epoch": 3.096774193548387, "grad_norm": 4.992405414581299, "learning_rate": 4.244604316546763e-06, "loss": 0.272, "step": 96 }, { "epoch": 3.129032258064516, "grad_norm": 5.053823471069336, "learning_rate": 4.172661870503597e-06, "loss": 0.3839, "step": 97 }, { "epoch": 3.161290322580645, "grad_norm": 9.411662101745605, "learning_rate": 4.100719424460432e-06, "loss": 0.2929, "step": 98 }, { "epoch": 3.193548387096774, "grad_norm": 3.2983734607696533, "learning_rate": 4.028776978417267e-06, "loss": 0.1844, "step": 99 }, { "epoch": 3.225806451612903, "grad_norm": 9.005085945129395, "learning_rate": 3.956834532374101e-06, "loss": 0.2335, "step": 100 }, { "epoch": 3.258064516129032, "grad_norm": 2.7752511501312256, "learning_rate": 3.884892086330936e-06, "loss": 0.1636, "step": 101 }, { "epoch": 3.2903225806451615, "grad_norm": 2.129927635192871, "learning_rate": 3.81294964028777e-06, "loss": 0.1674, "step": 102 }, { "epoch": 3.3225806451612905, "grad_norm": 3.201385736465454, "learning_rate": 3.741007194244605e-06, "loss": 0.2321, "step": 103 }, { "epoch": 3.3548387096774195, "grad_norm": 6.4212775230407715, "learning_rate": 3.669064748201439e-06, "loss": 0.3104, "step": 104 }, { "epoch": 3.3870967741935485, "grad_norm": 9.587445259094238, "learning_rate": 3.5971223021582737e-06, "loss": 0.3252, "step": 105 }, { "epoch": 3.4193548387096775, "grad_norm": 4.780144214630127, "learning_rate": 3.525179856115108e-06, "loss": 0.2342, "step": 106 }, { "epoch": 3.4516129032258065, "grad_norm": 6.432891845703125, "learning_rate": 3.453237410071943e-06, "loss": 0.3754, "step": 107 }, { "epoch": 3.4838709677419355, "grad_norm": 1.6028318405151367, "learning_rate": 3.381294964028777e-06, "loss": 0.0951, "step": 108 }, { "epoch": 3.5161290322580645, "grad_norm": 3.9140098094940186, "learning_rate": 3.309352517985612e-06, "loss": 0.1739, "step": 109 }, { "epoch": 3.5483870967741935, "grad_norm": 5.9437785148620605, "learning_rate": 3.237410071942446e-06, "loss": 0.3609, "step": 110 }, { "epoch": 3.5806451612903225, "grad_norm": 2.725947380065918, "learning_rate": 3.1654676258992807e-06, "loss": 0.2015, "step": 111 }, { "epoch": 3.6129032258064515, "grad_norm": 5.769122123718262, "learning_rate": 3.0935251798561158e-06, "loss": 0.2124, "step": 112 }, { "epoch": 3.6451612903225805, "grad_norm": 3.645589590072632, "learning_rate": 3.02158273381295e-06, "loss": 0.157, "step": 113 }, { "epoch": 3.6774193548387095, "grad_norm": 2.2191619873046875, "learning_rate": 2.949640287769784e-06, "loss": 0.1301, "step": 114 }, { "epoch": 3.709677419354839, "grad_norm": 2.9101061820983887, "learning_rate": 2.877697841726619e-06, "loss": 0.1281, "step": 115 }, { "epoch": 3.741935483870968, "grad_norm": 8.459104537963867, "learning_rate": 2.805755395683453e-06, "loss": 0.2189, "step": 116 }, { "epoch": 3.774193548387097, "grad_norm": 4.622738361358643, "learning_rate": 2.733812949640288e-06, "loss": 0.1275, "step": 117 }, { "epoch": 3.806451612903226, "grad_norm": 5.437811374664307, "learning_rate": 2.6618705035971228e-06, "loss": 0.2634, "step": 118 }, { "epoch": 3.838709677419355, "grad_norm": 1.6734240055084229, "learning_rate": 2.589928057553957e-06, "loss": 0.1186, "step": 119 }, { "epoch": 3.870967741935484, "grad_norm": 1.7176499366760254, "learning_rate": 2.5179856115107916e-06, "loss": 0.1065, "step": 120 }, { "epoch": 3.903225806451613, "grad_norm": Infinity, "learning_rate": 2.5179856115107916e-06, "loss": 0.2718, "step": 121 }, { "epoch": 3.935483870967742, "grad_norm": 5.398061752319336, "learning_rate": 2.4460431654676263e-06, "loss": 0.1756, "step": 122 }, { "epoch": 3.967741935483871, "grad_norm": 5.624890327453613, "learning_rate": 2.3741007194244605e-06, "loss": 0.177, "step": 123 }, { "epoch": 4.0, "grad_norm": 2.5228195190429688, "learning_rate": 2.302158273381295e-06, "loss": 0.1125, "step": 124 }, { "epoch": 4.0, "eval_accuracy": 0.9838709677419355, "eval_f1_macro": 0.983739837398374, "eval_f1_micro": 0.9838709677419355, "eval_f1_weighted": 0.9838709677419355, "eval_loss": 0.12518411874771118, "eval_precision_macro": 0.9841269841269842, "eval_precision_micro": 0.9838709677419355, "eval_precision_weighted": 0.9846390168970814, "eval_recall_macro": 0.9841269841269842, "eval_recall_micro": 0.9838709677419355, "eval_recall_weighted": 0.9838709677419355, "eval_runtime": 0.361, "eval_samples_per_second": 171.764, "eval_steps_per_second": 11.082, "step": 124 }, { "epoch": 4.032258064516129, "grad_norm": 7.587864398956299, "learning_rate": 2.23021582733813e-06, "loss": 0.3132, "step": 125 }, { "epoch": 4.064516129032258, "grad_norm": 9.148059844970703, "learning_rate": 2.158273381294964e-06, "loss": 0.3178, "step": 126 }, { "epoch": 4.096774193548387, "grad_norm": 1.7947614192962646, "learning_rate": 2.0863309352517987e-06, "loss": 0.112, "step": 127 }, { "epoch": 4.129032258064516, "grad_norm": 1.4539235830307007, "learning_rate": 2.0143884892086333e-06, "loss": 0.0894, "step": 128 }, { "epoch": 4.161290322580645, "grad_norm": 3.562119245529175, "learning_rate": 1.942446043165468e-06, "loss": 0.1412, "step": 129 }, { "epoch": 4.193548387096774, "grad_norm": 3.7732717990875244, "learning_rate": 1.8705035971223024e-06, "loss": 0.1285, "step": 130 }, { "epoch": 4.225806451612903, "grad_norm": 8.253504753112793, "learning_rate": 1.7985611510791368e-06, "loss": 0.1619, "step": 131 }, { "epoch": 4.258064516129032, "grad_norm": 1.0998966693878174, "learning_rate": 1.7266187050359715e-06, "loss": 0.0796, "step": 132 }, { "epoch": 4.290322580645161, "grad_norm": 1.808113694190979, "learning_rate": 1.654676258992806e-06, "loss": 0.0901, "step": 133 }, { "epoch": 4.32258064516129, "grad_norm": 1.456236720085144, "learning_rate": 1.5827338129496403e-06, "loss": 0.0853, "step": 134 }, { "epoch": 4.354838709677419, "grad_norm": 8.323379516601562, "learning_rate": 1.510791366906475e-06, "loss": 0.2426, "step": 135 }, { "epoch": 4.387096774193548, "grad_norm": 6.862564563751221, "learning_rate": 1.4388489208633094e-06, "loss": 0.179, "step": 136 }, { "epoch": 4.419354838709677, "grad_norm": 8.144110679626465, "learning_rate": 1.366906474820144e-06, "loss": 0.2786, "step": 137 }, { "epoch": 4.451612903225806, "grad_norm": 1.4601564407348633, "learning_rate": 1.2949640287769785e-06, "loss": 0.0765, "step": 138 }, { "epoch": 4.483870967741936, "grad_norm": 8.350536346435547, "learning_rate": 1.2230215827338131e-06, "loss": 0.1893, "step": 139 }, { "epoch": 4.516129032258064, "grad_norm": 6.004292011260986, "learning_rate": 1.1510791366906476e-06, "loss": 0.1283, "step": 140 }, { "epoch": 4.548387096774194, "grad_norm": 4.625925540924072, "learning_rate": 1.079136690647482e-06, "loss": 0.1, "step": 141 }, { "epoch": 4.580645161290323, "grad_norm": 4.438725471496582, "learning_rate": 1.0071942446043167e-06, "loss": 0.1802, "step": 142 }, { "epoch": 4.612903225806452, "grad_norm": 15.338831901550293, "learning_rate": 9.352517985611512e-07, "loss": 0.2354, "step": 143 }, { "epoch": 4.645161290322581, "grad_norm": 4.097437381744385, "learning_rate": 8.633093525179857e-07, "loss": 0.1042, "step": 144 }, { "epoch": 4.67741935483871, "grad_norm": 4.476722240447998, "learning_rate": 7.913669064748202e-07, "loss": 0.1137, "step": 145 }, { "epoch": 4.709677419354839, "grad_norm": 8.289087295532227, "learning_rate": 7.194244604316547e-07, "loss": 0.273, "step": 146 }, { "epoch": 4.741935483870968, "grad_norm": 1.9783574342727661, "learning_rate": 6.474820143884893e-07, "loss": 0.1018, "step": 147 }, { "epoch": 4.774193548387097, "grad_norm": 4.377150535583496, "learning_rate": 5.755395683453238e-07, "loss": 0.1195, "step": 148 }, { "epoch": 4.806451612903226, "grad_norm": 3.9018876552581787, "learning_rate": 5.035971223021583e-07, "loss": 0.1093, "step": 149 }, { "epoch": 4.838709677419355, "grad_norm": 13.00845718383789, "learning_rate": 4.3165467625899287e-07, "loss": 0.3638, "step": 150 }, { "epoch": 4.870967741935484, "grad_norm": 4.976847171783447, "learning_rate": 3.5971223021582736e-07, "loss": 0.0991, "step": 151 }, { "epoch": 4.903225806451613, "grad_norm": 10.16610336303711, "learning_rate": 2.877697841726619e-07, "loss": 0.1559, "step": 152 }, { "epoch": 4.935483870967742, "grad_norm": 1.9939817190170288, "learning_rate": 2.1582733812949643e-07, "loss": 0.1162, "step": 153 }, { "epoch": 4.967741935483871, "grad_norm": 5.235056400299072, "learning_rate": 1.4388489208633095e-07, "loss": 0.133, "step": 154 }, { "epoch": 5.0, "grad_norm": 2.054269790649414, "learning_rate": 7.194244604316547e-08, "loss": 0.0845, "step": 155 }, { "epoch": 5.0, "eval_accuracy": 0.9838709677419355, "eval_f1_macro": 0.983739837398374, "eval_f1_micro": 0.9838709677419355, "eval_f1_weighted": 0.9838709677419355, "eval_loss": 0.10599859803915024, "eval_precision_macro": 0.9841269841269842, "eval_precision_micro": 0.9838709677419355, "eval_precision_weighted": 0.9846390168970814, "eval_recall_macro": 0.9841269841269842, "eval_recall_micro": 0.9838709677419355, "eval_recall_weighted": 0.9838709677419355, "eval_runtime": 0.3592, "eval_samples_per_second": 172.607, "eval_steps_per_second": 11.136, "step": 155 } ], "logging_steps": 1, "max_steps": 155, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "total_flos": 161613108449280.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }