{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.01, "eval_steps": 500, "global_step": 1000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 1e-05, "grad_norm": 1.6317715205474108, "learning_rate": 3e-06, "loss": 10.867, "step": 1 }, { "epoch": 2e-05, "grad_norm": 1.6173147870740345, "learning_rate": 6e-06, "loss": 10.8657, "step": 2 }, { "epoch": 3e-05, "grad_norm": 1.6387509359885835, "learning_rate": 9e-06, "loss": 10.8658, "step": 3 }, { "epoch": 4e-05, "grad_norm": 1.597552357815991, "learning_rate": 1.2e-05, "loss": 10.865, "step": 4 }, { "epoch": 5e-05, "grad_norm": 1.6454471252189307, "learning_rate": 1.5e-05, "loss": 10.8617, "step": 5 }, { "epoch": 6e-05, "grad_norm": 1.6407925722175996, "learning_rate": 1.8e-05, "loss": 10.8593, "step": 6 }, { "epoch": 7e-05, "grad_norm": 1.6096088910322361, "learning_rate": 2.1000000000000002e-05, "loss": 10.8456, "step": 7 }, { "epoch": 8e-05, "grad_norm": 1.4682698515009915, "learning_rate": 2.4e-05, "loss": 10.8184, "step": 8 }, { "epoch": 9e-05, "grad_norm": 1.3934246427009196, "learning_rate": 2.7e-05, "loss": 10.8113, "step": 9 }, { "epoch": 0.0001, "grad_norm": 1.3326087040550991, "learning_rate": 3e-05, "loss": 10.7969, "step": 10 }, { "epoch": 0.00011, "grad_norm": 1.2173090118888668, "learning_rate": 3.2999999999999996e-05, "loss": 10.7801, "step": 11 }, { "epoch": 0.00012, "grad_norm": 1.176457724285593, "learning_rate": 3.6e-05, "loss": 10.7688, "step": 12 }, { "epoch": 0.00013, "grad_norm": 1.1304424318539597, "learning_rate": 3.9e-05, "loss": 10.7498, "step": 13 }, { "epoch": 0.00014, "grad_norm": 1.1158244568462428, "learning_rate": 4.2000000000000004e-05, "loss": 10.739, "step": 14 }, { "epoch": 0.00015, "grad_norm": 1.10654756415174, "learning_rate": 4.4999999999999996e-05, "loss": 10.7299, "step": 15 }, { "epoch": 0.00016, "grad_norm": 1.0751770608444569, "learning_rate": 4.8e-05, "loss": 10.7126, "step": 16 }, { "epoch": 0.00017, "grad_norm": 1.046143788290158, "learning_rate": 5.1000000000000006e-05, "loss": 10.6968, "step": 17 }, { "epoch": 0.00018, "grad_norm": 1.0230391412556632, "learning_rate": 5.4e-05, "loss": 10.6806, "step": 18 }, { "epoch": 0.00019, "grad_norm": 0.9869982733638543, "learning_rate": 5.7e-05, "loss": 10.6649, "step": 19 }, { "epoch": 0.0002, "grad_norm": 0.9728818553338922, "learning_rate": 6e-05, "loss": 10.6526, "step": 20 }, { "epoch": 0.00021, "grad_norm": 0.9439994749998407, "learning_rate": 6.3e-05, "loss": 10.639, "step": 21 }, { "epoch": 0.00022, "grad_norm": 0.9279528910342688, "learning_rate": 6.599999999999999e-05, "loss": 10.6244, "step": 22 }, { "epoch": 0.00023, "grad_norm": 0.9208663519602571, "learning_rate": 6.9e-05, "loss": 10.6103, "step": 23 }, { "epoch": 0.00024, "grad_norm": 0.9178551557561957, "learning_rate": 7.2e-05, "loss": 10.5993, "step": 24 }, { "epoch": 0.00025, "grad_norm": 0.9193923250060233, "learning_rate": 7.500000000000001e-05, "loss": 10.5847, "step": 25 }, { "epoch": 0.00026, "grad_norm": 0.9190901609677985, "learning_rate": 7.8e-05, "loss": 10.5717, "step": 26 }, { "epoch": 0.00027, "grad_norm": 0.913753327244254, "learning_rate": 8.1e-05, "loss": 10.5597, "step": 27 }, { "epoch": 0.00028, "grad_norm": 0.9119625217070774, "learning_rate": 8.400000000000001e-05, "loss": 10.5467, "step": 28 }, { "epoch": 0.00029, "grad_norm": 0.9131038863398008, "learning_rate": 8.7e-05, "loss": 10.5323, "step": 29 }, { "epoch": 0.0003, "grad_norm": 0.9186172829723749, "learning_rate": 8.999999999999999e-05, "loss": 10.517, "step": 30 }, { "epoch": 0.00031, "grad_norm": 0.9155453723962563, "learning_rate": 9.3e-05, "loss": 10.5024, "step": 31 }, { "epoch": 0.00032, "grad_norm": 0.909575589137279, "learning_rate": 9.6e-05, "loss": 10.4882, "step": 32 }, { "epoch": 0.00033, "grad_norm": 0.9102175928717151, "learning_rate": 9.900000000000001e-05, "loss": 10.4717, "step": 33 }, { "epoch": 0.00034, "grad_norm": 0.9103991171564593, "learning_rate": 0.00010200000000000001, "loss": 10.4554, "step": 34 }, { "epoch": 0.00035, "grad_norm": 0.9164468459870767, "learning_rate": 0.00010500000000000002, "loss": 10.4388, "step": 35 }, { "epoch": 0.00036, "grad_norm": 0.9129850089149896, "learning_rate": 0.000108, "loss": 10.4213, "step": 36 }, { "epoch": 0.00037, "grad_norm": 0.9029808327462479, "learning_rate": 0.000111, "loss": 10.4047, "step": 37 }, { "epoch": 0.00038, "grad_norm": 0.9074583652458479, "learning_rate": 0.000114, "loss": 10.3859, "step": 38 }, { "epoch": 0.00039, "grad_norm": 0.9101621604618185, "learning_rate": 0.000117, "loss": 10.3665, "step": 39 }, { "epoch": 0.0004, "grad_norm": 0.9144345472354501, "learning_rate": 0.00012, "loss": 10.3443, "step": 40 }, { "epoch": 0.00041, "grad_norm": 0.9045315909874942, "learning_rate": 0.000123, "loss": 10.3264, "step": 41 }, { "epoch": 0.00042, "grad_norm": 0.9101518170592343, "learning_rate": 0.000126, "loss": 10.3059, "step": 42 }, { "epoch": 0.00043, "grad_norm": 0.9147491310031046, "learning_rate": 0.000129, "loss": 10.2831, "step": 43 }, { "epoch": 0.00044, "grad_norm": 0.9138166723808987, "learning_rate": 0.00013199999999999998, "loss": 10.2617, "step": 44 }, { "epoch": 0.00045, "grad_norm": 0.9166634386783579, "learning_rate": 0.000135, "loss": 10.2388, "step": 45 }, { "epoch": 0.00046, "grad_norm": 0.9061319871747918, "learning_rate": 0.000138, "loss": 10.2171, "step": 46 }, { "epoch": 0.00047, "grad_norm": 0.9134144094551727, "learning_rate": 0.000141, "loss": 10.1906, "step": 47 }, { "epoch": 0.00048, "grad_norm": 0.9111015850262806, "learning_rate": 0.000144, "loss": 10.1669, "step": 48 }, { "epoch": 0.00049, "grad_norm": 0.9106086112424904, "learning_rate": 0.000147, "loss": 10.1439, "step": 49 }, { "epoch": 0.0005, "grad_norm": 0.9135108001899231, "learning_rate": 0.00015000000000000001, "loss": 10.1178, "step": 50 }, { "epoch": 0.00051, "grad_norm": 0.9152031108656089, "learning_rate": 0.000153, "loss": 10.0918, "step": 51 }, { "epoch": 0.00052, "grad_norm": 0.9133043896843657, "learning_rate": 0.000156, "loss": 10.0658, "step": 52 }, { "epoch": 0.00053, "grad_norm": 0.9039676544194273, "learning_rate": 0.000159, "loss": 10.0419, "step": 53 }, { "epoch": 0.00054, "grad_norm": 0.9245050218484777, "learning_rate": 0.000162, "loss": 10.0099, "step": 54 }, { "epoch": 0.00055, "grad_norm": 0.915653013423474, "learning_rate": 0.000165, "loss": 9.9858, "step": 55 }, { "epoch": 0.00056, "grad_norm": 0.90743999026624, "learning_rate": 0.00016800000000000002, "loss": 9.9567, "step": 56 }, { "epoch": 0.00057, "grad_norm": 0.9125740935300273, "learning_rate": 0.000171, "loss": 9.9292, "step": 57 }, { "epoch": 0.00058, "grad_norm": 0.9103134473221595, "learning_rate": 0.000174, "loss": 9.9046, "step": 58 }, { "epoch": 0.00059, "grad_norm": 0.91502039796166, "learning_rate": 0.000177, "loss": 9.8727, "step": 59 }, { "epoch": 0.0006, "grad_norm": 0.902549468432534, "learning_rate": 0.00017999999999999998, "loss": 9.8467, "step": 60 }, { "epoch": 0.00061, "grad_norm": 0.9011753674575653, "learning_rate": 0.000183, "loss": 9.8184, "step": 61 }, { "epoch": 0.00062, "grad_norm": 0.9073829944096542, "learning_rate": 0.000186, "loss": 9.7865, "step": 62 }, { "epoch": 0.00063, "grad_norm": 0.900277782228101, "learning_rate": 0.000189, "loss": 9.7594, "step": 63 }, { "epoch": 0.00064, "grad_norm": 0.8964836174343672, "learning_rate": 0.000192, "loss": 9.7292, "step": 64 }, { "epoch": 0.00065, "grad_norm": 0.9040253232587138, "learning_rate": 0.00019500000000000002, "loss": 9.6969, "step": 65 }, { "epoch": 0.00066, "grad_norm": 0.8969700410935529, "learning_rate": 0.00019800000000000002, "loss": 9.6728, "step": 66 }, { "epoch": 0.00067, "grad_norm": 0.8981910064021, "learning_rate": 0.000201, "loss": 9.643, "step": 67 }, { "epoch": 0.00068, "grad_norm": 0.9049113812894196, "learning_rate": 0.00020400000000000003, "loss": 9.61, "step": 68 }, { "epoch": 0.00069, "grad_norm": 0.8903730719674341, "learning_rate": 0.00020700000000000002, "loss": 9.5824, "step": 69 }, { "epoch": 0.0007, "grad_norm": 0.9034192157313848, "learning_rate": 0.00021000000000000004, "loss": 9.5456, "step": 70 }, { "epoch": 0.00071, "grad_norm": 0.8933343163190056, "learning_rate": 0.00021299999999999997, "loss": 9.5189, "step": 71 }, { "epoch": 0.00072, "grad_norm": 0.8968593008835964, "learning_rate": 0.000216, "loss": 9.4914, "step": 72 }, { "epoch": 0.00073, "grad_norm": 0.8960167287531013, "learning_rate": 0.00021899999999999998, "loss": 9.4574, "step": 73 }, { "epoch": 0.00074, "grad_norm": 0.8965354004907367, "learning_rate": 0.000222, "loss": 9.4304, "step": 74 }, { "epoch": 0.00075, "grad_norm": 0.8923958382533757, "learning_rate": 0.000225, "loss": 9.3946, "step": 75 }, { "epoch": 0.00076, "grad_norm": 0.9011054851249423, "learning_rate": 0.000228, "loss": 9.3667, "step": 76 }, { "epoch": 0.00077, "grad_norm": 0.8959973023769764, "learning_rate": 0.000231, "loss": 9.3374, "step": 77 }, { "epoch": 0.00078, "grad_norm": 0.8901362352309407, "learning_rate": 0.000234, "loss": 9.3055, "step": 78 }, { "epoch": 0.00079, "grad_norm": 0.8931237127830396, "learning_rate": 0.00023700000000000001, "loss": 9.2694, "step": 79 }, { "epoch": 0.0008, "grad_norm": 0.8937972757596185, "learning_rate": 0.00024, "loss": 9.2355, "step": 80 }, { "epoch": 0.00081, "grad_norm": 0.8984186266430717, "learning_rate": 0.00024300000000000002, "loss": 9.2033, "step": 81 }, { "epoch": 0.00082, "grad_norm": 0.9060711515558022, "learning_rate": 0.000246, "loss": 9.1789, "step": 82 }, { "epoch": 0.00083, "grad_norm": 0.8964035250431631, "learning_rate": 0.00024900000000000004, "loss": 9.1434, "step": 83 }, { "epoch": 0.00084, "grad_norm": 0.8986377804082708, "learning_rate": 0.000252, "loss": 9.1119, "step": 84 }, { "epoch": 0.00085, "grad_norm": 0.8933390971667627, "learning_rate": 0.000255, "loss": 9.0885, "step": 85 }, { "epoch": 0.00086, "grad_norm": 0.8953786211031305, "learning_rate": 0.000258, "loss": 9.0521, "step": 86 }, { "epoch": 0.00087, "grad_norm": 0.8876773666743288, "learning_rate": 0.000261, "loss": 9.0302, "step": 87 }, { "epoch": 0.00088, "grad_norm": 0.8929152197347487, "learning_rate": 0.00026399999999999997, "loss": 8.9927, "step": 88 }, { "epoch": 0.00089, "grad_norm": 0.8872457913370299, "learning_rate": 0.000267, "loss": 8.9669, "step": 89 }, { "epoch": 0.0009, "grad_norm": 0.8862509419810688, "learning_rate": 0.00027, "loss": 8.9393, "step": 90 }, { "epoch": 0.00091, "grad_norm": 0.8807880266520192, "learning_rate": 0.000273, "loss": 8.9089, "step": 91 }, { "epoch": 0.00092, "grad_norm": 0.8831207589668301, "learning_rate": 0.000276, "loss": 8.876, "step": 92 }, { "epoch": 0.00093, "grad_norm": 0.883866988348631, "learning_rate": 0.000279, "loss": 8.8462, "step": 93 }, { "epoch": 0.00094, "grad_norm": 0.8824853161719922, "learning_rate": 0.000282, "loss": 8.8199, "step": 94 }, { "epoch": 0.00095, "grad_norm": 0.8794572899807177, "learning_rate": 0.000285, "loss": 8.789, "step": 95 }, { "epoch": 0.00096, "grad_norm": 0.8879520784944948, "learning_rate": 0.000288, "loss": 8.7571, "step": 96 }, { "epoch": 0.00097, "grad_norm": 0.8798042444972031, "learning_rate": 0.000291, "loss": 8.7321, "step": 97 }, { "epoch": 0.00098, "grad_norm": 0.8802597894834375, "learning_rate": 0.000294, "loss": 8.6954, "step": 98 }, { "epoch": 0.00099, "grad_norm": 0.8806466323910314, "learning_rate": 0.000297, "loss": 8.6749, "step": 99 }, { "epoch": 0.001, "grad_norm": 0.8777097574069823, "learning_rate": 0.00030000000000000003, "loss": 8.6485, "step": 100 }, { "epoch": 0.00101, "grad_norm": 0.8786755147609817, "learning_rate": 0.00030300000000000005, "loss": 8.614, "step": 101 }, { "epoch": 0.00102, "grad_norm": 0.8680143868447665, "learning_rate": 0.000306, "loss": 8.5949, "step": 102 }, { "epoch": 0.00103, "grad_norm": 0.8748449452730288, "learning_rate": 0.000309, "loss": 8.5706, "step": 103 }, { "epoch": 0.00104, "grad_norm": 0.8670215859388973, "learning_rate": 0.000312, "loss": 8.5498, "step": 104 }, { "epoch": 0.00105, "grad_norm": 0.8687292008731472, "learning_rate": 0.000315, "loss": 8.5231, "step": 105 }, { "epoch": 0.00106, "grad_norm": 0.8611486197845404, "learning_rate": 0.000318, "loss": 8.4945, "step": 106 }, { "epoch": 0.00107, "grad_norm": 0.8521924348129856, "learning_rate": 0.000321, "loss": 8.4693, "step": 107 }, { "epoch": 0.00108, "grad_norm": 0.8581933475380797, "learning_rate": 0.000324, "loss": 8.4407, "step": 108 }, { "epoch": 0.00109, "grad_norm": 0.8524146875206363, "learning_rate": 0.000327, "loss": 8.421, "step": 109 }, { "epoch": 0.0011, "grad_norm": 0.8682563584613229, "learning_rate": 0.00033, "loss": 8.3983, "step": 110 }, { "epoch": 0.00111, "grad_norm": 0.8871469803064874, "learning_rate": 0.000333, "loss": 8.3685, "step": 111 }, { "epoch": 0.00112, "grad_norm": 0.9236879668329372, "learning_rate": 0.00033600000000000004, "loss": 8.3463, "step": 112 }, { "epoch": 0.00113, "grad_norm": 0.9129864456504505, "learning_rate": 0.000339, "loss": 8.3063, "step": 113 }, { "epoch": 0.00114, "grad_norm": 0.8352263501003522, "learning_rate": 0.000342, "loss": 8.2966, "step": 114 }, { "epoch": 0.00115, "grad_norm": 0.8592375580090957, "learning_rate": 0.00034500000000000004, "loss": 8.2718, "step": 115 }, { "epoch": 0.00116, "grad_norm": 0.8674294753896091, "learning_rate": 0.000348, "loss": 8.2506, "step": 116 }, { "epoch": 0.00117, "grad_norm": 0.8276917493567776, "learning_rate": 0.000351, "loss": 8.2188, "step": 117 }, { "epoch": 0.00118, "grad_norm": 0.8476744963131545, "learning_rate": 0.000354, "loss": 8.2045, "step": 118 }, { "epoch": 0.00119, "grad_norm": 0.844358071644388, "learning_rate": 0.000357, "loss": 8.1926, "step": 119 }, { "epoch": 0.0012, "grad_norm": 0.8166594269287538, "learning_rate": 0.00035999999999999997, "loss": 8.1658, "step": 120 }, { "epoch": 0.00121, "grad_norm": 0.8239930081024902, "learning_rate": 0.000363, "loss": 8.1389, "step": 121 }, { "epoch": 0.00122, "grad_norm": 0.8099951959348987, "learning_rate": 0.000366, "loss": 8.1225, "step": 122 }, { "epoch": 0.00123, "grad_norm": 0.830800320388625, "learning_rate": 0.000369, "loss": 8.1005, "step": 123 }, { "epoch": 0.00124, "grad_norm": 0.8139169053139192, "learning_rate": 0.000372, "loss": 8.0791, "step": 124 }, { "epoch": 0.00125, "grad_norm": 0.8112246790149765, "learning_rate": 0.000375, "loss": 8.0547, "step": 125 }, { "epoch": 0.00126, "grad_norm": 0.7922278873371895, "learning_rate": 0.000378, "loss": 8.0424, "step": 126 }, { "epoch": 0.00127, "grad_norm": 0.7755075943975184, "learning_rate": 0.000381, "loss": 8.0182, "step": 127 }, { "epoch": 0.00128, "grad_norm": 0.8028212720713388, "learning_rate": 0.000384, "loss": 8.0, "step": 128 }, { "epoch": 0.00129, "grad_norm": 0.8765129391436198, "learning_rate": 0.00038700000000000003, "loss": 7.9904, "step": 129 }, { "epoch": 0.0013, "grad_norm": 1.1486399983200042, "learning_rate": 0.00039000000000000005, "loss": 7.9724, "step": 130 }, { "epoch": 0.00131, "grad_norm": 1.0507777578095836, "learning_rate": 0.000393, "loss": 7.9382, "step": 131 }, { "epoch": 0.00132, "grad_norm": 0.7511305165281239, "learning_rate": 0.00039600000000000003, "loss": 7.9248, "step": 132 }, { "epoch": 0.00133, "grad_norm": 0.8108077692812641, "learning_rate": 0.00039900000000000005, "loss": 7.906, "step": 133 }, { "epoch": 0.00134, "grad_norm": 0.8301313874662418, "learning_rate": 0.000402, "loss": 7.8952, "step": 134 }, { "epoch": 0.00135, "grad_norm": 0.7123568807247732, "learning_rate": 0.00040500000000000003, "loss": 7.8651, "step": 135 }, { "epoch": 0.00136, "grad_norm": 0.7697556904537746, "learning_rate": 0.00040800000000000005, "loss": 7.8515, "step": 136 }, { "epoch": 0.00137, "grad_norm": 0.7190977621725152, "learning_rate": 0.000411, "loss": 7.8299, "step": 137 }, { "epoch": 0.00138, "grad_norm": 0.7147305335216294, "learning_rate": 0.00041400000000000003, "loss": 7.8066, "step": 138 }, { "epoch": 0.00139, "grad_norm": 0.747529428120578, "learning_rate": 0.00041700000000000005, "loss": 7.7954, "step": 139 }, { "epoch": 0.0014, "grad_norm": 0.6748198408281931, "learning_rate": 0.00042000000000000007, "loss": 7.7774, "step": 140 }, { "epoch": 0.00141, "grad_norm": 0.662142968172009, "learning_rate": 0.000423, "loss": 7.7644, "step": 141 }, { "epoch": 0.00142, "grad_norm": 0.6859796391897652, "learning_rate": 0.00042599999999999995, "loss": 7.7534, "step": 142 }, { "epoch": 0.00143, "grad_norm": 0.6828514310354903, "learning_rate": 0.00042899999999999997, "loss": 7.7255, "step": 143 }, { "epoch": 0.00144, "grad_norm": 0.6490687938821236, "learning_rate": 0.000432, "loss": 7.7078, "step": 144 }, { "epoch": 0.00145, "grad_norm": 0.6453156151137228, "learning_rate": 0.000435, "loss": 7.7035, "step": 145 }, { "epoch": 0.00146, "grad_norm": 0.7166050341593803, "learning_rate": 0.00043799999999999997, "loss": 7.6672, "step": 146 }, { "epoch": 0.00147, "grad_norm": 0.7443563124698165, "learning_rate": 0.000441, "loss": 7.6627, "step": 147 }, { "epoch": 0.00148, "grad_norm": 0.7124453009383569, "learning_rate": 0.000444, "loss": 7.648, "step": 148 }, { "epoch": 0.00149, "grad_norm": 0.6775306354557482, "learning_rate": 0.00044699999999999997, "loss": 7.6419, "step": 149 }, { "epoch": 0.0015, "grad_norm": 0.8906380813028638, "learning_rate": 0.00045, "loss": 7.6135, "step": 150 }, { "epoch": 0.00151, "grad_norm": 1.2892607147030477, "learning_rate": 0.000453, "loss": 7.605, "step": 151 }, { "epoch": 0.00152, "grad_norm": 0.6492280537852009, "learning_rate": 0.000456, "loss": 7.5875, "step": 152 }, { "epoch": 0.00153, "grad_norm": 0.8226748559002907, "learning_rate": 0.000459, "loss": 7.5783, "step": 153 }, { "epoch": 0.00154, "grad_norm": 1.073498340899344, "learning_rate": 0.000462, "loss": 7.5662, "step": 154 }, { "epoch": 0.00155, "grad_norm": 0.772416311968319, "learning_rate": 0.000465, "loss": 7.5481, "step": 155 }, { "epoch": 0.00156, "grad_norm": 0.7999364848474875, "learning_rate": 0.000468, "loss": 7.5209, "step": 156 }, { "epoch": 0.00157, "grad_norm": 0.9435570004081977, "learning_rate": 0.000471, "loss": 7.5196, "step": 157 }, { "epoch": 0.00158, "grad_norm": 0.8364296006508578, "learning_rate": 0.00047400000000000003, "loss": 7.4997, "step": 158 }, { "epoch": 0.00159, "grad_norm": 0.5596750978008624, "learning_rate": 0.000477, "loss": 7.4825, "step": 159 }, { "epoch": 0.0016, "grad_norm": 0.5748856794890025, "learning_rate": 0.00048, "loss": 7.4616, "step": 160 }, { "epoch": 0.00161, "grad_norm": 0.5986411025981033, "learning_rate": 0.00048300000000000003, "loss": 7.4419, "step": 161 }, { "epoch": 0.00162, "grad_norm": 0.5117747222458712, "learning_rate": 0.00048600000000000005, "loss": 7.4429, "step": 162 }, { "epoch": 0.00163, "grad_norm": 0.5509815158764758, "learning_rate": 0.0004890000000000001, "loss": 7.4259, "step": 163 }, { "epoch": 0.00164, "grad_norm": 0.46393808675226217, "learning_rate": 0.000492, "loss": 7.4236, "step": 164 }, { "epoch": 0.00165, "grad_norm": 0.5390679177469344, "learning_rate": 0.000495, "loss": 7.4006, "step": 165 }, { "epoch": 0.00166, "grad_norm": 0.645354949444588, "learning_rate": 0.0004980000000000001, "loss": 7.3773, "step": 166 }, { "epoch": 0.00167, "grad_norm": 1.0877276724000633, "learning_rate": 0.000501, "loss": 7.3741, "step": 167 }, { "epoch": 0.00168, "grad_norm": 1.0909247587015876, "learning_rate": 0.000504, "loss": 7.3697, "step": 168 }, { "epoch": 0.00169, "grad_norm": 0.5084368948335112, "learning_rate": 0.0005070000000000001, "loss": 7.3418, "step": 169 }, { "epoch": 0.0017, "grad_norm": 1.5053615947271437, "learning_rate": 0.00051, "loss": 7.3492, "step": 170 }, { "epoch": 0.00171, "grad_norm": 0.6120595065498764, "learning_rate": 0.000513, "loss": 7.3094, "step": 171 }, { "epoch": 0.00172, "grad_norm": 0.9401489103136018, "learning_rate": 0.000516, "loss": 7.3189, "step": 172 }, { "epoch": 0.00173, "grad_norm": 0.5558816443971099, "learning_rate": 0.0005189999999999999, "loss": 7.2906, "step": 173 }, { "epoch": 0.00174, "grad_norm": 0.7810658223347703, "learning_rate": 0.000522, "loss": 7.314, "step": 174 }, { "epoch": 0.00175, "grad_norm": 0.5409187817835382, "learning_rate": 0.000525, "loss": 7.2727, "step": 175 }, { "epoch": 0.00176, "grad_norm": 0.630636606756854, "learning_rate": 0.0005279999999999999, "loss": 7.2484, "step": 176 }, { "epoch": 0.00177, "grad_norm": 0.5206138592215499, "learning_rate": 0.000531, "loss": 7.2507, "step": 177 }, { "epoch": 0.00178, "grad_norm": 0.5577166707239251, "learning_rate": 0.000534, "loss": 7.2428, "step": 178 }, { "epoch": 0.00179, "grad_norm": 0.4980094016420717, "learning_rate": 0.000537, "loss": 7.2333, "step": 179 }, { "epoch": 0.0018, "grad_norm": 0.5436132024149352, "learning_rate": 0.00054, "loss": 7.2053, "step": 180 }, { "epoch": 0.00181, "grad_norm": 0.4850436660416719, "learning_rate": 0.000543, "loss": 7.1915, "step": 181 }, { "epoch": 0.00182, "grad_norm": 0.686701132410735, "learning_rate": 0.000546, "loss": 7.1909, "step": 182 }, { "epoch": 0.00183, "grad_norm": 0.68400997444651, "learning_rate": 0.000549, "loss": 7.1802, "step": 183 }, { "epoch": 0.00184, "grad_norm": 0.8396027257351396, "learning_rate": 0.000552, "loss": 7.171, "step": 184 }, { "epoch": 0.00185, "grad_norm": 0.7665365037410753, "learning_rate": 0.000555, "loss": 7.1476, "step": 185 }, { "epoch": 0.00186, "grad_norm": 0.5359175667027454, "learning_rate": 0.000558, "loss": 7.1506, "step": 186 }, { "epoch": 0.00187, "grad_norm": 0.5513800735908173, "learning_rate": 0.000561, "loss": 7.1261, "step": 187 }, { "epoch": 0.00188, "grad_norm": 0.5352093004255375, "learning_rate": 0.000564, "loss": 7.1044, "step": 188 }, { "epoch": 0.00189, "grad_norm": 0.5938457818726526, "learning_rate": 0.000567, "loss": 7.1004, "step": 189 }, { "epoch": 0.0019, "grad_norm": 0.5273842405001533, "learning_rate": 0.00057, "loss": 7.0834, "step": 190 }, { "epoch": 0.00191, "grad_norm": 0.47487115515279366, "learning_rate": 0.000573, "loss": 7.0721, "step": 191 }, { "epoch": 0.00192, "grad_norm": 0.6036837698051599, "learning_rate": 0.000576, "loss": 7.0655, "step": 192 }, { "epoch": 0.00193, "grad_norm": 0.3951184212196986, "learning_rate": 0.000579, "loss": 7.061, "step": 193 }, { "epoch": 0.00194, "grad_norm": 0.442083567688087, "learning_rate": 0.000582, "loss": 7.0548, "step": 194 }, { "epoch": 0.00195, "grad_norm": 0.439546882682468, "learning_rate": 0.000585, "loss": 7.0348, "step": 195 }, { "epoch": 0.00196, "grad_norm": 0.46247531692771043, "learning_rate": 0.000588, "loss": 7.0228, "step": 196 }, { "epoch": 0.00197, "grad_norm": 0.4140335072217301, "learning_rate": 0.000591, "loss": 7.0171, "step": 197 }, { "epoch": 0.00198, "grad_norm": 0.3685986320410548, "learning_rate": 0.000594, "loss": 7.0081, "step": 198 }, { "epoch": 0.00199, "grad_norm": 0.4020373564129086, "learning_rate": 0.0005970000000000001, "loss": 6.9898, "step": 199 }, { "epoch": 0.002, "grad_norm": 0.37126410475941546, "learning_rate": 0.0006000000000000001, "loss": 6.9867, "step": 200 }, { "epoch": 0.00201, "grad_norm": 0.3773154493828028, "learning_rate": 0.000603, "loss": 6.9617, "step": 201 }, { "epoch": 0.00202, "grad_norm": 0.3540017416986532, "learning_rate": 0.0006060000000000001, "loss": 6.9491, "step": 202 }, { "epoch": 0.00203, "grad_norm": 0.403279648640721, "learning_rate": 0.0006090000000000001, "loss": 6.9534, "step": 203 }, { "epoch": 0.00204, "grad_norm": 0.5112949618253247, "learning_rate": 0.000612, "loss": 6.9385, "step": 204 }, { "epoch": 0.00205, "grad_norm": 0.7200998739972175, "learning_rate": 0.000615, "loss": 6.931, "step": 205 }, { "epoch": 0.00206, "grad_norm": 1.209379808685074, "learning_rate": 0.000618, "loss": 6.9351, "step": 206 }, { "epoch": 0.00207, "grad_norm": 1.0040442357645134, "learning_rate": 0.000621, "loss": 6.9279, "step": 207 }, { "epoch": 0.00208, "grad_norm": 0.7878464521205251, "learning_rate": 0.000624, "loss": 6.9129, "step": 208 }, { "epoch": 0.00209, "grad_norm": 1.3096135062236434, "learning_rate": 0.000627, "loss": 6.9067, "step": 209 }, { "epoch": 0.0021, "grad_norm": 0.910504669978176, "learning_rate": 0.00063, "loss": 6.8878, "step": 210 }, { "epoch": 0.00211, "grad_norm": 1.3015817966038044, "learning_rate": 0.000633, "loss": 6.8987, "step": 211 }, { "epoch": 0.00212, "grad_norm": 0.5587575104994011, "learning_rate": 0.000636, "loss": 6.8696, "step": 212 }, { "epoch": 0.00213, "grad_norm": 0.8081412049208773, "learning_rate": 0.000639, "loss": 6.8608, "step": 213 }, { "epoch": 0.00214, "grad_norm": 0.6397075273457759, "learning_rate": 0.000642, "loss": 6.8511, "step": 214 }, { "epoch": 0.00215, "grad_norm": 0.5875043250740225, "learning_rate": 0.000645, "loss": 6.8482, "step": 215 }, { "epoch": 0.00216, "grad_norm": 0.6060774535893669, "learning_rate": 0.000648, "loss": 6.8413, "step": 216 }, { "epoch": 0.00217, "grad_norm": 0.5183751970166313, "learning_rate": 0.000651, "loss": 6.8119, "step": 217 }, { "epoch": 0.00218, "grad_norm": 0.5539011900924167, "learning_rate": 0.000654, "loss": 6.8171, "step": 218 }, { "epoch": 0.00219, "grad_norm": 0.5004312163301685, "learning_rate": 0.000657, "loss": 6.7882, "step": 219 }, { "epoch": 0.0022, "grad_norm": 0.4762494747244133, "learning_rate": 0.00066, "loss": 6.8062, "step": 220 }, { "epoch": 0.00221, "grad_norm": 0.353452768224107, "learning_rate": 0.0006630000000000001, "loss": 6.7814, "step": 221 }, { "epoch": 0.00222, "grad_norm": 0.3879096289808107, "learning_rate": 0.000666, "loss": 6.7696, "step": 222 }, { "epoch": 0.00223, "grad_norm": 0.4253518811476648, "learning_rate": 0.000669, "loss": 6.7658, "step": 223 }, { "epoch": 0.00224, "grad_norm": 0.4076338977034843, "learning_rate": 0.0006720000000000001, "loss": 6.7609, "step": 224 }, { "epoch": 0.00225, "grad_norm": 0.4491881195721512, "learning_rate": 0.000675, "loss": 6.7489, "step": 225 }, { "epoch": 0.00226, "grad_norm": 0.4195112098951784, "learning_rate": 0.000678, "loss": 6.7444, "step": 226 }, { "epoch": 0.00227, "grad_norm": 0.371663908330708, "learning_rate": 0.0006810000000000001, "loss": 6.7174, "step": 227 }, { "epoch": 0.00228, "grad_norm": 0.3462952066328263, "learning_rate": 0.000684, "loss": 6.7197, "step": 228 }, { "epoch": 0.00229, "grad_norm": 0.5288966810878937, "learning_rate": 0.000687, "loss": 6.7178, "step": 229 }, { "epoch": 0.0023, "grad_norm": 0.7623592789505088, "learning_rate": 0.0006900000000000001, "loss": 6.6993, "step": 230 }, { "epoch": 0.00231, "grad_norm": 0.9880780315432149, "learning_rate": 0.000693, "loss": 6.6923, "step": 231 }, { "epoch": 0.00232, "grad_norm": 1.028691756937799, "learning_rate": 0.000696, "loss": 6.7142, "step": 232 }, { "epoch": 0.00233, "grad_norm": 0.9501761132787978, "learning_rate": 0.0006990000000000001, "loss": 6.6946, "step": 233 }, { "epoch": 0.00234, "grad_norm": 0.9999298841530961, "learning_rate": 0.000702, "loss": 6.6948, "step": 234 }, { "epoch": 0.00235, "grad_norm": 0.6945930996245869, "learning_rate": 0.000705, "loss": 6.6675, "step": 235 }, { "epoch": 0.00236, "grad_norm": 0.6422608041910052, "learning_rate": 0.000708, "loss": 6.6513, "step": 236 }, { "epoch": 0.00237, "grad_norm": 0.47985797682797676, "learning_rate": 0.0007109999999999999, "loss": 6.657, "step": 237 }, { "epoch": 0.00238, "grad_norm": 0.6191094472441181, "learning_rate": 0.000714, "loss": 6.6482, "step": 238 }, { "epoch": 0.00239, "grad_norm": 0.5181523072026278, "learning_rate": 0.000717, "loss": 6.628, "step": 239 }, { "epoch": 0.0024, "grad_norm": 0.5894783578801835, "learning_rate": 0.0007199999999999999, "loss": 6.645, "step": 240 }, { "epoch": 0.00241, "grad_norm": 0.5639698176641863, "learning_rate": 0.000723, "loss": 6.6279, "step": 241 }, { "epoch": 0.00242, "grad_norm": 0.9095170652542525, "learning_rate": 0.000726, "loss": 6.6087, "step": 242 }, { "epoch": 0.00243, "grad_norm": 1.3373514416355459, "learning_rate": 0.000729, "loss": 6.6191, "step": 243 }, { "epoch": 0.00244, "grad_norm": 0.5837452319331187, "learning_rate": 0.000732, "loss": 6.5991, "step": 244 }, { "epoch": 0.00245, "grad_norm": 1.0261084178764917, "learning_rate": 0.000735, "loss": 6.6035, "step": 245 }, { "epoch": 0.00246, "grad_norm": 0.5663979442820745, "learning_rate": 0.000738, "loss": 6.574, "step": 246 }, { "epoch": 0.00247, "grad_norm": 0.5896636858778472, "learning_rate": 0.000741, "loss": 6.5719, "step": 247 }, { "epoch": 0.00248, "grad_norm": 0.47326080430149503, "learning_rate": 0.000744, "loss": 6.5757, "step": 248 }, { "epoch": 0.00249, "grad_norm": 0.6248164732961499, "learning_rate": 0.000747, "loss": 6.5582, "step": 249 }, { "epoch": 0.0025, "grad_norm": 0.48362025373458484, "learning_rate": 0.00075, "loss": 6.5567, "step": 250 }, { "epoch": 0.00251, "grad_norm": 0.4858269118610639, "learning_rate": 0.000753, "loss": 6.5227, "step": 251 }, { "epoch": 0.00252, "grad_norm": 0.41862369731289734, "learning_rate": 0.000756, "loss": 6.5302, "step": 252 }, { "epoch": 0.00253, "grad_norm": 0.47200180330590646, "learning_rate": 0.000759, "loss": 6.522, "step": 253 }, { "epoch": 0.00254, "grad_norm": 0.3651546358356223, "learning_rate": 0.000762, "loss": 6.5136, "step": 254 }, { "epoch": 0.00255, "grad_norm": 0.4285192093499382, "learning_rate": 0.0007650000000000001, "loss": 6.5149, "step": 255 }, { "epoch": 0.00256, "grad_norm": 0.42767490819993825, "learning_rate": 0.000768, "loss": 6.4917, "step": 256 }, { "epoch": 0.00257, "grad_norm": 0.48143237273687123, "learning_rate": 0.000771, "loss": 6.4974, "step": 257 }, { "epoch": 0.00258, "grad_norm": 0.5467090762573645, "learning_rate": 0.0007740000000000001, "loss": 6.4744, "step": 258 }, { "epoch": 0.00259, "grad_norm": 0.6750367098607939, "learning_rate": 0.000777, "loss": 6.4781, "step": 259 }, { "epoch": 0.0026, "grad_norm": 0.8320211245129605, "learning_rate": 0.0007800000000000001, "loss": 6.4681, "step": 260 }, { "epoch": 0.00261, "grad_norm": 0.905841412497731, "learning_rate": 0.0007830000000000001, "loss": 6.4809, "step": 261 }, { "epoch": 0.00262, "grad_norm": 1.0687360869781928, "learning_rate": 0.000786, "loss": 6.4647, "step": 262 }, { "epoch": 0.00263, "grad_norm": 1.2051408681263374, "learning_rate": 0.0007890000000000001, "loss": 6.4874, "step": 263 }, { "epoch": 0.00264, "grad_norm": 0.8690142485653533, "learning_rate": 0.0007920000000000001, "loss": 6.455, "step": 264 }, { "epoch": 0.00265, "grad_norm": 1.2774066489819682, "learning_rate": 0.000795, "loss": 6.4641, "step": 265 }, { "epoch": 0.00266, "grad_norm": 0.7288978979341997, "learning_rate": 0.0007980000000000001, "loss": 6.4454, "step": 266 }, { "epoch": 0.00267, "grad_norm": 0.6506287971604123, "learning_rate": 0.0008010000000000001, "loss": 6.4371, "step": 267 }, { "epoch": 0.00268, "grad_norm": 0.6866545943797145, "learning_rate": 0.000804, "loss": 6.4338, "step": 268 }, { "epoch": 0.00269, "grad_norm": 1.0440114151727509, "learning_rate": 0.0008070000000000001, "loss": 6.4227, "step": 269 }, { "epoch": 0.0027, "grad_norm": 1.170351969791303, "learning_rate": 0.0008100000000000001, "loss": 6.4362, "step": 270 }, { "epoch": 0.00271, "grad_norm": 0.6832222674646221, "learning_rate": 0.000813, "loss": 6.4253, "step": 271 }, { "epoch": 0.00272, "grad_norm": 0.617936452008115, "learning_rate": 0.0008160000000000001, "loss": 6.4063, "step": 272 }, { "epoch": 0.00273, "grad_norm": 0.6750066600068259, "learning_rate": 0.0008190000000000001, "loss": 6.4008, "step": 273 }, { "epoch": 0.00274, "grad_norm": 0.6315560378177079, "learning_rate": 0.000822, "loss": 6.379, "step": 274 }, { "epoch": 0.00275, "grad_norm": 0.6774266517760174, "learning_rate": 0.0008250000000000001, "loss": 6.3852, "step": 275 }, { "epoch": 0.00276, "grad_norm": 0.7272895822836024, "learning_rate": 0.0008280000000000001, "loss": 6.3794, "step": 276 }, { "epoch": 0.00277, "grad_norm": 0.6752471037637485, "learning_rate": 0.0008310000000000001, "loss": 6.3735, "step": 277 }, { "epoch": 0.00278, "grad_norm": 0.5678457826039285, "learning_rate": 0.0008340000000000001, "loss": 6.354, "step": 278 }, { "epoch": 0.00279, "grad_norm": 0.4611700189072147, "learning_rate": 0.0008370000000000001, "loss": 6.3529, "step": 279 }, { "epoch": 0.0028, "grad_norm": 0.343285643042232, "learning_rate": 0.0008400000000000001, "loss": 6.3329, "step": 280 }, { "epoch": 0.00281, "grad_norm": 0.4519631747446028, "learning_rate": 0.0008430000000000001, "loss": 6.3253, "step": 281 }, { "epoch": 0.00282, "grad_norm": 0.3255189118052276, "learning_rate": 0.000846, "loss": 6.3232, "step": 282 }, { "epoch": 0.00283, "grad_norm": 0.4297016476682907, "learning_rate": 0.0008489999999999999, "loss": 6.3149, "step": 283 }, { "epoch": 0.00284, "grad_norm": 0.37515936714697207, "learning_rate": 0.0008519999999999999, "loss": 6.3058, "step": 284 }, { "epoch": 0.00285, "grad_norm": 0.3458870104505622, "learning_rate": 0.000855, "loss": 6.3075, "step": 285 }, { "epoch": 0.00286, "grad_norm": 0.39943359732325345, "learning_rate": 0.0008579999999999999, "loss": 6.2857, "step": 286 }, { "epoch": 0.00287, "grad_norm": 0.3631751365570726, "learning_rate": 0.000861, "loss": 6.2875, "step": 287 }, { "epoch": 0.00288, "grad_norm": 0.552280032213235, "learning_rate": 0.000864, "loss": 6.2707, "step": 288 }, { "epoch": 0.00289, "grad_norm": 0.9396362484724781, "learning_rate": 0.0008669999999999999, "loss": 6.2923, "step": 289 }, { "epoch": 0.0029, "grad_norm": 1.3959756273937387, "learning_rate": 0.00087, "loss": 6.2945, "step": 290 }, { "epoch": 0.00291, "grad_norm": 0.6791923155853262, "learning_rate": 0.000873, "loss": 6.2663, "step": 291 }, { "epoch": 0.00292, "grad_norm": 0.9879732835250642, "learning_rate": 0.0008759999999999999, "loss": 6.2794, "step": 292 }, { "epoch": 0.00293, "grad_norm": 1.0140384817226566, "learning_rate": 0.000879, "loss": 6.2652, "step": 293 }, { "epoch": 0.00294, "grad_norm": 0.9889225551341856, "learning_rate": 0.000882, "loss": 6.2822, "step": 294 }, { "epoch": 0.00295, "grad_norm": 1.0429477871581094, "learning_rate": 0.0008849999999999999, "loss": 6.2468, "step": 295 }, { "epoch": 0.00296, "grad_norm": 1.1774929945794055, "learning_rate": 0.000888, "loss": 6.2705, "step": 296 }, { "epoch": 0.00297, "grad_norm": 0.7364156162079134, "learning_rate": 0.000891, "loss": 6.2278, "step": 297 }, { "epoch": 0.00298, "grad_norm": 0.8424638476384282, "learning_rate": 0.0008939999999999999, "loss": 6.2455, "step": 298 }, { "epoch": 0.00299, "grad_norm": 0.8668489286879963, "learning_rate": 0.000897, "loss": 6.2225, "step": 299 }, { "epoch": 0.003, "grad_norm": 0.9039057952602142, "learning_rate": 0.0009, "loss": 6.2236, "step": 300 }, { "epoch": 0.00301, "grad_norm": 1.0344858724084711, "learning_rate": 0.0009029999999999999, "loss": 6.2222, "step": 301 }, { "epoch": 0.00302, "grad_norm": 0.801399966246171, "learning_rate": 0.000906, "loss": 6.2007, "step": 302 }, { "epoch": 0.00303, "grad_norm": 0.7276255151675343, "learning_rate": 0.000909, "loss": 6.2106, "step": 303 }, { "epoch": 0.00304, "grad_norm": 0.8306162070729353, "learning_rate": 0.000912, "loss": 6.205, "step": 304 }, { "epoch": 0.00305, "grad_norm": 0.7650178489304597, "learning_rate": 0.000915, "loss": 6.2045, "step": 305 }, { "epoch": 0.00306, "grad_norm": 0.7024963687074245, "learning_rate": 0.000918, "loss": 6.1878, "step": 306 }, { "epoch": 0.00307, "grad_norm": 0.5687961336654864, "learning_rate": 0.000921, "loss": 6.1778, "step": 307 }, { "epoch": 0.00308, "grad_norm": 0.4515866134049927, "learning_rate": 0.000924, "loss": 6.1586, "step": 308 }, { "epoch": 0.00309, "grad_norm": 0.5454081565882548, "learning_rate": 0.000927, "loss": 6.1763, "step": 309 }, { "epoch": 0.0031, "grad_norm": 0.4033959865123679, "learning_rate": 0.00093, "loss": 6.1549, "step": 310 }, { "epoch": 0.00311, "grad_norm": 0.5004494164305024, "learning_rate": 0.000933, "loss": 6.1478, "step": 311 }, { "epoch": 0.00312, "grad_norm": 0.470361361901935, "learning_rate": 0.000936, "loss": 6.1275, "step": 312 }, { "epoch": 0.00313, "grad_norm": 0.4973667290148138, "learning_rate": 0.0009390000000000001, "loss": 6.1399, "step": 313 }, { "epoch": 0.00314, "grad_norm": 0.47398584855820086, "learning_rate": 0.000942, "loss": 6.1453, "step": 314 }, { "epoch": 0.00315, "grad_norm": 0.39081338977861474, "learning_rate": 0.000945, "loss": 6.1206, "step": 315 }, { "epoch": 0.00316, "grad_norm": 0.308452368547838, "learning_rate": 0.0009480000000000001, "loss": 6.0925, "step": 316 }, { "epoch": 0.00317, "grad_norm": 0.33735261718955184, "learning_rate": 0.000951, "loss": 6.1112, "step": 317 }, { "epoch": 0.00318, "grad_norm": 0.3843792856632324, "learning_rate": 0.000954, "loss": 6.1055, "step": 318 }, { "epoch": 0.00319, "grad_norm": 0.45015697169720664, "learning_rate": 0.0009570000000000001, "loss": 6.0951, "step": 319 }, { "epoch": 0.0032, "grad_norm": 0.6347836654615971, "learning_rate": 0.00096, "loss": 6.097, "step": 320 }, { "epoch": 0.00321, "grad_norm": 0.9264517947757075, "learning_rate": 0.000963, "loss": 6.0955, "step": 321 }, { "epoch": 0.00322, "grad_norm": 1.3311662164937155, "learning_rate": 0.0009660000000000001, "loss": 6.0984, "step": 322 }, { "epoch": 0.00323, "grad_norm": 0.7192768486088142, "learning_rate": 0.000969, "loss": 6.081, "step": 323 }, { "epoch": 0.00324, "grad_norm": 0.8815686637890671, "learning_rate": 0.0009720000000000001, "loss": 6.0921, "step": 324 }, { "epoch": 0.00325, "grad_norm": 0.9399377971403509, "learning_rate": 0.0009750000000000001, "loss": 6.0559, "step": 325 }, { "epoch": 0.00326, "grad_norm": 1.1161084705724094, "learning_rate": 0.0009780000000000001, "loss": 6.0866, "step": 326 }, { "epoch": 0.00327, "grad_norm": 0.9798577225908143, "learning_rate": 0.000981, "loss": 6.08, "step": 327 }, { "epoch": 0.00328, "grad_norm": 1.4594200076446016, "learning_rate": 0.000984, "loss": 6.0949, "step": 328 }, { "epoch": 0.00329, "grad_norm": 0.7530414187502621, "learning_rate": 0.000987, "loss": 6.0592, "step": 329 }, { "epoch": 0.0033, "grad_norm": 0.9983029089238832, "learning_rate": 0.00099, "loss": 6.0707, "step": 330 }, { "epoch": 0.00331, "grad_norm": 0.835334238631822, "learning_rate": 0.0009930000000000002, "loss": 6.0609, "step": 331 }, { "epoch": 0.00332, "grad_norm": 1.2633352311024129, "learning_rate": 0.0009960000000000001, "loss": 6.0632, "step": 332 }, { "epoch": 0.00333, "grad_norm": 0.9415989578137678, "learning_rate": 0.000999, "loss": 6.0388, "step": 333 }, { "epoch": 0.00334, "grad_norm": 0.752057040577036, "learning_rate": 0.001002, "loss": 6.0434, "step": 334 }, { "epoch": 0.00335, "grad_norm": 0.6368712711427076, "learning_rate": 0.001005, "loss": 6.039, "step": 335 }, { "epoch": 0.00336, "grad_norm": 0.5277321252477392, "learning_rate": 0.001008, "loss": 6.0255, "step": 336 }, { "epoch": 0.00337, "grad_norm": 0.545640031716998, "learning_rate": 0.0010110000000000002, "loss": 6.0051, "step": 337 }, { "epoch": 0.00338, "grad_norm": 0.5605817392374447, "learning_rate": 0.0010140000000000001, "loss": 6.0183, "step": 338 }, { "epoch": 0.00339, "grad_norm": 0.4963166785854256, "learning_rate": 0.0010170000000000001, "loss": 6.0214, "step": 339 }, { "epoch": 0.0034, "grad_norm": 0.496225629593103, "learning_rate": 0.00102, "loss": 5.9915, "step": 340 }, { "epoch": 0.00341, "grad_norm": 0.45013704339915594, "learning_rate": 0.001023, "loss": 5.9719, "step": 341 }, { "epoch": 0.00342, "grad_norm": 0.37899658405778774, "learning_rate": 0.001026, "loss": 5.9765, "step": 342 }, { "epoch": 0.00343, "grad_norm": 0.4216072327604619, "learning_rate": 0.0010290000000000002, "loss": 5.9773, "step": 343 }, { "epoch": 0.00344, "grad_norm": 0.4964437378733662, "learning_rate": 0.001032, "loss": 5.9678, "step": 344 }, { "epoch": 0.00345, "grad_norm": 0.6031142308234815, "learning_rate": 0.001035, "loss": 5.9605, "step": 345 }, { "epoch": 0.00346, "grad_norm": 0.6349778009023861, "learning_rate": 0.0010379999999999999, "loss": 5.9675, "step": 346 }, { "epoch": 0.00347, "grad_norm": 0.6213366174245898, "learning_rate": 0.001041, "loss": 5.9393, "step": 347 }, { "epoch": 0.00348, "grad_norm": 0.684857817555668, "learning_rate": 0.001044, "loss": 5.9499, "step": 348 }, { "epoch": 0.00349, "grad_norm": 0.8421661368559449, "learning_rate": 0.001047, "loss": 5.9398, "step": 349 }, { "epoch": 0.0035, "grad_norm": 0.8062802732253019, "learning_rate": 0.00105, "loss": 5.9568, "step": 350 }, { "epoch": 0.00351, "grad_norm": 0.7669273491138234, "learning_rate": 0.001053, "loss": 5.9549, "step": 351 }, { "epoch": 0.00352, "grad_norm": 1.0673384228730578, "learning_rate": 0.0010559999999999999, "loss": 5.9491, "step": 352 }, { "epoch": 0.00353, "grad_norm": 1.313057737158452, "learning_rate": 0.001059, "loss": 5.9596, "step": 353 }, { "epoch": 0.00354, "grad_norm": 0.8748293544965601, "learning_rate": 0.001062, "loss": 5.9228, "step": 354 }, { "epoch": 0.00355, "grad_norm": 0.7617258072087358, "learning_rate": 0.001065, "loss": 5.9348, "step": 355 }, { "epoch": 0.00356, "grad_norm": 0.9561923051718775, "learning_rate": 0.001068, "loss": 5.9306, "step": 356 }, { "epoch": 0.00357, "grad_norm": 0.8631216371911231, "learning_rate": 0.001071, "loss": 5.9199, "step": 357 }, { "epoch": 0.00358, "grad_norm": 0.9525953351992685, "learning_rate": 0.001074, "loss": 5.9184, "step": 358 }, { "epoch": 0.00359, "grad_norm": 0.6858307664974747, "learning_rate": 0.001077, "loss": 5.8973, "step": 359 }, { "epoch": 0.0036, "grad_norm": 0.7167037627907079, "learning_rate": 0.00108, "loss": 5.9065, "step": 360 }, { "epoch": 0.00361, "grad_norm": 0.7969047134484166, "learning_rate": 0.001083, "loss": 5.8986, "step": 361 }, { "epoch": 0.00362, "grad_norm": 0.8186204145394074, "learning_rate": 0.001086, "loss": 5.888, "step": 362 }, { "epoch": 0.00363, "grad_norm": 0.7436182013517663, "learning_rate": 0.001089, "loss": 5.8766, "step": 363 }, { "epoch": 0.00364, "grad_norm": 0.6658023261534547, "learning_rate": 0.001092, "loss": 5.8872, "step": 364 }, { "epoch": 0.00365, "grad_norm": 0.7206668214896482, "learning_rate": 0.001095, "loss": 5.8872, "step": 365 }, { "epoch": 0.00366, "grad_norm": 0.6345413922647961, "learning_rate": 0.001098, "loss": 5.8617, "step": 366 }, { "epoch": 0.00367, "grad_norm": 0.6094924231218852, "learning_rate": 0.001101, "loss": 5.8618, "step": 367 }, { "epoch": 0.00368, "grad_norm": 0.7732260654201254, "learning_rate": 0.001104, "loss": 5.8653, "step": 368 }, { "epoch": 0.00369, "grad_norm": 0.9451132874875877, "learning_rate": 0.001107, "loss": 5.8779, "step": 369 }, { "epoch": 0.0037, "grad_norm": 1.047387946938651, "learning_rate": 0.00111, "loss": 5.8603, "step": 370 }, { "epoch": 0.00371, "grad_norm": 0.8709497379832931, "learning_rate": 0.001113, "loss": 5.8262, "step": 371 }, { "epoch": 0.00372, "grad_norm": 0.6993621521276565, "learning_rate": 0.001116, "loss": 5.8434, "step": 372 }, { "epoch": 0.00373, "grad_norm": 0.6889019297226876, "learning_rate": 0.001119, "loss": 5.8344, "step": 373 }, { "epoch": 0.00374, "grad_norm": 0.8187808414499578, "learning_rate": 0.001122, "loss": 5.8424, "step": 374 }, { "epoch": 0.00375, "grad_norm": 0.7395556739972736, "learning_rate": 0.0011250000000000001, "loss": 5.8404, "step": 375 }, { "epoch": 0.00376, "grad_norm": 0.5027446756427529, "learning_rate": 0.001128, "loss": 5.8296, "step": 376 }, { "epoch": 0.00377, "grad_norm": 0.45297103786338255, "learning_rate": 0.001131, "loss": 5.8239, "step": 377 }, { "epoch": 0.00378, "grad_norm": 0.5150298222384522, "learning_rate": 0.001134, "loss": 5.82, "step": 378 }, { "epoch": 0.00379, "grad_norm": 0.4216428009711753, "learning_rate": 0.001137, "loss": 5.8036, "step": 379 }, { "epoch": 0.0038, "grad_norm": 0.43574801532624385, "learning_rate": 0.00114, "loss": 5.8234, "step": 380 }, { "epoch": 0.00381, "grad_norm": 0.4737821231317218, "learning_rate": 0.0011430000000000001, "loss": 5.7908, "step": 381 }, { "epoch": 0.00382, "grad_norm": 0.5236457098681065, "learning_rate": 0.001146, "loss": 5.7778, "step": 382 }, { "epoch": 0.00383, "grad_norm": 0.5006340354259897, "learning_rate": 0.001149, "loss": 5.787, "step": 383 }, { "epoch": 0.00384, "grad_norm": 0.5762330042414852, "learning_rate": 0.001152, "loss": 5.7915, "step": 384 }, { "epoch": 0.00385, "grad_norm": 0.6623935085819848, "learning_rate": 0.001155, "loss": 5.7808, "step": 385 }, { "epoch": 0.00386, "grad_norm": 0.6780931079980513, "learning_rate": 0.001158, "loss": 5.7864, "step": 386 }, { "epoch": 0.00387, "grad_norm": 0.6283557380392781, "learning_rate": 0.0011610000000000001, "loss": 5.76, "step": 387 }, { "epoch": 0.00388, "grad_norm": 0.8661947460887196, "learning_rate": 0.001164, "loss": 5.7642, "step": 388 }, { "epoch": 0.00389, "grad_norm": 1.1877567285678448, "learning_rate": 0.001167, "loss": 5.7898, "step": 389 }, { "epoch": 0.0039, "grad_norm": 0.5285139722189788, "learning_rate": 0.00117, "loss": 5.7491, "step": 390 }, { "epoch": 0.00391, "grad_norm": 0.6430118692881639, "learning_rate": 0.001173, "loss": 5.7622, "step": 391 }, { "epoch": 0.00392, "grad_norm": 0.6748100403552353, "learning_rate": 0.001176, "loss": 5.7713, "step": 392 }, { "epoch": 0.00393, "grad_norm": 0.7344199345621275, "learning_rate": 0.0011790000000000001, "loss": 5.7448, "step": 393 }, { "epoch": 0.00394, "grad_norm": 0.8611653440542995, "learning_rate": 0.001182, "loss": 5.7377, "step": 394 }, { "epoch": 0.00395, "grad_norm": 1.0200403565527223, "learning_rate": 0.001185, "loss": 5.7297, "step": 395 }, { "epoch": 0.00396, "grad_norm": 1.3219082304761296, "learning_rate": 0.001188, "loss": 5.7571, "step": 396 }, { "epoch": 0.00397, "grad_norm": 1.0648841185793536, "learning_rate": 0.001191, "loss": 5.7452, "step": 397 }, { "epoch": 0.00398, "grad_norm": 0.8818852137031271, "learning_rate": 0.0011940000000000002, "loss": 5.7476, "step": 398 }, { "epoch": 0.00399, "grad_norm": 0.7229082271352473, "learning_rate": 0.0011970000000000001, "loss": 5.7455, "step": 399 }, { "epoch": 0.004, "grad_norm": 0.7464785543636239, "learning_rate": 0.0012000000000000001, "loss": 5.7335, "step": 400 }, { "epoch": 0.00401, "grad_norm": 0.788008045277313, "learning_rate": 0.001203, "loss": 5.7147, "step": 401 }, { "epoch": 0.00402, "grad_norm": 0.8844811042429299, "learning_rate": 0.001206, "loss": 5.7255, "step": 402 }, { "epoch": 0.00403, "grad_norm": 0.8786909754931423, "learning_rate": 0.001209, "loss": 5.7201, "step": 403 }, { "epoch": 0.00404, "grad_norm": 0.9763167280044874, "learning_rate": 0.0012120000000000002, "loss": 5.7188, "step": 404 }, { "epoch": 0.00405, "grad_norm": 0.9708490732732808, "learning_rate": 0.0012150000000000002, "loss": 5.7216, "step": 405 }, { "epoch": 0.00406, "grad_norm": 0.7953769652219763, "learning_rate": 0.0012180000000000001, "loss": 5.6987, "step": 406 }, { "epoch": 0.00407, "grad_norm": 0.7231086032909518, "learning_rate": 0.0012209999999999999, "loss": 5.6945, "step": 407 }, { "epoch": 0.00408, "grad_norm": 0.6539281657127057, "learning_rate": 0.001224, "loss": 5.7108, "step": 408 }, { "epoch": 0.00409, "grad_norm": 0.48725126694443294, "learning_rate": 0.001227, "loss": 5.6827, "step": 409 }, { "epoch": 0.0041, "grad_norm": 0.5935989194477762, "learning_rate": 0.00123, "loss": 5.6771, "step": 410 }, { "epoch": 0.00411, "grad_norm": 0.6524672694061662, "learning_rate": 0.001233, "loss": 5.6945, "step": 411 }, { "epoch": 0.00412, "grad_norm": 0.5686252583506046, "learning_rate": 0.001236, "loss": 5.6796, "step": 412 }, { "epoch": 0.00413, "grad_norm": 0.5206794027062916, "learning_rate": 0.0012389999999999999, "loss": 5.676, "step": 413 }, { "epoch": 0.00414, "grad_norm": 0.5100681259570212, "learning_rate": 0.001242, "loss": 5.6561, "step": 414 }, { "epoch": 0.00415, "grad_norm": 0.5323494958966052, "learning_rate": 0.001245, "loss": 5.6572, "step": 415 }, { "epoch": 0.00416, "grad_norm": 0.593833682795931, "learning_rate": 0.001248, "loss": 5.6419, "step": 416 }, { "epoch": 0.00417, "grad_norm": 0.5881502982050868, "learning_rate": 0.001251, "loss": 5.6711, "step": 417 }, { "epoch": 0.00418, "grad_norm": 0.6571826832460801, "learning_rate": 0.001254, "loss": 5.6451, "step": 418 }, { "epoch": 0.00419, "grad_norm": 0.8242389183582979, "learning_rate": 0.0012569999999999999, "loss": 5.6393, "step": 419 }, { "epoch": 0.0042, "grad_norm": 0.7322278119135938, "learning_rate": 0.00126, "loss": 5.6531, "step": 420 }, { "epoch": 0.00421, "grad_norm": 0.6080486142741918, "learning_rate": 0.001263, "loss": 5.6415, "step": 421 }, { "epoch": 0.00422, "grad_norm": 0.6271805214461816, "learning_rate": 0.001266, "loss": 5.6327, "step": 422 }, { "epoch": 0.00423, "grad_norm": 0.5934268893913589, "learning_rate": 0.001269, "loss": 5.6286, "step": 423 }, { "epoch": 0.00424, "grad_norm": 0.6795296945522938, "learning_rate": 0.001272, "loss": 5.6299, "step": 424 }, { "epoch": 0.00425, "grad_norm": 0.7845276448161947, "learning_rate": 0.001275, "loss": 5.6344, "step": 425 }, { "epoch": 0.00426, "grad_norm": 0.9176562117983922, "learning_rate": 0.001278, "loss": 5.6402, "step": 426 }, { "epoch": 0.00427, "grad_norm": 1.0306205830362438, "learning_rate": 0.001281, "loss": 5.6293, "step": 427 }, { "epoch": 0.00428, "grad_norm": 1.0252381178312269, "learning_rate": 0.001284, "loss": 5.6086, "step": 428 }, { "epoch": 0.00429, "grad_norm": 1.3332162612637855, "learning_rate": 0.001287, "loss": 5.6393, "step": 429 }, { "epoch": 0.0043, "grad_norm": 0.8174291079939794, "learning_rate": 0.00129, "loss": 5.6202, "step": 430 }, { "epoch": 0.00431, "grad_norm": 0.821028463418781, "learning_rate": 0.001293, "loss": 5.6102, "step": 431 }, { "epoch": 0.00432, "grad_norm": 0.7475137161763143, "learning_rate": 0.001296, "loss": 5.6163, "step": 432 }, { "epoch": 0.00433, "grad_norm": 0.7571870052992741, "learning_rate": 0.001299, "loss": 5.604, "step": 433 }, { "epoch": 0.00434, "grad_norm": 1.0543560255015263, "learning_rate": 0.001302, "loss": 5.6184, "step": 434 }, { "epoch": 0.00435, "grad_norm": 0.8758758304266553, "learning_rate": 0.001305, "loss": 5.6066, "step": 435 }, { "epoch": 0.00436, "grad_norm": 0.9137062421440546, "learning_rate": 0.001308, "loss": 5.5859, "step": 436 }, { "epoch": 0.00437, "grad_norm": 1.301736875083812, "learning_rate": 0.001311, "loss": 5.6173, "step": 437 }, { "epoch": 0.00438, "grad_norm": 0.9356953917037294, "learning_rate": 0.001314, "loss": 5.5919, "step": 438 }, { "epoch": 0.00439, "grad_norm": 0.8522821992819578, "learning_rate": 0.001317, "loss": 5.6158, "step": 439 }, { "epoch": 0.0044, "grad_norm": 0.7182761753028103, "learning_rate": 0.00132, "loss": 5.5821, "step": 440 }, { "epoch": 0.00441, "grad_norm": 0.6293266892726601, "learning_rate": 0.001323, "loss": 5.577, "step": 441 }, { "epoch": 0.00442, "grad_norm": 0.8189921873128356, "learning_rate": 0.0013260000000000001, "loss": 5.5762, "step": 442 }, { "epoch": 0.00443, "grad_norm": 1.0212422821054057, "learning_rate": 0.001329, "loss": 5.5904, "step": 443 }, { "epoch": 0.00444, "grad_norm": 0.8951785018901781, "learning_rate": 0.001332, "loss": 5.5851, "step": 444 }, { "epoch": 0.00445, "grad_norm": 0.7868769140150608, "learning_rate": 0.001335, "loss": 5.5661, "step": 445 }, { "epoch": 0.00446, "grad_norm": 0.790102612629763, "learning_rate": 0.001338, "loss": 5.5718, "step": 446 }, { "epoch": 0.00447, "grad_norm": 0.8396194874372788, "learning_rate": 0.001341, "loss": 5.5716, "step": 447 }, { "epoch": 0.00448, "grad_norm": 0.9120841535821665, "learning_rate": 0.0013440000000000001, "loss": 5.5589, "step": 448 }, { "epoch": 0.00449, "grad_norm": 0.8573073152890212, "learning_rate": 0.001347, "loss": 5.5603, "step": 449 }, { "epoch": 0.0045, "grad_norm": 0.7420732830276576, "learning_rate": 0.00135, "loss": 5.5551, "step": 450 }, { "epoch": 0.00451, "grad_norm": 0.7054051762730813, "learning_rate": 0.001353, "loss": 5.5451, "step": 451 }, { "epoch": 0.00452, "grad_norm": 0.5383194985068459, "learning_rate": 0.001356, "loss": 5.5229, "step": 452 }, { "epoch": 0.00453, "grad_norm": 0.5845224529530345, "learning_rate": 0.001359, "loss": 5.5144, "step": 453 }, { "epoch": 0.00454, "grad_norm": 0.5197277965757966, "learning_rate": 0.0013620000000000001, "loss": 5.5269, "step": 454 }, { "epoch": 0.00455, "grad_norm": 0.44901215490386587, "learning_rate": 0.0013650000000000001, "loss": 5.5227, "step": 455 }, { "epoch": 0.00456, "grad_norm": 0.40823042735319937, "learning_rate": 0.001368, "loss": 5.5155, "step": 456 }, { "epoch": 0.00457, "grad_norm": 0.4077054091161063, "learning_rate": 0.001371, "loss": 5.5045, "step": 457 }, { "epoch": 0.00458, "grad_norm": 0.4051721280659754, "learning_rate": 0.001374, "loss": 5.5115, "step": 458 }, { "epoch": 0.00459, "grad_norm": 0.3819222938644956, "learning_rate": 0.0013770000000000002, "loss": 5.4986, "step": 459 }, { "epoch": 0.0046, "grad_norm": 0.4096791477469412, "learning_rate": 0.0013800000000000002, "loss": 5.4851, "step": 460 }, { "epoch": 0.00461, "grad_norm": 0.44167018915391465, "learning_rate": 0.0013830000000000001, "loss": 5.4805, "step": 461 }, { "epoch": 0.00462, "grad_norm": 0.49346826910496583, "learning_rate": 0.001386, "loss": 5.4928, "step": 462 }, { "epoch": 0.00463, "grad_norm": 0.463489633996114, "learning_rate": 0.001389, "loss": 5.4652, "step": 463 }, { "epoch": 0.00464, "grad_norm": 0.507205937400668, "learning_rate": 0.001392, "loss": 5.4859, "step": 464 }, { "epoch": 0.00465, "grad_norm": 0.5318639728571777, "learning_rate": 0.0013950000000000002, "loss": 5.4572, "step": 465 }, { "epoch": 0.00466, "grad_norm": 0.6489252257104292, "learning_rate": 0.0013980000000000002, "loss": 5.4792, "step": 466 }, { "epoch": 0.00467, "grad_norm": 0.7957789059873086, "learning_rate": 0.0014010000000000001, "loss": 5.4727, "step": 467 }, { "epoch": 0.00468, "grad_norm": 0.7621042481808248, "learning_rate": 0.001404, "loss": 5.462, "step": 468 }, { "epoch": 0.00469, "grad_norm": 0.573822393217689, "learning_rate": 0.001407, "loss": 5.457, "step": 469 }, { "epoch": 0.0047, "grad_norm": 0.7543746323219513, "learning_rate": 0.00141, "loss": 5.4709, "step": 470 }, { "epoch": 0.00471, "grad_norm": 0.8641651762100534, "learning_rate": 0.001413, "loss": 5.4567, "step": 471 }, { "epoch": 0.00472, "grad_norm": 0.9026674726231502, "learning_rate": 0.001416, "loss": 5.4699, "step": 472 }, { "epoch": 0.00473, "grad_norm": 0.8212885101559565, "learning_rate": 0.001419, "loss": 5.4778, "step": 473 }, { "epoch": 0.00474, "grad_norm": 0.8914030740906659, "learning_rate": 0.0014219999999999999, "loss": 5.461, "step": 474 }, { "epoch": 0.00475, "grad_norm": 1.0570094425693455, "learning_rate": 0.001425, "loss": 5.4652, "step": 475 }, { "epoch": 0.00476, "grad_norm": 0.9736444976311589, "learning_rate": 0.001428, "loss": 5.4875, "step": 476 }, { "epoch": 0.00477, "grad_norm": 1.1550380737092787, "learning_rate": 0.001431, "loss": 5.4568, "step": 477 }, { "epoch": 0.00478, "grad_norm": 0.848321570803796, "learning_rate": 0.001434, "loss": 5.4695, "step": 478 }, { "epoch": 0.00479, "grad_norm": 0.9517827225501269, "learning_rate": 0.001437, "loss": 5.4501, "step": 479 }, { "epoch": 0.0048, "grad_norm": 1.0883787540754652, "learning_rate": 0.0014399999999999999, "loss": 5.4562, "step": 480 }, { "epoch": 0.00481, "grad_norm": 0.9422991164230814, "learning_rate": 0.001443, "loss": 5.4516, "step": 481 }, { "epoch": 0.00482, "grad_norm": 0.9159499791385636, "learning_rate": 0.001446, "loss": 5.4273, "step": 482 }, { "epoch": 0.00483, "grad_norm": 0.9688645055474768, "learning_rate": 0.001449, "loss": 5.44, "step": 483 }, { "epoch": 0.00484, "grad_norm": 1.1114303023214132, "learning_rate": 0.001452, "loss": 5.4474, "step": 484 }, { "epoch": 0.00485, "grad_norm": 0.9051569573634253, "learning_rate": 0.001455, "loss": 5.4468, "step": 485 }, { "epoch": 0.00486, "grad_norm": 0.9247454458721566, "learning_rate": 0.001458, "loss": 5.4217, "step": 486 }, { "epoch": 0.00487, "grad_norm": 0.852928162562673, "learning_rate": 0.001461, "loss": 5.4339, "step": 487 }, { "epoch": 0.00488, "grad_norm": 0.8142513932978471, "learning_rate": 0.001464, "loss": 5.41, "step": 488 }, { "epoch": 0.00489, "grad_norm": 0.9329231917883894, "learning_rate": 0.001467, "loss": 5.403, "step": 489 }, { "epoch": 0.0049, "grad_norm": 0.9855383918059676, "learning_rate": 0.00147, "loss": 5.4217, "step": 490 }, { "epoch": 0.00491, "grad_norm": 0.7649189111405409, "learning_rate": 0.001473, "loss": 5.4131, "step": 491 }, { "epoch": 0.00492, "grad_norm": 0.7909135941762935, "learning_rate": 0.001476, "loss": 5.4088, "step": 492 }, { "epoch": 0.00493, "grad_norm": 0.9895574066894314, "learning_rate": 0.001479, "loss": 5.412, "step": 493 }, { "epoch": 0.00494, "grad_norm": 1.132991334341666, "learning_rate": 0.001482, "loss": 5.429, "step": 494 }, { "epoch": 0.00495, "grad_norm": 0.761507737933228, "learning_rate": 0.001485, "loss": 5.4134, "step": 495 }, { "epoch": 0.00496, "grad_norm": 0.9726159326361932, "learning_rate": 0.001488, "loss": 5.4067, "step": 496 }, { "epoch": 0.00497, "grad_norm": 1.1482864163713484, "learning_rate": 0.001491, "loss": 5.3997, "step": 497 }, { "epoch": 0.00498, "grad_norm": 0.8327332046897746, "learning_rate": 0.001494, "loss": 5.4147, "step": 498 }, { "epoch": 0.00499, "grad_norm": 0.725916611519047, "learning_rate": 0.001497, "loss": 5.3792, "step": 499 }, { "epoch": 0.005, "grad_norm": 0.6719782846245154, "learning_rate": 0.0015, "loss": 5.3842, "step": 500 }, { "epoch": 0.00501, "grad_norm": 0.7463076928465905, "learning_rate": 0.001503, "loss": 5.3796, "step": 501 }, { "epoch": 0.00502, "grad_norm": 0.8408104186601356, "learning_rate": 0.001506, "loss": 5.389, "step": 502 }, { "epoch": 0.00503, "grad_norm": 0.860909611402231, "learning_rate": 0.0015090000000000001, "loss": 5.3926, "step": 503 }, { "epoch": 0.00504, "grad_norm": 0.7026490274993983, "learning_rate": 0.001512, "loss": 5.3646, "step": 504 }, { "epoch": 0.00505, "grad_norm": 0.7184807991697565, "learning_rate": 0.001515, "loss": 5.3547, "step": 505 }, { "epoch": 0.00506, "grad_norm": 0.7839475253802514, "learning_rate": 0.001518, "loss": 5.3611, "step": 506 }, { "epoch": 0.00507, "grad_norm": 0.7039499110993044, "learning_rate": 0.001521, "loss": 5.3522, "step": 507 }, { "epoch": 0.00508, "grad_norm": 0.5587706859739108, "learning_rate": 0.001524, "loss": 5.3512, "step": 508 }, { "epoch": 0.00509, "grad_norm": 0.4952941325228141, "learning_rate": 0.0015270000000000001, "loss": 5.326, "step": 509 }, { "epoch": 0.0051, "grad_norm": 0.5131490664852795, "learning_rate": 0.0015300000000000001, "loss": 5.3428, "step": 510 }, { "epoch": 0.00511, "grad_norm": 0.5539213487194597, "learning_rate": 0.001533, "loss": 5.3196, "step": 511 }, { "epoch": 0.00512, "grad_norm": 0.5937876471208409, "learning_rate": 0.001536, "loss": 5.3158, "step": 512 }, { "epoch": 0.00513, "grad_norm": 0.5441672327389838, "learning_rate": 0.001539, "loss": 5.327, "step": 513 }, { "epoch": 0.00514, "grad_norm": 0.5371789671410057, "learning_rate": 0.001542, "loss": 5.3038, "step": 514 }, { "epoch": 0.00515, "grad_norm": 0.5194765862771661, "learning_rate": 0.0015450000000000001, "loss": 5.3109, "step": 515 }, { "epoch": 0.00516, "grad_norm": 0.5575198815834714, "learning_rate": 0.0015480000000000001, "loss": 5.3083, "step": 516 }, { "epoch": 0.00517, "grad_norm": 0.5237583962443445, "learning_rate": 0.001551, "loss": 5.2962, "step": 517 }, { "epoch": 0.00518, "grad_norm": 0.5237181167506534, "learning_rate": 0.001554, "loss": 5.2786, "step": 518 }, { "epoch": 0.00519, "grad_norm": 0.6652146969315359, "learning_rate": 0.001557, "loss": 5.2847, "step": 519 }, { "epoch": 0.0052, "grad_norm": 0.8817258231902962, "learning_rate": 0.0015600000000000002, "loss": 5.2966, "step": 520 }, { "epoch": 0.00521, "grad_norm": 0.9047592596397305, "learning_rate": 0.0015630000000000002, "loss": 5.2732, "step": 521 }, { "epoch": 0.00522, "grad_norm": 0.6139926424217688, "learning_rate": 0.0015660000000000001, "loss": 5.2701, "step": 522 }, { "epoch": 0.00523, "grad_norm": 0.6292464658556638, "learning_rate": 0.001569, "loss": 5.2826, "step": 523 }, { "epoch": 0.00524, "grad_norm": 0.752030715547053, "learning_rate": 0.001572, "loss": 5.2856, "step": 524 }, { "epoch": 0.00525, "grad_norm": 0.929770536798091, "learning_rate": 0.001575, "loss": 5.287, "step": 525 }, { "epoch": 0.00526, "grad_norm": 0.887370520628206, "learning_rate": 0.0015780000000000002, "loss": 5.2635, "step": 526 }, { "epoch": 0.00527, "grad_norm": 0.7819104471156305, "learning_rate": 0.0015810000000000002, "loss": 5.2728, "step": 527 }, { "epoch": 0.00528, "grad_norm": 0.9038037239389326, "learning_rate": 0.0015840000000000001, "loss": 5.2534, "step": 528 }, { "epoch": 0.00529, "grad_norm": 0.8898068580069259, "learning_rate": 0.001587, "loss": 5.265, "step": 529 }, { "epoch": 0.0053, "grad_norm": 1.0010848916209774, "learning_rate": 0.00159, "loss": 5.2764, "step": 530 }, { "epoch": 0.00531, "grad_norm": 1.0010984765594055, "learning_rate": 0.001593, "loss": 5.2677, "step": 531 }, { "epoch": 0.00532, "grad_norm": 1.026224264336229, "learning_rate": 0.0015960000000000002, "loss": 5.2779, "step": 532 }, { "epoch": 0.00533, "grad_norm": 0.9759289805508353, "learning_rate": 0.0015990000000000002, "loss": 5.2678, "step": 533 }, { "epoch": 0.00534, "grad_norm": 1.1376558518204782, "learning_rate": 0.0016020000000000001, "loss": 5.2612, "step": 534 }, { "epoch": 0.00535, "grad_norm": 1.0517455067238486, "learning_rate": 0.001605, "loss": 5.2672, "step": 535 }, { "epoch": 0.00536, "grad_norm": 0.9398243649272562, "learning_rate": 0.001608, "loss": 5.2627, "step": 536 }, { "epoch": 0.00537, "grad_norm": 0.9512995727424398, "learning_rate": 0.0016110000000000002, "loss": 5.2558, "step": 537 }, { "epoch": 0.00538, "grad_norm": 1.2746657535312511, "learning_rate": 0.0016140000000000002, "loss": 5.2451, "step": 538 }, { "epoch": 0.00539, "grad_norm": 1.067518036012326, "learning_rate": 0.0016170000000000002, "loss": 5.2415, "step": 539 }, { "epoch": 0.0054, "grad_norm": 1.1859628541053806, "learning_rate": 0.0016200000000000001, "loss": 5.2618, "step": 540 }, { "epoch": 0.00541, "grad_norm": 0.8882936824028492, "learning_rate": 0.001623, "loss": 5.2308, "step": 541 }, { "epoch": 0.00542, "grad_norm": 0.8517075205384302, "learning_rate": 0.001626, "loss": 5.2545, "step": 542 }, { "epoch": 0.00543, "grad_norm": 0.8283552605034004, "learning_rate": 0.0016290000000000002, "loss": 5.2098, "step": 543 }, { "epoch": 0.00544, "grad_norm": 0.9087829134911761, "learning_rate": 0.0016320000000000002, "loss": 5.2265, "step": 544 }, { "epoch": 0.00545, "grad_norm": 0.8034144620978907, "learning_rate": 0.0016350000000000002, "loss": 5.2348, "step": 545 }, { "epoch": 0.00546, "grad_norm": 0.7091235133563132, "learning_rate": 0.0016380000000000001, "loss": 5.2004, "step": 546 }, { "epoch": 0.00547, "grad_norm": 0.6683331586466694, "learning_rate": 0.001641, "loss": 5.1915, "step": 547 }, { "epoch": 0.00548, "grad_norm": 0.5441552662279447, "learning_rate": 0.001644, "loss": 5.1843, "step": 548 }, { "epoch": 0.00549, "grad_norm": 0.5462993858197037, "learning_rate": 0.0016470000000000002, "loss": 5.1806, "step": 549 }, { "epoch": 0.0055, "grad_norm": 0.5691406737163984, "learning_rate": 0.0016500000000000002, "loss": 5.1914, "step": 550 }, { "epoch": 0.00551, "grad_norm": 0.57998143312047, "learning_rate": 0.0016530000000000002, "loss": 5.1789, "step": 551 }, { "epoch": 0.00552, "grad_norm": 0.5286482487653069, "learning_rate": 0.0016560000000000001, "loss": 5.158, "step": 552 }, { "epoch": 0.00553, "grad_norm": 0.48759014943874474, "learning_rate": 0.001659, "loss": 5.1465, "step": 553 }, { "epoch": 0.00554, "grad_norm": 0.4792524526847805, "learning_rate": 0.0016620000000000003, "loss": 5.1537, "step": 554 }, { "epoch": 0.00555, "grad_norm": 0.5054837577144806, "learning_rate": 0.0016650000000000002, "loss": 5.1496, "step": 555 }, { "epoch": 0.00556, "grad_norm": 0.5148667639200912, "learning_rate": 0.0016680000000000002, "loss": 5.1288, "step": 556 }, { "epoch": 0.00557, "grad_norm": 0.5122706189056161, "learning_rate": 0.0016710000000000002, "loss": 5.1153, "step": 557 }, { "epoch": 0.00558, "grad_norm": 0.6046224327964763, "learning_rate": 0.0016740000000000001, "loss": 5.1358, "step": 558 }, { "epoch": 0.00559, "grad_norm": 0.6851636176151574, "learning_rate": 0.001677, "loss": 5.1268, "step": 559 }, { "epoch": 0.0056, "grad_norm": 0.6922547745331437, "learning_rate": 0.0016800000000000003, "loss": 5.096, "step": 560 }, { "epoch": 0.00561, "grad_norm": 0.7411075733344746, "learning_rate": 0.0016830000000000003, "loss": 5.1176, "step": 561 }, { "epoch": 0.00562, "grad_norm": 0.8981627852593407, "learning_rate": 0.0016860000000000002, "loss": 5.1206, "step": 562 }, { "epoch": 0.00563, "grad_norm": 1.1260200882948381, "learning_rate": 0.001689, "loss": 5.1239, "step": 563 }, { "epoch": 0.00564, "grad_norm": 1.1027210513289374, "learning_rate": 0.001692, "loss": 5.113, "step": 564 }, { "epoch": 0.00565, "grad_norm": 0.890213024695838, "learning_rate": 0.001695, "loss": 5.1071, "step": 565 }, { "epoch": 0.00566, "grad_norm": 0.8627979775394023, "learning_rate": 0.0016979999999999999, "loss": 5.0976, "step": 566 }, { "epoch": 0.00567, "grad_norm": 0.9823945407746334, "learning_rate": 0.0017009999999999998, "loss": 5.101, "step": 567 }, { "epoch": 0.00568, "grad_norm": 0.9715849000742567, "learning_rate": 0.0017039999999999998, "loss": 5.1139, "step": 568 }, { "epoch": 0.00569, "grad_norm": 0.9989913016974431, "learning_rate": 0.001707, "loss": 5.0861, "step": 569 }, { "epoch": 0.0057, "grad_norm": 1.0927877338999235, "learning_rate": 0.00171, "loss": 5.1262, "step": 570 }, { "epoch": 0.00571, "grad_norm": 1.0328737042176641, "learning_rate": 0.001713, "loss": 5.1192, "step": 571 }, { "epoch": 0.00572, "grad_norm": 1.3722388350456287, "learning_rate": 0.0017159999999999999, "loss": 5.1049, "step": 572 }, { "epoch": 0.00573, "grad_norm": 0.9525979203379623, "learning_rate": 0.0017189999999999998, "loss": 5.1081, "step": 573 }, { "epoch": 0.00574, "grad_norm": 1.0626128534442882, "learning_rate": 0.001722, "loss": 5.1048, "step": 574 }, { "epoch": 0.00575, "grad_norm": 0.9331527599734185, "learning_rate": 0.001725, "loss": 5.074, "step": 575 }, { "epoch": 0.00576, "grad_norm": 0.9277735357052385, "learning_rate": 0.001728, "loss": 5.0843, "step": 576 }, { "epoch": 0.00577, "grad_norm": 0.9070967517565243, "learning_rate": 0.001731, "loss": 5.0908, "step": 577 }, { "epoch": 0.00578, "grad_norm": 0.8451551366430134, "learning_rate": 0.0017339999999999999, "loss": 5.0704, "step": 578 }, { "epoch": 0.00579, "grad_norm": 0.7590562073625285, "learning_rate": 0.0017369999999999998, "loss": 5.058, "step": 579 }, { "epoch": 0.0058, "grad_norm": 0.6385326977156373, "learning_rate": 0.00174, "loss": 5.0662, "step": 580 }, { "epoch": 0.00581, "grad_norm": 0.5982129206576257, "learning_rate": 0.001743, "loss": 5.0624, "step": 581 }, { "epoch": 0.00582, "grad_norm": 0.7358321954275717, "learning_rate": 0.001746, "loss": 5.0371, "step": 582 }, { "epoch": 0.00583, "grad_norm": 0.790704997209863, "learning_rate": 0.001749, "loss": 5.0641, "step": 583 }, { "epoch": 0.00584, "grad_norm": 0.7351931257233056, "learning_rate": 0.0017519999999999999, "loss": 5.0549, "step": 584 }, { "epoch": 0.00585, "grad_norm": 0.6167540927458872, "learning_rate": 0.0017549999999999998, "loss": 5.0362, "step": 585 }, { "epoch": 0.00586, "grad_norm": 0.6143004439139178, "learning_rate": 0.001758, "loss": 5.0316, "step": 586 }, { "epoch": 0.00587, "grad_norm": 0.619830997492515, "learning_rate": 0.001761, "loss": 5.018, "step": 587 }, { "epoch": 0.00588, "grad_norm": 0.676754014852622, "learning_rate": 0.001764, "loss": 5.0113, "step": 588 }, { "epoch": 0.00589, "grad_norm": 0.8152367183756798, "learning_rate": 0.001767, "loss": 5.0063, "step": 589 }, { "epoch": 0.0059, "grad_norm": 0.8067730505459064, "learning_rate": 0.0017699999999999999, "loss": 5.0009, "step": 590 }, { "epoch": 0.00591, "grad_norm": 0.7857115694134444, "learning_rate": 0.001773, "loss": 5.0014, "step": 591 }, { "epoch": 0.00592, "grad_norm": 1.0305248086573016, "learning_rate": 0.001776, "loss": 5.0384, "step": 592 }, { "epoch": 0.00593, "grad_norm": 1.1034322728795254, "learning_rate": 0.001779, "loss": 5.0387, "step": 593 }, { "epoch": 0.00594, "grad_norm": 0.796782800390064, "learning_rate": 0.001782, "loss": 5.0138, "step": 594 }, { "epoch": 0.00595, "grad_norm": 0.947951394101208, "learning_rate": 0.001785, "loss": 5.0288, "step": 595 }, { "epoch": 0.00596, "grad_norm": 0.948001541672118, "learning_rate": 0.0017879999999999999, "loss": 5.0236, "step": 596 }, { "epoch": 0.00597, "grad_norm": 0.8614311642183788, "learning_rate": 0.001791, "loss": 4.983, "step": 597 }, { "epoch": 0.00598, "grad_norm": 0.7822502963641336, "learning_rate": 0.001794, "loss": 5.0032, "step": 598 }, { "epoch": 0.00599, "grad_norm": 0.662502314672726, "learning_rate": 0.001797, "loss": 4.9824, "step": 599 }, { "epoch": 0.006, "grad_norm": 0.612691797581944, "learning_rate": 0.0018, "loss": 4.9865, "step": 600 }, { "epoch": 0.00601, "grad_norm": 0.6872791409552744, "learning_rate": 0.001803, "loss": 4.9768, "step": 601 }, { "epoch": 0.00602, "grad_norm": 0.6805787880992017, "learning_rate": 0.0018059999999999999, "loss": 4.9611, "step": 602 }, { "epoch": 0.00603, "grad_norm": 0.6726518476757681, "learning_rate": 0.001809, "loss": 4.9863, "step": 603 }, { "epoch": 0.00604, "grad_norm": 0.6569683723636671, "learning_rate": 0.001812, "loss": 4.9594, "step": 604 }, { "epoch": 0.00605, "grad_norm": 0.7635898698708745, "learning_rate": 0.001815, "loss": 4.9733, "step": 605 }, { "epoch": 0.00606, "grad_norm": 0.7455554600622951, "learning_rate": 0.001818, "loss": 4.9743, "step": 606 }, { "epoch": 0.00607, "grad_norm": 0.7493825131119356, "learning_rate": 0.001821, "loss": 4.968, "step": 607 }, { "epoch": 0.00608, "grad_norm": 0.7957828614510537, "learning_rate": 0.001824, "loss": 4.9536, "step": 608 }, { "epoch": 0.00609, "grad_norm": 0.8304899179729943, "learning_rate": 0.001827, "loss": 4.9564, "step": 609 }, { "epoch": 0.0061, "grad_norm": 0.8109281284885322, "learning_rate": 0.00183, "loss": 4.9141, "step": 610 }, { "epoch": 0.00611, "grad_norm": 0.7476158062529945, "learning_rate": 0.001833, "loss": 4.936, "step": 611 }, { "epoch": 0.00612, "grad_norm": 0.7704555466663776, "learning_rate": 0.001836, "loss": 4.9262, "step": 612 }, { "epoch": 0.00613, "grad_norm": 0.7179188189545482, "learning_rate": 0.001839, "loss": 4.9285, "step": 613 }, { "epoch": 0.00614, "grad_norm": 0.6487628040476678, "learning_rate": 0.001842, "loss": 4.9483, "step": 614 }, { "epoch": 0.00615, "grad_norm": 0.6346341426047015, "learning_rate": 0.001845, "loss": 4.9336, "step": 615 }, { "epoch": 0.00616, "grad_norm": 0.6919482296512011, "learning_rate": 0.001848, "loss": 4.92, "step": 616 }, { "epoch": 0.00617, "grad_norm": 0.8470324326487232, "learning_rate": 0.001851, "loss": 4.937, "step": 617 }, { "epoch": 0.00618, "grad_norm": 0.9780451420741652, "learning_rate": 0.001854, "loss": 4.918, "step": 618 }, { "epoch": 0.00619, "grad_norm": 1.0393940937639112, "learning_rate": 0.001857, "loss": 4.9393, "step": 619 }, { "epoch": 0.0062, "grad_norm": 1.0864888594503201, "learning_rate": 0.00186, "loss": 4.9315, "step": 620 }, { "epoch": 0.00621, "grad_norm": 1.1243573748169302, "learning_rate": 0.001863, "loss": 4.931, "step": 621 }, { "epoch": 0.00622, "grad_norm": 1.2701399731469665, "learning_rate": 0.001866, "loss": 4.9502, "step": 622 }, { "epoch": 0.00623, "grad_norm": 0.8273605340089976, "learning_rate": 0.001869, "loss": 4.9139, "step": 623 }, { "epoch": 0.00624, "grad_norm": 0.9643311427939416, "learning_rate": 0.001872, "loss": 4.9253, "step": 624 }, { "epoch": 0.00625, "grad_norm": 1.2310098040523034, "learning_rate": 0.001875, "loss": 4.9539, "step": 625 }, { "epoch": 0.00626, "grad_norm": 0.8268848993203812, "learning_rate": 0.0018780000000000001, "loss": 4.9199, "step": 626 }, { "epoch": 0.00627, "grad_norm": 0.9575639973355906, "learning_rate": 0.001881, "loss": 4.909, "step": 627 }, { "epoch": 0.00628, "grad_norm": 0.9391450767019313, "learning_rate": 0.001884, "loss": 4.9487, "step": 628 }, { "epoch": 0.00629, "grad_norm": 0.8387975784827085, "learning_rate": 0.001887, "loss": 4.8976, "step": 629 }, { "epoch": 0.0063, "grad_norm": 0.7745062985579545, "learning_rate": 0.00189, "loss": 4.9062, "step": 630 }, { "epoch": 0.00631, "grad_norm": 0.8213366097020988, "learning_rate": 0.0018930000000000002, "loss": 4.9006, "step": 631 }, { "epoch": 0.00632, "grad_norm": 0.8670608316656828, "learning_rate": 0.0018960000000000001, "loss": 4.9226, "step": 632 }, { "epoch": 0.00633, "grad_norm": 0.9183940724690223, "learning_rate": 0.001899, "loss": 4.9055, "step": 633 }, { "epoch": 0.00634, "grad_norm": 0.8953336214105924, "learning_rate": 0.001902, "loss": 4.8931, "step": 634 }, { "epoch": 0.00635, "grad_norm": 0.8869843611928877, "learning_rate": 0.001905, "loss": 4.8851, "step": 635 }, { "epoch": 0.00636, "grad_norm": 1.014617544478078, "learning_rate": 0.001908, "loss": 4.911, "step": 636 }, { "epoch": 0.00637, "grad_norm": 0.8681767489690893, "learning_rate": 0.0019110000000000002, "loss": 4.909, "step": 637 }, { "epoch": 0.00638, "grad_norm": 0.9792692499883113, "learning_rate": 0.0019140000000000001, "loss": 4.9073, "step": 638 }, { "epoch": 0.00639, "grad_norm": 0.8702384519183142, "learning_rate": 0.001917, "loss": 4.8724, "step": 639 }, { "epoch": 0.0064, "grad_norm": 0.7656979903996737, "learning_rate": 0.00192, "loss": 4.8809, "step": 640 }, { "epoch": 0.00641, "grad_norm": 0.7895438458212233, "learning_rate": 0.001923, "loss": 4.8762, "step": 641 }, { "epoch": 0.00642, "grad_norm": 0.8281699495124627, "learning_rate": 0.001926, "loss": 4.8749, "step": 642 }, { "epoch": 0.00643, "grad_norm": 0.7845360432833325, "learning_rate": 0.0019290000000000002, "loss": 4.8856, "step": 643 }, { "epoch": 0.00644, "grad_norm": 0.6857666802048429, "learning_rate": 0.0019320000000000001, "loss": 4.8583, "step": 644 }, { "epoch": 0.00645, "grad_norm": 0.5056941566313082, "learning_rate": 0.001935, "loss": 4.8522, "step": 645 }, { "epoch": 0.00646, "grad_norm": 0.47964630717910517, "learning_rate": 0.001938, "loss": 4.8628, "step": 646 }, { "epoch": 0.00647, "grad_norm": 0.519046406898008, "learning_rate": 0.001941, "loss": 4.8552, "step": 647 }, { "epoch": 0.00648, "grad_norm": 0.4935185080219269, "learning_rate": 0.0019440000000000002, "loss": 4.8488, "step": 648 }, { "epoch": 0.00649, "grad_norm": 0.5205599594018984, "learning_rate": 0.0019470000000000002, "loss": 4.8333, "step": 649 }, { "epoch": 0.0065, "grad_norm": 0.45295352715192905, "learning_rate": 0.0019500000000000001, "loss": 4.8617, "step": 650 }, { "epoch": 0.00651, "grad_norm": 0.4378817965282719, "learning_rate": 0.001953, "loss": 4.84, "step": 651 }, { "epoch": 0.00652, "grad_norm": 0.48395025642257355, "learning_rate": 0.0019560000000000003, "loss": 4.8266, "step": 652 }, { "epoch": 0.00653, "grad_norm": 0.5194845024092328, "learning_rate": 0.0019590000000000002, "loss": 4.8294, "step": 653 }, { "epoch": 0.00654, "grad_norm": 0.5707068593609262, "learning_rate": 0.001962, "loss": 4.8026, "step": 654 }, { "epoch": 0.00655, "grad_norm": 0.6725172209521785, "learning_rate": 0.001965, "loss": 4.834, "step": 655 }, { "epoch": 0.00656, "grad_norm": 0.7729568432311581, "learning_rate": 0.001968, "loss": 4.8398, "step": 656 }, { "epoch": 0.00657, "grad_norm": 0.9182317536069756, "learning_rate": 0.001971, "loss": 4.8175, "step": 657 }, { "epoch": 0.00658, "grad_norm": 0.9241834469814327, "learning_rate": 0.001974, "loss": 4.8482, "step": 658 }, { "epoch": 0.00659, "grad_norm": 0.7554289117841486, "learning_rate": 0.001977, "loss": 4.8199, "step": 659 }, { "epoch": 0.0066, "grad_norm": 0.786222644002383, "learning_rate": 0.00198, "loss": 4.8336, "step": 660 }, { "epoch": 0.00661, "grad_norm": 0.7486956414167568, "learning_rate": 0.001983, "loss": 4.8035, "step": 661 }, { "epoch": 0.00662, "grad_norm": 0.8224936329580866, "learning_rate": 0.0019860000000000004, "loss": 4.8038, "step": 662 }, { "epoch": 0.00663, "grad_norm": 1.0333891900276588, "learning_rate": 0.0019890000000000003, "loss": 4.8216, "step": 663 }, { "epoch": 0.00664, "grad_norm": 1.0454184895583787, "learning_rate": 0.0019920000000000003, "loss": 4.826, "step": 664 }, { "epoch": 0.00665, "grad_norm": 0.9684168839706867, "learning_rate": 0.0019950000000000002, "loss": 4.7952, "step": 665 }, { "epoch": 0.00666, "grad_norm": 1.0353944208678731, "learning_rate": 0.001998, "loss": 4.8213, "step": 666 }, { "epoch": 0.00667, "grad_norm": 0.9360230286599108, "learning_rate": 0.002001, "loss": 4.8166, "step": 667 }, { "epoch": 0.00668, "grad_norm": 0.7256915823313501, "learning_rate": 0.002004, "loss": 4.83, "step": 668 }, { "epoch": 0.00669, "grad_norm": 0.8394069732991961, "learning_rate": 0.002007, "loss": 4.8494, "step": 669 }, { "epoch": 0.0067, "grad_norm": 1.077577977883947, "learning_rate": 0.00201, "loss": 4.8532, "step": 670 }, { "epoch": 0.00671, "grad_norm": 0.944781077919628, "learning_rate": 0.002013, "loss": 4.8236, "step": 671 }, { "epoch": 0.00672, "grad_norm": 0.7024086289087288, "learning_rate": 0.002016, "loss": 4.7966, "step": 672 }, { "epoch": 0.00673, "grad_norm": 0.6931174948166909, "learning_rate": 0.002019, "loss": 4.809, "step": 673 }, { "epoch": 0.00674, "grad_norm": 0.6108100930625565, "learning_rate": 0.0020220000000000004, "loss": 4.8005, "step": 674 }, { "epoch": 0.00675, "grad_norm": 0.5498762136107717, "learning_rate": 0.0020250000000000003, "loss": 4.8171, "step": 675 }, { "epoch": 0.00676, "grad_norm": 0.6244881777878202, "learning_rate": 0.0020280000000000003, "loss": 4.7951, "step": 676 }, { "epoch": 0.00677, "grad_norm": 0.5815667573949113, "learning_rate": 0.0020310000000000003, "loss": 4.7913, "step": 677 }, { "epoch": 0.00678, "grad_norm": 0.521491074875552, "learning_rate": 0.0020340000000000002, "loss": 4.7959, "step": 678 }, { "epoch": 0.00679, "grad_norm": 0.5801192243287052, "learning_rate": 0.002037, "loss": 4.7864, "step": 679 }, { "epoch": 0.0068, "grad_norm": 0.6398788453868361, "learning_rate": 0.00204, "loss": 4.7875, "step": 680 }, { "epoch": 0.00681, "grad_norm": 0.6456567912241706, "learning_rate": 0.002043, "loss": 4.7807, "step": 681 }, { "epoch": 0.00682, "grad_norm": 0.6889535888719373, "learning_rate": 0.002046, "loss": 4.7776, "step": 682 }, { "epoch": 0.00683, "grad_norm": 0.68132819276425, "learning_rate": 0.002049, "loss": 4.7612, "step": 683 }, { "epoch": 0.00684, "grad_norm": 0.6724100911918068, "learning_rate": 0.002052, "loss": 4.7814, "step": 684 }, { "epoch": 0.00685, "grad_norm": 0.815928121483196, "learning_rate": 0.0020550000000000004, "loss": 4.7913, "step": 685 }, { "epoch": 0.00686, "grad_norm": 0.7537150884203632, "learning_rate": 0.0020580000000000004, "loss": 4.7584, "step": 686 }, { "epoch": 0.00687, "grad_norm": 0.7326770278926477, "learning_rate": 0.0020610000000000003, "loss": 4.7655, "step": 687 }, { "epoch": 0.00688, "grad_norm": 0.7145131916015852, "learning_rate": 0.002064, "loss": 4.7631, "step": 688 }, { "epoch": 0.00689, "grad_norm": 0.694828893634318, "learning_rate": 0.002067, "loss": 4.7685, "step": 689 }, { "epoch": 0.0069, "grad_norm": 0.7402628986076095, "learning_rate": 0.00207, "loss": 4.7629, "step": 690 }, { "epoch": 0.00691, "grad_norm": 0.7809038473743037, "learning_rate": 0.0020729999999999998, "loss": 4.734, "step": 691 }, { "epoch": 0.00692, "grad_norm": 0.9321737974219415, "learning_rate": 0.0020759999999999997, "loss": 4.7623, "step": 692 }, { "epoch": 0.00693, "grad_norm": 1.0940875671633443, "learning_rate": 0.0020789999999999997, "loss": 4.788, "step": 693 }, { "epoch": 0.00694, "grad_norm": 0.9030072391970948, "learning_rate": 0.002082, "loss": 4.7582, "step": 694 }, { "epoch": 0.00695, "grad_norm": 0.8857004379777869, "learning_rate": 0.002085, "loss": 4.7649, "step": 695 }, { "epoch": 0.00696, "grad_norm": 0.8991843577664433, "learning_rate": 0.002088, "loss": 4.7712, "step": 696 }, { "epoch": 0.00697, "grad_norm": 1.0094726318481275, "learning_rate": 0.002091, "loss": 4.7869, "step": 697 }, { "epoch": 0.00698, "grad_norm": 1.09376996979269, "learning_rate": 0.002094, "loss": 4.7773, "step": 698 }, { "epoch": 0.00699, "grad_norm": 0.8766228601427865, "learning_rate": 0.002097, "loss": 4.7548, "step": 699 }, { "epoch": 0.007, "grad_norm": 0.8580408760742435, "learning_rate": 0.0021, "loss": 4.7811, "step": 700 }, { "epoch": 0.00701, "grad_norm": 0.7398805390269545, "learning_rate": 0.002103, "loss": 4.7496, "step": 701 }, { "epoch": 0.00702, "grad_norm": 0.8329034135969645, "learning_rate": 0.002106, "loss": 4.762, "step": 702 }, { "epoch": 0.00703, "grad_norm": 0.8505784581176613, "learning_rate": 0.0021089999999999998, "loss": 4.7427, "step": 703 }, { "epoch": 0.00704, "grad_norm": 0.8003313240374758, "learning_rate": 0.0021119999999999997, "loss": 4.7516, "step": 704 }, { "epoch": 0.00705, "grad_norm": 0.7109629500256387, "learning_rate": 0.002115, "loss": 4.7635, "step": 705 }, { "epoch": 0.00706, "grad_norm": 0.6289559048526784, "learning_rate": 0.002118, "loss": 4.7473, "step": 706 }, { "epoch": 0.00707, "grad_norm": 0.5482382237711074, "learning_rate": 0.002121, "loss": 4.7193, "step": 707 }, { "epoch": 0.00708, "grad_norm": 0.5598390325768088, "learning_rate": 0.002124, "loss": 4.7015, "step": 708 }, { "epoch": 0.00709, "grad_norm": 0.5529551760214286, "learning_rate": 0.002127, "loss": 4.7331, "step": 709 }, { "epoch": 0.0071, "grad_norm": 0.6671084870548735, "learning_rate": 0.00213, "loss": 4.7271, "step": 710 }, { "epoch": 0.00711, "grad_norm": 1.1336515667563218, "learning_rate": 0.002133, "loss": 4.7225, "step": 711 }, { "epoch": 0.00712, "grad_norm": 0.8878745395415352, "learning_rate": 0.002136, "loss": 4.761, "step": 712 }, { "epoch": 0.00713, "grad_norm": 0.6767692274074788, "learning_rate": 0.002139, "loss": 4.7547, "step": 713 }, { "epoch": 0.00714, "grad_norm": 0.7698487468076475, "learning_rate": 0.002142, "loss": 4.7544, "step": 714 }, { "epoch": 0.00715, "grad_norm": 0.6189384208707256, "learning_rate": 0.0021449999999999998, "loss": 4.7077, "step": 715 }, { "epoch": 0.00716, "grad_norm": 0.5968344999809513, "learning_rate": 0.002148, "loss": 4.7064, "step": 716 }, { "epoch": 0.00717, "grad_norm": 0.5117629580972962, "learning_rate": 0.002151, "loss": 4.7111, "step": 717 }, { "epoch": 0.00718, "grad_norm": 0.5586203984201494, "learning_rate": 0.002154, "loss": 4.7079, "step": 718 }, { "epoch": 0.00719, "grad_norm": 0.5075479633130432, "learning_rate": 0.002157, "loss": 4.7121, "step": 719 }, { "epoch": 0.0072, "grad_norm": 0.4840284295584723, "learning_rate": 0.00216, "loss": 4.7045, "step": 720 }, { "epoch": 0.00721, "grad_norm": 0.4971167150526394, "learning_rate": 0.002163, "loss": 4.7056, "step": 721 }, { "epoch": 0.00722, "grad_norm": 0.5347204422021664, "learning_rate": 0.002166, "loss": 4.6807, "step": 722 }, { "epoch": 0.00723, "grad_norm": 0.5424206330484346, "learning_rate": 0.002169, "loss": 4.6942, "step": 723 }, { "epoch": 0.00724, "grad_norm": 0.5219918511718452, "learning_rate": 0.002172, "loss": 4.6912, "step": 724 }, { "epoch": 0.00725, "grad_norm": 0.4975444213381179, "learning_rate": 0.002175, "loss": 4.6834, "step": 725 }, { "epoch": 0.00726, "grad_norm": 0.5840922524665152, "learning_rate": 0.002178, "loss": 4.7008, "step": 726 }, { "epoch": 0.00727, "grad_norm": 0.7633698368434719, "learning_rate": 0.0021809999999999998, "loss": 4.6777, "step": 727 }, { "epoch": 0.00728, "grad_norm": 0.9052599955462425, "learning_rate": 0.002184, "loss": 4.7084, "step": 728 }, { "epoch": 0.00729, "grad_norm": 0.9052862109699649, "learning_rate": 0.002187, "loss": 4.7065, "step": 729 }, { "epoch": 0.0073, "grad_norm": 0.9788098741089467, "learning_rate": 0.00219, "loss": 4.7163, "step": 730 }, { "epoch": 0.00731, "grad_norm": 0.9218858714566311, "learning_rate": 0.002193, "loss": 4.7193, "step": 731 }, { "epoch": 0.00732, "grad_norm": 1.1412145043722297, "learning_rate": 0.002196, "loss": 4.7513, "step": 732 }, { "epoch": 0.00733, "grad_norm": 1.0851884555194036, "learning_rate": 0.002199, "loss": 4.7135, "step": 733 }, { "epoch": 0.00734, "grad_norm": 0.9295060612046938, "learning_rate": 0.002202, "loss": 4.745, "step": 734 }, { "epoch": 0.00735, "grad_norm": 1.1744848346988581, "learning_rate": 0.002205, "loss": 4.766, "step": 735 }, { "epoch": 0.00736, "grad_norm": 0.8278931321658332, "learning_rate": 0.002208, "loss": 4.7413, "step": 736 }, { "epoch": 0.00737, "grad_norm": 0.8170748474782394, "learning_rate": 0.002211, "loss": 4.701, "step": 737 }, { "epoch": 0.00738, "grad_norm": 0.772640277783885, "learning_rate": 0.002214, "loss": 4.7083, "step": 738 }, { "epoch": 0.00739, "grad_norm": 0.6328057525928463, "learning_rate": 0.0022170000000000002, "loss": 4.6886, "step": 739 }, { "epoch": 0.0074, "grad_norm": 0.6456777453177256, "learning_rate": 0.00222, "loss": 4.7001, "step": 740 }, { "epoch": 0.00741, "grad_norm": 0.5912580097221507, "learning_rate": 0.002223, "loss": 4.6652, "step": 741 }, { "epoch": 0.00742, "grad_norm": 0.4622238730689768, "learning_rate": 0.002226, "loss": 4.6972, "step": 742 }, { "epoch": 0.00743, "grad_norm": 0.4904966133460914, "learning_rate": 0.002229, "loss": 4.6585, "step": 743 }, { "epoch": 0.00744, "grad_norm": 0.5611948954376664, "learning_rate": 0.002232, "loss": 4.6643, "step": 744 }, { "epoch": 0.00745, "grad_norm": 0.6222946629538655, "learning_rate": 0.002235, "loss": 4.6777, "step": 745 }, { "epoch": 0.00746, "grad_norm": 0.6678221674623648, "learning_rate": 0.002238, "loss": 4.6735, "step": 746 }, { "epoch": 0.00747, "grad_norm": 0.7945455347777014, "learning_rate": 0.002241, "loss": 4.6637, "step": 747 }, { "epoch": 0.00748, "grad_norm": 0.9027593299924651, "learning_rate": 0.002244, "loss": 4.683, "step": 748 }, { "epoch": 0.00749, "grad_norm": 0.8224517051186598, "learning_rate": 0.002247, "loss": 4.6612, "step": 749 }, { "epoch": 0.0075, "grad_norm": 0.6737194715747743, "learning_rate": 0.0022500000000000003, "loss": 4.6955, "step": 750 }, { "epoch": 0.00751, "grad_norm": 0.7232822830562855, "learning_rate": 0.0022530000000000002, "loss": 4.6893, "step": 751 }, { "epoch": 0.00752, "grad_norm": 0.6449225627821611, "learning_rate": 0.002256, "loss": 4.637, "step": 752 }, { "epoch": 0.00753, "grad_norm": 0.6558342053271724, "learning_rate": 0.002259, "loss": 4.6782, "step": 753 }, { "epoch": 0.00754, "grad_norm": 0.7056102529212358, "learning_rate": 0.002262, "loss": 4.6906, "step": 754 }, { "epoch": 0.00755, "grad_norm": 0.9043185481447295, "learning_rate": 0.002265, "loss": 4.6606, "step": 755 }, { "epoch": 0.00756, "grad_norm": 0.8770375990036015, "learning_rate": 0.002268, "loss": 4.6551, "step": 756 }, { "epoch": 0.00757, "grad_norm": 0.784655485815756, "learning_rate": 0.002271, "loss": 4.6655, "step": 757 }, { "epoch": 0.00758, "grad_norm": 0.6912079950422929, "learning_rate": 0.002274, "loss": 4.655, "step": 758 }, { "epoch": 0.00759, "grad_norm": 0.6281371231708264, "learning_rate": 0.002277, "loss": 4.6629, "step": 759 }, { "epoch": 0.0076, "grad_norm": 0.5992557531443571, "learning_rate": 0.00228, "loss": 4.6463, "step": 760 }, { "epoch": 0.00761, "grad_norm": 0.6034804931159479, "learning_rate": 0.002283, "loss": 4.6146, "step": 761 }, { "epoch": 0.00762, "grad_norm": 0.6060668832918386, "learning_rate": 0.0022860000000000003, "loss": 4.5953, "step": 762 }, { "epoch": 0.00763, "grad_norm": 0.5816892837581542, "learning_rate": 0.0022890000000000002, "loss": 4.6459, "step": 763 }, { "epoch": 0.00764, "grad_norm": 0.599786845424844, "learning_rate": 0.002292, "loss": 4.6554, "step": 764 }, { "epoch": 0.00765, "grad_norm": 0.6640243443171511, "learning_rate": 0.002295, "loss": 4.6451, "step": 765 }, { "epoch": 0.00766, "grad_norm": 0.6663176934335011, "learning_rate": 0.002298, "loss": 4.6438, "step": 766 }, { "epoch": 0.00767, "grad_norm": 0.6384162747773746, "learning_rate": 0.002301, "loss": 4.6392, "step": 767 }, { "epoch": 0.00768, "grad_norm": 0.8537658256313299, "learning_rate": 0.002304, "loss": 4.6457, "step": 768 }, { "epoch": 0.00769, "grad_norm": 1.0098120726516953, "learning_rate": 0.002307, "loss": 4.6694, "step": 769 }, { "epoch": 0.0077, "grad_norm": 0.7423003820485509, "learning_rate": 0.00231, "loss": 4.6223, "step": 770 }, { "epoch": 0.00771, "grad_norm": 0.6691351211965346, "learning_rate": 0.002313, "loss": 4.6569, "step": 771 }, { "epoch": 0.00772, "grad_norm": 0.8185609860643415, "learning_rate": 0.002316, "loss": 4.6531, "step": 772 }, { "epoch": 0.00773, "grad_norm": 1.0039613061235502, "learning_rate": 0.0023190000000000003, "loss": 4.6664, "step": 773 }, { "epoch": 0.00774, "grad_norm": 0.8500607234716588, "learning_rate": 0.0023220000000000003, "loss": 4.6845, "step": 774 }, { "epoch": 0.00775, "grad_norm": 0.7057192127032131, "learning_rate": 0.0023250000000000002, "loss": 4.6688, "step": 775 }, { "epoch": 0.00776, "grad_norm": 0.8527617729361273, "learning_rate": 0.002328, "loss": 4.6732, "step": 776 }, { "epoch": 0.00777, "grad_norm": 0.7987783133918631, "learning_rate": 0.002331, "loss": 4.6469, "step": 777 }, { "epoch": 0.00778, "grad_norm": 0.8221364860163118, "learning_rate": 0.002334, "loss": 4.6645, "step": 778 }, { "epoch": 0.00779, "grad_norm": 0.8901185821524535, "learning_rate": 0.002337, "loss": 4.6243, "step": 779 }, { "epoch": 0.0078, "grad_norm": 0.889956887452623, "learning_rate": 0.00234, "loss": 4.646, "step": 780 }, { "epoch": 0.00781, "grad_norm": 0.763195026878423, "learning_rate": 0.002343, "loss": 4.6465, "step": 781 }, { "epoch": 0.00782, "grad_norm": 0.6941360203492353, "learning_rate": 0.002346, "loss": 4.6378, "step": 782 }, { "epoch": 0.00783, "grad_norm": 0.6504570243922513, "learning_rate": 0.002349, "loss": 4.6195, "step": 783 }, { "epoch": 0.00784, "grad_norm": 0.5612286515783087, "learning_rate": 0.002352, "loss": 4.6293, "step": 784 }, { "epoch": 0.00785, "grad_norm": 0.6491841258536649, "learning_rate": 0.0023550000000000003, "loss": 4.6258, "step": 785 }, { "epoch": 0.00786, "grad_norm": 0.7618604140557245, "learning_rate": 0.0023580000000000003, "loss": 4.6353, "step": 786 }, { "epoch": 0.00787, "grad_norm": 0.9289251225559676, "learning_rate": 0.0023610000000000003, "loss": 4.6478, "step": 787 }, { "epoch": 0.00788, "grad_norm": 1.0160410825626025, "learning_rate": 0.002364, "loss": 4.6569, "step": 788 }, { "epoch": 0.00789, "grad_norm": 0.9904632547103471, "learning_rate": 0.002367, "loss": 4.6841, "step": 789 }, { "epoch": 0.0079, "grad_norm": 0.978549262142993, "learning_rate": 0.00237, "loss": 4.6525, "step": 790 }, { "epoch": 0.00791, "grad_norm": 0.7626000247604355, "learning_rate": 0.002373, "loss": 4.6008, "step": 791 }, { "epoch": 0.00792, "grad_norm": 0.7032433100729657, "learning_rate": 0.002376, "loss": 4.6361, "step": 792 }, { "epoch": 0.00793, "grad_norm": 0.6164393933780811, "learning_rate": 0.002379, "loss": 4.6263, "step": 793 }, { "epoch": 0.00794, "grad_norm": 0.6849838631141278, "learning_rate": 0.002382, "loss": 4.6402, "step": 794 }, { "epoch": 0.00795, "grad_norm": 0.7541453624973187, "learning_rate": 0.002385, "loss": 4.6181, "step": 795 }, { "epoch": 0.00796, "grad_norm": 0.7900314117496404, "learning_rate": 0.0023880000000000004, "loss": 4.6154, "step": 796 }, { "epoch": 0.00797, "grad_norm": 0.8032302751261208, "learning_rate": 0.0023910000000000003, "loss": 4.6598, "step": 797 }, { "epoch": 0.00798, "grad_norm": 0.7037506066431396, "learning_rate": 0.0023940000000000003, "loss": 4.6206, "step": 798 }, { "epoch": 0.00799, "grad_norm": 0.6873143294249761, "learning_rate": 0.0023970000000000003, "loss": 4.6565, "step": 799 }, { "epoch": 0.008, "grad_norm": 0.5611968184575689, "learning_rate": 0.0024000000000000002, "loss": 4.6002, "step": 800 }, { "epoch": 0.00801, "grad_norm": 0.6431524925156966, "learning_rate": 0.002403, "loss": 4.6326, "step": 801 }, { "epoch": 0.00802, "grad_norm": 0.664361851854149, "learning_rate": 0.002406, "loss": 4.624, "step": 802 }, { "epoch": 0.00803, "grad_norm": 0.7385532765716779, "learning_rate": 0.002409, "loss": 4.6034, "step": 803 }, { "epoch": 0.00804, "grad_norm": 0.6763586866376322, "learning_rate": 0.002412, "loss": 4.609, "step": 804 }, { "epoch": 0.00805, "grad_norm": 0.4984516329675046, "learning_rate": 0.002415, "loss": 4.5958, "step": 805 }, { "epoch": 0.00806, "grad_norm": 0.6422151194456196, "learning_rate": 0.002418, "loss": 4.6202, "step": 806 }, { "epoch": 0.00807, "grad_norm": 0.6520347813599764, "learning_rate": 0.0024210000000000004, "loss": 4.6111, "step": 807 }, { "epoch": 0.00808, "grad_norm": 0.6466117231395049, "learning_rate": 0.0024240000000000004, "loss": 4.5938, "step": 808 }, { "epoch": 0.00809, "grad_norm": 0.7173763198404414, "learning_rate": 0.0024270000000000003, "loss": 4.5875, "step": 809 }, { "epoch": 0.0081, "grad_norm": 0.7630708316428362, "learning_rate": 0.0024300000000000003, "loss": 4.6257, "step": 810 }, { "epoch": 0.00811, "grad_norm": 0.6597258634486322, "learning_rate": 0.0024330000000000003, "loss": 4.5916, "step": 811 }, { "epoch": 0.00812, "grad_norm": 0.5596135689444884, "learning_rate": 0.0024360000000000002, "loss": 4.5639, "step": 812 }, { "epoch": 0.00813, "grad_norm": 0.5455291997717118, "learning_rate": 0.0024389999999999998, "loss": 4.6039, "step": 813 }, { "epoch": 0.00814, "grad_norm": 0.643394229982758, "learning_rate": 0.0024419999999999997, "loss": 4.6125, "step": 814 }, { "epoch": 0.00815, "grad_norm": 0.7154153020141445, "learning_rate": 0.0024449999999999997, "loss": 4.6204, "step": 815 }, { "epoch": 0.00816, "grad_norm": 0.7061775094598466, "learning_rate": 0.002448, "loss": 4.6005, "step": 816 }, { "epoch": 0.00817, "grad_norm": 0.7497949470997279, "learning_rate": 0.002451, "loss": 4.6054, "step": 817 }, { "epoch": 0.00818, "grad_norm": 0.8303821582765404, "learning_rate": 0.002454, "loss": 4.6047, "step": 818 }, { "epoch": 0.00819, "grad_norm": 0.7870772660210968, "learning_rate": 0.002457, "loss": 4.6129, "step": 819 }, { "epoch": 0.0082, "grad_norm": 0.8045799279983024, "learning_rate": 0.00246, "loss": 4.5959, "step": 820 }, { "epoch": 0.00821, "grad_norm": 0.6847814476141592, "learning_rate": 0.002463, "loss": 4.5753, "step": 821 }, { "epoch": 0.00822, "grad_norm": 0.6767342952639601, "learning_rate": 0.002466, "loss": 4.5743, "step": 822 }, { "epoch": 0.00823, "grad_norm": 0.7814972414925468, "learning_rate": 0.002469, "loss": 4.5892, "step": 823 }, { "epoch": 0.00824, "grad_norm": 0.7970591329920511, "learning_rate": 0.002472, "loss": 4.6124, "step": 824 }, { "epoch": 0.00825, "grad_norm": 0.7237059608391628, "learning_rate": 0.0024749999999999998, "loss": 4.6158, "step": 825 }, { "epoch": 0.00826, "grad_norm": 0.8825063151832095, "learning_rate": 0.0024779999999999997, "loss": 4.5838, "step": 826 }, { "epoch": 0.00827, "grad_norm": 0.895197450200386, "learning_rate": 0.002481, "loss": 4.6224, "step": 827 }, { "epoch": 0.00828, "grad_norm": 0.7760924275661016, "learning_rate": 0.002484, "loss": 4.6087, "step": 828 }, { "epoch": 0.00829, "grad_norm": 0.8835354563964355, "learning_rate": 0.002487, "loss": 4.6157, "step": 829 }, { "epoch": 0.0083, "grad_norm": 0.8984013262238549, "learning_rate": 0.00249, "loss": 4.6232, "step": 830 }, { "epoch": 0.00831, "grad_norm": 0.8653834946214679, "learning_rate": 0.002493, "loss": 4.6038, "step": 831 }, { "epoch": 0.00832, "grad_norm": 0.8641563564291244, "learning_rate": 0.002496, "loss": 4.6089, "step": 832 }, { "epoch": 0.00833, "grad_norm": 0.8849940351412556, "learning_rate": 0.002499, "loss": 4.6069, "step": 833 }, { "epoch": 0.00834, "grad_norm": 0.9333438912141526, "learning_rate": 0.002502, "loss": 4.6165, "step": 834 }, { "epoch": 0.00835, "grad_norm": 0.8853225829488065, "learning_rate": 0.002505, "loss": 4.6108, "step": 835 }, { "epoch": 0.00836, "grad_norm": 0.9042165265638229, "learning_rate": 0.002508, "loss": 4.6433, "step": 836 }, { "epoch": 0.00837, "grad_norm": 0.8034248166459536, "learning_rate": 0.0025109999999999998, "loss": 4.5745, "step": 837 }, { "epoch": 0.00838, "grad_norm": 0.8086998635281932, "learning_rate": 0.0025139999999999997, "loss": 4.6114, "step": 838 }, { "epoch": 0.00839, "grad_norm": 0.587375456181757, "learning_rate": 0.002517, "loss": 4.5954, "step": 839 }, { "epoch": 0.0084, "grad_norm": 0.6561997865962053, "learning_rate": 0.00252, "loss": 4.5818, "step": 840 }, { "epoch": 0.00841, "grad_norm": 0.6860624256233325, "learning_rate": 0.002523, "loss": 4.5813, "step": 841 }, { "epoch": 0.00842, "grad_norm": 0.6384378914075728, "learning_rate": 0.002526, "loss": 4.5783, "step": 842 }, { "epoch": 0.00843, "grad_norm": 0.5397302637540236, "learning_rate": 0.002529, "loss": 4.5728, "step": 843 }, { "epoch": 0.00844, "grad_norm": 0.5753282450552862, "learning_rate": 0.002532, "loss": 4.5912, "step": 844 }, { "epoch": 0.00845, "grad_norm": 0.605188801053789, "learning_rate": 0.002535, "loss": 4.5588, "step": 845 }, { "epoch": 0.00846, "grad_norm": 0.523231421872222, "learning_rate": 0.002538, "loss": 4.5771, "step": 846 }, { "epoch": 0.00847, "grad_norm": 0.4442194357732046, "learning_rate": 0.002541, "loss": 4.5539, "step": 847 }, { "epoch": 0.00848, "grad_norm": 0.4340595323225716, "learning_rate": 0.002544, "loss": 4.5562, "step": 848 }, { "epoch": 0.00849, "grad_norm": 0.40987686572717924, "learning_rate": 0.002547, "loss": 4.5471, "step": 849 }, { "epoch": 0.0085, "grad_norm": 0.38823479672015115, "learning_rate": 0.00255, "loss": 4.5547, "step": 850 }, { "epoch": 0.00851, "grad_norm": 0.42997075207568086, "learning_rate": 0.002553, "loss": 4.5733, "step": 851 }, { "epoch": 0.00852, "grad_norm": 0.5192269912554525, "learning_rate": 0.002556, "loss": 4.5624, "step": 852 }, { "epoch": 0.00853, "grad_norm": 0.6626364019198889, "learning_rate": 0.002559, "loss": 4.5465, "step": 853 }, { "epoch": 0.00854, "grad_norm": 0.9475750496728129, "learning_rate": 0.002562, "loss": 4.5544, "step": 854 }, { "epoch": 0.00855, "grad_norm": 0.9196861875673602, "learning_rate": 0.002565, "loss": 4.5554, "step": 855 }, { "epoch": 0.00856, "grad_norm": 0.6972970172901616, "learning_rate": 0.002568, "loss": 4.5747, "step": 856 }, { "epoch": 0.00857, "grad_norm": 0.7425487517589463, "learning_rate": 0.002571, "loss": 4.5717, "step": 857 }, { "epoch": 0.00858, "grad_norm": 0.678911102843075, "learning_rate": 0.002574, "loss": 4.5576, "step": 858 }, { "epoch": 0.00859, "grad_norm": 0.6680695283315139, "learning_rate": 0.002577, "loss": 4.5494, "step": 859 }, { "epoch": 0.0086, "grad_norm": 0.7323389776740598, "learning_rate": 0.00258, "loss": 4.5797, "step": 860 }, { "epoch": 0.00861, "grad_norm": 0.6564756555111457, "learning_rate": 0.0025830000000000002, "loss": 4.5928, "step": 861 }, { "epoch": 0.00862, "grad_norm": 0.6458920384469554, "learning_rate": 0.002586, "loss": 4.5402, "step": 862 }, { "epoch": 0.00863, "grad_norm": 0.56307833723133, "learning_rate": 0.002589, "loss": 4.5659, "step": 863 }, { "epoch": 0.00864, "grad_norm": 0.6301527121328465, "learning_rate": 0.002592, "loss": 4.5169, "step": 864 }, { "epoch": 0.00865, "grad_norm": 0.5741592962502369, "learning_rate": 0.002595, "loss": 4.5342, "step": 865 }, { "epoch": 0.00866, "grad_norm": 0.6185631950272453, "learning_rate": 0.002598, "loss": 4.5624, "step": 866 }, { "epoch": 0.00867, "grad_norm": 0.7226342011827854, "learning_rate": 0.002601, "loss": 4.5538, "step": 867 }, { "epoch": 0.00868, "grad_norm": 0.6871063796143311, "learning_rate": 0.002604, "loss": 4.5354, "step": 868 }, { "epoch": 0.00869, "grad_norm": 0.6146544871717051, "learning_rate": 0.002607, "loss": 4.5411, "step": 869 }, { "epoch": 0.0087, "grad_norm": 0.5850681754527672, "learning_rate": 0.00261, "loss": 4.5279, "step": 870 }, { "epoch": 0.00871, "grad_norm": 0.6619050810997609, "learning_rate": 0.002613, "loss": 4.5496, "step": 871 }, { "epoch": 0.00872, "grad_norm": 0.6309510333250544, "learning_rate": 0.002616, "loss": 4.5312, "step": 872 }, { "epoch": 0.00873, "grad_norm": 0.620458951387438, "learning_rate": 0.0026190000000000002, "loss": 4.5409, "step": 873 }, { "epoch": 0.00874, "grad_norm": 0.7675582961358233, "learning_rate": 0.002622, "loss": 4.5679, "step": 874 }, { "epoch": 0.00875, "grad_norm": 0.8413199287183839, "learning_rate": 0.002625, "loss": 4.5217, "step": 875 }, { "epoch": 0.00876, "grad_norm": 0.6439305106538762, "learning_rate": 0.002628, "loss": 4.5435, "step": 876 }, { "epoch": 0.00877, "grad_norm": 0.6155973146282673, "learning_rate": 0.002631, "loss": 4.5607, "step": 877 }, { "epoch": 0.00878, "grad_norm": 0.6572664983016872, "learning_rate": 0.002634, "loss": 4.4959, "step": 878 }, { "epoch": 0.00879, "grad_norm": 0.678308084092591, "learning_rate": 0.002637, "loss": 4.526, "step": 879 }, { "epoch": 0.0088, "grad_norm": 0.682644138366955, "learning_rate": 0.00264, "loss": 4.5397, "step": 880 }, { "epoch": 0.00881, "grad_norm": 0.6916960221367301, "learning_rate": 0.002643, "loss": 4.5424, "step": 881 }, { "epoch": 0.00882, "grad_norm": 0.8625989079873627, "learning_rate": 0.002646, "loss": 4.5523, "step": 882 }, { "epoch": 0.00883, "grad_norm": 0.9115056000231379, "learning_rate": 0.002649, "loss": 4.5538, "step": 883 }, { "epoch": 0.00884, "grad_norm": 1.1386535317654105, "learning_rate": 0.0026520000000000003, "loss": 4.5528, "step": 884 }, { "epoch": 0.00885, "grad_norm": 0.9401738564991132, "learning_rate": 0.0026550000000000002, "loss": 4.5478, "step": 885 }, { "epoch": 0.00886, "grad_norm": 1.0261918984073035, "learning_rate": 0.002658, "loss": 4.5785, "step": 886 }, { "epoch": 0.00887, "grad_norm": 1.2427377192417122, "learning_rate": 0.002661, "loss": 4.6, "step": 887 }, { "epoch": 0.00888, "grad_norm": 1.061546131863703, "learning_rate": 0.002664, "loss": 4.5885, "step": 888 }, { "epoch": 0.00889, "grad_norm": 0.907593068920597, "learning_rate": 0.002667, "loss": 4.535, "step": 889 }, { "epoch": 0.0089, "grad_norm": 0.8129313603021157, "learning_rate": 0.00267, "loss": 4.5855, "step": 890 }, { "epoch": 0.00891, "grad_norm": 0.8985699151758613, "learning_rate": 0.002673, "loss": 4.5762, "step": 891 }, { "epoch": 0.00892, "grad_norm": 0.9328845387585828, "learning_rate": 0.002676, "loss": 4.5962, "step": 892 }, { "epoch": 0.00893, "grad_norm": 0.9658499469408992, "learning_rate": 0.002679, "loss": 4.5743, "step": 893 }, { "epoch": 0.00894, "grad_norm": 0.9262060176852961, "learning_rate": 0.002682, "loss": 4.5624, "step": 894 }, { "epoch": 0.00895, "grad_norm": 1.0340643070320608, "learning_rate": 0.0026850000000000003, "loss": 4.5798, "step": 895 }, { "epoch": 0.00896, "grad_norm": 0.961568519948311, "learning_rate": 0.0026880000000000003, "loss": 4.5724, "step": 896 }, { "epoch": 0.00897, "grad_norm": 0.9522979094666822, "learning_rate": 0.0026910000000000002, "loss": 4.5901, "step": 897 }, { "epoch": 0.00898, "grad_norm": 0.90558868323236, "learning_rate": 0.002694, "loss": 4.5802, "step": 898 }, { "epoch": 0.00899, "grad_norm": 0.7752291080346148, "learning_rate": 0.002697, "loss": 4.5848, "step": 899 }, { "epoch": 0.009, "grad_norm": 0.7198757797557334, "learning_rate": 0.0027, "loss": 4.5998, "step": 900 }, { "epoch": 0.00901, "grad_norm": 0.7328950523365488, "learning_rate": 0.002703, "loss": 4.5552, "step": 901 }, { "epoch": 0.00902, "grad_norm": 0.5974121750232518, "learning_rate": 0.002706, "loss": 4.558, "step": 902 }, { "epoch": 0.00903, "grad_norm": 0.5753314236304967, "learning_rate": 0.002709, "loss": 4.5534, "step": 903 }, { "epoch": 0.00904, "grad_norm": 0.6128643778394687, "learning_rate": 0.002712, "loss": 4.5329, "step": 904 }, { "epoch": 0.00905, "grad_norm": 0.5691916638541509, "learning_rate": 0.002715, "loss": 4.5446, "step": 905 }, { "epoch": 0.00906, "grad_norm": 0.5122473095411203, "learning_rate": 0.002718, "loss": 4.5538, "step": 906 }, { "epoch": 0.00907, "grad_norm": 0.4479956748150508, "learning_rate": 0.0027210000000000003, "loss": 4.5603, "step": 907 }, { "epoch": 0.00908, "grad_norm": 0.4387217017050332, "learning_rate": 0.0027240000000000003, "loss": 4.5304, "step": 908 }, { "epoch": 0.00909, "grad_norm": 0.4358600325368994, "learning_rate": 0.0027270000000000003, "loss": 4.5454, "step": 909 }, { "epoch": 0.0091, "grad_norm": 0.3849373702796595, "learning_rate": 0.0027300000000000002, "loss": 4.5477, "step": 910 }, { "epoch": 0.00911, "grad_norm": 0.4574591880095763, "learning_rate": 0.002733, "loss": 4.5038, "step": 911 }, { "epoch": 0.00912, "grad_norm": 0.6121326743360266, "learning_rate": 0.002736, "loss": 4.5131, "step": 912 }, { "epoch": 0.00913, "grad_norm": 0.6845161679319078, "learning_rate": 0.002739, "loss": 4.513, "step": 913 }, { "epoch": 0.00914, "grad_norm": 0.7333722528690678, "learning_rate": 0.002742, "loss": 4.5432, "step": 914 }, { "epoch": 0.00915, "grad_norm": 0.6714442153045657, "learning_rate": 0.002745, "loss": 4.5048, "step": 915 }, { "epoch": 0.00916, "grad_norm": 0.5861682792098877, "learning_rate": 0.002748, "loss": 4.4768, "step": 916 }, { "epoch": 0.00917, "grad_norm": 0.7568015257717708, "learning_rate": 0.002751, "loss": 4.4985, "step": 917 }, { "epoch": 0.00918, "grad_norm": 0.67845085346177, "learning_rate": 0.0027540000000000004, "loss": 4.5185, "step": 918 }, { "epoch": 0.00919, "grad_norm": 0.6376249584521229, "learning_rate": 0.0027570000000000003, "loss": 4.4976, "step": 919 }, { "epoch": 0.0092, "grad_norm": 0.5953295490629925, "learning_rate": 0.0027600000000000003, "loss": 4.5104, "step": 920 }, { "epoch": 0.00921, "grad_norm": 0.5470113372090761, "learning_rate": 0.0027630000000000003, "loss": 4.4892, "step": 921 }, { "epoch": 0.00922, "grad_norm": 0.5586775976924615, "learning_rate": 0.0027660000000000002, "loss": 4.5007, "step": 922 }, { "epoch": 0.00923, "grad_norm": 0.5602976125001545, "learning_rate": 0.002769, "loss": 4.4767, "step": 923 }, { "epoch": 0.00924, "grad_norm": 0.5512257756150963, "learning_rate": 0.002772, "loss": 4.4992, "step": 924 }, { "epoch": 0.00925, "grad_norm": 0.6104014154605217, "learning_rate": 0.002775, "loss": 4.4804, "step": 925 }, { "epoch": 0.00926, "grad_norm": 0.627882987139063, "learning_rate": 0.002778, "loss": 4.4994, "step": 926 }, { "epoch": 0.00927, "grad_norm": 0.5637389395797978, "learning_rate": 0.002781, "loss": 4.4582, "step": 927 }, { "epoch": 0.00928, "grad_norm": 0.5793012970445044, "learning_rate": 0.002784, "loss": 4.505, "step": 928 }, { "epoch": 0.00929, "grad_norm": 0.5538600481661593, "learning_rate": 0.0027870000000000004, "loss": 4.4867, "step": 929 }, { "epoch": 0.0093, "grad_norm": 0.4915384406563099, "learning_rate": 0.0027900000000000004, "loss": 4.4652, "step": 930 }, { "epoch": 0.00931, "grad_norm": 0.5291025545236148, "learning_rate": 0.0027930000000000003, "loss": 4.4796, "step": 931 }, { "epoch": 0.00932, "grad_norm": 0.5189072536981689, "learning_rate": 0.0027960000000000003, "loss": 4.4881, "step": 932 }, { "epoch": 0.00933, "grad_norm": 0.4391692113712953, "learning_rate": 0.0027990000000000003, "loss": 4.4495, "step": 933 }, { "epoch": 0.00934, "grad_norm": 0.5067956527040228, "learning_rate": 0.0028020000000000002, "loss": 4.4811, "step": 934 }, { "epoch": 0.00935, "grad_norm": 0.5752082887405049, "learning_rate": 0.002805, "loss": 4.4893, "step": 935 }, { "epoch": 0.00936, "grad_norm": 0.8597610157431208, "learning_rate": 0.002808, "loss": 4.5012, "step": 936 }, { "epoch": 0.00937, "grad_norm": 1.0923006968336397, "learning_rate": 0.002811, "loss": 4.5462, "step": 937 }, { "epoch": 0.00938, "grad_norm": 0.8211742236965874, "learning_rate": 0.002814, "loss": 4.5052, "step": 938 }, { "epoch": 0.00939, "grad_norm": 0.699189062953733, "learning_rate": 0.002817, "loss": 4.4989, "step": 939 }, { "epoch": 0.0094, "grad_norm": 0.6251265662758575, "learning_rate": 0.00282, "loss": 4.495, "step": 940 }, { "epoch": 0.00941, "grad_norm": 0.8614570213641495, "learning_rate": 0.002823, "loss": 4.4823, "step": 941 }, { "epoch": 0.00942, "grad_norm": 0.9710835908546839, "learning_rate": 0.002826, "loss": 4.5282, "step": 942 }, { "epoch": 0.00943, "grad_norm": 0.8255156288963836, "learning_rate": 0.002829, "loss": 4.5175, "step": 943 }, { "epoch": 0.00944, "grad_norm": 0.8529303551550831, "learning_rate": 0.002832, "loss": 4.5233, "step": 944 }, { "epoch": 0.00945, "grad_norm": 0.8805636588256177, "learning_rate": 0.002835, "loss": 4.5307, "step": 945 }, { "epoch": 0.00946, "grad_norm": 0.8815394682869104, "learning_rate": 0.002838, "loss": 4.545, "step": 946 }, { "epoch": 0.00947, "grad_norm": 0.9792537736888025, "learning_rate": 0.0028409999999999998, "loss": 4.5259, "step": 947 }, { "epoch": 0.00948, "grad_norm": 1.16480627354709, "learning_rate": 0.0028439999999999997, "loss": 4.5258, "step": 948 }, { "epoch": 0.00949, "grad_norm": 0.9311558570406092, "learning_rate": 0.002847, "loss": 4.5678, "step": 949 }, { "epoch": 0.0095, "grad_norm": 0.9151500308490546, "learning_rate": 0.00285, "loss": 4.5566, "step": 950 }, { "epoch": 0.00951, "grad_norm": 0.8788038193040383, "learning_rate": 0.002853, "loss": 4.539, "step": 951 }, { "epoch": 0.00952, "grad_norm": 0.6883343465241535, "learning_rate": 0.002856, "loss": 4.5492, "step": 952 }, { "epoch": 0.00953, "grad_norm": 0.7428316161957196, "learning_rate": 0.002859, "loss": 4.4912, "step": 953 }, { "epoch": 0.00954, "grad_norm": 0.6589075049765231, "learning_rate": 0.002862, "loss": 4.5478, "step": 954 }, { "epoch": 0.00955, "grad_norm": 0.6374906159512967, "learning_rate": 0.002865, "loss": 4.5305, "step": 955 }, { "epoch": 0.00956, "grad_norm": 0.7037593758453927, "learning_rate": 0.002868, "loss": 4.5123, "step": 956 }, { "epoch": 0.00957, "grad_norm": 0.7575837270625414, "learning_rate": 0.002871, "loss": 4.5364, "step": 957 }, { "epoch": 0.00958, "grad_norm": 0.8136537306721928, "learning_rate": 0.002874, "loss": 4.5221, "step": 958 }, { "epoch": 0.00959, "grad_norm": 0.7529303499273822, "learning_rate": 0.002877, "loss": 4.5134, "step": 959 }, { "epoch": 0.0096, "grad_norm": 0.7723176518474248, "learning_rate": 0.0028799999999999997, "loss": 4.5128, "step": 960 }, { "epoch": 0.00961, "grad_norm": 0.6826701565045171, "learning_rate": 0.002883, "loss": 4.4802, "step": 961 }, { "epoch": 0.00962, "grad_norm": 0.6022941634910601, "learning_rate": 0.002886, "loss": 4.5086, "step": 962 }, { "epoch": 0.00963, "grad_norm": 0.5423609760386227, "learning_rate": 0.002889, "loss": 4.5025, "step": 963 }, { "epoch": 0.00964, "grad_norm": 0.5430502092899187, "learning_rate": 0.002892, "loss": 4.493, "step": 964 }, { "epoch": 0.00965, "grad_norm": 0.5046734063237629, "learning_rate": 0.002895, "loss": 4.4643, "step": 965 }, { "epoch": 0.00966, "grad_norm": 0.4625107724617308, "learning_rate": 0.002898, "loss": 4.4758, "step": 966 }, { "epoch": 0.00967, "grad_norm": 0.4403845150464526, "learning_rate": 0.002901, "loss": 4.5117, "step": 967 }, { "epoch": 0.00968, "grad_norm": 0.40739936008519134, "learning_rate": 0.002904, "loss": 4.4717, "step": 968 }, { "epoch": 0.00969, "grad_norm": 0.41620271754109195, "learning_rate": 0.002907, "loss": 4.5059, "step": 969 }, { "epoch": 0.0097, "grad_norm": 0.3884080596123629, "learning_rate": 0.00291, "loss": 4.4561, "step": 970 }, { "epoch": 0.00971, "grad_norm": 0.40963591049792675, "learning_rate": 0.002913, "loss": 4.4624, "step": 971 }, { "epoch": 0.00972, "grad_norm": 0.427860611994679, "learning_rate": 0.002916, "loss": 4.4693, "step": 972 }, { "epoch": 0.00973, "grad_norm": 0.6266830722962007, "learning_rate": 0.002919, "loss": 4.4689, "step": 973 }, { "epoch": 0.00974, "grad_norm": 0.7795241673951117, "learning_rate": 0.002922, "loss": 4.4734, "step": 974 }, { "epoch": 0.00975, "grad_norm": 0.7392088002205494, "learning_rate": 0.002925, "loss": 4.4915, "step": 975 }, { "epoch": 0.00976, "grad_norm": 0.6090919745791438, "learning_rate": 0.002928, "loss": 4.4688, "step": 976 }, { "epoch": 0.00977, "grad_norm": 0.6721667952819499, "learning_rate": 0.002931, "loss": 4.5114, "step": 977 }, { "epoch": 0.00978, "grad_norm": 0.7157395676490952, "learning_rate": 0.002934, "loss": 4.4828, "step": 978 }, { "epoch": 0.00979, "grad_norm": 0.6673604586868628, "learning_rate": 0.002937, "loss": 4.4724, "step": 979 }, { "epoch": 0.0098, "grad_norm": 0.6241827666303977, "learning_rate": 0.00294, "loss": 4.4641, "step": 980 }, { "epoch": 0.00981, "grad_norm": 0.6477735028588306, "learning_rate": 0.002943, "loss": 4.479, "step": 981 }, { "epoch": 0.00982, "grad_norm": 0.5903229373466107, "learning_rate": 0.002946, "loss": 4.4717, "step": 982 }, { "epoch": 0.00983, "grad_norm": 0.5367081021491611, "learning_rate": 0.0029490000000000002, "loss": 4.4684, "step": 983 }, { "epoch": 0.00984, "grad_norm": 0.55855952891459, "learning_rate": 0.002952, "loss": 4.455, "step": 984 }, { "epoch": 0.00985, "grad_norm": 0.6566800838357816, "learning_rate": 0.002955, "loss": 4.4789, "step": 985 }, { "epoch": 0.00986, "grad_norm": 0.7602203626787503, "learning_rate": 0.002958, "loss": 4.4461, "step": 986 }, { "epoch": 0.00987, "grad_norm": 0.7515002206199092, "learning_rate": 0.002961, "loss": 4.4656, "step": 987 }, { "epoch": 0.00988, "grad_norm": 0.7510028045825011, "learning_rate": 0.002964, "loss": 4.497, "step": 988 }, { "epoch": 0.00989, "grad_norm": 0.6686524266681442, "learning_rate": 0.002967, "loss": 4.438, "step": 989 }, { "epoch": 0.0099, "grad_norm": 0.7110181433203292, "learning_rate": 0.00297, "loss": 4.4759, "step": 990 }, { "epoch": 0.00991, "grad_norm": 0.9487700940206629, "learning_rate": 0.002973, "loss": 4.4909, "step": 991 }, { "epoch": 0.00992, "grad_norm": 1.0128664486589338, "learning_rate": 0.002976, "loss": 4.4936, "step": 992 }, { "epoch": 0.00993, "grad_norm": 0.776315556008045, "learning_rate": 0.002979, "loss": 4.4927, "step": 993 }, { "epoch": 0.00994, "grad_norm": 0.6954881784010936, "learning_rate": 0.002982, "loss": 4.4894, "step": 994 }, { "epoch": 0.00995, "grad_norm": 0.8343170542469196, "learning_rate": 0.0029850000000000002, "loss": 4.4966, "step": 995 }, { "epoch": 0.00996, "grad_norm": 0.8064256733311401, "learning_rate": 0.002988, "loss": 4.4974, "step": 996 }, { "epoch": 0.00997, "grad_norm": 0.7600837252115415, "learning_rate": 0.002991, "loss": 4.4859, "step": 997 }, { "epoch": 0.00998, "grad_norm": 0.7413851219310601, "learning_rate": 0.002994, "loss": 4.4878, "step": 998 }, { "epoch": 0.00999, "grad_norm": 0.7789142980889687, "learning_rate": 0.002997, "loss": 4.4774, "step": 999 }, { "epoch": 0.01, "grad_norm": 0.5973093879700552, "learning_rate": 0.003, "loss": 4.4903, "step": 1000 } ], "logging_steps": 1, "max_steps": 100000, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 3.9643642855424e+16, "train_batch_size": 1024, "trial_name": null, "trial_params": null }